<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Epoch AI: Gradient Updates]]></title><description><![CDATA[Gradient Updates is a newsletter by Epoch AI, providing shorter-form commentary on important issues around AI.]]></description><link>https://epochai.substack.com/s/gradient-updates</link><image><url>https://substackcdn.com/image/fetch/$s_!ZsOK!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca617831-3128-496f-8aac-33d1fadda48f_176x176.png</url><title>Epoch AI: Gradient Updates</title><link>https://epochai.substack.com/s/gradient-updates</link></image><generator>Substack</generator><lastBuildDate>Sat, 18 Apr 2026 06:13:50 GMT</lastBuildDate><atom:link href="https://epochai.substack.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Epoch Artificial Intelligence, Inc]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[epochai@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[epochai@substack.com]]></itunes:email><itunes:name><![CDATA[Epoch AI]]></itunes:name></itunes:owner><itunes:author><![CDATA[Epoch AI]]></itunes:author><googleplay:owner><![CDATA[epochai@substack.com]]></googleplay:owner><googleplay:email><![CDATA[epochai@substack.com]]></googleplay:email><googleplay:author><![CDATA[Epoch AI]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[What does the war in Iran mean for AI?]]></title><description><![CDATA[A prolonged Hormuz crisis probably won't derail the compute buildout]]></description><link>https://epochai.substack.com/p/what-does-the-war-in-iran-mean-for</link><guid isPermaLink="false">https://epochai.substack.com/p/what-does-the-war-in-iran-mean-for</guid><dc:creator><![CDATA[Josh You]]></dc:creator><pubDate>Fri, 10 Apr 2026 16:40:42 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!ELM6!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6cadb5dd-d71c-4d3b-98e2-b873a74a503d_1362x876.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>This post is part of Epoch AI&#8217;s <a href="https://epoch.ai/gradient-updates">Gradient Updates</a> newsletter, which shares more opinionated or informal takes on big questions in AI progress. These posts solely represent the views of the authors, and do not necessarily reflect the views of Epoch AI as a whole.</em></p><p><em>Originally posted on <a href="https://epoch.ai/gradient-updates/war-in-iran-and-ai/">Epoch AI</a>.</em></p><div><hr></div><p><em>Disclaimer: My background is in the economics of AI and compute. I&#8217;m not an expert in war, diplomacy, or oil and gas, but I&#8217;m familiar with what economic inputs matter for AI. I am also writing about a very dynamic situation. So specific claims about the situation in Iran and Hormuz and its impacts on supply chains should be read as tentative and based on relatively quick research.</em></p><p>Since the US and Israel went to war with Iran at the end of February, shipping through the Strait of Hormuz &#8212; the sole sea route out of the Persian Gulf &#8212; has mostly shut down. This has disrupted around <a href="https://x.com/JavierBlas/status/2033540355216359476?s=20">10%</a> of the world&#8217;s supply of oil, as well as exports of natural gas, helium, urea, and aluminium, and others. Iran has also struck targets in the Gulf states, notably oil and gas facilities and a few <a href="https://www.datacenterdynamics.com/en/news/iran-attack-hits-an-oracle-data-center-in-dubai-causes-limited-damage/">data centers</a>.</p><p>On April 8, the US and Iran agreed to a <a href="https://apnews.com/article/iran-us-israel-trump-lebanon-april-8-2026-38d75d5e4f1c7339a1456fc99415bb2a">two-week ceasefire</a> that would reopen the Strait of Hormuz, though it is not clear how quickly shipping will resume or whether the ceasefire will hold. As of writing on April 9, Hormuz shipping has not <a href="https://www.nbcnews.com/world/iran/strait-hormuz-shipping-traffic-effectively-standstill-iran-ceasefire-rcna267391">meaningfully resumed</a> amid reports of possible ceasefire violations, and a prolonged conflict still seems very plausible to me.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> If the Iran war and Hormuz closure continues, how might this affect AI?</p><p>I will mainly take a compute-centric view to this question. Compute is the main physical input to AI, and other inputs like labor and data are much less likely to be disrupted by this conflict. So I&#8217;ll focus on how this conflict could affect the production and deployment of AI chips and data centers, assuming Hormuz stays closed for many more months. (But note, war is unpredictable, and I won&#8217;t pretend to know what the longer-term <em>n</em>th-order effects of this conflict may be.)</p><p>Several components of the AI compute supply chain might be directly affected by this war: energy shocks could make it harder to power chip fabrication plants (fabs) and data centers, helium shortages could impact fabs, and direct attacks on Gulf states and data center construction could disrupt data center construction and investment flows for AI.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ELM6!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6cadb5dd-d71c-4d3b-98e2-b873a74a503d_1362x876.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ELM6!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6cadb5dd-d71c-4d3b-98e2-b873a74a503d_1362x876.png 424w, https://substackcdn.com/image/fetch/$s_!ELM6!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6cadb5dd-d71c-4d3b-98e2-b873a74a503d_1362x876.png 848w, https://substackcdn.com/image/fetch/$s_!ELM6!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6cadb5dd-d71c-4d3b-98e2-b873a74a503d_1362x876.png 1272w, https://substackcdn.com/image/fetch/$s_!ELM6!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6cadb5dd-d71c-4d3b-98e2-b873a74a503d_1362x876.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ELM6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6cadb5dd-d71c-4d3b-98e2-b873a74a503d_1362x876.png" width="1362" height="876" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6cadb5dd-d71c-4d3b-98e2-b873a74a503d_1362x876.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:876,&quot;width&quot;:1362,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:173339,&quot;alt&quot;:&quot;Table rating five AI supply chain components by risk of disruption from a prolonged Iran War and their overall importance to AI/compute. Fabs in Asia: low risk (energy, helium), critically high importance. US data centers: low risk (energy), high importance. Data centers in Europe and Asia excluding the Gulf: medium risk (energy), medium importance. Gulf state data centers: medium-to-high risk (direct attacks, geopolitical), medium importance. Capital funding from Gulf states: high risk (oil/gas revenue disruption, risk aversion), medium-to-high importance.&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://epochai.substack.com/i/193726890?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6cadb5dd-d71c-4d3b-98e2-b873a74a503d_1362x876.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Table rating five AI supply chain components by risk of disruption from a prolonged Iran War and their overall importance to AI/compute. Fabs in Asia: low risk (energy, helium), critically high importance. US data centers: low risk (energy), high importance. Data centers in Europe and Asia excluding the Gulf: medium risk (energy), medium importance. Gulf state data centers: medium-to-high risk (direct attacks, geopolitical), medium importance. Capital funding from Gulf states: high risk (oil/gas revenue disruption, risk aversion), medium-to-high importance." title="Table rating five AI supply chain components by risk of disruption from a prolonged Iran War and their overall importance to AI/compute. Fabs in Asia: low risk (energy, helium), critically high importance. US data centers: low risk (energy), high importance. Data centers in Europe and Asia excluding the Gulf: medium risk (energy), medium importance. Gulf state data centers: medium-to-high risk (direct attacks, geopolitical), medium importance. Capital funding from Gulf states: high risk (oil/gas revenue disruption, risk aversion), medium-to-high importance." srcset="https://substackcdn.com/image/fetch/$s_!ELM6!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6cadb5dd-d71c-4d3b-98e2-b873a74a503d_1362x876.png 424w, https://substackcdn.com/image/fetch/$s_!ELM6!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6cadb5dd-d71c-4d3b-98e2-b873a74a503d_1362x876.png 848w, https://substackcdn.com/image/fetch/$s_!ELM6!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6cadb5dd-d71c-4d3b-98e2-b873a74a503d_1362x876.png 1272w, https://substackcdn.com/image/fetch/$s_!ELM6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6cadb5dd-d71c-4d3b-98e2-b873a74a503d_1362x876.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>These aren&#8217;t the only exports disrupted by the conflict (e.g. the Persian Gulf is also an important source of <a href="https://www.reuters.com/markets/commodities/iran-war-rattles-global-aluminium-supply-chain-2026-03-19/">aluminum</a>) and there&#8217;s certainly a risk that I missed something important. But the economics of chip fabrication and data centers will generalize to other inputs.</p><p>Overall, I don&#8217;t think a prolonged conflict in Iran will be a <em>huge</em> deal for the AI industry or for the global compute buildout, unless the war escalates dramatically. That said, an extended conflict will be broadly harmful to the global economy, with diffuse second-order effects on AI demand, revenue, or investments.</p><h1>The Iran War&#8217;s impact on energy</h1><p>Oil is not commonly used to generate electricity &#8212; the relevant fuel that normally passes through Hormuz is natural gas, which must be chilled into <em>liquefied natural gas</em> (LNG) to be transported by sea.</p><p>The closure of Hormuz disrupted <a href="https://www.brookings.edu/articles/the-iran-conflicts-energy-shocks-are-not-yet-fully-realized/#:~:text=However%2C%20the%20conflict%20will%20also,Rebuilding%20will%20take%20years">20%</a> of the world&#8217;s LNG supply, causing a ~70% increase in natural gas prices in <a href="https://apnews.com/article/energy-eu-oil-gas-iran-supply-65e520c30d94e7b6184e69d37a7cc09a?utm">Europe</a> and <a href="https://oilprice.com/Latest-Energy-News/World-News/Asia-Burns-More-Coal-as-Middle-East-War-Sends-LNG-Prices-to-3-Year-Highs.html">Asia</a> (live tracker <a href="https://www.reuters.com/markets/quote/JKMc1/">here</a>), which both depend on LNG imports for natural gas.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> This could significantly increase the cost of electricity: Taiwan, Japan, and South Korea all get around 30-40% of their electricity from gas (though this share is only 3% for mainland China), and much of Europe has a high share as well.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!pgS4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0244c40-352c-4e07-9449-f509a99b5ca9_1600x1130.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!pgS4!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0244c40-352c-4e07-9449-f509a99b5ca9_1600x1130.png 424w, https://substackcdn.com/image/fetch/$s_!pgS4!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0244c40-352c-4e07-9449-f509a99b5ca9_1600x1130.png 848w, https://substackcdn.com/image/fetch/$s_!pgS4!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0244c40-352c-4e07-9449-f509a99b5ca9_1600x1130.png 1272w, https://substackcdn.com/image/fetch/$s_!pgS4!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0244c40-352c-4e07-9449-f509a99b5ca9_1600x1130.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!pgS4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0244c40-352c-4e07-9449-f509a99b5ca9_1600x1130.png" width="1456" height="1028" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e0244c40-352c-4e07-9449-f509a99b5ca9_1600x1130.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1028,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!pgS4!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0244c40-352c-4e07-9449-f509a99b5ca9_1600x1130.png 424w, https://substackcdn.com/image/fetch/$s_!pgS4!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0244c40-352c-4e07-9449-f509a99b5ca9_1600x1130.png 848w, https://substackcdn.com/image/fetch/$s_!pgS4!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0244c40-352c-4e07-9449-f509a99b5ca9_1600x1130.png 1272w, https://substackcdn.com/image/fetch/$s_!pgS4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0244c40-352c-4e07-9449-f509a99b5ca9_1600x1130.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p style="text-align: center;"><em>Courtesy of Ember Energy Institute and OWID, who also provide a nice world <a href="https://ourworldindata.org/grapher/share-electricity-gas">map</a></em></p><p>By contrast, the US has been insulated from the LNG shock, with prices <a href="https://x.com/JavierBlas/status/2039415057210057111?s=20">stable</a> since before the war. This is because the US produces more gas than it consumes (it&#8217;s the <a href="https://www.eia.gov/todayinenergy/detail.php?id=64844">largest</a> LNG exporter in the world) <em>and</em> its LNG exports are bottlenecked by the capacity of its exporting <a href="https://www.marketplace.org/story/2026/04/03/why-are-natural-gas-prices-stable-in-the-us">infrastructure</a>. US LNG exports were already at capacity before the war, so in the short run European and Asian buyers cannot bid up US natural gas to the global average.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a> In short, <strong>the Hormuz closure has significantly increased electricity costs in much of Europe and Asia, but not in the US</strong>.</p><p>However, there are two major steps of the compute supply chain that depend on electricity in Asia and Europe: chip fabs and data centers.</p><h3>Energy for fabs</h3><p>Most of the world&#8217;s AI chips are manufactured in East Asia. TSMC in Taiwan handles nearly all leading-edge wafer fabrication and chip packaging, while South Korea hosts two of the three main players in the world&#8217;s memory fab capacity (SK Hynix and Samsung).<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a></p><p>However, despite both regions&#8217; dependence on LNG, I think it is pretty unlikely that a prolonged Hormuz crisis will interrupt AI chip production.</p><p>Fabs consume a lot of energy, but the returns to that energy are enormous. TSMC used <a href="https://esg.tsmc.com/file/public/e-appendix_6.pdf">25,000 GWh</a> of electricity in 2024, around <a href="https://www.datacenterdynamics.com/en/news/tsmc-could-account-for-24-of-taiwans-electricity-consumption-by-2030/#:~:text=Here's%20some%20information%20about%20TSMC's%20electricity%20consumption:,adopts%20the%20energy%2Dintensive%20extreme%20ultraviolet%20lithography%20processes.">8%</a> of Taiwan&#8217;s total electricity consumption. But at the <a href="https://www.taipower.com.tw/2764/2804/2805/59356/normalPost">average price for industrial customers</a> of NT$3.81 (~11 cents USD) per kWh, TSMC&#8217;s energy bill in 2024 would have been just $3 billion, or under 3% of its <a href="https://finance.yahoo.com/news/tsmc-achieves-record-revenue-2024-085914403.html">2024 revenue</a>.</p><p>Making AI chips is <em>very</em> profitable, so there&#8217;s a lot of headroom for power prices to increase. Chipmakers have very good margins (&gt;50% operating margins for both <a href="https://pr.tsmc.com/english/news/3281">TSMC</a> and <a href="https://mis-prod-koce-homepage-cdn-01-blob-ep.azureedge.net/web/attach/116964051831620258.pdf">SK Hynix</a>). And that&#8217;s before the also excellent margins at the chip designer/seller level, with Nvidia&#8217;s gross margins at <a href="https://nvidianews.nvidia.com/news/nvidia-announces-financial-results-for-fourth-quarter-and-fiscal-2026">70-75%</a>. Energy is such a small fraction of the final chip price that TSMC could absorb a doubling of power costs, or pass it to Nvidia, who could easily absorb it too.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a></p><p>So power prices in Asia would have to increase <em>a lot</em> to threaten chip fabs, e.g. by well over ten-fold. This is far greater than the ~70% increase we&#8217;ve seen so far.</p><p>Could gas prices continue to climb? I&#8217;m not qualified to say what the upper bound might be. The research group Oxford Institute for Energy Studies <a href="https://www.oxfordenergy.org/publications/modelling-the-impact-of-the-strait-of-hormuz-closure-on-global-gas-flows-and-prices">estimated</a> that a one-year closure could push gas prices in Europe to $40 per million BTU, almost <a href="https://ycharts.com/indicators/europe_natural_gas_price">quadruple</a> the pre-war price (though still lower than the 2022 peak following Russia&#8217;s invasion of Ukraine, which did not broadly shut down Europe&#8217;s most profitable factories to my knowledge). And Taiwan might be even more dependent on LNG than Europe. In general, demand for fuel or electricity is very <em><a href="https://fredblog.stlouisfed.org/2016/05/energy-demand-and-supply/">inelastic</a></em> in the short run, meaning that supply shocks can cause disproportionate effects on prices.</p><p>There are also policy dynamics at play: Taiwan <a href="https://www.taipower.com.tw/2764/2804/2805/59356/normalPost">subsidizes</a> residential power relative to industrial power, and in a severe crunch Taiwan and South Korea could ration power away from commercial customers to meet basic needs. But I suspect that chip fabs would be the highest-priority industrial customers in both countries, with other businesses rationed first.</p><p>And the US government could intervene in a severe shortage. The government of Taiwan has <a href="https://www.reuters.com/business/energy/taiwan-says-it-has-assurances-over-lng-supplies-major-country-2026-04-04/">said</a> that they have received supply assurances from a &#8220;major&#8221; LNG producing nation to meet its needs. They didn&#8217;t say which producer this was, but in any case AI chips are so important to the US&#8217;s economic and strategic interests that the US government may ensure that Taiwan gets the LNG they need if chip production is threatened.</p><p>Overall, it looks unlikely to me that the energy restrictions will disrupt AI chip production unless the scale of the energy shock dramatically escalates.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a> The energy crunch in Taiwan and Korea would have to be truly catastrophic, not just severe.</p><h3>Energy for data centers</h3><p>AI data centers are disproportionately located in the US, which is insulated from the LNG shock. But a significant share are in Europe and Asia (estimating the precise distribution is out of scope here), where the war could cause serious energy problems for the buildout of AI data centers.</p><p>AI data centers need much more electricity than fabs. The data centers running TSMC&#8217;s chips AI draw significantly more power than TSMC&#8217;s several gigawatts of direct electricity use.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-8" href="#footnote-8" target="_self">8</a></p><p>Still, as with chip fabs, it&#8217;s unlikely that <em>already-built</em> AI data centers in Europe or Asia will shut down. At normal prices, energy costs are a modest share of a data center&#8217;s total cost of ownership or revenue potential.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-9" href="#footnote-9" target="_self">9</a> Once the upfront capital cost to build an AI data center has been incurred, energy prices would have to massively surge to force a shutdown.</p><p>However, <em>planned</em> data centers are a different story. The hyperscaler clouds <a href="https://www.geekwire.com/2025/amazon-web-services-profits-squeezed-as-ai-arms-race-drives-spending-surge/">don&#8217;t</a> <a href="https://www.datacenterdynamics.com/en/news/google-cloud-sees-strong-quarter-revenue-up-28-yoy/">enjoy</a> Nvidia/TSMC-level margins, and -some &#8220;neoclouds&#8221; (smaller, newer AI cloud companies) are reportedly on at least <a href="https://www.reuters.com/commentary/breakingviews/neoclouds-fine-print-is-silver-lining-sorts-2025-10-22/">somewhat shaky</a> financial footing. Even if power is a minority of total costs, a 10&#8211;20% increase in total costs from an energy crunch could be enough to kill some projects.</p><p>It&#8217;s not clear if this would make a big difference in equilibrium. Nvidia could cut prices if enough customers can no longer afford to power its chips. <a href="https://newsletter.semianalysis.com/p/the-great-gpu-shortage-rental-capacity">Rising</a> GPU-hour prices due to demand growth could rescue marginal projects. And the decision to build a data center depends on forecasted energy costs over the coming years; over time Asia and Europe could build up alternative energy sources, or the US could ramp up LNG exports, reducing global prices. Finally, total data center capacity may be fungible by location, with the US eventually absorbing data centers diverted from Europe and Asia.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-10" href="#footnote-10" target="_self">10</a></p><p>Still, an extended Hormuz closure causing a large, longer-term energy shock in much of Asia and Europe would plausibly be at least moderately important in slowing the overall scale of AI compute investments in the medium term.</p><h1>Helium</h1><p>Helium is even more outside my wheelhouse than oil and gas, but it deserves a mention. Helium isn&#8217;t just used for party balloons and blimps. It&#8217;s <a href="https://www.instituteforenergyresearch.org/fossil-fuels/helium-is-instrumental-in-semiconductor-manufacturing/">essential</a> for chip fabrication, serving as an inert gas and a means to cool wafers, with one source <a href="https://j2sourcing.com/blog/helium-crisis-semiconductor-manufacturing-electronic-components-2026/">claiming</a> that 24% of global helium is used for semiconductors. Helium is also necessary for many other uses like MRI machines, cooling superconductors for scientific research, and fiber optics; see this <a href="https://www.construction-physics.com/p/helium-is-hard-to-replace">post</a> from Construction Physics for an overview.</p><p>The Persian Gulf supplies around one-third of the world&#8217;s helium, <a href="https://global.morningstar.com/en-gb/stocks/iran-war-threatens-ai-chip-supply-critical-minerals-risk">primarily from Qatar</a>. These exports were blocked by the Hormuz closure, so in relative terms this disruption is even greater than for oil and LNG.</p><p>However, the loss of Qatar&#8217;s helium supply doesn&#8217;t look likely to interrupt AI chip production. The New York Times spoke to <a href="https://www.nytimes.com/2026/03/27/business/helium-chips-iran-war.html">several experts</a> who believed that chipmakers would be able to outbid other users of helium, though the logistics may be challenging. South Korea claims it has <a href="https://www.reuters.com/business/energy/helium-stocks-south-koreas-chipmakers-last-until-june-sources-say-2026-03-31/">sufficient</a> helium reserves to last through June, with SK Hynix and Samsung working to secure supply from the US.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-11" href="#footnote-11" target="_self">11</a></p><p>I don&#8217;t know enough about helium to be <em>too</em> confident here. But, as with energy, it seems unlikely that the most profitable subsets of the semiconductor industry &#8212; AI chips and the requisite high-bandwidth memory &#8212; will be the ones squeezed out by a 33% helium shock.</p><h1>Gulf data centers and investment flows</h1><p>Over the past month, Iran has been targeting many neighboring Gulf states<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-12" href="#footnote-12" target="_self">12</a> with drone and missile attacks, including direct attacks on (non-AI) <a href="https://www.datacenterdynamics.com/en/news/iran-attack-hits-an-oracle-data-center-in-dubai-causes-limited-damage/">Oracle and AWS</a> data centers in the region. At least in the case of AWS, these attacks were serious enough to cause <a href="https://www.aboutamazon.com/news/aws-bahrain-region-middle-east-conflict">service disruptions</a>. And Iran&#8217;s Islamic Revolutionary Guard Corps has <a href="https://www.datacenterdynamics.com/en/news/iran-threatens-to-attack-openais-stargate-data-center-in-uae-war/">directly threatened</a> to destroy <a href="https://openai.com/index/introducing-stargate-uae/">Stargate UAE</a>, the gigawatt-scale data center that is currently <a href="https://epoch.ai/data/data-centers/satellite-explorer/OpenAIStargateUAE/">under construction</a> on behalf of OpenAI in the United Arab Emirates.</p><p>These attacks may cause disruptions or outright cancellations of AI data centers in the Gulf states. This could be due to direct damage, but more likely through the perception that the Gulf region will face <a href="https://www.reuters.com/world/middle-east/how-dubais-safe-haven-status-is-being-put-test-2026-03-02/">greater geopolitical risks</a> even after this conflict ends.</p><p>This is potentially quite significant. Stargate UAE alone is a large project, initially planned for &gt;1 GW of power capacity with an eventual goal of 5 GW. But I am personally somewhat skeptical that the Gulf was ever going to host more than, say, 10% of the world&#8217;s AI data centers. Gulf states like Saudi Arabia and the UAE are very wealthy and energy-abundant relative to their small populations, but they are fundamentally much smaller in terms of GDP (<a href="https://en.wikipedia.org/wiki/Gulf_Cooperation_Council#Member_states">~2%</a> of the world total), total electricity production, workforce, or geographic area than regions like the US, the rest of Asia, or Europe.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-13" href="#footnote-13" target="_self">13</a></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!h9RY!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c7cf286-bb8f-4578-af34-835ae928f1f7_1600x1130.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!h9RY!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c7cf286-bb8f-4578-af34-835ae928f1f7_1600x1130.png 424w, https://substackcdn.com/image/fetch/$s_!h9RY!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c7cf286-bb8f-4578-af34-835ae928f1f7_1600x1130.png 848w, https://substackcdn.com/image/fetch/$s_!h9RY!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c7cf286-bb8f-4578-af34-835ae928f1f7_1600x1130.png 1272w, https://substackcdn.com/image/fetch/$s_!h9RY!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c7cf286-bb8f-4578-af34-835ae928f1f7_1600x1130.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!h9RY!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c7cf286-bb8f-4578-af34-835ae928f1f7_1600x1130.png" width="1456" height="1028" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3c7cf286-bb8f-4578-af34-835ae928f1f7_1600x1130.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1028,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!h9RY!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c7cf286-bb8f-4578-af34-835ae928f1f7_1600x1130.png 424w, https://substackcdn.com/image/fetch/$s_!h9RY!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c7cf286-bb8f-4578-af34-835ae928f1f7_1600x1130.png 848w, https://substackcdn.com/image/fetch/$s_!h9RY!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c7cf286-bb8f-4578-af34-835ae928f1f7_1600x1130.png 1272w, https://substackcdn.com/image/fetch/$s_!h9RY!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c7cf286-bb8f-4578-af34-835ae928f1f7_1600x1130.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p style="text-align: center;"><em>Courtesy of Ember Energy Institute and <a href="https://ourworldindata.org/grapher/electricity-generation?tab=line&amp;country=USA~OWID_EU27~SAU~ARE~IND~JPN">Our World in Data</a></em></p><p>Indeed, the Gulf states don&#8217;t actually seem to be a great place to build data centers. Amelia Michael, a researcher with GovAI, <a href="https://ameliakmichael.substack.com/p/why-the-uae">observes</a> that the UAE has higher energy costs than the US, along with environmental challenges and a limited local workforce. </p><p>One possible driver of Gulf AI megaprojects is that the Gulf is a very important source of <em>capital funding</em> for AI: companies plan data centers there to cultivate ties with deep-pocketed investors and sovereign wealth funds. These Gulf AI investments have reportedly added up to <a href="https://www.theinformation.com/articles/iran-war-imperils-300-billion-gulf-ai-spending?rc=9mzoog">$300 billion</a> in total.</p><p>This spigot of funding may be interrupted if the Iran War continues, which may be more consequential than the direct impact on data centers. If the Gulf proves unsuitable for data centers, that capital can flow elsewhere &#8212; but only if the capital is still there. Gulf economies are built on oil, and a prolonged export disruption will massively crimp the cash flows that would have flowed into AI.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-14" href="#footnote-14" target="_self">14</a> Oil exports haven&#8217;t been <em>completely</em> cut off: Saudi Arabia and the UAE are still <a href="https://www.reuters.com/business/energy/saudi-crude-exports-yanbu-approach-capacity-hormuz-workaround-2026-03-30/">exporting</a> &gt;5M barrels of oil per day through pipelines bypassing Hormuz, but that&#8217;s a fraction of the <a href="https://x.com/JavierBlas/status/2033540355216359476?s=20">20M barrels/day</a> that formerly flowed through the strait.</p><p>Gulf investors may also become more risk-averse, prioritizing investments they perceive to be safer; though as of March 12, both the UAE and Saudi Arabia have <a href="https://www.reuters.com/world/middle-east/some-gulf-states-reviewing-sovereign-investments-offset-economic-shock-iran-war-2026-03-11/">said</a> they don&#8217;t plan to change their investment strategies.</p><p>In the near term, funding disruptions could affect OpenAI and Anthropic, both reportedly planning <a href="https://www.nytimes.com/2026/01/14/technology/ai-ipo-openai-anthropic-spacex.html">IPOs</a> as soon as this year, with potentially many billions of Gulf funding at stake. One sign that Gulf investments are important to AI is that Anthropic decided to seek Gulf investments despite moral and reputational risk. Anthropic CEO Dario Amodei wrote in a <a href="https://www.wired.com/story/anthropic-dario-amodei-gulf-state-leaked-memo/">leaked memo</a> last year about his decision to tap the &#8220;truly giant amount of capital in the Middle East, easily $100B or more&#8221;, despite his moral reservations about enriching &#8220;dictators&#8221; (and the subsequent leak suggests that this decision was controversial internally).</p><p>But I also don&#8217;t want to overstate the importance of Gulf funding. The Gulf states are a massive source of concentrated capital, with the sovereign wealth funds of the UAE, Saudi Arabia, Kuwait, and Qatar adding up to <a href="https://www.wsj.com/finance/a-guide-to-the-gulfs-trillions-of-dollars-of-sovereign-wealth-c17b505e">over $5 trillion</a>. Because many of these funds invest heavily in tech/venture startups, they are hugely important to the startup industry. But ultimately, the Gulf states aren&#8217;t <em>that</em> rich relative to the entire rest of the world.</p><p>The overall pace of the global compute buildout will be mostly determined by <a href="https://epoch.ai/data-insights/hyperscaler-capex-trend/">hyperscaler</a> capex, financed by the immense cash flows of tech giants like Microsoft, Google, Meta, and Amazon<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-15" href="#footnote-15" target="_self">15</a> and by a huge global <a href="https://www.fortunebusinessinsights.com/corporate-bond-market-113826">bond market</a>. Total AI capital spending may exceed $1 trillion per year soon. Gulf investment flows into AI are on the order of tens of billions annually: significant, but not critical.</p><h1>Takeaways</h1><p>Overall, I think the Iran War won&#8217;t be a huge deal for the AI sector, even if Hormuz stays closed for many more months. Chipmaking is massively profitable, so TSMC, Samsung, and SK Hynix will most likely be able to secure the power, helium, and other resources they need. Some data center plans in Europe or Asia may be cancelled due to rising costs, but in the US data centers will be insulated from the energy shock. I think the most important direct effect may be the disruption of Gulf investment flows, but they aren&#8217;t the biggest source of funding for AI.</p><p>But this all assumes the conflict doesn&#8217;t escalate dramatically. A wider regional war, or spillovers to <a href="https://www.nytimes.com/2026/04/07/world/europe/ukraine-attacks-russian-oil-exports.html">other theaters</a>, or severe second-order macroeconomic effects, could change the picture in ways I can&#8217;t model here. The Iran War (if it continues to be a war) is a dynamic situation, and there&#8217;s a real risk that this post will be outdated shortly after publication. I will keep &#8220;monitoring the situation&#8221; with interest.</p><div><hr></div><p><em>Thanks to Anson Ho, Luke Emberson, Jaime Seville, and Lynette Bye for their feedback on this post</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://epochai.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading! Subscribe to receive the latest from Epoch AI.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>And there will be prolonged <a href="https://www.reuters.com/business/energy/ceraweek-oil-execs-warn-long-term-damage-iran-war-us-downplays-crisis-2026-03-23/">after-effects</a> even if there&#8217;s an immediate, lasting peace deal and reopening of the Strait, due to damage to oil and gas infrastructure caused by the war, and impact of oil well &#8220;shut ins&#8221;: shutting down an oil well can cause damage to the oil. All my claims in this post are conditioned on &#8220;the war and Hormuz closure do not actually end this week&#8221;, but the distinction may not be too critical.</p><p></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>Not all natural gas is liquefied, so the 20% shock to LNG is a smaller share of overall world gas, but LNG is necessary to export gas to an island like Taiwan or a pseudo-island like South Korea.</p><p></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>In general, since gas is just one input into electricity generation into these countries, the percentage increase in gas prices should be an upper bound in the percentage increase in power prices. However, the respective price increases can be similar in a gas shortage: to clear an electricity market, prices need to be equal to the marginal cost of generating power, so power prices will be closely linked to the cost of gas generation (fuel plus other operating expenses).</p><p></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>This is not the case for oil; the price of oil <a href="https://www.cnbc.com/quotes/@CL.1">in the US</a> is similar to <a href="https://tradingeconomics.com/commodity/brent-crude-oil">global prices</a>.</p><p></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>This simplifies a very complex upstream supply chain including semiconductor manufacturing equipment. But TSMC and the memory companies are probably the most energy-intensive part of the chipmaking process.</p><p></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>If Nvidia sells a GPU for $40k at a gross margin of 75%, then it makes a gross profit of $30k and pays its suppliers (mainly TSMC) $10k. If TSMC has an <a href="https://www.investopedia.com/ask/answers/122314/what-difference-between-gross-margin-and-operating-margin.asp">operating margin</a> of 50%, that means it costs them $5k to produce the GPU including marginal production cost and overhead operating costs. I cite operating margin instead of <a href="https://www.investopedia.com/ask/answers/122314/what-difference-between-gross-margin-and-operating-margin.asp">gross margin</a> for TSMC/Hynix to ensure I&#8217;m including all fab-related costs.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>For example, if there are escalating direct attacks on oil and gas infrastructure on both sides&#8212;Iran has already attacked LNG in <a href="https://www.reuters.com/business/energy/iran-attack-damage-wipes-out-17-qatars-lng-capacity-three-five-years-qatarenergy-2026-03-19/">Qatar</a>&#8212;leading to severe long-term damage. There is a <a href="https://www.iea.org/about/oil-security-and-emergency-response/strait-of-hormuz">pipeline to Oman</a> bypassing Hormuz with a capacity of ~20% of Gulf LNG exports before the war, so these attacks could reduce Gulf gas exports even further in the short term. Or there could be spillover to other theaters; Ukraine has <a href="https://www.nytimes.com/2026/04/07/world/europe/ukraine-attacks-russian-oil-exports.html">escalated</a> its attacks on Russian oil production to try to prevent Russia from benefiting from the recent surge in oil prices.</p><p></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-8" href="#footnote-anchor-8" class="footnote-number" contenteditable="false" target="_self">8</a><div class="footnote-content"><p>This can be found by <a href="https://www.wolframalpha.com/input?i=25000+gigawatt-hours+%2F+8760+hours">dividing</a> TSMC&#8217;s consumption of <a href="https://esg.tsmc.com/file/public/e-appendix_6.pdf">25,000 GWh per year</a> by the number of hours in a year.</p><p></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-9" href="#footnote-anchor-9" class="footnote-number" contenteditable="false" target="_self">9</a><div class="footnote-content"><p>One <a href="https://newsletter.semianalysis.com/p/ai-datacenter-energy-dilemma-race">estimate</a> of the energy cost of running 20k Nvidia H100s, using prevailing power prices in the US, is $20.7M per year ($1000 per H100 per year), while an H100 can be rented out for roughly $2/hour, or over $15,000 per year.</p><p></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-10" href="#footnote-anchor-10" class="footnote-number" contenteditable="false" target="_self">10</a><div class="footnote-content"><p>Availability of suitable sites with grid connections and infrastructure may be a constraint as well as raw energy supply, so reducing the geographic scope of data center construction probably will have some effect on the global compute build out.</p><p></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-11" href="#footnote-anchor-11" class="footnote-number" contenteditable="false" target="_self">11</a><div class="footnote-content"><p>The US is the <a href="https://www.construction-physics.com/p/helium-is-hard-to-replace">largest producer</a> of helium outside of the Middle East, and also has around one-third of the world&#8217;s supply.</p><p></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-12" href="#footnote-anchor-12" class="footnote-number" contenteditable="false" target="_self">12</a><div class="footnote-content"><p>&#8220;Gulf states&#8221; is common shorthand for the <a href="https://en.wikipedia.org/wiki/Gulf_Cooperation_Council">six wealthy Arab-majority monarchies</a> that border the Persian Gulf: Saudi Arabia, United Arab Emirates, Bahrain, Qatar, Kuwait and Oman. Iraq and Iran itself are also major oil exporters on the Persian Gulf, but have a rather different vibe.</p><p></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-13" href="#footnote-anchor-13" class="footnote-number" contenteditable="false" target="_self">13</a><div class="footnote-content"><p>To clarify, while I would expect Asia (ex-Gulf and ex-China) to eventually build more AI data centers than the Gulf states even absent this war, the impact of the Iran war on Gulf data centers may be more significant than the energy impact on Asian data centers, so the disruption to Gulf data centers could still be larger.</p><p></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-14" href="#footnote-anchor-14" class="footnote-number" contenteditable="false" target="_self">14</a><div class="footnote-content"><p>My simple model here is that because profits from the Gulf oil and gas industry are reinvested into sovereign wealth funds or privately-held investment funds, they are the main source of outgoing capital flows from the Gulf. But it&#8217;s conceivable that even if oil money stops coming in for a while, Gulf investors that are highly bullish on AI could sell their non-AI investments so they can plow more into AI, though this seems quite risky.</p><p></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-15" href="#footnote-anchor-15" class="footnote-number" contenteditable="false" target="_self">15</a><div class="footnote-content"><p>See chart 12 in <a href="https://www.understandingai.org/p/16-charts-that-explain-the-ai-boom">Understanding AI&#8217;s</a> &#8220;16 charts that explain the AI boom&#8221;</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[Keeping up with the GPTs]]></title><description><![CDATA[Can Chinese and open model companies compete with the frontier through e.g. distillation and talent?]]></description><link>https://epochai.substack.com/p/keeping-up-with-the-gpts</link><guid isPermaLink="false">https://epochai.substack.com/p/keeping-up-with-the-gpts</guid><dc:creator><![CDATA[Anson Ho]]></dc:creator><pubDate>Wed, 08 Apr 2026 04:05:38 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!qvk2!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd1615edc-e04f-4313-a82c-5c02d2fc4016_1026x1283.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Gradient Updates shares more opinionated or informal takes on big questions in AI progress. These posts solely represent the views of the authors, and do not necessarily reflect the views of Epoch AI as a whole.</em></p><div><hr></div><p>If the last decade of AI has taught us one lesson, it&#8217;s that scaling compute builds better models. This sounds great &#8212; until you realize your competitors have ten times more compute than you.</p><p>This is the situation that many Chinese and open model companies find themselves in; relative to frontier companies, they&#8217;re &#8220;compute-poor&#8221;. Just last year, Anthropic spent over ten times more on compute than Minimax and Zhipu AI <em>combined</em>, and the gap is even wider for OpenAI:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!k6Iv!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a11b492-e8b0-4304-83e0-92881c0c72ab_1026x1283.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!k6Iv!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a11b492-e8b0-4304-83e0-92881c0c72ab_1026x1283.png 424w, https://substackcdn.com/image/fetch/$s_!k6Iv!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a11b492-e8b0-4304-83e0-92881c0c72ab_1026x1283.png 848w, https://substackcdn.com/image/fetch/$s_!k6Iv!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a11b492-e8b0-4304-83e0-92881c0c72ab_1026x1283.png 1272w, https://substackcdn.com/image/fetch/$s_!k6Iv!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a11b492-e8b0-4304-83e0-92881c0c72ab_1026x1283.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!k6Iv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a11b492-e8b0-4304-83e0-92881c0c72ab_1026x1283.png" width="1026" height="1283" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4a11b492-e8b0-4304-83e0-92881c0c72ab_1026x1283.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1283,&quot;width&quot;:1026,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!k6Iv!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a11b492-e8b0-4304-83e0-92881c0c72ab_1026x1283.png 424w, https://substackcdn.com/image/fetch/$s_!k6Iv!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a11b492-e8b0-4304-83e0-92881c0c72ab_1026x1283.png 848w, https://substackcdn.com/image/fetch/$s_!k6Iv!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a11b492-e8b0-4304-83e0-92881c0c72ab_1026x1283.png 1272w, https://substackcdn.com/image/fetch/$s_!k6Iv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4a11b492-e8b0-4304-83e0-92881c0c72ab_1026x1283.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Data from Epoch&#8217;s <a href="https://epoch.ai/data/ai-companies">data on AI companies</a> and <a href="https://epoch.ai/data-insights/company-spending-breakdown">Data Insights</a>.</em></figcaption></figure></div><p>You don&#8217;t need to be an AI expert to see that this is a huge handicap. With less compute, it&#8217;s harder to run experiments, train bigger models, and serve many users.</p><p>But compute-poor AI labs have an ace up their sleeve. Even lacking frontier-level compute, they can try to <a href="https://epoch.ai/gradient-updates/the-least-understood-driver-of-ai-progress">use theirs more efficiently</a> to punch well above their weight. That&#8217;s how DeepSeek was on the heels of OpenAI despite using a fraction of the training compute (at least on some benchmarks), driving the stock market bananas.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a></p><p>The big question is: are these efficiency gains big enough for compute-poor labs to <em>really </em>compete, potentially even leapfrogging the compute-rich labs?</p><h2>Breaking down the efficiency gains</h2><p>To figure this out, we need to look at how compute-poor labs can try to boost their compute efficiency. As far as I can tell, there are three main ways to do this:</p><ol><li><p>Develop new algorithms and improve data faster than compute-rich labs</p></li><li><p>Replicate innovations from frontier labs</p></li><li><p>Leverage compute-rich labs&#8217; model capabilities &#8212; e.g. train on synthetic data generated by rival models</p></li></ol><p>Out of these three approaches, the first is the only one that gives them a shot at <em>overtaking</em> the frontier &#8212; the latter two mainly help play catch-up. But catching up could still be important because it makes it easier to leapfrog the frontier later on. So in theory, all three approaches could matter a lot for competition.</p><p>Whether they matter a lot <em>in practice</em> depends on one key condition: whether they improve efficiency <em>asymmetrically</em>, helping compute-poor labs much more than compute-rich ones. After all, if both sides benefit equally from the same trick, a ten-fold compute gap stays a ten-fold compute gap. So let&#8217;s see how well each holds up.</p><h3>Approach 1: Innovate faster than the compute-rich labs</h3><p>If compute-poor companies can&#8217;t outspend the compute-rich frontier, maybe they can out-think them. Such labs often have many brilliant researchers spinning up new algorithmic innovations, like DeepSeek&#8217;s <a href="https://epoch.ai/gradient-updates/how-has-deepseek-improved-the-transformer-architecture">MLA</a> and <a href="https://arxiv.org/abs/2402.03300">GRPO</a> &#8212; things that <a href="https://x.com/dwarkesh_sp/status/1927485816307142737">demonstrate great research taste</a>.</p><p>Being compute-bottlenecked also pushes these labs to be efficient, experimenting with risky algorithmic &#8220;moonshots&#8221; that dramatically save compute. In contrast, frontier labs might tend to stick to tried-and-true recipes &#8212; if their training runs are many times larger, then it&#8217;s also many times more costly for them to take a risk that fails.</p><p>But compute-rich labs have their own efficiency advantages. For example, running experiments often requires lots of compute, like to <a href="https://www.latent.space/p/noam-brown?r=5fx8fq&amp;selection=93bcbca8-4261-49af-9e71-b5bbdfc0ad40&amp;utm_campaign=post-share-selection&amp;utm_medium=web&amp;aspectRatio=instagram&amp;textColor=%23ffffff&amp;triedRedirect=true&amp;_src_ref=epoch.ai#:~:text=Noam%20%5B01%3A04,inspiration%20for%20us.">test if</a> <a href="https://www.youtube.com/watch?v=6nJZopACRuQ&amp;t=1223s">algorithms scale well</a>. That&#8217;s why most compute spending at AI labs <a href="https://epoch.ai/data-insights/openai-compute-spend">goes to</a> <a href="https://epoch.ai/gradient-updates/r-and-d-vs-training-compute">experiments</a>:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!GUnu!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6e1474dc-b986-4cdb-b3bf-ce71828ae07c_1600x1459.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!GUnu!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6e1474dc-b986-4cdb-b3bf-ce71828ae07c_1600x1459.png 424w, https://substackcdn.com/image/fetch/$s_!GUnu!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6e1474dc-b986-4cdb-b3bf-ce71828ae07c_1600x1459.png 848w, https://substackcdn.com/image/fetch/$s_!GUnu!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6e1474dc-b986-4cdb-b3bf-ce71828ae07c_1600x1459.png 1272w, https://substackcdn.com/image/fetch/$s_!GUnu!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6e1474dc-b986-4cdb-b3bf-ce71828ae07c_1600x1459.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!GUnu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6e1474dc-b986-4cdb-b3bf-ce71828ae07c_1600x1459.png" width="1456" height="1328" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6e1474dc-b986-4cdb-b3bf-ce71828ae07c_1600x1459.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1328,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!GUnu!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6e1474dc-b986-4cdb-b3bf-ce71828ae07c_1600x1459.png 424w, https://substackcdn.com/image/fetch/$s_!GUnu!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6e1474dc-b986-4cdb-b3bf-ce71828ae07c_1600x1459.png 848w, https://substackcdn.com/image/fetch/$s_!GUnu!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6e1474dc-b986-4cdb-b3bf-ce71828ae07c_1600x1459.png 1272w, https://substackcdn.com/image/fetch/$s_!GUnu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6e1474dc-b986-4cdb-b3bf-ce71828ae07c_1600x1459.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>And as I explained in <a href="https://epoch.ai/gradient-updates/the-least-understood-driver-of-ai-progress#:~:text=III.%20What%20drives%20software%20progress%3F%20(Or%2C%20why%20all%20the%20estimates%20we%20just%20saw%20are%20misleading)">my last essay</a>, some innovations like the Transformer give bigger efficiency gains as you scale up training compute. So these innovations benefit the compute-rich more than the compute-poor.</p><p>What&#8217;s more, it&#8217;s not just the compute-poor labs that have talent &#8212; the compute-rich labs do as well, and they probably have the best talent. They can pay <a href="https://www.nytimes.com/2025/07/31/technology/ai-researchers-nba-stars.html">exorbitant</a> amounts for the cr&#232;me de la cr&#232;me, often <a href="https://www.wired.com/story/mark-zuckerberg-welcomes-superintelligence-team/">poaching</a> researchers from other labs.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> They also seem to be paying more for junior roles. For instance, early-career researcher salaries at OpenAI and Anthropic are around twice as high as at DeepSeek, even after accounting for <a href="https://wdi.worldbank.org/table/4.16">purchasing power</a>.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> I&#8217;d guess this is part of why the <a href="https://carnegieendowment.org/emissary/2025/12/china-ai-researchers-us-talent-pool">vast majority</a> of Chinese AI researchers who go to US institutions tend to stay there. And <a href="https://jasmi.news/p/china-2025">anecdotally</a>, Chinese AI researchers see the US as the most desirable place to work (at least for now).</p><p>In practice, I think this is why many of the <a href="https://epoch.ai/gradient-updates/the-least-understood-driver-of-ai-progress/#:~:text=Here%20you%20can,making%20further%20innovations.">most</a> <a href="https://epoch.ai/gradient-updates/quantifying-the-algorithmic-improvement-from-reasoning-models/">notable</a> innovations have come from those with the most compute. For example, innovations like modern Transformer-based LLMs, scaling laws, and reasoning models were developed by OpenAI and Google &#8212; organizations that fall squarely within the &#8220;compute-rich&#8221; category.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a></p><p>So I don&#8217;t see why I should expect compute-poor labs to find new software innovations <em>much</em> faster than compute-rich labs &#8212; on the contrary, I think the opposite is more likely.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a></p><h3>Approach 2: Replicate innovations from frontier labs</h3><p>If they can&#8217;t beat the frontier labs at developing <em>new</em> software innovations, why not replicate them? Knowledge (like algorithms) in frontier AI labs could &#8220;spill over&#8221; into compute-poor labs, keeping the latter in the competition without needing to pay big R&amp;D costs to find the innovations in the first place. That would be a big deal, because research costs are the main reason why AI companies <a href="https://epoch.ai/gradient-updates/can-ai-companies-become-profitable">aren&#8217;t yet profitable</a>, and this then makes it easier to leapfrog the frontier with new innovations or by closing the compute gap.</p><p>This could work in a few ways, including AI researchers gossiping, researchers moving from compute-rich to compute-poor labs, or researchers at compute-poor labs reconstructing innovations from frontier labs based on public information.</p><p>This could lead to especially large and fast spillovers if there are &#8220;<a href="https://x.com/sriramk/status/2037185774626509086">four minute mile</a>&#8221; effects &#8212; after one AI lab makes a breakthrough, other labs realise they can do it too, so they pour effort into reimplementation.</p><p>There is evidence this already happens, such as case studies of people broadly reimplementing frontier lab breakthroughs: despite limited public information,<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a> a small team of <a href="https://www.bakerlab.org/">academics</a> got pretty close to reproducing AlphaFold 2 within seven months of its release.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a> Another example is reasoning models: after OpenAI released the first reasoning model (o1), <a href="https://huggingface.co/Qwen/QwQ-32B">several</a> <a href="https://web.archive.org/web/20250128034747/https://ai.google.dev/gemini-api/docs/thinking">different</a> <a href="https://arxiv.org/abs/2501.12948">AI</a> <a href="https://www.anthropic.com/news/claude-3-7-sonnet">labs</a> followed suit within months, sometimes using several times less training compute.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-8" href="#footnote-8" target="_self">8</a></p><p>But are these dynamics big enough to bridge the gap? Probably not, for several reasons.</p><p>For starters, innovation economists seem to think that these kinds of spillover effects are <a href="https://cep.lse.ac.uk/textonly/_new/staff/vanreenen/pdf/ecta9466.pdf">a</a> <a href="https://www.ecb.europa.eu/press/conferences/ecbforum/shared/pdf/2024/EFCB_2024_Dyevre_paper.en.pdf">big</a> <a href="https://www.nber.org/system/files/chapters/c15155/c15155.pdf">deal</a>,<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-9" href="#footnote-9" target="_self">9</a> but I doubt even they would say these kinds of spillovers are strong enough to make up for a ten-fold difference in compute.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-10" href="#footnote-10" target="_self">10</a></p><p>Moreover, even if compute-poor companies <em>instantly</em> knew how to reimplement a closed model&#8217;s innovations upon release, they&#8217;d need to launch a new training run that takes (say) three <a href="https://epoch.ai/data-insights/longest-training-run/">months</a> to finish &#8211; three months where the compute-rich labs are gaining further ground. And this is ignoring the time it takes to figure out what things actually work, as in the case of reasoning models and AlphaFold 2. In practice, companies like MiniMax and Zhipu AI are still spending <a href="https://epoch.ai/gradient-updates/r-and-d-vs-training-compute">much more</a> on experiment compute than on final training runs, meaning spillovers aren&#8217;t enough to cut out those experiment compute costs:<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-11" href="#footnote-11" target="_self">11</a></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ewgY!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc704ccb7-10ef-44f8-999d-10dc941817a7_1026x1283.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ewgY!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc704ccb7-10ef-44f8-999d-10dc941817a7_1026x1283.png 424w, https://substackcdn.com/image/fetch/$s_!ewgY!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc704ccb7-10ef-44f8-999d-10dc941817a7_1026x1283.png 848w, https://substackcdn.com/image/fetch/$s_!ewgY!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc704ccb7-10ef-44f8-999d-10dc941817a7_1026x1283.png 1272w, https://substackcdn.com/image/fetch/$s_!ewgY!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc704ccb7-10ef-44f8-999d-10dc941817a7_1026x1283.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ewgY!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc704ccb7-10ef-44f8-999d-10dc941817a7_1026x1283.png" width="1026" height="1283" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c704ccb7-10ef-44f8-999d-10dc941817a7_1026x1283.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1283,&quot;width&quot;:1026,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ewgY!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc704ccb7-10ef-44f8-999d-10dc941817a7_1026x1283.png 424w, https://substackcdn.com/image/fetch/$s_!ewgY!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc704ccb7-10ef-44f8-999d-10dc941817a7_1026x1283.png 848w, https://substackcdn.com/image/fetch/$s_!ewgY!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc704ccb7-10ef-44f8-999d-10dc941817a7_1026x1283.png 1272w, https://substackcdn.com/image/fetch/$s_!ewgY!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc704ccb7-10ef-44f8-999d-10dc941817a7_1026x1283.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>And finally, the mechanisms for these spillovers don&#8217;t even favor compute-poor labs that much. Just as the compute-poor can copy the compute-rich, the compute-rich can copy the compute-poor, especially if their models are open &#8212; there&#8217;s a reason why <a href="https://www.latent.space/p/noam-brown#:~:text=Noam%20%5B01%3A04,inspiration%20for%20us.">big AI labs still follow the academic literature</a>. And while researchers often <a href="https://7min.ai/exodus/">move</a> from compute-rich to relatively compute-poor labs (e.g. from OpenAI to Mistral AI), they also go the other way. This is especially true for people who move from compute-poor academia to compute-rich industry.</p><p>So once again, I&#8217;m skeptical that &#8220;copying&#8221; innovations boosts compute efficiency that much more for the compute-poor than compute-rich &#8212; at the very least I doubt it&#8217;s enough for a ten-fold gap in compute.</p><h3>Approach 3: Leverage the capabilities of frontier models</h3><p>Compute-rich labs have already developed expensive frontier models, and they&#8217;re letting people use them. So why not use them to improve compute-poor models?</p><p>In practice, the biggest &#8220;oomph&#8221; here probably comes from getting frontier models to generate lots of outputs, and then training on that data. This is often (inaccurately) called &#8220;distillation&#8221;,<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-12" href="#footnote-12" target="_self">12</a> and it lets you transfer the capabilities of frontier models into smaller models. A notable recent example comes from Anthropic, who <a href="https://www.anthropic.com/news/detecting-and-preventing-distillation-attacks">accused</a> DeepSeek, Moonshot, and MiniMax of distilling from Claude&#8217;s outputs.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-13" href="#footnote-13" target="_self">13</a></p><p>The question is, just how much of a difference does distillation make in practice? The evidence is patchy, but if we squint at the numbers I think they weakly suggest that companies can potentially get several-fold compute efficiency gains.</p><p>One piece of evidence is what happens when you distill DeepSeek-R1 into Qwen2.5 models:<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-14" href="#footnote-14" target="_self">14</a></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!lPQl!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f60f08e-5c64-484a-af14-aba5972686e8_1026x1283.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!lPQl!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f60f08e-5c64-484a-af14-aba5972686e8_1026x1283.png 424w, https://substackcdn.com/image/fetch/$s_!lPQl!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f60f08e-5c64-484a-af14-aba5972686e8_1026x1283.png 848w, https://substackcdn.com/image/fetch/$s_!lPQl!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f60f08e-5c64-484a-af14-aba5972686e8_1026x1283.png 1272w, https://substackcdn.com/image/fetch/$s_!lPQl!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f60f08e-5c64-484a-af14-aba5972686e8_1026x1283.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!lPQl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f60f08e-5c64-484a-af14-aba5972686e8_1026x1283.png" width="1026" height="1283" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3f60f08e-5c64-484a-af14-aba5972686e8_1026x1283.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1283,&quot;width&quot;:1026,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!lPQl!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f60f08e-5c64-484a-af14-aba5972686e8_1026x1283.png 424w, https://substackcdn.com/image/fetch/$s_!lPQl!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f60f08e-5c64-484a-af14-aba5972686e8_1026x1283.png 848w, https://substackcdn.com/image/fetch/$s_!lPQl!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f60f08e-5c64-484a-af14-aba5972686e8_1026x1283.png 1272w, https://substackcdn.com/image/fetch/$s_!lPQl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3f60f08e-5c64-484a-af14-aba5972686e8_1026x1283.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>On MATH and GPQA Diamond, we see <em>enormous</em> efficiency gains &#8212; requiring several times less pre-training compute for GPQA, and multiple orders of magnitude less for MATH. So at first glance, distillation seems like it could be a huge deal for crossing the ten-fold compute gap to the frontier.</p><p>The problem is that this anchors too much on specific benchmarks. It&#8217;s suspicious that the efficiency gains vary dramatically between MATH and GPQA. So what if we retain capabilities across a <em>range</em> of benchmarks?<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-15" href="#footnote-15" target="_self">15</a> Once again the evidence is shoddy:</p><ul><li><p><a href="https://arxiv.org/abs/2512.13961">Olmo 3</a> was able to outperform the best open models at its scale across multiple benchmarks, while using 6&#215; fewer tokens. This involved a pipeline that included distillation, so we can interpret this as an upper bound of the efficiency gains in the paper.</p></li><li><p><a href="https://epoch.ai/blog/inference-economics-of-language-models">Based on how quickly they&#8217;re run</a>, it&#8217;s plausible that OpenAI mini models had several times fewer parameters than the original models they were distilled from. At the same time, they arguably perform <a href="https://openai.com/index/introducing-gpt-5-4-mini-and-nano/">only slightly worse</a> on various benchmarks, and are just edged out in head-to-head comparisons. For example, human voters preferred GPT-5 to GPT-5-mini around <a href="https://www.rival.tips/compare/gpt-5/gpt-5-mini">56% of the time.</a> (That said I think this framing is a bit optimistic &#8212; there are <a href="https://epoch.ai/models/gpt-5-mini/">other</a> <a href="https://epoch.ai/models/gpt-5/">benchmarks</a> where the mini models are quite a bit worse.)</p></li></ul><p>It&#8217;s not clear to me how comparable these examples are &#8212; the exact efficiency gains depend on how much data was used, how comprehensive the benchmarks are, and so on. Annoyingly they&#8217;re also based on different metrics &#8212; one helps you use less data, the other fewer parameters, and it&#8217;s not clear exactly how this maps onto <em>compute</em> savings. But I think they weakly suggest that you can potentially get several-fold compute efficiency gains, even if we account for a broad set of benchmarks or domains.</p><p>And you don&#8217;t need much data to make this work: it&#8217;s possible to distill <a href="https://arxiv.org/abs/2501.12948">DeepSeek-R1</a> with around 4 billion tokens.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-16" href="#footnote-16" target="_self">16</a> For comparison, MiniMax may have been able to get 100 billion tokens of data from interactions with Claude.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-17" href="#footnote-17" target="_self">17</a> Nathan Lambert <a href="https://www.interconnects.ai/p/how-much-does-distillation-really#:~:text=Estimates%20will%20vary%2C%20but%20if%20each%20response%20had%2010%2D25K%20tokens%20per%20exchange%2C%20the%20total%20tokens%20across%20these%20two%20labs%2C%20mostly%20with%20MiniMax%2C%20would%20be%20150%2D400%20billion%20tokens">points out</a> that this much data could &#8220;meaningfully improve a models&#8217; post-training&#8221;, which sounds about right to me. So this is probably the most compelling of all the potential explanations we&#8217;ve seen for how compute-poor AI labs could try to keep up with the frontier.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-18" href="#footnote-18" target="_self">18</a></p><p>That said, there are several caveats to bear in mind. First, there are limits to how well distillation works. As the model gets smaller it&#8217;s much harder to distill capabilities into it. We can see this when <a href="https://arxiv.org/pdf/2501.12948#page=60.62">distilling DeepSeek-R1</a> into different versions of <a href="https://arxiv.org/abs/2412.15115">Qwen2.5</a> &#8212; performance on benchmarks like AIME, MATH, and GPQA fall as the model gets smaller:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!_wZj!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fba06eed0-1020-4f45-8cf6-8b472ad967e8_1408x486.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!_wZj!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fba06eed0-1020-4f45-8cf6-8b472ad967e8_1408x486.png 424w, https://substackcdn.com/image/fetch/$s_!_wZj!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fba06eed0-1020-4f45-8cf6-8b472ad967e8_1408x486.png 848w, https://substackcdn.com/image/fetch/$s_!_wZj!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fba06eed0-1020-4f45-8cf6-8b472ad967e8_1408x486.png 1272w, https://substackcdn.com/image/fetch/$s_!_wZj!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fba06eed0-1020-4f45-8cf6-8b472ad967e8_1408x486.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!_wZj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fba06eed0-1020-4f45-8cf6-8b472ad967e8_1408x486.png" width="1408" height="486" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ba06eed0-1020-4f45-8cf6-8b472ad967e8_1408x486.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:486,&quot;width&quot;:1408,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:86554,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://epochai.substack.com/i/193530337?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fba06eed0-1020-4f45-8cf6-8b472ad967e8_1408x486.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!_wZj!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fba06eed0-1020-4f45-8cf6-8b472ad967e8_1408x486.png 424w, https://substackcdn.com/image/fetch/$s_!_wZj!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fba06eed0-1020-4f45-8cf6-8b472ad967e8_1408x486.png 848w, https://substackcdn.com/image/fetch/$s_!_wZj!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fba06eed0-1020-4f45-8cf6-8b472ad967e8_1408x486.png 1272w, https://substackcdn.com/image/fetch/$s_!_wZj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fba06eed0-1020-4f45-8cf6-8b472ad967e8_1408x486.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>To be more concrete, I&#8217;d be pretty surprised if a distilled model could be a hundred times smaller but still capture the same broad set of capabilities as its &#8220;teacher&#8221; model.</p><p>A second caveat is that distilled models may be more likely to be &#8220;<a href="https://www.ikot.blog/the-illusion-of-parity">benchmaxxed</a>&#8221; &#8212; they&#8217;re optimized for a specific set of benchmarks, and don&#8217;t do so well in general. For example, OpenAI&#8217;s &#8220;mini&#8221; models often do worse on things that involve lots of <a href="https://openai.com/index/introducing-simpleqa/">knowledge recall</a>.</p><p>Third, compute-rich frontier labs could keep the most powerful models to themselves or select customers, making it hard for the compute-poor to keep up via distillation.</p><p>Fourth and finally, it&#8217;s hard for distillation to be the full story for how models develop reasoning capabilities on a range of environments. As Nathan Lambert <a href="https://www.interconnects.ai/p/how-much-does-distillation-really">points out</a>, this typically requires large-scale RL where the model learns from its own trial-and-error, rather than purely distilling from a stronger model.So while distillation certainly helps reasoning, compute-poor labs will probably still need to establish <a href="https://epoch.ai/gradient-updates/state-of-rl-envs">RL environment</a> infrastructure and do the RL training itself.</p><p>Overall, leveraging frontier model capabilities probably does meaningfully boost efficiency for compute-poor labs relative to the compute-rich. I&#8217;d weakly guess that it doesn&#8217;t get them all the way to covering a 10&#215; compute gap &#8212; probably it narrows the gap several times.</p><h3>Putting things together</h3><p>Here&#8217;s where I land on the three approaches:</p><ol><li><p><strong>Coming up with new software innovations: </strong>This doesn&#8217;t seem to asymmetrically favor compute-poor AI labs &#8212; in fact it might favor the compute-rich.</p></li><li><p><strong>Replicating innovations from frontier labs: </strong>Spillovers often go in both directions, and in general I&#8217;m skeptical that the effects of this are large enough to make up for a ten-fold difference in compute.</p></li><li><p><strong>Distilling frontier model capabilities: </strong>This could help the compute-poor get very close to broad frontier capabilities with several times less compute. I doubt this is enough to save <em>ten</em> times less compute, but this is a meaningful efficiency boost, and you can get much bigger savings if you only care about specific narrow capabilities.</p></li></ol><p>So if we put everything together, I think compute-poor labs probably can&#8217;t fully make up for their 10&#215; compute disadvantage to compete at the frontier. The compute gap is just too large, and most approaches don&#8217;t help the compute-poor that much more relative to the compute-rich.</p><p>But it&#8217;s not something that we can totally rule out, especially if there are big algorithmic breakthroughs like reasoning models. And of course, I&#8217;m basing my opinion on pretty limited evidence &#8212; I&#8217;d love to hear if people know of good data or arguments for or against my position.</p><h2>What does this mean for the future of AI?</h2><p>So far, I&#8217;ve mostly been talking about &#8220;compute-poor&#8221; AI labs, which have around ten times less compute than those at the frontier. This framing helps us analyze the efficiency gains from things like distillation and replicating innovations, and helps us show that compute is the dominant factor in who &#8220;wins&#8221; in AI competition. But we still need to address a big question: who exactly are these &#8220;compute-poor&#8221; labs?</p><h3>Compute-poor = Chinese AI labs?</h3><p>One type of &#8220;compute-poor&#8221; AI lab, at least today, is Chinese AI companies. For example, we saw earlier that MiniMax and Zhipu AI collectively spent <a href="https://epoch.ai/data-insights/company-spending-breakdown">under ten times</a> the compute of Anthropic in 2025.</p><p>So does this mean that Chinese labs are destined to be relegated to the sidelines, unable to compete with American AI companies? Not necessarily, because it depends on how much compute American and Chinese AI labs can get their hands on. If the compute gap narrows, it becomes easier to compete with the frontier. If it widens, the opposite is true.</p><p>This of course hinges on a huge open question: how will the compute gap evolve over time? So far I think it&#8217;s not grown massively, and I think this is why they&#8217;ve not fallen far behind in capabilities to the frontier, as measured by the <a href="https://epoch.ai/benchmarks/eci">Epoch Capabilities Index</a> (a fancy aggregation of benchmark scores):</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!poEu!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F623fec00-b113-44b7-8d89-1142eb71e439_1026x1283.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!poEu!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F623fec00-b113-44b7-8d89-1142eb71e439_1026x1283.png 424w, https://substackcdn.com/image/fetch/$s_!poEu!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F623fec00-b113-44b7-8d89-1142eb71e439_1026x1283.png 848w, https://substackcdn.com/image/fetch/$s_!poEu!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F623fec00-b113-44b7-8d89-1142eb71e439_1026x1283.png 1272w, https://substackcdn.com/image/fetch/$s_!poEu!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F623fec00-b113-44b7-8d89-1142eb71e439_1026x1283.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!poEu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F623fec00-b113-44b7-8d89-1142eb71e439_1026x1283.png" width="1026" height="1283" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/623fec00-b113-44b7-8d89-1142eb71e439_1026x1283.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1283,&quot;width&quot;:1026,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!poEu!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F623fec00-b113-44b7-8d89-1142eb71e439_1026x1283.png 424w, https://substackcdn.com/image/fetch/$s_!poEu!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F623fec00-b113-44b7-8d89-1142eb71e439_1026x1283.png 848w, https://substackcdn.com/image/fetch/$s_!poEu!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F623fec00-b113-44b7-8d89-1142eb71e439_1026x1283.png 1272w, https://substackcdn.com/image/fetch/$s_!poEu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F623fec00-b113-44b7-8d89-1142eb71e439_1026x1283.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>But this interpretation is a little debatable. For example, the gap in training compute has grown, though not <em>dramatically</em> so:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!NgxV!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5d08502-4b73-4eab-b7cc-1a1f390c10ae_1026x1283.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!NgxV!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5d08502-4b73-4eab-b7cc-1a1f390c10ae_1026x1283.png 424w, https://substackcdn.com/image/fetch/$s_!NgxV!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5d08502-4b73-4eab-b7cc-1a1f390c10ae_1026x1283.png 848w, https://substackcdn.com/image/fetch/$s_!NgxV!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5d08502-4b73-4eab-b7cc-1a1f390c10ae_1026x1283.png 1272w, https://substackcdn.com/image/fetch/$s_!NgxV!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5d08502-4b73-4eab-b7cc-1a1f390c10ae_1026x1283.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!NgxV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5d08502-4b73-4eab-b7cc-1a1f390c10ae_1026x1283.png" width="1026" height="1283" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e5d08502-4b73-4eab-b7cc-1a1f390c10ae_1026x1283.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1283,&quot;width&quot;:1026,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!NgxV!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5d08502-4b73-4eab-b7cc-1a1f390c10ae_1026x1283.png 424w, https://substackcdn.com/image/fetch/$s_!NgxV!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5d08502-4b73-4eab-b7cc-1a1f390c10ae_1026x1283.png 848w, https://substackcdn.com/image/fetch/$s_!NgxV!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5d08502-4b73-4eab-b7cc-1a1f390c10ae_1026x1283.png 1272w, https://substackcdn.com/image/fetch/$s_!NgxV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5d08502-4b73-4eab-b7cc-1a1f390c10ae_1026x1283.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>But it&#8217;s rather confusing for a couple of reasons:</p><ul><li><p>The best US and Chinese models aren&#8217;t always trained with the most compute. For example, Grok 3 and 4 are probably also <a href="https://epoch.ai/gradient-updates/ai-progress-is-about-to-speed-up/#:~:text=I%20think%20the%20correct,training%20compute%20than%20them.">inefficient</a> for their size, and I&#8217;d guess the same is true for Doubao Pro.</p></li><li><p>This is looking at <em>training</em> compute rather than overall compute budgets. I&#8217;ve mostly been talking about the latter throughout this essay, but sadly we don&#8217;t have enough data on it.</p></li><li><p>Another confusing thing is that different AI labs were scrambling to figure out reasoning, which gives a <a href="https://epoch.ai/gradient-updates/quantifying-the-algorithmic-improvement-from-reasoning-models/">much larger efficiency boost</a> <a href="https://epoch.ai/gradient-updates/the-least-understood-driver-of-ai-progress/#:~:text=Here%20you%20can,making%20further%20innovations.">compared to most other innovations</a>.</p></li></ul><p>Overall I&#8217;d weakly guess that the overall compute gap hasn&#8217;t changed much over time, and this is why the capabilities gap hasn&#8217;t changed much either.</p><p>But there are some reasons to think that the gap will grow. As I&#8217;m writing this, American hyperscalers are driving a data center buildout that&#8217;s larger than the Manhattan Project and Apollo Program at their peaks. As part of this, frontier labs will soon have data centers that <a href="https://epoch.ai/data/data-centers">each cost tens of billions of dollars</a> and contain millions of GPUs. In contrast, Chinese AI companies are hampered by export controls, and being <a href="https://epoch.ai/gradient-updates/why-china-isnt-about-to-leap-ahead-of-the-west-on-compute">behind</a> in domestic AI chip production. This could widen the compute gap in the near term, and hence lead to a divergence in the capabilities of US and Chinese AI models. On some <a href="https://x.com/htihle/status/2020854971860746703/photo/1">measures</a> this is already happening, and Tang Jie (CEO of Zhipu AI) even recently <a href="https://www.wsj.com/tech/ai/china-ai-race-us-chips-9e74b957?mod=tech_lead_pos5">said</a> the following:</p><blockquote><p>&#8220;The truth may be that the gap [between US and Chinese AI] is actually widening.&#8221;</p></blockquote><p>On the other hand, Chinese AI labs could continue to <a href="https://www.the-substrate.net/p/where-will-china-get-its-compute">get more compute</a> by smuggling chips, or by renting it through the cloud. And the further we look into the future, the greater the odds of China catching up in domestic AI chip production.</p><p>So in my book, the overall picture of US-China AI competition looks like this: in the short run (say 2-3 years), I think it&#8217;s unlikely that the gap will narrow much, because of export controls and challenges in rapidly scaling up domestic chip production. But it&#8217;s more uncertain in the longer run. It depends on hard-to-predict chip export controls. It depends on how hard it is to smuggle and rent chips. It depends on how much China wants to <em>really</em> push for AGI, scaling up their compute stocks &#8212; and there are <a href="https://www.chinatalk.media/p/how-china-hopes-to-build-agi-through">signs</a> that the winds are shifting in this direction from China&#8217;s <a href="http://news.cn/politics/20260313/085af5de5a4b4268aa7d87d90817df2f/c.html">most recent Five-Year Plan</a>.</p><p>And there&#8217;s an additional wildcard that makes things even more uncertain &#8212; frontier AI companies can run more of the best AIs to speed up their own AI research, relative to their competitors. Right now these gains <a href="https://metr.org/blog/2026-02-24-uplift-update/">are</a> <em><a href="https://aleximas.substack.com/p/what-is-the-impact-of-ai-on-productivity">maybe</a></em><a href="https://aleximas.substack.com/p/what-is-the-impact-of-ai-on-productivity"> noticeable</a> but not game-changing, but that&#8217;ll probably change in the next few years. If so, this could widen the capability gap between the compute-rich and compute-poor, even if the compute gap stays fixed.</p><h3>Compute-poor = Open models?</h3><p>Another type of &#8220;compute-poor&#8221; AI lab is the labs developing open models. Like Chinese AI companies, their training compute hasn&#8217;t fallen <em>super</em> far behind the frontier, though the gap has grown&#8230;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!2AL1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02acb150-ae0e-43eb-b4bb-ab1a240e2150_1026x1283.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!2AL1!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02acb150-ae0e-43eb-b4bb-ab1a240e2150_1026x1283.png 424w, https://substackcdn.com/image/fetch/$s_!2AL1!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02acb150-ae0e-43eb-b4bb-ab1a240e2150_1026x1283.png 848w, https://substackcdn.com/image/fetch/$s_!2AL1!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02acb150-ae0e-43eb-b4bb-ab1a240e2150_1026x1283.png 1272w, https://substackcdn.com/image/fetch/$s_!2AL1!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02acb150-ae0e-43eb-b4bb-ab1a240e2150_1026x1283.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!2AL1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02acb150-ae0e-43eb-b4bb-ab1a240e2150_1026x1283.png" width="1026" height="1283" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/02acb150-ae0e-43eb-b4bb-ab1a240e2150_1026x1283.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1283,&quot;width&quot;:1026,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!2AL1!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02acb150-ae0e-43eb-b4bb-ab1a240e2150_1026x1283.png 424w, https://substackcdn.com/image/fetch/$s_!2AL1!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02acb150-ae0e-43eb-b4bb-ab1a240e2150_1026x1283.png 848w, https://substackcdn.com/image/fetch/$s_!2AL1!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02acb150-ae0e-43eb-b4bb-ab1a240e2150_1026x1283.png 1272w, https://substackcdn.com/image/fetch/$s_!2AL1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02acb150-ae0e-43eb-b4bb-ab1a240e2150_1026x1283.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>&#8230;and the capabilities gap looks like a constant-ish lag:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!qvk2!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd1615edc-e04f-4313-a82c-5c02d2fc4016_1026x1283.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!qvk2!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd1615edc-e04f-4313-a82c-5c02d2fc4016_1026x1283.png 424w, https://substackcdn.com/image/fetch/$s_!qvk2!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd1615edc-e04f-4313-a82c-5c02d2fc4016_1026x1283.png 848w, https://substackcdn.com/image/fetch/$s_!qvk2!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd1615edc-e04f-4313-a82c-5c02d2fc4016_1026x1283.png 1272w, https://substackcdn.com/image/fetch/$s_!qvk2!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd1615edc-e04f-4313-a82c-5c02d2fc4016_1026x1283.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!qvk2!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd1615edc-e04f-4313-a82c-5c02d2fc4016_1026x1283.png" width="1026" height="1283" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d1615edc-e04f-4313-a82c-5c02d2fc4016_1026x1283.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1283,&quot;width&quot;:1026,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!qvk2!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd1615edc-e04f-4313-a82c-5c02d2fc4016_1026x1283.png 424w, https://substackcdn.com/image/fetch/$s_!qvk2!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd1615edc-e04f-4313-a82c-5c02d2fc4016_1026x1283.png 848w, https://substackcdn.com/image/fetch/$s_!qvk2!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd1615edc-e04f-4313-a82c-5c02d2fc4016_1026x1283.png 1272w, https://substackcdn.com/image/fetch/$s_!qvk2!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd1615edc-e04f-4313-a82c-5c02d2fc4016_1026x1283.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Perhaps this isn&#8217;t so surprising. There&#8217;s quite a lot of overlap between open model labs and Chinese AI companies, so we should expect a lot of their challenges to be the same. For example, a big open question is how the compute gap between open model labs and frontier companies will change over time, just like what we saw with Chinese labs.</p><p>But the future of open models at the frontier depends on a separate open question (no pun intended): will future AI models be open-weight?</p><p>This is probably even harder to answer than the question about compute scaling. One issue is that companies seem to flip-flop on whether to release open models. For example, Alibaba has recently shown <a href="https://www.reuters.com/commentary/breakingviews/china-ai-labs-face-growing-open-source-dilemma-2026-03-25/">signs of</a> <a href="https://recodechinaai.substack.com/p/alibabas-qwen-lead-just-stepped-down">tilting</a> further toward closed releases.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-19" href="#footnote-19" target="_self">19</a></p><p>But my favourite example of this has got to be Meta &#8212; in July 2024, Mark Zuckerberg wrote an essay called &#8220;<a href="https://about.fb.com/news/2024/07/open-source-ai-is-the-path-forward/">Open Source AI is the Path Forward</a>&#8221;. He raised multiple arguments for open models, like privacy, safety through lots of scrutiny of AI models, and leveraging the broad AI community to improve Meta&#8217;s models. Now they seem to be backing away from this strategy &#8212; first with <a href="https://www.nytimes.com/2025/06/10/technology/meta-new-ai-lab-superintelligence.html">reported</a> discussion about abandoning the open-weight Llama 4 Behemoth in favor of a closed model, and now with their next-generation &#8220;Avocado&#8221; model <a href="https://www.bloomberg.com/news/articles/2025-12-10/inside-meta-s-pivot-from-open-source-to-money-making-ai-model">allegedly</a> to be released closed-weights. Someone has even written an article called &#8220;<a href="https://spyglass.org/open-source-ai-was-the-path-forward/">Open Source AI </a><em><a href="https://spyglass.org/open-source-ai-was-the-path-forward/">was</a></em><a href="https://spyglass.org/open-source-ai-was-the-path-forward/"> the Path Forward</a>&#8221;. (Last-minute pre-publication edit: apparently Meta now <a href="https://www.axios.com/2026/04/06/meta-open-source-ai-models">plans</a> to &#8220;eventually&#8221; offer open versions of their next generation of AI models, so maybe Open Source AI is <em>eventually</em> the Path Forward).</p><p>Sometimes labs also go the other way too. After DeepSeek-R1&#8217;s release as an open-weights model, several Chinese labs followed suit. For example, Baidu abandoned their <a href="https://www.yicaiglobal.com/news/future-of-llms-is-closed-source-baidu-ceo-says">commitments to closed models</a> with their ERNIE 4.5 model, <a href="https://www.fool.com/earnings/call-transcripts/2025/02/18/baidu-bidu-q4-2024-earnings-call-transcript/#:~:text=So%2C%20we%20have,stack%20AI%20capabilities.">citing DeepSeek as a motivator</a> driving this decision.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-20" href="#footnote-20" target="_self">20</a></p><p>The upshot is that the capability gap between open models and the frontier depends a lot on which labs choose to be &#8220;open&#8221;. This depends on things like <a href="https://www.chinatalk.media/i/172613241/open-source-as-strategic-imperative">government strategies</a>, the ideologies of company CEOs, and also business incentives, as Nathan Lambert <a href="https://www.interconnects.ai/p/the-next-phase-of-open-models">points</a> <a href="https://x.com/natolambert/status/2029049751472357631">out</a>:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!cNhU!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ff1aa56-8bdf-48c6-9c68-6dc450f66831_1190x772.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!cNhU!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ff1aa56-8bdf-48c6-9c68-6dc450f66831_1190x772.png 424w, https://substackcdn.com/image/fetch/$s_!cNhU!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ff1aa56-8bdf-48c6-9c68-6dc450f66831_1190x772.png 848w, https://substackcdn.com/image/fetch/$s_!cNhU!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ff1aa56-8bdf-48c6-9c68-6dc450f66831_1190x772.png 1272w, https://substackcdn.com/image/fetch/$s_!cNhU!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ff1aa56-8bdf-48c6-9c68-6dc450f66831_1190x772.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!cNhU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ff1aa56-8bdf-48c6-9c68-6dc450f66831_1190x772.png" width="1190" height="772" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2ff1aa56-8bdf-48c6-9c68-6dc450f66831_1190x772.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:772,&quot;width&quot;:1190,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!cNhU!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ff1aa56-8bdf-48c6-9c68-6dc450f66831_1190x772.png 424w, https://substackcdn.com/image/fetch/$s_!cNhU!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ff1aa56-8bdf-48c6-9c68-6dc450f66831_1190x772.png 848w, https://substackcdn.com/image/fetch/$s_!cNhU!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ff1aa56-8bdf-48c6-9c68-6dc450f66831_1190x772.png 1272w, https://substackcdn.com/image/fetch/$s_!cNhU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ff1aa56-8bdf-48c6-9c68-6dc450f66831_1190x772.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3>The bottom line</h3><p>So what have we learned from all of this discussion? For me the primary takeaway is this: compute is the biggest factor for which companies can compete at the capabilities frontier &#8212; efficiency matters too, but it&#8217;s probably not enough to make up for ten times less compute.</p><p>At first glance this seems to paint a gloomy picture for Chinese and open model labs, but I think it&#8217;s not so clear cut. This is because the label of &#8220;compute-poor&#8221; may not map onto them very cleanly in the future. Maybe it&#8217;s hard for them to compete with the &#8220;compute-rich&#8221; frontier labs <em>today</em>, but in the future this&#8217;ll depend on how much compute they have, and which labs choose to be &#8220;open&#8221; or &#8220;closed&#8221;.</p><p>Moreover, I&#8217;ve been framing the whole analysis around competition at the frontier of capabilities, but strictly speaking that might not even be the only priority of these AI labs! For example, we could imagine Chinese models simply finding a different part of the &#8220;capabilities vs cheapness&#8221; pareto frontier:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!QhHz!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F288686f9-a909-4347-9c22-15f3a7ea5f51_1026x1283.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!QhHz!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F288686f9-a909-4347-9c22-15f3a7ea5f51_1026x1283.png 424w, https://substackcdn.com/image/fetch/$s_!QhHz!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F288686f9-a909-4347-9c22-15f3a7ea5f51_1026x1283.png 848w, https://substackcdn.com/image/fetch/$s_!QhHz!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F288686f9-a909-4347-9c22-15f3a7ea5f51_1026x1283.png 1272w, https://substackcdn.com/image/fetch/$s_!QhHz!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F288686f9-a909-4347-9c22-15f3a7ea5f51_1026x1283.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!QhHz!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F288686f9-a909-4347-9c22-15f3a7ea5f51_1026x1283.png" width="1026" height="1283" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/288686f9-a909-4347-9c22-15f3a7ea5f51_1026x1283.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1283,&quot;width&quot;:1026,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!QhHz!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F288686f9-a909-4347-9c22-15f3a7ea5f51_1026x1283.png 424w, https://substackcdn.com/image/fetch/$s_!QhHz!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F288686f9-a909-4347-9c22-15f3a7ea5f51_1026x1283.png 848w, https://substackcdn.com/image/fetch/$s_!QhHz!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F288686f9-a909-4347-9c22-15f3a7ea5f51_1026x1283.png 1272w, https://substackcdn.com/image/fetch/$s_!QhHz!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F288686f9-a909-4347-9c22-15f3a7ea5f51_1026x1283.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In fact, if you look at token usage on <a href="https://openrouter.ai/rankings?view=month">OpenRouter</a>, you&#8217;ll notice that a bunch of the most popular language models are actually Chinese (though this is kinda misleading &#8212; for example, most people who use Claude probably don&#8217;t use it through OpenRouter):</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!i2t0!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd0295cc-d39f-4242-9819-3283b08f872b_1600x734.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!i2t0!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd0295cc-d39f-4242-9819-3283b08f872b_1600x734.png 424w, https://substackcdn.com/image/fetch/$s_!i2t0!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd0295cc-d39f-4242-9819-3283b08f872b_1600x734.png 848w, https://substackcdn.com/image/fetch/$s_!i2t0!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd0295cc-d39f-4242-9819-3283b08f872b_1600x734.png 1272w, https://substackcdn.com/image/fetch/$s_!i2t0!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd0295cc-d39f-4242-9819-3283b08f872b_1600x734.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!i2t0!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd0295cc-d39f-4242-9819-3283b08f872b_1600x734.png" width="1456" height="668" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/dd0295cc-d39f-4242-9819-3283b08f872b_1600x734.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:668,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!i2t0!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd0295cc-d39f-4242-9819-3283b08f872b_1600x734.png 424w, https://substackcdn.com/image/fetch/$s_!i2t0!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd0295cc-d39f-4242-9819-3283b08f872b_1600x734.png 848w, https://substackcdn.com/image/fetch/$s_!i2t0!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd0295cc-d39f-4242-9819-3283b08f872b_1600x734.png 1272w, https://substackcdn.com/image/fetch/$s_!i2t0!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd0295cc-d39f-4242-9819-3283b08f872b_1600x734.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>I happen to think that competing along the dimension of raw capabilities matters a ton as the world gets closer and closer to building AGI. But it&#8217;s not the only thing that matters, and it&#8217;s very possible that different AI labs might not even be &#8220;racing&#8221; toward the same thing at all.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://epochai.substack.com/p/keeping-up-with-the-gpts?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://epochai.substack.com/p/keeping-up-with-the-gpts?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><div><hr></div><p><em>I&#8217;d like to thank JS Denain, Jaime Sevilla, David Owen, Ben Cottier, Luke Emberson, Lynette Bye, and Stefania Guerra for their feedback and support on this post.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://epochai.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading! Subscribe to receive the latest from Epoch AI.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>Notably, my colleagues at Epoch <a href="https://epoch.ai/data/ai-models">estimate</a> that DeepSeek-R1 was trained with 4 &#215; 10<sup>24</sup> FLOP, compared to a speculative guesstimate of 4 &#215; 10<sup>25</sup> FLOP for o1, assuming the base model is GPT-4o and post-training compute costs are relatively small. That said, it&#8217;s worth saying that many claims about how DeepSeek-R1 was super cheap are <a href="https://newsletter.semianalysis.com/p/deepseek-debates">overstated</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>Though I suppose we still need to see what happens with Meta Superintelligence Labs after their big hiring spree.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>For example, junior research scientists at OpenAI and Anthropic tend to earn in the ballpark of <a href="https://web.archive.org/web/20260326192700/https://openai.com/careers/research-scientist-research-engineer-early-career-cohort-san-francisco/">$350,000</a> <a href="https://web.archive.org/web/20260326193447/https://job-boards.greenhouse.io/anthropic/jobs/5076616008">per year</a>. In contrast, junior researchers at DeepSeek earn around <a href="https://www.zhipin.com/zhaopin/e7b8a2f92a29adbf1nN72tW-Ew~~/">&#165;50,000-100,000 per month</a>, which in US dollars works out to $100,000-200,000 per year (note that the annual compensation is for 14 months rather than 12, because there are two extra salary-equivalent payments, e.g. at year-end or from bonuses).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>Perhaps there&#8217;s a counterargument that you didn&#8217;t necessarily <em>need</em> a lot of compute to come up with these innovations in the first place. But even if that&#8217;s true, it&#8217;s still telling that the innovations were developed by compute-rich AI labs in actuality.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>But even if I&#8217;m wrong about this, a ten-fold gap in compute is <em>a lot</em>, and it&#8217;s rare to find original software innovations that can make up for it. One exception <em>might</em> be <a href="https://epoch.ai/gradient-updates/quantifying-the-algorithmic-improvement-from-reasoning-models">reasoning models</a>, but notably these were invented by OpenAI &#8212; hardly a compute-poor AI lab!</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>As I understand it, the main thing that was <a href="https://fabianfuchsml.github.io/alphafold2/">known</a> was some high-level details about the model architecture. RoseTTAFold was developed knowing that AlphaFold 2 used end-to-end prediction from MSAs with attention and an equivariant structure module, but not the specific architecture, loss functions, or training innovations, which were only revealed when the AlphaFold2 paper was published on the same day as RoseTTAFold (July 15, 2021).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>I say &#8220;pretty close&#8221; because their attempt (<a href="https://www.science.org/doi/10.1126/science.abj8754">RoseTTAFold</a>) doesn&#8217;t quite match AlphaFold 2&#8217;s performance &#8212; if we fish out the numbers using <a href="https://automeris.io/">WebPlotDigitizer</a>, we see that RoseTTAFold scores 80.3% compared to AlphaFold 2&#8217;s 90.2% on the &#8220;Average TM-score&#8221; metric in the CASP14 competition. On the other hand, RoseTTAFold may have used several times less compute &#8212; according to the paper&#8217;s <a href="https://www.biorxiv.org/content/10.1101/2022.12.09.519842v1.supplementary-material?versioned=true">supplementary material</a>, it was trained for 4 weeks on 64 NVIDIA V100 GPUs. <a href="https://epoch.ai/blog/estimating-training-compute">Assuming</a> 16-bit precision and a utilization of 30%, this works out to 1.4e21 FLOP. In contrast, my colleagues <a href="https://epoch.ai/data/ai-models">estimate</a> that AlphaFold 2 needed 3e21 FLOP, which is around twice that. I don&#8217;t know how the performance of AlphaFold 2 and RosettaFold scale with more training compute, but I&#8217;d guess that AlphaFold 2 was more compute-efficient overall.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-8" href="#footnote-anchor-8" class="footnote-number" contenteditable="false" target="_self">8</a><div class="footnote-content"><p>A caveat: I&#8217;ve heard through word of mouth that Anthropic was actually able to reimplement reasoning models within about two weeks of o1&#8217;s announcement, which doesn&#8217;t sound crazy to me. I&#8217;m not sure what this means in terms of the time lag between o1 and other reasoning models though &#8212; probably OpenAI had &#8220;finished&#8221; developing o1 some time before they announced it.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-9" href="#footnote-anchor-9" class="footnote-number" contenteditable="false" target="_self">9</a><div class="footnote-content"><p>Note that I&#8217;m referring to spillovers between firms in the same industry (&#8220;technological spillovers&#8221;) rather than spillovers from firms in the same region (&#8220;<a href="https://www.nber.org/system/files/working_papers/w3993/w3993.pdf">geographical spillovers</a>&#8221;). Both matter, and in fact the latter probably differentially benefits the cluster of compute-rich AI labs in the Bay Area relative to other AI companies.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-10" href="#footnote-anchor-10" class="footnote-number" contenteditable="false" target="_self">10</a><div class="footnote-content"><p>It&#8217;s not directly comparable to the metrics we&#8217;re looking at here, but we can try to get a sense of the effect sizes.  According to one <a href="https://cep.lse.ac.uk/textonly/_new/staff/vanreenen/pdf/ecta9466.pdf">paper</a>, if other firms working on similar tech increase their total R&amp;D by 10%, your own firm&#8217;s market value also grows by 2-3%. That doesn&#8217;t sound like the kind of thing that means you can compete with ten-fold less compute!</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-11" href="#footnote-anchor-11" class="footnote-number" contenteditable="false" target="_self">11</a><div class="footnote-content"><p>If it did, it would also disincentivize frontier labs from doing R&amp;D <a href="https://www.nber.org/system/files/chapters/c2144/c2144.pdf">a la Kenneth Arrow</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-12" href="#footnote-anchor-12" class="footnote-number" contenteditable="false" target="_self">12</a><div class="footnote-content"><p>Strictly speaking this is an abuse of terminology &#8212; distillation <a href="https://arxiv.org/abs/1503.02531">as originally defined</a> means training a smaller model on a &#8220;teacher&#8221; model&#8217;s full probability distribution over all possible outputs, not just its final generated outputs.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-13" href="#footnote-anchor-13" class="footnote-number" contenteditable="false" target="_self">13</a><div class="footnote-content"><p>Another piece of evidence is &#8220;<a href="https://www.dbreunig.com/2025/05/30/using-slop-forensics-to-determine-model-ancestry.html">slop forensics</a>&#8221; to figure out which models were related to one another (and hence which models inherited capabilities from frontier models).</p><p>This works by looking at which words or phrases show up disproportionately frequently in some models, to pick up on its &#8220;signature&#8221;. Then it compares these signatures to see the relations between models, which gives something like an evolutionary tree. If you stare at the outputs, you could <a href="https://x.com/sam_paech/status/1928187246689112197">conclude</a> that DeepSeek switched from training on synthetic data from OpenAI models, to synthetic outputs from Gemini. I&#8217;m not sure how much to update on this kind of evidence though. For example, I&#8217;ve also heard <a href="https://x.com/teortaxestex/status/2026130112685416881">suggestions</a> that Anthropic&#8217;s models might be trained on data from DeepSeek&#8217;s models! So what&#8217;s the overall picture?</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-14" href="#footnote-anchor-14" class="footnote-number" contenteditable="false" target="_self">14</a><div class="footnote-content"><p>Note that this is using the more accurate definition of &#8220;distillation&#8221;, with access to probability distributions over tokens, rather than just final output tokens.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-15" href="#footnote-anchor-15" class="footnote-number" contenteditable="false" target="_self">15</a><div class="footnote-content"><p>Here the constraint is to maintain around 95% of original benchmark performance, and broad capabilities across a range of different domains. You can probably do a lot better than that if you only keep capabilities on a particular set of benchmarks.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-16" href="#footnote-anchor-16" class="footnote-number" contenteditable="false" target="_self">16</a><div class="footnote-content"><p>The <a href="https://arxiv.org/abs/2501.12948">paper</a> describes using 800,000 examples with around 5,000 tokens each, which works out to around 4 billion tokens.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-17" href="#footnote-anchor-17" class="footnote-number" contenteditable="false" target="_self">17</a><div class="footnote-content"><p>MiniMax <a href="https://www.anthropic.com/news/detecting-and-preventing-distillation-attacks">allegedly</a> had at least 10 million exchanges with Claude to gather its data. If each &#8220;exchange&#8221; had around <a href="https://www.interconnects.ai/p/how-much-does-distillation-really#:~:text=Estimates%20will%20vary%2C%20but%20if%20each%20response%20had%2010%2D25K%20tokens%20per%20exchange%2C%20the%20total%20tokens%20across%20these%20two%20labs%2C%20mostly%20with%20MiniMax%2C%20would%20be%20150%2D400%20billion%20tokens">10,000 tokens</a>, as Nathan Lambert suggests, this works out to 100 billion tokens in total.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-18" href="#footnote-anchor-18" class="footnote-number" contenteditable="false" target="_self">18</a><div class="footnote-content"><p>Some kinds of distillation involve much more data. For example, some Llama 4 models were <a href="https://ai.meta.com/blog/llama-4-multimodal-intelligence/">trained with &#8220;codistillation&#8221;</a>, where the smaller model is trained alongside a much larger one that&#8217;s still learning itself, so it sees as much data as the larger model. This kind of distillation is probably much harder for compute-poor labs to replicate just by externally accessing frontier models.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-19" href="#footnote-anchor-19" class="footnote-number" contenteditable="false" target="_self">19</a><div class="footnote-content"><p>As far as I can tell, it&#8217;s not the case that Alibaba now only releases closed models &#8212; they have a hybrid structure where most models are open, but their best models are closed. This isn&#8217;t totally new either &#8212; it&#8217;s been the case <a href="https://www.cometapi.com/alibabas-qwen-is-it-truly-open-source/">since mid-2024</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-20" href="#footnote-anchor-20" class="footnote-number" contenteditable="false" target="_self">20</a><div class="footnote-content"><p>Though I&#8217;m not sure if they&#8217;ve reverted back to a closed-weight approach with <a href="https://ernie.baidu.com/blog/posts/ernie5.0/">ERNIE 5.0</a>.</p></div></div>]]></content:encoded></item><item><title><![CDATA[What do frontier AI companies' job postings reveal about their plans?]]></title><description><![CDATA[A fast increase in go-to-market roles, and hints about upcoming products]]></description><link>https://epochai.substack.com/p/what-do-frontier-ai-companies-job</link><guid isPermaLink="false">https://epochai.substack.com/p/what-do-frontier-ai-companies-job</guid><dc:creator><![CDATA[JSD]]></dc:creator><pubDate>Thu, 26 Mar 2026 15:02:53 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!O98n!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd4478e5-f99a-4ac2-a843-2725b74978f7_1026x1283.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>This post is part of Epoch AI&#8217;s <a href="https://epoch.ai/gradient-updates">Gradient Updates</a> newsletter, which shares more opinionated or informal takes on big questions in AI progress. These posts solely represent the views of the authors, and do not necessarily reflect the views of Epoch AI as a whole.</em></p><p><em>Originally posted on <a href="https://epoch.ai/gradient-updates/ai-lab-job-postings">Epoch AI</a>.</em></p><div><hr></div><p>AI companies guard their strategies closely. Their hiring pages, however, are public.</p><p>And those posts contain clues about what products a company is developing, who it hopes to sell them to, and which bottlenecks it sees coming. A posting for a &#8220;Camera ISP Software Engineer&#8221; suggests a device with a camera. A search for &#8220;Forward Deployed Engineers&#8221; hints at the challenges of deploying AI inside companies. A cluster of roles mentioning robotics implies ambitions well beyond chatbots.</p><p style="text-align: justify;">We analyzed open roles at the leading foundation labs, including OpenAI, Anthropic, xAI and Google DeepMind. Here is what we found:</p><ul><li><p style="text-align: justify;">First, <a href="https://docs.google.com/document/d/19i41rG9NR6rV0JkHoyVMX06h_GmdrtXBvFfUiRY84-Y/edit?tab=t.262k7agn5hfz#heading=h.5brt0rid9fsu">sales and sales-related hiring has increased sharply</a> at both Anthropic and OpenAI over the past year. Anthropic&#8217;s go-to-market share of open roles grew from 17% to 31% and OpenAI&#8217;s from 18% to 28%. This increase has been particularly concentrated in technical roles that help clients deploy AI to their companies.</p></li><li><p style="text-align: justify;">Second, open roles can provide insight into the <a href="https://docs.google.com/document/d/19i41rG9NR6rV0JkHoyVMX06h_GmdrtXBvFfUiRY84-Y/edit?tab=t.262k7agn5hfz#heading=h.pezvcegrhike">product roadmap at the labs</a>. For example, OpenAI and DeepMind are both investing in hardware products, such as robotics and consumer devices. In contrast, Anthropic appears to be focusing on improvements to its core products instead.</p></li><li><p style="text-align: justify;">Third, career pages can help give insight into how the leading AI companies are using different strategies to <a href="https://docs.google.com/document/d/19i41rG9NR6rV0JkHoyVMX06h_GmdrtXBvFfUiRY84-Y/edit?tab=t.262k7agn5hfz#heading=h.ealsrd3co8wv">acquire key inputs such as compute or data</a>. For example, OpenAI has 21 open roles related to custom silicon, while Anthropic, which does not have an internal effort aimed at developing its own chips, has none.</p></li></ul><p>A few caveats before we dive in. Job postings tell us about roles a company is trying to fill, not about its current staff. For example, a team with 20 open roles could be large and growing, or a brand-new team that doesn&#8217;t exist yet. We also can&#8217;t tell how many people will be hired for each posting: a single &#8220;Research Engineer&#8221; listing might yield one hire or ten or never get filled.</p><h1><strong>Go-to-market is the top hiring category at OpenAI and Anthropic</strong></h1><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!O98n!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd4478e5-f99a-4ac2-a843-2725b74978f7_1026x1283.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!O98n!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd4478e5-f99a-4ac2-a843-2725b74978f7_1026x1283.png 424w, https://substackcdn.com/image/fetch/$s_!O98n!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd4478e5-f99a-4ac2-a843-2725b74978f7_1026x1283.png 848w, https://substackcdn.com/image/fetch/$s_!O98n!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd4478e5-f99a-4ac2-a843-2725b74978f7_1026x1283.png 1272w, https://substackcdn.com/image/fetch/$s_!O98n!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd4478e5-f99a-4ac2-a843-2725b74978f7_1026x1283.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!O98n!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd4478e5-f99a-4ac2-a843-2725b74978f7_1026x1283.png" width="1026" height="1283" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/dd4478e5-f99a-4ac2-a843-2725b74978f7_1026x1283.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1283,&quot;width&quot;:1026,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!O98n!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd4478e5-f99a-4ac2-a843-2725b74978f7_1026x1283.png 424w, https://substackcdn.com/image/fetch/$s_!O98n!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd4478e5-f99a-4ac2-a843-2725b74978f7_1026x1283.png 848w, https://substackcdn.com/image/fetch/$s_!O98n!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd4478e5-f99a-4ac2-a843-2725b74978f7_1026x1283.png 1272w, https://substackcdn.com/image/fetch/$s_!O98n!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd4478e5-f99a-4ac2-a843-2725b74978f7_1026x1283.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p style="text-align: justify;">Sales and sales-related hiring has increased sharply at both Anthropic and at OpenAI over the past year: Anthropic&#8217;s share of open roles dedicated to go-to-market grew from 17% to 31%, while OpenAI&#8217;s grew from 18% to 28%. Perhaps unsurprisingly for companies with rapidly increasing revenue competing over a largely untapped market, sales-related roles now represent the largest category of open roles at both companies. Research, by comparison, now makes up only 12% of open roles at Anthropic and only 7% at OpenAI.</p><p style="text-align: justify;">One subcategory has grown especially fast: technical roles dedicated to helping customers actually adopt AI. Both companies are now hiring roles such as <a href="https://web.archive.org/web/20260306135207/https://openai.com/careers/ai-success-engineer-edu-san-francisco/">AI Success Engineers</a>, <a href="https://web.archive.org/web/20260310210044/https://openai.com/careers/partner-ai-deployment-engineer-san-francisco/">Partner AI Deployment Engineers</a>, <a href="https://web.archive.org/web/20260310210050mp_/https://job-boards.greenhouse.io/anthropic/jobs/4977624008">Solutions Architects</a> and <a href="https://web.archive.org/web/20260309022933mp_/https://job-boards.greenhouse.io/anthropic/jobs/5012991008">Forward Deployed Engineers</a>, designed to help customers find opportunities to use AI and integrate it into their businesses. Over the past year, the share of open roles dedicated to adoption increased from 5% to 11% at Anthropic and from 11% to 17% at OpenAI. This suggests customers are struggling to fully utilize OpenAI and Anthropic&#8217;s products, and an important aspect of bridging this gap is to educate customers on what is possible.</p><p style="text-align: justify;">The geographic distribution of sales roles also tells us something about where these companies see their market. More than half of each company&#8217;s sales roles are located in the U.S. with 52% of open roles at Anthropic and 55% at OpenAI. Neither company discloses its geographic revenue split, but the hiring concentration suggests the U.S. remains the primary market for both Anthropic and OpenAI by a wide margin.</p><p style="text-align: justify;">Internationally, both companies are hiring aggressively across Europe and Asia-Pacific. Anthropic tilts more toward Europe, where 29% of its open sales roles are located versus 21% for OpenAI. OpenAI tilts more toward Asia-Pacific, at 24% versus Anthropic&#8217;s 19%. Within Asia-Pacific, the growth is concentrated in <a href="https://web.archive.org/web/20260303204816/https://job-boards.greenhouse.io/anthropic/jobs/5055488008">Japan</a>, <a href="https://web.archive.org/web/20260225002840/https://job-boards.greenhouse.io/anthropic/jobs/5014500008">South Korea</a>, <a href="https://web.archive.org/web/20260225002927/https://job-boards.greenhouse.io/anthropic/jobs/5117581008">India</a>, <a href="https://web.archive.org/web/20260307235220/https://openai.com/careers/ai-success-engineer-singapore-singapore/">Singapore</a>, and <a href="https://web.archive.org/web/20260210184604/https://openai.com/careers/ai-deployment-engineer-sydney-australia/">Australia</a>. Notably absent are China, the Middle East, Latin America, and Africa. The labs&#8217; focus on global sales suggests that they do not expect to be crowded out of the European and Asia-Pacific markets by national champions.</p><p style="text-align: justify;">Government sales appear to be another important priority. OpenAI and Anthropic both have 10 <a href="https://web.archive.org/web/20260316200123/https://openai.com/careers/search/?c=9dd96fff-cd6b-431f-ba41-7a73ab3dea57">sales roles</a> covering federal civilian, defense, and state and local government. Of these roles, 1 of OpenAI&#8217;s roles specifically <a href="https://web.archive.org/web/20260312013141/https://openai.com/careers/government-account-director-defense-washington-dc/">targets national security</a> and <a href="https://web.archive.org/web/20260225003908/https://job-boards.greenhouse.io/anthropic/jobs/5118189008">2</a> <a href="https://web.archive.org/web/20260227015238/https://job-boards.greenhouse.io/anthropic/jobs/5126080008">of</a> Anthropic&#8217;s roles specifically target national security. xAI has 2 open sales roles targeting international governments, located in <a href="https://web.archive.org/web/20260317133006/https://job-boards.greenhouse.io/xai/jobs/5074107007">London</a> and <a href="https://web.archive.org/web/20260306021853/https://job-boards.greenhouse.io/xai/jobs/5050658007">Dubai</a>, and 1 open role targeting the <a href="https://web.archive.org/web/20260306021850/https://job-boards.greenhouse.io/xai/jobs/4793939007">US government</a>. These roles demonstrate the importance of government as a future revenue stream for the foundation labs.</p><p style="text-align: justify;">Unlike Anthropic, OpenAI and xAI, where open sales roles can give insight into the current go-to-market priorities of the AI labs, DeepMind&#8217;s job ads reveal little about their sales motion at all, since Google&#8217;s existing sales organization handles distribution for Gemini.</p><h1><strong>Job postings shed light on new product bets at OpenAI and DeepMind</strong></h1><p>Open roles can also offer a window into what each lab is building. Anthropic has 5 <a href="https://web.archive.org/web/20260310171808/https://job-boards.greenhouse.io/anthropic/jobs/4816198008">open</a> <a href="https://web.archive.org/web/20260310171808/https://job-boards.greenhouse.io/anthropic/jobs/5126691008">product</a> <a href="https://web.archive.org/web/20260303165343/https://job-boards.greenhouse.io/anthropic/jobs/5098025008">and</a> <a href="http://web.archive.org/web/20260310171808/https://job-boards.greenhouse.io/anthropic/jobs/5098506008">engineering</a> <a href="http://web.archive.org/web/20260225003250/https://job-boards.greenhouse.io/anthropic/jobs/4985920008">roles</a> aimed at improving Claude Code, while OpenAI has <a href="https://web.archive.org/web/20260323165515/https://openai.com/careers/software-engineer-codex-app-san-francisco/">10</a> <a href="https://web.archive.org/web/20260323165504/https://openai.com/careers/software-engineer-codex-cloud-san-francisco/">similar</a> <a href="https://web.archive.org/web/20260219030422/https://openai.com/careers/product-manager-codex-san-francisco/">roles</a> <a href="https://web.archive.org/web/20260224171653/https://openai.com/careers/research-engineer-codex-san-francisco/">aimed</a> at improving Codex. Both OpenAI and <a href="https://web.archive.org/web/20260323185508/https://job-boards.greenhouse.io/anthropic/jobs/5140492008">Anthropic</a> each have an engineering role dedicated to building product add-ons for financial services. OpenAI also has 3 <a href="https://web.archive.org/web/20260210185436/https://openai.com/careers/full-stack-engineer-health-ai-san-francisco/">roles</a> <a href="https://web.archive.org/web/20260210185157/https://openai.com/careers/lead-product-designer-health-ai-san-francisco/">targeting</a> <a href="https://web.archive.org/web/20260210185207/https://openai.com/careers/researcher-health-ai-san-francisco/">new</a> features for <a href="https://openai.com/index/introducing-chatgpt-health/">ChatGPT Health</a> and <a href="https://openai.com/index/openai-for-healthcare/">OpenAI for Healthcare</a>.</p><p>That said, the view from job postings is imperfect for existing products. It can be hard to tell whether a role is expanding an existing feature or building something new, and platform or infrastructure roles often span multiple products. Therefore, job postings are most informative when they reveal new bets.</p><p style="text-align: justify;">First, OpenAI&#8217;s postings show it is investing heavily in a consumer hardware <a href="https://www.theinformation.com/articles/inside-openai-team-developing-ai-devices">device</a>, with 15 open roles for the project. The postings reveal a fair amount about what the device looks like:  a <a href="https://web.archive.org/web/20260310212010/https://openai.com/careers/camera-isp-software-engineer-consumer-devices-san-francisco/">Camera ISP Software Engineer role</a> describes building imaging systems for a battery-powered portable device, a <a href="https://web.archive.org/web/20260311150716/https://openai.com/careers/senior-research-engineerscientist-on-device-transformer-models-san-francisco/">Research Engineer role</a> focuses on running transformer models directly on the device, and an <a href="https://web.archive.org/web/20260323190250/https://openai.com/careers/operating-systems-engineer-or-consumer-devices-san-francisco/?__cf_chl_rt_tk=ONdDc4l7KYnkrBlzRW6buU8uAGCermtHQ5G7tlbnwWo-1774292570-1.0.1.1-SyJT9CzE1IfnvbMXrDQ.r9s88.ZvPX2V3RncK6_6H2s">Operating Systems Engineer role</a> references custom silicon. Taken together, these roles suggest something like a portable device equipped with a camera, running its own AI chip, and designed to run AI models on the edge. Moreover, two Singapore-based <a href="https://web.archive.org/web/20260310212337/https://openai.com/careers/hardware-exploratory-singapore-singapore/">hardware</a> and <a href="https://web.archive.org/web/20260310212234/https://openai.com/careers/operations-exploratory-singapore/">operations</a> hires suggest a move toward manufacturing. DeepMind is also making a hardware bet, with two open roles for the <a href="https://web.archive.org/web/20260310213231/https://job-boards.greenhouse.io/deepmind/jobs/7094457">development of XR glasses</a>, one of which suggests <a href="https://web.archive.org/web/20260310214125/https://job-boards.greenhouse.io/deepmind/jobs/7651945">voice commands</a> will be a key interaction mode.</p><p style="text-align: justify;">Second, both OpenAI and DeepMind are betting on robotics. OpenAI has seven robotics roles that indicate work on <a href="https://web.archive.org/web/20260310212421/https://openai.com/careers/simulation-environments-engineer-san-francisco/">training robots in simulation</a> at scale and <a href="https://web.archive.org/web/20260210184843/https://openai.com/careers/simulation-realism-engineer-san-francisco/">improving simulation realism</a>. The postings also suggest some robots may have <a href="https://web.archive.org/web/20260210185227mp_/https://openai.com/careers/technical-soft-goods-designer-san-francisco/">soft components or coverings</a>, and that production is scaling up. DeepMind is hiring 9 roles dedicated to its robotics program, which suggest that they are <a href="https://web.archive.org/web/20260310213305/https://job-boards.greenhouse.io/deepmind/jobs/7576917">building a humanoid robot</a>, with <a href="https://web.archive.org/web/20260310213222/https://job-boards.greenhouse.io/deepmind/jobs/7647241">dexterous hands</a>.</p><p style="text-align: justify;">Beyond hardware, OpenAI has 2 incubation-stage <a href="https://web.archive.org/web/20260210185245/https://openai.com/careers/full-stack-software-engineer-social-products-san-francisco/">roles</a> focused on social products, and <a href="https://web.archive.org/web/20260323190541/https://openai.com/careers/full-stack-software-engineer-jobs-platform-san-francisco/">one</a> for a <a href="https://openai.com/index/expanding-economic-opportunity-with-ai/">jobs platform</a> designed to help people train, certify their skills, and get matched with employers. Anthropic has a <a href="https://web.archive.org/web/20260225003329/https://job-boards.greenhouse.io/anthropic/jobs/5096878008">general research product manager</a> role dedicated to creating entirely new product categories, and <a href="https://web.archive.org/web/20260227062635/https://job-boards.greenhouse.io/anthropic/jobs/5127559008">another</a> aimed at developing new consumer products.</p><h1 style="text-align: justify;"><strong>Job postings also offer clues about how labs secure compute and data</strong></h1><p style="text-align: justify;">Open roles at the labs also give insight into how each company approaches its raw inputs: compute and data. The clearest divide is between labs that are building their own compute infrastructure and labs that are contracting for it.</p><p style="text-align: justify;">OpenAI has 21 open roles, mostly engineering, associated with its internal chip development effort, which represents 3% of current listed roles. Anthropic, which is not developing its own custom silicon, has taken a different path: it has multiple roles focused on overseeing datacenter design and construction with external partners, including a <a href="https://web.archive.org/web/20260309001936/https://job-boards.greenhouse.io/anthropic/jobs/5140329008">Data Center Mechanical Engineer</a> responsible for directing cooling and mechanical system design produced by external firms, and a <a href="https://web.archive.org/web/20260323234145/https://job-boards.greenhouse.io/anthropic/jobs/5157023008">Data Center Design Execution Lead</a> who owns the bridge between Anthropic&#8217;s technical requirements and third-party delivery partners. Anthropic also has 3 open legal roles dedicated to <a href="https://web.archive.org/web/20260217070930/https://job-boards.greenhouse.io/anthropic/jobs/5044994008">negotiating datacenter</a>, or <a href="https://web.archive.org/web/20260224233700/https://job-boards.greenhouse.io/anthropic/jobs/5045003008">colocation</a> contracts.</p><p style="text-align: justify;">Another input that features prominently in hiring is <a href="https://epoch.ai/gradient-updates/state-of-rl-envs">RL training environments</a>. Anthropic has multiple roles focused on building environments for training models on new capabilities, including an <a href="https://web.archive.org/web/20260226201551/https://job-boards.greenhouse.io/anthropic/jobs/4951064008">Environment Scaling</a> team that builds RL environments and manages vendor relationships, and a <a href="https://web.archive.org/web/20260321020533/https://job-boards.greenhouse.io/anthropic/jobs/5061517008">Universes</a> team building ultra-realistic long-horizon settings for agentic training. OpenAI is also hiring researchers for a <a href="https://web.archive.org/web/20260221181508/https://openai.com/careers/researcher-synthetic-rl-san-francisco/">Synthetic RL</a> team developing RL training methods using self-play, simulators, and synthetic feedback.</p><p style="text-align: justify;">In contrast to OpenAI, Anthropic and DeepMind, which do not have dedicated human labeler roles, xAI&#8217;s hiring suggests a distinct model with respect to <a href="https://x.ai/careers/open-roles?dept=4025087007,4064287007,4064300007,4064289007,4064288007,4064290007,4064286007">human data</a>. It has 27 open human data roles, which suggests that it prefers keeping its data labeling operations in-house. It&#8217;s also notable that xAI is comfortable advertising these roles publicly. Other labs also rely on human labeling at scale, but tend to outsource it or keep the hiring less visible.</p><h1 style="text-align: justify;"><strong>Conclusion</strong></h1><p style="text-align: justify;">Job postings are an imperfect signal, but they&#8217;re one of the few public windows into how the leading AI labs are evolving. The picture they paint right now is of companies that are investing heavily in selling and deploying their products, diversifying into new product categories, and competing for key inputs like compute and data. As the labs continue to grow and their strategies diverge, their career pages will remain one of the best places to watch it happen.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://epochai.substack.com/p/what-do-frontier-ai-companies-job?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://epochai.substack.com/p/what-do-frontier-ai-companies-job?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://epochai.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading! Subscribe to receive the latest from Epoch AI.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p><em>Thanks to Lynette Bye for helpful comments and editing.</em></p>]]></content:encoded></item><item><title><![CDATA[Final training runs account for a minority of R&D compute spending]]></title><description><![CDATA[New evidence following the MiniMax and Z.ai IPOs]]></description><link>https://epochai.substack.com/p/final-training-runs-account-for-a</link><guid isPermaLink="false">https://epochai.substack.com/p/final-training-runs-account-for-a</guid><dc:creator><![CDATA[JSD]]></dc:creator><pubDate>Tue, 24 Mar 2026 16:16:01 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!WC1W!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0309abb9-75f4-44bf-b6aa-35bf13eaced4_1026x1283.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>This post is part of Epoch AI&#8217;s <a href="https://epoch.ai/gradient-updates">Gradient Updates</a> newsletter, which shares more opinionated or informal takes on big questions in AI progress. These posts solely represent the views of the authors, and do not necessarily reflect the views of Epoch AI as a whole.</em></p><p><em>Originally posted on <a href="https://epoch.ai/gradient-updates/r-and-d-vs-training-compute">Epoch AI</a>.</em></p><div><hr></div><p>In the popular picture of how AI companies use compute, there are two big buckets: training and inference.</p><p>But in reality, the R&amp;D side is more complex. The final training run &#8212; the one that produces the model with a name &#8212; is only the last step in a long, expensive process of exploration. Before that run begins, companies burn through compute on: running experiments at various scales, generating synthetic data, testing which ideas work before committing to a final run, and training models that are never released.</p><p>This distinction matters. When people discuss <a href="https://epoch.ai/gradient-updates/three-issues-undermining-compute-based-ai-policies">compute thresholds</a> or the cost of training a frontier model, they often mean the final training run. However, the full cost of <em>developing </em>that model is much higher. And if most of the spending is exploration rather than execution, then a competitor who learns what works from the frontier could replicate the results for a fraction of the original cost.</p><p>So there&#8217;s more to R&amp;D compute than final training runs, but how much? Last year, we <a href="https://epoch.ai/data-insights/openai-compute-spend">estimated</a> the breakdown of OpenAI&#8217;s 2024 compute spending at around $5 billion on R&amp;D compute that year, with only about $500 million &#8211; roughly 10% &#8211; going to the final training runs that produced released models such as GPT-4.5. The rest went to scaling experiments, synthetic data generation, basic research, or other R&amp;D workloads.</p><p>However, that estimate only looked at a single company (OpenAI) for one year (2024). It&#8217;s hard to know how general this pattern is, since companies rarely disclose how they allocate their compute.</p><p>But in early January, two smaller Chinese AI companies, MiniMax and Z.ai, disclosed their R&amp;D compute spending as part of their IPO processes. Since most of MiniMax and Z.ai&#8217;s models are also published, we can estimate how much compute went to their final training runs, and compare.</p><p>We find that the pattern holds: for all three companies, final training runs account for a minority of R&amp;D compute spending despite major differences in scale, country, and business model. We also explore what these ratios might tell us about catch-up growth: companies that can learn from the frontier should in theory need less experimentation and devote more of their R&amp;D compute to training. MiniMax&#8217;s higher ratio is consistent with this, but Z.ai&#8217;s is not, and with only three data points, the evidence remains inconclusive.</p><h1><strong>Breaking down MiniMax and Z.ai&#8217;s compute spending</strong></h1><p>Our analysis relies on two types of data: the R&amp;D compute spending and the final training run compute spending.</p><p>The R&amp;D compute spending comes directly from MiniMax&#8217;s and Z.ai&#8217;s IPO documents. Both companies went public on the Hong Kong Stock Exchange in early January 2026. Since their disclosed financials do not cover the full year 2025, we have to instead rely on the most recent periods available: For MiniMax, we use reported compute spending from Q4 2024 to Q3 2025, and for Z.ai, we use H2 2024 to H1 2025.</p><p>Following the procedure in <a href="https://epoch.ai/data-insights/openai-compute-spend">our OpenAI estimate</a>, we match the reported spending window to the training compute of models released one quarter later. For MiniMax, we look for the training compute of models released in 2025. For Z.ai, we look for the training compute of models released in 2024 Q4 &#8211; 2025 Q3. This accounts for the fact that firms spend the R&amp;D compute for a model several months before its public release.</p><p>Most of the final training run compute spending comes from our <a href="https://epoch.ai/data/ai-models">database of AI models</a>, with several assumptions. Epoch reports the models&#8217; training run compute in FLOP. To convert FLOP into USD, we use the following equation:</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ixj4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9169c538-715c-4ce4-95ee-a63ea9874d46_1033x98.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ixj4!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9169c538-715c-4ce4-95ee-a63ea9874d46_1033x98.png 424w, https://substackcdn.com/image/fetch/$s_!ixj4!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9169c538-715c-4ce4-95ee-a63ea9874d46_1033x98.png 848w, https://substackcdn.com/image/fetch/$s_!ixj4!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9169c538-715c-4ce4-95ee-a63ea9874d46_1033x98.png 1272w, https://substackcdn.com/image/fetch/$s_!ixj4!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9169c538-715c-4ce4-95ee-a63ea9874d46_1033x98.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ixj4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9169c538-715c-4ce4-95ee-a63ea9874d46_1033x98.png" width="1033" height="98" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9169c538-715c-4ce4-95ee-a63ea9874d46_1033x98.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:98,&quot;width&quot;:1033,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:21424,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://epochai.substack.com/i/191875493?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9169c538-715c-4ce4-95ee-a63ea9874d46_1033x98.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ixj4!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9169c538-715c-4ce4-95ee-a63ea9874d46_1033x98.png 424w, https://substackcdn.com/image/fetch/$s_!ixj4!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9169c538-715c-4ce4-95ee-a63ea9874d46_1033x98.png 848w, https://substackcdn.com/image/fetch/$s_!ixj4!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9169c538-715c-4ce4-95ee-a63ea9874d46_1033x98.png 1272w, https://substackcdn.com/image/fetch/$s_!ixj4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9169c538-715c-4ce4-95ee-a63ea9874d46_1033x98.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>Where:</p><ul><li><p>GPU peak FLOP/s = 9.89e14 (H800 dense BF16/FP16), unless developers specifically state the model is trained with FP8 precision.</p></li><li><p>MFU (Model FLOP Utilization) is assumed to be 0.15-0.35 for pre-training and supervised fine-tuning and 0.05 - 0.20 for RL (reinforcement learning), since RL is inference-heavy and inference has a lower utilization than training. We propagate this MFU uncertainty by Monte Carlo simulation (a way to estimate uncertainty intervals through a series of calculations).</p></li><li><p>Price/GPU hour is assumed to be $1.50 - $3.00. We also propagate this price uncertainty by Monte Carlo.</p></li></ul><p>Sometimes, firms do not disclose sufficient information to estimate training compute with precision in our <a href="https://epoch.ai/data/ai-models">database of AI models</a>. If so, we estimate a 90% confidence interval for the training compute of those models.</p><p>Most model estimates do not significantly affect the final total training compute, except for MiniMax&#8217;s Hailuo 02. We estimate a 90% CI for its compute spending of [$5M, $51M], with a median of $18M &#8211; about 40% of Minimax&#8217;s total 2025 training compute. We think this is reasonable, since Hailuo accounted for 32.7% of MiniMax&#8217;s 2025 revenue and training spend on a model tends to correlate with its commercial importance. Note that if we exclude the Hailuo model family entirely from the final training run compute, MiniMax&#8217;s training-to-R&amp;D ratio is more on par with OpenAI&#8217;s training-to-R&amp;D ratio, with a median of 12.2% and a 90% CI of [6.4%, 23.5%].</p><p>Lastly, data for OpenAI&#8217;s compute allocation comes from our previous <a href="https://epoch.ai/data-insights/openai-compute-spend">data insight</a>.</p><h1><strong>Final training runs are a small fraction of R&amp;D compute spending</strong></h1><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!WC1W!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0309abb9-75f4-44bf-b6aa-35bf13eaced4_1026x1283.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!WC1W!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0309abb9-75f4-44bf-b6aa-35bf13eaced4_1026x1283.png 424w, https://substackcdn.com/image/fetch/$s_!WC1W!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0309abb9-75f4-44bf-b6aa-35bf13eaced4_1026x1283.png 848w, https://substackcdn.com/image/fetch/$s_!WC1W!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0309abb9-75f4-44bf-b6aa-35bf13eaced4_1026x1283.png 1272w, https://substackcdn.com/image/fetch/$s_!WC1W!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0309abb9-75f4-44bf-b6aa-35bf13eaced4_1026x1283.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!WC1W!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0309abb9-75f4-44bf-b6aa-35bf13eaced4_1026x1283.png" width="1026" height="1283" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0309abb9-75f4-44bf-b6aa-35bf13eaced4_1026x1283.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1283,&quot;width&quot;:1026,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!WC1W!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0309abb9-75f4-44bf-b6aa-35bf13eaced4_1026x1283.png 424w, https://substackcdn.com/image/fetch/$s_!WC1W!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0309abb9-75f4-44bf-b6aa-35bf13eaced4_1026x1283.png 848w, https://substackcdn.com/image/fetch/$s_!WC1W!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0309abb9-75f4-44bf-b6aa-35bf13eaced4_1026x1283.png 1272w, https://substackcdn.com/image/fetch/$s_!WC1W!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0309abb9-75f4-44bf-b6aa-35bf13eaced4_1026x1283.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>With these estimates in hand, here&#8217;s what we find. For all of OpenAI, Z.ai, and MiniMax, final training runs account for a small fraction of total R&amp;D compute spending: 9.6% for OpenAI, 22.6% for MiniMax, and 12.3% for Z.ai. This pattern holds despite major differences between the three companies. OpenAI spent $5 billion on R&amp;D compute in 2024, while MiniMax spent $141 million and Z.ai $216 million. They operate in different countries, under different regulatory environments, and with different business models. Yet in all three cases, the bulk of R&amp;D compute spending went to something other than the final training runs for released models. This shows that our <a href="https://epoch.ai/data-insights/openai-compute-spend">earlier findings</a> on OpenAI in 2024 were not a fluke.</p><p>An important caveat is that these figures measure compute spending, not compute in terms of actual FLOP performed. This is particularly important because training runs and research often have different levels of MFU (Model FLOP Utilization), or how efficiently the GPUs are used. R&amp;D workloads tend to achieve lower GPU utilization than final training runs, because experiments involve more overhead like waiting between runs, failed jobs, debugging, and other idle periods that lower GPU utilization. In contrast, final training runs optimize to keep hardware busy more consistently. As a result, even if the two activities incur the same spending, R&amp;D could use fewer actual FLOP due to lower MFU.</p><h1><strong>R&amp;D compute and catch-up growth</strong></h1><p>So far, we&#8217;ve focused on what these three companies have in common: final training runs are a small share of R&amp;D compute spending in all three cases. But the ratios aren&#8217;t identical. Can we learn anything from the differences?</p><p>A major difference between OpenAI and Z.ai/MiniMax is that OpenAI operates at the frontier of AI capabilities, and is spending an order of magnitude more on compute. At the frontier, companies face many plausible research directions, so they need to spend heavily on experiments to figure out which ideas work at scale.</p><p>If a company is farther from the frontier, it should be able to learn what works from leading companies, skip much of this costly experimentation, and allocate more of its R&amp;D compute to actual training runs. There are many ways this can happen: ideas spread through publications and rumors, competitors can train on data generated by frontier models, they can analyze a model&#8217;s behavior to reverse-engineer how it was built, or they simply learn that something is <em>possible</em> and focus their efforts accordingly.</p><p>This predicts that lagging companies should have higher training-to-R&amp;D ratios than frontier companies. MiniMax fits this prediction: its ratio appears higher than OpenAI&#8217;s or Z.ai&#8217;s. This is also consistent with Anthropic&#8217;s<a href="https://www.anthropic.com/news/detecting-and-preventing-distillation-attacks"> disclosure</a> that MiniMax ran a large-scale distillation campaign against Claude, extracting over 13 million exchanges (more than DeepSeek or Moonshot). Z.ai, on the other hand, doesn&#8217;t clearly fit the catch-up prediction, since its training-to-R&amp;D ratio looks similar to OpenAI&#8217;s.</p><p>Overall, we still only have three data points, with wide error bars on each. So these results don&#8217;t provide particularly strong evidence for or against the catch-up story. The pattern is suggestive in MiniMax&#8217;s case, but we can&#8217;t draw confident conclusions about what drives the differences in training-to-R&amp;D ratios across companies. Better understanding the business models and incentives for Z.ai and MiniMax could help explain why the two companies seemingly differ in their R&amp;D compute allocation.</p><h1><strong>Conclusion</strong></h1><p>Final training runs account for a minority of R&amp;D compute spending across all three companies we examined: this confirms that the pattern we found for OpenAI in 2024 extends to smaller companies in a different country. The differences between companies are harder to interpret with so few data points.</p><p>We&#8217;d like to do this analysis for more companies, but the binding constraint is getting reliable data on compute spending. IPO disclosures have been our best source so far, and there may be more opportunities soon when OpenAI, SpaceX, and Anthropic IPO.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://epochai.substack.com/p/final-training-runs-account-for-a?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://epochai.substack.com/p/final-training-runs-account-for-a?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://epochai.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading! Subscribe to receive the latest from Epoch AI.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p><em>Thanks to Konstantin Pilz and Lynette Bye for helpful comments and editing.</em></p><h1>Appendix</h1><h2>R&amp;D Compute Spending</h2><p>Minimax and Z.ai disclose the R&amp;D compute in their IPO documents. In particular, Z.ai&#8217;s R&amp;D compute is reported in RMB. To convert that to USD, we use an exchange rate of 7.2 RMB/USD, which was the average exchange rate in 2025.</p><h2>Final Training Run Compute Spending</h2><p>By the &#8220;Final training run compute&#8221; of a model, we mean the compute used to train that model specifically, including pre-training and post-training. If a base model as well as a fine-tuned version of it were both released, we do not double-count the compute used to train the base model.</p><p>To estimate the cost of final training runs in USD, we use the aforementioned equation:</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!zQon!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbcb5a179-3a04-48df-ac31-7ae32a924f89_1044x75.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!zQon!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbcb5a179-3a04-48df-ac31-7ae32a924f89_1044x75.png 424w, https://substackcdn.com/image/fetch/$s_!zQon!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbcb5a179-3a04-48df-ac31-7ae32a924f89_1044x75.png 848w, https://substackcdn.com/image/fetch/$s_!zQon!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbcb5a179-3a04-48df-ac31-7ae32a924f89_1044x75.png 1272w, https://substackcdn.com/image/fetch/$s_!zQon!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbcb5a179-3a04-48df-ac31-7ae32a924f89_1044x75.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!zQon!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbcb5a179-3a04-48df-ac31-7ae32a924f89_1044x75.png" width="1044" height="75" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bcb5a179-3a04-48df-ac31-7ae32a924f89_1044x75.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:75,&quot;width&quot;:1044,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:20841,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://epochai.substack.com/i/191875493?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbcb5a179-3a04-48df-ac31-7ae32a924f89_1044x75.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!zQon!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbcb5a179-3a04-48df-ac31-7ae32a924f89_1044x75.png 424w, https://substackcdn.com/image/fetch/$s_!zQon!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbcb5a179-3a04-48df-ac31-7ae32a924f89_1044x75.png 848w, https://substackcdn.com/image/fetch/$s_!zQon!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbcb5a179-3a04-48df-ac31-7ae32a924f89_1044x75.png 1272w, https://substackcdn.com/image/fetch/$s_!zQon!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbcb5a179-3a04-48df-ac31-7ae32a924f89_1044x75.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>We pair this with a Monte Carlo simulation of 50,000 draws to propagate uncertainty across our three input parameters:</p><ul><li><p>MFU, modeled as log-normally distributed, with the 10th and 90th percentiles anchored at 0.15 and 0.35 for pre-training and supervised fine-tuning, and 0.01 and 0.10 for RL post-training.</p></li><li><p>Price/GPU hour, modeled as log-normally distributed, with the 10th and 90th percentiles anchored at $1.50 and $3.00.</p></li><li><p>FLOP counts for models missing training compute information. We will explain in more detail the distribution below.</p></li></ul><p>We also assume GPU peak FLOP/s = 9.89e14 FLOP/s (H800 dense BF16/FP16), unless developers specifically state the model is trained with FP8 precision, in which case we use GPU peak FLOP/s = 1.513e15 FLOP/s. Minimax-M2 is the only model trained with FP8 precision in our sample.</p><p>A key methodological choice concerns how uncertainty in MFU and price/GPU hour is handled across models. We assume these are largely market- or period-level conditions rather than model-specific parameters. Thus, in each Monte Carlo draw, we sample one common MFU and one common price/GPU hour and apply them to all models.</p><p>FLOP counts are primarily from our <a href="https://epoch.ai/data/ai-models">AI models dataset</a>. Caveats and details are in <a href="https://github.com/CherylWu3/rd-vs-training-compute/blob/9abbcace984a76127854e8c35aaa5fbc15982a3c/model_flop.md">model_flop.md</a> in our <a href="https://github.com/CherylWu3/rd-vs-training-compute">GitHub repository</a>.</p>]]></content:encoded></item><item><title><![CDATA[The least understood driver of AI progress]]></title><description><![CDATA[An opinionated guide to &#8220;algorithmic progress&#8221; and why it matters]]></description><link>https://epochai.substack.com/p/the-least-understood-driver-of-ai</link><guid isPermaLink="false">https://epochai.substack.com/p/the-least-understood-driver-of-ai</guid><dc:creator><![CDATA[Anson Ho]]></dc:creator><pubDate>Thu, 26 Feb 2026 16:44:02 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!2OnP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F25c5ffda-685c-472d-80e3-aaa90a59469e_1024x820.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>This post is part of Epoch AI&#8217;s <a href="https://epoch.ai/gradient-updates">Gradient Updates</a> newsletter, which shares more opinionated or informal takes on big questions in AI progress. These posts solely represent the views of the authors, and do not necessarily reflect the views of Epoch AI as a whole.</em></p><p><em>Originally posted on <a href="https://epoch.ai/gradient-updates/the-least-understood-driver-of-ai-progress">Epoch AI</a>.</em></p><div><hr></div><p>AI software progress is one of those things that everyone vaguely knows about, but only a handful of people in the world truly understand its significance. Consider that many of the most fervent debates in AI to date depend enormously on it: How did <a href="https://peterwildeford.substack.com/p/ten-takes-on-deepseek">DeepSeek</a> seem to catch up to OpenAI&#8217;s o1 within months while using less training compute? <a href="https://www.astralcodexten.com/p/what-happened-with-bio-anchors">When will the world develop AGI</a>? And if we automate AI research, will AI progress accelerate like crazy a la <a href="https://situational-awareness.ai/">Situational Awareness</a> and <a href="https://ai-2027.com/">AI 2027</a>?</p><p>I don&#8217;t know your stances on these questions, but I do know that you can&#8217;t have a well-informed opinion on them without understanding software progress. So I figured I should write a post describing the most important things that you need to know, starting from the basics and leading up to the current frontier.</p><p>Here are the main takeaways, one for each section of the post:</p><ol><li><p><strong>AI software progress is about reducing the training compute you need to get to the same level of capability</strong>, through better algorithms or data. This is commonly called &#8220;algorithmic progress&#8221; including by myself, but as we&#8217;ll see, this is probably a bit of a misnomer.</p></li><li><p><strong>Almost all the evidence points to very fast software progress</strong>: each year, the training compute needed to get to the same capability declines several times &#8212; possibly even ten times or more. But existing estimates are also incredibly uncertain, because they depend on limited observational data and dubious statistical assumptions. Relaxing some of these assumptions could change some estimates by close to an order of magnitude!</p></li><li><p><strong>Recent evidence suggests that these estimates might not measure what we thought they did. </strong>Software progress has often been framed as continually finding new algorithmic innovations, which cumulatively bring massive efficiency gains. But <strong>most software progress might actually be due to data quality improvements</strong> (hence &#8220;algorithmic progress&#8221; may be a misnomer). And <strong>much of the measured efficiency gains might come from scaling up just a small handful of &#8220;scale-dependent&#8221; algorithmic changes</strong> &#8212; innovations that have a greater impact at higher training compute scales &#8212; rather than from discovering lots of new algorithms. Both of these effects were mostly unaccounted for in prior literature.</p></li><li><p>Some argue that automating AI research could lead to a <strong>&#8220;software intelligence explosion&#8221;</strong>, with AIs recursively improving themselves. Previous analyses suggest this is plausible but uncertain. However, some of these analyses are based on <strong>overly conservative estimates of software progress, which makes an explosion seem more likely </strong>when corrected. On the other hand, they also ignore scale-dependent innovations, which pose a<strong> compute bottleneck &#8212; it may be hard to get fast software progress without also scaling up training compute, potentially making an explosion less likely. </strong>The net effect isn&#8217;t totally clear: the compute bottleneck argument is suggestive but the empirical evidence is shaky, and there are plausible reasons automated AI researchers could overcome it.</p></li><li><p><strong>There are still big open problems about AI software progress</strong>, like the rate of progress in post-training, and the strength of compute bottlenecks. I think it&#8217;ll be hard to make progress on these questions, but it&#8217;s still worth trying to answer them because AI software progress is so overwhelmingly important.</p></li></ol><p>Now let&#8217;s dig into the details.</p><h2>I. AI software progress: Doing more with what we have</h2><p>First things first, what do we mean by &#8220;AI software progress&#8221;? You&#8217;re probably familiar with some examples: when the field of AI switched from LSTMs to Transformers, that was software progress. When OpenAI released the first reasoning models that took time to think before responding to users, that was software progress. It seems simple enough: with better algorithms and data (&#8220;software&#8221;), you can train better AI systems.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> There are other kinds of software progress too (like in inference), but for this post I&#8217;ll focus primarily on training.</p><p>But what exactly does it mean to have &#8220;better&#8221; software? Historically, people have defined this through the lens of AI scaling, where using more training compute is the main thing that increases capabilities. Then you define software progress relative to this: if you have better software, then you need <a href="https://arxiv.org/abs/2005.04305">less compute to reach the same level of capabilities</a>.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a></p><p>To see what I mean, let&#8217;s say we&#8217;re training a Transformer with a certain amount of compute, shown as a red dot below. If we change the amount of training compute, we get a relationship between training compute and capabilities, which we can draw as a line:<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!p9bd!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb8e413eb-5367-462d-be4e-7668f90861c8_1024x820.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!p9bd!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb8e413eb-5367-462d-be4e-7668f90861c8_1024x820.png 424w, https://substackcdn.com/image/fetch/$s_!p9bd!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb8e413eb-5367-462d-be4e-7668f90861c8_1024x820.png 848w, https://substackcdn.com/image/fetch/$s_!p9bd!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb8e413eb-5367-462d-be4e-7668f90861c8_1024x820.png 1272w, https://substackcdn.com/image/fetch/$s_!p9bd!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb8e413eb-5367-462d-be4e-7668f90861c8_1024x820.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!p9bd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb8e413eb-5367-462d-be4e-7668f90861c8_1024x820.png" width="1024" height="820" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b8e413eb-5367-462d-be4e-7668f90861c8_1024x820.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:820,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!p9bd!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb8e413eb-5367-462d-be4e-7668f90861c8_1024x820.png 424w, https://substackcdn.com/image/fetch/$s_!p9bd!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb8e413eb-5367-462d-be4e-7668f90861c8_1024x820.png 848w, https://substackcdn.com/image/fetch/$s_!p9bd!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb8e413eb-5367-462d-be4e-7668f90861c8_1024x820.png 1272w, https://substackcdn.com/image/fetch/$s_!p9bd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb8e413eb-5367-462d-be4e-7668f90861c8_1024x820.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Now let&#8217;s introduce a new software innovation, say <a href="https://arxiv.org/abs/2104.09864">rotary embeddings</a> or training data filtering. Under this definition of software progress, this reduces the training compute we need to reach a certain capability, and so we shift this curve to the left by some factor, shown as the &#8220;compute efficiency gain&#8221;:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!tAPH!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e1f3787-5bdb-4fe7-ab02-0fd54292b7af_1024x820.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!tAPH!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e1f3787-5bdb-4fe7-ab02-0fd54292b7af_1024x820.png 424w, https://substackcdn.com/image/fetch/$s_!tAPH!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e1f3787-5bdb-4fe7-ab02-0fd54292b7af_1024x820.png 848w, https://substackcdn.com/image/fetch/$s_!tAPH!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e1f3787-5bdb-4fe7-ab02-0fd54292b7af_1024x820.png 1272w, https://substackcdn.com/image/fetch/$s_!tAPH!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e1f3787-5bdb-4fe7-ab02-0fd54292b7af_1024x820.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!tAPH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e1f3787-5bdb-4fe7-ab02-0fd54292b7af_1024x820.png" width="1024" height="820" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5e1f3787-5bdb-4fe7-ab02-0fd54292b7af_1024x820.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:820,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!tAPH!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e1f3787-5bdb-4fe7-ab02-0fd54292b7af_1024x820.png 424w, https://substackcdn.com/image/fetch/$s_!tAPH!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e1f3787-5bdb-4fe7-ab02-0fd54292b7af_1024x820.png 848w, https://substackcdn.com/image/fetch/$s_!tAPH!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e1f3787-5bdb-4fe7-ab02-0fd54292b7af_1024x820.png 1272w, https://substackcdn.com/image/fetch/$s_!tAPH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e1f3787-5bdb-4fe7-ab02-0fd54292b7af_1024x820.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>What&#8217;s more, this algorithmic shift also means that we can do more with the same amount of compute, reaching the blue level of capabilities:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!2OnP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F25c5ffda-685c-472d-80e3-aaa90a59469e_1024x820.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!2OnP!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F25c5ffda-685c-472d-80e3-aaa90a59469e_1024x820.png 424w, https://substackcdn.com/image/fetch/$s_!2OnP!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F25c5ffda-685c-472d-80e3-aaa90a59469e_1024x820.png 848w, https://substackcdn.com/image/fetch/$s_!2OnP!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F25c5ffda-685c-472d-80e3-aaa90a59469e_1024x820.png 1272w, https://substackcdn.com/image/fetch/$s_!2OnP!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F25c5ffda-685c-472d-80e3-aaa90a59469e_1024x820.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!2OnP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F25c5ffda-685c-472d-80e3-aaa90a59469e_1024x820.png" width="1024" height="820" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/25c5ffda-685c-472d-80e3-aaa90a59469e_1024x820.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:820,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!2OnP!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F25c5ffda-685c-472d-80e3-aaa90a59469e_1024x820.png 424w, https://substackcdn.com/image/fetch/$s_!2OnP!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F25c5ffda-685c-472d-80e3-aaa90a59469e_1024x820.png 848w, https://substackcdn.com/image/fetch/$s_!2OnP!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F25c5ffda-685c-472d-80e3-aaa90a59469e_1024x820.png 1272w, https://substackcdn.com/image/fetch/$s_!2OnP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F25c5ffda-685c-472d-80e3-aaa90a59469e_1024x820.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>If we had wanted to reach that blue level of capabilities using the original transformer, we would&#8217;ve needed to use &#8220;compute efficiency gain&#8221; times more training compute:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Kv0v!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4f6914cb-c8f7-437d-8415-446a52055fca_1024x820.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Kv0v!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4f6914cb-c8f7-437d-8415-446a52055fca_1024x820.png 424w, https://substackcdn.com/image/fetch/$s_!Kv0v!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4f6914cb-c8f7-437d-8415-446a52055fca_1024x820.png 848w, https://substackcdn.com/image/fetch/$s_!Kv0v!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4f6914cb-c8f7-437d-8415-446a52055fca_1024x820.png 1272w, https://substackcdn.com/image/fetch/$s_!Kv0v!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4f6914cb-c8f7-437d-8415-446a52055fca_1024x820.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Kv0v!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4f6914cb-c8f7-437d-8415-446a52055fca_1024x820.png" width="1024" height="820" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4f6914cb-c8f7-437d-8415-446a52055fca_1024x820.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:820,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Kv0v!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4f6914cb-c8f7-437d-8415-446a52055fca_1024x820.png 424w, https://substackcdn.com/image/fetch/$s_!Kv0v!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4f6914cb-c8f7-437d-8415-446a52055fca_1024x820.png 848w, https://substackcdn.com/image/fetch/$s_!Kv0v!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4f6914cb-c8f7-437d-8415-446a52055fca_1024x820.png 1272w, https://substackcdn.com/image/fetch/$s_!Kv0v!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4f6914cb-c8f7-437d-8415-446a52055fca_1024x820.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>This is why researchers often define the &#8220;effective compute&#8221; = &#8220;compute efficiency gain&#8221; &#215; training compute. Because of the software innovation, we <em>in effect</em> have a larger compute budget that we can use to reach higher levels of capability. We can do more with the compute we currently have.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a></p><p>Of course, this picture is oversimplified. For example, this assumes that the lines are parallel &#8212; more on that later. But even though it&#8217;s oversimplified, I think it helps clarify why it&#8217;s so important to understand software progress. For example, if AI software progress is really fast, then frontier labs could build AGI with much less compute than today&#8217;s software, which could mean building AGI much sooner. And if these innovations are easily replicable by many actors, the implications extend further: shortly after one lab makes a big algorithmic breakthrough, another lab catches up. Shortly after frontier LLMs become good enough to help novices develop bioweapons, the novices will have enough compute to train these LLMs themselves. No wonder some people are so nervous about this!</p><p>But whether these implications actually hold depends on the numbers. How fast is software progress in practice? If we&#8217;re right about this theoretical framing, then this could be one of the most important questions for what AI will look like over the next few years. So let&#8217;s take a look at the evidence.</p><h2>II. How fast is AI software progress?</h2><p>To answer this question, I searched through the literature for AI software progress estimates, some of which I contributed to.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a> In the case of pre-training, this looks <em>very</em> fast, on the order of several times per year:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!XqDP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5ca2e96-dd6f-4318-8c99-317111c247f4_1200x1500.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!XqDP!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5ca2e96-dd6f-4318-8c99-317111c247f4_1200x1500.png 424w, https://substackcdn.com/image/fetch/$s_!XqDP!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5ca2e96-dd6f-4318-8c99-317111c247f4_1200x1500.png 848w, https://substackcdn.com/image/fetch/$s_!XqDP!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5ca2e96-dd6f-4318-8c99-317111c247f4_1200x1500.png 1272w, https://substackcdn.com/image/fetch/$s_!XqDP!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5ca2e96-dd6f-4318-8c99-317111c247f4_1200x1500.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!XqDP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5ca2e96-dd6f-4318-8c99-317111c247f4_1200x1500.png" width="1200" height="1500" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e5ca2e96-dd6f-4318-8c99-317111c247f4_1200x1500.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1500,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!XqDP!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5ca2e96-dd6f-4318-8c99-317111c247f4_1200x1500.png 424w, https://substackcdn.com/image/fetch/$s_!XqDP!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5ca2e96-dd6f-4318-8c99-317111c247f4_1200x1500.png 848w, https://substackcdn.com/image/fetch/$s_!XqDP!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5ca2e96-dd6f-4318-8c99-317111c247f4_1200x1500.png 1272w, https://substackcdn.com/image/fetch/$s_!XqDP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5ca2e96-dd6f-4318-8c99-317111c247f4_1200x1500.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>You can find the full data and sources in the Appendix.</em></figcaption></figure></div><p>Again, this means that we need <em>several times</em> less training compute each year to reach the same capability &#8212; that&#8217;s super fast! But I don&#8217;t think we should trust these specific numbers very much because they&#8217;re full of problems, and you don&#8217;t have to squint very hard to see them. First of all, the error bars look almost comically wide in the graph above &#8212; across the different estimates, they range from around 1.1&#215; to 300&#215; per year! That&#8217;s why I say &#8220;several times per year&#8221; rather than any particular number &#8212; it seems pretty plausible based on the range of estimates, but I don&#8217;t want to signal too much confidence.</p><p>Why are these estimates so uncertain? I think the biggest reason is that there&#8217;s not enough high-quality data out there. All the estimates in the graph above are based on observational approaches, which build a dataset of existing AI models and then try to back out an estimate from that. To do this, each model in the dataset needs three ingredients: (1) a training compute estimate, (2) a measure of its &#8220;capabilities&#8221;, and (3) a description of when it was released.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a> But in practice we often don&#8217;t know the training compute, and it&#8217;s hard to find a capability metric that&#8217;s used for a long time, because things like benchmarks saturate too quickly to capture long-run trends. So we&#8217;re left with less data than we&#8217;d like, and more uncertainty than we&#8217;d want.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a></p><p>These uncertainties are compounded by how we estimate the rate of progress. For example, one approach is what I like to call the &#8220;lines on a graph&#8221; approach &#8212; pick a capability level, find models that achieve it with less training compute over time, and measure how fast that compute requirement falls.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-8" href="#footnote-8" target="_self">8</a> That looks like this (hence the name of the approach):</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!41iy!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99ce995a-5000-427e-b1ca-fefddafaee11_1600x1156.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!41iy!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99ce995a-5000-427e-b1ca-fefddafaee11_1600x1156.png 424w, https://substackcdn.com/image/fetch/$s_!41iy!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99ce995a-5000-427e-b1ca-fefddafaee11_1600x1156.png 848w, https://substackcdn.com/image/fetch/$s_!41iy!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99ce995a-5000-427e-b1ca-fefddafaee11_1600x1156.png 1272w, https://substackcdn.com/image/fetch/$s_!41iy!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99ce995a-5000-427e-b1ca-fefddafaee11_1600x1156.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!41iy!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99ce995a-5000-427e-b1ca-fefddafaee11_1600x1156.png" width="1456" height="1052" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/99ce995a-5000-427e-b1ca-fefddafaee11_1600x1156.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1052,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!41iy!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99ce995a-5000-427e-b1ca-fefddafaee11_1600x1156.png 424w, https://substackcdn.com/image/fetch/$s_!41iy!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99ce995a-5000-427e-b1ca-fefddafaee11_1600x1156.png 848w, https://substackcdn.com/image/fetch/$s_!41iy!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99ce995a-5000-427e-b1ca-fefddafaee11_1600x1156.png 1272w, https://substackcdn.com/image/fetch/$s_!41iy!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99ce995a-5000-427e-b1ca-fefddafaee11_1600x1156.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>(<a href="https://epoch.ai/blog/a-rosetta-stone-for-ai-benchmarks">Source</a>) The &#8220;lines on a graph&#8221; approach to estimating the rate of AI software progress. The rate of progress is given by the slope of the line.</em></figcaption></figure></div><p>Unfortunately, because we don&#8217;t have large datasets, these &#8220;lines on a graph&#8221; are fit with only a handful of data points, making the estimates really noisy. On occasion this can yield ludicrously fast estimates, like <a href="https://www.lesswrong.com/posts/yXLqrpfFwBW5knpgc/catch-up-algorithmic-progress-might-actually-be-60-per-year#:~:text=And%20the%20accompanying%20table%3A">30,000&#215; per year</a> or even <a href="https://arxiv.org/html/2512.00193v1#:~:text=to%20data%20limitations-,.23,-To%20make%20the">5<sup>12</sup>&#215; per year</a>!<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-9" href="#footnote-9" target="_self">9</a> We can mark these sorts of cases as obviously dubious outliers and focus more on the median trend across the lines-on-graphs, but it somewhat undermines my trust in this approach.</p><p>The other way to estimate the rate of software progress is what I&#8217;ll call the &#8220;fancy statistics&#8221; approach. Here the idea is to have some more complicated model relating compute, capabilities, and time, fit the parameters to the data, and then use that to back out an estimate. The key difference is that it&#8217;s more explicit about what this relationship looks like, whereas the &#8220;lines on a graph&#8221; approach is a bit more agnostic about that. So if the assumed relationship is wrong, that could throw off the estimates.</p><p>But for the most part, both approaches share similar weaknesses. For example, in <a href="https://arxiv.org/abs/2403.05812">one of my own papers</a> I modeled software progress as growing exponentially and only depending on time. In doing so, I implicitly assumed that the &#8220;software quality&#8221; is the same across all labs at any point in time, which is clearly false &#8212; some labs have better algorithms and data than others. Parker Whitfill <a href="https://arxiv.org/abs/2508.11033">argued</a> that this could bias the estimates to be substantially faster or slower &#8212; for instance, growth rates might be as much as 9&#215; slower than what my paper estimates! Sadly for me, that&#8217;s potentially quite a devastating rebuttal.</p><p>Even if we set aside these uncertainties, there are still plenty of problems with all the estimates I showed you. For example, all the estimates fail to fully capture software improvements in post-training,<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-10" href="#footnote-10" target="_self">10</a> which is especially egregious because reasoning post-training is now all the rage among frontier models. But the sad reality is that there&#8217;s not enough public post-training compute data to let us reliably work out a growth trend.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-11" href="#footnote-11" target="_self">11</a></p><p>One last problem I&#8217;ll raise has to do with how we measure &#8220;capabilities&#8221;. Typically this means looking at AI benchmarks, but models can be &#8220;bench maxxed&#8221; to do well on benchmarks in a way that fails to generalize to real-world use-cases. So if we measure software progress using only these metrics, we&#8217;d end up overstating growth rates &#8212; what if we&#8217;re just getting much better on the things that we can easily measure, and not so much elsewhere? My guess is that this should make us slightly more skeptical about fast progress, but not by that much, because we do see models generalizing to some extent. I wish I knew how to be more precise than that.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-12" href="#footnote-12" target="_self">12</a></p><p>Okay, so there are myriad problems and uncertainties with existing software progress estimates, some of them potentially quite devastating. Now what? If we could be so far off the mark, should we trust these numbers at all?</p><p>I think there&#8217;s still a decent case for software progress being &#8220;several times per year&#8221;. For one, the estimates above are based on different approaches and all suggest that these rapid growth rates are possible &#8212; I doubt that&#8217;s just a coincidence. It&#8217;s also consistent with <a href="https://cdn.openai.com/papers/ai_and_efficiency.pdf">numbers</a> in computer vision reported by frontier AI companies, and <a href="https://www.darioamodei.com/post/on-deepseek-and-export-controls#:~:text=But%20what%27s%20important,maybe%20~4x/year.">corroborates Dario Amodei</a>. What&#8217;s more, we can just look at example case studies. Here&#8217;s one <a href="https://x.com/EpochAIResearch/status/1953212437794021561">example</a> from last year, which suggests an annual growth rate of 2-5&#215; per year:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!b_wk!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71e3e937-9d8d-473a-90f9-0bf0543d1c19_1188x760.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!b_wk!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71e3e937-9d8d-473a-90f9-0bf0543d1c19_1188x760.png 424w, https://substackcdn.com/image/fetch/$s_!b_wk!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71e3e937-9d8d-473a-90f9-0bf0543d1c19_1188x760.png 848w, https://substackcdn.com/image/fetch/$s_!b_wk!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71e3e937-9d8d-473a-90f9-0bf0543d1c19_1188x760.png 1272w, https://substackcdn.com/image/fetch/$s_!b_wk!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71e3e937-9d8d-473a-90f9-0bf0543d1c19_1188x760.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!b_wk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71e3e937-9d8d-473a-90f9-0bf0543d1c19_1188x760.png" width="1188" height="760" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/71e3e937-9d8d-473a-90f9-0bf0543d1c19_1188x760.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:760,&quot;width&quot;:1188,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!b_wk!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71e3e937-9d8d-473a-90f9-0bf0543d1c19_1188x760.png 424w, https://substackcdn.com/image/fetch/$s_!b_wk!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71e3e937-9d8d-473a-90f9-0bf0543d1c19_1188x760.png 848w, https://substackcdn.com/image/fetch/$s_!b_wk!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71e3e937-9d8d-473a-90f9-0bf0543d1c19_1188x760.png 1272w, https://substackcdn.com/image/fetch/$s_!b_wk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71e3e937-9d8d-473a-90f9-0bf0543d1c19_1188x760.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>(<a href="https://x.com/EpochAIResearch/status/1953212437794021561">Source</a>) gpt-oss-20b does substantially better than GPT-3 on MMLU, despite using the same amount of training compute. If we look at the relationship between pre-training compute and MMLU performance among non-reasoning GPT models, we can back out a rate of algorithmic improvement from this &#8212; this works out to around 2-5&#215; per year.</em></figcaption></figure></div><p>Overall I&#8217;d say that the evidence looks pretty shoddy, so I don&#8217;t think it makes sense to <em>confidently</em> declare that &#8220;software progress is 3&#215; per year&#8221; or any particular number. But I think we have enough evidence to think that software progress might really be several times a year, and to make a best guess contextualized with a lot of uncertainty.</p><p>So here&#8217;s my best guess: after accounting for all training compute (including post-training), I think we&#8217;re seeing software progress at around 10&#215; per year, and my 80% credible interval would probably range from 2&#215; to 50&#215; per year. But again, this is the kind of thing that might change depending on the day and what I had for breakfast that morning.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://epochai.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://epochai.substack.com/subscribe?"><span>Subscribe now</span></a></p><h2>III. What drives software progress? (Or, why all the estimates we just saw are misleading)</h2><p>If we think about it a bit more, these estimates seem quite fishy: have model architectures and training algorithms really changed enough to justify something like a 10&#215; per year rate of progress?</p><p>Consider <a href="https://cdn.openai.com/better-language-models/language_models_are_unsupervised_multitask_learners.pdf">GPT-2</a> and <a href="https://arxiv.org/abs/2412.19437">DeepSeek-V3</a>. These two models were released almost six years apart, which at 10&#215; per year would imply a million-fold difference in compute efficiency. That seems crazy, especially because they don&#8217;t look <em>that</em> different on an algorithmic level. Sure, DeepSeek-V3 is a mixture of expert model, uses different <a href="https://epoch.ai/gradient-updates/how-has-deepseek-improved-the-transformer-architecture">variants of Attention</a> and positional encoding schemes, and so on &#8212; but how is that supposed to explain a factor of <em>one million</em>?</p><p>I think there are two plausible factors, both of which play an important role. One is that these improvements aren&#8217;t just coming from better algorithms &#8212; the bulk of this comes from better <em>data</em>. To my knowledge, the best <a href="https://www.beren.io/2025-08-02-Most-Algorithmic-Progress-is-Data-Progress/">argument</a> for this position comes from Beren Millidge, who brings several pieces of evidence to the table:</p><ul><li><p>We&#8217;ve shifted from gnarly, uncurated web data to heavily processed and often <a href="https://vintagedata.org/blog/posts/synthetic-pretraining">synthetic data</a>, which is strongly targeted towards the tasks that we want models to do well on. This could include <a href="https://arxiv.org/abs/2406.11794">techniques</a> like filtering data and finding the right data mix to get better performance.</p></li><li><p>Researchers have been throwing tons of effort into getting better training data. For example, Surge AI had a revenue of <a href="https://www.inc.com/brian-contreras/surge-ai-startup-bootstrapped-1-billion-revenue-venture-capital-vc/91205888">over $1 billion</a> last August, and Scale AI was probably in a similar boat. I&#8217;d also add that this checks out with how OpenAI and Anthropic are spending in the <a href="https://www.theinformation.com/articles/anthropic-openai-developing-ai-co-workers?rc=spkbjw">billions</a> on data-related costs &#8212; this includes <a href="https://epoch.ai/gradient-updates/state-of-rl-envs">RL environments</a>, which are now very much a part of the AI zeitgeist.</p></li></ul><p>To my mind, the strongest point here is about synthetic data (and distillation). These are techniques that could help models reach certain capabilities with much less training compute.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-13" href="#footnote-13" target="_self">13</a> Here&#8217;s an illustrative example from an <a href="https://epoch.ai/gradient-updates/three-issues-undermining-compute-based-ai-policies">article</a> I co-authored last year:</p><blockquote><p>The efficiency gains from distillation can be dramatic. <a href="https://arxiv.org/pdf/1910.01108v4">DistilBERT</a>, an early demonstration of this technique, preserved 97% of its teacher model&#8217;s capabilities while using 40% fewer parameters and requiring merely 3% of the compute budget that went into training the original BERT model.</p></blockquote><p>In other words, in this case distillation made it possible to reduce training compute by something like thirty to forty times while keeping roughly the same capabilities.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-14" href="#footnote-14" target="_self">14</a> Synthetic data can help push beyond this &#8212; a good <a href="https://arxiv.org/abs/2412.08905">example</a> that Millidge raises is the Phi series of models. So if you stack these kinds of gains, you can sort of see how this could help explain orders-of-magnitude improvements in software efficiency.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-15" href="#footnote-15" target="_self">15</a> But I&#8217;m still not totally sure, because <a href="https://epoch.ai/blog/algorithmic-progress-in-language-models">some</a> <a href="https://epoch.ai/blog/a-rosetta-stone-for-ai-benchmarks">of</a> the studies I listed in the previous section deliberately (try to) exclude distilled models and synthetic data while still finding absurdly fast rates of software progress. I&#8217;m not yet totally convinced that the numbers add up.</p><p>This brings us to the second explanation for ultra-fast-software-progress, namely that software improvements depend on the scale of training compute. To see what I mean, let&#8217;s go back to the picture we drew earlier to illustrate software progress:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!tAPH!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e1f3787-5bdb-4fe7-ab02-0fd54292b7af_1024x820.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!tAPH!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e1f3787-5bdb-4fe7-ab02-0fd54292b7af_1024x820.png 424w, https://substackcdn.com/image/fetch/$s_!tAPH!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e1f3787-5bdb-4fe7-ab02-0fd54292b7af_1024x820.png 848w, https://substackcdn.com/image/fetch/$s_!tAPH!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e1f3787-5bdb-4fe7-ab02-0fd54292b7af_1024x820.png 1272w, https://substackcdn.com/image/fetch/$s_!tAPH!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e1f3787-5bdb-4fe7-ab02-0fd54292b7af_1024x820.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!tAPH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e1f3787-5bdb-4fe7-ab02-0fd54292b7af_1024x820.png" width="1024" height="820" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5e1f3787-5bdb-4fe7-ab02-0fd54292b7af_1024x820.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:820,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!tAPH!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e1f3787-5bdb-4fe7-ab02-0fd54292b7af_1024x820.png 424w, https://substackcdn.com/image/fetch/$s_!tAPH!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e1f3787-5bdb-4fe7-ab02-0fd54292b7af_1024x820.png 848w, https://substackcdn.com/image/fetch/$s_!tAPH!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e1f3787-5bdb-4fe7-ab02-0fd54292b7af_1024x820.png 1272w, https://substackcdn.com/image/fetch/$s_!tAPH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5e1f3787-5bdb-4fe7-ab02-0fd54292b7af_1024x820.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>This depicts software improvements as a parallel shift to the left. But what if it&#8217;s not parallel? What if the slope changes, for example like this?</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!pNx0!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42f07c84-c74a-4f3e-9db0-1d51fd81dca9_1024x820.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!pNx0!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42f07c84-c74a-4f3e-9db0-1d51fd81dca9_1024x820.png 424w, https://substackcdn.com/image/fetch/$s_!pNx0!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42f07c84-c74a-4f3e-9db0-1d51fd81dca9_1024x820.png 848w, https://substackcdn.com/image/fetch/$s_!pNx0!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42f07c84-c74a-4f3e-9db0-1d51fd81dca9_1024x820.png 1272w, https://substackcdn.com/image/fetch/$s_!pNx0!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42f07c84-c74a-4f3e-9db0-1d51fd81dca9_1024x820.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!pNx0!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42f07c84-c74a-4f3e-9db0-1d51fd81dca9_1024x820.png" width="1024" height="820" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/42f07c84-c74a-4f3e-9db0-1d51fd81dca9_1024x820.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:820,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!pNx0!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42f07c84-c74a-4f3e-9db0-1d51fd81dca9_1024x820.png 424w, https://substackcdn.com/image/fetch/$s_!pNx0!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42f07c84-c74a-4f3e-9db0-1d51fd81dca9_1024x820.png 848w, https://substackcdn.com/image/fetch/$s_!pNx0!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42f07c84-c74a-4f3e-9db0-1d51fd81dca9_1024x820.png 1272w, https://substackcdn.com/image/fetch/$s_!pNx0!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42f07c84-c74a-4f3e-9db0-1d51fd81dca9_1024x820.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>This illustrates how software progress can depend on the scale of training compute &#8212; in this case, models trained with more compute see larger efficiency gains (and also larger capability gains).</p><p>Now the line could also become less steep, and it doesn&#8217;t have to be a straight line &#8212; maybe there are some diminishing returns to log(training compute), for example. But the important thing is that the size of the compute efficiency gain could change with scale.</p><p>So how might this explain the really fast estimates of software progress? The idea is that software improvements may have made the line steeper, and since we&#8217;ve been scaling up training compute, we take advantage of these larger gains at scale. So when we measure software progress, this shows up as rapid efficiency gains.</p><p>The most direct evidence for this is the paper <em><a href="https://arxiv.org/abs/2511.21622">On the Origin of Algorithmic Progress in AI</a></em>.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-16" href="#footnote-16" target="_self">16</a> Here&#8217;s one example where they compare the scaling of a modern Transformer architecture (purple) and an old-school LSTM (green):</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!WGff!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc3eaa8d6-094c-4fd4-ae12-c303dc40ad35_1272x1074.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!WGff!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc3eaa8d6-094c-4fd4-ae12-c303dc40ad35_1272x1074.png 424w, https://substackcdn.com/image/fetch/$s_!WGff!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc3eaa8d6-094c-4fd4-ae12-c303dc40ad35_1272x1074.png 848w, https://substackcdn.com/image/fetch/$s_!WGff!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc3eaa8d6-094c-4fd4-ae12-c303dc40ad35_1272x1074.png 1272w, https://substackcdn.com/image/fetch/$s_!WGff!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc3eaa8d6-094c-4fd4-ae12-c303dc40ad35_1272x1074.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!WGff!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc3eaa8d6-094c-4fd4-ae12-c303dc40ad35_1272x1074.png" width="1272" height="1074" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c3eaa8d6-094c-4fd4-ae12-c303dc40ad35_1272x1074.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1074,&quot;width&quot;:1272,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!WGff!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc3eaa8d6-094c-4fd4-ae12-c303dc40ad35_1272x1074.png 424w, https://substackcdn.com/image/fetch/$s_!WGff!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc3eaa8d6-094c-4fd4-ae12-c303dc40ad35_1272x1074.png 848w, https://substackcdn.com/image/fetch/$s_!WGff!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc3eaa8d6-094c-4fd4-ae12-c303dc40ad35_1272x1074.png 1272w, https://substackcdn.com/image/fetch/$s_!WGff!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc3eaa8d6-094c-4fd4-ae12-c303dc40ad35_1272x1074.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>This meshes pretty well with the stylized stick-figure diagrams that I&#8217;ve been showing you. The main difference is that the y-axis is flipped, where a lower &#8220;validation loss&#8221; means a higher capability. We can see that the Transformer has a steeper scaling slope, and the efficiency gains grow with training compute scale. At 10<sup>15</sup> FLOP the efficiency gain is 6.3&#215;, whereas at 3 &#215; 10<sup>16</sup> FLOP the efficiency gain is 26&#215;.</p><p>And these gains grow <em>a lot</em> as you scale up training compute &#8212; if you naively extrapolated the purple line to 10<sup>23</sup> FLOP the total efficiency gain is over 20,000&#215;! If you work through the numbers more carefully, this works out to an annual improvement of 2.23&#215; per year at 10<sup>23</sup> FLOP scales &#8212; so it&#8217;s easier to see how you can measure compute efficiency gains of several times per year.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-17" href="#footnote-17" target="_self">17</a></p><p>To my mind, these results lend quite a bit of weight toward the &#8220;scale-dependence&#8221; argument for how measured software progress has been so fast. I also think it has some intuitive appeal to it, because there&#8217;s been evidence for a while now that some AI architectures <a href="https://arxiv.org/abs/2106.09488">scale better</a> <a href="https://arxiv.org/abs/2207.10551">than others</a>.</p><p>But I don&#8217;t think this is really a slam-dunk argument either. For starters, I don&#8217;t want to over-index on one paper. The estimate of 2.23&#215; per year is also substantially lower than my 10&#215; per year best guess earlier, though this could change if the authors look at a broader set of innovations, or if they consider higher compute scales.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-18" href="#footnote-18" target="_self">18</a> There are also some suspicious aspects to this study: for example, it uses scaling experiments between 10<sup>13</sup> FLOP and 10<sup>18</sup> FLOP (or under 10<sup>17</sup> FLOP for LSTMs) to infer efficiency improvements as far as 10<sup>23</sup> FLOP &#8212; so it involves heroically extrapolating out five orders of magnitude. This is a problem because the result about scale-dependence might itself depend on which compute scale we&#8217;re looking at, which is very meta.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-19" href="#footnote-19" target="_self">19</a> Ideally, it&#8217;d be great to do something like this experiment but at larger compute scales, and also experimentally accounting for the gains from better data and post-training. Both of these are things that the paper glosses over.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-20" href="#footnote-20" target="_self">20</a></p><p>Overall, I&#8217;d say that both data quality improvements and scale-dependence play an important role &#8212; it&#8217;s not clear to me that the numbers totally add up with either approach. And each approach fails to explain certain observations &#8212; for example, capabilities have improved in the last year <a href="https://epoch.ai/gradient-updates/why-gpt5-used-less-training-compute-than-gpt45-but-gpt6-probably-wont">without much training compute growth</a>. I think this is probably better explained by data quality improvements than scale-dependent innovations (though these aren&#8217;t mutually exclusive, since better data could itself change the scaling slope).</p><p>This means that almost all existing estimates of software progress were misleading. Many efficiency gains that we thought were from algorithmic improvements actually came from better data. And many efficiency gains that we thought came from continually inventing lots of new algorithms could&#8217;ve come from scaling up a tiny handful of algorithmic innovations!</p><p>In fact, this seems to be the case in practice, if you break down the 20,000&#215; gains between the LSTM and Transformer that we saw earlier:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!2PyT!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e2f2a9b-d4dd-4f7c-bce7-17c308350a96_1600x460.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!2PyT!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e2f2a9b-d4dd-4f7c-bce7-17c308350a96_1600x460.png 424w, https://substackcdn.com/image/fetch/$s_!2PyT!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e2f2a9b-d4dd-4f7c-bce7-17c308350a96_1600x460.png 848w, https://substackcdn.com/image/fetch/$s_!2PyT!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e2f2a9b-d4dd-4f7c-bce7-17c308350a96_1600x460.png 1272w, https://substackcdn.com/image/fetch/$s_!2PyT!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e2f2a9b-d4dd-4f7c-bce7-17c308350a96_1600x460.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!2PyT!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e2f2a9b-d4dd-4f7c-bce7-17c308350a96_1600x460.png" width="1456" height="419" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2e2f2a9b-d4dd-4f7c-bce7-17c308350a96_1600x460.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:419,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!2PyT!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e2f2a9b-d4dd-4f7c-bce7-17c308350a96_1600x460.png 424w, https://substackcdn.com/image/fetch/$s_!2PyT!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e2f2a9b-d4dd-4f7c-bce7-17c308350a96_1600x460.png 848w, https://substackcdn.com/image/fetch/$s_!2PyT!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e2f2a9b-d4dd-4f7c-bce7-17c308350a96_1600x460.png 1272w, https://substackcdn.com/image/fetch/$s_!2PyT!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e2f2a9b-d4dd-4f7c-bce7-17c308350a96_1600x460.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Note that the <a href="https://arxiv.org/abs/2511.21622v1">paper</a> in question uses the Deep Learning Era trendline from Epoch&#8217;s &#8220;<a href="https://epoch.ai/data/ai-models">Notable AI models</a>&#8221; dataset as the frontier of training compute. This is why the &#8220;compute frontier&#8221; is close to 10<sup>23</sup> FLOP in 2024-2025.</em></figcaption></figure></div><p>Here you can see that <em>most efficiency gains came from two scale-dependent innovations</em> &#8212; (1) the shift from LSTMs to the <a href="https://arxiv.org/abs/2001.08361">Kaplan</a> <a href="https://arxiv.org/abs/2005.14165">Transformer</a>, and (2) &#8220;Chinchilla Rebalancing&#8221;.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-21" href="#footnote-21" target="_self">21</a> Other innovations seem scale-independent and also contribute very little overall &#8212; they barely shift the curve to the left and don&#8217;t change the slope. So even if we measure fast software progress, it might just be because we&#8217;re scaling quickly &#8212; it may not have much to do with making further innovations.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://epochai.substack.com/p/the-least-understood-driver-of-ai?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://epochai.substack.com/p/the-least-understood-driver-of-ai?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><h2>IV. How this impacts the software intelligence explosion debate</h2><p>Everything we&#8217;ve seen so far bears on one of the most important debates in AI &#8212; namely, how likely is a <a href="https://www.forethought.org/research/three-types-of-intelligence-explosion">software intelligence explosion</a>? The idea is that once you automate AI research,<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-22" href="#footnote-22" target="_self">22</a> you&#8217;ll get a rapid feedback loop: the AIs get smarter (or more numerous), their research efforts drive more AI software progress, making the AIs even smarter (or more numerous) still, and so on.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-23" href="#footnote-23" target="_self">23</a></p><p>This matters a lot because if it&#8217;s true, it could engender an enormous amount of AI and societal change in the span of years. What does &#8220;enormous&#8221; mean? Well opinions on this vary a lot, but they often include things like <a href="https://www.darioamodei.com/essay/machines-of-loving-grace">eliminating most cancers</a>, <a href="https://www.forethought.org/research/preparing-for-the-intelligence-explosion">compressing a century of technological progress into a decade</a>, or <a href="https://ai-2027.com/summary">the extinction of humanity</a>. Saying that the stakes are high would be the understatement of the century.</p><p>So how do the results we&#8217;ve seen &#8212; about the rate of software progress and its drivers &#8212; impact this debate? The short answer is that some <a href="https://epoch.ai/blog/do-the-returns-to-software-rnd-point-towards-a-singularity">existing</a> <a href="https://epoch.ai/gradient-updates/the-software-intelligence-explosion-debate-needs-experiments">analyses</a> of the software intelligence explosion are based on overly conservative estimates of software progress, and essentially all of them are based on faulty assumptions.</p><p>To see why, consider <a href="https://epoch.ai/gradient-updates/the-software-intelligence-explosion-debate-needs-experiments#:~:text=Prior%20work%20has%20used%20the%20%E2%80%9CJones%20model%E2%80%9D%2C%20which%20tells%20us%20how%20R%26D%20inputs%20translate%20into%20outputs.%20For%20example%2C%20the%20R%26D%20input%20might%20be%20the%20number%20of%20AI%20researchers%2C%20and%20the%20R%26D%20output%20could%20measure%20how%20efficiently%20an%20AI%20algorithm%20uses%20training%20compute%20(%E2%80%9Csoftware%20efficiency%E2%80%9D).2">the</a> <a href="https://epoch.ai/epoch-after-hours/economics-of-ai#:~:text=For%20example%2C%20the,also%20increase%20dramatically.">model</a> that people usually use to study the software intelligence explosion, which relates AI research effort to AI software progress. The core parameter that relates these two is the &#8220;returns to AI software R&amp;D&#8221;, which tells you how well you can turn more research effort into more software progress.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-24" href="#footnote-24" target="_self">24</a></p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!bWtC!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6cebce4f-a2a0-4cf2-9d25-c62b6dfee96f_1500x240.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!bWtC!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6cebce4f-a2a0-4cf2-9d25-c62b6dfee96f_1500x240.png 424w, https://substackcdn.com/image/fetch/$s_!bWtC!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6cebce4f-a2a0-4cf2-9d25-c62b6dfee96f_1500x240.png 848w, https://substackcdn.com/image/fetch/$s_!bWtC!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6cebce4f-a2a0-4cf2-9d25-c62b6dfee96f_1500x240.png 1272w, https://substackcdn.com/image/fetch/$s_!bWtC!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6cebce4f-a2a0-4cf2-9d25-c62b6dfee96f_1500x240.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!bWtC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6cebce4f-a2a0-4cf2-9d25-c62b6dfee96f_1500x240.png" width="1456" height="233" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6cebce4f-a2a0-4cf2-9d25-c62b6dfee96f_1500x240.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:233,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!bWtC!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6cebce4f-a2a0-4cf2-9d25-c62b6dfee96f_1500x240.png 424w, https://substackcdn.com/image/fetch/$s_!bWtC!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6cebce4f-a2a0-4cf2-9d25-c62b6dfee96f_1500x240.png 848w, https://substackcdn.com/image/fetch/$s_!bWtC!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6cebce4f-a2a0-4cf2-9d25-c62b6dfee96f_1500x240.png 1272w, https://substackcdn.com/image/fetch/$s_!bWtC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6cebce4f-a2a0-4cf2-9d25-c62b6dfee96f_1500x240.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>This parameter has a key threshold: if it&#8217;s greater than 1, then you get a software intelligence explosion. And in practice, <a href="https://epoch.ai/blog/do-the-returns-to-software-rnd-point-towards-a-singularity">literature</a> <a href="https://epoch.ai/gradient-updates/the-software-intelligence-explosion-debate-needs-experiments">estimates</a> suggest that might be the case (though it&#8217;s really uncertain):</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!eeQL!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6d9fb9c5-4ce5-40e1-8053-11e29e9c9e70_1024x1280.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!eeQL!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6d9fb9c5-4ce5-40e1-8053-11e29e9c9e70_1024x1280.png 424w, https://substackcdn.com/image/fetch/$s_!eeQL!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6d9fb9c5-4ce5-40e1-8053-11e29e9c9e70_1024x1280.png 848w, https://substackcdn.com/image/fetch/$s_!eeQL!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6d9fb9c5-4ce5-40e1-8053-11e29e9c9e70_1024x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!eeQL!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6d9fb9c5-4ce5-40e1-8053-11e29e9c9e70_1024x1280.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!eeQL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6d9fb9c5-4ce5-40e1-8053-11e29e9c9e70_1024x1280.png" width="1024" height="1280" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6d9fb9c5-4ce5-40e1-8053-11e29e9c9e70_1024x1280.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1280,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!eeQL!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6d9fb9c5-4ce5-40e1-8053-11e29e9c9e70_1024x1280.png 424w, https://substackcdn.com/image/fetch/$s_!eeQL!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6d9fb9c5-4ce5-40e1-8053-11e29e9c9e70_1024x1280.png 848w, https://substackcdn.com/image/fetch/$s_!eeQL!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6d9fb9c5-4ce5-40e1-8053-11e29e9c9e70_1024x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!eeQL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6d9fb9c5-4ce5-40e1-8053-11e29e9c9e70_1024x1280.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>(<a href="https://epoch.ai/gradient-updates/the-software-intelligence-explosion-debate-needs-experiments">Source</a>) Estimates of the returns to AI software R&amp;D are very uncertain but straddle 1. This suggests that an intelligence explosion is plausible, but it&#8217;s also plausible that the feedback loop of AIs improving themselves simply fizzles out.</em></figcaption></figure></div><p>However, as <a href="https://www.forethought.org/research/will-ai-r-and-d-automation-cause-a-software-intelligence-explosion#you-might-need-fast-growing-computing-power-to-discover-better-algorithms">pointed out</a> by Daniel Eth and Tom Davidson, these estimates rely on an overly conservative estimate of software progress of 3&#215; per year (this is the <a href="https://epoch.ai/blog/algorithmic-progress-in-language-models">Ho et al. 2024</a> estimate that we saw earlier). I now think that software progress is closer to 10&#215; per year, so this change should increase the estimates of this parameter several-fold,<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-25" href="#footnote-25" target="_self">25</a> making a software intelligence explosion seem much more likely.</p><p>But there&#8217;s also a good reason to think that these are <em>drastic</em> <em>overestimates</em>. That&#8217;s because existing models of the software intelligence explosion don&#8217;t account for how a lot of software progress could come from scaling training compute. Recall how scaling up training compute for just two innovations led to most observed efficiency gains between LSTMs and Modern Transformers, rather than humans discovering tons of new innovations.</p><p>Instead, existing models either assume that research effort is the <a href="https://epoch.ai/gradient-updates/the-software-intelligence-explosion-debate-needs-experiments">only</a> <a href="https://epoch.ai/blog/do-the-returns-to-software-rnd-point-towards-a-singularity">thing</a> that matters, or assume that it depends on <a href="https://epoch.ai/gradient-updates/the-software-intelligence-explosion-debate-needs-experiments#:~:text=Proxy%202%3A%20Some%20function%20of%20cognitive%20labor%20and%20experimental%20compute">a</a> <a href="https://www.forethought.org/research/will-ai-r-and-d-automation-cause-a-software-intelligence-explosion#you-might-need-fast-growing-computing-power-to-discover-better-algorithms">mix</a> <a href="https://www.aifuturesmodel.com/#section-experimentthroughput">of</a> research effort and the compute used to run experiments.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-26" href="#footnote-26" target="_self">26</a> But to my knowledge, none explicitly account for training compute scaling being a <em>source</em> of software progress, so they could heavily overstate the importance of research effort.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!DAM9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f16ae81-e068-4067-98b4-d316a164b65a_1500x858.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!DAM9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f16ae81-e068-4067-98b4-d316a164b65a_1500x858.png 424w, https://substackcdn.com/image/fetch/$s_!DAM9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f16ae81-e068-4067-98b4-d316a164b65a_1500x858.png 848w, https://substackcdn.com/image/fetch/$s_!DAM9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f16ae81-e068-4067-98b4-d316a164b65a_1500x858.png 1272w, https://substackcdn.com/image/fetch/$s_!DAM9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f16ae81-e068-4067-98b4-d316a164b65a_1500x858.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!DAM9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f16ae81-e068-4067-98b4-d316a164b65a_1500x858.png" width="1456" height="833" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1f16ae81-e068-4067-98b4-d316a164b65a_1500x858.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:833,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!DAM9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f16ae81-e068-4067-98b4-d316a164b65a_1500x858.png 424w, https://substackcdn.com/image/fetch/$s_!DAM9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f16ae81-e068-4067-98b4-d316a164b65a_1500x858.png 848w, https://substackcdn.com/image/fetch/$s_!DAM9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f16ae81-e068-4067-98b4-d316a164b65a_1500x858.png 1272w, https://substackcdn.com/image/fetch/$s_!DAM9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f16ae81-e068-4067-98b4-d316a164b65a_1500x858.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In effect, this means that existing estimates overstate the returns to software R&amp;D, and makes the software intelligence explosion seem much less likely. One way to think of this is as a &#8220;compute bottleneck&#8221; &#8212; it&#8217;s hard to make super fast software progress without also scaling up training compute.</p><p>There are some reasons to expect that this training compute bottleneck can be overcome. For example, in principle there may be innovations that <em>really</em> boost performance at the same training compute, hence driving a lot of software progress. This could look like software innovations that scale really well, even if they don&#8217;t look great at small scales:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!bWik!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6a73939c-559a-450f-a5be-f9a4c9c9d2be_1025x820.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!bWik!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6a73939c-559a-450f-a5be-f9a4c9c9d2be_1025x820.png 424w, https://substackcdn.com/image/fetch/$s_!bWik!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6a73939c-559a-450f-a5be-f9a4c9c9d2be_1025x820.png 848w, https://substackcdn.com/image/fetch/$s_!bWik!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6a73939c-559a-450f-a5be-f9a4c9c9d2be_1025x820.png 1272w, https://substackcdn.com/image/fetch/$s_!bWik!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6a73939c-559a-450f-a5be-f9a4c9c9d2be_1025x820.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!bWik!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6a73939c-559a-450f-a5be-f9a4c9c9d2be_1025x820.png" width="1025" height="820" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6a73939c-559a-450f-a5be-f9a4c9c9d2be_1025x820.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:820,&quot;width&quot;:1025,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!bWik!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6a73939c-559a-450f-a5be-f9a4c9c9d2be_1025x820.png 424w, https://substackcdn.com/image/fetch/$s_!bWik!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6a73939c-559a-450f-a5be-f9a4c9c9d2be_1025x820.png 848w, https://substackcdn.com/image/fetch/$s_!bWik!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6a73939c-559a-450f-a5be-f9a4c9c9d2be_1025x820.png 1272w, https://substackcdn.com/image/fetch/$s_!bWik!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6a73939c-559a-450f-a5be-f9a4c9c9d2be_1025x820.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Alternatively, you could in principle find a ton of scale-independent innovations with big capability gains:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!kqLZ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F167a2aac-d92e-4f07-918b-b46690a8f2d9_1024x820.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!kqLZ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F167a2aac-d92e-4f07-918b-b46690a8f2d9_1024x820.png 424w, https://substackcdn.com/image/fetch/$s_!kqLZ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F167a2aac-d92e-4f07-918b-b46690a8f2d9_1024x820.png 848w, https://substackcdn.com/image/fetch/$s_!kqLZ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F167a2aac-d92e-4f07-918b-b46690a8f2d9_1024x820.png 1272w, https://substackcdn.com/image/fetch/$s_!kqLZ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F167a2aac-d92e-4f07-918b-b46690a8f2d9_1024x820.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!kqLZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F167a2aac-d92e-4f07-918b-b46690a8f2d9_1024x820.png" width="1024" height="820" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/167a2aac-d92e-4f07-918b-b46690a8f2d9_1024x820.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:820,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!kqLZ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F167a2aac-d92e-4f07-918b-b46690a8f2d9_1024x820.png 424w, https://substackcdn.com/image/fetch/$s_!kqLZ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F167a2aac-d92e-4f07-918b-b46690a8f2d9_1024x820.png 848w, https://substackcdn.com/image/fetch/$s_!kqLZ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F167a2aac-d92e-4f07-918b-b46690a8f2d9_1024x820.png 1272w, https://substackcdn.com/image/fetch/$s_!kqLZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F167a2aac-d92e-4f07-918b-b46690a8f2d9_1024x820.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>And of course you could get huge capability gains from a combination of scale-dependent and scale-independent improvements. If doing these things is easy, then you could skyrocket capabilities without increasing training compute very much.</p><p>My personal inclination is to say that these things are probably quite hard &#8212; otherwise, why haven&#8217;t we found these innovations already? But I can&#8217;t theoretically rule it out, and you could always say &#8220;maybe <em>you</em> can&#8217;t find these innovations, but if you had a country of geniuses in a data center, they certainly would! Not to mention what would happen if AGI is as smart as Von Neumann+++&#8230;&#8221;.</p><p>Moreover, the strength of training compute bottlenecks depends on the &#8220;scale-dependence&#8221; explanation for rapid AI software progress. But it may apply less well to software progress that&#8217;s driven by data quality improvements &#8212; perhaps you could have a software intelligence explosion just by building better RL environments! I don&#8217;t know to what extent this is possible, and in principle data quality innovations could also be scale-dependent, but I think it&#8217;s still a very important consideration.</p><p>There&#8217;s a lot more to the debate about the software intelligence explosion and its bottlenecks which is too long to get into here. But overall I think the upshot is this: some prior work on the software intelligence explosion is based on outdated estimates of software progress, which make an &#8220;explosion&#8221; seem more likely. On the other hand, essentially all existing work is also based on a heavily flawed model, making the &#8220;explosion&#8221; seem less likely. Personally I now think the software intelligence explosion is less likely than before looking into this, though I also think that the bottlenecks aren&#8217;t strong enough to preclude it altogether.</p><h2>V. Conclusion</h2><p>Let&#8217;s recap all the main things that we&#8217;ve seen:</p><ul><li><p>Software improvements are about reducing the training compute you need to get to the same capability, or doing more with the training compute you have.</p></li><li><p>It progresses very quickly, with efficiency gains increasing several times per year. But these estimates have gigantic error bars, due to data limitations and statistical assumptions.</p></li><li><p>The numbers might also be somewhat misleading. Much of the measured progress is plausibly driven by data quality improvements rather than algorithmic innovations. And it&#8217;s likely that a large share of measured efficiency gains reflects scaling up a small handful of scale-dependent innovations, rather than the continual discovery of new algorithms.</p></li><li><p>If most efficiency improvements came from a small handful of scale-dependent innovations, then existing models of the software intelligence explosion may be flawed. Correcting for underestimated software progress makes an explosion seem more likely, but accounting for training compute bottlenecks from scale-dependent innovations pushes in the other direction.</p></li></ul><p>Given all of this background, we finally have enough ammunition to answer the three questions I raised at the start of the essay:</p><blockquote><p><strong>Q: How did DeepSeek seem to catch up to OpenAI&#8217;s o1 within months, despite being trained with less compute?</strong></p><p>A: If software really improves several times a year, it&#8217;s less of a mystery why DeepSeek was able to get close to o1 after a few months, <a href="https://www.anthropic.com/news/detecting-and-preventing-distillation-attacks">especially if there was a lot of distillation involved</a>.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-27" href="#footnote-27" target="_self">27</a> Peter Wildeford describes this in more detail <a href="https://peterwildeford.substack.com/p/ten-takes-on-deepseek">here</a>.</p><p><strong>Q: When will the world develop AGI?</strong></p><p>A: No one knows for sure, but it&#8217;s likely that software progress matters tremendously. For example, one of the most prominent models of AGI timelines is Ajeya Cotra&#8217;s <a href="https://www.alignmentforum.org/posts/KrJfoZzpSDpnrv9va/draft-report-on-ai-timelines">bioanchors model</a>, which originally had a median timeline of around 2050. But as Scott Alexander <a href="https://www.astralcodexten.com/p/what-happened-with-bio-anchors">points out</a>, this may have been over a decade too long because Cotra underestimated AI software progress (Cotra&#8217;s most recent estimate is <a href="https://80000hours.org/podcast/episodes/ajeya-cotra-transformative-ai-crunch-time/#:~:text=Ajeya%E2%80%99s%20timeline%3A%20AI%20R%26D%20automation%20in%20the%20early%202030s%2C%20then%20things%20move%20fast">probably somewhere in the 2030s</a>).</p><p><strong>Q: If we automate AI research, will AI progress accelerate like crazy a la <a href="https://situational-awareness.ai/">Situational Awareness</a> and <a href="https://ai-2027.com/">AI 2027</a>?</strong></p><p>A: The answer is essentially Section IV. That is, nobody really knows, but you can form your opinion on it by answering two subquestions: (1) &#8220;What are the returns to AI software R&amp;D?&#8221;, and (2) &#8220;Will compute be a major bottleneck to the software intelligence explosion?&#8221;. I lean towards the answer being &#8220;no&#8221; because I&#8217;d guess compute bottlenecks will be strong, but I think there&#8217;s quite a lot of room for disagreement.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-28" href="#footnote-28" target="_self">28</a></p></blockquote><p>I hope these questions also help reiterate why AI software progress is such a big deal. These are foundational questions about competition between AI firms, and what happens to the world if we develop what would be the most influential technology in history. And the answers to them can be enormously swayed by the speed and nature of software progress &#8212; the problem is that we understand it so poorly that we can&#8217;t narrow down our uncertainty.</p><p>So what should we do? As a starting point, I want people to truly appreciate just how important software progress is. After that, we can try to do the best we can to understand it better. Perhaps that means running ablation experiments at larger compute scales, seeing what happens to software progress if researchers are compute-bottlenecked, or estimating software progress in post-training. And maybe some AI researchers already have good answers to some of this stuff, what do I know?</p><p>I think it&#8217;s going to be very hard to reduce our uncertainty, but at least in my mind, this topic is important enough that it&#8217;s worth giving it a shot.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://epochai.substack.com/p/the-least-understood-driver-of-ai?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://epochai.substack.com/p/the-least-understood-driver-of-ai?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://epochai.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading! Subscribe to receive the latest from Epoch AI.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p><em>I&#8217;d like to thank Aaron Scher, JS Denain, Josh You, Luke Emberson, Jaime Sevilla, David Owen, Parker Whitfill, Hans Gundlach, Brendan Halstead, Joshua Turner, Eli Lifland, Alex Fogelson, and Lynette Bye for helpful feedback and support. Special thanks to Aaron inspiring me to write this post in the first place.</em></p><h2>Appendix: Estimates of software progress</h2><p>The table below shows all the estimates used to generate the bar chart at the beginning of section II, as well as some additional estimates from post-training.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-29" href="#footnote-29" target="_self">29</a></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!BxFp!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fccfca31e-cb7a-4e06-80a8-955e63c5447e_1438x1504.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!BxFp!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fccfca31e-cb7a-4e06-80a8-955e63c5447e_1438x1504.png 424w, https://substackcdn.com/image/fetch/$s_!BxFp!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fccfca31e-cb7a-4e06-80a8-955e63c5447e_1438x1504.png 848w, https://substackcdn.com/image/fetch/$s_!BxFp!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fccfca31e-cb7a-4e06-80a8-955e63c5447e_1438x1504.png 1272w, https://substackcdn.com/image/fetch/$s_!BxFp!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fccfca31e-cb7a-4e06-80a8-955e63c5447e_1438x1504.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!BxFp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fccfca31e-cb7a-4e06-80a8-955e63c5447e_1438x1504.png" width="1438" height="1504" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ccfca31e-cb7a-4e06-80a8-955e63c5447e_1438x1504.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1504,&quot;width&quot;:1438,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:335092,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://epochai.substack.com/i/189189605?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fccfca31e-cb7a-4e06-80a8-955e63c5447e_1438x1504.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!BxFp!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fccfca31e-cb7a-4e06-80a8-955e63c5447e_1438x1504.png 424w, https://substackcdn.com/image/fetch/$s_!BxFp!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fccfca31e-cb7a-4e06-80a8-955e63c5447e_1438x1504.png 848w, https://substackcdn.com/image/fetch/$s_!BxFp!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fccfca31e-cb7a-4e06-80a8-955e63c5447e_1438x1504.png 1272w, https://substackcdn.com/image/fetch/$s_!BxFp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fccfca31e-cb7a-4e06-80a8-955e63c5447e_1438x1504.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>Sometimes people call this &#8220;<a href="https://epoch.ai/blog/algorithmic-progress-in-language-models">algorithmic progress</a>&#8221; but I want to account for the possibility that these improvements are due to better data.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>There are other ways to define &#8220;software progress&#8221; besides looking at training compute, which I&#8217;ll get to later in the post.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>Here I&#8217;m just using &#8220;capabilities&#8221; as an abstract concept to illustrate what I mean in a simple way. But if you want something more concrete, you can consider the <a href="https://epoch.ai/benchmarks/eci">Epoch Capabilities Index</a>, or log(<a href="https://metr.org/blog/2025-03-19-measuring-ai-ability-to-complete-long-tasks/">METR time horizon</a>). You could also consider something like benchmark performance, but then the assumption of a straight-line relationship doesn&#8217;t make that much sense there.</p><p>Another note: in principle it&#8217;s also possible that some capabilities can&#8217;t be accessed with infinite training compute &#8212; this is just one of the (many) assumptions of this simplified picture.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>This definition of effective compute breaks down somewhat if you include distilled models &#8212; that&#8217;s because the compute efficiency gain largely comes from being able to use data from an existing larger model. In particular, if you&#8217;ve only trained models up to 10<sup>25</sup> FLOP, distillation would shift the curve left, but it doesn&#8217;t allow you to access higher capabilities with the same 10<sup>25</sup> FLOP.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>I&#8217;ll primarily focus on LLMs for this post, because these are the most prominent AI systems (e.g. ChatGPT probably has close to a <a href="https://epochai.substack.com/p/the-changing-drivers-of-llm-adoption">1 billion weekly active users</a> by now), and are the main kind of model that are pushing towards &#8220;AGI&#8221;.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>Strictly speaking, <a href="https://epoch.ai/blog/algorithmic-progress-in-language-models">Ho et al. 2024</a> uses data on the parameters and data rather than the training compute (which is related to the product of the two).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>It&#8217;s also hard to get data about counterfactuals just from observational data &#8212; for example, it doesn&#8217;t really tell us what happens if we scale up GPT-4 by ten times, using the techniques available at the time. We could try to figure this out by running experiments, but that&#8217;s a different approach altogether.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-8" href="#footnote-anchor-8" class="footnote-number" contenteditable="false" target="_self">8</a><div class="footnote-content"><p>Specifically we consider a narrow <em>range</em> of capabilities rather than a single capability level, otherwise we wouldn&#8217;t find any models at all!</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-9" href="#footnote-anchor-9" class="footnote-number" contenteditable="false" target="_self">9</a><div class="footnote-content"><p>The 30,000&#215; number comes from the &#8220;Results table&#8221; in the <a href="https://www.lesswrong.com/posts/yXLqrpfFwBW5knpgc/catch-up-algorithmic-progress-might-actually-be-60-per-year">post</a>, with a capability threshold of 45. The 5<sup>12</sup>&#215; per year improvement arises due to the following problem: GPT-5 was released about a month after Grok 4 and used around 5&#215; less training compute, while achieving similar performance, implying a rate of 5<sup>12</sup>&#215; per year. But it&#8217;s not clear that this actually reflects genuine software progress.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-10" href="#footnote-anchor-10" class="footnote-number" contenteditable="false" target="_self">10</a><div class="footnote-content"><p><a href="https://arxiv.org/abs/2511.19492">Whitfill, Snodin, and Becker 2025</a>, <a href="https://arxiv.org/abs/2512.00193">Ho et al. 2025</a>, and <a href="https://www.lesswrong.com/posts/yXLqrpfFwBW5knpgc/catch-up-algorithmic-progress-might-actually-be-60-per-year">Scher 2025</a> include some gains from post-training, whereas <a href="https://epoch.ai/blog/algorithmic-progress-in-language-models">Ho et al. 2024</a> explicitly only considers pre-training. In practice, all papers may reflect pre-training to a large extent, because there is little public data on post-training compute, which is needed to properly estimate post-training software progress.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-11" href="#footnote-anchor-11" class="footnote-number" contenteditable="false" target="_self">11</a><div class="footnote-content"><p>The main estimate I know of is one from Anthropic&#8217;s <a href="https://www-cdn.anthropic.com/872c653b2d0501d6ab44cf87f43e1dc4853e4d37.pdf#page=20.43">Responsible Scaling Policy</a>, which incidentally is also 3&#215; per year. Sadly this number doesn&#8217;t look like it&#8217;s meant to be taken seriously &#8212; Anthropic describes it as an &#8220;informal estimate&#8221;, and gives no reasoning for it.</p><p>There is some other work that looks at the gains from individual post-training innovations rather than a trend. For example, one <a href="https://epoch.ai/blog/ai-capabilities-can-be-significantly-improved-without-expensive-retraining">analysis</a> looks at the efficiency gain from a range of post-training innovations, and finds that they are usually between 5&#215; and 30&#215; on relevant benchmarks. In some cases, the software efficiency gain can even go up to 100&#215;! But I think this is likely very specific to the task, and is also outdated since the paper was written before reasoning models were a thing.</p><p>More recently, Arden Berg and I <a href="https://epoch.ai/gradient-updates/quantifying-the-algorithmic-improvement-from-reasoning-models">looked</a> specifically at the software efficiency gain from early reasoning models, such as o1 and DeepSeek-R1. After some detective work, we estimated that you&#8217;d need around 10&#215; more pre-training compute to match the same performance without RL training, at least on benchmarks like GPQA diamond and MATH Level 5. In practice this&#8217;ll depend on how much RL training you do and the amount of inference scaling, the domain, and in any case it doesn&#8217;t give us a trend in the software efficiency over time.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-12" href="#footnote-anchor-12" class="footnote-number" contenteditable="false" target="_self">12</a><div class="footnote-content"><p>Incidentally, this is one of the reasons that Ajeya Cotra shifts her estimate of software progress downward in her <a href="https://niplav.site/doc/cs/ai/alignment/policy/forecasting/forecasting_tai_cotra_2020.pdf#page=125.55">bioanchors report</a>, and Tom Davidson employs a similar argument in his work on <a href="https://coefficientgiving.org/research/what-a-compute-centric-framework-says-about-takeoff-speeds/#:~:text=There%E2%80%99s%20massive%20uncertainty,speed%20up%20takeoff.">takeoff speeds</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-13" href="#footnote-anchor-13" class="footnote-number" contenteditable="false" target="_self">13</a><div class="footnote-content"><p>This depends on how exactly we <a href="https://epoch.ai/gradient-updates/three-issues-undermining-compute-based-ai-policies">measure &#8220;training compute&#8221;</a>. For example, both model distillation and synthetic data generation often rely on using a larger model to help develop a smaller one. But typically when we count training compute, we focus only on the small model, which neglects the compute needed to train the large model. When we then use these (understated) training compute estimates to measure software progress, we then pick up a big efficiency gain.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-14" href="#footnote-anchor-14" class="footnote-number" contenteditable="false" target="_self">14</a><div class="footnote-content"><p>This is probably less relevant for frontier models, though I&#8217;m not sure that&#8217;s universally true. For example, I think there&#8217;s some chance that o3 was in part distilled from GPT-4.5.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-15" href="#footnote-anchor-15" class="footnote-number" contenteditable="false" target="_self">15</a><div class="footnote-content"><p>Interestingly enough, my coauthors and I <a href="https://epoch.ai/blog/algorithmic-progress-in-language-models">tried decomposing software progress in language models</a> and found that most of it comes from data efficiency improvements (as opposed to parameter efficiency), which is consistent with this hypothesis. But I think this is pretty weak evidence because the confidence intervals are super wide, and sometimes the fit would suggest most gains were from parameter efficiency instead, after making small tweaks to the optimization setup. The overall compute efficiency result seemed a lot more stable though.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-16" href="#footnote-anchor-16" class="footnote-number" contenteditable="false" target="_self">16</a><div class="footnote-content"><p>Another example that shows scale-dependence of software progress in RL is Figure 2 in <a href="https://arxiv.org/pdf/2510.13786">Khatri et al. 2025</a>, which shows how improvements to RL algorithms change the slope of training curves.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-17" href="#footnote-anchor-17" class="footnote-number" contenteditable="false" target="_self">17</a><div class="footnote-content"><p>Unfortunately, existing estimates just don&#8217;t really account for this. The &#8220;fancy statistics&#8221; approaches tend to assume away scale dependence, and the &#8220;lines on a graph&#8221; approaches don&#8217;t have enough high-quality data to reliably identify it. I only know of <a href="https://arxiv.org/abs/2505.04075">two</a> <a href="https://www.lesswrong.com/posts/tJAD2LG9uweeEfjwq/estimating-efficiency-improvements-in-llm-pre-training">exceptions</a> to this, that try to account for efficiency contributions in a very granular way.</p><p>Note that the &#8220;lines on a graph&#8221; approach could try to measure this in principle, where you see faster software progress at larger scales than smaller scales. But besides being limited by limited noisy data, there&#8217;s an additional complication: at smaller scales you can do more experiments, and that pushes towards faster progress at small scales. This is perhaps a part of the reason why <a href="https://x.com/JasonBotterill/status/2017818802457874786">Claude Sonnet models can end up better than Opus</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-18" href="#footnote-anchor-18" class="footnote-number" contenteditable="false" target="_self">18</a><div class="footnote-content"><p>My current impression is that there&#8217;s no clean way to directly compare the numbers from existing estimates (which ignore scale-dependence) and the numbers from this paper &#8212; they&#8217;re measuring somewhat different things. The paper instead proposes a plausible justification for several-times-per-year estimates of software progress.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-19" href="#footnote-anchor-19" class="footnote-number" contenteditable="false" target="_self">19</a><div class="footnote-content"><p>Having spoken with the first-author of the paper (Hans Gundlach), my impression is that the project was pretty compute-constrained, hence the small-scale experiments. Somebody give them more compute!</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-20" href="#footnote-anchor-20" class="footnote-number" contenteditable="false" target="_self">20</a><div class="footnote-content"><p>Specifically, the paper doesn&#8217;t account for post-training improvements, and it uses a literature estimate for data improvements (rather than accounting for this experimentally).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-21" href="#footnote-anchor-21" class="footnote-number" contenteditable="false" target="_self">21</a><div class="footnote-content"><p>The latter refers to the shift from using <a href="https://arxiv.org/abs/2001.08361">Kaplan scaling laws</a> to <a href="https://arxiv.org/abs/2203.15556">Chinchilla scaling laws</a> in Transformers &#8212; Chinchilla uses a different ratio of training data and parameters to get higher performance with the same compute.</p><p>Interestingly, the estimate shown in the figure (from <a href="https://arxiv.org/abs/2511.21622">Gundlach et al. 2025</a>) estimates a higher compute efficiency gain shifting from Kaplan to Chinchilla scaling laws compared to <a href="https://epoch.ai/blog/algorithmic-progress-in-language-models">Ho et al. 2024</a>. I suspect this is because these were estimated in different ways &#8212; the former estimates the gain using the loss function / scaling law reported in <a href="https://arxiv.org/abs/2001.08361">Kaplan et al. 2020</a>, whereas the latter uses the <em>stated recommendations </em>for how to allocate training compute across parameters and data. In principle these should be the same, but perhaps there&#8217;s a difference between them in practice.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-22" href="#footnote-anchor-22" class="footnote-number" contenteditable="false" target="_self">22</a><div class="footnote-content"><p>I think the proponents of this view also think that AI will lead to substantial speedups well before AI research is fully automated, but the real hyperspeed feedback loop really comes when AI can fully do AI R&amp;D, removing the last human bottlenecks in the chain.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-23" href="#footnote-anchor-23" class="footnote-number" contenteditable="false" target="_self">23</a><div class="footnote-content"><p>Note that the case of more numerous AIs involves improvements in inference efficiency, which is separate from the training efficiency estimates I discussed earlier.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-24" href="#footnote-anchor-24" class="footnote-number" contenteditable="false" target="_self">24</a><div class="footnote-content"><p>Technically speaking, this parameter tells us how compute efficiency changes when you double cumulative research effort. If doubling cumulative research effort also doubles compute efficiency, then the returns to R&amp;D are 1. If it quadruples, then the returns are 2. In general, if the output increases by 2<sup>r</sup>, then the returns are r.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-25" href="#footnote-anchor-25" class="footnote-number" contenteditable="false" target="_self">25</a><div class="footnote-content"><p>One caveat is that I&#8217;m uncertain what the rate of software progress was prior to the recent rise in post-training. I think it&#8217;s possible that progress has sped up in recent years.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-26" href="#footnote-anchor-26" class="footnote-number" contenteditable="false" target="_self">26</a><div class="footnote-content"><p>Why include experiment compute at all? The idea is that if you don&#8217;t scale up your innovations, you don&#8217;t know how it&#8217;s going to work, so you need <a href="https://epoch.ai/data-insights/openai-compute-spend">a lot of compute</a> to do large-scale scaling experiments to test this out. This would thus pose a big compute bottleneck, and it seems to be backed up by claims from frontier lab AI researchers. For example, Alex Paino (who helped <a href="https://www.youtube.com/watch?v=6nJZopACRuQ&amp;t=1223s">pre-train GPT-4.5</a>) describes how lots of software innovations &#8220;look good at small scale, but don&#8217;t look good at large scale.&#8221; Relatedly, some innovations like <a href="https://www.forethought.org/research/could-advanced-ai-accelerate-the-pace-of-ai-progress-interviews-with-ai#:~:text=For%20example%2C%20Reinforcement%20Learning%20from%20Human%20Feedback%20may%20not%20improve%20GPT%2D2%20or%20GPT%2D1%2C%20even%20though%20it%20works%20with%20GPT%2D3%20or%20GPT%2D4.">RLHF</a> and <a href="https://www.latent.space/p/noam-brown#:~:text=happened%20earlier%2C%20but-,if%20you%20try%20to%20do%20the%20reasoning%20paradigm%20on%20top%20of%20GPT%2D2%2C%20I%20don%27t%20think%20it%20would%20have%20gotten%20you%20almost%20anything,-%E2%80%A6%20if%20you%20ask">reasoning training</a> plausibly would&#8217;ve been a waste of time with small models like GPT-2, but likely become very important at the scales of GPT-4 and GPT-5. The ideas for both of these could&#8217;ve come about (<a href="https://arxiv.org/abs/1706.03741">or actually did come about</a>) at smaller scales, but scaling up was essential to seeing if they would actually work.</p><p>But note that some argue that experiment compute bottlenecks haven&#8217;t actually been that strong historically, <a href="https://arxiv.org/abs/2507.10618">if you look at the numbers</a>. Moreover, you might expect AI systems to substantially reduce experiment compute costs because they have much better &#8220;<a href="https://blog.ai-futures.org/p/ai-futures-model-dec-2025-update#:~:text=Davidson%20and%20Houlden%20focuses,improvements%20in%20research%20taste.">research taste</a>&#8221;, further making it more likely that you&#8217;ll find innovations that scale super well with training compute.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-27" href="#footnote-anchor-27" class="footnote-number" contenteditable="false" target="_self">27</a><div class="footnote-content"><p>Though catch-up between US and Chinese models isn&#8217;t all about distillation either, as Nathan Lambert recently <a href="https://www.interconnects.ai/p/how-much-does-distillation-really">argued</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-28" href="#footnote-anchor-28" class="footnote-number" contenteditable="false" target="_self">28</a><div class="footnote-content"><p>It also depends on what &#8220;accelerate like crazy&#8221; means &#8212; I think AI progress would definitely speed up, but I&#8217;m more skeptical of something like &#8220;5 orders of magnitude of software progress in a year&#8221;. I&#8217;d probably still give it something like a 15% chance though!</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-29" href="#footnote-anchor-29" class="footnote-number" contenteditable="false" target="_self">29</a><div class="footnote-content"><p>Note that Parker Whitfill noticed a bug in the code for <a href="https://epoch.ai/blog/algorithmic-progress-in-language-models">Ho et al. 2024</a>, although this didn&#8217;t change the numbers very much. Moreover, the reported confidence interval for this study was calculated after using updated data, so it&#8217;s wider than the numbers shown in the original paper (although the original central estimate remains the same).</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[How persistent is the inference cost burden?]]></title><description><![CDATA[RL scaling might be better than it looks, and inference costs to reach a capability level fall fast]]></description><link>https://epochai.substack.com/p/how-persistent-is-the-inference-cost</link><guid isPermaLink="false">https://epochai.substack.com/p/how-persistent-is-the-inference-cost</guid><dc:creator><![CDATA[JSD]]></dc:creator><pubDate>Mon, 16 Feb 2026 16:47:19 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!CPij!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6fb5830e-a8be-471c-a4a5-5f41ac2bc602_1600x680.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>This post is part of Epoch AI&#8217;s <a href="https://epoch.ai/gradient-updates">Gradient Updates</a> newsletter, which shares more opinionated or informal takes on big questions in AI progress. These posts solely represent the views of the authors, and do not necessarily reflect the views of Epoch AI as a whole.</em></p><p><em>Originally posted on <a href="https://epoch.ai/gradient-updates/how-persistent-is-the-inference-cost-burden">Epoch AI</a>.</em></p><div><hr></div><p>Toby Ord has written a <a href="https://www.tobyord.com/writing/how-well-does-rl-scale">thoughtful post</a> on how RL and inference compute scale for frontier AI models.</p><p>As I understand it, the core of his argument is</p><ul><li><p>(1) RL scaling primarily bears fruit by enabling models to productively use longer outputs, which means you need to scale inference compute to realize the gains</p></li><li><p>(2) RL scaling itself delivers poor returns, requiring roughly 10,000x more compute to match what 100x more inference provides.</p></li></ul><p>Combined with the fact that inference costs are per-use and can&#8217;t be amortized like training costs, this paints a picture of a significant and persistent economic burden as we shift away from pretraining scaling.</p><p>There&#8217;s a lot I agree with in Toby&#8217;s analysis, and I find the framing useful. However, I think both claims above may be overstated. On (1): even though inference costs are per-use, the dollar cost to reach a given capability level falls rapidly over time, so the burden might not be so persistent. On (2): the RL scaling data is thin, and there&#8217;s likely been substantial compute efficiency progress in RL since o1 and o3.</p><h2>I. What I agree with</h2><p>I agree with Toby that for tasks that models can solve at all by increasing inference compute, it frequently seems more efficient to scale inference compute than to increase RL compute.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a></p><p>Scaling reinforcement learning also makes it possible to perform new tasks the model could not achieve before with any amount of test-time compute (for example, many <a href="https://evaluations.metr.org/gpt-5-1-codex-max-report/#:~:text=Comparison%20of%20time%20horizon%20achieved%20by%20tokens%20used%20for%20different%20models">METR tasks</a>). However, as Toby <a href="https://www.lesswrong.com/posts/xpj6KhDM9bJybdnEe/how-well-does-rl-scale?commentId=hxrE5JmLwjbe5qXvF">argues</a>, RL unlocks these new tasks in large part by allowing the model to productively use longer outputs (chain of thought, tool calls, or writing) when completing them. Often, these harder tasks inherently require more steps to complete, and just scaling reinforcement learning does not particularly seem to make more steps happen within individual forward passes (see also this related <a href="https://blog.redwoodresearch.org/p/distinguish-between-inference-scaling">post</a> by Ryan Greenblatt).</p><p>Thus, to solve these new, harder tasks with your newly RL-scaled model, you have to pay a large inference cost &#8212; a variable cost that unlike training costs is not amortized by serving more users.</p><h2>II. Fixed-capability costs fall fast</h2><p>Toby <a href="https://www.tobyord.com/writing/how-well-does-rl-scale#:~:text=But%20perhaps%20more,into%20more%20intelligence.">suggests</a> that with RL scaling approaching its limits, inference scaling is all that remains<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a>, and inference scaling gives models more time to think rather than making them smarter per step. He argues this means the compute scaling that has driven recent AI progress will substantially weaken, while also imposing persistent per-use costs.</p><p>I agree that per-use costs can be significant in the short run<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a>, but I&#8217;m not convinced they&#8217;re persistent, since the price of inference to reach a given capability level drops rapidly. In a related <a href="https://www.tobyord.com/writing/hourly-costs-for-ai-agents">post</a>, Toby looks at the cost of running models on the METR suite, and how that has changed as time horizons have increased. However, all the models considered in this analysis were state-of-the-art at release, so this analysis does not account for how much cheaper inference can become.</p><p>First, individual forward passes can become less expensive over time. Improvements to model training mean that smaller models can match the capability of earlier larger ones. Distillation is a salient example: once a frontier model can generate high-quality reasoning traces, you can train smaller, cheaper models to imitate those traces. On top of that, a steady stream of algorithmic improvements in inference efficiency (e.g. <a href="https://arxiv.org/abs/2211.17192">speculative decoding</a>, <a href="https://docs.vllm.ai/en/latest/design/paged_attention/">paged attention</a>, <a href="https://arxiv.org/abs/2512.02556">sparse</a> <a href="https://arxiv.org/abs/2502.11089">attention</a>, <a href="https://blog.vllm.ai/2026/01/08/kv-offloading-connector.html">offloading</a> or <a href="https://github.com/NVIDIA/kvpress">compressing</a> the KV-cache, etc.) all reduce the cost per token.</p><p>At the same time, over time models can reach the same level of capabilities while needing to do fewer forward passes. Models can be trained to reason more concisely: Anthropic, for instance, <a href="https://newsletter.semianalysis.com/p/deepseek-debrief-128-days-later#:~:text=Claude%20has%20the%20lowest%20amount%20of%20total%20output%20tokens%20for%20leading%20reasoning%20models%20and%20showed%20an%20impressive%20improvement%20over%20Claude%203.7%20Sonnet.%C2%A0%C2%A0">substantially reduced</a> the verbosity of Claude&#8217;s reasoning between Claude Sonnet 3.7 and Claude Sonnet 4<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a>.</p><p>Hardware improvements also help: the cost per FLOP <a href="https://epoch.ai/data/machine-learning-hardware">decreases</a> with each new generation of GPU. Combined with the algorithmic gains above, we can see the net effect on concrete tasks. On <a href="https://epoch.ai/benchmarks/frontiermath">FrontierMath</a>, reaching roughly 27% accuracy required about 43 million output tokens with o4-mini with high reasoning effort in April 2025, but only about 5 million tokens with GPT-5.2 with low reasoning effort in December 2025. Even accounting for the difference in price per output token, that&#8217;s roughly a 3x cost reduction over eight months. Overall, the trend seems to be very roughly a <a href="https://arxiv.org/abs/2511.23455">5&#8211;10x cost reduction per year</a> for reaching a given capability level<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a>. Concretely, suppose the first AI that can <a href="https://epochai.substack.com/i/187149776/task-1-replicating-an-interactive-web-interface-for-an-economic-model">perfectly replicate the web interface for a complex economic model</a> can only do so for $50,000. While that is a high initial price, if the most aggressive cost reductions hold, it would become $5,000 a year later, and $500 after two years.</p><p>This changes the economic picture: even if it is initially costly for a user to reach a given capability level, this gets much cheaper over time. Nevertheless, this decrease might slow down in the future. There is likely a <a href="https://x.com/karpathy/status/1938626382248149433">number of parameters</a> below which models are simply too small to have strong general agentic capabilities regardless of how they are distilled. And heavily distilled models do seem to be more brittle overall. The fast cost reductions we observe might partly be an artifact of measuring capabilities through benchmarks, where distilled models tend to perform disproportionately well. That said, even with these caveats, 5&#8211;10x per year is a very fast trend. Overall, this makes me more interested in tracking the capabilities of smaller models and trends in falling inference costs: stay tuned as we will have more rigorous analysis on this soon!</p><h2>III. The returns to RL scaling might be higher</h2><p>Toby&#8217;s core estimate comes from comparing the two panels in OpenAI&#8217;s original o1 announcement chart. Both panels show performance on AIME with a logarithmic x-axis spanning roughly two orders of magnitude (100x). The left panel shows RL training scaling, while the right panel shows inference scaling. Toby observes that the slope of the RL panel is roughly half that of the inference panel: 100x more inference takes you from around 20% to around 80%, while 100x more RL training only takes you from around 33% to around 66%. At that rate, matching the jump from 20% to 80% would require 10,000x more RL compute. Toby then checks this against the o3 training curve and the o1-to-o3 and o3-to-GPT-5 comparisons, which are broadly consistent.</p><p>This is an interesting observation, but I find it hard to draw strong conclusions from it. The x-axis numbers are removed, we&#8217;re reading slopes off a small number of data points, and scaling trends are frequently sensitive to details we don&#8217;t have access to from the outside.</p><p>More importantly, algorithmic progress can shift a given RL scaling curve substantially. For example, <a href="https://arxiv.org/pdf/2510.13786">in Figure 2 of this paper studying RL scaling laws, Scaled RL</a> achieves more than twice the slope of GRPO on comparable benchmarks (see figure below), including when the y-axis is linear while the x-axis is logarithmic. This is at much lower compute scales than o1 and o3, so it&#8217;s not directly comparable, but it illustrates how much the efficiency of RL training can vary with the algorithm.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!CPij!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6fb5830e-a8be-471c-a4a5-5f41ac2bc602_1600x680.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!CPij!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6fb5830e-a8be-471c-a4a5-5f41ac2bc602_1600x680.png 424w, https://substackcdn.com/image/fetch/$s_!CPij!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6fb5830e-a8be-471c-a4a5-5f41ac2bc602_1600x680.png 848w, https://substackcdn.com/image/fetch/$s_!CPij!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6fb5830e-a8be-471c-a4a5-5f41ac2bc602_1600x680.png 1272w, https://substackcdn.com/image/fetch/$s_!CPij!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6fb5830e-a8be-471c-a4a5-5f41ac2bc602_1600x680.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!CPij!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6fb5830e-a8be-471c-a4a5-5f41ac2bc602_1600x680.png" width="1456" height="619" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6fb5830e-a8be-471c-a4a5-5f41ac2bc602_1600x680.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:619,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!CPij!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6fb5830e-a8be-471c-a4a5-5f41ac2bc602_1600x680.png 424w, https://substackcdn.com/image/fetch/$s_!CPij!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6fb5830e-a8be-471c-a4a5-5f41ac2bc602_1600x680.png 848w, https://substackcdn.com/image/fetch/$s_!CPij!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6fb5830e-a8be-471c-a4a5-5f41ac2bc602_1600x680.png 1272w, https://substackcdn.com/image/fetch/$s_!CPij!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6fb5830e-a8be-471c-a4a5-5f41ac2bc602_1600x680.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>It also makes sense to me that OpenAI wouldn&#8217;t have been laser-focused on maximizing RL compute efficiency in the early days of o1 and o3. RL compute was still a small fraction of total training cost, so the returns were good enough without needing to optimize compute aggressively. Since then, there&#8217;s likely been considerable algorithmic progress (in large part through better and more diverse <a href="https://epoch.ai/gradient-updates/state-of-rl-envs">RL tasks and environments</a>), which likely improved the slope of RL scaling. Much of the effort since o1 and o3 has also gone into applying RL to new domains beyond math and coding, rather than pushing RL as far as possible within those two domains.</p><h2>Conclusion</h2><p>Toby's discussion of RL scaling versus inference scaling is useful, and the core observation that RL gains come largely with longer chains of thought is well-taken. But the picture he paints may overstate how much of a bottleneck this will be for AI progress. The cost to reach a given capability level falls fast, so the inference cost burden is more transient than it might appear from looking only at frontier models at launch. And the RL scaling data we have is still thin, so I would treat the 10,000x figure with a grain of salt. This has made me more interested in studying how quickly cheaper models catch up to frontier capability levels, and how inference costs for a fixed task decrease over time.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://epochai.substack.com/p/how-persistent-is-the-inference-cost?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:&quot;button-wrapper&quot;}" data-component-name="ButtonCreateButton"><a class="button primary button-wrapper" href="https://epochai.substack.com/p/how-persistent-is-the-inference-cost?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><p><em>Thanks to Toby Ord, Ryan Greenblatt, Greg Burnham, Anson Ho, David Owen, Tao Lin and Josh You for helpful comments.</em></p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://epochai.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading! Subscribe to receive the latest from Epoch AI.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>In fact, the returns to inference compute scaling might be even better than Toby mentions, since his estimates focus on increasing reasoning effort but mostly do not account for other ways of increasing inference compute, such as best-of-n sampling.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>At least, unless or until there is a <a href="https://forum.effectivealtruism.org/posts/JAcueP8Dh6db6knBK/the-scaling-series-discussion-thread-with-toby-ord?commentId=eHEfpwDgbkCnv3gzP#:~:text=It%20is%20possible%20that%20there%20will%20always%20be%20a%20new%20paradigm%20of%20compute%20scaling">new paradigm of compute scaling</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>As Ryan Greenblatt <a href="https://blog.redwoodresearch.org/p/distinguish-between-inference-scaling">points out</a>, even for these initial costs, it matters how they compare to the cost of humans solving the task.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>Relatedly, single-forward-pass reasoning capabilities appear to be<a href="https://blog.redwoodresearch.org/p/measuring-no-cot-math-time-horizon"> increasing exponentially</a> (thanks to Greg Burnham for pointing this out). This also suggests that tasks that once required long chains of thought could be handled with much fewer tokens. Note that this trend does not seem to be driven by RL itself, and might trade off with the previous trend, to the extent that it requires larger models.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>Our<a href="https://epoch.ai/data-insights/llm-inference-price-trends"> earlier data insight</a> on inference price trends isn&#8217;t ideal evidence here, since it doesn&#8217;t account for reasoning models or tasks where longer outputs provide genuine value, and probably is unusually vulnerable to benchmark climbing/contamination.</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[How close is AI to taking my job?]]></title><description><![CDATA[Beyond benchmarks as leading indicators for task automation]]></description><link>https://epochai.substack.com/p/how-close-is-ai-to-taking-my-job</link><guid isPermaLink="false">https://epochai.substack.com/p/how-close-is-ai-to-taking-my-job</guid><dc:creator><![CDATA[Anson Ho]]></dc:creator><pubDate>Sat, 07 Feb 2026 02:54:58 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/eaa93762-5658-409a-b3ff-430f7549ab8e_2454x992.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>This post is part of Epoch AI&#8217;s <a href="https://epoch.ai/gradient-updates">Gradient Updates</a> newsletter, which shares more opinionated or informal takes on big questions in AI progress. These posts solely represent the views of the authors, and do not necessarily reflect the views of Epoch AI as a whole.</em></p><p><em>Originally posted on <a href="https://epoch.ai/gradient-updates/how-close-is-ai-to-taking-my-job">Epoch AI</a>.</em></p><div><hr></div><h2>I. Searching under the streetlight</h2><p>How can we anticipate when AI will be able to do our jobs? AI researchers have mainly tried to answer this question by building complex AI benchmarks. The problem is that this approach is fundamentally flawed.</p><p>A good example of this is OpenAI&#8217;s <a href="https://openai.com/index/gdpval/">GDPval</a>. On paper, it&#8217;s a cool benchmark that captures AI performance on a wide range of real-world job tasks in the US economy. The benchmark tasks were meticulously constructed to be realistic, involving the hard work of hundreds of experts and likely millions of dollars &#8212; placing it among the most expensive economics papers of all time.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> If there&#8217;s one benchmark that could be the leading indicator of AI job automation, it&#8217;s GDPval.</p><p>Unfortunately, the benchmark seems to have fallen prey to the same issue plaguing most other benchmarks. Shortly after release, AI models have beaten the human baseline &#8212; GPT-5.2 reached <a href="https://evals.openai.com/gdpval/leaderboard">parity</a> with industry experts, and Claude Opus 4.6 likely does <a href="https://www.anthropic.com/news/claude-opus-4-6">even better</a>. And yet, the actual economic impacts of AI remain muted. The benchmark doesn&#8217;t fully reflect the economic effects, and so it&#8217;s falling short in its role as a leading indicator for automation.</p><p>This isn&#8217;t really OpenAI&#8217;s fault, rather it&#8217;s a fundamental challenge with AI benchmarks. These need to be designed so that you can automatically run evaluations quickly and regularly, but this constrains us to tasks that are &#8220;clean&#8221; and close-ended. So when we try to generalize these to the messy workflows and fuzzy tasks of the real world, we inevitably encounter many issues.</p><p>If this is right, then the problem is that we&#8217;ve mostly been searching under the streetlight, primarily tracking progress on the things that benchmarks can easily measure. But it also suggests that there&#8217;s a way to search beyond the streetlight: we just need to sacrifice a bit of rigor and ease of evaluation, and in exchange we can get a better sense of automation on <em>truly</em> real-world tasks.</p><p>Here&#8217;s the idea: pick out a set of <em>actual</em> work tasks that people do on a regular basis, and then see how well AI is able to do them, and regularly check how much better AI is getting. This kind of subjective AI evaluation on realistic tasks isn&#8217;t a totally novel idea &#8212; for example, in AI safety, models are often &#8220;<a href="https://red.anthropic.com/">red-teamed</a>&#8221; in a way that involves subjectively rating their risk levels. There are also <a href="https://aleximas.substack.com/p/what-is-the-impact-of-ai-on-productivity">many studies</a> looking at AI uplift on real-world tasks. But I think there&#8217;s room for a lot more of this kind of analysis, especially because all the benchmarks seem to be running into the same bottlenecks. So I think it&#8217;s an underrated way of tracking AI&#8217;s potential impacts in actual jobs.</p><p>So to demonstrate how this could work, I decided to try getting today&#8217;s best AI agents to do my job.</p><h2>II. Trying to automate my job away (for science)</h2><p>The first step is to pick out three tasks that people have actually done at Epoch, with a heavy bias towards my own work. I thought it&#8217;d be good to focus on things that capture the full gamut of important AI bottlenecks &#8212; long-horizon planning and execution, creativity and taste, and so on. So these are tasks that should be relatively difficult for AIs today, but play an important role in my work.</p><p>After that, I spent 30-60min trying to get AIs to autonomously solve each task and analyzed where current models seem to be falling short. I could&#8217;ve done a lot more to improve AI performance on each task, but I wanted this to broadly reflect how I&#8217;ve been using AI in practice given my time and work constraints. Then I roughly forecasted when I thought AIs would do them to my liking (i.e. such that I&#8217;d personally use their outputs in practice), and how well AIs would do by the end of the year.</p><p>So if we put things together, this should give a more concrete picture of how close AI is to taking my job. Here&#8217;s where I ended up:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!WoqK!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0f0cca8-74a5-4397-b8b1-90a0e75308a6_2558x1340.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!WoqK!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0f0cca8-74a5-4397-b8b1-90a0e75308a6_2558x1340.png 424w, https://substackcdn.com/image/fetch/$s_!WoqK!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0f0cca8-74a5-4397-b8b1-90a0e75308a6_2558x1340.png 848w, https://substackcdn.com/image/fetch/$s_!WoqK!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0f0cca8-74a5-4397-b8b1-90a0e75308a6_2558x1340.png 1272w, https://substackcdn.com/image/fetch/$s_!WoqK!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0f0cca8-74a5-4397-b8b1-90a0e75308a6_2558x1340.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!WoqK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0f0cca8-74a5-4397-b8b1-90a0e75308a6_2558x1340.png" width="1456" height="763" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a0f0cca8-74a5-4397-b8b1-90a0e75308a6_2558x1340.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:763,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:464216,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://epochai.substack.com/i/187149776?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0f0cca8-74a5-4397-b8b1-90a0e75308a6_2558x1340.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!WoqK!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0f0cca8-74a5-4397-b8b1-90a0e75308a6_2558x1340.png 424w, https://substackcdn.com/image/fetch/$s_!WoqK!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0f0cca8-74a5-4397-b8b1-90a0e75308a6_2558x1340.png 848w, https://substackcdn.com/image/fetch/$s_!WoqK!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0f0cca8-74a5-4397-b8b1-90a0e75308a6_2558x1340.png 1272w, https://substackcdn.com/image/fetch/$s_!WoqK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa0f0cca8-74a5-4397-b8b1-90a0e75308a6_2558x1340.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>My probabilities that an AI successfully completes the given task by the specified date. For a task to be &#8220;completed&#8221;, the final product needs to be good enough for me to want to use it in practice. Tasks are done autonomously by the AI except for the second task, where two rounds of comments are provided.</em></figcaption></figure></div><p>To be clear, this doesn&#8217;t mean that I expect to lose my job by early 2029 &#8212; more on that later. But first, let&#8217;s dive into the details of the tasks.</p><h3>Task 1: Replicating an interactive web interface for an economic model</h3><p>About a year ago, we at Epoch released a complex economic model of AI automation, which we called the GATE model. And by &#8220;complex&#8221;, I mean that it has over forty parameters, is solved by some clunky variant of gradient descent, and has a bunch of add-ons that can make the simulation especially unstable.</p><p>To accompany the release, we provided an <a href="https://epoch.ai/gate">interactive web interface</a> that lets users input their own parameters, look at a bunch of fancy graphs about how economic growth could go crazy, and read some documentation. The point is, it&#8217;s by no means a simple thing to implement, and the frontend isn&#8217;t a piece of cake either.</p><p>Here&#8217;s what that looks like:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Ole_!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F73dc92a7-f3a3-48fe-9535-fd2aefe14be7_1600x900.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Ole_!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F73dc92a7-f3a3-48fe-9535-fd2aefe14be7_1600x900.png 424w, https://substackcdn.com/image/fetch/$s_!Ole_!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F73dc92a7-f3a3-48fe-9535-fd2aefe14be7_1600x900.png 848w, https://substackcdn.com/image/fetch/$s_!Ole_!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F73dc92a7-f3a3-48fe-9535-fd2aefe14be7_1600x900.png 1272w, https://substackcdn.com/image/fetch/$s_!Ole_!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F73dc92a7-f3a3-48fe-9535-fd2aefe14be7_1600x900.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Ole_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F73dc92a7-f3a3-48fe-9535-fd2aefe14be7_1600x900.png" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/73dc92a7-f3a3-48fe-9535-fd2aefe14be7_1600x900.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Ole_!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F73dc92a7-f3a3-48fe-9535-fd2aefe14be7_1600x900.png 424w, https://substackcdn.com/image/fetch/$s_!Ole_!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F73dc92a7-f3a3-48fe-9535-fd2aefe14be7_1600x900.png 848w, https://substackcdn.com/image/fetch/$s_!Ole_!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F73dc92a7-f3a3-48fe-9535-fd2aefe14be7_1600x900.png 1272w, https://substackcdn.com/image/fetch/$s_!Ole_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F73dc92a7-f3a3-48fe-9535-fd2aefe14be7_1600x900.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>So the first task is this: <strong>Can an AI model build a functional replica that&#8217;s as good as the published web playground?</strong><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> The AI is given access to the academic <a href="https://arxiv.org/abs/2503.04941">paper</a> and the original webpage, which includes a bunch of documentation. It&#8217;s also provided with descriptions on which graphs need to be created, as well as parameter ranges for each of the parameters.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a></p><p>I asked Claude Code to implement this, mainly because it&#8217;s widely regarded as the best tool for coding. Initially it complained about the message being too long if I dumped in the whole paper PDF, so I gave it the <a href="https://arxiv.org/abs/2503.04941">arXiv URL</a>. It then spent a while looking up the equations in the paper, quickly wrote some HTML and Javascript, and this was the result:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Gyim!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8028af8-9e8c-4420-80ac-9114b92661cc_1600x1178.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Gyim!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8028af8-9e8c-4420-80ac-9114b92661cc_1600x1178.png 424w, https://substackcdn.com/image/fetch/$s_!Gyim!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8028af8-9e8c-4420-80ac-9114b92661cc_1600x1178.png 848w, https://substackcdn.com/image/fetch/$s_!Gyim!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8028af8-9e8c-4420-80ac-9114b92661cc_1600x1178.png 1272w, https://substackcdn.com/image/fetch/$s_!Gyim!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8028af8-9e8c-4420-80ac-9114b92661cc_1600x1178.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Gyim!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8028af8-9e8c-4420-80ac-9114b92661cc_1600x1178.png" width="1456" height="1072" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e8028af8-9e8c-4420-80ac-9114b92661cc_1600x1178.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1072,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Gyim!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8028af8-9e8c-4420-80ac-9114b92661cc_1600x1178.png 424w, https://substackcdn.com/image/fetch/$s_!Gyim!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8028af8-9e8c-4420-80ac-9114b92661cc_1600x1178.png 848w, https://substackcdn.com/image/fetch/$s_!Gyim!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8028af8-9e8c-4420-80ac-9114b92661cc_1600x1178.png 1272w, https://substackcdn.com/image/fetch/$s_!Gyim!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8028af8-9e8c-4420-80ac-9114b92661cc_1600x1178.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>This looks pretty cool to start off with, and it&#8217;s certainly a functioning website, but it goes downhill from there. The most crucial issue is that many of the model predictions differ vastly from the official simulation, which suggests something went wrong in the reimplementation. For example, the GWP change is way lower than what you observe in the mainline scenario in our actual GATE model.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a> Another issue is that the website has missing features, like a &#8220;comparison mode&#8221; that allows users to compare different parameter settings.</p><p>The big question is: how well will AI systems do on this task in the future? I think we can think about this in several ways.</p><p>One is to consider the trend in <a href="https://metr.org/blog/2025-03-19-measuring-ai-ability-to-complete-long-tasks/">METR&#8217;s time horizons</a>, which I think applies reasonably well given that this is a software task that&#8217;s pretty well-defined. By the end of the year, we expect AI to be able to do tasks roughly one day long with a 50% success rate.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a> In comparison, I&#8217;d guess that this task would take several days for a person familiar with the paper and is able to play around with the web interface. So this task seems plausibly solvable by December, but it&#8217;s less likely than not.</p><p>Still, it&#8217;s a clunky task with a lot of moving parts. Perhaps the hardest part is the backend, which involves some nitty-gritty optimization with numerical stability problems &#8212; you&#8217;d probably need to stare at many graphs and iterate to get it right. The frontend has its own challenges, but I think Claude Code&#8217;s implementation isn&#8217;t so bad on this front.</p><p>Putting things together, I think that models have about a 10% chance of implementing this as well as the existing live version by the end of the year. In my mind, this&#8217;ll probably be because models flunk some part of the code implementation for the economic model&#8217;s dynamics, or there are some non-obvious bugs that need a bunch of people trying diverse things on the website to identify. Fixing these issues takes a bit more time, and by late 2027 or early 2028, I think there&#8217;s a roughly even chance that this task gets solved.</p><p>So my guess is that our web dev team will continue to produce interactive web interfaces that are of a similar complexity, like Epoch&#8217;s <a href="https://epoch.ai/tools/distributed-training">distributed training simulator</a> and webpage for <a href="https://epoch.ai/data/data-centers#">frontier data centers</a>, at least in terms of core functionality. And if you work on things of a similar difficulty, I think you&#8217;ll probably be in the same boat.</p><h3>Task 2: Writing an article</h3><p>Now let&#8217;s move on to the second task: Can AI write articles for this newsletter?</p><p>To test this, I asked Claude Opus 4.5 to try and write one of my previous <a href="https://epoch.ai/gradient-updates/how-well-did-forecasters-predict-2025-ai-progress">articles</a>, which summarized how well people forecast AI progress in 2025 (notably, this means that Claude didn&#8217;t need to come up with the post idea itself). I gave it the list of questions and resolutions from the AI Digest (who ran the survey), all the results, and asked it to look at our previous posts to get a sense of our style.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a></p><p>Its initial attempt at this wasn&#8217;t very good &#8212; it didn&#8217;t include graphs, links to relevant sources, and even left out some of the questions from the survey entirely! The writing style was a tad stiff, the structure was odd (e.g. the survey demographics were put near the end of the article), and the results weren&#8217;t very well justified. For example, here&#8217;s Claude explaining why forecasters significantly underestimated progress on the cybersecurity benchmark Cybench:</p><blockquote><p><em>Cybersecurity evaluations have received less public attention than coding or math benchmarks, and forecasters may have had weaker mental models of the rate of progress in this domain. It also suggests that capability improvements on agentic tasks may be occurring faster than the broader forecasting community appreciates.</em></p></blockquote><p>As far as I can tell, there isn&#8217;t any justification for any of the claims in the paragraph &#8212; don&#8217;t ask me how Claude came to these conclusions!</p><p>That said, it&#8217;s a bit unfair of me to judge it purely on its zero-shot performance &#8212; when I&#8217;m writing I usually get lots of helpful comments from others. So I gave Claude two rounds of feedback, totalling around forty comments, each about one to two sentences long. For example:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Gw10!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff14bbecc-8ea8-422a-8002-87905a687805_1600x751.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Gw10!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff14bbecc-8ea8-422a-8002-87905a687805_1600x751.png 424w, https://substackcdn.com/image/fetch/$s_!Gw10!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff14bbecc-8ea8-422a-8002-87905a687805_1600x751.png 848w, https://substackcdn.com/image/fetch/$s_!Gw10!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff14bbecc-8ea8-422a-8002-87905a687805_1600x751.png 1272w, https://substackcdn.com/image/fetch/$s_!Gw10!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff14bbecc-8ea8-422a-8002-87905a687805_1600x751.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Gw10!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff14bbecc-8ea8-422a-8002-87905a687805_1600x751.png" width="1456" height="683" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f14bbecc-8ea8-422a-8002-87905a687805_1600x751.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:683,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Gw10!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff14bbecc-8ea8-422a-8002-87905a687805_1600x751.png 424w, https://substackcdn.com/image/fetch/$s_!Gw10!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff14bbecc-8ea8-422a-8002-87905a687805_1600x751.png 848w, https://substackcdn.com/image/fetch/$s_!Gw10!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff14bbecc-8ea8-422a-8002-87905a687805_1600x751.png 1272w, https://substackcdn.com/image/fetch/$s_!Gw10!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff14bbecc-8ea8-422a-8002-87905a687805_1600x751.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Unfortunately, while this solved some problems (like adding graphs and links), new issues emerged. Sometimes Claude would fail to resolve subtle graphical issues that would normally involve multiple rounds of iteration, like putting text in the right part of the graph. And there were many such problems littered throughout the article &#8212; even after my feedback, the writing was still stilted, had many inaccuracies, and terms that weren&#8217;t explained to my liking.</p><p>In the end, the post wasn&#8217;t unreadable, but there were so many small errors that I felt better off going back to square one and rewriting the post from scratch. So, at least in the near future, I&#8217;ll continue to be writing articles.</p><p>But just how near is the &#8220;near future&#8221;? <strong>When will AI essentially be able to write a full article of this complexity that I&#8217;ll really like, after giving a few rounds of comments?</strong> Unlike Task 1, I don&#8217;t see a clear trend that we can extrapolate to forecast progress. So I think we&#8217;ll probably need to turn to the age-old tradition of &#8220;vibes-based forecasting&#8221;.</p><p>On the one hand, there are a bunch of reasons to expect AI to get a lot better on this task very soon. One is that AI writing has improved a ton over the last few years, and it&#8217;s not crazy at all to expect that this continues. Labs have also been making efforts to improve writing, as we&#8217;ve seen with models like GPT-4.5, and people are increasingly building <a href="https://surgehq.ai/blog/hemingway-bench-ai-writing-leaderboard">writing</a> <a href="https://eqbench.com/creative_writing.html">benchmarks</a>.</p><p>What&#8217;s more, the task of writing an article isn&#8217;t just about the quality of the <em>writing</em> itself. In many ways, Claude was actually more bottlenecked on things like data analysis and graphing, as well as insufficient context about my writing preferences. But I expect AIs to get vastly better at practically any kind of coding-related task over the next few years, and they&#8217;ll also have a <a href="https://epochai.substack.com/p/the-huge-potential-implications-of">greater degree of continual learning</a>, which would help surmount these barriers. Surely it&#8217;s not that hard to give AIs context about my writing preferences and Epoch&#8217;s worldview?</p><p>On the other hand, there are also reasons for slower progress, at least relative to domains like coding. For one, labs are probably going to focus on things that are more lucrative than writing, like finance and coding. Consider that there are around <a href="https://www.bls.gov/ooh/Computer-and-Information-Technology/Software-developers.htm">1.7 million</a> software engineering jobs in the US, with a median pay of <a href="https://www.bls.gov/ooh/Computer-and-Information-Technology/Software-developers.htm">$133,000</a>. That&#8217;s several times higher than both the total writing jobs (around 350,000) and median pay (in the ballpark of $70,000).<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a></p><p>This relative focus on coding becomes even stronger because writing is one of those fuzzy tasks where it&#8217;s hard to say what&#8217;s good or bad. So it&#8217;s hard to measure improvements in writing, making it hard for AI companies to &#8220;hill-climb&#8221; their way to becoming Ernest Hemingway. This is also part of the reason why reasoning <a href="https://epoch.ai/gradient-updates/quantifying-the-algorithmic-improvement-from-reasoning-models">seems to help more</a> on tasks like math than writing.</p><p>Personally, I find the arguments for slower progress more compelling. I&#8217;d probably give a 5% chance that AI will be better at writing these posts than me by the end of the year, and maybe a 50% around late 2028/early 2029. For posts that are more complex and require coming up with novel ideas, I&#8217;d probably go a bit later than that, but that depends on how well I can actually write those posts!</p><p>This means that I&#8217;m more bearish on AI progress for this task compared to the previous one &#8212; that&#8217;s interesting because I actually find this task a lot easier. Perhaps this is a case of &#8220;<a href="https://epoch.ai/gradient-updates/moravec-s-paradox">Moravec&#8217;s Paradox</a> strikes again&#8221; &#8212; AI is often good at the things humans find hard, and bad at the things we find easy.</p><h3>Task 3: Publishing an article</h3><p>The third and final task may be even more susceptible to Moravec&#8217;s Paradox: <strong>Can AI port an article from Google Docs to Substack and Epoch&#8217;s website?</strong> This requires agentic computer use, which seems to trip up AI models in strange ways.</p><p>Most of this task is boring grunt work, spent going through many tedious steps, across different platforms. We start off with a finalized draft of the post and thread written up in Google Docs, which then gets ported to two places:</p><ul><li><p>Epoch&#8217;s website: This involves making a new branch on the GitHub repo for Epoch&#8217;s website, creating a new markdown document with the right metadata, copying all the content in, formatting the tables, attaching images, and so on.</p></li><li><p>Substack: This one is easier, but there are some oddities of Substack. You can&#8217;t make tables, so we have to take screenshots of the tables from our website&#8217;s preview. And footnotes don&#8217;t get copied over from Google Docs, so you need to add them manually.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-8" href="#footnote-8" target="_self">8</a></p></li></ul><p>Overall, it usually takes me about two hours to do this task. If only it were as simple as a single copy and paste, life would be so much easier &#8212; or so I thought. To my surprise, Claude with its Chrome browser extension struggles to copy+paste from Google Docs to Substack:</p><ul><li><p>First it tried to download the file. It clicks on <code>File</code>, but then complains that no window opened, even though I could see it opening. Not sure what&#8217;s going on there.</p></li><li><p>Then it tried to highlight everything in the Google Doc and do a copy+paste operation. But for some reason it wasn&#8217;t able to do this for all the text, and had to resort to copying text block by block, scrolling down the page.</p></li><li><p>This was taking forever, so I tried restarting the task, (naively?) hoping that things would be better this time. Unfortunately it got worse: it ran into the same issue, but instead of copying text block by block, it opted for the strategy of scrolling down the page and <em>reading </em>the text screenshot by screenshot! </p></li></ul><p>After it started doing that, I threw in the towel and decided I&#8217;d try another agent. This time it was ChatGPT Atlas&#8217; turn. Initially it seemed to get stuck identifying which tabs were open, but after another try it managed to get going. To my surprise, it successfully copied the main post onto Substack &#8212; after watching Claude&#8217;s agonizing failed attempts to copy and paste text, this was mindblowing.</p><p>Alas, I only managed to feel the AGI for about fifteen more seconds before ChatGPT started making some really silly mistakes. It was transferring the footnotes, but it didn&#8217;t use Substack&#8217;s footnote feature, and it didn&#8217;t add all the footnotes. Even worse, the footnotes were all made up, and there was a bizarre formatting issue where ChatGPT placed the cursor in the wrong place:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!JUTR!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6b4e822-86e9-47df-8927-8cce35cc3f35_884x990.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!JUTR!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6b4e822-86e9-47df-8927-8cce35cc3f35_884x990.png 424w, https://substackcdn.com/image/fetch/$s_!JUTR!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6b4e822-86e9-47df-8927-8cce35cc3f35_884x990.png 848w, https://substackcdn.com/image/fetch/$s_!JUTR!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6b4e822-86e9-47df-8927-8cce35cc3f35_884x990.png 1272w, https://substackcdn.com/image/fetch/$s_!JUTR!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6b4e822-86e9-47df-8927-8cce35cc3f35_884x990.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!JUTR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6b4e822-86e9-47df-8927-8cce35cc3f35_884x990.png" width="884" height="990" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b6b4e822-86e9-47df-8927-8cce35cc3f35_884x990.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:990,&quot;width&quot;:884,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!JUTR!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6b4e822-86e9-47df-8927-8cce35cc3f35_884x990.png 424w, https://substackcdn.com/image/fetch/$s_!JUTR!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6b4e822-86e9-47df-8927-8cce35cc3f35_884x990.png 848w, https://substackcdn.com/image/fetch/$s_!JUTR!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6b4e822-86e9-47df-8927-8cce35cc3f35_884x990.png 1272w, https://substackcdn.com/image/fetch/$s_!JUTR!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6b4e822-86e9-47df-8927-8cce35cc3f35_884x990.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>ChatGPT Agent successfully ports the main text of our <a href="https://epoch.ai/gradient-updates/can-ai-companies-become-profitable">last Gradient Update</a>, but messes up the footnotes.</em></figcaption></figure></div><p>So at the moment, AI agents seem to fail quite miserably at this task, and are also extremely slow. This corroborates pretty well with Claude&#8217;s <a href="https://www.twitch.tv/claudeplayspokemon">performance</a> on Pok&#233;mon, as well as the finding that METR&#8217;s time horizons are <a href="https://metr.org/blog/2025-07-14-how-does-time-horizon-vary-across-domains/">40-100&#215;</a> <a href="https://metr.org/notes/2026-01-22-time-horizon-limitations/">shorter</a> for visual computer use tasks compared to their <a href="https://metr.org/blog/2025-03-19-measuring-ai-ability-to-complete-long-tasks/">headline numbers</a>. For example, this could bring a five hour (300 minute) time horizon down to a three minute time horizon.</p><p>But while the time horizons are much shorter, the <em>growth rate</em> is about the same as the METR&#8217;s main results, with roughly two doublings each year. Today&#8217;s agents are also <a href="https://time.com/7345903/ai-chatgpt-claude-gemini-pokemon/">much better</a> at Pok&#233;mon compared to a year ago. If these trends hold, we should expect to see very fast progress in the near future.</p><p>And people are trying to make it happen &#8212; labs have been creating agents like <a href="https://openai.com/index/introducing-chatgpt-agent/">ChatGPT Agent</a>, <a href="https://www.kimi.com/blog/kimi-k2-5.html">Kimi K2.5&#8217;s Agent Swarm</a>, and <a href="https://x.com/claudeai/status/2010805682434666759">Claude Cowork</a>, and we&#8217;re still in the early days of AI agents. There are also huge incentives for labs to push in this direction, overcoming bottlenecks to AI adoption among <a href="https://www.itu.int/en/mediacentre/Pages/PR-2025-11-17-Facts-and-Figures.aspx">billions</a> of computer users worldwide.</p><p>Of course, there are some bottlenecks that we need to account for. One bottleneck that I learned the hard way is reliability. When ChatGPT Agent was messing around on Substack, there was a point when the cursor hovered over the button to publish the post. I was practically freaking out at the idea of sending an already-published post with garbled footnotes to over ten thousand subscribers, but fortunately nothing happened, and I was able to intervene.</p><p>By the end of the year, I&#8217;d probably give roughly a 10% chance of AIs completing the task autonomously, though they might need to do it very very slowly. In the median case, I think AI agents will get much further in the task &#8212; porting most of the footnotes correctly, adding content to an IDE and attaching the images, and so on. But I think it&#8217;ll be very slow, there&#8217;ll be missed steps, and more likely it&#8217;ll get stuck on something that&#8217;s trivial to most computer-literate humans.</p><p>Looking further out, I think there&#8217;s a 50-50 chance that AI agents can complete this task by around mid-2028.</p><h2>III. What this all means for my job, and perhaps yours too</h2><p>Let&#8217;s take stock of what we&#8217;ve seen so far. Based on the three tasks, today&#8217;s AI seems unable to fully generate complex interactive web simulations, is okay at writing articles, and very bad at publishing them. But I also think that we&#8217;re only a few years away from AIs that can solve these tasks, and the tasks that are hard for me might not be the ones that are hard for the AIs. It depends on how amenable those tasks are to things like RL, and on how economically valuable they are.</p><p>Looking at the forecasts, the last task that I expect to fall comes in early 2029, but that doesn&#8217;t necessarily mean that AI will <em>replace</em> me by 2029. The obvious reason is that these three tasks don&#8217;t comprehensively represent my job, and I wasn&#8217;t trying to pick out the hardest-to-automate tasks. So even if AI is able to solve them, there&#8217;s a good chance that I&#8217;ll just work on other things instead. The bottleneck shifts to something new.</p><p>For example, we at Epoch have mostly prepared for podcast episodes by brainstorming questions in advance. Now let&#8217;s say that Claude becomes better than us at coming up with good podcast questions in advance. Then the bottleneck shifts from writing questions to being able to understand them and ask good follow-ups during the recording itself, in a way that understands what kinds of things humans might be curious about on the fly.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!B1sc!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed908753-3ee8-4618-9f55-1c29af2fd4b8_1600x900.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!B1sc!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed908753-3ee8-4618-9f55-1c29af2fd4b8_1600x900.png 424w, https://substackcdn.com/image/fetch/$s_!B1sc!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed908753-3ee8-4618-9f55-1c29af2fd4b8_1600x900.png 848w, https://substackcdn.com/image/fetch/$s_!B1sc!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed908753-3ee8-4618-9f55-1c29af2fd4b8_1600x900.png 1272w, https://substackcdn.com/image/fetch/$s_!B1sc!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed908753-3ee8-4618-9f55-1c29af2fd4b8_1600x900.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!B1sc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed908753-3ee8-4618-9f55-1c29af2fd4b8_1600x900.png" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ed908753-3ee8-4618-9f55-1c29af2fd4b8_1600x900.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!B1sc!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed908753-3ee8-4618-9f55-1c29af2fd4b8_1600x900.png 424w, https://substackcdn.com/image/fetch/$s_!B1sc!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed908753-3ee8-4618-9f55-1c29af2fd4b8_1600x900.png 848w, https://substackcdn.com/image/fetch/$s_!B1sc!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed908753-3ee8-4618-9f55-1c29af2fd4b8_1600x900.png 1272w, https://substackcdn.com/image/fetch/$s_!B1sc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed908753-3ee8-4618-9f55-1c29af2fd4b8_1600x900.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>The bottleneck is shifting from us Epochians preparing good questions in advance, to being able to understand them and ask good follow ups during real conversation. But there may still be some time before Humanity&#8217;s Last Podcast. (Source: <a href="https://epoch.ai/epoch-after-hours/daniel-litt-ai-math-capabilities-could-be-jagged-for-a-long-time">Epoch After Hours</a>)</em></figcaption></figure></div><p>So if we want to know when AI takes my job, rather than just &#8220;do a bunch of tasks&#8221;, we also need to know how far these bottlenecks can shift. My &#8220;strong opinion, loosely held&#8221; is that this can go surprisingly far. One reason is that Moravec&#8217;s Paradox really messes with our intuitions. AI capabilities are very spiky, and if we don&#8217;t know what this spikiness looks like, then it becomes hard to predict what kinds of things will stump future AIs. This becomes doubly hard because it&#8217;s often hard to know what bottlenecks exist until you actually encounter them. The last bottlenecks will eventually fall. But if we forecast job automation by saying &#8220;job X is just doing Y&#8221; and predicting when AI can do Y, we&#8217;ll likely produce overly aggressive timelines.</p><p>In practice, I imagine the path towards automating my job probably looks something like this: in the next year or two, my day-to-day work will look pretty similar to today at a high level. But if we narrow down to subtasks, we&#8217;ll see a lot more AI use, like how GPT-5.2 is now essentially my default search engine instead of Google. Between 2027 and 2029, AI will continue to get much better at coding, writing, and agentic computer use, until eventually it&#8217;s able to do large chunks of what I do today. I&#8217;ll increasingly shift my attention to things that are hard to stuff into AI contexts, manage multiple AIs, deal with bizarre AI weaknesses, and do more high-level ideation or other fuzzy tasks. This continues for a handful more years before it&#8217;s finally game over for my job.</p><h3>What about you and your job?</h3><p>For the most part, I&#8217;ve framed this post around AI progress on my job, and how I personally expect to be using AIs. But what about everyone else and their jobs?</p><p>First of all, while I probably use AI more than 99% of people on Earth, there are some people who use AI much more than me. I suspect this is especially true for coders based in the Bay Area. As a particularly extreme example, Boris Cherny (the creator of Claude Code) puts my puny Claude + Chrome extension setup to shame:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ernJ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe88b63b7-ff78-4efb-9fc5-d32ad9897d75_1194x974.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ernJ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe88b63b7-ff78-4efb-9fc5-d32ad9897d75_1194x974.png 424w, https://substackcdn.com/image/fetch/$s_!ernJ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe88b63b7-ff78-4efb-9fc5-d32ad9897d75_1194x974.png 848w, https://substackcdn.com/image/fetch/$s_!ernJ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe88b63b7-ff78-4efb-9fc5-d32ad9897d75_1194x974.png 1272w, https://substackcdn.com/image/fetch/$s_!ernJ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe88b63b7-ff78-4efb-9fc5-d32ad9897d75_1194x974.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ernJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe88b63b7-ff78-4efb-9fc5-d32ad9897d75_1194x974.png" width="1194" height="974" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e88b63b7-ff78-4efb-9fc5-d32ad9897d75_1194x974.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:974,&quot;width&quot;:1194,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ernJ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe88b63b7-ff78-4efb-9fc5-d32ad9897d75_1194x974.png 424w, https://substackcdn.com/image/fetch/$s_!ernJ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe88b63b7-ff78-4efb-9fc5-d32ad9897d75_1194x974.png 848w, https://substackcdn.com/image/fetch/$s_!ernJ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe88b63b7-ff78-4efb-9fc5-d32ad9897d75_1194x974.png 1272w, https://substackcdn.com/image/fetch/$s_!ernJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe88b63b7-ff78-4efb-9fc5-d32ad9897d75_1194x974.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>(<a href="https://x.com/bcherny/status/2007179833990885678">Image source</a>)</em></figcaption></figure></div><p>This can strongly influence just what kinds of tasks AI systems can or cannot do. Some AI users might be able to elicit AI capabilities a lot better than I can &#8212; this in turn impacts how AI changes their work, and how far AI is from taking their jobs.</p><p>There&#8217;s clearly also a ton of variation across different occupations, which you can see from resources like <a href="https://cdn.openai.com/pdf/a253471f-8260-40c6-a2cc-aa93fe9f142e/economic-research-chatgpt-usage-paper.pdf">How People Use ChatGPT</a> and the <a href="https://www.anthropic.com/economic-index">Anthropic Economic Index</a>, or even just by asking your friends. So unless you and I work on similar things, I can&#8217;t say much about AI timelines on your specific job.</p><p>The solution is to have lots of people from various AI-exposed fields do a similar analysis to what I&#8217;ve done &#8212; that means economists, lawyers, mathematicians, and so on. The procedure is simple: pick out three work tasks that you do regularly (and ideally spend a large fraction of your time on), and then spend a bit of time getting AI to attempt the task. Document how much progress the AI makes, look at the bottlenecks, and make your forecasts for when they&#8217;ll be overcome.</p><p>I&#8217;m hoping that this&#8217;ll give a lot more concrete evidence about how AI can be used and is bottlenecked in truly real-world tasks, especially in your own personal situation. So it might be worthwhile trying to get AI to automate your job &#8212; you might learn things about both your life decisions and the future of labor automation. That&#8217;s something that even a million-dollar benchmark may not be able to do.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://epochai.substack.com/p/how-close-is-ai-to-taking-my-job?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:&quot;button-wrapper&quot;}" data-component-name="ButtonCreateButton"><a class="button primary button-wrapper" href="https://epochai.substack.com/p/how-close-is-ai-to-taking-my-job?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><p><em>I&#8217;d like to thank JS Denain, David Owen, Greg Burnham, Luke Emberson, Markov Grey, Jaime Sevilla, and Lynette Bye for helpful feedback and support.</em></p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://epochai.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading! Subscribe to receive the latest from Epoch AI.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>I don&#8217;t know of an official source for how much this cost, but we can do a ballpark estimate: The full <a href="https://arxiv.org/abs/2510.04374v1">benchmark</a> has 1320 tasks, each of which needs 7 hours of expert time to complete (on average). Assuming that experts are paid around $75 per hour, that adds up to around $700,000, just accounting for the costs of measuring human performance. But this ignores costs from building the benchmark, operations, and multiple rounds of expert review, which I think are likely to raise the costs into the millions.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>Hat tip to J&#233;r&#233;my and Tangui for the inspiration behind this question.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>So far, the code for both the model and the web implementation have been private or shared with a handful of collaborators, so I think the chances of test set leakage are quite low. Note also that this task isn&#8217;t 100% the same as the one that our web dev team had to deal with &#8212; it&#8217;s easier in some ways (e.g. there&#8217;s a completed paper and webpage which the AI can reference), but it&#8217;s harder in some other ways (e.g. it doesn&#8217;t have the benefit of iterating many times with the people who designed the economic model).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>Funnily enough, this trajectory probably seems more plausible to most economists, who are extremely skeptical about the plausibility of explosive economic growth. That said, even to them, the GWP growth rate should look quite odd &#8212; it starts off at around 5-6% without much automation, then goes down to 3% with more automation, and then goes back up to 5-6% around the time of full automation!</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>To give a sense of where this number comes from, the current time horizon at a 50% success rate is around <a href="https://metr.org/blog/2025-03-19-measuring-ai-ability-to-complete-long-tasks/">6 hours</a> &#8212; Claude Opus 4.5&#8217;s time horizon is 5.3 hours, and GPT-5.2 (high)&#8217;s time horizon is <a href="https://x.com/METR_Evals/status/2019169900317798857">6.6 hours</a>. Each year we might see roughly two doublings of the time horizon, so by the end of the year this might be about a day long.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>I gave it links to several of our most recent Gradient Updates articles, and asked it to look at them to get a sense of the formatting and style. I also asked it not to look at any other Gradient Updates, so that it wouldn&#8217;t look at the already-published version of the target post &#8212; probably there&#8217;s a smarter way of doing this but I think it should do the trick for the sake of this post.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>Here&#8217;s a rough justification: according to the Bureau of Labor Statistics, there are <a href="https://www.bls.gov/ooh/media-and-communication/writers-and-authors.htm">135,400</a> &#8220;writers and authors&#8221; jobs, <a href="https://www.bls.gov/ooh/Media-and-Communication/Technical-writers.htm">56,400</a> &#8220;technical writer&#8221; jobs, <a href="https://www.bls.gov/ooh/media-and-communication/editors.htm">115,800</a> &#8220;editors&#8221;, and  <a href="https://www.bls.gov/ooh/media-and-communication/reporters-correspondents-and-broadcast-news-analysts.htm">49,300</a> &#8220;news analysts, reporters, and journalists&#8221;. In total that&#8217;s around 350,000 writers, and if we eyeball the numbers in these links the median pay is probably in the ballpark of $70,000.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-8" href="#footnote-anchor-8" class="footnote-number" contenteditable="false" target="_self">8</a><div class="footnote-content"><p>If you know of a better way to do this, <em>please</em> let me know.</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[Can AI companies become profitable?]]></title><description><![CDATA[Lessons from GPT-5&#8217;s economics]]></description><link>https://epochai.substack.com/p/can-ai-companies-become-profitable</link><guid isPermaLink="false">https://epochai.substack.com/p/can-ai-companies-become-profitable</guid><dc:creator><![CDATA[Jaime Sevilla]]></dc:creator><pubDate>Wed, 28 Jan 2026 23:50:19 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!aqoY!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c851485-87e2-4efa-ad20-02255014ecd8_1280x1280.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>This post was written in collaboration with <a href="https://www.exponentialview.co/">Exponential View</a>. It is part of Epoch AI&#8217;s <a href="https://epoch.ai/gradient-updates">Gradient Updates</a> newsletter, which shares more opinionated or informal takes on big questions in AI progress. These posts solely represent the views of the authors, and do not necessarily reflect the views of Epoch AI as a whole.</em></p><p>Originally posted on <a href="https://epoch.ai/gradient-updates/can-ai-companies-become-profitable">Epoch AI</a>.</p><p><em><strong>Update (March 6, 2026):</strong> We&#8217;ve revised our estimates based on new information and feedback from people familiar with the matter. In particular, we&#8217;ve 1) increased our estimate of inference costs given new information, and 2) lowered our estimate of sales and marketing spending after excluding inference compute costs associated with serving free users. This article reflects these updated figures.</em></p><div><hr></div><p>Are AI models profitable? If you ask <a href="https://www.axios.com/2025/08/15/sam-altman-gpt5-launch-chatgpt-future">Sam Altman</a> and <a href="https://open.substack.com/pub/cheekypint/p/a-cheeky-pint-with-anthropic-ceo?r=5erk95&amp;selection=97f244b2-8e86-48a6-a4d4-ea366c90bc83&amp;utm_campaign=post-share-selection&amp;utm_medium=web&amp;aspectRatio=instagram&amp;textColor=%23ffffff&amp;bgImage=true">Dario Amodei</a>, the answer seems to be yes &#8212; it just doesn&#8217;t appear that way on the surface.</p><p>Here&#8217;s the idea: running each AI model generates enough revenue to cover its own R&amp;D costs. But that surplus gets outweighed by the costs of developing the <em>next</em> big model. So, despite making money on each model, companies can lose money each year.</p><p>This is big if true. In fast-growing tech sectors, investors typically accept losses today in exchange for big profits down the line. So if AI models are already covering their own costs, that would paint a healthy financial outlook for AI companies.</p><p>But we can&#8217;t take Altman and Amodei at their word &#8212; you&#8217;d expect CEOs to paint a rosy picture of their company&#8217;s finances. And even if they&#8217;re right, we don&#8217;t know just <em>how</em> profitable models are.</p><p>To shed light on this, we looked into a notable case study: using public reporting on OpenAI&#8217;s finances,<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> we made an educated guess on the profits from running GPT-5, and whether that was enough to recoup its R&amp;D costs. Here&#8217;s what we found:</p><ul><li><p><strong>Whether GPT-5 was profitable to run depends on which profit margin you&#8217;re talking about. </strong>If we subtract the cost of compute from revenue to calculate the gross margin (on an accounting basis),<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> it seems to be about 30% &#8212; lower than the norm for software companies (where 60-80% is typical) but still higher than many industries.</p></li><li><p>But if you also subtract other operating costs, including salaries and marketing, <strong>then OpenAI just barely broke even</strong>, even without including R&amp;D. And they likely ended up at a loss after accounting for their revenue-sharing deal with Microsoft.</p></li><li><p><strong>Moreover, OpenAI likely failed to recoup the costs of developing GPT-5 during its 4-month lifetime. </strong>Even using gross profit, GPT-5&#8217;s tenure was too short to bring in enough revenue to offset OpenAI&#8217;s R&amp;D costs in the four months prior to launch. So if GPT-5 is at all representative, then at least for now, developing and running AI models is loss-making.</p></li></ul><p>This doesn&#8217;t necessarily mean that models like GPT-5 are a bad investment. Even an unprofitable model demonstrates progress, which attracts customers and helps labs raise money to train future models &#8212; and that next generation may earn far more. What&#8217;s more, the R&amp;D that went into GPT-5 likely informs future models like GPT-6. So these labs might have a much better financial outlook than it might initially seem.</p><p>Let&#8217;s dig into the details.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://epochai.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://epochai.substack.com/subscribe?"><span>Subscribe now</span></a></p><h2>Part I: How profitable is running AI models?</h2><p>To answer this question, we consider a case study which we call the &#8220;GPT-5 bundle&#8221;.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> This includes all of OpenAI&#8217;s offerings available during GPT-5&#8217;s lifetime as the flagship model &#8212; GPT-5 and GPT-5.1, GPT-4o, ChatGPT, the API, and so on.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a> We then estimate the revenue and costs of running the bundle.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a></p><p>Revenue is relatively straightforward: since the bundle includes all of OpenAI&#8217;s models, this is just <a href="https://epoch.ai/data/ai-companies#explore-the-data">their total revenue</a> over GPT-5&#8217;s lifetime, from August to December last year.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a> This works out to $6 billion.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a></p><p>At first glance, $6 billion sounds healthy, until you juxtapose it with the costs of running the GPT-5 bundle. These costs come from four main sources:<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-8" href="#footnote-8" target="_self">8</a></p><ol><li><p><strong>Inference compute</strong>: <strong>$4 billion</strong>. This is based on <a href="https://www.theinformation.com/articles/openai-boost-revenue-forecasts-predicts-112-billion-cash-burn-2030">public</a> <a href="https://www.theinformation.com/articles/openai-spend-100-billion-backup-servers-ai-breakthroughs?rc=spkbjw">estimates</a> of OpenAI&#8217;s total inference compute spend in 2025, and assuming that the allocation of compute during GPT-5&#8217;s tenure was proportional to the fraction of the year&#8217;s revenue raised in that period.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-9" href="#footnote-9" target="_self">9</a> This cost covers serving both paid and free users.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-10" href="#footnote-10" target="_self">10</a></p></li><li><p><strong>Staff compensation: $1 billion</strong>, which we can back out from <a href="https://epoch.ai/data/ai-companies?dataView=staff&amp;yAxis=Staff+count#explore-the-data">OpenAI staff counts</a>, reports on <a href="https://www.wsj.com/tech/ai/openai-is-paying-employees-more-than-any-major-tech-startup-in-history-23472527">stock compensation</a>, and things like <a href="https://h1bgrader.com/h1b-sponsors/openai-opco-llc-60d97wl6k8/salaries/2025">H1B filings</a>. One big uncertainty with this: how much of the stock compensation goes toward running models, rather than R&amp;D? We assume 40%, matching the fraction of <em><a href="https://www.theinformation.com/articles/openai-spend-100-billion-backup-servers-ai-breakthroughs">compute</a></em> that goes to inference. Whether staffing follows the same split is uncertain, but it&#8217;s our best guess.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-11" href="#footnote-11" target="_self">11</a><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-12" href="#footnote-12" target="_self">12</a></p></li><li><p><strong>Sales and marketing (S&amp;M): $0.5 billion</strong>. This would include expenses such as their SuperBowl ad and business sale campaigns, promotions and discounts, and other related expenses. Notably, it excludes the compute costs of serving free users.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-13" href="#footnote-13" target="_self">13</a><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-14" href="#footnote-14" target="_self">14</a></p></li><li><p><strong>Legal, office, and administrative costs: $0.2 billion</strong>, assuming this grew between 1.6&#215; and 2&#215; relative to their <a href="https://www.theinformation.com/articles/openai-projections-imply-losses-tripling-to-14-billion-in-2026">2024 expenses</a>. This accounts for <a href="https://aibusiness.com/responsible-ai/openai-signs-partnership-with-uk-government-plans-office-expansion">office expansions</a>, <a href="https://www.reuters.com/world/asia-pacific/openai-open-office-seoul-amid-growing-demand-chatgpt-2025-05-26/">new office setups</a>, and rising <a href="https://www.reuters.com/world/asia-pacific/openai-open-office-seoul-amid-growing-demand-chatgpt-2025-05-26/">administrative costs</a> with their growing workforce.</p></li></ol><p>So what are the profits? One option is to look at gross profits. This only considers the direct cost of running a model, which in this case is just the inference compute cost of $4 billion. Since the revenue was $6 billion, this leads to a profit of around $2 billion, or gross profit margin of around 30%.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-15" href="#footnote-15" target="_self">15</a> This is lower than other software businesses (typically <a href="https://www.key.com/content/dam/kco/documents/businesses___institutions/2024_kbcm_sapphire_saas_survey.pdf">70-80%</a>) but high enough to eventually build a business on.</p><p>On the other hand, if we add up all four cost types, we get close to $6 billion. That&#8217;s close to the revenue, so on operating profit terms, we think the GPT-5 bundle just about broke even.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-16" href="#footnote-16" target="_self">16</a></p><p>Stress-testing the analysis with more aggressive or conservative assumptions doesn&#8217;t change the picture much:<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-17" href="#footnote-17" target="_self">17</a></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!sHkA!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ea97ebf-f57b-4855-b0a3-6aff602b280f_1256x404.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!sHkA!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ea97ebf-f57b-4855-b0a3-6aff602b280f_1256x404.png 424w, https://substackcdn.com/image/fetch/$s_!sHkA!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ea97ebf-f57b-4855-b0a3-6aff602b280f_1256x404.png 848w, https://substackcdn.com/image/fetch/$s_!sHkA!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ea97ebf-f57b-4855-b0a3-6aff602b280f_1256x404.png 1272w, https://substackcdn.com/image/fetch/$s_!sHkA!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ea97ebf-f57b-4855-b0a3-6aff602b280f_1256x404.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!sHkA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ea97ebf-f57b-4855-b0a3-6aff602b280f_1256x404.png" width="1256" height="404" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8ea97ebf-f57b-4855-b0a3-6aff602b280f_1256x404.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:404,&quot;width&quot;:1256,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:102920,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://epochai.substack.com/i/186119616?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ea97ebf-f57b-4855-b0a3-6aff602b280f_1256x404.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!sHkA!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ea97ebf-f57b-4855-b0a3-6aff602b280f_1256x404.png 424w, https://substackcdn.com/image/fetch/$s_!sHkA!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ea97ebf-f57b-4855-b0a3-6aff602b280f_1256x404.png 848w, https://substackcdn.com/image/fetch/$s_!sHkA!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ea97ebf-f57b-4855-b0a3-6aff602b280f_1256x404.png 1272w, https://substackcdn.com/image/fetch/$s_!sHkA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ea97ebf-f57b-4855-b0a3-6aff602b280f_1256x404.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Confidence intervals are obtained from a Monte Carlo analysis. Margins are rounded to the nearest 5%.</em></figcaption></figure></div><p>And there&#8217;s one more hiccup: OpenAI signed a deal with Microsoft to hand over about <a href="https://www.theinformation.com/articles/openai-gain-50-billion-cutting-revenue-share-microsoft-partners?rc=spkbjw">20%</a> of their $6 billion revenue,<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-18" href="#footnote-18" target="_self">18</a> making them end up at a loss.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-19" href="#footnote-19" target="_self">19</a> This doesn&#8217;t mean that the revenue deal is entirely harmful to OpenAI &#8212; for example, Microsoft also shares revenue back to OpenAI.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-20" href="#footnote-20" target="_self">20</a> And the deal probably shouldn&#8217;t significantly affect how we see model profitability &#8212; it seems more to do with OpenAI&#8217;s economic structure rather than something fundamental to AI models. But the fact that OpenAI and Microsoft <a href="https://techcrunch.com/2025/03/05/u-k-s-competition-authority-says-microsofts-openai-partnership-doesnt-quality-for-investigation/">have been</a> <a href="https://blogs.microsoft.com/blog/2025/10/28/the-next-chapter-of-the-microsoft-openai-partnership/">renegotiating</a> this deal suggests it&#8217;s a real drag on OpenAI&#8217;s path to profitability.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!teMP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2c55a78-fa4f-47c7-97b5-18628bfe3b0f_5120x5120.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!teMP!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2c55a78-fa4f-47c7-97b5-18628bfe3b0f_5120x5120.png 424w, https://substackcdn.com/image/fetch/$s_!teMP!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2c55a78-fa4f-47c7-97b5-18628bfe3b0f_5120x5120.png 848w, https://substackcdn.com/image/fetch/$s_!teMP!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2c55a78-fa4f-47c7-97b5-18628bfe3b0f_5120x5120.png 1272w, https://substackcdn.com/image/fetch/$s_!teMP!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2c55a78-fa4f-47c7-97b5-18628bfe3b0f_5120x5120.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!teMP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2c55a78-fa4f-47c7-97b5-18628bfe3b0f_5120x5120.png" width="1456" height="1456" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b2c55a78-fa4f-47c7-97b5-18628bfe3b0f_5120x5120.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1456,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:544880,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://epochai.substack.com/i/186119616?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2c55a78-fa4f-47c7-97b5-18628bfe3b0f_5120x5120.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!teMP!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2c55a78-fa4f-47c7-97b5-18628bfe3b0f_5120x5120.png 424w, https://substackcdn.com/image/fetch/$s_!teMP!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2c55a78-fa4f-47c7-97b5-18628bfe3b0f_5120x5120.png 848w, https://substackcdn.com/image/fetch/$s_!teMP!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2c55a78-fa4f-47c7-97b5-18628bfe3b0f_5120x5120.png 1272w, https://substackcdn.com/image/fetch/$s_!teMP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2c55a78-fa4f-47c7-97b5-18628bfe3b0f_5120x5120.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In short, running AI models is likely profitable in the sense of having decent gross margins. But OpenAI&#8217;s operating margin, which includes marketing and staffing, is likely close to zero. For a fast-growing company, though, operating margins can be misleading &#8212; S&amp;M costs typically grow sublinearly with revenue, so gross margins are arguably a better proxy for long-run profitability.</p><p>So our numbers don&#8217;t necessarily contradict Altman and Amodei yet. But so far we&#8217;ve only seen half the story &#8212; we still need to account for R&amp;D costs, which we&#8217;ll turn to now.</p><h2>Part II: Are models profitable over their lifecycle?</h2><p>Let&#8217;s say we buy the argument that we should look at gross margins. On those terms, it was profitable to run the GPT-5 bundle. But was it profitable enough to recoup the costs of developing it?</p><p>In theory, yes &#8212; you just have to keep running them, and sooner or later you&#8217;ll earn enough revenue to recoup these costs. But in practice, models might have too short a lifetime to make enough revenue. For example, they could be outcompeted by products from rival labs, forcing them to be replaced.</p><p>So to figure out the answer, let&#8217;s go back to the GPT-5 bundle. We&#8217;ve already figured out its gross profits to be around $2 billion. So how do these compare to its R&amp;D costs?</p><p>Estimating this turns out to be a finicky business. We estimate that OpenAI spent $15 billion on R&amp;D in 2025,<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-21" href="#footnote-21" target="_self">21</a> but there&#8217;s no conceptually clean way to attribute some fraction of this to the GPT-5 bundle. We&#8217;d need to make several arbitrary choices: should we count the R&amp;D effort that went into earlier reasoning models, like o1 and o3? Or what if experiments failed, and didn&#8217;t directly change how GPT-5 was trained? Depending on how you answer these questions, the development cost could vary significantly.</p><p>But we can still do an illustrative calculation: let&#8217;s conservatively assume that OpenAI started R&amp;D on GPT-5 after o3&#8217;s <a href="https://openai.com/index/introducing-o3-and-o4-mini/">release</a> last April. Then there&#8217;d still be four months between then and GPT-5&#8217;s release in August,<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-22" href="#footnote-22" target="_self">22</a> during which OpenAI spent around $5 billion on R&amp;D.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-23" href="#footnote-23" target="_self">23</a> But that&#8217;s <em>still</em> higher than the $2 billion of gross profits. In other words, <strong>OpenAI spent more on R&amp;D in the four months preceding GPT-5, than it made in gross profits during GPT-5&#8217;s four-month tenure.</strong><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-24" href="#footnote-24" target="_self">24</a></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!mrWO!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6a736f8b-7579-4706-bb1d-37d479827003_4096x5120.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!mrWO!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6a736f8b-7579-4706-bb1d-37d479827003_4096x5120.png 424w, https://substackcdn.com/image/fetch/$s_!mrWO!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6a736f8b-7579-4706-bb1d-37d479827003_4096x5120.png 848w, https://substackcdn.com/image/fetch/$s_!mrWO!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6a736f8b-7579-4706-bb1d-37d479827003_4096x5120.png 1272w, https://substackcdn.com/image/fetch/$s_!mrWO!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6a736f8b-7579-4706-bb1d-37d479827003_4096x5120.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!mrWO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6a736f8b-7579-4706-bb1d-37d479827003_4096x5120.png" width="1456" height="1820" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6a736f8b-7579-4706-bb1d-37d479827003_4096x5120.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1820,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:454685,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://epochai.substack.com/i/186119616?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6a736f8b-7579-4706-bb1d-37d479827003_4096x5120.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!mrWO!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6a736f8b-7579-4706-bb1d-37d479827003_4096x5120.png 424w, https://substackcdn.com/image/fetch/$s_!mrWO!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6a736f8b-7579-4706-bb1d-37d479827003_4096x5120.png 848w, https://substackcdn.com/image/fetch/$s_!mrWO!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6a736f8b-7579-4706-bb1d-37d479827003_4096x5120.png 1272w, https://substackcdn.com/image/fetch/$s_!mrWO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6a736f8b-7579-4706-bb1d-37d479827003_4096x5120.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>So in practice, it seems like model tenures might indeed be too short to recoup R&amp;D costs. Indeed, GPT-5&#8217;s short tenure was driven by external competition &#8212; <a href="https://www.wired.com/story/openai-gpt-launch-gemini-code-red/">Gemini 3 Pro</a> had arguably surpassed the GPT-5 base model within three months.</p><p>One way to think about this is to treat frontier models like <strong>rapidly-depreciating infrastructure</strong>: their value must be extracted before competitors or successors render them obsolete. So to evaluate AI products, we need to look at both profit margins in inference as well as the time it takes for users to migrate to something better. In the case of the GPT-5 bundle, we find that it&#8217;s decidedly unprofitable over its full lifecycle, even from a gross margin perspective.</p><h2>Part III: Will AI models become profitable?</h2><p>So the finances of the GPT-5 bundle are less rosy than Altman and Amodei suggest. And while we don&#8217;t have as much direct evidence on other models from other labs, they&#8217;re plausibly in a similar boat &#8212; for instance, Anthropic has <a href="https://www.theinformation.com/articles/anthropic-lowers-profit-margin-projection-revenue-skyrockets">reported</a> similar gross margins to OpenAI. So it&#8217;s worth thinking about what it means if the GPT-5 bundle is at all representative of other models.</p><p>The most crucial point is that these model lifecycle losses aren&#8217;t necessarily cause for alarm. AI models don&#8217;t need to be profitable <em>today</em>, as long as companies can convince investors that they will be in the future. That&#8217;s standard for fast-growing tech companies.</p><p>Early on, investors value growth over profit, believing that once a company has captured the market, they&#8217;ll eventually figure out how to make it profitable. The archetypal example of this is Uber &#8212; they accumulated a <a href="https://www.sec.gov/Archives/edgar/data/1543151/000154315123000010/uber-20221231.htm">$32.5 billion deficit</a> over 14 years of net losses, before their first profitable year in 2023. By that measure, OpenAI is thriving: revenues are tripling annually, and <a href="https://epoch.ai/gradient-updates/openai-is-projecting-unprecedented-revenue-growth">projections</a> show continued growth. If that trajectory holds, profitability looks very likely.</p><p>And there are reasons to even be <em>really</em> bullish about AI&#8217;s long-run profitability &#8212; most notably, the sheer scale of value that AI could create. <a href="https://www.wired.com/story/sam-altman-says-the-gpt-5-haters-got-it-all-wrong/">Many</a> <a href="https://www.darioamodei.com/essay/machines-of-loving-grace#:~:text=However%2C%20I%20do%20think%20in%20the%20long%20run%20AI%20will%20become%20so%20broadly%20effective%20and%20so%20cheap%20that%20this%20will%20no%20longer%20apply.%20At%20that%20point%20our%20current%20economic%20setup%20will%20no%20longer%20make%20sense%2C%20and%20there%20will%20be%20a%20need%20for%20a%20broader%20societal%20conversation%20about%20how%20the%20economy%20should%20be%20organized.">higher</a>-<a href="https://www.youtube.com/watch?v=PqVbypvxDto&amp;t=2309s">ups</a> <a href="https://x.com/elonmusk/status/1980765809338147193">at</a> AI companies expect AI systems to outcompete humans across virtually all economically valuable tasks. If you truly believe that in your heart of hearts, that means potentially capturing <a href="https://epoch.ai/epoch-after-hours/ai-in-2030#:~:text=So%20I%20think%20one,trying%20to%20capture%20that.">trillions of dollars</a> from labor automation. The resulting revenue growth could dwarf development costs even with thin margins and short model lifespans.</p><p>That&#8217;s a big leap, and some investors won&#8217;t buy the vision. Or they might doubt that massive revenue growth automatically means huge profits &#8212; what if R&amp;D costs scale up like revenue? These investors might pay special attention to the profit margins of current AI, and want a more concrete picture of how AI companies could be profitable in the near term.</p><p>There&#8217;s an answer for these investors, too. Even if you doubt that AI will become good enough to spark the intelligence explosion or <a href="https://www.darioamodei.com/essay/machines-of-loving-grace#1-biology-and-health">double human lifespans</a>, there are still ways that AI companies could turn a profit. For example, OpenAI is now <a href="https://x.com/openai/status/2012223373489614951">rolling out ads</a> to some ChatGPT users, which could add between <a href="https://www.theinformation.com/articles/openais-international-conundrum">$2 to 15 billion</a> in yearly revenue even without any user growth.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-25" href="#footnote-25" target="_self">25</a> They&#8217;re moving beyond individual consumers and increasingly <a href="https://www.youtube.com/watch?v=tUVSuFT301U">leaning on enterprise adoption</a>. Algorithmic innovations mean that running models could get many times <a href="https://epoch.ai/data-insights/llm-inference-price-trends">cheaper</a> each year, and <a href="https://www.exponentialview.co/p/ai-is-ready-is-your-company">possibly much faster</a>. And there&#8217;s still a lot of <a href="https://www.exponentialview.co/p/can-openai-reach-100-billion-by-2027">room to grow</a> their user base and usage intensity &#8212; for example, ChatGPT has close to <a href="https://epochai.substack.com/p/the-changing-drivers-of-llm-adoption">a billion users</a>, compared to around <a href="https://www.itu.int/itu-d/reports/statistics/2025/10/15/ff25-internet-use/">six billion</a> internet users. Combined, these could add many <a href="https://www.exponentialview.co/p/can-openai-reach-100-billion-by-2027">tens of billions of revenue</a>.</p><p>It won&#8217;t necessarily be easy for AI companies to do this, especially because individual labs will need to come face-to-face with AI&#8217;s &#8220;depreciating infrastructure&#8221; problem. In practice, the &#8220;state-of-the-art&#8221; is often <a href="https://www.exponentialview.co/p/openais-narrow-100-billion-path">challenged</a> within months of a model&#8217;s release, and it&#8217;s hard to make a profit from the latest GPT if Claude and Gemini keep drawing users away.</p><p>But this inter-lab competition doesn&#8217;t stop all AI models from being profitable. Profits are often high in oligopolies because consumers have limited alternatives to switch to. One lab could also pull ahead because they have some kind of algorithmic &#8220;secret sauce&#8221;, or they have more compute.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-26" href="#footnote-26" target="_self">26</a> Or they develop <a href="https://epoch.ai/gradient-updates/the-huge-potential-implications-of-long-context-inference">continual learning</a> techniques that make it <a href="https://epoch.ai/epoch-after-hours/luis-garicano-not-so-simple-macroeconomics-of-ai#:~:text=So%20I%20would%20think%20that%20that%20layer%20remains%20quite%20competitive%2C%20with%20one%20caveat%2C%20which%20is%20the%20introduction%20of%20switching%20costs%20through%20memory.%20If%20the%20system%20starts%20to%20remember%20you%20and%20starts%20to%20know%20who%20you%20are%2C%20then%20switching%20systems%20is%20going%20to%20be%20costly.">harder for consumers to switch</a> between model providers.<br></p><p>These competitive barriers can also be circumvented. Companies could form their own niches, and we&#8217;ve already seen that to some degree: Anthropic is <a href="https://www.businessinsider.com/anthropic-ceo-dario-amodei-drags-openai-and-google-code-red-2025-12">pursuing</a> something akin to a &#8220;code is all you need&#8221; mission, Google DeepMind wants to &#8220;<a href="https://www.technologyreview.com/2016/03/31/161234/how-google-plans-to-solve-artificial-intelligence/">solve intelligence</a>&#8221; and use that to solve everything from cancer to climate change, and Meta strives to make <a href="https://www.meta.com/superintelligence/">AI friends too cheap to meter</a>. This lets individual companies gain revenue for longer.</p><p>So will AI models (and hence AI companies) become profitable? We think it&#8217;s very possible. While our analysis of the GPT-5 bundle is more conservative than Altman and Amodei hint at, what matters more is the trend: Compute margins are falling, enterprise deals are stickier, and models can stay relevant longer than the GPT-5 cycle suggests.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://epochai.substack.com/p/can-ai-companies-become-profitable?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://epochai.substack.com/p/can-ai-companies-become-profitable?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><div><hr></div><p><em>We&#8217;d like to thank JS Denain, Josh You, David Owen, Yafah Edelman, Ricardo Pimentel, Marija Gavrilov, Caroline Falkman Olsson, Lynette Bye, Jay Tate, Dwarkesh Patel, Juan Garc&#237;a, Charles Dillon, Brendan Halstead, Isabel Johnson, Markov Gray and Stefania Guerra for their feedback and support on this post. Special thanks to Azeem Azhar for initiating this collaboration and vital input, and Benjamin Todd for in-depth feedback and discussion.</em></p><p><em>Epoch AI has commercial relationships with multiple AI companies, including OpenAI and Anthropic. We disclose these relationships and associated conflicts of interest on <a href="https://epoch.ai/our-funding">our funding page</a>. The views and analysis presented here are independent and not reviewed or endorsed by these companies.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://epochai.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading! Subscribe to receive the latest from Epoch AI.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>Our main sources of information include claims by OpenAI and their staff, and reporting by <a href="https://www.theinformation.com/">The Information</a>, <a href="https://www.cnbc.com/">CNBC</a> and the <a href="https://www.wsj.com/">Wall Street Journal</a>. We&#8217;ve linked our primary sources though the document.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>Technically, gross margins should also account for staff costs that were essential to delivering the product, such as customer service. But these are likely a small fraction of salaries, which are in turn dominated by compute costs &#8212; so it won&#8217;t affect our analysis much, as we&#8217;ll see.</p><p>Similarly, OpenAI&#8217;s revenue sharing agreement with Microsoft should, in theory, be counted as cost of goods sold and included in the gross margin per usual GAAP accounting. But the revenue sharing reflects more how the profits of the operation are divided between OpenAI and Microsoft, rather than an actual cost, so we have excluded it from the analysis.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>We focus on OpenAI models because we have the most financial data available on them.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>Should we include Sora 2 in this bundle? You could argue that we shouldn&#8217;t, because it runs on its own platform and is heavily subsidized to kickstart a <a href="https://www.bloomberg.com/news/newsletters/2025-12-05/enthusiasm-for-openai-s-sora-fades-after-initial-creative-burst">new social network</a>, making its economics quite different. However, we find that it&#8217;s likely a rounding error for revenues, since people <a href="https://www.bloomberg.com/news/newsletters/2025-12-05/enthusiasm-for-openai-s-sora-fades-after-initial-creative-burst">don&#8217;t use it much</a>. In particular, the Sora app had close to <a href="https://sherwood.news/tech/openais-sora-2-started-off-scorching-hot-things-have-slowed-down-since/">9 million downloads</a> by December, compared to around <a href="https://epochai.substack.com/p/the-changing-drivers-of-llm-adoption">900 million weekly active users</a> of ChatGPT.</p><p>Now, while it likely didn&#8217;t make much revenue, it might have been costly to serve &#8212; apparently making <a href="https://openai.com/index/sora-2/">TikTok-esque</a> <a href="https://www.nytimes.com/2025/10/19/opinion/ai-sora-slop.html">AI short-form videos</a> using Sora 2 cost OpenAI several hundred million dollars. Here&#8217;s a rough estimate: In November (when app downloads <a href="https://sherwood.news/tech/openais-sora-2-started-off-scorching-hot-things-have-slowed-down-since/">peaked</a>), Sora 2 had &#8220;<a href="https://sequoiacap.com/podcast/openai-sora-2-team-how-generative-video-will-unlock-creativity-and-world-models/#:~:text=I%20think%20we%20have%20almost%20seven%20million%20generations%20happening%20a%20day">almost seven million generations happening a day</a>&#8221;. Assuming generations were proportional to <a href="https://www.similarweb.com/app/google/com.openai.sora/#overview">weekly active users</a> over time, this would mean 330 million videos in total. The API cost is <a href="https://web.archive.org/web/20260117001407/https://openai.com/api/pricing/">$0.1/s</a>, so if the average video was 10s long, and assuming the API compute profit margin was 20%, this adds up to 330 million &#215; $0.1 &#215; 10 / 1.2 &#8776; $250 million. This is significant, but it&#8217;s minor compared to OpenAI&#8217;s overall inference compute spend.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>Ideally we&#8217;d have only looked at a single model, but we only have data on costs and revenues at the company-level, not at the release-level, so we do the next best thing.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>For the purposes of this post, we assume that GPT-5&#8217;s lifetime started when GPT-5 was released (<a href="https://openai.com/index/introducing-gpt-5/">Aug 7th</a>) and ended when GPT-5.2 was released (<a href="https://openai.com/index/introducing-gpt-5-2/">Dec 11th</a>). That might seem a bit odd &#8212; after all, isn&#8217;t GPT-5.2 based on GPT-5? We thought so too, but GPT-5.2 has a new knowledge cutoff, and is <a href="https://azure.microsoft.com/en-us/blog/introducing-gpt-5-2-in-microsoft-foundry-the-new-standard-for-enterprise-ai/#:~:text=The%20GPT%2D5.2%20series%20is%20built%20on%20new%20architecture">apparently</a> &#8220;built on a new architecture&#8221;, so it might have a different base model from the other models under the GPT-5 moniker.</p><p>Admittedly, we don&#8217;t know for sure that GPT-5.2 uses a different base model, but it&#8217;s a convenient way to bound the timeframe of our analysis. And it shouldn&#8217;t matter much for our estimates of profit margins, because we&#8217;re simply comparing revenues and costs over the same time period.</p><p>Also note that GPT-5 and GPT-5.1 are still available through ChatGPT and OpenAI&#8217;s API, so their useful life hasn&#8217;t strictly ended. We assume for simplicity that usage has been largely displaced by GPT-5.2.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>In July, OpenAI had its first month with over <a href="https://www.cnbc.com/2025/08/20/openai-compute-ai.html">$1 billion</a> in revenue, and it closed the year with an annualized revenue of over <a href="https://openai.com/index/a-business-that-scales-with-the-value-of-intelligence/">$20 billion</a> ($1.7 billion per month). If this grew exponentially, the average revenue over the four months of GPT-5&#8217;s tenure would&#8217;ve been close to $1.5 billion, giving a total of $6 billion during the period.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-8" href="#footnote-anchor-8" class="footnote-number" contenteditable="false" target="_self">8</a><div class="footnote-content"><p>One category of loss we are ignoring is the <a href="https://www.theinformation.com/articles/openais-first-half-results-4-3-billion-sales-2-5-billion-cash-burn">remeasurement of convertible interest rights</a>, which purportedly amounted to half of OpenAI&#8217;s net loss in the first half of 2025. As with the revenue sharing deal, this only affects how profits would be divided between investors, not the product margins, so it is not relevant to the overall profitability of AI.</p><p>Another category of expense we didn&#8217;t account for is the cost of acquisitions. We also consider these not overly relevant to determine the overall profitability of operating OpenAI&#8217;s bundle of products, though it could be relevant to their R&amp;D.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-9" href="#footnote-anchor-9" class="footnote-number" contenteditable="false" target="_self">9</a><div class="footnote-content"><p>Last year, OpenAI earned about <a href="https://www.theinformation.com/articles/openai-boost-revenue-forecasts-predicts-112-billion-cash-burn-2030">$13.1 billion</a> in full year revenue, compared to $6.1 billion for the GPT-5 bundle. At the same time, it was reported they achieved a <a href="https://www.theinformation.com/articles/openai-boost-revenue-forecasts-predicts-112-billion-cash-burn-2030">33% gross margin</a> on inference compute, implying they spent around $8.8 billion running all models last year, so if we assume revenue and inference compute are proportional throughout the year, they spent 6.1 billion / 13.1 billion &#215; 8.8 billion &#8776; $4.1 billion.</p><p>This is then increased by an additional $200 million from other sources of IT spending, including e.g. servers and networking equipment. The total is then still around $4.1 billion + $0.2 billion = $4.3 billion.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-10" href="#footnote-anchor-10" class="footnote-number" contenteditable="false" target="_self">10</a><div class="footnote-content"><p>The Information <a href="https://www.theinformation.com/articles/openai-boost-revenue-forecasts-predicts-112-billion-cash-burn-2030">reported</a> that OpenAI spent around $4.5bn in serving paid users in 2025 of over $8bn spent in inference, implying they spent around $4bn serving free users. That&#8217;s it, they spent 40% to 50% of their compute serving free users.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-11" href="#footnote-anchor-11" class="footnote-number" contenteditable="false" target="_self">11</a><div class="footnote-content"><p><a href="https://h1bgrader.com/h1b-sponsors/openai-opco-llc-60d97wl6k8/salaries/2025">H1B filings</a> suggest an average base salary of $310,000 in 2025, ranging from $150,000 to $685,000. This seems broadly consistent with data from <a href="https://www.levels.fyi/en-gb/companies/openai/salaries">levels.fyi</a>, which reports salaries ranging from $144,275 to $1,274,139 as we&#8217;re writing this. Overall, let&#8217;s go with an average of $310,000 plus around <a href="https://www.bbgbroker.com/cost-of-employee-benefits-2022/">40%</a> <a href="https://web.mit.edu/e-club/hadzima/how-much-does-an-employee-cost.html">in benefits</a>. We also know that OpenAI&#8217;s staff counts surged from <a href="https://calv.info/openai-reflections#footnote-fnref-1">3,000 in mid-2025</a> to <a href="https://www.wsj.com/tech/ai/openai-is-paying-employees-more-than-any-major-tech-startup-in-history-23472527?gaa_at=eafs&amp;gaa_n=AWEtsqe_DtK3xu-WaMFy0UhvovnVcFhPcrhxbGZ63SUUds2O7rDMoNsaWhaI2G4yq5k%3D&amp;gaa_ts=6974ee2a&amp;gaa_sig=w8SmjoETgCEEqEpi_mznXf95P9mcVjOP6AKTUrMiooJctsgLb1eCoGtHJUeAFisybVP-t4MEvAiDqvcLisinmQ%3D%3D">4,000 by the end of 2025</a>. We smoothly interpolate between these to get an average staff count of 3,500 employees during GPT-5&#8217;s lifetime.</p><p>Then the base salary comes to: 3,500 employees &#215; $310,000 base salary &#215; 1.4 benefits &#215; 40% share of employees working on serving GPT-5 &#215; 127 / 365 period serving &#8776; $0.2 billion (the 127 comes from the number of days in GPT-5&#8217;s lifetime).</p><p>We then need to account for stock compensation. In 2025, OpenAI awarded <a href="https://www.wsj.com/tech/ai/openai-is-paying-employees-more-than-any-major-tech-startup-in-history-23472527">$6 billion to employees</a> in stock compensation. Assuming they awarded them proportionally to staff count over the year, and given the exponential increase of staff counts, that would indicate that over 42% of the stock was awarded during GPT-5&#8217;s lifetime. Assuming 40% goes to operations as before, that results in $6 billion x 42% x 40% = $1 billion stock expense for operating the GPT-5 bundle. The total staff compensation would then be around $1.2 billion.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-12" href="#footnote-anchor-12" class="footnote-number" contenteditable="false" target="_self">12</a><div class="footnote-content"><p>It&#8217;s debatable whether the <a href="https://www.cnbc.com/2025/09/06/ai-talent-war-tech-giants-pay-talent-millions-of-dollars.html">very high compensation packages</a> for technical staff will continue as the industry matures.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-13" href="#footnote-anchor-13" class="footnote-number" contenteditable="false" target="_self">13</a><div class="footnote-content"><p>This is the line item we are most uncertain about. The Information reported that <a href="https://www.theinformation.com/articles/openais-first-half-results-4-3-billion-sales-2-5-billion-cash-burn">OpenAI had spent $2bn in sales and marketing during the first half of 2025</a>, &#8220;nearly doubling what it spent in all of 2024.&#8221; However, they previously reported only <a href="https://www.theinformation.com/articles/openai-projections-imply-losses-tripling-to-14-billion-in-2026">$300mn in S&amp;M spending during 2024</a>. Our explanation of the discrepancy is that the 2025 H1 S&amp;M spending includes the cost of inference for free users, OpenAI&#8217;s primary way of advertising their products, while the 2024 figure does not.</p><p>To estimate the S&amp;M spending excluding compute, we assume it grew proportionally to revenue compared to their 2024 expenses. This would result in an estimate of $300mn 2024 S&amp;M budget x $6.1bn GPT-5 tenure revenue / <a href="https://www.theinformation.com/articles/openai-boost-revenue-forecasts-predicts-112-billion-cash-burn-2030">$4bn 2024 revenue</a> = $500mn.</p><p>It&#8217;s possible this could have grown much more given OpenAI&#8217;s growing focus on B2B, which could be especially S&amp;M expensive and include hefty commissions for closed deals.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-14" href="#footnote-anchor-14" class="footnote-number" contenteditable="false" target="_self">14</a><div class="footnote-content"><p>This corresponds to around 10% of revenue during the period, which <a href="https://www.benchmarkit.ai/2025benchmarks">isn&#8217;t</a> <a href="https://www.saas-capital.com/blog-posts/spending-benchmarks-for-private-b2b-saas-companies/">unusual</a> compared to other large software companies. For example, <a href="https://www.adobe.com/cc-shared/assets/investor-relations/pdfs/adbe-2024-annual-report.pdf">Adobe</a>, <a href="https://investors.intuit.com/sec-filings/all-sec-filings/content/0000896878-25-000035/intu-20250731.htm">Intuit</a>, <a href="https://www.sec.gov/Archives/edgar/data/1108524/000110852425000006/crm-20250131.htm">Salesforce</a> and <a href="https://www.sec.gov/Archives/edgar/data/1373715/000137371525000010/now-20241231.htm">ServiceNow</a> all spent around 27% to 35% of their 2024-2025 revenue in S&amp;M. That said, there are certainly examples with lower spends &#8212; for example, <a href="https://www.microsoft.com/investor/reports/ar25/index.html">Microsoft</a> and <a href="https://d18rn0p25nwr6d.cloudfront.net/CIK-0001341439/7455eba6-bb80-41d3-96b7-12111eae648c.pdf">Oracle</a> spend 9 to 15% of their revenue on marketing, though note that these are relatively mature firms &#8212; younger firms may spend higher fractions on S&amp;M.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-15" href="#footnote-anchor-15" class="footnote-number" contenteditable="false" target="_self">15</a><div class="footnote-content"><p>In comparison, Anthropic projected in mid-December a gross margin <a href="https://www.theinformation.com/articles/anthropic-lowers-profit-margin-projection-revenue-skyrockets">around 40%</a>, suggesting gross margins in the 30-40% range might be representative of the industry.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-16" href="#footnote-anchor-16" class="footnote-number" contenteditable="false" target="_self">16</a><div class="footnote-content"><p>How does this compare to previous years? <a href="https://www.theinformation.com/articles/openai-projections-imply-losses-tripling-to-14-billion-in-2026">The Information</a> reported that in 2024 OpenAI made $4 billion in revenue, and spent $2.4 billion in inference compute and hosting, $700 million in employee salaries, $600 million in G&amp;A, and $300 million in S&amp;M. This implies a gross margin of 40% and an operating margin of 0% (excluding stock compensation).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-17" href="#footnote-anchor-17" class="footnote-number" contenteditable="false" target="_self">17</a><div class="footnote-content"><p>In broad strokes, we perform a sensitivity analysis by considering a range of possible values for each cost component, then sampling from each to consider a range of plausible scenarios (a Monte Carlo analysis). The largest uncertainties that feed into this analysis are how much staff compensation goes to inference instead of R&amp;D, S&amp;M spending in the second half of 2025, and revenue during GPT-5&#8217;s tenure.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-18" href="#footnote-anchor-18" class="footnote-number" contenteditable="false" target="_self">18</a><div class="footnote-content"><p>Two more caveats to add: first, this 20% rate isn&#8217;t publicly confirmed by OpenAI or Microsoft, at least in our knowledge. Second, the revenue sharing agreement is also <a href="https://blogs.microsoft.com/blog/2025/10/28/the-next-chapter-of-the-microsoft-openai-partnership/">more</a> <a href="https://openai.com/index/next-chapter-of-microsoft-openai-partnership/">complex</a> than just this number. Microsoft put a lot of money and compute into OpenAI, and in return it gets a significant ownership stake, special rights to use OpenAI&#8217;s technology, and some of OpenAI&#8217;s revenue. There also isn&#8217;t a single well-defined &#8220;end date&#8221;: some rights are set to last into the early 2030s, while other parts (including revenue sharing) continue until an independent panel confirms OpenAI has reached &#8220;AGI&#8221;.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-19" href="#footnote-anchor-19" class="footnote-number" contenteditable="false" target="_self">19</a><div class="footnote-content"><p>Strictly speaking, a revenue share agreement is often seen as an expense that would impact gross margins. But we&#8217;re more interested in the unit economics that generalize across models, rather than those that are unique to OpenAI&#8217;s financial situation.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-20" href="#footnote-anchor-20" class="footnote-number" contenteditable="false" target="_self">20</a><div class="footnote-content"><p>The deal was signed in <a href="https://news.microsoft.com/source/2019/07/22/openai-forms-exclusive-computing-partnership-with-microsoft-to-build-new-azure-ai-supercomputing-technologies/">2019</a>, a year before GPT-3 was released, and at this time it may have been an effective way to access compute resources and get commercial distribution. This could&#8217;ve been important for OpenAI to develop GPT-5 in the first place.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-21" href="#footnote-anchor-21" class="footnote-number" contenteditable="false" target="_self">21</a><div class="footnote-content"><p>OpenAI&#8217;s main R&amp;D spending is on compute, salaries and data. In 2025, they spent <a href="https://www.theinformation.com/articles/openai-spend-100-billion-backup-servers-ai-breakthroughs">$9 billion</a> on R&amp;D AI compute, and about <a href="https://www.theinformation.com/articles/anthropic-openai-developing-ai-co-workers?rc=spkbjw">$1 billion</a> on data (which includes paying for human experts and <a href="https://epochai.substack.com/p/an-faq-on-reinforcement-learning">RL environments</a>). We can estimate salary payouts in the same way we did in the previous section on inference, except we consider 60% of staff compensation rather than 40%, resulting in a expense of $4.6 billion. Finally, we add about $400 million in offices and administrative expenses, and $600 million in other compute costs (including e.g. networking costs). This adds up to about $16 billion.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-22" href="#footnote-anchor-22" class="footnote-number" contenteditable="false" target="_self">22</a><div class="footnote-content"><p>In fact, we could be substantially lowballing the R&amp;D costs. GPT-5 has been in the works for a long time &#8212; for example, early reasoning models like <a href="https://openai.com/o1/">o1</a> probably helped develop GPT-5&#8217;s reasoning abilities. GPT-5.1 was probably being developed between August and November, covering a good chunk of the GPT-5 bundle&#8217;s tenure. But there&#8217;s a countervailing consideration: some of the R&amp;D costs for GPT-5 probably help develop <em>future</em> models like &#8220;GPT-6&#8221;. So it&#8217;s hard to say what the exact numbers are, but we&#8217;re pretty confident that our overall point still stands.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-23" href="#footnote-anchor-23" class="footnote-number" contenteditable="false" target="_self">23</a><div class="footnote-content"><p>Because OpenAI&#8217;s expenses are growing exponentially, we can&#8217;t just estimate the share of R&amp;D spending in this period as one third of the annual total. Assuming a 2.3&#215; annual growth rate in R&amp;D expenses &#8212; comparable to the <a href="https://epoch.ai/data/ai-companies?dataView=compute&amp;yAxis=Annual+compute+spend+%28USD%29#explore-the-data">increase in OpenAI&#8217;s R&amp;D compute spending from 2024 to 2025</a> &#8212; the costs incurred between April 16 and August 7 would account for approximately 35% of total yearly R&amp;D expenses.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-24" href="#footnote-anchor-24" class="footnote-number" contenteditable="false" target="_self">24</a><div class="footnote-content"><p>Are our results sensitive to whether we try to account for the costs and revenue of GPT-5.2? In December 2025, OpenAI made around $1.7bn in revenue, corresponding to about $600mn in gross profit, while in February 2026, they made around $2.1bn in revenue, or $700mn in gross profit. So in the three months that GPT-5.2 has been the default model, OpenAI has made about $2bn in gross profit. But of course, GPT-5.2 likely took some non-trivial effort to develop: GPT-5.2 has a <a href="https://developers.openai.com/api/docs/models/gpt-5.2">different knowledge cutoff</a>, a <a href="https://www.lesswrong.com/posts/gJyGQrGrojNnCQZgD?commentId=xPZ6Jxgd9mxtmxyPv">distinct knowledge profile</a>, and was the alleged result of a <a href="https://www.theverge.com/report/838857/openai-gpt-5-2-release-date-code-red-google-response">&#8220;code red&#8221; effort by OpenAI to respond to competitors</a>. If in the four months between April and August OpenAI spent around $5.3bn in R&amp;D, in each subsequent month they likely spent at least $1bn more. So if we think GPT-5.2 took at least two additional months' worth of R&amp;D spending to develop, then that would put the expense to develop GPT-5.2 at over $7bn, while the gross profit of the GPT-5+GPT-5.1+5.2 bundle would currently be at around $4bn. Still not enough to break even, though depending on how tight this is as a lower bound it might get there with three more months of operations.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-25" href="#footnote-anchor-25" class="footnote-number" contenteditable="false" target="_self">25</a><div class="footnote-content"><p>OpenAI was approaching <a href="https://www.theinformation.com/newsletters/ai-agenda/chatgpt-nears-900-million-weekly-active-users-gemini-catching">900 million weekly active users</a> in December last year. For ads, they project a revenue of <a href="https://www.theinformation.com/articles/openais-international-conundrum">$2 per free user for 2026, and up to $15 per free user for 2030</a>. Combining these numbers gives our estimate of around $2 billion to $15 billion.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-26" href="#footnote-anchor-26" class="footnote-number" contenteditable="false" target="_self">26</a><div class="footnote-content"><p>For the investors who are willing to entertain more extreme scenarios, an even stronger effect is when &#8220;<a href="https://situational-awareness.ai/the-free-world-must-prevail/#:~:text=If%20there%20is%20a%20rapid%20intelligence%20explosion%2C%20it%E2%80%99s%20plausible%20a%20lead%20of%20mere%20months%20could%20be%20decisive">intelligence</a> <a href="https://ai-2027.com/#narrative-2027-08-31">explosion</a>&#8221; dynamics kick in &#8212; if OpenAI pulls ahead at the right time, they could use their better AIs to accelerate their own research, <a href="https://epoch.ai/gradient-updates/the-software-intelligence-explosion-debate-needs-experiments">amplifying</a> a small edge into a huge lead. This might sound like science fiction to a lot of readers, but representatives from some AI companies have publicly set these as goals. For instance, Sam Altman <a href="https://www.youtube.com/watch?v=ngDCxlZcecw&amp;t=2110s">claims</a> that one of OpenAI&#8217;s goals is to have a &#8220;true automated AI researcher&#8221; by March 2028.</p></div></div>]]></content:encoded></item><item><title><![CDATA[How well did forecasters predict 2025 AI progress?]]></title><description><![CDATA[Mostly right about benchmarks, mixed results on real-world impacts]]></description><link>https://epochai.substack.com/p/how-well-did-forecasters-predict</link><guid isPermaLink="false">https://epochai.substack.com/p/how-well-did-forecasters-predict</guid><dc:creator><![CDATA[Anson Ho]]></dc:creator><pubDate>Fri, 16 Jan 2026 19:11:46 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!ooY5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02e4ee45-0f87-4f52-a6ea-ca2b06dacee2_836x755.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>This post was written in collaboration between the AI Futures Project, the AI Digest, and Epoch AI. It analyzes the results of the <a href="https://ai2025.org/">2025 AI Digest survey</a>. You can take the 2026 AI forecasting survey <a href="https://forecast2026.ai/">here</a>.</em></p><p><em>This post is also part of Epoch AI&#8217;s <a href="https://epoch.ai/gradient-updates">Gradient Updates</a> newsletter, which shares more opinionated or informal takes on big questions in AI progress. These posts solely represent the views of the authors, and do not necessarily reflect the views of Epoch AI as a whole.</em></p><p><em>Originally posted on <a href="https://epoch.ai/gradient-updates/how-well-did-forecasters-predict-2025-ai-progress">Epoch AI</a>.</em></p><div><hr></div><p>Every other AI paper I read seems to start with some version of &#8220;AI progress has been fast&#8221;. And sure, that&#8217;s obviously true &#8212; a year ago there was no GPT-5, no DeepSeek-R1, and not even Claude 3.7 Sonnet! But few people seem to say <em>exactly</em> how fast things have been, and whether people saw it coming. So when the <a href="https://theaidigest.org/">AI Digest</a> released a <a href="https://ai2025.org/">survey</a> for people to forecast AI progress over the last year, I was excited.</p><p>The survey helps track something akin to an &#8220;<a href="https://ai-2027.com/">AI 2027</a> worldview&#8221;, where automating AI R&amp;D leads to a surge in AI capabilities and hence a range of risks to humanity, especially from handing power off to AI systems. You can see this in the question topics: about half of the survey is about forecasting performance on benchmarks related to AI R&amp;D. The other half looks at real-world impacts &#8212; think AI-enabled cyberattacks, humans losing control of AI, and more &#8220;mundane&#8221; things like revenue and public perception. If this worldview is even partially right (and I think it clears that bar), these questions capture some of the most consequential topics about the future of AI.</p><p>So how did forecasters do on these questions?</p><h2>Demographics: Junior, short-ish timelines, high risk of AI catastrophe</h2><p>To answer this question, we first need to know who these forecasters were. They were primarily recruited through the AI Digest&#8217;s outreach on X and their monthly newsletter. In total there were 421 of them, some with serious forecasting credentials. For example, there&#8217;s regular top forecaster <a href="https://x.com/peterwildeford/status/1873760001464189437">Peter Wildeford</a> (who came in 14th), as well as AI forecasting researcher <a href="https://x.com/ajeya_cotra/status/1867813307073409333">Ajeya Cotra</a> (who came in 3rd).</p><p>The first thing to know about the forecasters is that they were pretty junior. 73% claimed to have some professional or academic experience in AI, of which 80% claimed to have five years of experience or less.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!qzNB!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fedafcd16-6388-455a-b068-1422c6456f60_1024x1280.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!qzNB!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fedafcd16-6388-455a-b068-1422c6456f60_1024x1280.png 424w, https://substackcdn.com/image/fetch/$s_!qzNB!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fedafcd16-6388-455a-b068-1422c6456f60_1024x1280.png 848w, https://substackcdn.com/image/fetch/$s_!qzNB!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fedafcd16-6388-455a-b068-1422c6456f60_1024x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!qzNB!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fedafcd16-6388-455a-b068-1422c6456f60_1024x1280.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!qzNB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fedafcd16-6388-455a-b068-1422c6456f60_1024x1280.png" width="1024" height="1280" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/edafcd16-6388-455a-b068-1422c6456f60_1024x1280.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1280,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!qzNB!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fedafcd16-6388-455a-b068-1422c6456f60_1024x1280.png 424w, https://substackcdn.com/image/fetch/$s_!qzNB!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fedafcd16-6388-455a-b068-1422c6456f60_1024x1280.png 848w, https://substackcdn.com/image/fetch/$s_!qzNB!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fedafcd16-6388-455a-b068-1422c6456f60_1024x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!qzNB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fedafcd16-6388-455a-b068-1422c6456f60_1024x1280.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The next thing to know is that they had very short timelines to &#8220;<a href="https://arxiv.org/pdf/2401.02843">High-Level Machine Intelligence (HLMI)</a>&#8221; &#8212;  that is, &#8220;unaided machines that can accomplish every task better and more cheaply than human workers&#8221;, seemingly including robotics. Half of them expected HLMI to be developed by 2030, and 90% of them expected it by 2040.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!qzFP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb0b001ba-5c82-4c4a-9bda-cf079835ec28_1024x1280.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!qzFP!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb0b001ba-5c82-4c4a-9bda-cf079835ec28_1024x1280.png 424w, https://substackcdn.com/image/fetch/$s_!qzFP!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb0b001ba-5c82-4c4a-9bda-cf079835ec28_1024x1280.png 848w, https://substackcdn.com/image/fetch/$s_!qzFP!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb0b001ba-5c82-4c4a-9bda-cf079835ec28_1024x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!qzFP!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb0b001ba-5c82-4c4a-9bda-cf079835ec28_1024x1280.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!qzFP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb0b001ba-5c82-4c4a-9bda-cf079835ec28_1024x1280.png" width="1024" height="1280" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b0b001ba-5c82-4c4a-9bda-cf079835ec28_1024x1280.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1280,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!qzFP!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb0b001ba-5c82-4c4a-9bda-cf079835ec28_1024x1280.png 424w, https://substackcdn.com/image/fetch/$s_!qzFP!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb0b001ba-5c82-4c4a-9bda-cf079835ec28_1024x1280.png 848w, https://substackcdn.com/image/fetch/$s_!qzFP!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb0b001ba-5c82-4c4a-9bda-cf079835ec28_1024x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!qzFP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb0b001ba-5c82-4c4a-9bda-cf079835ec28_1024x1280.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Thirdly, respondents were pretty &#8220;doom-y&#8221;: three-quarters of them gave over a 10% chance that HLMI would lead to &#8220;extremely bad&#8221; outcomes like human extinction.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!H2W0!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04c301a4-5b89-48ee-907b-287c6555f283_1024x1280.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!H2W0!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04c301a4-5b89-48ee-907b-287c6555f283_1024x1280.png 424w, https://substackcdn.com/image/fetch/$s_!H2W0!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04c301a4-5b89-48ee-907b-287c6555f283_1024x1280.png 848w, https://substackcdn.com/image/fetch/$s_!H2W0!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04c301a4-5b89-48ee-907b-287c6555f283_1024x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!H2W0!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04c301a4-5b89-48ee-907b-287c6555f283_1024x1280.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!H2W0!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04c301a4-5b89-48ee-907b-287c6555f283_1024x1280.png" width="1024" height="1280" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/04c301a4-5b89-48ee-907b-287c6555f283_1024x1280.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1280,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!H2W0!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04c301a4-5b89-48ee-907b-287c6555f283_1024x1280.png 424w, https://substackcdn.com/image/fetch/$s_!H2W0!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04c301a4-5b89-48ee-907b-287c6555f283_1024x1280.png 848w, https://substackcdn.com/image/fetch/$s_!H2W0!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04c301a4-5b89-48ee-907b-287c6555f283_1024x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!H2W0!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04c301a4-5b89-48ee-907b-287c6555f283_1024x1280.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>So to oversimplify, we can think of these forecasters as &#8220;junior AI professionals who expect HLMI by 2040, and think it might lead to horrible consequences&#8221; &#8212; fairly common within the AI safety community. And it&#8217;s worth thinking about what this means: if they&#8217;re right about AI progress and its impacts over the next fifteen-ish years, it could make even the Industrial Revolution seem like a footnote in a history book.</p><p>The million-dollar question: were they right about AI progress over the <em>last</em> year?</p><h2>Benchmarks related to AI R&amp;D: The median forecast was on the money (for the most part)</h2><p>The first half of the survey asked participants to forecast performance on five separate benchmarks, each of which is related to AI R&amp;D. Here&#8217;s how they did:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!B6JU!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff53079b2-35d2-4851-ac30-2a75a594aef0_1798x902.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!B6JU!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff53079b2-35d2-4851-ac30-2a75a594aef0_1798x902.png 424w, https://substackcdn.com/image/fetch/$s_!B6JU!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff53079b2-35d2-4851-ac30-2a75a594aef0_1798x902.png 848w, https://substackcdn.com/image/fetch/$s_!B6JU!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff53079b2-35d2-4851-ac30-2a75a594aef0_1798x902.png 1272w, https://substackcdn.com/image/fetch/$s_!B6JU!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff53079b2-35d2-4851-ac30-2a75a594aef0_1798x902.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!B6JU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff53079b2-35d2-4851-ac30-2a75a594aef0_1798x902.png" width="1456" height="730" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f53079b2-35d2-4851-ac30-2a75a594aef0_1798x902.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:730,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:201230,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://epochai.substack.com/i/184793687?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff53079b2-35d2-4851-ac30-2a75a594aef0_1798x902.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!B6JU!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff53079b2-35d2-4851-ac30-2a75a594aef0_1798x902.png 424w, https://substackcdn.com/image/fetch/$s_!B6JU!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff53079b2-35d2-4851-ac30-2a75a594aef0_1798x902.png 848w, https://substackcdn.com/image/fetch/$s_!B6JU!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff53079b2-35d2-4851-ac30-2a75a594aef0_1798x902.png 1272w, https://substackcdn.com/image/fetch/$s_!B6JU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff53079b2-35d2-4851-ac30-2a75a594aef0_1798x902.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Note: You can see the reasoning for the resolutions in the original survey <a href="https://ai2025.org/">link</a>. [Edit: The OSWorld performance was updated from 66.3% to 72.6% after this post was written.]</figcaption></figure></div><p>For the most part, the median forecast looks pretty good &#8212; they were basically right on RE-Bench and FrontierMath, and fairly close for OSWorld and SWE-Bench Verified. The one exception seems to be Cybench, where the median forecast underestimated progress by 20 percentage points.</p><p>So what went wrong with Cybench? Looking at the few dozen respondents that shared rationales for their forecasts, I&#8217;d guess that there are two main reasons at play. The first is not extrapolating existing benchmark trends, and gesturing to how they expect the benchmark to be quite hard (though it&#8217;s not very clear to me how this translates into quantitative forecasts). The second reason is overrating how reluctant labs would be to report improvements on Cybench &#8212; many respondents thought that high scores on a cybersecurity benchmark would indicate strong cyberattack abilities, putting labs under more scrutiny, so they&#8217;d have an incentive to not report on it.</p><p>Although the forecasters weren&#8217;t as far off, another interesting case is SWE-Bench Verified, where unlike Cybench, forecasters somewhat <em>over</em>estimated progress &#8212; the median forecast was 88% instead of the final 81%. So why did forecasters overshoot? Some of them argued that SWE-Bench&#8217;s coding tasks are especially amenable to reasoning models. Some argued that progress would be fast because many labs optimize their models specifically to do well on the benchmark &#8212; Anthropic often uses it as the <a href="https://www.anthropic.com/news/claude-opus-4-5">first benchmark</a> in new model announcements, and some people even have the <a href="https://job-boards.greenhouse.io/warp/jobs/5542429004">job</a> of hill-climbing it. Finally, some pointed out that the original state-of-the-art was 72%, and appealed to how benchmarks typically saturate quickly once scores start climbing the &#8220;steep part of the s-curve&#8221;. This led to forecasts close to the benchmark&#8217;s saturation point at 90%ish.</p><p>Broadly speaking, I think all of these arguments are right, so it&#8217;s slightly puzzling to me why SWE-Bench Verified scores didn&#8217;t end up closer to 90%. I doubt flawed benchmark questions capped scores to 80%, because <a href="https://epoch.ai/blog/what-skills-does-swe-bench-verified-evaluate">separate analyses</a> suggest the error rate is likely between 5 and 10%. Furthermore, at Epoch we&#8217;ve run many different models on the benchmark, and if we look at the fraction of problems that have ever been solved by <em>any</em> model, we get 87%. So maybe it&#8217;s just that models still fall short on some problems, or it might even just be random chance. In any case, I don&#8217;t think forecasters as a whole should update much from overshooting.</p><p>This doesn&#8217;t mean that <em>individual</em> forecasters shouldn&#8217;t update &#8212; not all the forecasters have the same beliefs after all! For example, people who expected HLMI to be developed before 2030 tended to have more aggressive benchmark forecasts than those who expected HLMI after 2030. Here&#8217;s an example of this on FrontierMath:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!cOjU!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa3ed1211-c067-4c02-b3a2-2b0da639861c_1024x1280.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!cOjU!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa3ed1211-c067-4c02-b3a2-2b0da639861c_1024x1280.png 424w, https://substackcdn.com/image/fetch/$s_!cOjU!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa3ed1211-c067-4c02-b3a2-2b0da639861c_1024x1280.png 848w, https://substackcdn.com/image/fetch/$s_!cOjU!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa3ed1211-c067-4c02-b3a2-2b0da639861c_1024x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!cOjU!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa3ed1211-c067-4c02-b3a2-2b0da639861c_1024x1280.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!cOjU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa3ed1211-c067-4c02-b3a2-2b0da639861c_1024x1280.png" width="1024" height="1280" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a3ed1211-c067-4c02-b3a2-2b0da639861c_1024x1280.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1280,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!cOjU!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa3ed1211-c067-4c02-b3a2-2b0da639861c_1024x1280.png 424w, https://substackcdn.com/image/fetch/$s_!cOjU!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa3ed1211-c067-4c02-b3a2-2b0da639861c_1024x1280.png 848w, https://substackcdn.com/image/fetch/$s_!cOjU!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa3ed1211-c067-4c02-b3a2-2b0da639861c_1024x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!cOjU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa3ed1211-c067-4c02-b3a2-2b0da639861c_1024x1280.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>You might be thinking, &#8220;of course people with longer timelines expect slower benchmark improvements, isn&#8217;t that common sense?&#8221;. I actually don&#8217;t think so &#8212; you could have longer timelines not because you expect slower benchmark improvements, but because you think benchmarks are less meaningful a proxy of progress towards HLMI. For example, this is a crucial difference between the timelines of Eli Lifland (who <a href="https://ai-2027.com/">co-authored AI 2027</a>) and Mechanize co-founder Ege Erdil, who&#8217;s <a href="https://www.dwarkesh.com/p/ege-tamay">publicly argued</a> <a href="https://epoch.ai/gradient-updates/the-case-for-multi-decade-ai-timelines">for multi-decade AI timelines</a>.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!d8Bq!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F447aac3d-93cf-4b3a-b48b-7f28a264d5f6_1178x350.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!d8Bq!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F447aac3d-93cf-4b3a-b48b-7f28a264d5f6_1178x350.png 424w, https://substackcdn.com/image/fetch/$s_!d8Bq!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F447aac3d-93cf-4b3a-b48b-7f28a264d5f6_1178x350.png 848w, https://substackcdn.com/image/fetch/$s_!d8Bq!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F447aac3d-93cf-4b3a-b48b-7f28a264d5f6_1178x350.png 1272w, https://substackcdn.com/image/fetch/$s_!d8Bq!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F447aac3d-93cf-4b3a-b48b-7f28a264d5f6_1178x350.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!d8Bq!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F447aac3d-93cf-4b3a-b48b-7f28a264d5f6_1178x350.png" width="1178" height="350" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/447aac3d-93cf-4b3a-b48b-7f28a264d5f6_1178x350.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:350,&quot;width&quot;:1178,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!d8Bq!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F447aac3d-93cf-4b3a-b48b-7f28a264d5f6_1178x350.png 424w, https://substackcdn.com/image/fetch/$s_!d8Bq!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F447aac3d-93cf-4b3a-b48b-7f28a264d5f6_1178x350.png 848w, https://substackcdn.com/image/fetch/$s_!d8Bq!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F447aac3d-93cf-4b3a-b48b-7f28a264d5f6_1178x350.png 1272w, https://substackcdn.com/image/fetch/$s_!d8Bq!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F447aac3d-93cf-4b3a-b48b-7f28a264d5f6_1178x350.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>(<a href="https://x.com/eli_lifland/status/1916524442844598749">Source</a>). Note that at the time of writing the tweet, Eli had a median AGI timeline of around 2031, which still counts as having short timelines, even if he doesn&#8217;t fall into the &#8220;HLMI-by-2030&#8221; group of respondents.</em></figcaption></figure></div><p>Indeed, in the case of this survey, those with 2030-or-earlier timelines performed fairly similarly to those with post-2030 timelines (those with longer timelines did a tad better, but it&#8217;s not statistically significant).<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a></p><h2>OpenAI preparedness scores: mixed results on risks</h2><p>The second half of the survey increasingly turns towards broader societal impacts besides AI R&amp;D, which as we just alluded to, may or may not follow from substantial benchmark improvements. The first part of this was to look at potential AI-enabled risks. The results were as follows:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ZRa_!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F614cd693-ad72-447d-87fa-5f32f06a4e7b_2280x1290.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ZRa_!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F614cd693-ad72-447d-87fa-5f32f06a4e7b_2280x1290.png 424w, https://substackcdn.com/image/fetch/$s_!ZRa_!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F614cd693-ad72-447d-87fa-5f32f06a4e7b_2280x1290.png 848w, https://substackcdn.com/image/fetch/$s_!ZRa_!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F614cd693-ad72-447d-87fa-5f32f06a4e7b_2280x1290.png 1272w, https://substackcdn.com/image/fetch/$s_!ZRa_!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F614cd693-ad72-447d-87fa-5f32f06a4e7b_2280x1290.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ZRa_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F614cd693-ad72-447d-87fa-5f32f06a4e7b_2280x1290.png" width="1456" height="824" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/614cd693-ad72-447d-87fa-5f32f06a4e7b_2280x1290.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:824,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:303331,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://epochai.substack.com/i/184793687?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F614cd693-ad72-447d-87fa-5f32f06a4e7b_2280x1290.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ZRa_!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F614cd693-ad72-447d-87fa-5f32f06a4e7b_2280x1290.png 424w, https://substackcdn.com/image/fetch/$s_!ZRa_!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F614cd693-ad72-447d-87fa-5f32f06a4e7b_2280x1290.png 848w, https://substackcdn.com/image/fetch/$s_!ZRa_!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F614cd693-ad72-447d-87fa-5f32f06a4e7b_2280x1290.png 1272w, https://substackcdn.com/image/fetch/$s_!ZRa_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F614cd693-ad72-447d-87fa-5f32f06a4e7b_2280x1290.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Note: You can see the reasoning for the resolutions in the original survey <a href="https://ai2025.org/">link</a>.</em></figcaption></figure></div><p>The survey captured potential risks by looking at OpenAI&#8217;s 2023 <a href="https://cdn.openai.com/openai-preparedness-framework-beta.pdf">preparedness framework</a>, which describes how they &#8220;track, evaluate, forecast, and protect against catastrophic risks posed by increasingly powerful models.&#8221; This is split into four risk areas &#8212; besides Chemical, Biological, Radiological, and Nuclear (CBRN) risks, there&#8217;s also cybersecurity, persuading humans, and model autonomy (which allows models to do things like avoid being shut down by human operators and self-improve). Each risk area then has four risk levels &#8212; low, medium, high, and critical &#8212; and the survey asked how likely each of the four risk areas were to increase its risk level.</p><p>From the table above, it looks like the median forecaster correctly surmised that risk levels would increase in cybersecurity and model autonomy, from &#8220;low&#8221; to &#8220;medium&#8221; risk. This perhaps isn&#8217;t so surprising, because these are heavily related to coding and computer-use, which forecasters expected AIs to get much better at (as we saw in the last section).</p><p>But they underestimated CBRN risks, which went from &#8220;medium&#8221; to &#8220;high&#8221; &#8212; so what went wrong? One small part of this may be that, like with Cybench, respondents overrated OpenAI&#8217;s incentive to underreport increases in these risks. But this can&#8217;t be the full story &#8212; the same argument would apply to all other risk areas as well. The bigger reason seems to be that they specifically underestimated how much LLMs could provide uplift on biological capabilities. Consider the conditions needed to go from a &#8220;medium&#8221; to a &#8220;high&#8221; on CBRN risks:</p><blockquote><p>&#8220;Model enables an expert to develop a novel threat vector OR model provides meaningfully improved assistance that enables anyone with basic training in a relevant field (e.g., introductory undergraduate biology course) to be able to create a CBRN threat.&#8221;</p></blockquote><p>Many forecasters argued it&#8217;d be hard for models to develop &#8220;truly novel threat vectors&#8221;, which is right as far as I&#8217;m aware. But they also didn&#8217;t expect models to be able to provide meaningful &#8220;uplift&#8221; to non-experts in creating CBRN threats, which turned out wrong &#8212; <a href="https://cdn.openai.com/pdf/3a4153c8-c748-4b71-8e31-aecbde944f8d/oai_5_2_system-card.pdf#page=12.33">GPT-5.2</a> seems to have met this bar. Interestingly, this seems in line with a <a href="https://www.governance.ai/research-paper/forecasting-llm-enabled-biorisk-and-the-efficacy-of-safeguards">study</a> earlier this year, which finds superforecasters and biology experts drastically underestimating uplift.</p><p>Putting these results together suggests a mixed record forecasting OpenAI&#8217;s risk levels (the final risk area of &#8220;persuasion&#8221; was removed from OpenAI&#8217;s risk framework in <a href="https://openai.com/index/updating-our-preparedness-framework/">April</a>, so the question was resolved inconclusively).</p><h2>AI&#8217;s prominence: underestimated revenue, and overestimated public perception</h2><p>This brings us to the final part of the survey, which looks at the real-world impacts of AI, namely annualized revenue and public concern about AI.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!qN1Z!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F08cd6b16-d506-4f11-a94d-e1268f0e0f24_2288x836.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!qN1Z!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F08cd6b16-d506-4f11-a94d-e1268f0e0f24_2288x836.png 424w, https://substackcdn.com/image/fetch/$s_!qN1Z!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F08cd6b16-d506-4f11-a94d-e1268f0e0f24_2288x836.png 848w, https://substackcdn.com/image/fetch/$s_!qN1Z!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F08cd6b16-d506-4f11-a94d-e1268f0e0f24_2288x836.png 1272w, https://substackcdn.com/image/fetch/$s_!qN1Z!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F08cd6b16-d506-4f11-a94d-e1268f0e0f24_2288x836.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!qN1Z!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F08cd6b16-d506-4f11-a94d-e1268f0e0f24_2288x836.png" width="1456" height="532" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/08cd6b16-d506-4f11-a94d-e1268f0e0f24_2288x836.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:532,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:214829,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://epochai.substack.com/i/184793687?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F08cd6b16-d506-4f11-a94d-e1268f0e0f24_2288x836.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!qN1Z!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F08cd6b16-d506-4f11-a94d-e1268f0e0f24_2288x836.png 424w, https://substackcdn.com/image/fetch/$s_!qN1Z!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F08cd6b16-d506-4f11-a94d-e1268f0e0f24_2288x836.png 848w, https://substackcdn.com/image/fetch/$s_!qN1Z!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F08cd6b16-d506-4f11-a94d-e1268f0e0f24_2288x836.png 1272w, https://substackcdn.com/image/fetch/$s_!qN1Z!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F08cd6b16-d506-4f11-a94d-e1268f0e0f24_2288x836.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Note: You can see the reasoning for the resolutions in the original survey <a href="https://ai2025.org/">link</a>.</em></figcaption></figure></div><p>Like with the preparedness framework questions, there was a mixed bag: forecasters substantially underestimated revenue (by two-fold), and overestimated public concern about AI.</p><h3>Frontier lab revenues</h3><p>For me, the most striking result of the whole survey was that people heavily underestimated AI lab revenue. The median forecaster expected OpenAI, Anthropic, and xAI to have an annualized revenue of $16 billion by the end of last year, whereas it was almost double that &#8212; $30.4 billion, to be exact. The forecast was only a year out, so how did we end up with such a big difference?</p><p>One issue could be misinterpreting the question. Importantly, the survey asked for <em>annualized </em>revenue rather than <em>total revenue in 2025</em>, so if forecasters misunderstood the question they might&#8217;ve given way too low of an estimate.</p><p>But I doubt this really explains the difference. Eli Lifland <a href="https://x.com/eli_lifland/status/2009397564282687554">provides</a> one reason for this: the survey itself gave numbers for annualized revenue. So anyone reasoning &#8220;revenue will grow by N times&#8221; from those provided figures would arrive at the same forecast regardless of the terminology confusion.</p><p>Another reason is that we look at the forecasters&#8217; reasoning on these questions. Out of the 33 provided rationales, nine of them explicitly referenced annualized revenue, whereas only two of them were clearly about full year revenue. The remaining 22 were pretty vague, like claiming to &#8220;extrapolate previous trends&#8221; without saying what exactly the trend is. So I&#8217;m not totally sure, but I think this weakly suggests that most respondents correctly looked at annualized revenue.</p><p>One final possibility is confusion about the baseline annualized revenue figure provided in the survey, i.e. $4.7 billion. This number relied on OpenAI revenue numbers from August rather than December 2024, so forecasters may have anchored on an already-outdated starting point. Instead, if they had access to a more updated baseline, their forecasts would&#8217;ve been less far off.</p><p>But I don&#8217;t think this is the full story either. Suppose the forecasters had an updated figure, which we can get with the benefit of hindsight: At the end of 2024, OpenAI&#8217;s annualized revenue was around <a href="https://www.reuters.com/business/media-telecom/openais-annualized-revenue-hits-10-billion-up-55-billion-december-2024-2025-06-09/">$5.5 billion</a>, and Anthropic&#8217;s was around <a href="https://www.theinformation.com/articles/anthropic-projects-soaring-growth-to-34-5-billion-in-2027-revenue?rc=spkbjw">$0.9 billion</a>. This gives us a total of $6.4 billion in annualized revenue. So what would their forecasts have been? From reading the forecasters&#8217; rationales, many respondents seemed to be extrapolating &#8220;2-3&#215; per year&#8221; growth rates, which on the high end of things would&#8217;ve led to a forecast of $19.2 billion. That&#8217;s a lot closer to the $30.4 billion resolution, but still would&#8217;ve fallen short.</p><p>If I&#8217;m right, then the reason these forecasters were off was more about how OpenAI and Anthropic revenues grew like crazy. And I really mean <em>crazy</em> &#8212; the annualized revenue increased from around $6.4 billion to $30.4 billion, which is around 4.8&#215; in a year!</p><h3>Public attention on AI</h3><p>The last question of the survey was about public concern about AI, specifically: &#8220;What percentage of Americans will identify computers/technology advancement as the US&#8217;s most important problem by December 31st 2025?&#8221;.</p><p>The median forecaster ended up overestimating somewhat. Depending on what time period you average over, the answer was around 0.625%, compared to around 0.4% or 0.45% by the end of last year. In contrast, the median forecast was around 2%. So there&#8217;s some degree of overestimating public concern going on here.</p><p>This overestimation seems more egregious for the handful of people who gave substantially higher forecasts, say north of 5%. Some of these respondents thought that public attention would really spike because of job market impacts and increasing concerns about AI safety, which hasn&#8217;t really panned out.</p><p>Maybe it&#8217;s a mistake for me to read too much into this &#8212; I suspect people didn&#8217;t put too much thought into the exact numbers, and instead just chose fairly round numbers like 1% or 2%, thinking that public attention would increase but remain small. But overall I&#8217;d say this looks a bit like what economists frequently accuse me of &#8212; being a &#8220;technologist&#8221; who overestimates the importance of their technology.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://epochai.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://epochai.substack.com/subscribe?"><span>Subscribe now</span></a></p><h2>Takeaways from the survey</h2><p>If I had to summarize the results in a single sentence, I&#8217;d say that the forecasters were mostly right about benchmark scores, but had mixed results with the risks in OpenAI&#8217;s preparedness framework and on real-world impacts. Most notably, they underestimated revenue and CBRN risks, and overestimated public concern about AI. We can see this in the following image:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ooY5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02e4ee45-0f87-4f52-a6ea-ca2b06dacee2_836x755.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ooY5!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02e4ee45-0f87-4f52-a6ea-ca2b06dacee2_836x755.png 424w, https://substackcdn.com/image/fetch/$s_!ooY5!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02e4ee45-0f87-4f52-a6ea-ca2b06dacee2_836x755.png 848w, https://substackcdn.com/image/fetch/$s_!ooY5!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02e4ee45-0f87-4f52-a6ea-ca2b06dacee2_836x755.png 1272w, https://substackcdn.com/image/fetch/$s_!ooY5!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02e4ee45-0f87-4f52-a6ea-ca2b06dacee2_836x755.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ooY5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02e4ee45-0f87-4f52-a6ea-ca2b06dacee2_836x755.png" width="836" height="755" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/02e4ee45-0f87-4f52-a6ea-ca2b06dacee2_836x755.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:755,&quot;width&quot;:836,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ooY5!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02e4ee45-0f87-4f52-a6ea-ca2b06dacee2_836x755.png 424w, https://substackcdn.com/image/fetch/$s_!ooY5!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02e4ee45-0f87-4f52-a6ea-ca2b06dacee2_836x755.png 848w, https://substackcdn.com/image/fetch/$s_!ooY5!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02e4ee45-0f87-4f52-a6ea-ca2b06dacee2_836x755.png 1272w, https://substackcdn.com/image/fetch/$s_!ooY5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02e4ee45-0f87-4f52-a6ea-ca2b06dacee2_836x755.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>For the most updated version of this, see the &#8220;Summary&#8221; tab of the <a href="https://ai2025.org/">original forecasting survey</a>.</em></figcaption></figure></div><p>To me, these results really bring an important question back to center stage: &#8220;how much do benchmarks reflect broader societal impacts?&#8221;. Survey respondents did pretty well forecasting improvements on AI R&amp;D benchmarks, but we should always be clear about what these benchmarks really tell us. For example, while RE-Bench is designed to track AI R&amp;D, the tasks <a href="https://metr.org/AI_R_D_Evaluation_Report.pdf">tend</a> to be very &#8220;clean&#8221;, with nice feedback loops and few tasks occurring in parallel, which is quite unlike real-world AI research with millions of lines of code. It&#8217;s a useful benchmark, but doing well on it doesn&#8217;t imply, for example, that AI R&amp;D is automatable.</p><p>It&#8217;s also a good time for researchers and forecasters to figure out how we can improve our forecasts of real-world impacts. One relatively obvious answer is to gather more data about these impacts &#8212; the forecasters in this survey probably did worse on real-world impacts in part because they had less high-quality data to extrapolate. So if we want to improve our future forecasts, we should probably gather data on a range of real-world indicators, especially as AI&#8217;s impacts continue expanding beyond the Silicon Valley sphere.</p><p>So what does this all mean for AI progress and its societal impacts? Dwarkesh <a href="https://open.substack.com/pub/dwarkesh/p/thoughts-on-ai-progress-dec-2025?r=5fx8fq&amp;selection=dd356fbe-43e5-475c-83ad-d99732910ca5&amp;utm_campaign=post-share-selection&amp;utm_medium=web&amp;aspectRatio=instagram&amp;bgColor=%23f3c016&amp;textColor=%23ffffff">writes</a>, &#8220;Models keep getting more impressive at the rate the short timelines people predict, but more useful at the rate the long timelines people predict.&#8221; I think the results of this survey largely support this, though maybe the point about revenue is a counterpoint &#8212; the annualized revenue numbers seem to suggest they&#8217;re getting more useful faster than both the short and long timelines people predict!</p><p>Another way to look at it is this: the survey respondents were probably more bullish on rapid near-term AI progress than almost anyone else in the world, and yet they substantially underestimated AI revenue! So perhaps, despite all the hype, AI is an even bigger deal than we thought.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://epochai.substack.com/p/how-well-did-forecasters-predict?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://epochai.substack.com/p/how-well-did-forecasters-predict?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p><p><em>I&#8217;d like to thank Eli Lifland, Adam Binksmith, JS Denain, Lynette Bye, Josh You, Tao Lin, and Greg Burnham for their feedback and support on this post.</em></p><h2>Appendix</h2><p>In the main post I focused on the main high-level takeaways from the forecasting survey, for a general audience. In contrast, this appendix is for those of you who are also interested in the nitty-gritty details, like on the distribution of responses and how we should interpret the data.</p><h3>Benchmarks relating to AI R&amp;D</h3><p>The survey focused on the following five benchmarks:</p><ol><li><p><strong>RE-Bench</strong>: Seven open-ended machine learning problems like &#8220;fine-tune GPT-2&#8221;.</p></li><li><p><strong>SWE-Bench Verified:</strong> Resolving small real-world GitHub issues in Python.</p></li><li><p><strong>Cybench: </strong>Professional cybersecurity competition tasks like exploiting vulnerabilities in web apps.</p></li><li><p><strong>OSWorld: </strong>Simple computer-use tasks like duplicating powerpoint slides.</p></li><li><p><strong>FrontierMath: </strong>Extremely challenging math problems with automatically verifiable answers.</p></li></ol><p>The first three of these seem pretty clearly related to AI R&amp;D, involving tasks with lots of coding. OSWorld and FrontierMath are more debatable, but they do capture skills that are probably quite important for this, namely computer-use and deep technical reasoning. Performance on FrontierMath and OSWorld are overall also quite correlated with the first three benchmarks, perhaps because they capture the same <a href="https://epoch.ai/benchmarks/eci">latent</a> &#8220;<a href="https://epoch.ai/blog/a-rosetta-stone-for-ai-benchmarks">capability</a>&#8221; factor across different models.</p><h4>RE-Bench</h4><p>Research Engineering Bench (<a href="https://metr.org/blog/2024-11-22-evaluating-r-d-capabilities-of-llms/">RE-Bench</a>) is a benchmark created by METR to measure automation of AI R&amp;D. It consists of seven challenging open-ended machine learning problems, things like &#8220;optimizing a GPU kernel&#8221; and &#8220;finetuning GPT-2&#8221;. Each task has some starting code which achieves a certain baseline performance, and the goal is to improve performance as far as possible.</p><p>Trouble is, the most natural measures of performance look quite different across each of these problems. To make them more comparable in an apples-to-apples way, the METR researchers had to normalize these different metrics, where 0 is the starting performance of the reference code, and 1 corresponds to a human baseline set by human experts. Then they average across the tasks.</p><p>At the end of 2024, the highest AI score was 0.61, held by Claude 3.5 Sonnet. The median forecast was 1.1, which is very close to the final resolution of 1.13. That said, this final resolution is based on scores from Gemini 3 Pro &#8212; it&#8217;s plausible that Claude Opus 4.5 or GPT-5.2 would&#8217;ve done better had they been evaluated, making the forecast an underestimate. Note that this evaluation with Gemini only considers 5 of the 7 RE-Bench tasks.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Uxru!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f6dcb41-4873-4d38-8e5a-06d177c4c7a9_940x460.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Uxru!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f6dcb41-4873-4d38-8e5a-06d177c4c7a9_940x460.png 424w, https://substackcdn.com/image/fetch/$s_!Uxru!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f6dcb41-4873-4d38-8e5a-06d177c4c7a9_940x460.png 848w, https://substackcdn.com/image/fetch/$s_!Uxru!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f6dcb41-4873-4d38-8e5a-06d177c4c7a9_940x460.png 1272w, https://substackcdn.com/image/fetch/$s_!Uxru!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f6dcb41-4873-4d38-8e5a-06d177c4c7a9_940x460.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Uxru!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f6dcb41-4873-4d38-8e5a-06d177c4c7a9_940x460.png" width="940" height="460" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1f6dcb41-4873-4d38-8e5a-06d177c4c7a9_940x460.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:460,&quot;width&quot;:940,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Uxru!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f6dcb41-4873-4d38-8e5a-06d177c4c7a9_940x460.png 424w, https://substackcdn.com/image/fetch/$s_!Uxru!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f6dcb41-4873-4d38-8e5a-06d177c4c7a9_940x460.png 848w, https://substackcdn.com/image/fetch/$s_!Uxru!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f6dcb41-4873-4d38-8e5a-06d177c4c7a9_940x460.png 1272w, https://substackcdn.com/image/fetch/$s_!Uxru!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f6dcb41-4873-4d38-8e5a-06d177c4c7a9_940x460.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Looking at the available rationales reveals a couple of interesting takeaways. One observation is that some estimates seem to exceed what is theoretically possible. We can estimate the theoretical maximum score based on the &#8220;more info&#8221; section of the <a href="https://ai2025.org/">original survey</a>. Each task has its own maximum, namely the following: 1.36, 2.53, 1.25, 1.81, 2.51, 1.8, and 1.48. If we then take the average of this we get 1.82, but some forecasted higher than this. This may be because they disagree with the maximum bound, but from reading the rationales, I suspect it&#8217;s because they weren&#8217;t considering this bound at all.</p><p>We can also look at the factors influencing how aggressive and conservative the forecasts were. Those with more aggressive forecasts (e.g. with a score of 1.2 or higher) tended to think that RE-Bench-esque tasks would be heavily optimized for, are amenable to reasoning model improvements, and have fast feedback loops. Those who forecasted below 0.9 seemed to not provide very substantive rationales (e.g. saying things like &#8220;this was a naive guess&#8221;). And those with forecasts in between these two ranges tended to anchor a bit more to previous rates of improvement &#8212; that said, it&#8217;s not always clear what data they were using to determine these rates, because different extrapolations seemed to yield different forecasts.</p><h4>SWE-Bench Verified</h4><p>The <a href="https://openai.com/index/introducing-swe-bench-verified/">SWE-bench Verified</a> benchmark consists of 500 software engineering problems drawn from real GitHub issues. These are Python-specific and verified by human annotators to remove errors.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!GWXp!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a58c539-d851-4738-a600-7a4c23f66365_940x460.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!GWXp!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a58c539-d851-4738-a600-7a4c23f66365_940x460.png 424w, https://substackcdn.com/image/fetch/$s_!GWXp!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a58c539-d851-4738-a600-7a4c23f66365_940x460.png 848w, https://substackcdn.com/image/fetch/$s_!GWXp!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a58c539-d851-4738-a600-7a4c23f66365_940x460.png 1272w, https://substackcdn.com/image/fetch/$s_!GWXp!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a58c539-d851-4738-a600-7a4c23f66365_940x460.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!GWXp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a58c539-d851-4738-a600-7a4c23f66365_940x460.png" width="940" height="460" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3a58c539-d851-4738-a600-7a4c23f66365_940x460.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:460,&quot;width&quot;:940,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!GWXp!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a58c539-d851-4738-a600-7a4c23f66365_940x460.png 424w, https://substackcdn.com/image/fetch/$s_!GWXp!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a58c539-d851-4738-a600-7a4c23f66365_940x460.png 848w, https://substackcdn.com/image/fetch/$s_!GWXp!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a58c539-d851-4738-a600-7a4c23f66365_940x460.png 1272w, https://substackcdn.com/image/fetch/$s_!GWXp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a58c539-d851-4738-a600-7a4c23f66365_940x460.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Since this was discussed in depth in the main post, I won&#8217;t elaborate much on this here. However I&#8217;ll add that some of the rationales for slower forecasts seemed to be reasoning from a lower baseline than o3&#8217;s 71% (instead they were a 55% baseline, from the Amazon Q Developer Agent). This is likely because the reference numbers provided in the survey were <a href="https://x.com/aidigest_/status/1871471620247720289">updated</a> several weeks after the original release. This would tend to lower the observed forecasts, but it seems this wasn&#8217;t enough to prevent the overall forecast from overshooting the observed final performance of around 80%.</p><h4>Cybench</h4><p><a href="https://cybench.github.io/">Cybench</a> is a benchmark with cybersecurity tasks. These were taken from professional-level &#8220;Capture the Flag&#8221; competitions, where competitors attempt to identify specific vulnerabilities in computer programs.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Ubc5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbcbf19d6-f657-4637-8ecf-8e7406af5579_940x460.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Ubc5!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbcbf19d6-f657-4637-8ecf-8e7406af5579_940x460.png 424w, https://substackcdn.com/image/fetch/$s_!Ubc5!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbcbf19d6-f657-4637-8ecf-8e7406af5579_940x460.png 848w, https://substackcdn.com/image/fetch/$s_!Ubc5!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbcbf19d6-f657-4637-8ecf-8e7406af5579_940x460.png 1272w, https://substackcdn.com/image/fetch/$s_!Ubc5!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbcbf19d6-f657-4637-8ecf-8e7406af5579_940x460.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Ubc5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbcbf19d6-f657-4637-8ecf-8e7406af5579_940x460.png" width="940" height="460" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bcbf19d6-f657-4637-8ecf-8e7406af5579_940x460.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:460,&quot;width&quot;:940,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Ubc5!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbcbf19d6-f657-4637-8ecf-8e7406af5579_940x460.png 424w, https://substackcdn.com/image/fetch/$s_!Ubc5!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbcbf19d6-f657-4637-8ecf-8e7406af5579_940x460.png 848w, https://substackcdn.com/image/fetch/$s_!Ubc5!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbcbf19d6-f657-4637-8ecf-8e7406af5579_940x460.png 1272w, https://substackcdn.com/image/fetch/$s_!Ubc5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbcbf19d6-f657-4637-8ecf-8e7406af5579_940x460.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The scores are determined using the &#8220;unguided&#8221; evaluation setting. This is where the AI model is expected to perform the whole task end-to-end; without guidance on which subtasks to complete, which serve as intermediate checkpoints.</p><h4>OSWorld</h4><p>OSWorld is a notable benchmark consisting of 361 computer-use tasks. These tend to focus on simple realistic tasks in Linux-based environments, like adding page numbers to a document.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Deh-!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2e15a19-0cfc-4f28-b4dc-cc975c0b3188_940x460.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Deh-!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2e15a19-0cfc-4f28-b4dc-cc975c0b3188_940x460.png 424w, https://substackcdn.com/image/fetch/$s_!Deh-!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2e15a19-0cfc-4f28-b4dc-cc975c0b3188_940x460.png 848w, https://substackcdn.com/image/fetch/$s_!Deh-!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2e15a19-0cfc-4f28-b4dc-cc975c0b3188_940x460.png 1272w, https://substackcdn.com/image/fetch/$s_!Deh-!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2e15a19-0cfc-4f28-b4dc-cc975c0b3188_940x460.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Deh-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2e15a19-0cfc-4f28-b4dc-cc975c0b3188_940x460.png" width="940" height="460" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b2e15a19-0cfc-4f28-b4dc-cc975c0b3188_940x460.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:460,&quot;width&quot;:940,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Deh-!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2e15a19-0cfc-4f28-b4dc-cc975c0b3188_940x460.png 424w, https://substackcdn.com/image/fetch/$s_!Deh-!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2e15a19-0cfc-4f28-b4dc-cc975c0b3188_940x460.png 848w, https://substackcdn.com/image/fetch/$s_!Deh-!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2e15a19-0cfc-4f28-b4dc-cc975c0b3188_940x460.png 1272w, https://substackcdn.com/image/fetch/$s_!Deh-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2e15a19-0cfc-4f28-b4dc-cc975c0b3188_940x460.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>At the time of forecasting, the state-of-the-art performance was 24.5%, and the median forecast was 60%. This is fairly close to the actual performance of 66.3%, though it looks as though forecasters were quite uncertain about what performance would be achieved.</p><p>I think a lot of this boils down to different intuitions on how hard OSWorld tasks are. Those with aggressive forecasts tended to think that the benchmark tasks were like &#8220;engineering problems&#8221;, where bottlenecks to higher performance would probably be resolved just by spending more researcher time and effort. Those with conservative forecasts tended to think that OSWorld required multimodal skills, which language models tend to struggle with. I think the latter argument is overstated &#8212; much of OSWorld can be done <a href="https://epoch.ai/blog/what-does-osworld-tell-us-about-ais-ability-to-use-computers">without a graphical user interface</a>, though this likely wasn&#8217;t widely known when the forecasts were being made.</p><p>One final issue is that these performance numbers are harder to interpret than they might initially seem. Greg Burnham succinctly <a href="https://epochai.substack.com/p/osworld-ai-computer-use-capabilities">describes</a> one such issue: &#8220;Contrary to standard practice, OSWorld is updated continuously. A major update in July changed most tasks, and 10% of task instructions have been updated since then. Furthermore, 10% of tasks rely on live data from the web. This makes it hard to compare results over time.&#8221;</p><h4>FrontierMath</h4><p>FrontierMath is a benchmark with several hundred extremely challenging math problems. Each one contains automatically verifiable answers, like numbers or matrices or SymPy objects.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ZPjL!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F465a8439-9b06-4f01-9ec4-b6262cf344ef_940x460.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ZPjL!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F465a8439-9b06-4f01-9ec4-b6262cf344ef_940x460.png 424w, https://substackcdn.com/image/fetch/$s_!ZPjL!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F465a8439-9b06-4f01-9ec4-b6262cf344ef_940x460.png 848w, https://substackcdn.com/image/fetch/$s_!ZPjL!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F465a8439-9b06-4f01-9ec4-b6262cf344ef_940x460.png 1272w, https://substackcdn.com/image/fetch/$s_!ZPjL!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F465a8439-9b06-4f01-9ec4-b6262cf344ef_940x460.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ZPjL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F465a8439-9b06-4f01-9ec4-b6262cf344ef_940x460.png" width="940" height="460" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/465a8439-9b06-4f01-9ec4-b6262cf344ef_940x460.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:460,&quot;width&quot;:940,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ZPjL!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F465a8439-9b06-4f01-9ec4-b6262cf344ef_940x460.png 424w, https://substackcdn.com/image/fetch/$s_!ZPjL!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F465a8439-9b06-4f01-9ec4-b6262cf344ef_940x460.png 848w, https://substackcdn.com/image/fetch/$s_!ZPjL!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F465a8439-9b06-4f01-9ec4-b6262cf344ef_940x460.png 1272w, https://substackcdn.com/image/fetch/$s_!ZPjL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F465a8439-9b06-4f01-9ec4-b6262cf344ef_940x460.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Here it seems like the median forecast was very good &#8212; 40%, compared to the true answer of 40.7%. More interesting to me is why this number wasn&#8217;t substantially higher.</p><p>For instance, in May I <a href="https://epoch.ai/gradient-updates/is-ai-already-superhuman-on-frontiermath">wrote</a> about how I expected FrontierMath performance to clearly exceed 50% by the end of the year. In fact, I think a forecast like 75% wasn&#8217;t unreasonable &#8212; the benchmark problems are roughly such that 25% are at &#8220;undergrad level&#8221;, 50% are at &#8220;graduate level&#8221;, and the remaining 25% is &#8220;research level&#8221;. So if we expect the models to be able to do most of the &#8220;graduate level&#8221; problems, we might expect a score close to 75%. And there are other reasons too, like running simulations of benchmark progress given the limited data available at the time of forecasting. The only problem is that this turned out to be wrong.</p><p>In retrospect, I think I may have made two mistakes. The first is that these difficulty tiers don&#8217;t very cleanly delineate what problems AI systems are or aren&#8217;t able to successfully solve. This is especially because the benchmark problems can be &#8220;<a href="https://epoch.ai/gradient-updates/beyond-benchmark-scores-analysing-o3-mini-math-reasoning">cheesed</a>&#8221;, where the answer is guessed using an informal argument.</p><p>My second mistake was to update too much on o3&#8217;s release, which suddenly brought state-of-the-art performance from 2% to around 25% within a month of FrontierMath&#8217;s release. In fact, (slight) overupdating seems to have been a common issue: the median prior to o3&#8217;s release was around 30% below the resolution value, and the median after o3 was around 10% too high.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Cx21!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5186d0f3-bfb4-40a1-a5cf-b68c5cac163d_752x561.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Cx21!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5186d0f3-bfb4-40a1-a5cf-b68c5cac163d_752x561.png 424w, https://substackcdn.com/image/fetch/$s_!Cx21!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5186d0f3-bfb4-40a1-a5cf-b68c5cac163d_752x561.png 848w, https://substackcdn.com/image/fetch/$s_!Cx21!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5186d0f3-bfb4-40a1-a5cf-b68c5cac163d_752x561.png 1272w, https://substackcdn.com/image/fetch/$s_!Cx21!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5186d0f3-bfb4-40a1-a5cf-b68c5cac163d_752x561.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Cx21!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5186d0f3-bfb4-40a1-a5cf-b68c5cac163d_752x561.png" width="752" height="561" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5186d0f3-bfb4-40a1-a5cf-b68c5cac163d_752x561.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:561,&quot;width&quot;:752,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Cx21!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5186d0f3-bfb4-40a1-a5cf-b68c5cac163d_752x561.png 424w, https://substackcdn.com/image/fetch/$s_!Cx21!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5186d0f3-bfb4-40a1-a5cf-b68c5cac163d_752x561.png 848w, https://substackcdn.com/image/fetch/$s_!Cx21!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5186d0f3-bfb4-40a1-a5cf-b68c5cac163d_752x561.png 1272w, https://substackcdn.com/image/fetch/$s_!Cx21!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5186d0f3-bfb4-40a1-a5cf-b68c5cac163d_752x561.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Source: <em><a href="https://theaidigest.org/ai2025-analysis-may#5-what-will-be-the-best-performance-on-frontiermath-by-december-31st-2025">AI 2025 Forecasts - May Update</a></em></figcaption></figure></div><p>One more small thing that makes this tricky is that the ~25% figure provided at the start of forecasts is based on OpenAI&#8217;s internal evaluations, whereas the final resolution was based on Epoch AI&#8217;s evaluations. The latter is often somewhat lower &#8212; for example, in the case of o3 the highest score from Epoch&#8217;s numbers is closer to 18.7%.</p><h2>OpenAI&#8217;s preparedness scores, frontier lab revenue, public attention on AI</h2><p>Since these were discussed quite thoroughly in the main article, I won&#8217;t add much more detail beyond sharing the main graphs.</p><h3>OpenAI preparedness scores</h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!xZ45!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd62a3c4-b975-4b70-a306-852ad79a2237_940x460.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!xZ45!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd62a3c4-b975-4b70-a306-852ad79a2237_940x460.png 424w, https://substackcdn.com/image/fetch/$s_!xZ45!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd62a3c4-b975-4b70-a306-852ad79a2237_940x460.png 848w, https://substackcdn.com/image/fetch/$s_!xZ45!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd62a3c4-b975-4b70-a306-852ad79a2237_940x460.png 1272w, https://substackcdn.com/image/fetch/$s_!xZ45!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd62a3c4-b975-4b70-a306-852ad79a2237_940x460.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!xZ45!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd62a3c4-b975-4b70-a306-852ad79a2237_940x460.png" width="940" height="460" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/fd62a3c4-b975-4b70-a306-852ad79a2237_940x460.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:460,&quot;width&quot;:940,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!xZ45!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd62a3c4-b975-4b70-a306-852ad79a2237_940x460.png 424w, https://substackcdn.com/image/fetch/$s_!xZ45!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd62a3c4-b975-4b70-a306-852ad79a2237_940x460.png 848w, https://substackcdn.com/image/fetch/$s_!xZ45!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd62a3c4-b975-4b70-a306-852ad79a2237_940x460.png 1272w, https://substackcdn.com/image/fetch/$s_!xZ45!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffd62a3c4-b975-4b70-a306-852ad79a2237_940x460.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!1Vo9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F74f4945c-7fa2-4b49-95d8-2201b08f514d_940x460.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!1Vo9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F74f4945c-7fa2-4b49-95d8-2201b08f514d_940x460.png 424w, https://substackcdn.com/image/fetch/$s_!1Vo9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F74f4945c-7fa2-4b49-95d8-2201b08f514d_940x460.png 848w, https://substackcdn.com/image/fetch/$s_!1Vo9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F74f4945c-7fa2-4b49-95d8-2201b08f514d_940x460.png 1272w, https://substackcdn.com/image/fetch/$s_!1Vo9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F74f4945c-7fa2-4b49-95d8-2201b08f514d_940x460.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!1Vo9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F74f4945c-7fa2-4b49-95d8-2201b08f514d_940x460.png" width="940" height="460" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/74f4945c-7fa2-4b49-95d8-2201b08f514d_940x460.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:460,&quot;width&quot;:940,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!1Vo9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F74f4945c-7fa2-4b49-95d8-2201b08f514d_940x460.png 424w, https://substackcdn.com/image/fetch/$s_!1Vo9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F74f4945c-7fa2-4b49-95d8-2201b08f514d_940x460.png 848w, https://substackcdn.com/image/fetch/$s_!1Vo9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F74f4945c-7fa2-4b49-95d8-2201b08f514d_940x460.png 1272w, https://substackcdn.com/image/fetch/$s_!1Vo9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F74f4945c-7fa2-4b49-95d8-2201b08f514d_940x460.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!CEoN!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fafd86659-ef0f-4609-8015-69173f1038f5_940x460.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!CEoN!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fafd86659-ef0f-4609-8015-69173f1038f5_940x460.png 424w, https://substackcdn.com/image/fetch/$s_!CEoN!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fafd86659-ef0f-4609-8015-69173f1038f5_940x460.png 848w, https://substackcdn.com/image/fetch/$s_!CEoN!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fafd86659-ef0f-4609-8015-69173f1038f5_940x460.png 1272w, https://substackcdn.com/image/fetch/$s_!CEoN!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fafd86659-ef0f-4609-8015-69173f1038f5_940x460.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!CEoN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fafd86659-ef0f-4609-8015-69173f1038f5_940x460.png" width="940" height="460" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/afd86659-ef0f-4609-8015-69173f1038f5_940x460.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:460,&quot;width&quot;:940,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!CEoN!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fafd86659-ef0f-4609-8015-69173f1038f5_940x460.png 424w, https://substackcdn.com/image/fetch/$s_!CEoN!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fafd86659-ef0f-4609-8015-69173f1038f5_940x460.png 848w, https://substackcdn.com/image/fetch/$s_!CEoN!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fafd86659-ef0f-4609-8015-69173f1038f5_940x460.png 1272w, https://substackcdn.com/image/fetch/$s_!CEoN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fafd86659-ef0f-4609-8015-69173f1038f5_940x460.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!zTpy!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F211fea19-7966-4f41-a474-a0a501257754_940x460.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!zTpy!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F211fea19-7966-4f41-a474-a0a501257754_940x460.png 424w, https://substackcdn.com/image/fetch/$s_!zTpy!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F211fea19-7966-4f41-a474-a0a501257754_940x460.png 848w, https://substackcdn.com/image/fetch/$s_!zTpy!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F211fea19-7966-4f41-a474-a0a501257754_940x460.png 1272w, https://substackcdn.com/image/fetch/$s_!zTpy!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F211fea19-7966-4f41-a474-a0a501257754_940x460.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!zTpy!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F211fea19-7966-4f41-a474-a0a501257754_940x460.png" width="940" height="460" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/211fea19-7966-4f41-a474-a0a501257754_940x460.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:460,&quot;width&quot;:940,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!zTpy!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F211fea19-7966-4f41-a474-a0a501257754_940x460.png 424w, https://substackcdn.com/image/fetch/$s_!zTpy!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F211fea19-7966-4f41-a474-a0a501257754_940x460.png 848w, https://substackcdn.com/image/fetch/$s_!zTpy!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F211fea19-7966-4f41-a474-a0a501257754_940x460.png 1272w, https://substackcdn.com/image/fetch/$s_!zTpy!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F211fea19-7966-4f41-a474-a0a501257754_940x460.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h4>Frontier lab revenue</h4><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!t-l_!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7241a1cc-01a5-4164-8189-003e320ca47d_940x460.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!t-l_!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7241a1cc-01a5-4164-8189-003e320ca47d_940x460.png 424w, https://substackcdn.com/image/fetch/$s_!t-l_!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7241a1cc-01a5-4164-8189-003e320ca47d_940x460.png 848w, https://substackcdn.com/image/fetch/$s_!t-l_!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7241a1cc-01a5-4164-8189-003e320ca47d_940x460.png 1272w, https://substackcdn.com/image/fetch/$s_!t-l_!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7241a1cc-01a5-4164-8189-003e320ca47d_940x460.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!t-l_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7241a1cc-01a5-4164-8189-003e320ca47d_940x460.png" width="940" height="460" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7241a1cc-01a5-4164-8189-003e320ca47d_940x460.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:460,&quot;width&quot;:940,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!t-l_!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7241a1cc-01a5-4164-8189-003e320ca47d_940x460.png 424w, https://substackcdn.com/image/fetch/$s_!t-l_!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7241a1cc-01a5-4164-8189-003e320ca47d_940x460.png 848w, https://substackcdn.com/image/fetch/$s_!t-l_!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7241a1cc-01a5-4164-8189-003e320ca47d_940x460.png 1272w, https://substackcdn.com/image/fetch/$s_!t-l_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7241a1cc-01a5-4164-8189-003e320ca47d_940x460.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h4>Public attention on AI</h4><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!NnIA!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F868e10f6-adbf-4d02-b836-154926269ca7_940x460.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!NnIA!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F868e10f6-adbf-4d02-b836-154926269ca7_940x460.png 424w, https://substackcdn.com/image/fetch/$s_!NnIA!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F868e10f6-adbf-4d02-b836-154926269ca7_940x460.png 848w, https://substackcdn.com/image/fetch/$s_!NnIA!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F868e10f6-adbf-4d02-b836-154926269ca7_940x460.png 1272w, https://substackcdn.com/image/fetch/$s_!NnIA!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F868e10f6-adbf-4d02-b836-154926269ca7_940x460.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!NnIA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F868e10f6-adbf-4d02-b836-154926269ca7_940x460.png" width="940" height="460" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/868e10f6-adbf-4d02-b836-154926269ca7_940x460.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:460,&quot;width&quot;:940,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!NnIA!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F868e10f6-adbf-4d02-b836-154926269ca7_940x460.png 424w, https://substackcdn.com/image/fetch/$s_!NnIA!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F868e10f6-adbf-4d02-b836-154926269ca7_940x460.png 848w, https://substackcdn.com/image/fetch/$s_!NnIA!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F868e10f6-adbf-4d02-b836-154926269ca7_940x460.png 1272w, https://substackcdn.com/image/fetch/$s_!NnIA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F868e10f6-adbf-4d02-b836-154926269ca7_940x460.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>Data on confidence intervals</h2><p>In filling out the survey, participants were optionally allowed to add 80% credible intervals for their forecasts. We present some of the data on this in the table below:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!h2-l!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F47e0c65c-ec33-4663-8286-9bcf514a6870_2288x1398.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!h2-l!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F47e0c65c-ec33-4663-8286-9bcf514a6870_2288x1398.png 424w, https://substackcdn.com/image/fetch/$s_!h2-l!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F47e0c65c-ec33-4663-8286-9bcf514a6870_2288x1398.png 848w, https://substackcdn.com/image/fetch/$s_!h2-l!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F47e0c65c-ec33-4663-8286-9bcf514a6870_2288x1398.png 1272w, https://substackcdn.com/image/fetch/$s_!h2-l!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F47e0c65c-ec33-4663-8286-9bcf514a6870_2288x1398.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!h2-l!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F47e0c65c-ec33-4663-8286-9bcf514a6870_2288x1398.png" width="1456" height="890" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/47e0c65c-ec33-4663-8286-9bcf514a6870_2288x1398.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:890,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:334028,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://epochai.substack.com/i/184793687?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F47e0c65c-ec33-4663-8286-9bcf514a6870_2288x1398.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!h2-l!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F47e0c65c-ec33-4663-8286-9bcf514a6870_2288x1398.png 424w, https://substackcdn.com/image/fetch/$s_!h2-l!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F47e0c65c-ec33-4663-8286-9bcf514a6870_2288x1398.png 848w, https://substackcdn.com/image/fetch/$s_!h2-l!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F47e0c65c-ec33-4663-8286-9bcf514a6870_2288x1398.png 1272w, https://substackcdn.com/image/fetch/$s_!h2-l!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F47e0c65c-ec33-4663-8286-9bcf514a6870_2288x1398.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://epochai.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading! Subscribe to receive the latest from Epoch AI.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>Users with 2030-or-earlier timelines had a mean score of -0.014 on the AI R&amp;D-related benchmarks, compared to +0.059 for those with post-2030 timelines. The medians for the same groups were +0.023 and +0.088 respectively. So the post-2030 group did slightly better, but this difference isn&#8217;t statistically significant based on a t-test (p = 0.39). You can also just see that these differences are small by looking at the <a href="https://ai2025.org/">leaderboard</a>, where scores go from around -6 to 2.</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[An FAQ on Reinforcement Learning Environments]]></title><description><![CDATA[We interviewed 18 people across RL environment startups, neolabs, and frontier labs about the state of the field and where it's headed.]]></description><link>https://epochai.substack.com/p/an-faq-on-reinforcement-learning</link><guid isPermaLink="false">https://epochai.substack.com/p/an-faq-on-reinforcement-learning</guid><dc:creator><![CDATA[JSD]]></dc:creator><pubDate>Mon, 12 Jan 2026 20:27:19 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!bv6y!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21b4870f-4a72-47e3-be91-36ee75329c81_1280x720.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>This post is a collaboration between guest author <a href="https://x.com/chrisbarber">Chris Barber</a> and <a href="https://jsdena.in/">JS Denain</a> from Epoch AI. This post is part of our <a href="https://epoch.ai/gradient-updates">Gradient Updates</a> newsletter, which shares more opinionated or informal takes about big questions in AI progress. These posts solely represent the views of the authors, and do not necessarily reflect the views of Epoch AI as a whole.</em></p><p><em>Originally posted on <a href="https://epoch.ai/gradient-updates/state-of-rl-envs">Epoch AI</a>.</em></p><div><hr></div><p>Reinforcement learning (RL) environments have become central to how frontier AI labs train their models. In September 2025, The Information <a href="https://www.theinformation.com/articles/anthropic-openai-developing-ai-co-workers?rc=dp0mql">reported</a> that Anthropic had discussed spending over $1 billion on RL environments over the following year. As Andrej Karpathy put it in his <a href="https://karpathy.bearblog.dev/year-in-review-2025/">2025 year-in-review</a>: by training LLMs on a wide range of verifiable tasks across different environments, &#8220;the LLMs spontaneously develop strategies that look like &#8216;reasoning&#8217; to humans.&#8221;</p><p>This wave of RL for capabilities started with OpenAI&#8217;s o1, which was trained on math and coding problems with verifiable answers. Since then, labs have expanded the range of tasks they train on, all the while scaling the amount of compute spent on RL training.</p><p>Without diverse, high-quality environments and tasks to train on, throwing more compute at RL <a href="https://www.mechanize.work/blog/cheap-rl-tasks-will-waste-compute/">risks wasting much of it</a>. As a result, creating those tasks and environments has become a key bottleneck for scaling capabilities, and a growing market that remains largely behind closed doors.</p><p>To understand the emerging industry of building environments and tasks that labs use to RL-train their models, we interviewed<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> 18 people across RL environment startups, <a href="https://www.theinformation.com/articles/investors-chase-neolabs-outflank-openai-anthropic">neolabs</a>, and frontier labs. We asked them what RL environments and tasks look like, how labs use them, what makes a good one, and where the field is headed.</p><p><strong>Main takeaways:</strong></p><ul><li><p><strong>Enterprise workflows are a major growth area</strong>. Math and coding tasks came first, but we&#8217;re now seeing significant growth in enterprise workflows: tasks like navigating Salesforce, filing reports, or manipulating spreadsheets.</p></li><li><p><strong>Reward hacking is a top concern</strong>. Interviewees consistently cited robustness against reward hacking as a key quality criterion. Models find ways to game graders, and preventing this requires extensive iteration on both environments and tasks.</p></li><li><p><strong>Scaling without sacrificing quality is hard</strong>. A major challenge is scaling the quantity of environments and tasks without sacrificing quality. The hard parts are management (coordinating a growing number of task builders) and maintaining good quality assessment processes.</p></li></ul><h2><strong>What are RL environments and tasks?</strong></h2><p>In modern reinforcement learning for language models, the model is given a task to accomplish and a set of actions it can take. The model attempts the task, and a grader (typically automated, such as a unit test or an LLM judging against a rubric) assigns a score to its attempts. These scored attempts are then used to update the model&#8217;s weights, reinforcing successful strategies.</p><p>The RL environment is defined by the set of actions the model can take (running code, thinking out loud, clicking buttons, searching documents) and the surrounding context that determines the effect of these actions (environment variables, file systems, the state of a simulated application). In practice, the environment is often delivered as a Docker container, but not always<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a>.</p><p>Each task consists of a prompt instructing the model to achieve an objective, and a grader that determines whether (or to what extent) the objective was met. Terminology in this space isn&#8217;t fully standardized, and the boundary between &#8220;environment&#8221; and &#8220;task&#8221; is somewhat fuzzy<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a>. In this piece we discuss both environments and tasks, since they&#8217;re often built and sold together.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!bv6y!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21b4870f-4a72-47e3-be91-36ee75329c81_1280x720.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!bv6y!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21b4870f-4a72-47e3-be91-36ee75329c81_1280x720.png 424w, https://substackcdn.com/image/fetch/$s_!bv6y!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21b4870f-4a72-47e3-be91-36ee75329c81_1280x720.png 848w, https://substackcdn.com/image/fetch/$s_!bv6y!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21b4870f-4a72-47e3-be91-36ee75329c81_1280x720.png 1272w, https://substackcdn.com/image/fetch/$s_!bv6y!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21b4870f-4a72-47e3-be91-36ee75329c81_1280x720.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!bv6y!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21b4870f-4a72-47e3-be91-36ee75329c81_1280x720.png" width="1280" height="720" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/21b4870f-4a72-47e3-be91-36ee75329c81_1280x720.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:720,&quot;width&quot;:1280,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:75889,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://epochai.substack.com/i/184345628?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21b4870f-4a72-47e3-be91-36ee75329c81_1280x720.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!bv6y!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21b4870f-4a72-47e3-be91-36ee75329c81_1280x720.png 424w, https://substackcdn.com/image/fetch/$s_!bv6y!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21b4870f-4a72-47e3-be91-36ee75329c81_1280x720.png 848w, https://substackcdn.com/image/fetch/$s_!bv6y!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21b4870f-4a72-47e3-be91-36ee75329c81_1280x720.png 1272w, https://substackcdn.com/image/fetch/$s_!bv6y!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F21b4870f-4a72-47e3-be91-36ee75329c81_1280x720.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p>Here are some examples of environments and the kinds of tasks they could support:</p><ul><li><p>A git repository, with tasks like fixing a bug so that unit tests pass, similar to benchmarks like <a href="https://epoch.ai/blog/what-skills-does-swe-bench-verified-evaluate">SWE-bench Verified</a>. The task specifies a git repository at a specific commit with a failing test suite; the environment provides the operating system and tools to interact with the repo; the grader runs the tests and checks that they pass.</p></li><li><p>An Airbnb clone, with tasks like finding the cheapest two-bedroom listing in a given city for specific dates. The environment is a simulated website with realistic listings, prices, and filters; the agent sees a structured representation of the page (like a DOM or accessibility tree) and outputs actions like clicking elements or typing into fields. The grader verifies the final answer.</p></li><li><p>A Bloomberg terminal clone, with tasks like finding the 5-year compound annual growth rate for a list of companies. The environment simulates the terminal&#8217;s interface and data; the grader checks whether the returned figures match the correct values.</p></li><li><p>An Excel clone, with tasks like creating a pivot table showing revenue by region from a raw dataset. The environment provides a spreadsheet application with realistic functionality; the grader compares the output against a reference solution.</p></li></ul><p>For computer use environments like the Excel clone, a single environment might support hundreds of different tasks. For coding environments, it&#8217;s more common (but not always the case) for each environment to contain just one task, since setting up a repo state is relatively cheap<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a>.</p><h2><strong>How are RL environments used by labs?</strong></h2><p>Each environment and task can be used in three main ways: for reinforcement learning, for benchmarking, or for supervised fine-tuning on trajectories that solve the task.</p><p>Reinforcement learning remains the primary use case. As one RL environment startup employee put it: &#8220;RL is the main use. We have some requests for creating envs for benchmarking. I&#8217;d say perhaps 10-20x more the former vs the latter.&#8221; One difference is that benchmarks are typically built for single-turn evaluation, whereas there&#8217;s growing interest in RL environments that capture multi-turn interactions between agent and user.</p><p>Environments are also used to generate data for supervised learning, by using successful RL trajectories as training examples during midtraining. One interviewee noted: &#8220;While it might not drive purchasing today, a well-designed environment can be used as an effective mechanism for synthetic data generation. I feel this will be increasingly important as env development matures and designers target this use-case.&#8221;</p><p>An interviewee noted that supervised fine-tuning (SFT) had been growing especially for interleaved thinking and tool calling. With SFT, you can choose a single good trajectory and train on that, whereas RL requires multiple trajectories with enough differentiation between them to provide a learning signal. This makes SFT more practical when it&#8217;s relatively easy to produce good trajectories but hard to get a reliable grader or enough variation between attempts.</p><p>Usage patterns also vary by lab. One RL environment startup founder noted that trailing labs tend to be more interested in SFT than RL: &#8220;It&#8217;s quite hard to get your really large scale RL pipelines and it&#8217;s easy to throw stuff into your SFT mix.&#8221;</p><p>RL environment companies don&#8217;t always have full visibility into how their products are ultimately used, and usage likely changes over time as labs&#8217; training pipelines evolve. As one founder put it when asked how customers use the environments they build: &#8220;To what extent are they using it to generate gold trajectories<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a> versus doing RL? I think that would be pretty hard to answer.&#8221;</p><h2><strong>Which companies build RL environments?</strong></h2><p>A growing number of <strong>specialized startups</strong> focus specifically on building RL environments. They cover a range of domains, from software engineering tasks to computer use to math and finance. Chris compiled a list of startups in this space <a href="https://pavlovslist.com/">here</a>.</p><p><strong>Traditional data providers</strong> like Mercor, Surge, Handshake, and Turing, who used to primarily provide human-labeled data, now also sell RL environments. Part of what you pay for is their QA processes and supervision infrastructure, but as one founder put it, the main value add is &#8220;they have the guys&#8221;: if you need to scale up task creation quickly, they can staff a project faster than you could hire in-house.</p><p><strong>In-house teams</strong> at model developers are also building environments (<em>cf.</em> job postings by <a href="https://x.ai/careers/open-roles?dept=4025087007">xAI</a> or <a href="https://www.anthropic.com/careers/jobs/5061517008">Anthropic</a>). This includes both frontier labs and neolabs like Cursor, who can leverage user data to build training tasks. One founder noted there&#8217;s been &#8220;substantially more in-housing&#8221; recently, as labs build out their own data teams. Reasons to do things in-house include avoiding margins paid to vendors, keeping training priorities confidential, and having direct expertise in some domains like AI research.</p><p>Finally, <strong>product companies</strong> are natural partners for building environments around their own software. If you&#8217;re Salesforce or Slack, you understand your product&#8217;s interface and edge cases better than anyone. We&#8217;re seeing a wave of partnerships between labs and product companies: Benchling and Anthropic for biology workflows, OpenAI with Shopify and Stripe for shopping, OpenAI&#8217;s recent health integrations. While we don&#8217;t know the exact details of how these partnerships work, it seems likely that the product companies are often at least collaborating on environment or task creation. In some cases, though, companies are hostile to agent traffic: Amazon has <a href="https://www.reuters.com/business/retail-consumer/perplexity-receives-legal-threat-amazon-over-agentic-ai-shopping-tool-2025-11-04/">sued</a> Perplexity over its agentic shopping tool and blocked most AI agents from accessing its site.</p><h2><strong>How much do environments and tasks cost?</strong></h2><p>Costs are highly variable, depending on the domain, complexity, and number of tasks.</p><p>Contract sizes are often six to seven figures per quarter. For example, one RL environment founder noted that contracts are often seven figures per quarter or more. A neolab researcher mentioned seeing contracts in the $300k-$500k range, adding that it varies a lot depending on the number of tasks.</p><p>Environment costs depend on fidelity.<a href="https://newsletter.semianalysis.com/p/rl-environments-and-rl-for-science"> SemiAnalysis reported</a> that website replicas (&#8221;UI gyms&#8221;) cost around $20k each. But higher-quality replicas of complex products like Slack can cost significantly more: one interviewee mentioned figures around $300k for these.</p><p>Task costs also vary, but multiple interviewees agreed on a ballpark of $200-$2,000 per task. As one RL environment founder put it: &#8220;I&#8217;ve seen $200 to $2000 mostly. $20k per task would be rare but possible.&#8221; A neolab researcher confirmed this range aligns with their experience. The $20k figure comes up for especially complex software engineering tasks, but it&#8217;s rare. To put task costs in perspective:<a href="https://www.mechanize.work/blog/cheap-rl-tasks-will-waste-compute"> Mechanize estimates</a> that around $2,400 in compute is spent per task during RL training. This suggests that cheaper, lower-quality tasks might end up wasting that compute.</p><p>Exclusivity affects pricing significantly. Environments and tasks can be sold exclusively to one customer or non-exclusively to multiple labs. Two RL environment founders independently agreed that exclusive deals are roughly 4-5x more expensive than non-exclusive ones.</p><p>Overall spending in this space is growing rapidly. As mentioned above, The Information <a href="https://www.theinformation.com/articles/anthropic-openai-developing-ai-co-workers?rc=dp0mql">reported</a> that Anthropic had discussed spending over $1 billion on RL environments over the following year. However, this is still a fraction of compute costs. OpenAI&#8217;s R&amp;D compute spend in 2026 is <a href="https://www.theinformation.com/articles/openai-spend-100-billion-backup-servers-ai-breakthroughs?rc=1hiqbr">projected</a> to be around $19 billion, roughly double the 2025 figure. Even accounting for the fact that Anthropic is smaller than OpenAI, compute spending will still dwarf RL environment spending.</p><h2><strong>What domains do RL environments cover?</strong></h2><p>Originally, the main domains were mathematics and coding<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a>. As one neolab researcher put it: &#8220;Coding and in some sense math kickstarted all the RL environments explorations. So code and math are most plentiful.&#8221;</p><p>Mathematics tasks are relatively easy to produce, since you don&#8217;t need to build a complex environment, just tasks with verifiable answers. However, emphasis on math may be declining more recently. One interviewee noted that &#8220;math might be shrinking,&#8221; and an RL environment founder observed that math tasks are easy to create but don&#8217;t transfer as well to other capabilities.</p><p>Coding remains a major source of demand. It&#8217;s a huge market, and there&#8217;s already been demonstrated uplift from training on code tasks. We&#8217;re also seeing coding environments go beyond SWE-bench-style tasks. One RL environment founder noted: &#8220;I&#8217;ve seen code environments go from more simple PASS_TO_PASS and FAIL_TO_PASS<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a> type SWE-bench Verified tasks more towards being productionized. So how does a SWE actually work in an environment? They have GitHub, they have Linear, they have a code IDE.&#8221; A key challenge is that many tasks don&#8217;t have clean verification criteria beyond &#8220;passes unit tests.&#8221; As one neolab researcher noted: &#8220;There&#8217;s probably many solutions that pass the unit tests. But there&#8217;s some nicer solutions that have better trade-offs in terms of your engineering design decisions.&#8221; This poses a challenge on the grading front.</p><p>One of the major growth areas is enterprise workflows: tasks like filing an expense report, creating a pivot table in a spreadsheet, generating slides from a brief, or navigating a CRM to update customer records. One RL environment founder told us: &#8220;I think enterprise workflows are going to explode this year. I think labs index very heavily on what&#8217;s valuable and what is quantifiable, and enterprise workflows are perfect for that. By enterprise workflows I mean things that are sometimes computer use (e.g. a specific ERP<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-8" href="#footnote-8" target="_self">8</a> that has no backend) or sometimes things that involve APIs that can be exposed to an agent and used without computer use.&#8221;</p><p>Environments for enterprise workflows can take various forms: MCP-style tool integrations, Playwright-style browser interactions, or screenshot-based computer use. Many rely on clones of websites and apps like Slack or SAP. One lab researcher cautioned: &#8220;There&#8217;s a lot of good reasons to use clones of websites, but what everyone does is vibe code a buggy website which isn&#8217;t useful. There&#8217;s a large amount of useless bad environments out there for that reason.&#8221;</p><p><a href="https://www.primeintellect.ai/blog/scaling-environments-program">Prime Intellect&#8217;s environment bounties</a> give a sense of some of the domains covered by RL environments and how they&#8217;re built. They include environments based on prominent benchmarks, environments for interacting with MCP servers, AI research tasks like<a href="https://docs.google.com/spreadsheets/d/13UDfRDjgIZXsMI2s9-Lmn8KSMMsgk2_zsfju6cx_pNU/edit?gid=650541192#gid=650541192&amp;range=A110"> nanoGPT Speedrunning</a>, and tasks based on questions sourced from textbooks.</p><p>Overall, the main directions in the near future are coding and enterprise workflows. Across both, there&#8217;s growing interest in longer-horizon tasks. As one RL environment founder put it: &#8220;Long horizon is I think the future direction. Having agents do full end-to-end tasks that involve navigating through multiple tabs, browsers, and then submitting something that involves multi-hop steps.&#8221;</p><p>A few other directions came up in our interviews: training for multi-turn user interactions rather than just single-turn performance, environments that optimize for multiple goals at once, and environments with tooling allowing researchers to inspect and change trajectories to provide feedback.</p><h2><strong>What are the top priorities and challenges?</strong></h2><p>Interviewees consistently cited robustness against reward hacking as the most important quality criterion. As one neolab researcher put it: &#8220;Reward hacking is a big issue. The model might cheat by searching up a solution, or if you&#8217;re not careful with how you script the repo, by checking out future commits<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-9" href="#footnote-9" target="_self">9</a>. It needs to be robust. That&#8217;s the minimum.&#8221; Another framed it similarly: &#8220;Soundness matters most: high reward must mean the task was actually solved, not hacked.&#8221; Creating robust graders rarely works on the first pass: as one RL environment founder noted, &#8220;It takes many many iterations to check against reward hacking.&#8221; Notably, this remains a top priority even though models have become<a href="https://assets.anthropic.com/m/64823ba7485345a7/Claude-Opus-4-5-System-Card.pdf#page=103"> better at not reward hacking over the past year</a>.</p><p>Difficulty calibration also matters. Tasks need to be challenging but not impossible, since if the pass rate is 0% or 100%, the model won&#8217;t learn<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-10" href="#footnote-10" target="_self">10</a>. Moreover, as one RL environment founder noted: &#8220;You don&#8217;t want it to be zero percent because then there&#8217;s a possibility that the task is infeasible, unless you&#8217;ve had another annotator do a blind solve and they succeed.&#8221; Multiple interviewees mentioned wanting a minimum pass rate of around 2-3%, or at least one success out of 64 or 128 rollouts.</p><p>Beyond individual task difficulty, the overall distribution matters. As one neolab researcher put it: &#8220;A very important feature of the RL environment is a smooth gradient: diversity in the difficulty of the tasks.&#8221; One person noted you might want a mix: some tasks at 0%, some at 5%, and some at 30%. After some training steps, the 0% tasks become learnable. Once a task reaches around 70% pass rate, you might discard it and move on to harder tasks.</p><p>Separately, tasks should also be compositional. As one neolab researcher noted: &#8220;The skills should not be disparate tasks, they should be similar to each other and leverage common skills.&#8221;</p><p>Scaling while maintaining quality is the core operational bottleneck. As<a href="https://kevinlu.ai/the-only-important-technology-is-the-internet"> Kevin Lu has argued</a>, scaling RL environments is one of the key challenges for continued AI progress. But scaling task creation while maintaining quality is very hard. One RL environment founder noted: &#8220;Maintaining quality while scaling is the number one bottleneck that people see. Finding the experts isn&#8217;t that hard, but managing them and doing quality control is hard.&#8221; A neolab researcher emphasized the management challenge: &#8220;It&#8217;s not easy to find people to oversee this data construction, the RL environment construction process. The contractors, you need to motivate them. Sure, you&#8217;re paying them money. But how do you make sure they&#8217;re not just using LLMs? How do you make sure they&#8217;re actually verified? Motivating the contractors and doing the quality control is the grunt work.&#8221; One RL environment founder noted that their constraint on making more revenue is simply the difficulty of scaling up task creation at the required quality level.</p><p>The skills required are a mix of domain expertise, prompting ability, and product sense. Building environments mostly involves engineering skills, but creating good tasks requires something different. As one RL environment founder put it: &#8220;Domain knowledge and expert level prompting is more important than ML skills for creating tasks.&#8221; A neolab researcher added that product sense matters too: &#8220;You want to know how people are actually using these tools.&#8221; Having a lot of experience interacting with frontier models is key, as one neolab researcher noted: &#8220;You don&#8217;t necessarily need to be an AI researcher, but perhaps a very heavy Claude Code user, a prompt whisperer like Riley Goodside,  can be better at figuring out what the frontier is than an AI researcher.&#8221; Another put it simply: &#8220;The people who are probably best at this are the people who create benchmarks that actually get used.&#8221;</p><p>RL environments have quickly become a major input to frontier AI training. Key challenges include preventing reward hacking and scaling production without sacrificing quality, while demand is growing for enterprise workflows and longer-horizon tasks. This is a fast-moving space, and we expect this picture to look quite different in a year.</p><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://epochai.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading! Subscribe to receive the latest from Epoch AI.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p><em>Thank you to all the interviewees, as well as to Jaime Sevilla, Josh You, and Lynette Bye for comments and editing help on this piece.</em></p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>We had calls with 9 different people, got text/email input from an additional 9 people, and sanity checks but no input from 4 more people.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>For example, the environments on Prime Intellect&#8217;s environment hub don&#8217;t rely on Docker, instead using a more lightweight approach based on their verifiers library. In some domains, for example most mathematics problems, there isn&#8217;t even a need for an environment since the model does not have to use tools.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>In traditional RL terminology, the &#8220;environment&#8221; encompasses the transition dynamics (how actions affect state), the reward function, and the distribution over starting states. A &#8220;task&#8221; in that framing is simply a specific starting state. In practice, the way people use these terms varies.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>Here again, the terminology isn&#8217;t fully standardized, and varies across domains: some people use &#8220;environment&#8221; to refer to a single Docker container with one task; others use it to mean a collection of containers with many tasks.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>Gold trajectories are human-verified examples of successfully completing a task, which can be used as training data for supervised fine-tuning.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>In a sense, RLHF and Constitutional AI came before math and coding environments (and, broadening still, physics engines like MuJoCo came even earlier). They both involve training language models with RL, although typically there's been less emphasis on tool use and environments. In practice, people tend to think of them as separate categories, and the actors involved are often different.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>In SWE-bench Verified, PASS_TO_PASS refers to tests that should continue passing after the model&#8217;s edits, while FAIL_TO_PASS refers to tests that were failing before and should pass after the fix.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-8" href="#footnote-anchor-8" class="footnote-number" contenteditable="false" target="_self">8</a><div class="footnote-content"><p>ERP stands for Enterprise Resource Planning, software used by businesses to manage day-to-day activities like accounting, procurement, and project management. Examples include SAP and Oracle.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-9" href="#footnote-anchor-9" class="footnote-number" contenteditable="false" target="_self">9</a><div class="footnote-content"><p>See <a href="https://x.com/tmkadamcz/status/1963996138044096969?s=46">here</a> for an example of the model looking at future commits on SWE-bench Verified.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-10" href="#footnote-anchor-10" class="footnote-number" contenteditable="false" target="_self">10</a><div class="footnote-content"><p>RL algorithms like <a href="https://arxiv.org/abs/2501.12948">GRPO</a> improve by comparing different attempts at the same task. If one attempt succeeds and another fails, the model learns to favor the successful strategy. But if every attempt gets the same score (all passing or all failing), there&#8217;s no signal about which approaches work better.</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[How far can decentralized training over the internet scale?]]></title><description><![CDATA[Decentralized AI training is growing fast, and while it likely won&#8217;t catch up to frontier models this decade, even narrowing the gap could have major implications for AI policy.]]></description><link>https://epochai.substack.com/p/how-far-can-decentralized-training</link><guid isPermaLink="false">https://epochai.substack.com/p/how-far-can-decentralized-training</guid><dc:creator><![CDATA[Jaime Sevilla]]></dc:creator><pubDate>Mon, 29 Dec 2025 17:06:57 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!sMDM!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48476ea1-d58f-4d9a-91c6-d440a8328621_1024x1280.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>This post is part of our <a href="https://epochai.substack.com/s/gradient-updates">Gradient Updates</a> newsletter, which shares more opinionated or informal takes about big questions in AI progress. These posts solely represent the views of the authors, and do not necessarily reflect the views of Epoch AI as a whole.</em></p><div><hr></div><p>Previously, I <a href="https://epoch.ai/blog/could-decentralized-training-solve-ais-power-problem">discussed</a> decentralized training in the context of hyperscalers. Microsoft, Google and other giants are building interconnected gigawatt scale datacenters, which could be used to train models at an unprecedented computational scale. The decentralization could sidestep the difficulty of securing 10 GW of power in a single location by splitting one massive run into ten more manageable gigawatt-scale blocks.</p><p>But when people think of decentralized training, they don&#8217;t first think of gigantic datacenters, owned by the same company, training models across large distances. Instead, <a href="https://www.galaxy.com/insights/research/decentralized-ai-training">they</a> imagine <a href="https://blog.pluralis.ai/p/decentralized-ai-looms">thousands</a> of small datacenters, or individual consumers, <a href="https://www.primeintellect.ai/blog/our-approach-to-decentralized-training">pooling</a> their spare compute over the internet to orchestrate a training run larger than any single actor could manage alone.</p><p>Many companies are pursuing this vision: <a href="https://pluralis.ai/">Pluralis Research</a>, <a href="https://www.primeintellect.ai/">Prime Intellect</a> and <a href="https://nousresearch.com/">Nous Research</a> have already successfully decentrally trained models at scale. But in practice, training decentrally over the internet has lagged far behind more centralized training. Even their largest models (Pluralis&#8217; 8B Protocol Model, Prime Intellect&#8217;s INTELLECT-1, and Nous&#8217; Consilience 40B) have been trained with 1,000x less compute than <a href="https://epoch.ai/data/ai-models">today&#8217;s frontier models</a> (such as xAI&#8217;s Grok 4).</p><p>And, importantly, many proposals for monitoring and regulating AI <a href="https://jack-clark.net/2024/06/03/import-ai-375-gpt-2-five-years-later-decentralized-training-new-ways-of-thinking-about-consciousness-and-ai/">depend</a> on internet decentralized training continuing to lag far behind the frontier. The compute is easy to <a href="https://www.transformernews.ai/p/decentralized-training-policy-implications">track</a> when it&#8217;s centralized, either in a few massive data centers or many smaller (but still large) data centers connected via a dedicated network of fiber optic cables. As long as it only makes economic sense for frontier models to be trained thusly by a handful of the biggest companies, governments can relatively easily regulate them.</p><p>However, if that assumption &#8211; that training over the internet across thousands of computers isn&#8217;t feasible &#8211; breaks, then regulation might be left scrambling to catch up.</p><p>In this article, I review and critically examine the state of the field of decentralized training over the internet. I conclude that while decentralized training is growing fast (around 20x per year) and technically feasible at frontier scale, it&#8217;s unlikely that decentralized developers will amass frontier amounts of compute this decade.</p><h2>Is decentralized training over the internet feasible?</h2><p>Decentralized training over the internet is a strictly harder engineering task than centralized development.</p><p>The three big additional challenges that decentralized training poses are: low-bandwidth communication, managing a network of heterogeneous devices with inconsistent availability, and trust and coordination issues. While important, I believe the latter two won&#8217;t ultimately impede training at scale.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a></p><p>Low bandwidth poses a more important problem. Typical internet upload bandwidths for consumers are around 60 Mbps. If using naive data parallelism to distribute training across nodes with this much bandwidth, it would take 5,000 years to train a DeepSeek v3 style model with 671B parameters.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a></p><p>But Pluralis, Prime Intellect and Nous Research have already trained (and postrained) models with billions of parameters. The key is in a family of techniques to reduce bandwidth requirements, and enable different kinds of decentralized training: data parallelism, model parallelism, and RL training.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!j4Fi!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4afdc73f-ce39-4dbc-a75f-018a70d4b543_1024x1280.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!j4Fi!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4afdc73f-ce39-4dbc-a75f-018a70d4b543_1024x1280.png 424w, https://substackcdn.com/image/fetch/$s_!j4Fi!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4afdc73f-ce39-4dbc-a75f-018a70d4b543_1024x1280.png 848w, https://substackcdn.com/image/fetch/$s_!j4Fi!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4afdc73f-ce39-4dbc-a75f-018a70d4b543_1024x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!j4Fi!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4afdc73f-ce39-4dbc-a75f-018a70d4b543_1024x1280.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!j4Fi!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4afdc73f-ce39-4dbc-a75f-018a70d4b543_1024x1280.png" width="1024" height="1280" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4afdc73f-ce39-4dbc-a75f-018a70d4b543_1024x1280.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1280,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:114519,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://epochai.substack.com/i/182853461?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4afdc73f-ce39-4dbc-a75f-018a70d4b543_1024x1280.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!j4Fi!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4afdc73f-ce39-4dbc-a75f-018a70d4b543_1024x1280.png 424w, https://substackcdn.com/image/fetch/$s_!j4Fi!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4afdc73f-ce39-4dbc-a75f-018a70d4b543_1024x1280.png 848w, https://substackcdn.com/image/fetch/$s_!j4Fi!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4afdc73f-ce39-4dbc-a75f-018a70d4b543_1024x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!j4Fi!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4afdc73f-ce39-4dbc-a75f-018a70d4b543_1024x1280.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><h3>Decentralized data parallelism</h3><p>When using data parallelism to distribute training across multiple data centers, we have the centers communicate once every batch to synchronize gradient updates.</p><p>When training over the internet, that communication cost would be prohibitive. For instance, with typical upload speeds reaching around 60 Mbps, the largest model we could train while keeping synchronization times under ten minutes would be a rather small 600M parameter model with 32-bit gradients.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a></p><p>One solution is to reduce the frequency with which each data center or individual computer (each node) talks to each other. So instead of synchronizing weights with every batch, we let each node process multiple batches independently (these are called &#8220;<em>inner steps</em>&#8221;), then aggregate the cumulative gradient updates each node has taken in the synchronization step (an &#8220;<em>outer step</em>&#8221;).</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!dtn8!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd879a68-a0db-4b8f-a533-f9331e7045d8_1316x403.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!dtn8!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd879a68-a0db-4b8f-a533-f9331e7045d8_1316x403.png 424w, https://substackcdn.com/image/fetch/$s_!dtn8!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd879a68-a0db-4b8f-a533-f9331e7045d8_1316x403.png 848w, https://substackcdn.com/image/fetch/$s_!dtn8!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd879a68-a0db-4b8f-a533-f9331e7045d8_1316x403.png 1272w, https://substackcdn.com/image/fetch/$s_!dtn8!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd879a68-a0db-4b8f-a533-f9331e7045d8_1316x403.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!dtn8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd879a68-a0db-4b8f-a533-f9331e7045d8_1316x403.png" width="1316" height="403" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bd879a68-a0db-4b8f-a533-f9331e7045d8_1316x403.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:403,&quot;width&quot;:1316,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!dtn8!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd879a68-a0db-4b8f-a533-f9331e7045d8_1316x403.png 424w, https://substackcdn.com/image/fetch/$s_!dtn8!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd879a68-a0db-4b8f-a533-f9331e7045d8_1316x403.png 848w, https://substackcdn.com/image/fetch/$s_!dtn8!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd879a68-a0db-4b8f-a533-f9331e7045d8_1316x403.png 1272w, https://substackcdn.com/image/fetch/$s_!dtn8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd879a68-a0db-4b8f-a533-f9331e7045d8_1316x403.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure from <a href="https://arxiv.org/abs/2311.08105">Douillard et al. (2023)</a>. Each node holds a replica of the model, and trains independently for a number of inner steps before synchronizing across nodes.</figcaption></figure></div><p>This process was demonstrated in the DiLoCo paper by <a href="https://arxiv.org/abs/2311.08105">Douillard et </a>al. (2023), building on previous work, such as that by Sticht et al.<a href="https://arxiv.org/abs/1805.09767"> (2018)</a>. This process is not equivalent to the usual training process, and as such it can harm performance. However, the DiLoCo authors find that they can conduct training with up to 500 inner steps with only a small reduction in performance, for a 500x reduction in bandwidth requirements.</p><p>One key weakness of DiLoCo worth stressing is that the performance is reduced as the number of nodes increases. In particular, increasing from 1 to 8 nodes is equivalent to a 1.5x decrease in training compute (see table 4 and 11 of <a href="https://arxiv.org/abs/2503.09799v1">Charles et al, 2025</a>). Naively, this suggests that scaling to 10,000 nodes would require 6x as much FLOP as centralized training for the same performance. This is not trivial, though it can be compensated for by training for longer, or with larger clusters.</p><p>To further reduce the strain on the bandwidth, we can reduce the size of the gradients. The main way this is done in practice is <em>quantization</em>, where the gradients are communicated on a lower bit precision. By reducing the gradients to 8-bits, as in the <a href="https://arxiv.org/abs/2412.01152">INTELLECT-1</a> run, or to 4 bits, as in the <a href="https://arxiv.org/abs/2501.18512">Streaming DiLoCo</a> paper, we can reduce bandwidth requirements by 2-4x compared to typical 32-bit gradient synchronization. Some even have experimented with 1-bit gradient updates (<a href="https://arxiv.org/abs/2102.02888">Tang et al, 2021</a>; <a href="https://arxiv.org/abs/2202.06009">Lu et al, 2022</a>), though their efficacy is yet to be shown at scale.</p><p>Another type of compression is <em>sparsification</em>, in which only the most significant (top-k sparsification) or a random subset of the gradients (random sparsification) are communicated in each round. The <a href="https://arxiv.org/abs/2411.19870">decoupled momentum (DeMo) optimizer</a> by Nous Research is a variant of this, in which larger updates are prioritized, while smaller gradients are accumulated until they reach a threshold of significance.</p><p>Both quantization and sparsification degrade the quality of the gradients communicated. To fight this, one technique often used is <em>error feedback accumulation</em>: the difference between the gradients computed and communicated is stored, to be communicated in future updates.</p><p>These methods can be combined. <a href="https://proceedings.mlr.press/v202/wang23t.html">CocktailSGD</a> experiments with finetuning models with up to 20B parameters combining quantization and sparsification. <a href="https://arxiv.org/abs/2505.23725">MuLoCo</a> further combines these techniques with DiLoCo and a Muon optimizer. <a href="https://templarresearch.substack.com/p/checkpoint-one">SparseLoCo</a> takes a similar approach &#8211; and it is now being applied at scale as part of the <a href="https://www.tplr.ai/dashboard">Covenant72B</a> decentralized training run by Covenant AI.</p><p>We can optimize DiLoCO further. In the Streaming DiLoCo paper, <a href="https://arxiv.org/abs/2501.18512">Douillard et al (2025)</a> explain how to sequence synchronization on different subsets of parameters while training. This enables overlapping computation with communication, which together with 4-bit gradient quantization, reduces bandwidth strain by 100x compared to naive data parallelism.</p><h3>Decentralized model parallelism</h3><p>While most decentralized training today leverages data parallelism, others &#8212; notably Pluralis Research &#8212; have experimented with <em>model parallelism</em>.</p><p>In model parallelism, the parameters of the models are themselves split between nodes. This has one major advantage: since each node doesn&#8217;t need to hold the whole model, nodes with small memory such as consumer GPUs can be part of the network without constraining the model size.</p><p>For instance, the SWARM parallelism paper by<a href="https://arxiv.org/abs/2301.11913"> Ryabinin et al (2023)</a> proposes decentralized pipeline parallelism, where nodes process subsets of model layers, propagating activations and gradients between them. The paper notes the <em>square-cube law</em>: as we increase the model dimension,  communication scales linearly, but compute scales <em>quadratically</em>. This means that larger models suffer less from the communication overhead, making pipeline parallelism a practical choice in low-bandwidth setups.</p><p>Pluralis&#8217; <a href="https://arxiv.org/abs/2506.01260">Protocol Models</a> further optimize pipeline parallelism by constraining each layer&#8217;s activations to a small subspace, for up to a 100x reduction in size. This allows them to train a 8B Llama-like model in a decentralized fashion, as well as a 7.5B model <a href="https://dashboard.pluralis.ai/">through a volunteer network</a>.</p><p>Other forms of model parallelism have been experimented with. Prime Intellect used pipeline parallelism for <a href="https://www.primeintellect.ai/blog/synthetic-2">SYNTHETIC-2</a> &#8211; a decentralized synthetic data generation project. The learning@home paradigm by <a href="https://arxiv.org/abs/2002.04013">Ryabinin and Gusev (2020)</a> and the SPES protocol by <a href="https://openreview.net/forum?id=x3qnrhfhX0">Zhang et al (2025)</a> splits different experts from a mixture-of-experts (MoE) across different nodes for training and inference. <a href="https://openreview.net/forum?id=x3qnrhfhX0">Pluralis</a> has also experimented with decentralized context parallelism in <a href="https://openreview.net/forum?id=x3qnrhfhX0">Ramasinghe et al (2025)</a>, in which different nodes process different parts of a sequence, and need to coordinate to compute the attention context.</p><h3>Decentralized RL training</h3><p>Since last year we have seen a rise in the prominence of RL postraining methods, often called reasoning training.</p><p>In a typical RL training scheme (such as <a href="https://arxiv.org/abs/2402.03300">GRPO</a>) we generate many reasoning traces with the latest model checkpoint. Each trace is automatically scored relative to the other traces in the group. These scores are used to update the model weights, increasing the relative log-likelihood of the highest scoring traces relative to the worse ones.</p><p>In asynchronous RL, we allow the trace generation to be done with an outdated version of the model, often 2 to 5 steps behind the current checkpoint. This has the advantage that we can overlap communicating the updates with generating the traces, mitigating communication bottlenecks.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!2iyB!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4eb474ce-dc72-440f-9eaf-8b854914ab87_1600x784.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!2iyB!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4eb474ce-dc72-440f-9eaf-8b854914ab87_1600x784.png 424w, https://substackcdn.com/image/fetch/$s_!2iyB!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4eb474ce-dc72-440f-9eaf-8b854914ab87_1600x784.png 848w, https://substackcdn.com/image/fetch/$s_!2iyB!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4eb474ce-dc72-440f-9eaf-8b854914ab87_1600x784.png 1272w, https://substackcdn.com/image/fetch/$s_!2iyB!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4eb474ce-dc72-440f-9eaf-8b854914ab87_1600x784.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!2iyB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4eb474ce-dc72-440f-9eaf-8b854914ab87_1600x784.png" width="1456" height="713" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4eb474ce-dc72-440f-9eaf-8b854914ab87_1600x784.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:713,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!2iyB!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4eb474ce-dc72-440f-9eaf-8b854914ab87_1600x784.png 424w, https://substackcdn.com/image/fetch/$s_!2iyB!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4eb474ce-dc72-440f-9eaf-8b854914ab87_1600x784.png 848w, https://substackcdn.com/image/fetch/$s_!2iyB!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4eb474ce-dc72-440f-9eaf-8b854914ab87_1600x784.png 1272w, https://substackcdn.com/image/fetch/$s_!2iyB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4eb474ce-dc72-440f-9eaf-8b854914ab87_1600x784.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure from the <a href="https://www.primeintellect.ai/blog/intellect-2-release">INTELLECT-2 release post</a>, illustrating asynchronous RL training.</figcaption></figure></div><p>The largest decentralized RL training run to date is <a href="https://arxiv.org/abs/2505.07291">INTELLECT-2</a> by Prime Intellect, which coordinated the output of over 800+ nodes to generate traces and post train a <a href="https://qwenlm.github.io/blog/qwq-32b/">QwQ-32B</a> base model using asynchronous RL.</p><p>In synchronous RL, the burden of compute is tilted more towards inference than in traditional pretraining. The ratio of compute in FLOP is the same &#8212; a forward pass to compute the activations is matched by twice as much compute in the backward pass to compute the gradients &#8212; but trace generation is memory-bound, making it take more GPU-hours than the backward pass. In asynchronous RL, because the forward pass must be repeated with up-to-date model weights, the amount of inference compute is even higher.</p><p>INTELLECT-2 employs <em>Online Advantage Filtering</em> to shift the burden even more towards inference: groups of reasoning traces with no advantage are discarded, since they generate no training signal. This allowed for a decentralized inference network that used 4x more compute than the training nodes.</p><p>More recently, Prime Intellect unveiled their <a href="https://www.primeintellect.ai/blog/intellect-3">INTELLECT-3</a> run. This was a larger scale finetuning and asynchronous RL post-training run, albeit run in a centralized setting. The success of this run paves the way for larger decentralized RL training runs in the future.</p><h2>Putting it all together: decentralized internet training at frontier scale is likely feasible</h2><p>My overall impression is that all the techniques we have discussed could be combined to train much larger decentralized models than exist today.</p><p>Taking for example the INTELLECT-1 model, one of the largest decentralized pretraining runs to date. It is a 10B parameter model trained on 1T tokens, which amounts to 6e22 FLOP. And it essentially only leveraged two basic decentralized training techniques to manage bandwidth: DiLoCo with 100 inner steps per synchronization, and 8 bit quantization.</p><p>If we further quantize to 4-bits and apply 25% sparsification as in <a href="https://arxiv.org/abs/2508.15706">SparseLoCo</a>, we would theoretically be able to train a model 8x larger, which per <a href="https://epoch.ai/publications/chinchilla-scaling-a-replication-attempt">Chinchilla scaling</a> would allow for training runs with about 64x more compute. A similar strategy is how Templar is currently decentrally training a <a href="https://templarresearch.substack.com/p/checkpoint-one">72B parameter model</a>.</p><p>Going further, during the INTELLECT-1 training run, the contributor nodes ran independently for about 38 minutes before synchronizing for 7 minutes. Applying <a href="https://arxiv.org/abs/2501.18512">streaming DiLoCo</a> could hide a large part of that communication.  With an optimized interleaving schedule, this could allow for training a 2x larger model, i.e., 4x more compute.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a></p><p>As we alluded to, many of these techniques moderately harm performance. While not trivial, the performance reduction looks modest in practice. For example, based on the evaluations results reported by developers, INTELLECT-1 would achieve an <a href="https://epoch.ai/benchmarks/eci">Epoch Capability Score (ECI)</a> around 98. This is e.g. similar to Qwen2.5-Coder 1.5B, which was also trained with a similar amount of compute around the same time INTELLECT-1 was released.</p><p>My conclusion is that there exists plenty of room to experiment with bandwidth reduction techniques, such that much larger training runs are technically feasible over the internet. These techniques compromise somewhat on quality compared to centralized training, and many are not yet tested at scale nor in combination with each other, so not all will work. But there are enough avenues that I am optimistic bandwidth won&#8217;t limit the scale of decentralized training anytime soon.</p><h2>Can decentralized developers amass the necessary compute?</h2><p>While technically feasible, reaching the frontier of compute requires an astounding amount of resources.</p><p>The largest decentralized pretraining runs to date are INTELLECT-1, Protocol Model 8B, and Consilience 40B. These span the 6e22-6e23 FLOP range &#8212; 1000x less compute than what we estimated was used to train the largest models today, such as <a href="https://epoch.ai/trends">Grok 4</a>.</p><p>In order to train them, decentralized networks have been set up, such as <a href="https://app.primeintellect.ai/intelligence/">Prime Intellect&#8217;s platform</a>, the <a href="https://psyche.network/">Psyche Network by Nous Research</a>, and the <a href="https://dashboard.pluralis.ai/">Pluralis dashboard</a>. The largest such active network we&#8217;ve found is Covenant AI&#8217;s Templar, which is currently achieving an effective throughput of 9e17 FLOP/s respectively. This is about 300x smaller than frontier AI datacenters today, which have a theoretical training throughput of about 3e20 effective FLOP/s (eg the Microsoft Fairwater Atlanta site, assuming 30% MFU and 8-bit training).</p><p>While they are relatively small for the moment, it is worth appreciating that the scale of decentralized training runs has grown astoundingly fast. Since 2020, we have seen a 600,000x increase in the computational scale of decentralized training projects, for an implied growth rate of about 20x/year.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!sMDM!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48476ea1-d58f-4d9a-91c6-d440a8328621_1024x1280.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!sMDM!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48476ea1-d58f-4d9a-91c6-d440a8328621_1024x1280.png 424w, https://substackcdn.com/image/fetch/$s_!sMDM!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48476ea1-d58f-4d9a-91c6-d440a8328621_1024x1280.png 848w, https://substackcdn.com/image/fetch/$s_!sMDM!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48476ea1-d58f-4d9a-91c6-d440a8328621_1024x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!sMDM!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48476ea1-d58f-4d9a-91c6-d440a8328621_1024x1280.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!sMDM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48476ea1-d58f-4d9a-91c6-d440a8328621_1024x1280.png" width="1024" height="1280" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/48476ea1-d58f-4d9a-91c6-d440a8328621_1024x1280.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1280,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:101939,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://epochai.substack.com/i/182853461?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48476ea1-d58f-4d9a-91c6-d440a8328621_1024x1280.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!sMDM!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48476ea1-d58f-4d9a-91c6-d440a8328621_1024x1280.png 424w, https://substackcdn.com/image/fetch/$s_!sMDM!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48476ea1-d58f-4d9a-91c6-d440a8328621_1024x1280.png 848w, https://substackcdn.com/image/fetch/$s_!sMDM!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48476ea1-d58f-4d9a-91c6-d440a8328621_1024x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!sMDM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48476ea1-d58f-4d9a-91c6-d440a8328621_1024x1280.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p>This is a rate of growth that even dwarfs the rate of growth of frontier AI training, <a href="https://epoch.ai/trends#compute">recently growing as 5x/year</a>. If both trends held, it would take five and a half years for decentralized training to catch up to the scale of centralized training.</p><p>But can this growth be maintained? We look into three reference classes that inform the largest scale decentralized training could reach: the largest volunteer computing project to date, the scale of the largest cryptocurrency networks, and an estimate of the amount of spare compute capacity today.</p><p>Starting with the last one: today, there are about ~<a href="https://epoch.ai/data/ai-chip-sales">15.7M H100-equivalents of AI compute across NVIDIA, TPU, and Trainium devices</a>. In comparison, the largest datacenters today, such as <a href="https://epoch.ai/data/data-centers/satellite-explorer/AnthropicAmazonProjectRainierNewCarlisleIndiana/2025-09-26">Anthropic&#8217;s New Carlisle</a> site, host about 300k H100-equivalents, ie less than 2% of the AI compute stock total. A significant fraction of total compute could be idle for long periods, making it available for decentralized internet-based training.</p><p>So in theory, there exists a massive amount of AI compute which could be leveraged for decentralized AI training at a larger scale than has been achieved to date.</p><p>However, leveraging that amount of compute is likely unfeasible. Decentralized training project will only be able to access a fraction of existing computational resources. Determining exactly how much is complicated.</p><p>One possibility is by comparison to the largest decentralized computing projects. The largest collaborative computational project to date is <a href="https://foldingathome.org/">folding@home</a>, a volunteer computing network for simulating protein dynamics relevant to drug discovery. At its peak, folding@home attained a <a href="https://www.techspot.com/news/84832-foldinghome-project-passes-24-exaflops-more-than-top.html?utm_source=chatgpt.com">peak throughput of 2.43e18 FLOP/s</a> &#8211; more than the Top 500 supercomputers combined at the time. At 2.43e18 FLOP/s, a decentralized network would be able to conduct a 2.43e18 FLOP/s x 100 days &#8776; 2e25 FLOP training run. Not enough to reach the frontier today, but similar in scale to the previous generation of frontier AI models, including eg Llama 3, GPT-4, Gemini 1.0 Ultra and Claude 3 Opus.</p><p>Alternatively, we could compare to Bitcoin, whose decentralized hashing network encompasses $30&#8239;billion of existing infrastructure. For AI, that would be about enough for a gigawatt scale site, which could be used to train models at the 3e27 FLOP, still enough for a frontier scale run today.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!VJXd!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbb467ff-fcff-40b9-9be2-6e8d99e95811_1031x1280.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!VJXd!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbb467ff-fcff-40b9-9be2-6e8d99e95811_1031x1280.png 424w, https://substackcdn.com/image/fetch/$s_!VJXd!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbb467ff-fcff-40b9-9be2-6e8d99e95811_1031x1280.png 848w, https://substackcdn.com/image/fetch/$s_!VJXd!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbb467ff-fcff-40b9-9be2-6e8d99e95811_1031x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!VJXd!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbb467ff-fcff-40b9-9be2-6e8d99e95811_1031x1280.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!VJXd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbb467ff-fcff-40b9-9be2-6e8d99e95811_1031x1280.png" width="1031" height="1280" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/dbb467ff-fcff-40b9-9be2-6e8d99e95811_1031x1280.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1280,&quot;width&quot;:1031,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:97770,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://epochai.substack.com/i/182853461?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbb467ff-fcff-40b9-9be2-6e8d99e95811_1031x1280.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!VJXd!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbb467ff-fcff-40b9-9be2-6e8d99e95811_1031x1280.png 424w, https://substackcdn.com/image/fetch/$s_!VJXd!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbb467ff-fcff-40b9-9be2-6e8d99e95811_1031x1280.png 848w, https://substackcdn.com/image/fetch/$s_!VJXd!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbb467ff-fcff-40b9-9be2-6e8d99e95811_1031x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!VJXd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdbb467ff-fcff-40b9-9be2-6e8d99e95811_1031x1280.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p>The folding@home and Bitcoin reference classes suggest that today&#8217;s largest decentralized training networks (such as Prime Intellect&#8217;s) could be expanded 30-3,000x in scale, enough to train models on 50-5,000x more compute than today, in combination with longer training durations. This suggests the current fast rate of growth of decentralized training runs could last 3 to 6 years.</p><h2>Conclusion</h2><p>Decentralized training over the internet has captured the imagination of many developers, due to its potential to leverage large amounts of idle compute worldwide.</p><p>Technical feasibility is an important concern. But while bandwidth imposes an important technical challenge, enough techniques are being experimented with that I am optimistic about the prospects of much larger models being trained.</p><p>Frontier companies are unlikely to pursue internet decentralized training, since they can afford to build more efficient data centers. So internet decentralized training is likely to remain the domain of smaller companies focused on decentralized networks.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a></p><p>Hence, my uncertainty is about the compute such decentralized networks will be able to muster. They have grown remarkably quickly so far, though they are still far from the frontier. Looking at past volunteer computing projects such as folding@home, and the scale of decentralized computing networks such as Bitcoin, I see room for decentralized training networks to grow 30 to 3,000x in scale in coming years. If unimpeded, at the current rate of growth, we won&#8217;t see decentralized training runs catch up to the frontier of training in scale this decade. Even if they did, the performance loss from the bandwidth reduction techniques would set them back compared to centralized training.</p><p>But decentralized training could still be a very important part of AI. To the extent that decentralized networks remain associated with open weights, they could lead to larger open models to exist trailing the frontier. And while it&#8217;s unlikely that the small decentralized training projects will amass large frontier amounts of compute, thanks to <a href="https://epoch.ai/blog/algorithmic-progress-in-language-models">compute efficiency advances</a> and the increased <a href="https://epoch.ai/data/machine-learning-hardware">efficiency of hardware</a>, I expect decentralized training runs to not trail that far behind the frontier.</p><p>One practical implication is that decentralized training projects put a limit on the scale of models that can be affected by regulation, at least insofar as enforcing such regulations relies on training happening in large data centers.</p><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://epochai.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading! Subscribe to receive the latest from Epoch AI.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p><em>Thank you to Arthur Douillard, Max Ryabinin, Eugene Belilovsky, Sami Jaghouar, Fares Obeid, Josh You, and Aaron Scher for comments and suggestions. Jaeho Lee and Venkat Somala collected data for the charts. Lynette Bye edited the piece.</em></p><p><em>As part of the research that went into this piece we investigated over 100 papers related to decentralized training. You can find an annotated database of these papers <a href="https://bit.ly/4pe5DJm">here</a>.</em></p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>Managing a heterogeneous, inconsistent network requires important engineering resources, but through clever engineering it is ultimately feasible (see for example the work on <a href="https://arxiv.org/abs/2301.11913">SWARM</a> or <a href="https://arxiv.org/abs/2206.01288">Tasklets</a>).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>DeepSeek v3 was trained on 14.8T tokens using a batch size of 63M, for a total of 14.8T / 63M &#8776; 235k updates. At 32-bit precision, each update takes 2 x 671B parameters x 32 bits per parameter / 60 Mbps &#8776; 8 days, and the whole training would take 235k updates x 8 days / update &#8776; 5,000 years.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>60 Mpbs * 10 minutes / 2 uplink and downlink / 32 bits per parameter &#8776; 600M parameters.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p> With a proper streaming schedule, we could train the same model with 38 minutes / 7 minutes &#8776; 5x less bandwidth while still completely hiding the communication time behind computation. Naively, this suggests we could train a model with 5x more parameters. However, a larger model would be trained with a larger cluster, shortening the computation time per outer step. The communication time grows linearly with model size. The computation time grows with the batch size and the model size, and decreases with the computational power. The batch size grows with the <a href="https://arxiv.org/abs/2505.13738">0.3 power of the dataset size</a>, which itself grows proportionally to the model size under Chinchilla scaling. And the computational power grows proportionally to the square of model size under Chinchilla scaling. So if we increase the model size by a factor n, the computation time decreases by a factor </p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot; \n\\frac{n \\,(\\text{model size}) \\cdot n^{0.3} \\,(\\text{batch size})}{n^{2} \\,(\\text{cluster throughput})}\n= \\frac{n^{1.3}}{n^{2}}\n= \\frac{1}{n^{0.7}}\n&quot;,&quot;id&quot;:&quot;FUJYWHPVBL&quot;}" data-component-name="LatexBlockToDOM"></div><p>and the synchronization time grows by a factor <em>n</em>. The synchronization time can be almost fully masked while its below the computation time, so we can grow the model at the same bandwidth until it is</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;38\\ \\text{minutes} \\times \\frac{1}{n^{0.7}}\n= 7\\ \\text{minutes} \\times n&quot;,&quot;id&quot;:&quot;MLEQLTUWEG&quot;}" data-component-name="LatexBlockToDOM"></div><p>Solving for <em>n</em>, this allows for training a model</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;n = \\left(\\frac{35\\ \\text{minutes}}{7\\ \\text{minutes}}\\right)^{\\frac{1}{1.7}} \\approx 2.7\n\\;  times \\; larger.&quot;,&quot;id&quot;:&quot;MIYGOWCCNW&quot;}" data-component-name="LatexBlockToDOM"></div></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p><a href="https://www.sec.gov/Archives/edgar/data/1878848/000187884824000043/nov2024monthlyupdatevdc.htm">IREN (Iris Energy) &#8212; &#8220;November 2024 Monthly Investor Update&#8221;</a> filed on the SEC&#8217;s EDGAR system says: &#8220;1 EH/s = $30m cost to deliver&#8221; (including mining hardware + infrastructure capex) and lists the assumptions (fleet efficiency 15 J/TH, $18.9/TH ASIC pricing, and $750k/MW infrastructure capex). <a href="https://ycharts.com/indicators/bitcoin_network_hash_rate">YCharts</a> shows 942.95 EH/s. Combining both numbers, Bitcoin infra is valued at about $30bn.</p></div></div>]]></content:encoded></item><item><title><![CDATA[Why benchmarking is hard]]></title><description><![CDATA[Running benchmarks involves many moving parts, each of which can influence the final score. The two most impactful components are scaffolds and API providers.]]></description><link>https://epochai.substack.com/p/why-benchmarking-is-hard</link><guid isPermaLink="false">https://epochai.substack.com/p/why-benchmarking-is-hard</guid><dc:creator><![CDATA[Florian Brand]]></dc:creator><pubDate>Tue, 23 Dec 2025 22:04:20 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!2hBX!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8531e764-b5d9-44ba-9d88-71c9672e1100_1024x1280.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>This post is part of our <a href="https://epochai.substack.com/s/gradient-updates">Gradient Updates</a> newsletter, which shares more opinionated or informal takes about big questions in AI progress. These posts solely represent the views of the authors, and do not necessarily reflect the views of Epoch AI as a whole.</em></p><div><hr></div><p>Benchmarks play a crucial role in the AI landscape: They inform everyone, from AI researchers to the general public, about the current state of capabilities and the overall rate of progress. Third-party organizations, such as Epoch AI, independently run and collate benchmark results on a page like <a href="https://epoch.ai/benchmarks">the benchmarking hub</a>.</p><p>However, benchmarking isn&#8217;t easy: at each stage of the benchmarking pipeline, there are many moving parts and degrees of freedom that can affect the final result: this makes it hard to compare any two evaluation scores. Moreover, each stage can introduce bugs or mistakes that make the results costly to obtain or invalid.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!2hBX!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8531e764-b5d9-44ba-9d88-71c9672e1100_1024x1280.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!2hBX!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8531e764-b5d9-44ba-9d88-71c9672e1100_1024x1280.png 424w, https://substackcdn.com/image/fetch/$s_!2hBX!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8531e764-b5d9-44ba-9d88-71c9672e1100_1024x1280.png 848w, https://substackcdn.com/image/fetch/$s_!2hBX!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8531e764-b5d9-44ba-9d88-71c9672e1100_1024x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!2hBX!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8531e764-b5d9-44ba-9d88-71c9672e1100_1024x1280.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!2hBX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8531e764-b5d9-44ba-9d88-71c9672e1100_1024x1280.png" width="1024" height="1280" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8531e764-b5d9-44ba-9d88-71c9672e1100_1024x1280.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1280,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!2hBX!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8531e764-b5d9-44ba-9d88-71c9672e1100_1024x1280.png 424w, https://substackcdn.com/image/fetch/$s_!2hBX!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8531e764-b5d9-44ba-9d88-71c9672e1100_1024x1280.png 848w, https://substackcdn.com/image/fetch/$s_!2hBX!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8531e764-b5d9-44ba-9d88-71c9672e1100_1024x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!2hBX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8531e764-b5d9-44ba-9d88-71c9672e1100_1024x1280.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In this post, we dive into the different steps of the benchmarking process, which we split into two main parts:</p><ul><li><p>Benchmark Setup: all the steps related to how the benchmark is run; for example, the prompt that describes the task to the model and instructs it to answer each question, or the methodology used to score each sample.</p></li><li><p>Model Access: all the settings involved in accessing the evaluated model itself; for example, which API provider to call the model from.</p></li></ul><h3>Main takeaways</h3><p>Differences in how the benchmark is set up and how the model is accessed make it hard to compare any two scores on the same benchmark. For the benchmark setup, scaffolds have a huge impact on agentic benchmarks, especially on weaker models. To access the models, API providers are commonly used. Bugs and instabilities in the chosen API provider are the biggest source of evaluation errors, which particularly affects newer models.</p><h2>The Benchmark Setup</h2><p>Even though everyone uses the same name for a benchmark, this does not mean that everyone runs the same version (<a href="https://epoch.ai/blog/what-does-osworld-tell-us-about-ais-ability-to-use-computers#osworld-task-instructions-are-continually-updated-making-through-time-comparisons-of-uncertain-value">if versioning exists</a>), nor that they implement the benchmark in the same way. To illustrate this, we use the well-established <a href="https://github.com/idavidrein/gpqa">GPQA-Diamond benchmark</a> as an example throughout this blog post.</p><h3>Prompts &amp; Sampling Parameters</h3><p>GPQA-Diamond&#8217;s initial release was accompanied by an <a href="https://github.com/idavidrein/gpqa">analysis repository</a> containing code to evaluate models on the benchmark. However, this repo is not used to run the benchmark these days. Instead, practitioners typically re-implement GPQA-Diamond&#8217;s evaluation as part of a larger standardized infrastructure to run multiple models on multiple benchmarks.</p><p>Compared to many other benchmarks, GPQA-Diamond is relatively simple to (re-)implement, with few moving parts. Indeed, it just consists of the following:</p><ul><li><p>A template which instructs the model to answer the question</p></li><li><p>The question and four possible answers to said question</p></li></ul><p>As the question and answers are fixed (and re-used through all implementations), the component that can be different is the prompt template. Evaluators must also decide on the sampling parameters that models will be run with, in particular the temperature and potential <code>top_p</code> or <code>top_k</code> parameters. Let&#8217;s look at the prompt templates and sampling parameters used by some popular benchmarking libraries:</p><ul><li><p><a href="https://github.com/EleutherAI/lm-evaluation-harness/blob/01e412089bdb7590cb75d43310ac0b67c72938fa/lm_eval/tasks/gpqa/cot_zeroshot/_gpqa_cot_zeroshot_yaml#L9">EleutherAI/lm-evaluation-harness</a>: <code>What is the correct answer to this question:{{Question}}\nChoices:\n(A) {{choice1}}\n(B) {{choice2}}\n(C) {{choice3}}\n(D) {{choice4}}\nLet&#8217;s think step by step:</code></p><ul><li><p>No system message</p></li><li><p>Default temperature is 0.0 for API-based models</p></li><li><p><em>Note:</em> There are different <a href="https://github.com/EleutherAI/lm-evaluation-harness/blob/01e412089bdb7590cb75d43310ac0b67c72938fa/lm_eval/tasks/gpqa/zeroshot/_gpqa_zeroshot_yaml#L9C15-L9C176">templates and setups</a> depending on the evaluation and model type.</p></li></ul></li><li><p><a href="https://github.com/openai/simple-evals">OpenAI/simple-evals</a>: <code>Answer the following multiple choice question. The last line of your response should be of the following format: &#8216;Answer: $LETTER&#8217; (without quotes) where LETTER is one of ABCD. Think step by step before answering.\n\n{Question}\n\nA) {A}\nB) {B}\nC) {C}\nD) {D}</code></p><ul><li><p>System message set depending on the model: <code>You are a helpful assistant.</code> for GPT-3.5 until GPT-4.1; <a href="https://github.com/openai/simple-evals/blob/ee3b0318d8d1d9d72755a4120879be65f7c07e9e/sampler/chat_completion_sampler.py#L10">this</a> message for ChatGPT-based models and <a href="https://github.com/openai/simple-evals/blob/ee3b0318d8d1d9d72755a4120879be65f7c07e9e/sampler/claude_sampler.py#L9">this</a> lengthy system prompt for Claude 3 Opus and Claude 3.7 Sonnet. No system message for reasoning models.</p></li><li><p>Default temperature is 0.5</p></li></ul></li><li><p><a href="https://github.com/openai/gpt-oss/blob/48db88d8e29f48493fe75f084a8c9bd900a2b92f/gpt_oss/evals/gpqa_eval.py#L16">OpenAI/gpt-oss</a>: <code>{Question}\n\n(A) {A}\n(B) {B}\n(C) {C}\n(D) {D}\n\nExpress your final answer as the corresponding option &#8216;A&#8217;, &#8216;B&#8217;, &#8216;C&#8217;, or &#8216;D&#8217;.</code></p><ul><li><p>No system message set</p></li><li><p>Default temperature is 1.0 (when invoking the script via command line)</p></li></ul></li><li><p><a href="https://github.com/groq/openbench/blob/f8e643bc3593e3e47e8fade3121b49e00f222bc5/src/openbench/utils/text.py#L136">groq/openbench</a>: <code>Answer the following multiple choice question. The last line of your response should be of the following format: &#8216;Answer: $LETTER&#8217; (without quotes) where LETTER is one of ABCD.\n\n{Question}\n\nA) {A}\nB) {B}\nC) {C}\nD) {D}</code></p><ul><li><p><a href="https://github.com/groq/openbench/blob/f8e643bc3593e3e47e8fade3121b49e00f222bc5/src/openbench/utils/text.py#L122">System message</a>: <code>You are a helpful assistant.</code></p></li><li><p>Default temperature is 0.5</p></li></ul></li></ul><p>So, basically everyone is doing their own thing with implementations of even simple evaluations. Fortunately, <a href="https://epoch.ai/data-insights/self-reported-gpqa">the scores reported by AI developers on GPQA-Diamond match the results of independent runs</a>. We also tested the effect of using different prompts and temperatures with gpt-oss on GPQA-Diamond (with high reasoning effort): while we found differences between the average scores across settings (ranging from 74% to 80%), these differences were not statistically significant given the small size of GPQA-Diamond (only 198 questions). Even with the same implementation, we observe variance in this range (<em>cf.</em> GPT-OSS evaluation in the Appendix).</p><p>While the prompt typically has a small impact for modern reasoning models on simple benchmarks like GPQA-Diamond, this <a href="https://arxiv.org/pdf/2311.12022#page=10">was not always the case</a>. Moreover, as we will now see, changing the prompt can have a large impact when it comes to more complex, agentic evaluations.</p><h3>Scaffolds continue to have an outsized impact</h3><p>As agentic evals, such as <a href="https://epoch.ai/blog/what-skills-does-swe-bench-verified-evaluate">SWE-bench Verified</a> or <a href="https://www.remotelabor.ai/">RLI</a>, become more common, one component becomes increasingly important: The scaffold, i.e., the software that operates the agent, usually a CLI such as <a href="https://claude.com/product/claude-code">Claude Code</a>, <a href="https://github.com/OpenHands/OpenHands">OpenHands</a>, etc.</p><p>In particular, the scaffold includes all the components mentioned above, i.e., sampling parameters and prompt templates. Even more importantly, it also gives the agent a set of tools, i.e., access to specialized applications or capabilities. The exact tools and prompts change drastically depending on the chosen scaffold.</p><p>On SWE-bench Verified, a popular agentic coding benchmark, simply switching the scaffold makes up to an 11% difference for GPT-5 and up to a 15% difference for Kimi K2 Thinking. We cover the effect of the scaffold <a href="https://epoch.ai/blog/what-skills-does-swe-bench-verified-evaluate#scaffolds-matter-as-much-as-models">in our SWE-bench Verified review</a>. The choice of scaffold has the single biggest impact on the overall performance.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!uMgf!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13f81da4-59a7-4d8a-a1bb-83a251063732_1200x1500.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!uMgf!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13f81da4-59a7-4d8a-a1bb-83a251063732_1200x1500.png 424w, https://substackcdn.com/image/fetch/$s_!uMgf!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13f81da4-59a7-4d8a-a1bb-83a251063732_1200x1500.png 848w, https://substackcdn.com/image/fetch/$s_!uMgf!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13f81da4-59a7-4d8a-a1bb-83a251063732_1200x1500.png 1272w, https://substackcdn.com/image/fetch/$s_!uMgf!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13f81da4-59a7-4d8a-a1bb-83a251063732_1200x1500.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!uMgf!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13f81da4-59a7-4d8a-a1bb-83a251063732_1200x1500.png" width="1200" height="1500" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/13f81da4-59a7-4d8a-a1bb-83a251063732_1200x1500.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1500,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!uMgf!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13f81da4-59a7-4d8a-a1bb-83a251063732_1200x1500.png 424w, https://substackcdn.com/image/fetch/$s_!uMgf!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13f81da4-59a7-4d8a-a1bb-83a251063732_1200x1500.png 848w, https://substackcdn.com/image/fetch/$s_!uMgf!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13f81da4-59a7-4d8a-a1bb-83a251063732_1200x1500.png 1272w, https://substackcdn.com/image/fetch/$s_!uMgf!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F13f81da4-59a7-4d8a-a1bb-83a251063732_1200x1500.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Customizing the harness for each model risks hill-climbing on the evaluation and makes direct comparisons between models difficult. Therefore, there are two ways to evaluate agentic benchmarks: If you care about comparing models, a standardized scaffold (like mini-SWE-agent) is usually enough. On the other hand, <a href="https://x.com/sayashk/status/1996334941832089732">assessing frontier capabilities</a> requires the usage of leading products like Claude Code. Both these choices come with very obvious trade-offs, especially in terms of reproducibility, and require additional engineering efforts.</p><h3>Execution Environment</h3><p>The LLM, especially in agentic evaluations, has to operate in an execution context or environment. Concretely, this means things like the virtual machine for computer use evals, or the Docker container for coding evals. Creating and maintaining these environments <a href="https://epoch.ai/blog/swebench-docker">is hard</a>: OpenAI only ran 477 of the 500 problems in SWE-bench Verified in their<a href="https://openai.com/index/o3-o4-mini-system-card/"> o3 and o4-mini</a> evaluations due to infrastructure challenges.</p><p>Sometimes the environments <a href="https://github.com/SWE-bench/SWE-bench/issues/465">contain critical errors</a>, which means that agents are able to hack the eval (or are unable to fulfill a task due to bugs). This is especially true for evaluations which allow the agent to search the web for information: In the worst case, the agent is able to find the original dataset or sites which re-host parts of a problem. As the open web is so vast, it means that a banlist of domains needs to be continuously maintained.</p><p>The impact of the environment is moderate compared to the scaffold. Typically, the environment has little impact on the results, unless the agent is able to hack a large percentage of samples or if the environment is actually broken.</p><h3>Scoring</h3><p>The final step of every evaluation is to score the given sample. To extract and score the answer, implementations of GPQA-Diamond typically use regular expression parsing, which can be tweaked to catch all possible edge cases and model answers. Coding benchmarks such as SWE-bench Verified get scored by using a test suite, which runs the model-generated code.</p><p>Other evaluations, such as SimpleQA, use a second LLM to extract the answer and grade its correctness; while evals like tau-bench use a second LLM to simulate a user that the evaluated model interacts with. The choice of this second LLM can have a <a href="https://arxiv.org/abs/2511.21140">sizable impact</a> on the benchmark score.</p><h2>Model Access</h2><p>While the choices made when implementing the benchmark are within the control of the evaluator, they are not the only source of variance, as the models themselves have to be accessed in some way.</p><h3>API &amp; SDK</h3><p>The prompt first gets sent through the SDK and API endpoints that are called during the evaluation run. Using the standard OpenAI ChatCompletions makes it easy to use a lot of different models, as all services and providers offer an OpenAI-compatible endpoint.</p><p>However, this might leave performance on the table &#8211; OpenAI reports up to 3% improvements on SWE-bench Verified <a href="https://cookbook.openai.com/examples/responses_api/reasoning_items#:~:text=On%20a%20more%20rigorous%20benchmark%20like%20SWE%2Dbench%2C%20including%20reasoning%20items%20led%20to%20about%20a%203%25%20improvement%20for%20the%20same%20prompt%20and%20setup.">by using their Responses API</a>, while other providers, such as <a href="https://platform.claude.com/docs/en/api/openai-sdk#important-openai-compatibility-limitations">Anthropic</a>, only offer a subset of features in their OpenAI ChatCompletions compatible endpoint. <a href="https://x.com/MiniMax__AI/status/1985375617622454566">Minimax</a> reports an astronomical 23 percentage point difference in performance on tau-bench when using their API implementation compared to the standard ChatCompletions API. Therefore, to avoid undereliciting models&#8217; capabilities, it&#8217;s important to use the correct SDK.</p><h3>API Aggregator</h3><p>Of course, implementing all providers yourself is a tedious task, and there are solutions for this, such as <a href="https://github.com/BerriAI/litellm">LiteLLM</a> or <a href="https://inspect.aisi.org.uk/providers.html">Inspect AI</a>, which offer a unified interface to interact with any model from any provider. Aside from this, services such as <a href="https://huggingface.co/docs/inference-providers/en/index">HuggingFace Inference Providers</a> or <a href="https://openrouter.ai/">OpenRouter</a> offer a unified interface to various models and their providers. Naturally, they have to build a layer on top of the APIs, <a href="https://x.com/xeophon_/status/1994296820713894026">which may result in new bugs</a><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a>.</p><h3>Model Provider</h3><p>The biggest factor of variance in evaluation results, however, is the provider of the model itself, especially for open models. The choice of provider impacts both which sampling parameters are supported, as well as the <a href="https://x.com/andonlabs/status/1989862276137119799">downstream performance</a>.</p><p>To test the effect of the API provider ourselves, we ran several popular open models on GPQA-Diamond and averaged their scores over 4 runs<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a>. We retry each sample up to three times in case of errors and score (API) errors as failed samples<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a>.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!lU9O!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbaa54886-715a-4722-a82d-babdeb429da2_1024x1280.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!lU9O!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbaa54886-715a-4722-a82d-babdeb429da2_1024x1280.png 424w, https://substackcdn.com/image/fetch/$s_!lU9O!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbaa54886-715a-4722-a82d-babdeb429da2_1024x1280.png 848w, https://substackcdn.com/image/fetch/$s_!lU9O!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbaa54886-715a-4722-a82d-babdeb429da2_1024x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!lU9O!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbaa54886-715a-4722-a82d-babdeb429da2_1024x1280.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!lU9O!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbaa54886-715a-4722-a82d-babdeb429da2_1024x1280.png" width="1024" height="1280" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/baa54886-715a-4722-a82d-babdeb429da2_1024x1280.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1280,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!lU9O!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbaa54886-715a-4722-a82d-babdeb429da2_1024x1280.png 424w, https://substackcdn.com/image/fetch/$s_!lU9O!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbaa54886-715a-4722-a82d-babdeb429da2_1024x1280.png 848w, https://substackcdn.com/image/fetch/$s_!lU9O!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbaa54886-715a-4722-a82d-babdeb429da2_1024x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!lU9O!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbaa54886-715a-4722-a82d-babdeb429da2_1024x1280.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!EdOF!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b756026-9956-40f7-bea1-8a2d5dd95be4_1024x1280.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!EdOF!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b756026-9956-40f7-bea1-8a2d5dd95be4_1024x1280.png 424w, https://substackcdn.com/image/fetch/$s_!EdOF!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b756026-9956-40f7-bea1-8a2d5dd95be4_1024x1280.png 848w, https://substackcdn.com/image/fetch/$s_!EdOF!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b756026-9956-40f7-bea1-8a2d5dd95be4_1024x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!EdOF!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b756026-9956-40f7-bea1-8a2d5dd95be4_1024x1280.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!EdOF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b756026-9956-40f7-bea1-8a2d5dd95be4_1024x1280.png" width="1024" height="1280" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7b756026-9956-40f7-bea1-8a2d5dd95be4_1024x1280.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1280,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!EdOF!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b756026-9956-40f7-bea1-8a2d5dd95be4_1024x1280.png 424w, https://substackcdn.com/image/fetch/$s_!EdOF!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b756026-9956-40f7-bea1-8a2d5dd95be4_1024x1280.png 848w, https://substackcdn.com/image/fetch/$s_!EdOF!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b756026-9956-40f7-bea1-8a2d5dd95be4_1024x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!EdOF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b756026-9956-40f7-bea1-8a2d5dd95be4_1024x1280.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>We find differences in benchmark scores depending on the model provider for all tested models, though the effect size varies between models. More results can be found in the Appendix<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a>. We observed various sources of error:</p><ul><li><p>Some providers return RateLimitErrors, especially when accessing GLM-4.6 over OpenRouter. The providers mainly affected are Fireworks (via OpenRouter), GMICloud, Mancer, Parasail, DeepInfra, SiliconFlow. As these samples get scored as failure, the scores for the corresponding provider are affected substantially.</p></li><li><p>Some providers return empty or cut-off responses, although the <code>max_tokens</code> are not reached. This includes AtlasCloud, Mancer, Fireworks (via OpenRouter).</p></li><li><p>Some providers have lower <code>max_tokens</code> than advertised, resulting in cut-off responses, even though a higher limit was set via the API request. This affects SiliconFlow, Friendly and Cerebras.</p></li><li><p>Some providers have <code>max_tokens</code> limits which are lower than needed to evaluate the corresponding model. These providers were dropped completely for the given eval. An example of this is gpt-oss 120B, which needs more than 32K tokens to properly evaluate GPQA-Diamond; providers such as Novita, SiliconFlow or Cerebras have lower limits.</p></li><li><p>A few responses run into a timeout, which is set to 10 minutes by default.</p></li><li><p>Rarely, model responses run into &#8220;doom loops&#8221;, i.e., the model re-generates part of its response endlessly, until it reaches the <code>max_tokens</code> limit.</p></li><li><p>For GPT-OSS, one provider (Amazon Bedrock) does not get the correct <code>reasoning_effort</code> from OpenRouter, resulting in the same performance for all three levels of reasoning.</p></li></ul><p>Aside from these errors, we find that providers are noticeably worse at serving newer models, in our case GLM-4.6, compared to established models such as Qwen3. This is consistent with <a href="https://simonwillison.net/2025/Aug/15/inconsistent-performance/">other model releases</a>, which are accompanied <a href="https://github.com/MoonshotAI/K2-Vendor-Verifier">by bugs</a> that are then <a href="https://blog.vllm.ai/2025/10/28/Kimi-K2-Accuracy.html">fixed over time</a>. However, independent benchmark organizations want to evaluate new models <a href="https://x.com/CFGeek/status/1985766672406626348">as soon as possible</a>.</p><p>The selection of an appropriate provider has the biggest impact on model performance. The model developer usually hosts their own API, which could serve as a reference. However, this hosted API often offers worse data security guarantees than third party APIs, so evaluators might <a href="https://x.com/EpochAIResearch/status/2003178198146891876">prefer not to use it</a> for private benchmarks.</p><p>In practice, this means that those running the benchmark have to experiment a lot, trying to replicate known scores with third-party APIs<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a> and hope that it is implemented correctly. This is a laborious and costly effort and one of the main reasons evaluations (of open models) take a lot of time.</p><h3>Model Deployment</h3><p>Going under the hood, there are multiple reasons why the performance for a model differs from the reported score or between different model providers. They can range from software, such as a different inference engine (vLLM, SGLang, private forks), <a href="https://x.com/deepseek_ai/status/1990616182161068094?s=20">bugs in the reference implementation</a>, or <a href="https://adamkarvonen.github.io/machine_learning/2025/11/28/difr.html#:~:text=we%20found%20that%20at%20least%20two%20providers%20were%20using%20an%20older%20Llama%2D3.1%2D8B%20chat%20template%20that%20doesn%E2%80%99t%20match%20the%20latest%20version%20on%20HuggingFace">outdated chat templates</a>, down to the used hardware and the corresponding deployment. Yes, even using different parallelism setups, i.e., how you distribute and split up a model over multiple GPUs, has a <a href="https://youtube.com/watch?v=uaZ3yRdYg8A&amp;t=3096">measurable, but small, impact</a>.</p><h2>Conclusion</h2><p>Running benchmarks isn&#8217;t easy, and there are a lot of variables that affect the headline numbers you will see on a graph or in a table. A lot of these variables don&#8217;t seem to matter in isolation. However, they can add up quickly over the whole stack, which results in numbers that differ substantially from the scores reported by model developers.</p><p>Some of these variables, such as the used prompts and scaffolds, are in the control of those who create, re-implement and run the benchmark, while others aren&#8217;t. This is especially true for the model access by using (third-party) APIs, which slows down lower resource actors, such as hobbyists or academics. Errors in evaluations also make it hard for others, from AI researchers to decision makers and the general public, to accurately assess the current progress of capabilities.</p><p><em>OpenRouter graciously sponsored the credits for the experiments on their platform. We maintained full editorial control over the output.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://epochai.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading! Subscribe to receive the latest from Epoch AI.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2>Appendix</h2><p>We evaluated GPT-OSS 120B at all three reasoning levels, GLM-4.6, Qwen3 235B-A22B Instruct 2507, Kimi K2 Thinking and Kimi K2 0905 Instruct on GPQA-Diamond while varying the API provider and keeping everything else constant.</p><p>We evaluated GLM-4.6 (as both the agent and the user model) on tau-bench Retail while varying the API provider (keeping the same provider for both agent and user), while keeping everything else the same.</p><p>We use <a href="https://github.com/groq/openbench/pull/291">this</a> implementation for GPQA-Diamond and <a href="https://github.com/groq/openbench/pull/294">this</a> re-implementation of tau-bench. We also make all the data for the analysis available under <a href="https://huggingface.co/datasets/xeophon/epoch-benchmark-sweep">this link</a>.</p><h4>GPT-OSS 120B</h4><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!cde0!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc533cba0-c039-4701-b80d-0f67d9219b91_1024x1281.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!cde0!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc533cba0-c039-4701-b80d-0f67d9219b91_1024x1281.png 424w, https://substackcdn.com/image/fetch/$s_!cde0!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc533cba0-c039-4701-b80d-0f67d9219b91_1024x1281.png 848w, https://substackcdn.com/image/fetch/$s_!cde0!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc533cba0-c039-4701-b80d-0f67d9219b91_1024x1281.png 1272w, https://substackcdn.com/image/fetch/$s_!cde0!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc533cba0-c039-4701-b80d-0f67d9219b91_1024x1281.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!cde0!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc533cba0-c039-4701-b80d-0f67d9219b91_1024x1281.png" width="1024" height="1281" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c533cba0-c039-4701-b80d-0f67d9219b91_1024x1281.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1281,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!cde0!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc533cba0-c039-4701-b80d-0f67d9219b91_1024x1281.png 424w, https://substackcdn.com/image/fetch/$s_!cde0!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc533cba0-c039-4701-b80d-0f67d9219b91_1024x1281.png 848w, https://substackcdn.com/image/fetch/$s_!cde0!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc533cba0-c039-4701-b80d-0f67d9219b91_1024x1281.png 1272w, https://substackcdn.com/image/fetch/$s_!cde0!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc533cba0-c039-4701-b80d-0f67d9219b91_1024x1281.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h4>Kimi K2 Thinking</h4><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!db9-!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d7fe44e-958d-44eb-9b6f-20e588300ba3_1024x1280.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!db9-!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d7fe44e-958d-44eb-9b6f-20e588300ba3_1024x1280.png 424w, https://substackcdn.com/image/fetch/$s_!db9-!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d7fe44e-958d-44eb-9b6f-20e588300ba3_1024x1280.png 848w, https://substackcdn.com/image/fetch/$s_!db9-!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d7fe44e-958d-44eb-9b6f-20e588300ba3_1024x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!db9-!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d7fe44e-958d-44eb-9b6f-20e588300ba3_1024x1280.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!db9-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d7fe44e-958d-44eb-9b6f-20e588300ba3_1024x1280.png" width="1024" height="1280" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4d7fe44e-958d-44eb-9b6f-20e588300ba3_1024x1280.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1280,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!db9-!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d7fe44e-958d-44eb-9b6f-20e588300ba3_1024x1280.png 424w, https://substackcdn.com/image/fetch/$s_!db9-!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d7fe44e-958d-44eb-9b6f-20e588300ba3_1024x1280.png 848w, https://substackcdn.com/image/fetch/$s_!db9-!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d7fe44e-958d-44eb-9b6f-20e588300ba3_1024x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!db9-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d7fe44e-958d-44eb-9b6f-20e588300ba3_1024x1280.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h4>Qwen3 235B-A22B Instruct 2507</h4><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!6JH8!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3deb60fa-c150-4fb3-bdf0-477b8b9e58c9_1024x1280.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!6JH8!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3deb60fa-c150-4fb3-bdf0-477b8b9e58c9_1024x1280.png 424w, https://substackcdn.com/image/fetch/$s_!6JH8!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3deb60fa-c150-4fb3-bdf0-477b8b9e58c9_1024x1280.png 848w, https://substackcdn.com/image/fetch/$s_!6JH8!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3deb60fa-c150-4fb3-bdf0-477b8b9e58c9_1024x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!6JH8!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3deb60fa-c150-4fb3-bdf0-477b8b9e58c9_1024x1280.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!6JH8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3deb60fa-c150-4fb3-bdf0-477b8b9e58c9_1024x1280.png" width="1024" height="1280" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3deb60fa-c150-4fb3-bdf0-477b8b9e58c9_1024x1280.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1280,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!6JH8!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3deb60fa-c150-4fb3-bdf0-477b8b9e58c9_1024x1280.png 424w, https://substackcdn.com/image/fetch/$s_!6JH8!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3deb60fa-c150-4fb3-bdf0-477b8b9e58c9_1024x1280.png 848w, https://substackcdn.com/image/fetch/$s_!6JH8!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3deb60fa-c150-4fb3-bdf0-477b8b9e58c9_1024x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!6JH8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3deb60fa-c150-4fb3-bdf0-477b8b9e58c9_1024x1280.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h4>GLM-4.6</h4><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!sIrF!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F74964066-bac1-48a4-9eb6-50ff016af0fe_1024x1280.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!sIrF!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F74964066-bac1-48a4-9eb6-50ff016af0fe_1024x1280.png 424w, https://substackcdn.com/image/fetch/$s_!sIrF!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F74964066-bac1-48a4-9eb6-50ff016af0fe_1024x1280.png 848w, https://substackcdn.com/image/fetch/$s_!sIrF!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F74964066-bac1-48a4-9eb6-50ff016af0fe_1024x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!sIrF!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F74964066-bac1-48a4-9eb6-50ff016af0fe_1024x1280.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!sIrF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F74964066-bac1-48a4-9eb6-50ff016af0fe_1024x1280.png" width="1024" height="1280" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/74964066-bac1-48a4-9eb6-50ff016af0fe_1024x1280.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1280,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!sIrF!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F74964066-bac1-48a4-9eb6-50ff016af0fe_1024x1280.png 424w, https://substackcdn.com/image/fetch/$s_!sIrF!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F74964066-bac1-48a4-9eb6-50ff016af0fe_1024x1280.png 848w, https://substackcdn.com/image/fetch/$s_!sIrF!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F74964066-bac1-48a4-9eb6-50ff016af0fe_1024x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!sIrF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F74964066-bac1-48a4-9eb6-50ff016af0fe_1024x1280.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>This particular bug has since then been (mostly) fixed.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>Usually, benchmarks such as GPQA-Diamond are run between 8-32 times and then averaged. We opt for 4 runs for cost reasons. We use <a href="https://github.com/groq/openbench/pull/291">this</a> implementation for GPQA-Diamond and <a href="https://github.com/groq/openbench/pull/294">this</a> re-implementation of tau-bench.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>Of course, this is not the policy we use when evaluating models on the <a href="https://epoch.ai/benchmarks">Epoch AI Benchmarking Hub</a>, where except in <a href="https://epoch.ai/benchmarks/frontiermath#:~:text=28%2Dprivate).-,Gemini%20models,-We%20faced%20difficulties">pathological cases</a> we ensure that there are no more errors before reporting a score. However, we believe this convention makes sense for the purpose of evaluating providers and how they complicate the benchmarking process.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p> We also make all the data for the analysis available at <a href="https://huggingface.co/datasets/xeophon/epoch-benchmark-sweep">this link</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>While there are <a href="https://arxiv.org/abs/2511.20621">methods</a> to detect levels of quantization and possibly deployment bugs, they require a correctly set up local deployment, knowledge of the inference engine of the provider as well as exposed sampling parameters such as the seed (which not all providers offer).</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[The changing drivers of LLM adoption]]></title><description><![CDATA[Public data as well as our original polling suggest LLM adoption is roughly on trend, but the underlying drivers are shifting.]]></description><link>https://epochai.substack.com/p/the-changing-drivers-of-llm-adoption</link><guid isPermaLink="false">https://epochai.substack.com/p/the-changing-drivers-of-llm-adoption</guid><dc:creator><![CDATA[JSD]]></dc:creator><pubDate>Sat, 20 Dec 2025 00:16:22 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!HjMs!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa1416009-1664-49c7-a0f1-917724eafc58_2048x2560.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>This post is part of our <a href="https://epochai.substack.com/s/gradient-updates">Gradient Updates</a> newsletter, which shares more opinionated or informal takes about big questions in AI progress. These posts solely represent the views of the authors, and do not necessarily reflect the views of Epoch AI as a whole.</em></p><p><em>Originally posted on <a href="https://epoch.ai/gradient-updates/the-changing-drivers-of-llm-adoption">Epoch AI</a>.</em></p><div><hr></div><p>In the world of AI, half a year is a very long time. Back in July, we <a href="https://epoch.ai/gradient-updates/after-the-chatgpt-moment-measuring-ais-adoption">saw</a> LLMs being adopted faster than almost any other technology in history. Five months later we&#8217;re still seeing rapid growth, but we&#8217;re also seeing early winds of change &#8212; both in who uses AI and how they do so.</p><p>Using the latest public data,<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> and a <a href="https://epoch.ai/data/polling">poll</a> of US adults we conducted with Blue Rose Research, this post shares an updated picture of the state of LLM adoption.</p><h2>How quickly are consumers adopting LLMs?</h2><h3>More people are using LLMs &#8212; but they&#8217;re increasingly using different LLMs, different products, and in different places</h3><p>Through the first half of 2025, ChatGPT&#8217;s user base grew at a remarkable pace, from under 400 million weekly active users in January to nearly 800 million by August &#8212; roughly 50 million new users per month. Since then, growth may have slowed slightly, though it&#8217;s a bit soon to tell how much of this is noise versus a lasting trend change:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!HjMs!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa1416009-1664-49c7-a0f1-917724eafc58_2048x2560.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!HjMs!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa1416009-1664-49c7-a0f1-917724eafc58_2048x2560.png 424w, https://substackcdn.com/image/fetch/$s_!HjMs!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa1416009-1664-49c7-a0f1-917724eafc58_2048x2560.png 848w, https://substackcdn.com/image/fetch/$s_!HjMs!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa1416009-1664-49c7-a0f1-917724eafc58_2048x2560.png 1272w, https://substackcdn.com/image/fetch/$s_!HjMs!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa1416009-1664-49c7-a0f1-917724eafc58_2048x2560.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!HjMs!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa1416009-1664-49c7-a0f1-917724eafc58_2048x2560.png" width="1456" height="1820" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a1416009-1664-49c7-a0f1-917724eafc58_2048x2560.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1820,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:159825,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://epochai.substack.com/i/182110144?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa1416009-1664-49c7-a0f1-917724eafc58_2048x2560.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!HjMs!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa1416009-1664-49c7-a0f1-917724eafc58_2048x2560.png 424w, https://substackcdn.com/image/fetch/$s_!HjMs!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa1416009-1664-49c7-a0f1-917724eafc58_2048x2560.png 848w, https://substackcdn.com/image/fetch/$s_!HjMs!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa1416009-1664-49c7-a0f1-917724eafc58_2048x2560.png 1272w, https://substackcdn.com/image/fetch/$s_!HjMs!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa1416009-1664-49c7-a0f1-917724eafc58_2048x2560.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Does this mean that LLM adoption has slowed down? Not necessarily &#8212; even if ChatGPT user growth is <a href="https://techcrunch.com/2025/12/05/chatgpts-user-growth-has-slowed-report-finds">slowing</a>, other LLMs have recently seen faster consumer growth. For example, between August and November, Gemini monthly active users <a href="https://www.theinformation.com/articles/chatgpt-nears-900-million-weekly-active-users-gemini-catching">increased</a> about 30%, compared to 5% for ChatGPT. And that&#8217;s not because Gemini has very few users &#8212; both Gemini and ChatGPT have hundreds of millions of monthly active users. That said, ChatGPT still maintains the larger userbase: in <a href="https://epoch.ai/data/polling#ai-services">our poll</a>, about 35% of registered voters reported having used ChatGPT over the past week, compared to 24% for Gemini.</p><p>Increasingly, this consumer growth is coming from countries besides the US. According to OpenAI&#8217;s &#8220;How People Use ChatGPT&#8221; <a href="https://cdn.openai.com/pdf/a253471f-8260-40c6-a2cc-aa93fe9f142e/economic-research-chatgpt-usage-paper.pdf">report</a>, 30% of internet users in high-income countries were already using ChatGPT weekly by mid-2025<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a>. With less room to grow, we should expect slower user growth in the US compared to the global average<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a>. And there are hints that this is the case &#8212; for example, ChatGPT daily users in India grew <a href="https://www.reuters.com/world/india/with-freebies-openai-google-vie-indian-users-training-data-2025-12-17/">sevenfold</a> over the past year, and the number of daily Gemini users in India grew by 15% in November alone. As a result, India now has over twice the ChatGPT users of the US.</p><p>Finally, user growth could continue or even accelerate through a different route: integrating AI into widely-used products, rather than using standalone chatbot apps. One example is Google Search, which now includes an &#8220;AI mode&#8221; that&#8217;s <a href="https://blog.google/products/search/gemini-3-search-ai-mode/">powered by their state-of-the-art model</a>, Gemini 3, and was made available to the public in December. Another example is Meta, which has been pushing Meta AI across WhatsApp, Messenger, and Instagram &#8212; this integration likely explains where Meta AI&#8217;s <a href="https://www.cnbc.com/2025/05/28/zuckerberg-meta-ai-one-billion-monthly-users.html">alleged</a> 1 billion monthly active users comes from<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a>. And if Meta succeeds in making AI a habitual part of messaging, Meta AI could reach an enormous user base &#8212; a possibility <a href="https://www.justice.gov/atr/media/1397596/dl">not lost on OpenAI</a>.</p><p>Overall, the prior trend in global user growth seems more or less on track. Growth might be slower in saturated markets like the US, but this is offset by faster growth elsewhere.</p><h3>Consumers are using LLMs much more intensely on AI apps, while web traffic has stagnated</h3><p>Even without new LLM users, we could still see an increase in adoption. That&#8217;s because each user could still use LLMs more <em>intensely</em>, such as by sending more messages. Indeed, between June 2024 and 2025, the number of ChatGPT messages <a href="https://cdn.openai.com/pdf/a253471f-8260-40c6-a2cc-aa93fe9f142e/economic-research-chatgpt-usage-paper.pdf">grew faster</a> than the number of weekly active users.</p><p>So what&#8217;s happened more recently? The answer depends on what metric you look at, and also on the specific LLM.</p><p>For example, one metric we can check is recent web traffic. According to <a href="https://x.com/Similarweb/status/1995136322097529180">SimilarWeb</a>, ChatGPT&#8217;s web traffic has been essentially flat since September 2025. In contrast, Gemini&#8217;s traffic has grown somewhat over that same period, though it started from a smaller base, such that the combined web traffic across the two models has stagnated.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!UJ32!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7030715-4ab7-458f-9052-a2b26c4bde0c_2048x2560.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!UJ32!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7030715-4ab7-458f-9052-a2b26c4bde0c_2048x2560.png 424w, https://substackcdn.com/image/fetch/$s_!UJ32!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7030715-4ab7-458f-9052-a2b26c4bde0c_2048x2560.png 848w, https://substackcdn.com/image/fetch/$s_!UJ32!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7030715-4ab7-458f-9052-a2b26c4bde0c_2048x2560.png 1272w, https://substackcdn.com/image/fetch/$s_!UJ32!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7030715-4ab7-458f-9052-a2b26c4bde0c_2048x2560.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!UJ32!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7030715-4ab7-458f-9052-a2b26c4bde0c_2048x2560.png" width="1456" height="1820" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d7030715-4ab7-458f-9052-a2b26c4bde0c_2048x2560.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1820,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:139552,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://epochai.substack.com/i/182110144?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7030715-4ab7-458f-9052-a2b26c4bde0c_2048x2560.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!UJ32!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7030715-4ab7-458f-9052-a2b26c4bde0c_2048x2560.png 424w, https://substackcdn.com/image/fetch/$s_!UJ32!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7030715-4ab7-458f-9052-a2b26c4bde0c_2048x2560.png 848w, https://substackcdn.com/image/fetch/$s_!UJ32!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7030715-4ab7-458f-9052-a2b26c4bde0c_2048x2560.png 1272w, https://substackcdn.com/image/fetch/$s_!UJ32!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7030715-4ab7-458f-9052-a2b26c4bde0c_2048x2560.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>But the story looks quite different over a longer period. Between November 2024 and November 2025, ChatGPT&#8217;s web traffic <a href="https://x.com/Similarweb/status/1998677459382710656">grew</a> by around 1.5&#215;. That&#8217;s <a href="https://x.com/Similarweb/status/1998677459382710656">faster than most websites</a>, but surprisingly slow compared to the roughly 3&#215; growth in weekly active users over a similar period. This discrepancy might be due to a shift from web use to faster-growing chatbot apps &#8212; SimilarWeb captures traffic from the former, but not the latter. Notably, the ChatGPT app was <a href="https://x.com/Similarweb/status/1991130907382620610">downloaded</a> 1.9 billion times between October 2024 and September 2025 &#8212; the most downloaded app of the year.</p><p>We also see people using models more intensely within these apps. As of November, ChatGPT users <a href="https://techcrunch.com/2025/12/05/chatgpts-user-growth-has-slowed-report-finds">spent</a> 17 minutes per day in the app, up about 6% from March. Gemini users averaged about 11 minutes per day, up roughly 120% from March, but again starting from a smaller base.</p><p>But while looking at web traffic and time use is helpful, it misses out on a crucial way by which LLMs could be used more intensely &#8212; processing more tokens <em>even with the same number of user sessions. </em>For example, reasoning models tend to have <a href="https://epoch.ai/data-insights/output-length">longer responses</a>, and some changes in <a href="https://www.anthropic.com/research/claude-character">model</a> <a href="https://techcrunch.com/2025/09/05/openai-reorganizes-research-team-behind-chatgpts-personality/">personality</a> could lead to more messages per session. So ideally, we&#8217;d also look at the number of messages sent and the tokens processed per user.  Alas, we don&#8217;t know how much this impacts user on average, because <a href="https://epoch.ai/data/polling#reasoning-frequency">only a small fraction of users</a> regularly use the &#8220;reasoning&#8221; functionalities of frontier models (e.g. ChatGPT&#8217;s &#8220;thinking&#8221; mode). Moreover, we don&#8217;t know how many of the total tokens processed by ChatGPT are driven by a small fraction of ChatGPT &#8220;power users&#8221;.</p><p>Putting everything together, there&#8217;s still only limited evidence on how intensely people use LLMs and how this has changed over time. But if we squint at the data we have, it seems that individuals are likely using LLMs much more intensely than before.</p><h3>AI company revenues have continued to grow incredibly fast, in line with previous trends</h3><p>So far we&#8217;ve looked at two dimensions of increasing adoption: more users, and more intense use. Combining these two lets us measure total LLM use, which we can capture using revenue. This is a metric which has <a href="https://epoch.ai/gradient-updates/openai-is-projecting-unprecedented-revenue-growth">grown so fast</a> that it puts companies like Google and Uber to shame &#8212; so you&#8217;d have a really hard time arguing that LLM adoption has been slow.</p><p>To put things into perspective, OpenAI&#8217;s annualized revenue was $13 billion back in August, and <a href="https://epoch.ai/data/ai-companies">growing</a> at about 4.3&#215; a year. On that trend, we&#8217;d expect an annualized revenue of around $21 billion by the end of the year, which is essentially <a href="https://www.theinformation.com/articles/openais-organizational-problems-hurt-chatgpt">what we&#8217;ve seen</a><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a>. So under this lens, revenue (which roughly tracks adoption) has continued following extremely fast historical trends.</p><h2>How embedded is AI in daily tasks and jobs?</h2><p>Knowing about the rate of adoption already tells us a lot, but to understand how AI is impacting society, we need to look further. We want to understand how AI is actually being used, and by whom.</p><h3>AI has entered the workplace beyond formal enterprise adoption</h3><p>The first step is to move beyond individual consumers to the workplace. This is important because an increasing share of OpenAI&#8217;s revenue has come from enterprise products like <a href="https://chatgpt.com/business/enterprise">ChatGPT Enterprise</a> and <a href="https://chatgpt.com/business/">ChatGPT Business</a><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a>.</p><p>The obvious way to measure work-related AI adoption is to look at enterprise deployments. For example, the <a href="https://ramp.com/data/ai-index">Ramp AI Index</a> we <a href="https://epoch.ai/gradient-updates/after-the-chatgpt-moment-measuring-ais-adoption">cited</a> back in July measures the &#8220;share of U.S. businesses with paid subscriptions to AI models, platforms, and tools&#8221;.</p><p>But this misses a large share of actual work usage &#8212; many people are using AI for their jobs, even if their employers don&#8217;t provide them access. We can see this by looking at our survey data. On the one hand, 36% of respondents who used AI in the past week said they had used it to assist with work. On the other hand, only about 18% of those same respondents said that their job provides access to an AI chatbot, and another 18% weren&#8217;t sure!</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!N-yr!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F332c12b5-ddda-4e28-8584-911316e0decd_1280x800.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!N-yr!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F332c12b5-ddda-4e28-8584-911316e0decd_1280x800.png 424w, https://substackcdn.com/image/fetch/$s_!N-yr!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F332c12b5-ddda-4e28-8584-911316e0decd_1280x800.png 848w, https://substackcdn.com/image/fetch/$s_!N-yr!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F332c12b5-ddda-4e28-8584-911316e0decd_1280x800.png 1272w, https://substackcdn.com/image/fetch/$s_!N-yr!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F332c12b5-ddda-4e28-8584-911316e0decd_1280x800.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!N-yr!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F332c12b5-ddda-4e28-8584-911316e0decd_1280x800.png" width="1280" height="800" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/332c12b5-ddda-4e28-8584-911316e0decd_1280x800.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:800,&quot;width&quot;:1280,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:57301,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://epochai.substack.com/i/182110144?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F332c12b5-ddda-4e28-8584-911316e0decd_1280x800.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!N-yr!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F332c12b5-ddda-4e28-8584-911316e0decd_1280x800.png 424w, https://substackcdn.com/image/fetch/$s_!N-yr!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F332c12b5-ddda-4e28-8584-911316e0decd_1280x800.png 848w, https://substackcdn.com/image/fetch/$s_!N-yr!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F332c12b5-ddda-4e28-8584-911316e0decd_1280x800.png 1272w, https://substackcdn.com/image/fetch/$s_!N-yr!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F332c12b5-ddda-4e28-8584-911316e0decd_1280x800.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>This gap suggests that a substantial share of work-related AI use is happening on workers&#8217; own initiative, using free tiers or personal subscriptions rather than employer-provided tools. While the proportion of respondents who reported using AI for work does vary across job categories, the effect size is lower than we expected: less than 10 percentage points between &#8220;Office job&#8221; workers (34.6%) and workers in &#8220;Manual, non customer-facing&#8221; jobs (25.8%).</p><h3>Most consumers use AI to seek information</h3><p>On paper, modern frontier LLMs are optimized for a wide range of tasks, such as coding and writing. But there&#8217;s one use case that stands out by far &#8212; looking up information. About 58% of AI users <a href="https://epoch.ai/data/polling#usage-category">in our survey</a> reported using it for this purpose in the past week. Writing assistance (32%) and advice (30%) came next, followed by technical tasks (18%), multimedia (17%), and self-expression (16%).</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!lS_X!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7fdf06c0-e5c5-4575-8d2b-ae17dc06efd8_2048x2560.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!lS_X!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7fdf06c0-e5c5-4575-8d2b-ae17dc06efd8_2048x2560.png 424w, https://substackcdn.com/image/fetch/$s_!lS_X!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7fdf06c0-e5c5-4575-8d2b-ae17dc06efd8_2048x2560.png 848w, https://substackcdn.com/image/fetch/$s_!lS_X!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7fdf06c0-e5c5-4575-8d2b-ae17dc06efd8_2048x2560.png 1272w, https://substackcdn.com/image/fetch/$s_!lS_X!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7fdf06c0-e5c5-4575-8d2b-ae17dc06efd8_2048x2560.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!lS_X!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7fdf06c0-e5c5-4575-8d2b-ae17dc06efd8_2048x2560.png" width="1456" height="1820" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7fdf06c0-e5c5-4575-8d2b-ae17dc06efd8_2048x2560.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1820,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:136377,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://epochai.substack.com/i/182110144?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7fdf06c0-e5c5-4575-8d2b-ae17dc06efd8_2048x2560.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!lS_X!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7fdf06c0-e5c5-4575-8d2b-ae17dc06efd8_2048x2560.png 424w, https://substackcdn.com/image/fetch/$s_!lS_X!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7fdf06c0-e5c5-4575-8d2b-ae17dc06efd8_2048x2560.png 848w, https://substackcdn.com/image/fetch/$s_!lS_X!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7fdf06c0-e5c5-4575-8d2b-ae17dc06efd8_2048x2560.png 1272w, https://substackcdn.com/image/fetch/$s_!lS_X!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7fdf06c0-e5c5-4575-8d2b-ae17dc06efd8_2048x2560.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>This pattern aligns with OpenAI&#8217;s &#8220;<a href="https://cdn.openai.com/pdf/a253471f-8260-40c6-a2cc-aa93fe9f142e/economic-research-chatgpt-usage-paper.pdf">How People Use ChatGPT</a>&#8220; report, which found that &#8220;Practical Guidance,&#8221; &#8220;Seeking Information,&#8221; and &#8220;Writing&#8221; accounted for the largest share of ChatGPT messages<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a>. It also matches the results of another <a href="https://x.com/charlotteeffect/status/2001340374841987387">recent survey</a>, which found &#8220;General information or answering question&#8221; as the most common use case by far. For most users, these tools function as an enhanced search engine and writing aid, rather than as autonomous agents or virtual companions.</p><p>This raises a question: how much does progress on frontier capabilities actually matter for typical users? The benchmarks that AI labs optimize for &#8212; <a href="https://openai.com/index/introducing-swe-bench-verified/">SWE-bench Verified</a> (coding), <a href="https://epoch.ai/benchmarks/gpqa-diamond">GPQA Diamond</a> (advanced scientific reasoning), <a href="https://epoch.ai/frontiermath">FrontierMath</a> (hard math problems), <a href="https://metr.org/blog/2025-03-19-measuring-ai-ability-to-complete-long-tasks/">METR&#8217;s suite</a> and <a href="https://openai.com/index/gdpval/">GDPVal</a> (autonomous task completion) &#8212; seem quite distant from &#8220;seeking information (e.g. to look up a fact, as a replacement for Google, to get product recommendations),&#8221; which is how we described the category to poll respondents. The closest benchmark to this dominant use case might be something like OpenAI&#8217;s <a href="https://openai.com/index/browsecomp/">BrowseComp</a>, which tests web browsing and information retrieval, but usually gets less airtime.</p><p>In practice, <a href="https://epoch.ai/gradient-updates/benchmark-scores-general-capability-claudiness">different capabilities have been highly correlated</a>: models that score well on coding benchmarks also tend to score well on reasoning and information retrieval. So progress on frontier benchmarks may benefit typical users anyway, <em>if</em> the correlation in benchmark scores actually reflects a deep shared latent ability factor. On the other hand, <a href="https://epoch.ai/gradient-updates/benchmark-scores-general-capability-claudiness#:~:text=Is%20the%20%E2%80%9Cgeneral%20capability%E2%80%9D%20dimension%20deep%2C%20or%20contingent%3F">if benchmark scores are correlated for more contingent reasons</a>, there could be a real tradeoff between pushing frontier capabilities in specific domains and catering to the most common use cases.</p><h3>AI use is stratified by age, income and job type, and less so by gender</h3><p>Who is using AI? We found that higher-income respondents were substantially more likely to have used AI in the past week, and the pattern holds across services: ChatGPT, Gemini, and Copilot usage all skew toward upper-income groups.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!gSqP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a80d25c-3484-4c3b-9eb8-cc60c5d04a84_2048x2560.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!gSqP!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a80d25c-3484-4c3b-9eb8-cc60c5d04a84_2048x2560.png 424w, https://substackcdn.com/image/fetch/$s_!gSqP!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a80d25c-3484-4c3b-9eb8-cc60c5d04a84_2048x2560.png 848w, https://substackcdn.com/image/fetch/$s_!gSqP!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a80d25c-3484-4c3b-9eb8-cc60c5d04a84_2048x2560.png 1272w, https://substackcdn.com/image/fetch/$s_!gSqP!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a80d25c-3484-4c3b-9eb8-cc60c5d04a84_2048x2560.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!gSqP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a80d25c-3484-4c3b-9eb8-cc60c5d04a84_2048x2560.png" width="1456" height="1820" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9a80d25c-3484-4c3b-9eb8-cc60c5d04a84_2048x2560.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1820,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:155184,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://epochai.substack.com/i/182110144?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a80d25c-3484-4c3b-9eb8-cc60c5d04a84_2048x2560.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!gSqP!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a80d25c-3484-4c3b-9eb8-cc60c5d04a84_2048x2560.png 424w, https://substackcdn.com/image/fetch/$s_!gSqP!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a80d25c-3484-4c3b-9eb8-cc60c5d04a84_2048x2560.png 848w, https://substackcdn.com/image/fetch/$s_!gSqP!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a80d25c-3484-4c3b-9eb8-cc60c5d04a84_2048x2560.png 1272w, https://substackcdn.com/image/fetch/$s_!gSqP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a80d25c-3484-4c3b-9eb8-cc60c5d04a84_2048x2560.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Age also matters: younger respondents (18-34) are more likely to use AI than those over 65, with the gap especially pronounced for ChatGPT.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!cmrW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F46cef60a-ebd8-4e38-98c5-3e5c8e2e62c5_2048x2560.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!cmrW!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F46cef60a-ebd8-4e38-98c5-3e5c8e2e62c5_2048x2560.png 424w, https://substackcdn.com/image/fetch/$s_!cmrW!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F46cef60a-ebd8-4e38-98c5-3e5c8e2e62c5_2048x2560.png 848w, https://substackcdn.com/image/fetch/$s_!cmrW!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F46cef60a-ebd8-4e38-98c5-3e5c8e2e62c5_2048x2560.png 1272w, https://substackcdn.com/image/fetch/$s_!cmrW!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F46cef60a-ebd8-4e38-98c5-3e5c8e2e62c5_2048x2560.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!cmrW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F46cef60a-ebd8-4e38-98c5-3e5c8e2e62c5_2048x2560.png" width="1456" height="1820" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/46cef60a-ebd8-4e38-98c5-3e5c8e2e62c5_2048x2560.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1820,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:156613,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://epochai.substack.com/i/182110144?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F46cef60a-ebd8-4e38-98c5-3e5c8e2e62c5_2048x2560.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!cmrW!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F46cef60a-ebd8-4e38-98c5-3e5c8e2e62c5_2048x2560.png 424w, https://substackcdn.com/image/fetch/$s_!cmrW!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F46cef60a-ebd8-4e38-98c5-3e5c8e2e62c5_2048x2560.png 848w, https://substackcdn.com/image/fetch/$s_!cmrW!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F46cef60a-ebd8-4e38-98c5-3e5c8e2e62c5_2048x2560.png 1272w, https://substackcdn.com/image/fetch/$s_!cmrW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F46cef60a-ebd8-4e38-98c5-3e5c8e2e62c5_2048x2560.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>How does this age gap compare to previous technologies? Among respondents 65 and older, about 34% reported using an AI service in the past week. According to<a href="https://www.pewresearch.org/chart/internet-use-by-age-2/"> Pew Research data</a>, that&#8217;s roughly comparable to internet use among the same age group around 2007, when about 35% of Americans over 65 were online.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-8" href="#footnote-8" target="_self">8</a> It&#8217;s also similar to smartphone ownership among over-65s around<a href="https://www.pewresearch.org/internet/2015/04/01/chapter-one-a-portrait-of-smartphone-ownership/"> 2015</a>&#8211;<a href="https://www.pewresearch.org/short-reads/2017/01/12/evolution-of-technology">2016</a>, when about 30&#8211;40% owned one. In other words, AI adoption among older adults today looks similar to where internet and smartphone adoption were roughly 18 and 10 years ago respectively &#8212; past the early enthusiast phase, but still with substantial room to grow.</p><p>Gender, by contrast, shows <a href="https://epoch.ai/data/polling#ai-services">relatively little difference</a> in our data: 59.8% of men had used at least one AI tool over the past week, compared with 54.4% of women. This is notable given that early ChatGPT adoption skewed heavily male; OpenAI researchers <a href="https://cdn.openai.com/pdf/a253471f-8260-40c6-a2cc-aa93fe9f142e/economic-research-chatgpt-usage-paper.pdf#page=26">documented</a> a dramatic narrowing of this gender gap over time, and our polling also suggests it has largely closed for general usage.</p><h2>Conclusion</h2><p>In summary, consumer AI adoption remains strong, but the picture is more nuanced than headline user counts suggest. ChatGPT remains dominant and keeps acquiring users, but Gemini&#8217;s user growth has been faster over the last few months. Web traffic is plateauing, but usage has shifted to apps. OpenAI&#8217;s revenue seems to be on track, but consumer revenue is likely decreasing as a share.</p><p>Meanwhile, the way people use these tools is becoming clearer. Most consumers treat AI as a search-and-writing assistant, not as an autonomous agent completing tasks end-to-end or as a virtual partner. And a substantial share of workplace AI use is happening bottom-up, with workers adopting tools on their own rather than waiting for employer-provided access.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://epochai.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading! Subscribe to receive the latest from Epoch AI.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>In particular, this includes data from <a href="https://cdn.openai.com/pdf/a253471f-8260-40c6-a2cc-aa93fe9f142e/economic-research-chatgpt-usage-paper.pdf">OpenAI</a>, Google, as well as third-party analytics firms like <a href="https://sensortower.com/">SensorTower</a> or <a href="http://similarweb.com">SimilarWeb</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>This is shown in Figure 21 of the <a href="https://cdn.openai.com/pdf/a253471f-8260-40c6-a2cc-aa93fe9f142e/economic-research-chatgpt-usage-paper.pdf">report</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>Our polling data is consistent with this: as part of our polling partnership with Blue Rose Research, they asked US adults how often they use AI tools, across three waves from June through November. After reweighting to be representative of US registered voters, the share reporting they use AI tools &#8220;once or twice a week&#8221; or more frequently was essentially flat: 36.5% in June to July, 37.0% in mid-July to August, and 38.9% in November.</p><p>This is suggestive evidence that the US user acquisition is slowing down substantially, though we&#8217;re cautious about reading too much into it. Notably, the growth in Gemini users we see in global data doesn&#8217;t clearly show up in our survey, which is somewhat puzzling and suggests the polling may not be capturing all the dynamics.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>Unfortunately, it&#8217;s hard to be sure how many of these 1 billion users come from deliberate adoption rather than accidental one-time triggers. We might also be skeptical of this number for other reasons &#8212; for example, our US survey <a href="https://epoch-website--polling-preview-f9wnyczc.web.app/data/polling#ai-services">suggests</a> Meta AI still lags far behind ChatGPT and Gemini in use. But it&#8217;s hard to be certain because our survey data applies to US data, whereas Meta&#8217;s<a href="https://josephlevine.substack.com/p/meta"> strength</a> is in markets outside the US (where WhatsApp dominates).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>Strictly speaking, OpenAI&#8217;s annualized revenue is <a href="https://www.theinformation.com/articles/openais-organizational-problems-hurt-chatgpt">$19 billion</a> as of mid-December. That&#8217;s slightly below the extrapolated $21 billion, but annualized revenues are quite a noisy metric, so these two numbers seem close enough to say that revenue growth has been roughly on trend.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>ChatGPT had <a href="https://www.theinformation.com/articles/openais-organizational-problems-hurt-chatgpt">35 million consumer subscribers</a> in July 2025. Assuming all subscriptions were ChatGPT Plus at $20/month (a lower bound, since some are ChatGPT Pro at $200/month), this implies consumer annualized revenue of at least $8.4 billion. Since total annualized revenue at the end of July <a href="https://epoch.ai/data/ai-companies#explore-the-data">was</a> $12 billion, consumer subscriptions were at least 70% of revenue, most likely more.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>Unlike our poll, &#8220;How people use ChatGPT&#8221; found that &#8220;Practical guidance&#8221; accounts for a higher share of messages than &#8220;Seeking information&#8221;. We think this might be because &#8220;Practical guidance&#8221; queries involve more messages per session, as users are more likely to ask follow-up questions.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-8" href="#footnote-anchor-8" class="footnote-number" contenteditable="false" target="_self">8</a><div class="footnote-content"><p>The Pew data surveys U.S. adults, while our poll&#8217;s percentages apply to U.S. registered voters. This likely doesn&#8217;t affect the comparison too much: the largest difference between registered voters and the general adult population is age composition, and here we&#8217;re looking specifically at the 65+ group.</p></div></div>]]></content:encoded></item><item><title><![CDATA[Is almost everyone wrong about America’s AI power problem?]]></title><description><![CDATA[Why power is less of a bottleneck than you think.]]></description><link>https://epochai.substack.com/p/is-almost-everyone-wrong-about-americas</link><guid isPermaLink="false">https://epochai.substack.com/p/is-almost-everyone-wrong-about-americas</guid><dc:creator><![CDATA[Anson Ho]]></dc:creator><pubDate>Wed, 17 Dec 2025 16:22:34 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!JLF4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6c72ae49-7852-485f-aa64-de835edb80d8_1024x802.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>This post is part of our <a href="https://epochai.substack.com/s/gradient-updates">Gradient Updates</a> newsletter, which shares more opinionated or informal takes about big questions in AI progress. These posts solely represent the views of the authors, and do not necessarily reflect the views of Epoch AI as a whole.</em></p><div><hr></div><p>In AI circles, there&#8217;s a common argument that goes: &#8220;The US is horrible at building power, but China&#8217;s great at it. And since power is so important for the AI race, China wins by default.&#8221;</p><p>This line of reasoning is everywhere. NVIDIA CEO Jensen Huang used it to <a href="https://www.ft.com/content/53295276-ba8d-4ec2-b0de-081e73b3ba43?utm_source=chatgpt.com">argue</a> that &#8220;China is going to win the AI race&#8221; last month. It features in <em><a href="https://situational-awareness.ai/the-free-world-must-prevail/">Situational Awareness</a>, </em>a series of essays about how the world&#8217;s in a fierce race to AGI, which received a <a href="https://x.com/IvankaTrump/status/1839002887600370145">seal of endorsement</a> from Ivanka Trump. There&#8217;s even an entire Dwarkesh podcast <a href="https://www.dwarkesh.com/p/casey-handmer">episode</a> called &#8220;China is killing the US on energy. Does that mean they&#8217;ll win AGI?&#8221;.</p><p>But we think this argument is overstated &#8212; power bottlenecks likely won&#8217;t dramatically or permanently impede the data center buildout in the US. Claims about America&#8217;s AI power predicament are partially based on a misunderstanding, and there are multiple promising approaches to meet America&#8217;s AI&#8217;s power demands. That means that people are overrating the strength of the power bottleneck, and how much that impacts the &#8220;race to AGI&#8221;.</p><p>So why do we believe this, and how could everyone be wrong about America&#8217;s AI power problem?</p><h2>America&#8217;s AI power predicament &#8212; or not?</h2><p>You don&#8217;t have to look hard to find horror stories of US power infrastructure. There are transmission lines that need <a href="https://austinvernon.site/blog/sitingpowerlines.html">approval from 1,700 landowners</a>, electric truck charging depots that need to wait <a href="https://www.canarymedia.com/articles/ev-charging/a-big-barrier-to-californias-electric-truck-goals-a-backlogged-power-grid?utm_source=chatgpt.com">up to a decade</a> to connect to the grid, and weather events that plunge <a href="https://www.ferc.gov/news-events/news/final-report-february-2021-freeze-underscores-winterization-recommendations?utm_source=chatgpt.com">millions of Texans</a> into rolling blackouts.</p><p>And we can also just look at the numbers &#8212; over the last four decades, total US power supply has been flat as a board, whereas China&#8217;s has far surpassed it:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!JLF4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6c72ae49-7852-485f-aa64-de835edb80d8_1024x802.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!JLF4!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6c72ae49-7852-485f-aa64-de835edb80d8_1024x802.png 424w, https://substackcdn.com/image/fetch/$s_!JLF4!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6c72ae49-7852-485f-aa64-de835edb80d8_1024x802.png 848w, https://substackcdn.com/image/fetch/$s_!JLF4!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6c72ae49-7852-485f-aa64-de835edb80d8_1024x802.png 1272w, https://substackcdn.com/image/fetch/$s_!JLF4!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6c72ae49-7852-485f-aa64-de835edb80d8_1024x802.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!JLF4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6c72ae49-7852-485f-aa64-de835edb80d8_1024x802.png" width="1024" height="802" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6c72ae49-7852-485f-aa64-de835edb80d8_1024x802.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:802,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!JLF4!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6c72ae49-7852-485f-aa64-de835edb80d8_1024x802.png 424w, https://substackcdn.com/image/fetch/$s_!JLF4!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6c72ae49-7852-485f-aa64-de835edb80d8_1024x802.png 848w, https://substackcdn.com/image/fetch/$s_!JLF4!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6c72ae49-7852-485f-aa64-de835edb80d8_1024x802.png 1272w, https://substackcdn.com/image/fetch/$s_!JLF4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6c72ae49-7852-485f-aa64-de835edb80d8_1024x802.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Source: <a href="https://situational-awareness.ai/the-free-world-must-prevail/">Situational Awareness</a></figcaption></figure></div><p>So there&#8217;s a natural story here &#8212; regulations have killed the US&#8217; ability to build power and &#8220;win the AI race&#8221;. And there&#8217;s some truth to this; legal hurdles can impede power buildout.</p><p>But this story points the finger at the wrong thing. It assumes that US power capacity has stagnated because America <em>can&#8217;t</em> build &#8212; without first checking whether demand actually warranted more capacity! Maybe Americans just haven&#8217;t <em>wanted</em> much more power!</p><p>Indeed, we see that <a href="https://fred.stlouisfed.org/series/APU000072610#">real energy prices</a> have been basically stagnant over the last twenty years.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!paCA!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F84909a0d-b2d3-4373-9159-a9b5cd283b8c_1024x1280.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!paCA!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F84909a0d-b2d3-4373-9159-a9b5cd283b8c_1024x1280.png 424w, https://substackcdn.com/image/fetch/$s_!paCA!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F84909a0d-b2d3-4373-9159-a9b5cd283b8c_1024x1280.png 848w, https://substackcdn.com/image/fetch/$s_!paCA!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F84909a0d-b2d3-4373-9159-a9b5cd283b8c_1024x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!paCA!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F84909a0d-b2d3-4373-9159-a9b5cd283b8c_1024x1280.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!paCA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F84909a0d-b2d3-4373-9159-a9b5cd283b8c_1024x1280.png" width="1024" height="1280" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/84909a0d-b2d3-4373-9159-a9b5cd283b8c_1024x1280.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1280,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!paCA!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F84909a0d-b2d3-4373-9159-a9b5cd283b8c_1024x1280.png 424w, https://substackcdn.com/image/fetch/$s_!paCA!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F84909a0d-b2d3-4373-9159-a9b5cd283b8c_1024x1280.png 848w, https://substackcdn.com/image/fetch/$s_!paCA!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F84909a0d-b2d3-4373-9159-a9b5cd283b8c_1024x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!paCA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F84909a0d-b2d3-4373-9159-a9b5cd283b8c_1024x1280.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Data from the Federal Reserve Bank of St Louis, dividing <a href="https://fred.stlouisfed.org/series/APU000072610#">average electricity prices</a> by the <a href="https://fred.stlouisfed.org/series/CPIAUCSL">Consumer Price Index</a> (with January 2025 = 100).</figcaption></figure></div><p>Moreover, as we&#8217;ll see, there&#8217;s actually excess capacity available on the grid. So this suggests that the problem wasn&#8217;t an inability to expand supply; rather it was that demand remained stagnant for most of the last few decades.</p><p>You might be thinking, &#8220;there&#8217;s a big elephant in the room here &#8212; isn&#8217;t AI going to hugely increase US power demand?&#8221; And you&#8217;d be right. AI has driven recent spikes in power demand, and AI data centers in the US will need <a href="https://epoch.ai/blog/power-demands-of-frontier-ai-training">30 to 80 GW</a> of power by 2030.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!B9k4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d3c7ec6-0a58-4a2b-9c6a-1d96ccfdbeec_1600x949.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!B9k4!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d3c7ec6-0a58-4a2b-9c6a-1d96ccfdbeec_1600x949.png 424w, https://substackcdn.com/image/fetch/$s_!B9k4!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d3c7ec6-0a58-4a2b-9c6a-1d96ccfdbeec_1600x949.png 848w, https://substackcdn.com/image/fetch/$s_!B9k4!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d3c7ec6-0a58-4a2b-9c6a-1d96ccfdbeec_1600x949.png 1272w, https://substackcdn.com/image/fetch/$s_!B9k4!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d3c7ec6-0a58-4a2b-9c6a-1d96ccfdbeec_1600x949.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!B9k4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d3c7ec6-0a58-4a2b-9c6a-1d96ccfdbeec_1600x949.png" width="1456" height="864" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3d3c7ec6-0a58-4a2b-9c6a-1d96ccfdbeec_1600x949.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:864,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!B9k4!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d3c7ec6-0a58-4a2b-9c6a-1d96ccfdbeec_1600x949.png 424w, https://substackcdn.com/image/fetch/$s_!B9k4!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d3c7ec6-0a58-4a2b-9c6a-1d96ccfdbeec_1600x949.png 848w, https://substackcdn.com/image/fetch/$s_!B9k4!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d3c7ec6-0a58-4a2b-9c6a-1d96ccfdbeec_1600x949.png 1272w, https://substackcdn.com/image/fetch/$s_!B9k4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d3c7ec6-0a58-4a2b-9c6a-1d96ccfdbeec_1600x949.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Let&#8217;s assume that AI&#8217;s power demand grows aggressively, reaching 100 GW by 2030.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> That means the US would need to build out four times more power than <a href="https://www.caiso.com/Documents/SummerMarketPerformanceReportforSeptember2022.pdf?utm_source=chatgpt.com">California needs on average</a>. We haven&#8217;t seen this much relative growth in electricity demand <a href="https://www.eia.gov/totalenergy/data/annual/txt/ptb0809.html">since the 1980s</a>.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a></p><p>But we think the US can likely meet this demand anyway. As long as there&#8217;s enough will to invest in AI scaling, there&#8217;ll be enough will to pay for the power. After all, power is a relatively small slice of the cost of running a data center &#8212; chips are far more expensive. If companies are willing to shell out for massive AI chip clusters, they won&#8217;t balk at paying a premium for electricity.</p><h2>Plucking fruit from the power tree</h2><p>This sudden spike in power demand after decades of stagnation means that the US is off to a standing start. But the flip side is that America hasn&#8217;t really <em>tried</em> to meet this kind of demand surge, so there&#8217;s still a bunch of relatively low-hanging fruit to be picked.</p><h3>Natural gas: Digging power out of the ground</h3><p>Probably the most salient way to supply power is to build out natural gas &#8212; it&#8217;s <a href="https://epoch.ai/blog/what-you-need-to-know-about-ai-data-centers">relatively cheap</a> and can be built fast (e.g. without immediately connecting to the electrical grid), which is why it powers <a href="https://epoch.ai/data/data-centers">frontier data centers</a> like <a href="https://cdn.prod.website-files.com/6855c1aa175582ee23e0aa19/68710273a71bb7c11049ba28_Crusoe_Impact_Report_2024.pdf#page=19">OpenAI&#8217;s Stargate Abilene</a> and <a href="https://www.datacenterdynamics.com/en/news/meta-announces-4-million-sq-ft-louisiana-data-center-campus/">Meta&#8217;s Hyperion</a>.</p><p>So how much natural gas could we build out? As a starting point, natural gas providers already aim to grow their production a lot. Consider the three largest gas turbine manufacturers, which collectively make up something like 85% of <a href="https://gasturbineworld.com/market-forecast/">market share</a>:</p><ul><li><p>GE Vernova <a href="https://www.gevernova.com/sites/default/files/gev_webcast_presentation_12092025.pdf">says</a> they&#8217;ll build 20 GW per year by 2026 and 24 GW per year by 2028.</p></li><li><p>Mitsubishi Heavy plans to <a href="https://www.bloomberg.com/news/articles/2025-08-31/mitsubishi-heavy-to-double-gas-turbine-capacity-as-demand-soars">double</a> their gas turbine production over two years, though sadly we don&#8217;t know what this means in terms of absolute power capacity</p></li><li><p>Siemens <a href="https://p3.aprimocdn.net/siemensenergy/5fc093ca-4fd8-482c-8390-b39b004f66f3/251120_CMD-2025_Gas_Services_FINAL-pdf_Original%20file.pdf">plans to</a> boost their gas turbine production capacity from 17 GW per year in 2024, to 22 GW per year between 2025 and 2027, and over 30 GW a year by 2028.</p></li></ul><p>So it&#8217;s hard to say overall, but even if we just sum up the power contributions from GE Vernova and Siemens until 2030, that brings us beyond 200 GW. Accounting for Mitsubishi Heavy and other producers should bring us well beyond that. This power isn&#8217;t necessarily produced for US consumers (last year US consumption contributed 20% to 40% of the revenues of each company),<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a> but it does strongly suggest that these companies would be <em>able</em> to supply the AI power demand for several years to come.</p><p>This is also broadly consistent with <a href="https://rmi.org/gas-turbine-supply-constraints-threaten-grid-reliability-more-affordable-near-term-solutions-can-help/">older forecasts</a> for natural gas production in the US, with over 100 GW produced by 2030.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!5cjr!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1da0d50-8088-40f5-9271-d4d81cf6b6e8_892x1158.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!5cjr!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1da0d50-8088-40f5-9271-d4d81cf6b6e8_892x1158.png 424w, https://substackcdn.com/image/fetch/$s_!5cjr!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1da0d50-8088-40f5-9271-d4d81cf6b6e8_892x1158.png 848w, https://substackcdn.com/image/fetch/$s_!5cjr!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1da0d50-8088-40f5-9271-d4d81cf6b6e8_892x1158.png 1272w, https://substackcdn.com/image/fetch/$s_!5cjr!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1da0d50-8088-40f5-9271-d4d81cf6b6e8_892x1158.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!5cjr!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1da0d50-8088-40f5-9271-d4d81cf6b6e8_892x1158.png" width="892" height="1158" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f1da0d50-8088-40f5-9271-d4d81cf6b6e8_892x1158.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1158,&quot;width&quot;:892,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!5cjr!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1da0d50-8088-40f5-9271-d4d81cf6b6e8_892x1158.png 424w, https://substackcdn.com/image/fetch/$s_!5cjr!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1da0d50-8088-40f5-9271-d4d81cf6b6e8_892x1158.png 848w, https://substackcdn.com/image/fetch/$s_!5cjr!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1da0d50-8088-40f5-9271-d4d81cf6b6e8_892x1158.png 1272w, https://substackcdn.com/image/fetch/$s_!5cjr!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1da0d50-8088-40f5-9271-d4d81cf6b6e8_892x1158.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Source: <a href="https://rmi.org/gas-turbine-supply-constraints-threaten-grid-reliability-more-affordable-near-term-solutions-can-help/">Cohen, Fitch, and Shwisberg</a></figcaption></figure></div><p>Now we can&#8217;t quite declare victory yet, because of two issues. First, natural gas capacity doesn&#8217;t translate directly into data center capacity. So if there&#8217;s 100 GW of added natural gas capacity, grid planners might only expect around <a href="https://stanwichenergy.com/insights/understanding-effective-load-carrying-capability-elcc-how-renewable-reliability-impacts-costs-for-energy-users">80 GW</a> of that to be reliably available during peak demand.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a></p><p>The second (and trickier) issue is that not all of this power capacity goes to AI &#8212; but we don&#8217;t know how much. GE Vernova <a href="https://gasoutlook.com/analysis/costs-to-build-gas-plants-triple-says-ceo-of-nextera-energy/">states</a> that their backlog goes into 2028, and &#8220;only a small portion of those orders relate to data centers and AI&#8221;. But what&#8217;s a &#8220;small portion&#8221;? 3%? 30%? At the very least, AI probably makes up a pretty substantial portion of the most recent orders &#8212; <a href="https://p3.aprimocdn.net/siemensenergy/870b5cc4-8054-4705-9d6b-b331005a2187/Presentation-Q3-Website-English-%28final%29-pdf_Original%20file.pdf">65%</a> in the case of Siemens. And if people are willing to spend big on AI scaling, many of these orders could potentially be bought from other use cases.</p><p>Even after we account for these two issues, it seems like natural gas could cover a big chunk of expected AI power demand. And planned natural gas buildouts are actually quite modest: companies are planning to spend <a href="https://www.gevernova.com/news/press-releases/ge-vernova-raises-multi-year-financial-outlook-initiates-dividend-authorizes-buyback">several billions</a> on equipment for natural gas through 2028, which sounds like a lot but pales in comparison to the trillions needed for IT equipment. If companies are going to spend a lot on buying GPUs anyway, they&#8217;re going to be willing to pay a fraction of that amount for power. So there&#8217;s plenty of room to increase turbine production even more if they really wanted to.</p><p>This can mean taking steps that aren&#8217;t the most cost-effective by traditional power industry metrics, because the time-to-power is what matters most for AI. For example, rather than using standard &#8220;combined-cycle&#8221; gas turbines, AI companies are turning to smaller &#8220;simple-cycle <a href="https://www.gevernova.com/content/dam/gepower-microsites/global/en_US/documents/avr/GEA34130%20AeroderivativeGT_Whitepaper_R5.pdf">aeroderivative gas turbines&#8221;</a>. Although these jet engine-based turbines can cost more than their combined cycle counterparts, they&#8217;re also much faster to build.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a></p><p>But there&#8217;s perhaps an even better example of how gas turbines could be rapidly scaled up &#8212; Elon Musk&#8217;s xAI. Founded in <a href="https://en.wikipedia.org/wiki/XAI_(company)">2023</a>, xAI needed to catch up quickly with the likes of Anthropic and OpenAI. That meant building data centers <em>fast</em> and powering them in pretty out-of-the-box ways &#8212; <a href="https://www.offgridai.us/">renting gas</a> before connecting to the electrical grid, deploying natural gas turbines <a href="https://www.datacenterdynamics.com/en/news/elon-musk-xai-gas-turbines-memphis/">prior to receiving permits</a>, and even <a href="https://www.datacenterdynamics.com/en/news/xai-importing-power-plant-from-abroad-for-new-memphis-data-center-elon-musk-claims/?utm_source=chatgpt.com">importing and disassembling a gas plant from overseas</a>!</p><h3>Solar: Raining power down from the sky</h3><p>While there&#8217;s still room for Elon Musk-esque &#8220;just get the power ASAP&#8221; approaches, we should also consider other sources of power. And the best candidate for this is solar energy &#8212; it&#8217;s more environmentally friendly, and the price has famously been dropping <a href="https://www.dwarkesh.com/p/casey-handmer">monstrously</a> <a href="https://ourworldindata.org/learning-curve">fast</a> compared to other energy sources:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Aoeb!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02f8e4a5-0e57-4520-8a8f-1dcb5b9de719_1242x1600.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Aoeb!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02f8e4a5-0e57-4520-8a8f-1dcb5b9de719_1242x1600.png 424w, https://substackcdn.com/image/fetch/$s_!Aoeb!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02f8e4a5-0e57-4520-8a8f-1dcb5b9de719_1242x1600.png 848w, https://substackcdn.com/image/fetch/$s_!Aoeb!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02f8e4a5-0e57-4520-8a8f-1dcb5b9de719_1242x1600.png 1272w, https://substackcdn.com/image/fetch/$s_!Aoeb!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02f8e4a5-0e57-4520-8a8f-1dcb5b9de719_1242x1600.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Aoeb!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02f8e4a5-0e57-4520-8a8f-1dcb5b9de719_1242x1600.png" width="1242" height="1600" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/02f8e4a5-0e57-4520-8a8f-1dcb5b9de719_1242x1600.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1600,&quot;width&quot;:1242,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Aoeb!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02f8e4a5-0e57-4520-8a8f-1dcb5b9de719_1242x1600.png 424w, https://substackcdn.com/image/fetch/$s_!Aoeb!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02f8e4a5-0e57-4520-8a8f-1dcb5b9de719_1242x1600.png 848w, https://substackcdn.com/image/fetch/$s_!Aoeb!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02f8e4a5-0e57-4520-8a8f-1dcb5b9de719_1242x1600.png 1272w, https://substackcdn.com/image/fetch/$s_!Aoeb!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02f8e4a5-0e57-4520-8a8f-1dcb5b9de719_1242x1600.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>More importantly, solar power can be scaled <em>a lot</em>. Solar installations have grown <a href="https://www.reuters.com/markets/commodities/world-solar-generation-set-eclipse-nuclear-first-time-maguire-2025-05-21/">roughly exponentially</a> for at least the last decade, and even if US <a href="https://seia.org/research-resources/solar-market-insight-report-q3-2025/">policy headwinds</a> stopped us from going beyond 2024&#8217;s rate of <a href="https://www.irena.org/-/media/Files/IRENA/Agency/Publication/2025/Jul/IRENA_DAT_RE_Statistics_2025.pdf">40 GW</a> per year, we&#8217;d end up with 200 GW over the next five years.</p><p>Once again, we shouldn&#8217;t jump the gun and declare victory just yet &#8212; not all of this solar capacity translates into data center capacity. In practice, we might have to cut this by a <a href="https://www.reuters.com/markets/commodities/world-solar-generation-set-eclipse-nuclear-first-time-maguire-2025-05-21/?utm_source=chatgpt.com">factor of five</a> to <a href="https://www.rand.org/pubs/research_reports/RRA3845-1.html">40 GW</a> over the next five years. This is a huge drop, but it still covers a good chunk of AI&#8217;s power demand, and like with natural gas, there&#8217;s a lot of room to grow the supply of solar power.</p><p>Importantly, the core bottlenecks are things like installation costs, permitting costs, and connecting solar power to the grid. It&#8217;s less about manufacturing &#8212; the US can basically manufacture or import as many solar panels and batteries as they could realistically want. For instance, US solar manufacturing capacity reached <a href="https://seia.org/research-resources/solar-market-insight-report-q3-2025/">55 GW</a> in the second quarter of 2025 alone, which is more than enough to supply installations for the full year. So if only you could bypass these bottlenecks, you could unlock a lot of power.</p><p>This is exactly what was <a href="http://offgridai.us">proposed</a> in a report late last year &#8212; rather than waiting to connect renewables to the grid, simply move off-grid. The idea is to have many &#8220;microgrids&#8221; that are powered by solar and a bit of natural gas backup. Each microgrid would individually power data centers, such as from 100 MW to 1 GW in size. According to them, this could unlock <em>over</em> <em>1000 GW </em>of potential power for data centers in the US Southwest alone. And they think this could be built out fast:</p><blockquote><p>&#8220;Estimated time to operation for a large off-grid solar microgrid could be around 2 years (1-2 years for site acquisition and permitting plus 1-2 years for site buildout), though there&#8217;s no obvious reason why this couldn&#8217;t be done faster by very motivated and competent builders.&#8221;</p></blockquote><p>That said, this approach is still pretty speculative, and has a bunch of hiccups. For one, you need a lot of land &#8212; 2 GW of solar power needs as much space as all of Manhattan!<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a> Moreover, while you might want your data center to run 24/7, you can&#8217;t generate solar power at night. If you&#8217;re connected to the grid this isn&#8217;t too big of a deal because you can draw on spare grid capacity from other power sources, but going off-grid makes this more challenging. These reasons are probably why (to our knowledge) there haven&#8217;t yet been plans for gigawatt-scale data centers to be mainly solar-powered.</p><p>But as with natural gas, these obstacles don&#8217;t seem insurmountable. For one, a lot of US land gets a ton of sun &#8212; enough land for over ten thousand Manhattan-sized solar arrays.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-8" href="#footnote-8" target="_self">8</a> Moreover, batteries can help substantially support off-grid approaches. There was a roughly <a href="https://www.eia.gov/todayinenergy/detail.php?id=64705">20 GW</a> increase in battery capacity from 2024 to 2025, and if this is sustained this would bring us to 100 GW by 2030.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-9" href="#footnote-9" target="_self">9</a> And in many ways this is too conservative &#8212; batteries could be imported, and also repurposed from Electric Vehicles, which currently make up <a href="https://www.iea.org/reports/global-ev-outlook-2025/electric-vehicle-batteries">the vast majority of battery demand</a>. In fact there&#8217;s already some precedent for this &#8212; companies like Redwood Materials and Crusoe were <a href="https://www.crusoe.ai/resources/newsroom/crusoe-and-redwood-materials-power-the-future-of-ai?utm_source=chatgpt.com">exploring</a> this approach earlier this year.</p><h3>Demand response: Acquiring power &#8220;out of thin air&#8221;</h3><p>So far we&#8217;ve looked at two ways of building power &#8212; but there&#8217;s one more wildcard that could make a huge difference. This is what Dean Ball cheekily <a href="https://www.hyperdimensional.co/p/out-of-thin-air">describes</a> as pulling &#8220;gigawatts out of thin air&#8221;, the more technical term for which is &#8220;demand response&#8221; (yet another name is &#8220;data center flexibility&#8221;).</p><p>As crazy as it might sound, the US electricity grid is oversupplied most of the time. That&#8217;s because power infrastructure is built to handle <em>peak</em> demand, like when everyone&#8217;s turned up their air conditioning during a heatwave.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-10" href="#footnote-10" target="_self">10</a> But most of the time, there&#8217;s actually abundant spare power in the grid, enough to run (say) a 1 GW data center.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-11" href="#footnote-11" target="_self">11</a></p><p>To make this work, data centers would need to reduce power use during peak demand. This can be tricky &#8212; for one, data centers would need to credibly commit to decreasing their power draw when necessary. This probably also involves deliberately curtailing AI workloads, which raises technical challenges &#8212; but not insurmountable ones. Even today, large training runs have to contend with <a href="https://epoch.ai/blog/hardware-failures-wont-limit-ai-scaling">spontaneous AI chip failures</a> that force the entire process to stop, but data center operators have been able to manage these quite successfully.</p><p>If demand response actually works, it could unlock a lot of power: One <a href="https://nicholasinstitute.duke.edu/articles/significant-flexible-load-potential-us-grid-finds-duke-study">study</a> finds that if data centers could cut power demand by just 0.25% across the year, they could get around 76 GW of spare capacity.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-12" href="#footnote-12" target="_self">12</a> That&#8217;s <a href="https://www.eia.gov/todayinenergy/detail.php?id=65104">75%</a> of the power of the entire US nuclear power plant fleet, and gets us a lot of the way to meeting AI&#8217;s power demand. If you&#8217;re willing to curtail power demand even more, you could get substantially more power &#8212; e.g. a 1.0% power cut across the year gives you 126 GW of power.</p><p>This almost sounds too good to be true, but is already gaining traction. Google has <a href="https://blog.google/inside-google/infrastructure/how-were-making-data-centers-more-flexible-to-benefit-power-grids/">agreed</a> to implement this in Indiana and Tennessee. The Electric Power Research Institute launched a <a href="https://dcflex.sf.epri.com/">project</a> last year to demonstrate demand response in real markets. And <a href="https://www.emeraldai.co/">Emerald AI</a> has been working with companies like Oracle and Nvidia to implement it in practice.</p><p>This doesn&#8217;t mean it&#8217;ll all be smooth sailing. The estimates in the aforementioned study might be too optimistic.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-13" href="#footnote-13" target="_self">13</a> And moreover, a demand response <a href="https://www.pjm.com/-/media/DotCom/committees-groups/cifp-lla/2025/20250915/20250915-item-07---pjm-initial-proposal-and-alternatives-considered---pjm-presentation.pdf">proposal</a> by the <a href="https://www.pjm.com/about-pjm">PJM</a> (the largest grid operator in the US) has been met with some strong <a href="https://www.pjm.com/-/media/DotCom/committees-groups/cifp-lla/postings/20250828-stakeholder-comments-cifp-lla.pdf?utm_source=chatgpt.com">pushback</a> from big tech and data centers. But the pushback seems mostly about the <em>specific </em>proposal,<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-14" href="#footnote-14" target="_self">14</a> and given that multiple actors have their own <a href="https://www.ercot.com/mktrules/issues/PGRR134#keydocs">proposals</a>, this doesn&#8217;t seem like a big strike against demand response in general.</p><h3>Adding up the numbers</h3><p>Now let&#8217;s put together all the numbers. Our best projections suggest that the US will need around 100 GW of power by 2030. And if we look at individual power sources, we see there&#8217;s a lot of potential to meet this demand. Demand response unlocks tens of gigawatts of power, and plausibly over 100 GW. Solar and natural gas <em>individually</em> contribute many tens more gigawatts under existing projections, and there&#8217;s likely substantial room to grow by going off-grid or just scaling up equipment spending.</p><p>These different approaches can complement each other to some extent. For example, traditional buildouts of solar power could require five to seven years to connect to the grid, which is too slow. But if data centers procure their own solar power generation capacity and combine this with demand response, they could <a href="https://cdn.prod.website-files.com/60dbdcca2e4b1919e8894fa5/6930abf1be0f36db6fc27157_Whitepaper%20-%20With%20Appendix.pdf">cut this wait time down to two years</a>.</p><p>And let&#8217;s not forget that there are other power sources too. For example, next-generation geothermal power seems close to medium-scale commercial deployment &#8212; one pilot project is already <a href="https://www.worldoil.com/magazine/2025/november-2025/features/breakthroughs-in-geothermal-drilling-eavor-s-foak-closed-loop-commercial-project-in-germany/">online</a> and there are several <a href="https://www.datacenterdynamics.com/en/news/geothermal-developer-fervo-raises-432m-with-backing-from-google/">startups</a> backed by AI &#8220;hyperscalers&#8221; that promise hundreds of megawatts by 2028. By 2035, these buildouts could plausibly add <a href="https://www.mckinsey.com/industries/electric-power-and-natural-gas/our-insights/is-geothermal-energy-ready-to-make-its-mark-in-the-us-power-mix">40 GW</a> of geothermal power.</p><p>So if we combine all these approaches, it looks quite likely that the US will have enough power for AI scaling until 2030.</p><p>Strictly speaking this isn&#8217;t quite enough because we don&#8217;t just care about the <em>total </em>power &#8212; to train AI systems, this power often needs to be concentrated in a single location. And if current trends persist, this might mean truly massive training runs with <a href="https://epoch.ai/blog/power-demands-of-frontier-ai-training">4 to 16 GW</a> by 2030. However, this is a big if, because a lot of model development is now <a href="https://newsletter.semianalysis.com/p/scaling-reinforcement-learning-environments-reward-hacking-agents-scaling-data">less focused</a> on performing a single large training run in one location.</p><p>But we doubt even this will be that much of a blocker. Frontier AI companies seem to be doing a pretty good job at quickly finding locations for <a href="https://epoch.ai/data/data-centers">gigawatt-scale data centers</a>. Consider how OpenAI and Oracle were able to do this for <a href="https://www.reuters.com/business/media-telecom/openai-oracle-softbank-plan-five-new-ai-data-centers-500-billion-stargate-2025-09-23/">multiple sites</a> within roughly a year, as part of their <a href="https://openai.com/index/announcing-the-stargate-project/">Stargate project</a>. And even if we can&#8217;t find individual locations that fit the bill, companies could split <a href="https://epoch.ai/blog/could-decentralized-training-solve-ais-power-problem">training across multiple locations</a> to circumvent local power constraints.</p><h2>So is everyone wrong about this?</h2><p>Taking a step back, what we&#8217;re saying is perhaps rather outrageous. Are we <em>really</em> saying that everyone is wrong? Do we <em>really</em> think that we know better than all the frontier AI companies, all the AI policy researchers, and all the utility companies CEOs? If these approaches are so good, why isn&#8217;t everybody implementing them?</p><p>The thing is, they <em>are</em> implementing them! As we saw earlier, new frontier AI data centers <a href="https://www.latitudemedia.com/news/behind-the-meter-generation-is-picking-up-traction/">often use natural gas</a>, and companies like Google are now <a href="https://blog.google/inside-google/infrastructure/how-were-making-data-centers-more-flexible-to-benefit-power-grids/">implementing</a> demand response.</p><p>And to the extent that these approaches aren&#8217;t being put into practice, remember that a lot of this is very new. The US has seen &#8220;stagnant power demand for a few decades followed by a huge AI-shaped spike, roughly since ChatGPT was released three years ago&#8221;. And three years is ancient history by AI standards, but a blink of an eye for the power industry. So we shouldn&#8217;t be surprised that many people have been thinking about power through an outdated lens, tailored to a world without several gigawatt-scale megaprojects spun up each year.</p><p>The pressures of AI competition have transformed the picture. Consider electrical transformers and transmission lines: connecting a new facility to the grid traditionally requires building these out, a process which can take years. But in an industry where multi-year waits are a death sentence, companies are increasingly turning to off-grid generation and demand response programs which sidestep connecting new power to the grid.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-15" href="#footnote-15" target="_self">15</a></p><p>Why then are <a href="https://openai.com/global-affairs/seizing-the-ai-opportunity/">so</a> <a href="https://ifp.org/future-of-ai-compute/">many</a> <a href="https://www.anthropic.com/news/build-ai-in-america">actors</a> <a href="https://www.energy.gov/sites/default/files/2025-10/403%20Large%20Loads%20Letter.pdf">complaining</a> about the size of the power bottleneck? We think the obvious reason is true &#8212; there are indeed challenges to accessing power for AI scaling! Overcoming these challenges often means using out-of-the-box techniques for acquiring power, and might end up costing more. So there are <a href="https://x.com/ShanuMathew93/status/1991686457438712037?s=20">incentives</a> to try and reduce the price of power as much as possible.</p><p>The upshot: we doubt that these challenges alone are strong enough to impede AI scaling. Power bottlenecks will likely add months &#8212; but not years &#8212; to build-out times. Even if power prices tripled, they would still be substantially lower than the costs for AI compute &#8212; in current frontier data centers, the cost of AI chips is roughly <a href="https://docs.google.com/spreadsheets/d/1_hu9Nczi4ob4tyRsPH78e3fFw-xWMzgTTv8ZF76SHwM/edit?gid=0#gid=0">ten times higher</a> than that of power! If people are willing to build out huge data centers with loads of chips, then they&#8217;re also willing to pay a premium to build out power.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Qep1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F792ae2e0-5519-4b6d-966c-d70e1b927f5d_1024x1280.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Qep1!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F792ae2e0-5519-4b6d-966c-d70e1b927f5d_1024x1280.png 424w, https://substackcdn.com/image/fetch/$s_!Qep1!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F792ae2e0-5519-4b6d-966c-d70e1b927f5d_1024x1280.png 848w, https://substackcdn.com/image/fetch/$s_!Qep1!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F792ae2e0-5519-4b6d-966c-d70e1b927f5d_1024x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!Qep1!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F792ae2e0-5519-4b6d-966c-d70e1b927f5d_1024x1280.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Qep1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F792ae2e0-5519-4b6d-966c-d70e1b927f5d_1024x1280.png" width="1024" height="1280" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/792ae2e0-5519-4b6d-966c-d70e1b927f5d_1024x1280.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1280,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Qep1!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F792ae2e0-5519-4b6d-966c-d70e1b927f5d_1024x1280.png 424w, https://substackcdn.com/image/fetch/$s_!Qep1!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F792ae2e0-5519-4b6d-966c-d70e1b927f5d_1024x1280.png 848w, https://substackcdn.com/image/fetch/$s_!Qep1!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F792ae2e0-5519-4b6d-966c-d70e1b927f5d_1024x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!Qep1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F792ae2e0-5519-4b6d-966c-d70e1b927f5d_1024x1280.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The question that matters the most, then, is whether people will be willing to spend on AI scaling. So far, spending on building power has been fairly conservative, perhaps <a href="https://heatmap.news/ideas/natural-gas-turbine-crisis">due to</a> <a href="https://www.youtube.com/watch?v=o7TUaV4E5Uo">uncertainty in future power demand</a>.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-16" href="#footnote-16" target="_self">16</a> But if there continues to be intense economic and geopolitical incentives to build AI infrastructure in the US, then the AI industry will probably be able to secure a large share of new power generation, especially for smaller-scale AI data centers used for inference or decentralized training.</p><p>This finally brings us back to the original argument &#8212; does China&#8217;s power advantage mean that they &#8220;win by default&#8221; (whatever that means)? Not necessarily. It&#8217;s still the case that China can build power faster than the US, but America&#8217;s power bottleneck is much weaker than many people make it out to be. Moreover, the whole reason electrical power matters for AI is that it&#8217;s used to run AI chips. But as things currently stand, China still has to overcome serious <a href="https://epoch.ai/gradient-updates/why-china-isnt-about-to-leap-ahead-of-the-west-on-compute">deficits in the quantity and quality of their AI chips</a> to gain a compute advantage over the US &#8212; and that would be true even with <em>unlimited</em> power capacity.</p><p>The upshot is this: power does pose a real challenge, but it matters a lot how big this challenge is. At the very least, we doubt that power bottlenecks will be strong enough to substantially slow down US AI scaling. For that, we&#8217;re not holding our breath.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://epochai.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading! Subscribe to receive the latest from Epoch AI.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p><em>We&#8217;d like to thank Lynette Bye, Yann Riviere and Konstantin Pilz for their feedback and support on this post.</em></p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>The same is true for <a href="https://www.eia.gov/electricity/data/browser/#/topic/5?agg=0,1&amp;geo=g&amp;endsec=vg&amp;linechart=ELEC.SALES.US-ALL.A~~~~~&amp;columnchart=ELEC.SALES.US-ALL.A~ELEC.SALES.US-RES.A~ELEC.SALES.US-COM.A~ELEC.SALES.US-IND.A&amp;map=ELEC.SALES.US-ALL.A&amp;freq=A&amp;start=2003&amp;end=2023&amp;chartindexed=0&amp;ctype=columnchart&amp;ltype=pin&amp;columnvalues=0&amp;rtype=s&amp;pin=&amp;rse=0&amp;maptype=0">retail sales of electricity</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>This isn&#8217;t a strict upper bound, for example <a href="https://www.rand.org/pubs/research_reports/RRA3572-1.html">some estimates</a> are substantially higher, in the ballpark of 250 GW (and if building data centers costs <a href="https://epoch.ai/blog/what-you-need-to-know-about-ai-data-centers">$30 billion per GW</a>, this amounts to $7.5 trillion!). We&#8217;re more skeptical that we&#8217;ll reach these scales, but it&#8217;s possible. </p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>100 GW would be around 10% of America&#8217;s peak power capacity. If we spread this out over four to five years, this works out to a growth rate of 2% to 3%, which we haven&#8217;t seen <a href="https://www.eia.gov/totalenergy/data/annual/txt/ptb0809.html">since roughly the 1980s</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>For example, US consumption contributed around <a href="https://www.gevernova.com/sites/default/files/gevernova_2024_form10-k.pdf">40%</a> of revenue for GE Vernova, about <a href="https://p3.aprimocdn.net/siemensenergy/6974f976-432a-4ed9-ba61-b24301167dc8/A-Siemens-Energy-Annual-Report-2024-pdf_Original%20file.pdf">20%</a> for Siemens, and roughly <a href="https://www.mhi.com/finance/library/annual/pdf/report_2024_financial.pdf">20%</a> for Mitsubishi Heavy.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>This is captured more formally by the <a href="https://stanwichenergy.com/insights/understanding-effective-load-carrying-capability-elcc-how-renewable-reliability-impacts-costs-for-energy-users">Effective Load Carrying Capacity (ELCC)</a>, which is around 80% for natural gas.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>Good data is somewhat sparse, but there is some suggestive evidence. For example, a <a href="https://www.eia.gov/analysis/studies/powerplants/capitalcost/pdf/capital_cost_AEO2025.pdf">report</a> from the Energy Information Administration shows a handful of examples where simple-cycle aeroderivative turbines cost over 50% more than their combined-cycle counterparts. Similarly, smaller simple-cycle gas turbines can <a href="https://www.gevernova.com/content/dam/gepower-new/global/en_US/downloads/gas-new-site/resources/GEA31885B-TM2500DLE-brochure.pdf">take</a> 18 to 24 months to go from order to operation, and larger combined-cycle gas turbines take 36 to 48 months.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>Each GW of solar power needs around <a href="https://docs.nrel.gov/docs/fy13osti/56290.pdf#page=1.00&amp;gsr=0">30 square kilometers</a> of land, so for 2 GW you&#8217;d need around 60 square km, which is around the <a href="https://en.wikipedia.org/wiki/Manhattan">land area of Manhattan</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-8" href="#footnote-anchor-8" class="footnote-number" contenteditable="false" target="_self">8</a><div class="footnote-content"><p>There&#8217;s around <a href="https://atb.nrel.gov/electricity/2024/2023/utility-scale_pv?utm_source=chatgpt.com">1.5 million square km</a> of US land that receives over 5 kWh/m<sup>2</sup>/day. That&#8217;s roughly the equivalent of 5 hours a day of full midday sun &#8212; for comparison, Los Angeles has an average of <a href="https://www.solarenergylocal.com/states/california/los-angeles/">5.3 kWh/m<sup>2</sup>/day</a>. Given that Manhattan&#8217;s land area is around <a href="https://en.wikipedia.org/wiki/Manhattan">60 square km</a>, this is enough for 25,000 Manhattans.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-9" href="#footnote-anchor-9" class="footnote-number" contenteditable="false" target="_self">9</a><div class="footnote-content"><p>This is consistent with <a href="https://publications.anl.gov/anlpubs/2024/03/187735.pdf">projections</a> from the US Department of Energy&#8217;s Argonne National Laboratory, which expects US battery cell capacity to grow by around 930 GWh through 2030. If batteries are 10-hour-systems, this would be enough to support around 90 GW of power.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-10" href="#footnote-anchor-10" class="footnote-number" contenteditable="false" target="_self">10</a><div class="footnote-content"><p>A better example comes from the UK. During commercial breaks, many TV-watchers simultaneously take the opportunity to boil water to prepare another cup of tea or coffee. That leads to a predictable <a href="https://en.wikipedia.org/wiki/TV_pickup">power spike</a> of around 200 to 400 MW.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-11" href="#footnote-anchor-11" class="footnote-number" contenteditable="false" target="_self">11</a><div class="footnote-content"><p>This doesn&#8217;t necessarily preclude training runs that need much more than 1 GW of power. For example, it&#8217;s likely possible to train AI models across multiple data centers in <a href="https://epoch.ai/blog/could-decentralized-training-solve-ais-power-problem">decentralized</a> fashion, even if each data center needs less than a gigawatt.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-12" href="#footnote-anchor-12" class="footnote-number" contenteditable="false" target="_self">12</a><div class="footnote-content"><p>In practice, this doesn&#8217;t quite mean cutting power 0.25% of the time (i.e. 22 hours per year) &#8212; these power reductions are usually partial rather than fully cutting out the data center&#8217;s power load. Instead, the authors estimate this would be spread out over 85 hours. But that&#8217;s still just under 1% of the total hours in a year.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-13" href="#footnote-anchor-13" class="footnote-number" contenteditable="false" target="_self">13</a><div class="footnote-content"><p>In general, we worry it&#8217;s <a href="https://nicholasinstitute.duke.edu/sites/default/files/documents/rethinking-load-growth-webinar-presentation.pdf">hard</a> <a href="https://cdn.prod.website-files.com/60dbdcca2e4b1919e8894fa5/6930abf1be0f36db6fc27157_Whitepaper%20-%20With%20Appendix.pdf#page=44.08">to know</a> what happens when you apply something as novel as demand response at close to 100 GW scales (whereas building out this much power is something with much more historical precedent).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-14" href="#footnote-anchor-14" class="footnote-number" contenteditable="false" target="_self">14</a><div class="footnote-content"><p>PJM proposed that data centers could avoid capacity charges by accepting priority curtailment during emergencies. But if not enough people volunteer, PJM would mandate participation. Data centers objected that this singled them out for inferior service while existing customers faced no equivalent obligation, and utilities argued PJM lacked authority to create what amounts to a new class of retail service.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-15" href="#footnote-anchor-15" class="footnote-number" contenteditable="false" target="_self">15</a><div class="footnote-content"><p>In the longer run, more substantial buildouts should be possible &#8212; for example, there&#8217;s historical precedent for much larger supply of <a href="https://www.energy.gov/sites/prod/files/2014/04/f15/LPTStudyUpdate-040914.pdf">transformers</a> and investment into <a href="https://cleanenergygrid.org/wp-content/uploads/2024/07/GS_ACEG-Fewer-New-Miles-Report-July-2024.pdf">transmission lines</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-16" href="#footnote-anchor-16" class="footnote-number" contenteditable="false" target="_self">16</a><div class="footnote-content"><p>This in turn depends on the <a href="https://epoch.ai/gradient-updates/openai-is-projecting-unprecedented-revenue-growth">economic returns</a> to further compute scaling &#8212; if these are high enough, this would help justify substantial investments.</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[Benchmark Scores = General Capability + Claudiness]]></title><description><![CDATA[Is this because skills generalize very well, or because developers are pushing on all benchmarks at once?]]></description><link>https://epochai.substack.com/p/benchmark-scores-general-capability</link><guid isPermaLink="false">https://epochai.substack.com/p/benchmark-scores-general-capability</guid><dc:creator><![CDATA[Greg Burnham]]></dc:creator><pubDate>Thu, 20 Nov 2025 21:09:33 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!i7Ha!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9dce396d-38f9-43f0-95ca-da6c5844bb64_1023x1279.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The Gemini 3 release included a <a href="https://storage.googleapis.com/gweb-uniblog-publish-prod/original_images/gemini_3_table_final_HLE_Tools_on.gif">massive table</a> showing how the model was state-of-the-art on nineteen diverse benchmarks. Such tables are commonplace by now, but they add up to an odd statistical situation. Benchmarks ostensibly measure different things, but since models tend to improve on many benchmarks at once, the dataset of benchmark scores is dominated by a single &#8220;General Capability&#8221; dimension.</p><p>In this post, I&#8217;ll describe the statistics of this dataset, look into what&#8217;s left when you factor out this dominant dimension (hint: it&#8217;s &#8220;Claudiness&#8221;), and discuss how this relates to an important key question about cross-task generalization.</p><h2>Benchmarking data is dominated by a single underlying dimension</h2><p>This is one of the lessons of our recent work on the <a href="https://epoch.ai/benchmarks/eci">Epoch Capabilities Index</a> (ECI), which combines thirty-nine benchmarks into a single capabilities score. If benchmarks were generally uncorrelated with each other, you&#8217;d expect to see large residuals: the benchmark scores predicted by a model&#8217;s ECI number wouldn&#8217;t match the model&#8217;s actual benchmark scores. As it turns out, we see a very good match. In other words, our nominally high-dimensional dataset is well-approximated by just a single dimension.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!i7Ha!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9dce396d-38f9-43f0-95ca-da6c5844bb64_1023x1279.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!i7Ha!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9dce396d-38f9-43f0-95ca-da6c5844bb64_1023x1279.png 424w, https://substackcdn.com/image/fetch/$s_!i7Ha!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9dce396d-38f9-43f0-95ca-da6c5844bb64_1023x1279.png 848w, https://substackcdn.com/image/fetch/$s_!i7Ha!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9dce396d-38f9-43f0-95ca-da6c5844bb64_1023x1279.png 1272w, https://substackcdn.com/image/fetch/$s_!i7Ha!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9dce396d-38f9-43f0-95ca-da6c5844bb64_1023x1279.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!i7Ha!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9dce396d-38f9-43f0-95ca-da6c5844bb64_1023x1279.png" width="1023" height="1279" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9dce396d-38f9-43f0-95ca-da6c5844bb64_1023x1279.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1279,&quot;width&quot;:1023,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:210093,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://epochai.substack.com/i/179472410?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9dce396d-38f9-43f0-95ca-da6c5844bb64_1023x1279.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!i7Ha!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9dce396d-38f9-43f0-95ca-da6c5844bb64_1023x1279.png 424w, https://substackcdn.com/image/fetch/$s_!i7Ha!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9dce396d-38f9-43f0-95ca-da6c5844bb64_1023x1279.png 848w, https://substackcdn.com/image/fetch/$s_!i7Ha!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9dce396d-38f9-43f0-95ca-da6c5844bb64_1023x1279.png 1272w, https://substackcdn.com/image/fetch/$s_!i7Ha!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9dce396d-38f9-43f0-95ca-da6c5844bb64_1023x1279.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>To look beyond this dimension, we can do a <a href="https://en.wikipedia.org/wiki/Principal_component_analysis">Principal Component Analysis</a> (PCA). This basically asks: if we make synthetic &#8220;components&#8221; by taking weighted sums of the different benchmark scores, what&#8217;s the most variance in the dataset we can account for with the fewest number of components?</p><p>When we do this on the raw data underlying ECI, the first component captures about half the variance of the dataset.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> The table below shows the weights on the different benchmarks in this component, accounting for 80% of the total weight. Note that the weights are all positive and not very dispersed. That is, PCA also finds a single &#8220;general capability&#8221; component.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!LPXS!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F281dd18e-8b93-430d-9a01-ea8b38280ea5_1434x1028.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!LPXS!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F281dd18e-8b93-430d-9a01-ea8b38280ea5_1434x1028.png 424w, https://substackcdn.com/image/fetch/$s_!LPXS!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F281dd18e-8b93-430d-9a01-ea8b38280ea5_1434x1028.png 848w, https://substackcdn.com/image/fetch/$s_!LPXS!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F281dd18e-8b93-430d-9a01-ea8b38280ea5_1434x1028.png 1272w, https://substackcdn.com/image/fetch/$s_!LPXS!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F281dd18e-8b93-430d-9a01-ea8b38280ea5_1434x1028.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!LPXS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F281dd18e-8b93-430d-9a01-ea8b38280ea5_1434x1028.png" width="1434" height="1028" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/281dd18e-8b93-430d-9a01-ea8b38280ea5_1434x1028.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1028,&quot;width&quot;:1434,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:122616,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://epochai.substack.com/i/179472410?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F281dd18e-8b93-430d-9a01-ea8b38280ea5_1434x1028.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!LPXS!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F281dd18e-8b93-430d-9a01-ea8b38280ea5_1434x1028.png 424w, https://substackcdn.com/image/fetch/$s_!LPXS!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F281dd18e-8b93-430d-9a01-ea8b38280ea5_1434x1028.png 848w, https://substackcdn.com/image/fetch/$s_!LPXS!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F281dd18e-8b93-430d-9a01-ea8b38280ea5_1434x1028.png 1272w, https://substackcdn.com/image/fetch/$s_!LPXS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F281dd18e-8b93-430d-9a01-ea8b38280ea5_1434x1028.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Moving beyond the first principal component, the chart below shows the magnitudes of all the principal components, plotted against the size of components found in randomly generated data of the same shape (i.e., a <a href="https://en.wikipedia.org/wiki/Parallel_analysis">parallel analysis</a>). We see the single large component mentioned above, a second component that is borderline significant, and the rest having small sizes, consistent with noise.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!I2V9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a0135cf-34f6-4ab9-9a6d-a54d906d4805_1023x1279.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!I2V9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a0135cf-34f6-4ab9-9a6d-a54d906d4805_1023x1279.png 424w, https://substackcdn.com/image/fetch/$s_!I2V9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a0135cf-34f6-4ab9-9a6d-a54d906d4805_1023x1279.png 848w, https://substackcdn.com/image/fetch/$s_!I2V9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a0135cf-34f6-4ab9-9a6d-a54d906d4805_1023x1279.png 1272w, https://substackcdn.com/image/fetch/$s_!I2V9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a0135cf-34f6-4ab9-9a6d-a54d906d4805_1023x1279.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!I2V9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a0135cf-34f6-4ab9-9a6d-a54d906d4805_1023x1279.png" width="1023" height="1279" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1a0135cf-34f6-4ab9-9a6d-a54d906d4805_1023x1279.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1279,&quot;width&quot;:1023,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:81938,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://epochai.substack.com/i/179472410?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a0135cf-34f6-4ab9-9a6d-a54d906d4805_1023x1279.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!I2V9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a0135cf-34f6-4ab9-9a6d-a54d906d4805_1023x1279.png 424w, https://substackcdn.com/image/fetch/$s_!I2V9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a0135cf-34f6-4ab9-9a6d-a54d906d4805_1023x1279.png 848w, https://substackcdn.com/image/fetch/$s_!I2V9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a0135cf-34f6-4ab9-9a6d-a54d906d4805_1023x1279.png 1272w, https://substackcdn.com/image/fetch/$s_!I2V9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a0135cf-34f6-4ab9-9a6d-a54d906d4805_1023x1279.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>Benchmarking data shows a smaller &#8220;Claudiness&#8221; dimension</h2><p>What is this second component? Here are the top weights, by absolute value, that this component assigns to the benchmarks, again accounting for over 80% of the total weight. By construction, this component is orthogonal to the main &#8220;general capability&#8221; component. When I first saw this, I said it looked something like, &#8220;good at agentic tasks, but bad at vision&#8230; and also bad at math?&#8221;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!FokV!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d17195f-eff6-461f-b73f-4382efea0117_1450x660.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!FokV!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d17195f-eff6-461f-b73f-4382efea0117_1450x660.png 424w, https://substackcdn.com/image/fetch/$s_!FokV!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d17195f-eff6-461f-b73f-4382efea0117_1450x660.png 848w, https://substackcdn.com/image/fetch/$s_!FokV!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d17195f-eff6-461f-b73f-4382efea0117_1450x660.png 1272w, https://substackcdn.com/image/fetch/$s_!FokV!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d17195f-eff6-461f-b73f-4382efea0117_1450x660.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!FokV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d17195f-eff6-461f-b73f-4382efea0117_1450x660.png" width="1450" height="660" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7d17195f-eff6-461f-b73f-4382efea0117_1450x660.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:660,&quot;width&quot;:1450,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:80424,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://epochai.substack.com/i/179472410?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d17195f-eff6-461f-b73f-4382efea0117_1450x660.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!FokV!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d17195f-eff6-461f-b73f-4382efea0117_1450x660.png 424w, https://substackcdn.com/image/fetch/$s_!FokV!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d17195f-eff6-461f-b73f-4382efea0117_1450x660.png 848w, https://substackcdn.com/image/fetch/$s_!FokV!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d17195f-eff6-461f-b73f-4382efea0117_1450x660.png 1272w, https://substackcdn.com/image/fetch/$s_!FokV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d17195f-eff6-461f-b73f-4382efea0117_1450x660.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>But I showed it to a colleague and he just said, &#8220;it&#8217;s Claude&#8221;. He was right. Here are the top five models on this component, as well as the bottom five.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!2wdx!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60f2d973-3675-4bc1-bf80-922134159cfa_1514x486.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!2wdx!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60f2d973-3675-4bc1-bf80-922134159cfa_1514x486.png 424w, https://substackcdn.com/image/fetch/$s_!2wdx!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60f2d973-3675-4bc1-bf80-922134159cfa_1514x486.png 848w, https://substackcdn.com/image/fetch/$s_!2wdx!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60f2d973-3675-4bc1-bf80-922134159cfa_1514x486.png 1272w, https://substackcdn.com/image/fetch/$s_!2wdx!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60f2d973-3675-4bc1-bf80-922134159cfa_1514x486.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!2wdx!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60f2d973-3675-4bc1-bf80-922134159cfa_1514x486.png" width="1456" height="467" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/60f2d973-3675-4bc1-bf80-922134159cfa_1514x486.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:467,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:76528,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://epochai.substack.com/i/179472410?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60f2d973-3675-4bc1-bf80-922134159cfa_1514x486.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!2wdx!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60f2d973-3675-4bc1-bf80-922134159cfa_1514x486.png 424w, https://substackcdn.com/image/fetch/$s_!2wdx!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60f2d973-3675-4bc1-bf80-922134159cfa_1514x486.png 848w, https://substackcdn.com/image/fetch/$s_!2wdx!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60f2d973-3675-4bc1-bf80-922134159cfa_1514x486.png 1272w, https://substackcdn.com/image/fetch/$s_!2wdx!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F60f2d973-3675-4bc1-bf80-922134159cfa_1514x486.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>I think this second component shows that benchmarks aren&#8217;t <em>entirely</em> &#8220;general capability&#8221; plus &#8220;noise&#8221;, even if that is a pretty good approximation. Even though this component isn&#8217;t so statistically significant, I think it&#8217;s fair to say that it aligns with the general public sense of Anthropic&#8217;s priorities, i.e. they seem to be making Claude like this on purpose. This updated my thinking a bit on a broader question, as I&#8217;ll explain next.</p><h2>Is the &#8220;general capability&#8221; dimension deep, or contingent?</h2><p>The big question here is <em>why</em> a single dimension captures so much of the variance in benchmark scores. I can think of two possible reasons, corresponding to two possible worlds we may be in. I&#8217;ll call these worlds &#8220;deep&#8221; and &#8220;contingent&#8221;.</p><p>In the &#8220;deep&#8221; world, there is a single underlying ability that governs how well models do at superficially unrelated tasks. In this world, the only thing a model developer can do is make this ability go up. If they succeed, their model gets better at everything.</p><p>In the &#8220;contingent&#8221; world, there are many orthogonal abilities that models can have. These are orthogonal in the sense that model developers have to do completely unrelated work to get a model to improve on each ability. Still, in the world I&#8217;m imagining, customers demand models with many capabilities, and so developers put in the work to make this happen.</p><p>Which world more resembles our own? Sometimes in the history of AI, things have looked like the contingent world. AlphaGo was superhuman at go but it was nonsense to ask it to do anything else. At other times, things have looked like the deep world. When LLMs were picking up steam, next-token prediction on relatively uncurated web text was tearing through NLP tasks that had previously been dominated by specialized models.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a></p><p>To a first approximation, benchmark scores look the same in both worlds. But the existence of the Claudiness dimension feels to me like a bit of evidence for the &#8220;contingent&#8221; world. Anthropic has focused on making models that are state-of-the-art at agentic coding. Without additional focused investment, the models turn out not to be exceptional at advanced math. There is surely some generalization across tasks, but perhaps this is a sign of its limits.</p><h2>A trillion dollar question</h2><p>The Claudiness dimension is not very strong evidence for the contingent view. Stronger evidence might be how model developers are investing heavily in collecting specialized data, like reinforcement learning (RL) environments for industry verticals. Even so, it&#8217;s possible that they&#8217;re doing that <em>and</em> that RL shows excellent cross-task generalization.</p><p>One way to test this would be to find an uncontaminated benchmark that measures something unusual, and see if it correlates with the &#8220;General Capability&#8221; dimension. Unfortunately, we don&#8217;t know what counts as &#8220;unusual&#8221; for models because we don&#8217;t know what they saw in training. Also, I suspect there&#8217;s a selection effect where benchmarks that show top models scoring poorly tend to capture attention. Still, this seems worth pursuing.</p><p>Even if the explanation for what we see in benchmark data is that model developers are pursuing an &#8220;everything at once&#8221; strategy, they have the resources and the scalable architectures necessary to keep it going. In other words, they can keep making all the benchmarks go up so long as they can get the right training data plus enough compute to make use of it.</p><p>What does this mean for the future? I like how <a href="https://x.com/snewmanpv/status/1990193674161189009">Steve Newman put it recently</a>: <em>&#8220;how far can you get by simply putting an insane number of things in distribution&#8221; is one of the trillion dollar questions.</em></p><p>I doubt that there are in-principle limits to putting everything in distribution, but if we&#8217;re more in the contingent world then we shouldn&#8217;t expect much of a tailwind from generalization either. Every percentage point of improvement on every benchmark must be paid for. Here I think we should expect to see capabilities continue to improve quite generally, but only so long as the flywheel of growth and investment continues to allow developers to devote resources to actively making this happen.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://epochai.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading! Subscribe to receive the latest from Epoch AI.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2></h2><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>Methodology: we filter our dataset to benchmarks created in 2023 and beyond, and models with at least 8 benchmark scores. We combine different reasoning settings for the same model, taking the max scores. We use k-nearest neighbors to impute missing data, transforming [0-1] scores by a logit first, weighting by distance, and using 5 neighbors. We then do PCA. Data and code can be found <a href="https://colab.research.google.com/drive/1sV1TmAxclpb6rwrRSQf-g-vGyP6uE6ml#scrollTo=_ZS8B7Ba5d4K">here</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>This main finding accords with <a href="https://arxiv.org/abs/2405.10938">previous work</a>, although we now have a much larger dataset of benchmarks.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>Even now there are some prominent specialized models, like Cursor&#8217;s Tab or OpenAI&#8217;s Codex series. But it seems fair to characterize the current landscape as dominated by models that at least try to &#8220;do it all&#8221;.</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[The software intelligence explosion debate needs experiments]]></title><description><![CDATA[The existing debate rests on data and assumptions that are shakier than most people realize. To make progress, we need better evidence, and experiments are the best way to get it on the margin.]]></description><link>https://epochai.substack.com/p/the-software-intelligence-explosion</link><guid isPermaLink="false">https://epochai.substack.com/p/the-software-intelligence-explosion</guid><dc:creator><![CDATA[Anson Ho]]></dc:creator><pubDate>Sat, 15 Nov 2025 01:07:35 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!SbjF!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f248a1d-1d9f-4bab-923a-4b7e4b0ccc7e_1024x1280.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Originally posted on <a href="https://epoch.ai/gradient-updates/the-software-intelligence-explosion-debate-needs-experiments">Epoch AI</a>.</em></p><div><hr></div><p>Suppose you had a million AIs, each surpassing humanity&#8217;s best AI researchers. If they all worked on advancing AI, how much would AI progress accelerate?</p><p>This might sound like science fiction, but it may be the most consequential question about the future of AI. The problem is that the experts disagree wildly on the answer.</p><p><a href="https://ai-2027.com/research/takeoff-forecast">Some</a> <a href="https://www.forethought.org/research/will-ai-r-and-d-automation-cause-a-software-intelligence-explosion">foresee</a> <a href="https://intelligence.org/files/IEM.pdf">a</a> positive feedback loop. These AIs are smart enough to find new algorithms to make smarter AIs, which make <em>even</em> smarter AIs, and so on. Very soon, we could see <a href="https://www.forethought.org/research/how-quick-and-big-would-a-software-intelligence-explosion-be">multiple years of AI progress</a> compressed into a single year just through software advances &#8212; a &#8220;software intelligence explosion&#8221;.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a></p><p><a href="https://epoch.ai/gradient-updates/most-ai-value-will-come-from-broad-automation-not-from-r-d">Others</a> <a href="https://epoch.ai/gradient-updates/the-case-for-multi-decade-ai-timelines">agree</a> that AI progress would speed up, but think that something will block the explosive feedback loop. For example, increasing difficulty in finding new algorithms might bottleneck AI self-improvement, or software improvements might depend heavily on physical resources like compute, which can&#8217;t be scaled as easily.</p><p>And we really need to know who&#8217;s right. If a software intelligence explosion is likely, economic and military power might become much <a href="https://situational-awareness.ai/from-agi-to-superintelligence/">more</a> <a href="https://www.forethought.org/research/ai-enabled-coups-how-a-small-group-could-use-ai-to-seize-power">concentrated</a> in a few companies or nations. They might also focus more on automating AI research rather than driving <a href="https://epoch.ai/gradient-updates/most-ai-value-will-come-from-broad-automation-not-from-r-d">broad economic impacts</a> &#8212; so AI progress could rapidly accelerate before most of the world knows what&#8217;s happening.</p><p>So why do the experts disagree so much, and how can we figure out who&#8217;s right?</p><h2>Flawed data</h2><p>The core reason for this disagreement is a lack of strong evidence &#8212; thus far, empirical work on the software intelligence explosion has used flawed data and models.</p><p>Prior work has used the &#8220;<a href="https://en.wikipedia.org/wiki/Jones_model">Jones model</a>&#8221;, which tells us how R&amp;D inputs translate into outputs. For example, the R&amp;D input might be the number of AI researchers, and the R&amp;D output could measure how efficiently an AI algorithm uses training compute (&#8220;software efficiency&#8221;).<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a></p><p>The core parameter in this model is the &#8220;returns to R&amp;D&#8221; &#8212; in our example, this tells us how software efficiency changes when we double the number of researchers. If it also doubles, the returns are just 1. If it quadruples, then the returns are 2. In general, if doubling the inputs increases the output by 2<sup>r</sup>, then the returns are <em>r</em>.</p><p>This parameter dictates the long-run dynamics of the software intelligence explosion.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> We especially care if we have increasing returns (i.e.<em> r &gt; 1</em>) &#8212; this is when you get more than what you put in, resulting in the positive feedback loop where smarter AIs recursively build smarter AIs.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a> <a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a></p><p>So far, so good. But the issues show up when we apply this in practice. Consider our estimates of these returns in several domains of AI software:<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!SbjF!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f248a1d-1d9f-4bab-923a-4b7e4b0ccc7e_1024x1280.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!SbjF!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f248a1d-1d9f-4bab-923a-4b7e4b0ccc7e_1024x1280.png 424w, https://substackcdn.com/image/fetch/$s_!SbjF!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f248a1d-1d9f-4bab-923a-4b7e4b0ccc7e_1024x1280.png 848w, https://substackcdn.com/image/fetch/$s_!SbjF!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f248a1d-1d9f-4bab-923a-4b7e4b0ccc7e_1024x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!SbjF!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f248a1d-1d9f-4bab-923a-4b7e4b0ccc7e_1024x1280.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!SbjF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f248a1d-1d9f-4bab-923a-4b7e4b0ccc7e_1024x1280.png" width="1024" height="1280" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8f248a1d-1d9f-4bab-923a-4b7e4b0ccc7e_1024x1280.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1280,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!SbjF!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f248a1d-1d9f-4bab-923a-4b7e4b0ccc7e_1024x1280.png 424w, https://substackcdn.com/image/fetch/$s_!SbjF!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f248a1d-1d9f-4bab-923a-4b7e4b0ccc7e_1024x1280.png 848w, https://substackcdn.com/image/fetch/$s_!SbjF!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f248a1d-1d9f-4bab-923a-4b7e4b0ccc7e_1024x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!SbjF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f248a1d-1d9f-4bab-923a-4b7e4b0ccc7e_1024x1280.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Estimates of the returns to software R&amp;D for three different domains of AI: (1) computer vision, (2) reinforcement learning, and (3) language models. Technical details are provided in the appendix.</em></figcaption></figure></div><p>Here we see a lot of probability on compounding returns. But the uncertainty is very large, and our range of estimates straddle the key threshold of <em>r = 1</em>. So at face value, the software intelligence explosion seems quite plausible &#8212; at the very least, it&#8217;s hard to rule out.</p><p>But if we dig into the details, we see a ton of big problems. To estimate <em>r</em>, we need to collect data on R&amp;D inputs and outputs, but our measures of these are very imperfect.</p><p>For example, to estimate the returns in language models, we first looked at the number of published papers in natural language processing. This proxies for the number of researchers, accounting for how some researchers might be more productive than others (and hence publish more papers).</p><p>But this ignores all the R&amp;D effort that spills over from other domains of AI. For example, it&#8217;s common for language models like the <a href="https://arxiv.org/abs/1512.03385">Transformer</a> to use &#8220;residual connections&#8221;, but this innovation was heavily inspired by work in <a href="https://arxiv.org/abs/1512.03385">computer vision</a>.</p><p>So we also considered a second measure of R&amp;D inputs, which is a function of human labor and AI compute at AI labs. Unfortunately, this is flawed for a similar reason &#8212; it ignores the R&amp;D effort that spills over from researchers outside of a collection of AI companies.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a> For instance, OpenAI&#8217;s work is often <a href="https://www.latent.space/p/noam-brown#:~:text=Noam%20%5B01%3A04,inspiration%20for%20us">inspired</a> by academic research.</p><p>So these &#8220;spillover effects&#8221; cause a mismatch between R&amp;D inputs and what we <em>measure</em> it to be. But it&#8217;s not clear how best to adjust for this &#8212; we could include papers from computer vision and other domains of AI, but many of these papers will be unrelated to language model improvements! We could also try to account for academic R&amp;D inputs, but we don&#8217;t know how much academic research compute there was over time. In general, we don&#8217;t have good data on the inputs and need to rely on proxies or best-guesses with limited data.</p><p>What does this mean for our estimates of the returns to R&amp;D? Possibly this means that the &#8220;true&#8221; returns should be higher than what we observe. In particular, both the number of researchers and supply of compute probably grows more slowly in academia compared to AI labs.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-8" href="#footnote-8" target="_self">8</a> So if we broaden our metric to include academia, we&#8217;d measure slower R&amp;D input growth. This means fewer input doublings for the same number of output doublings, so the estimated returns to R&amp;D become higher.</p><p>There are similar issues with the R&amp;D outputs. Ideally we&#8217;d like to have a fine-grained measure of software quality over time, but unfortunately we only have an average growth rate. This forces us to make strong assumptions to get estimates &#8212; for instance, we assume that software progress has proceeded at a constant rate. This is plausibly <a href="https://www.lesswrong.com/posts/ssNSCaug5p3xnNDSd/how-fast-is-algorithmic-progress-in-ai-inference">consistent with</a> <a href="https://epoch.ai/blog/algorithmic-progress-in-language-models">existing empirical evidence</a>, but we think the evidence on this is quite weak. In fact, the <a href="https://epoch.ai/gradient-updates/quantifying-the-algorithmic-improvement-from-reasoning-models">rise of reasoning models</a> might&#8217;ve marked an acceleration in algorithmic progress.</p><p>In general, there are a <a href="https://arxiv.org/abs/2405.10494">whole host of data-related issues</a> that we need to contend with when estimating these returns, which might bias our estimates upward or downward.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-9" href="#footnote-9" target="_self">9</a></p><h2>Flawed models</h2><p>Even without data issues, we might still end up with misleading estimates. The problem is that we&#8217;re taking the Jones model way beyond what has been empirically observed.</p><p>The model was originally written for the whole economy, where the main input, researchers, and the main output, productivity, are growing at a few percent a year. But we&#8217;re trying to apply to a situation where the inputs are growing by orders of magnitude in a few years &#8212; i.e. the software intelligence explosion, where you get many more AI workers that become increasingly intelligent. So we shouldn&#8217;t be too surprised if the Jones law of motion is pretty badly misspecified for thinking about the software intelligence explosion.</p><p>One issue is that finding software improvements might require running <a href="https://epoch.ai/gradient-updates/most-ai-value-will-come-from-broad-automation-not-from-r-d">experiments close to the size of frontier training runs</a>. This can make a huge difference &#8212; for instance, in <a href="https://arxiv.org/abs/2507.23181">previous empirical work</a>, such compute requirements could prevent a software intelligence explosion.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-10" href="#footnote-10" target="_self">10</a> But some <a href="https://epoch.ai/blog/do-the-returns-to-software-rnd-point-towards-a-singularity">previous work</a> neglects this dynamic.</p><p>Another issue is that the Jones model <a href="https://epoch.ai/epoch-after-hours/economics-of-ai">doesn&#8217;t include a hard limit on how much research can be parallelized</a> at a given time.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-11" href="#footnote-11" target="_self">11</a> So if you scale up the number of AI researchers really fast, the model predicts that software progress should go to infinity essentially immediately. But this seems implausible &#8212; you couldn&#8217;t get infinite software progress in five minutes even if you had infinitely many Alec Radfords doing AI research. You still need to run experiments and wait for them to finish, and some software innovations might need to happen in sequence. So if we naively apply the Jones model, we might vastly overstate how much software progress we see when AI R&amp;D is automated.</p><p>Finally, when most people think of an intelligence explosion, they think of AIs getting smarter and smarter, but most models of the intelligence explosion consider making AI more and more numerous. It&#8217;s not clear that this is what people actually care about, as having infinite copies of GPT-2 would arguably have little impact on the world. The difficulty with modeling &#8220;intelligence&#8221; increases is that it&#8217;s not clear how intelligence translates into effective research input. Is being twice as smart as good as having twice as many researchers? We have fairly minimal data on this front, and adjusting for it has unclear implications. On the one hand, the returns to intelligence in R&amp;D might be extremely high, effectively increasing <em>r</em>. On the other hand, if we changed the output of R&amp;D to intelligence, this has plausibly grown much slower than effective compute, lowering <em>r</em>.</p><p>Part of the issue here is that we don&#8217;t know enough about the nature of algorithmic progress to know which model is right. This can lead us to over or underestimate the impact of things like scaling up R&amp;D investments.</p><h2>To make progress, we need experiments</h2><p>Most of the approaches at trying to get a handle on the software intelligence explosion so far have been either through theoretical modelling, conceptual arguments, or doing statistics on messy real-world data. In part, this is because few people have been thinking about the software intelligence explosion in rigorous terms. And of the people that have, these approaches better fit their skillset, or are cheaper for them to implement.</p><p>But we think we&#8217;ve squeezed out much of the value from these approaches given existing data, and on the margin there&#8217;s a better approach &#8212; experiments. This helps us get much higher quality data that more directly tests the hypothesis and removes confounding factors. It also lets us test whether key assumptions will hold to help inform how we think about the software intelligence explosion, like in terms of models.</p><p>Here are some examples that we think are especially important:</p><p><strong>Studying how much software progress is due to data</strong>. If we learn that <a href="https://www.beren.io/2025-08-02-Most-Algorithmic-Progress-is-Data-Progress/">most software improvement</a> comes from increases in data quality, that could reduce our credences in the chances of a software intelligence explosion.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-12" href="#footnote-12" target="_self">12</a> However, existing approaches are based on high-level statistical analyses of actual papers, where we can&#8217;t control which algorithms or training data are used. So it&#8217;s hard to say which innovations are most important, or how important data is. In contrast, experiments would offer a much higher degree of control.</p><p><strong>Understanding how much compute is a bottleneck to software progress. </strong>In the software intelligence explosion, we can get lots of software progress even if we don&#8217;t use a lot more compute. So we want to know how hard it is to find new algorithms that push the state of the art, even without using much compute to run experiments. However, there isn&#8217;t much real-world data that tells us how easy this is. So we want to run experiments to see whether this is possible &#8212; this might mean randomly allocating compute budgets to different researchers to see how that impacts the amount of software innovation they&#8217;re able to achieve. We might also want to perform scaling experiments to understand how well algorithms that work with small training runs to large ones &#8212; if this happens easily, compute might not be as big of a bottleneck as we think.</p><p>Importantly, just doing experiments isn&#8217;t going to be a silver bullet. We might still work with flawed models. Experiments might also be impractical &#8212; for example, if we want to test how well software innovations apply at frontier training scales, we need enough compute to do a frontier training run! That restricts the number of actors that can study this kind of question dramatically. So we&#8217;re skeptical that this will &#8220;solve&#8221; questions about the software intelligence explosion.</p><p>And there&#8217;s still some role for other approaches. For example, we might want to know which input most bottlenecks software progress &#8212; is it intelligence, compute, engineers, or something else? Besides experimental approaches, we could try and see if software progress mostly comes from large labs that are compute rich, or from the same set of researchers over and over.</p><p>But on the margin, we think experiments are the most promising avenue for improving our understanding about the software singularity. And if the software intelligence explosion is actually real, it might be especially important to gather evidence on these questions as soon as possible.</p><p><em>We&#8217;d like to thank Greg Burnham, JS Denain, Jaime Sevilla, Lynette Bye and Phil Trammell for feedback on this post.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://epochai.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading! Subscribe to receive the latest from Epoch AI.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2>Technical appendix</h2><p>In the main post we elided over many of the technical details to make our results widely accessible. But if you&#8217;re one of the more technically-minded readers of our newsletter, you&#8217;ll also be interested in our modeling and data choices, so we specify these here. Note that you might want to <a href="https://epoch.ai/gradient-updates/the-software-intelligence-explosion-debate-needs-experiments">read this on our website</a>, since mathematical symbols will be rendered more nicely.</p><h3>Estimating the returns to R&amp;D using the Jones model</h3><p>Let A denote some measure of software quality, which is improved via a Jones-style idea production function:</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\frac{dA}{dt} = A^{1-\\beta} F(L,K)^\\lambda; \\beta, \\lambda > 0&quot;,&quot;id&quot;:&quot;JUVHPBOOUL&quot;}" data-component-name="LatexBlockToDOM"></div><p><em>A<sup>1-&#946;</sup></em> represents ideas possibly getting harder to find over time, while <em>F(L,K)</em> denotes the effective research input given experimental compute <em>K</em>  and cognitive labor <em>L</em>. Suppose research is automated in such a way that <em>F(L,K)</em> shifts to be proportional to the effective number of simulated AI researchers, which is given by <em>Ac</em> for some constant <em>c</em>.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-13" href="#footnote-13" target="_self">13</a></p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\frac{dA}{dt} \\propto A^{\\lambda + 1 - \\beta} c&quot;,&quot;id&quot;:&quot;PFMXTOVQLU&quot;}" data-component-name="LatexBlockToDOM"></div><p>This model says software progress growth is hyperbolic if &#955;/&#946; &gt; 1. In the terminology of the post, <em>r = </em>&#955;/&#946; is the &#8220;returns to AI research.&#8221; Therefore, we have a software-only intelligence explosion if the returns to AI software research are sufficiently high.</p><p>In the following section, we describe how we estimate the returns to AI research for frontier language models. The key to measuring this is getting some input measure of <em>F(L,K)</em>.</p><h3>Approach 1: The number of papers</h3><p>One approach is to proxy for <em>F(L,K)</em> based on the number of papers that are produced.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-14" href="#footnote-14" target="_self">14</a> This is the same approach as used in <a href="https://epoch.ai/blog/do-the-returns-to-software-rnd-point-towards-a-singularity">previous work</a>.</p><p>We do this in domains of AI for which we have estimates of the relevant rates of software progress &#8212; namely <a href="https://epoch.ai/blog/revisiting-algorithmic-progress">computer vision</a>, <a href="https://arxiv.org/abs/2102.04881">reinforcement learning</a>, and <a href="https://epoch.ai/blog/algorithmic-progress-in-language-models">language modeling</a>. We use these estimates to define the &#8220;software quality&#8221;,<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-15" href="#footnote-15" target="_self">15</a> which we use as our output metric.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-16" href="#footnote-16" target="_self">16</a></p><p>We get data on the number of papers in different years from <a href="https://openalex.org/">OpenAlex</a>, which we identify using the following &#8220;concept&#8221; codes:</p><ul><li><p><strong>Computer vision</strong>: Computer vision (C31972630) and Deep learning (C108583219)</p></li><li><p><strong>Reinforcement learning</strong>: Reinforcement learning (C97541855)</p></li><li><p><strong>Language models</strong>: Natural language processing (C204321447) and Deep learning (C108583219)</p></li></ul><p>The resulting estimates of the parameters are shown in the table below:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!j9sJ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1259caa9-aed2-409c-904e-86e83a3def17_1624x496.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!j9sJ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1259caa9-aed2-409c-904e-86e83a3def17_1624x496.png 424w, https://substackcdn.com/image/fetch/$s_!j9sJ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1259caa9-aed2-409c-904e-86e83a3def17_1624x496.png 848w, https://substackcdn.com/image/fetch/$s_!j9sJ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1259caa9-aed2-409c-904e-86e83a3def17_1624x496.png 1272w, https://substackcdn.com/image/fetch/$s_!j9sJ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1259caa9-aed2-409c-904e-86e83a3def17_1624x496.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!j9sJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1259caa9-aed2-409c-904e-86e83a3def17_1624x496.png" width="1456" height="445" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1259caa9-aed2-409c-904e-86e83a3def17_1624x496.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:445,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:88243,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://epochai.substack.com/i/178906633?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1259caa9-aed2-409c-904e-86e83a3def17_1624x496.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!j9sJ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1259caa9-aed2-409c-904e-86e83a3def17_1624x496.png 424w, https://substackcdn.com/image/fetch/$s_!j9sJ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1259caa9-aed2-409c-904e-86e83a3def17_1624x496.png 848w, https://substackcdn.com/image/fetch/$s_!j9sJ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1259caa9-aed2-409c-904e-86e83a3def17_1624x496.png 1272w, https://substackcdn.com/image/fetch/$s_!j9sJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1259caa9-aed2-409c-904e-86e83a3def17_1624x496.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Estimates of key parameters in the Jones model, proxying R&amp;D inputs with the number of papers in certain domains. We provide median estimates, and show 90% credible intervals in brackets. Code for this analysis can be found in this <a href="https://github.com/parkerwhitfill/epoch_RRD/">github repository</a>.</em></figcaption></figure></div><p>We then estimate the returns to R&amp;D using Bayesian inference.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-17" href="#footnote-17" target="_self">17</a> Like in computer vision and reinforcement learning, this yields central estimates of <em>r </em>that exceed 1, and in fact the central estimate is higher in language models. Moreover, the 90% credible interval exceeds one as well &#8212; pointing towards a software intelligence explosion! On the other hand, all of these estimates have very large uncertainties, so the confidence intervals across these different domains overlap quite a lot.</p><h3>Approach 2: Some function of cognitive labor and experimental compute</h3><p>In reality, we might think that papers are a poor proxy for R&amp;D inputs. For example, AI R&amp;D involves lots of people training models, but the &#8220;number of published papers&#8221; doesn&#8217;t directly model these people at all.</p><p>So we consider an alternative approach that treats the R&amp;D input as some function of training compute and AI researcher labor at OpenAI from 2022-2025.</p><p>The next approach will try to directly measure <em>F(L,K)</em> instead of proxying it. Suppose that software progress is growing at a constant rate, which is consistent with the empirical evidence we have so far (see <a href="https://epoch.ai/blog/algorithmic-progress-in-language-models">here</a>, and <a href="https://www.lesswrong.com/posts/ssNSCaug5p3xnNDSd/how-fast-is-algorithmic-progress-in-ai-inference">here</a>). That implies the Jones equation is at a steady state, i.e.,</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;g_A = \\frac{\\lambda}{\\beta} [\\epsilon_K g_K + (1 - \\epsilon_K) g_L]&quot;,&quot;id&quot;:&quot;LXIXYYWMGN&quot;}" data-component-name="LatexBlockToDOM"></div><p>In this equation, <em>g<sub>A</sub></em>, <em>g<sub>K</sub></em>, and <em>g<sub>L</sub></em> denote the growth rates of <em>A</em>, <em>K</em>, and <em>L</em> respectively. <em>&#1013;<sub>K</sub></em> denotes the elasticity of <em>F</em> with respect to <em>K</em> &#8212; so if you increase <em>K</em> by 1%, <em>F</em> increases by <em>&#1013;<sub>K</sub></em> percent. Re-arranging this equation results in the following:</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;\\frac{\\lambda}{\\beta} = \\frac{g_A}{\\epsilon_K g_K + (1-\\epsilon_K) g_L}&quot;,&quot;id&quot;:&quot;NBTWHMKRMY&quot;}" data-component-name="LatexBlockToDOM"></div><p>Intuitively, this equation says that &#955;/&#946; is equal to the ratio of research output divided by research inputs. We can now estimate the terms on the right-hand side by considering frontier AI labs after November 2022, when ChatGPT was first released. Note that we give growth rates in base <em>e</em>.</p><ul><li><p><em>g<sub>K</sub></em> &#8776; 1.3 from <a href="https://epoch.ai/data/ai-companies">data</a> on OpenAI&#8217;s annual compute spend</p></li><li><p><em>g<sub>L</sub></em> &#8776; 0.85 from <a href="https://epoch.ai/data/ai-companies">data</a> on OpenAI&#8217;s staff counts over time</p></li><li><p><em>&#1013;<sub>K</sub></em> &#8776; 0.67 based on the fact that if markets are competitive, the elasticity of output with respect to capital should equal the compute share, the share of research money spent on compute instead of researchers. We guess the compute share is about 0.59 to 0.75.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-18" href="#footnote-18" target="_self">18</a> As a central estimate, we take the average of 0.67.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-19" href="#footnote-19" target="_self">19</a></p></li><li><p><em>g<sub>A</sub></em> &#8776; 1.1 based on <a href="https://epoch.ai/blog/algorithmic-progress-in-language-models">estimates</a> of training compute efficiency improvements in language models.</p></li></ul><p>The nice thing about this approach is that it doesn&#8217;t rely on the weak proxy of the &#8220;number of papers&#8221;. But it instead requires making an additional assumption, which is that software efficiency has grown at a constant rate.</p><h3>Compute bottlenecks</h3><p>One issue with the Jones model, pointed out in <a href="https://arxiv.org/abs/2507.23181">Whitfill and Wu 2025</a> and in <a href="https://epoch.ai/gradient-updates/most-ai-value-will-come-from-broad-automation-not-from-r-d">Erdil and Barnett 2025</a>, is that there may be compute bottlenecks.</p><p>Rigorously, if <em>F(L,K)</em> is a production function such that both inputs are necessary for sustained growth, then even if <em>L</em> is increasing due to more simulated AIs, if experimental compute <em>K</em> is not growing quickly, then <em>F(L,K)</em> becomes proportional to experimental compute, not proportional to the number of simulated AIs, which breaks the argument for an intelligence explosion no matter what value &#955;/&#946; is.</p><p>One baseline case to consider is a Cobb-Douglas baseline:</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;F(L,K) = L^{1 - \\epsilon_K} K^{\\epsilon_K}&quot;,&quot;id&quot;:&quot;KAEWQQQQKY&quot;}" data-component-name="LatexBlockToDOM"></div><p>Here, the lack of <em>K</em> doesn&#8217;t fully bottleneck progress as you can always still increase <em>L</em> to get more output, but there is some penalty for increasing only <em>L</em> instead of both <em>L </em>and <em>K</em>. Under this assumption, the key intelligence explosion parameter becomes (1 - <em>&#1013;<sub>K</sub></em>) &#955;/&#946;.</p><p>Given our estimate that <em>&#1013;<sub>K</sub> </em>&#8776; 2/3, that means all our estimates of &#955;/&#946; should be cut by a factor of three, which puts them all below 1.</p><h3>Parallelizability in the Jones model</h3><p>To address the issue of parallelizability in the Jones model, Phil Trammell <a href="https://drive.google.com/file/d/1IM3plm5m5FaGaR-NLdhtIjSMa2b0Gs-y/view">proposes an alternative model</a> where the impact of R&amp;D inputs is mediated by the current level of software quality. This is shown in the equation below, where <em>A</em> is software quality and <em>R</em> is research input:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!MopU!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5b1a2e3-1ced-4ebd-afdd-97cf8940d9af_750x270.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!MopU!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5b1a2e3-1ced-4ebd-afdd-97cf8940d9af_750x270.png 424w, https://substackcdn.com/image/fetch/$s_!MopU!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5b1a2e3-1ced-4ebd-afdd-97cf8940d9af_750x270.png 848w, https://substackcdn.com/image/fetch/$s_!MopU!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5b1a2e3-1ced-4ebd-afdd-97cf8940d9af_750x270.png 1272w, https://substackcdn.com/image/fetch/$s_!MopU!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5b1a2e3-1ced-4ebd-afdd-97cf8940d9af_750x270.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!MopU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5b1a2e3-1ced-4ebd-afdd-97cf8940d9af_750x270.png" width="750" height="270" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d5b1a2e3-1ced-4ebd-afdd-97cf8940d9af_750x270.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:270,&quot;width&quot;:750,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!MopU!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5b1a2e3-1ced-4ebd-afdd-97cf8940d9af_750x270.png 424w, https://substackcdn.com/image/fetch/$s_!MopU!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5b1a2e3-1ced-4ebd-afdd-97cf8940d9af_750x270.png 848w, https://substackcdn.com/image/fetch/$s_!MopU!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5b1a2e3-1ced-4ebd-afdd-97cf8940d9af_750x270.png 1272w, https://substackcdn.com/image/fetch/$s_!MopU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5b1a2e3-1ced-4ebd-afdd-97cf8940d9af_750x270.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Here the CES function between <em>A</em> and <em>R</em> on the right-hand side prevents you from just being able to immediately increase <em>A</em> by a ton just by increasing <em>R</em> by a ton, because &#961; &lt; 0. But while this deals with the parallelizability issue, it unfortunately also makes things harder to estimate!</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>This is sometimes called the &#8220;<a href="https://epoch.ai/gradient-updates/the-case-for-multi-decade-ai-timelines">software-only singularity</a>&#8221;.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>We describe the full model in the technical appendix.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>Things may be more complicated in the short-run &#8212; this parameter primarily determines the long-run asymptotic dynamics of AI R&amp;D, rather than the short-run.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>If <em>r &lt; 1</em>, this feedback loop fizzles out. If <em>r = 1</em>, we&#8217;re in the special case where growth in inputs results in proportional growth in outputs &#8211; but that also means that we don&#8217;t get a massive &#8220;intelligence explosion&#8221;.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>This doesn&#8217;t mean that AIs only improve AIs when there&#8217;s a software intelligence explosion. This happens regardless &#8212; the question is whether these improvements happen fast enough to lead to an &#8220;explosion&#8221;, as opposed to &#8220;fizzling out&#8221;.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>Technical details on data and the model are provided in the technical appendix.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>One counterargument is that it&#8217;s possible that algorithmic innovations proposed outside of major AI companies tend to be much less useful, because it&#8217;s not clear that they&#8217;ll work at the compute scale of frontier AI models.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-8" href="#footnote-anchor-8" class="footnote-number" contenteditable="false" target="_self">8</a><div class="footnote-content"><p>To get a sense of the numbers, we know that OpenAI has <a href="https://epoch.ai/data/ai-companies">more than doubled</a> their compute stock in the last year, and AI companies are <a href="https://epoch.ai/data/ai-companies">almost doubling</a> their staff headcounts each year. Relatedly, we also know that private sector companies own an<a href="https://epoch.ai/data-insights/ai-supercomputers-performance-share-by-sector"> increasing share of GPU clusters</a>, and the academic share of notable machine learning models has <a href="https://arxiv.org/abs/2401.02452">decreased over time</a>. So if we look at the aggregate of OpenAI and academia, we&#8217;d measure slower growth in both labor and compute, and thus overall R&amp;D inputs.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-9" href="#footnote-anchor-9" class="footnote-number" contenteditable="false" target="_self">9</a><div class="footnote-content"><p>Moreover, the input and output metrics are themselves often extremely uncertain. For instance, existing <a href="https://epoch.ai/trends">estimates of software progress</a> are themselves based on observational estimates. Prior estimates of the returns to R&amp;D are also unrealistic &#8212; for example, the exponent on research inputs in the Jones law of motion is greater than 1, so doubling researchers could more than double software efficiency! Part of the issue here is that these individual exponents cannot be estimated well due to data constraints, but the hope is that the overall returns can still be estimated reasonably well.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-10" href="#footnote-anchor-10" class="footnote-number" contenteditable="false" target="_self">10</a><div class="footnote-content"><p>In particular, <a href="https://arxiv.org/abs/2507.23181">Whitfill and Wu 2025</a> estimate whether compute and labor are complements or substitutes. If you don&#8217;t account for the dynamic where frontier scale is necessary, you find that compute and labor are substitutes. But if you do account for this you find that they&#8217;re very strong complements!</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-11" href="#footnote-anchor-11" class="footnote-number" contenteditable="false" target="_self">11</a><div class="footnote-content"><p>Yet another issue is that there&#8217;s no explicit data in the model proxying for research inputs based on compute and labor. In practice, developing better capabilities also depends on data.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-12" href="#footnote-anchor-12" class="footnote-number" contenteditable="false" target="_self">12</a><div class="footnote-content"><p>On the other hand, AI systems could improve data quality through things like verification and filtering, generating more diverse problems, or helping build high-quality RL environments. So even if most software improvements boil down to data quality, we could conceivably see a software intelligence explosion.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-13" href="#footnote-anchor-13" class="footnote-number" contenteditable="false" target="_self">13</a><div class="footnote-content"><p>If <em>A</em> is inference compute efficiency, then this equation holds exactly. If <em>A</em> is training compute efficiency, then this equation only holds if the effective number of AI researchers (because they are getting smarter) is linear in training compute.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-14" href="#footnote-anchor-14" class="footnote-number" contenteditable="false" target="_self">14</a><div class="footnote-content"><p>This data is taken from <a href="https://openalex.org/">OpenAlex</a>, based on papers grouped into both &#8220;Natural Language Processing&#8221; and &#8220;Deep learning&#8221;.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-15" href="#footnote-anchor-15" class="footnote-number" contenteditable="false" target="_self">15</a><div class="footnote-content"><p>Software quality is tricky to define, but can be defined with reference to algorithms of previous years. For example, suppose we choose 2024 as our start year, where software quality is defined as 1. Compared to models in 2024, models in 2025 might need around <a href="https://epoch.ai/blog/algorithmic-progress-in-language-models">three times</a> less training compute to achieve the same performance. So we then say that algorithms in 2025 have three times higher software quality than algorithms in 2024. Note that this <a href="https://arxiv.org/abs/2508.11033">dubiously assumes</a> that all software (algorithms and data) in a particular year have the same quality.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-16" href="#footnote-anchor-16" class="footnote-number" contenteditable="false" target="_self">16</a><div class="footnote-content"><p>Strictly speaking, these works provide a single growth rate of the R&amp;D output metric, rather than the output metric itself. Other domains like computer chess have much higher quality R&amp;D output data, since we can look at chess engine ELO scores. This gives a time series rather than a single average trend.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-17" href="#footnote-anchor-17" class="footnote-number" contenteditable="false" target="_self">17</a><div class="footnote-content"><p>This approach is detailed in <a href="https://arxiv.org/abs/2405.10494">Erdil, Besiroglu and Ho 2024</a>. Based on the paper, we select relatively uninformative priors to avoid biasing results, where exponential growth is the typical outcome. We need to use this approach rather than things like maximum likelihood estimation because we only have a single trend over time, rather than an actual time series.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-18" href="#footnote-anchor-18" class="footnote-number" contenteditable="false" target="_self">18</a><div class="footnote-content"><p>We take this estimate from OpenAI&#8217;s <a href="https://www.theinformation.com/articles/openai-projections-imply-losses-tripling-to-14-billion-in-2026?rc=spkbjw">2024 budget</a>, which shows 700 million spent on salaries and 1 billion on research compute, giving a na&#239;ve capital share of 59%. If we do a more sophisticated analysis that adjusts for spend on research staff only, include <a href="https://epoch.ai/data-insights/openai-compute-spend">equity and a broader set of R&amp;D,</a> then we get about 75%.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-19" href="#footnote-anchor-19" class="footnote-number" contenteditable="false" target="_self">19</a><div class="footnote-content"><p>Note this term is not necessarily constant over time, so consider it a single snapshot.</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[Less than 70% of FrontierMath is within reach for today’s models]]></title><description><![CDATA[57% of problems have been solved at least once]]></description><link>https://epochai.substack.com/p/less-than-70-of-frontiermath-is-within</link><guid isPermaLink="false">https://epochai.substack.com/p/less-than-70-of-frontiermath-is-within</guid><dc:creator><![CDATA[Greg Burnham]]></dc:creator><pubDate>Fri, 17 Oct 2025 16:53:02 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!gGhh!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F46792651-ef3b-4baf-bc9f-6fdf3e55658c_1024x1280.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Originally posted on <a href="https://epoch.ai/gradient-updates/less-than-70-percent-of-frontiermath-is-within-reach-for-todays-models">Epoch AI</a>.</em></p><div><hr></div><p>The best we have seen a model perform on a single run of FrontierMath is <a href="https://epoch.ai/blog/deep-think-math">29%</a>.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> If you want to use a model to solve FrontierMath-style problems, that&#8217;s the number to consider.</p><p>But there&#8217;s another way to gauge state-of-the-art performance: how many FrontierMath problems have been solved by any model, on any run, even once? This tells us more about what is &#8220;within reach&#8221; for today&#8217;s models. It&#8217;s also more forward-looking: if today&#8217;s models can generate the right ideas to solve a problem <em>at all</em>, then that makes it more likely that tomorrow&#8217;s models will be able to solve the problem reliably.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> In other words, we can get a view of the future that&#8217;s a bit more concrete than just extrapolating accuracy trends.</p><p>To make matters more interesting, there&#8217;s some <a href="https://arxiv.org/pdf/2407.21787">empirical evidence</a> that if you run an LLM on a benchmark N times, the percentage of problems correctly solved at least once (known as pass@N) increases proportionally to log(N). If that&#8217;s true in general, then, since log(N) is unbounded, we should expect pass@N to approach 100% as the models are given more tries. Could FrontierMath&#8217;s saturation already be so clearly within sight?</p><p>To investigate this, I conducted 32 runs of GPT-5 (medium).<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a> In short: FrontierMath isn&#8217;t dead yet. Pass@N on these runs show <em>sub</em>-logarithmic growth and appears to cap out below 50%.</p><p>That&#8217;s just one model, though: we have a bunch of other model runs sitting around that seem just as relevant for understanding what&#8217;s likely to be reliably solvable soon. Throwing them all together gives a <em>pass@the-kitchen-sink</em> of 57% &#8212; that is, how many problems were solved by any model on any run. Moving into guesswork territory, we estimate that even repeatedly running all of these models would cap out below 70%.</p><p>Going forward, it will be interesting to see how much progress on FrontierMath accuracy comes from shoring up reliability on this 57% of problems vs. solving problems models haven&#8217;t solved before.</p><h2>GPT-5&#8217;s pass@N caps out below 50%</h2><p>The chart below shows pass@N across our 32 runs of GPT-5 on our scaffold. Note how the pass@N curve is concave down compared to the straight line of a logarithmic fit.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!gGhh!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F46792651-ef3b-4baf-bc9f-6fdf3e55658c_1024x1280.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!gGhh!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F46792651-ef3b-4baf-bc9f-6fdf3e55658c_1024x1280.png 424w, https://substackcdn.com/image/fetch/$s_!gGhh!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F46792651-ef3b-4baf-bc9f-6fdf3e55658c_1024x1280.png 848w, https://substackcdn.com/image/fetch/$s_!gGhh!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F46792651-ef3b-4baf-bc9f-6fdf3e55658c_1024x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!gGhh!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F46792651-ef3b-4baf-bc9f-6fdf3e55658c_1024x1280.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!gGhh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F46792651-ef3b-4baf-bc9f-6fdf3e55658c_1024x1280.png" width="1024" height="1280" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/46792651-ef3b-4baf-bc9f-6fdf3e55658c_1024x1280.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1280,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:63606,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://epochai.substack.com/i/176416854?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F46792651-ef3b-4baf-bc9f-6fdf3e55658c_1024x1280.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!gGhh!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F46792651-ef3b-4baf-bc9f-6fdf3e55658c_1024x1280.png 424w, https://substackcdn.com/image/fetch/$s_!gGhh!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F46792651-ef3b-4baf-bc9f-6fdf3e55658c_1024x1280.png 848w, https://substackcdn.com/image/fetch/$s_!gGhh!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F46792651-ef3b-4baf-bc9f-6fdf3e55658c_1024x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!gGhh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F46792651-ef3b-4baf-bc9f-6fdf3e55658c_1024x1280.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Below is the same data in a table. Each doubling of N increases pass@N, but by a smaller amount than the previous doubling. At the extremes, doubling N from one to two increases pass@N by 5.4%, but doubling N from sixteen to thirty-two increases pass@N by only 1.5%.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!peIA!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6d8cea0-31eb-467b-9e26-e59bc220a9d1_2256x980.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!peIA!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6d8cea0-31eb-467b-9e26-e59bc220a9d1_2256x980.png 424w, https://substackcdn.com/image/fetch/$s_!peIA!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6d8cea0-31eb-467b-9e26-e59bc220a9d1_2256x980.png 848w, https://substackcdn.com/image/fetch/$s_!peIA!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6d8cea0-31eb-467b-9e26-e59bc220a9d1_2256x980.png 1272w, https://substackcdn.com/image/fetch/$s_!peIA!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6d8cea0-31eb-467b-9e26-e59bc220a9d1_2256x980.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!peIA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6d8cea0-31eb-467b-9e26-e59bc220a9d1_2256x980.png" width="1456" height="632" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e6d8cea0-31eb-467b-9e26-e59bc220a9d1_2256x980.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:632,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:219222,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://epochai.substack.com/i/176416854?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6d8cea0-31eb-467b-9e26-e59bc220a9d1_2256x980.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!peIA!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6d8cea0-31eb-467b-9e26-e59bc220a9d1_2256x980.png 424w, https://substackcdn.com/image/fetch/$s_!peIA!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6d8cea0-31eb-467b-9e26-e59bc220a9d1_2256x980.png 848w, https://substackcdn.com/image/fetch/$s_!peIA!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6d8cea0-31eb-467b-9e26-e59bc220a9d1_2256x980.png 1272w, https://substackcdn.com/image/fetch/$s_!peIA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6d8cea0-31eb-467b-9e26-e59bc220a9d1_2256x980.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>How much more would pass@N increase with additional doublings? On average, the gains from doubling decreased by about 1% with each doubling. If we extrapolate that, we would see a gain of 0.5% as N goes from thirty-two to sixty-four, and that would be it. Maybe there&#8217;s a bit of a longer tail, but as a rough round-number guess, it seems likely that pass@N here would cap out below 50%.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a></p><p>To sanity check this, I randomly drew ten problems that were not solved in any of these runs, and sampled them each 100 more times, for 132 total i.e. about two more doublings. None of these problems were solved even once. This is consistent with the observed data: we wouldn&#8217;t expect a draw of ten problems to hit one of the few problems that may be within reach of two more doublings. But this at least rules out the possibility that there&#8217;s a gold mine of problems that GPT-5 can get right if only it&#8217;s given 100 more chances.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a></p><h2>Pass@the-kitchen-sink likely caps out below 70%</h2><p>What about models other than GPT-5, or scaffolds other than Epoch&#8217;s? It&#8217;s a bit messy, but we can at least look at the assortment of FrontierMath runs Epoch has accumulated over time. We call this <em>pass@the-kitchen-sink</em>, and at the moment our kitchen sink is filled from the following six buckets:</p><ul><li><p>32 runs of GPT-5, described above</p></li><li><p>52 runs of various models from various developers from Epoch&#8217;s <a href="https://epoch.ai/benchmarks/frontiermath">benchmarking hub</a></p></li><li><p>16 runs of ChatGPT Agent <a href="https://x.com/EpochAIResearch/status/1945905802998423867">conducted by OpenAI and graded by Epoch</a></p></li><li><p>1 run of a Gemini 2.5 Deep Think, <a href="https://epoch.ai/blog/deep-think-math">evaluated manually</a></p></li><li><p>20 runs of o4-mini conducted a few months ago as part of an open-ended exploration</p></li><li><p>6 miscellaneous runs conducted while experimenting with o4-mini, Gemini 2.5 Pro, and Grok 4</p></li></ul><p>All together this gives a pass@the-kitchen-sink of 57%, or 165/290 problems solved.</p><p>Of these problems, 140 are solved in at least two buckets. This suggests that models do not have very different skill profiles, even ones from different developers: if one model can solve it, another probably can too. The table below shows how many problems are solved uniquely within a given bucket, as well as the bucket&#8217;s overall pass@N rate. Note in particular that our 32 runs of GPT-5 only solved a single problem that we hadn&#8217;t seen solved in some other run before!</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!QSxg!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fab7042f0-e1d7-4bd8-94a8-442a13b53c64_1692x664.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!QSxg!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fab7042f0-e1d7-4bd8-94a8-442a13b53c64_1692x664.png 424w, https://substackcdn.com/image/fetch/$s_!QSxg!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fab7042f0-e1d7-4bd8-94a8-442a13b53c64_1692x664.png 848w, https://substackcdn.com/image/fetch/$s_!QSxg!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fab7042f0-e1d7-4bd8-94a8-442a13b53c64_1692x664.png 1272w, https://substackcdn.com/image/fetch/$s_!QSxg!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fab7042f0-e1d7-4bd8-94a8-442a13b53c64_1692x664.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!QSxg!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fab7042f0-e1d7-4bd8-94a8-442a13b53c64_1692x664.png" width="1456" height="571" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ab7042f0-e1d7-4bd8-94a8-442a13b53c64_1692x664.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:571,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:153361,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://epochai.substack.com/i/176416854?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fab7042f0-e1d7-4bd8-94a8-442a13b53c64_1692x664.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!QSxg!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fab7042f0-e1d7-4bd8-94a8-442a13b53c64_1692x664.png 424w, https://substackcdn.com/image/fetch/$s_!QSxg!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fab7042f0-e1d7-4bd8-94a8-442a13b53c64_1692x664.png 848w, https://substackcdn.com/image/fetch/$s_!QSxg!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fab7042f0-e1d7-4bd8-94a8-442a13b53c64_1692x664.png 1272w, https://substackcdn.com/image/fetch/$s_!QSxg!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fab7042f0-e1d7-4bd8-94a8-442a13b53c64_1692x664.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The one notable deviation from this is ChatGPT Agent, which solved 14 problems (5%) that no other model solved. There&#8217;s a salient explanation for this: it&#8217;s the only model here with access to a web search tool. While FrontierMath doesn&#8217;t contain pure &#8220;look-up&#8221; problems, it does have some problems that are meant to be solved by appropriately adapting somewhat obscure knowledge, so we expect web search to help.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-8" href="#footnote-8" target="_self">8</a></p><p>Since we have 16 runs of ChatGPT Agent, we can look at how Pass@N scales as above. Doing this, we see a similar sub-logarithmic pattern: the returns to doubling N decrease by about 1% (absolute) with each doubling. Simple extrapolation suggests there might be another 6% to gain from additional doublings.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!gn0o!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79e31062-78a0-4d51-9cf9-ca78d2122f0c_2250x864.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!gn0o!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79e31062-78a0-4d51-9cf9-ca78d2122f0c_2250x864.png 424w, https://substackcdn.com/image/fetch/$s_!gn0o!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79e31062-78a0-4d51-9cf9-ca78d2122f0c_2250x864.png 848w, https://substackcdn.com/image/fetch/$s_!gn0o!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79e31062-78a0-4d51-9cf9-ca78d2122f0c_2250x864.png 1272w, https://substackcdn.com/image/fetch/$s_!gn0o!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79e31062-78a0-4d51-9cf9-ca78d2122f0c_2250x864.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!gn0o!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79e31062-78a0-4d51-9cf9-ca78d2122f0c_2250x864.png" width="1456" height="559" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/79e31062-78a0-4d51-9cf9-ca78d2122f0c_2250x864.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:559,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:177575,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://epochai.substack.com/i/176416854?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79e31062-78a0-4d51-9cf9-ca78d2122f0c_2250x864.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!gn0o!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79e31062-78a0-4d51-9cf9-ca78d2122f0c_2250x864.png 424w, https://substackcdn.com/image/fetch/$s_!gn0o!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79e31062-78a0-4d51-9cf9-ca78d2122f0c_2250x864.png 848w, https://substackcdn.com/image/fetch/$s_!gn0o!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79e31062-78a0-4d51-9cf9-ca78d2122f0c_2250x864.png 1272w, https://substackcdn.com/image/fetch/$s_!gn0o!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79e31062-78a0-4d51-9cf9-ca78d2122f0c_2250x864.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Where would pass@the-kitchen-sink cap out?</p><p>We have clean estimates for GPT-5 and ChatGPT Agent. Conservatively assuming the gains from additional runs of these would be disjoint, that would add 7% in the limit, for a running tally of 64% total.</p><p>We don&#8217;t have as good a way to predict what all the other models would add, if re-run repeatedly. We can at least note that, across all our existing runs for them, they solve 46% of problems, but only 13 problems (5%) that neither GPT-5 nor ChatGPT Agent solved. This suggests fairly low marginal diversity, though we can&#8217;t rule out that one of these models would break into new territory if re-run repeatedly.</p><p>For the sake of a bottom-line number, I&#8217;ll assume that this pool has as much juice left to squeeze as ChatGPT Agent, the larger of our previous two estimates. Again conservatively assuming that these gains would be disjoint, this gives an all-in cap of 70%.</p><p>In the original release of FrontierMath, we estimated that 7% of problems were erroneous, either due to fatal ambiguities in the problem statement or issues in the answer key or verification script. We collect errors as we find them and will update the benchmark periodically, so this number will decrease over time. Currently we are only aware of uncorrected errors in 2% of problems. Conservatively taking the error rate to be 10%, we are thus left with 20% of FrontierMath problems being likely out of reach for current models.</p><p>Returning to the 57% of solved problems, we observe the expected pattern across the FrontierMath difficulty tiers: lower tiers have higher solve rates.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!c6Fm!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe36c604e-eb4f-493f-a660-5795684c69da_1025x1280.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!c6Fm!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe36c604e-eb4f-493f-a660-5795684c69da_1025x1280.png 424w, https://substackcdn.com/image/fetch/$s_!c6Fm!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe36c604e-eb4f-493f-a660-5795684c69da_1025x1280.png 848w, https://substackcdn.com/image/fetch/$s_!c6Fm!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe36c604e-eb4f-493f-a660-5795684c69da_1025x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!c6Fm!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe36c604e-eb4f-493f-a660-5795684c69da_1025x1280.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!c6Fm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe36c604e-eb4f-493f-a660-5795684c69da_1025x1280.png" width="1025" height="1280" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e36c604e-eb4f-493f-a660-5795684c69da_1025x1280.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1280,&quot;width&quot;:1025,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:80146,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://epochai.substack.com/i/176416854?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe36c604e-eb4f-493f-a660-5795684c69da_1025x1280.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!c6Fm!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe36c604e-eb4f-493f-a660-5795684c69da_1025x1280.png 424w, https://substackcdn.com/image/fetch/$s_!c6Fm!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe36c604e-eb4f-493f-a660-5795684c69da_1025x1280.png 848w, https://substackcdn.com/image/fetch/$s_!c6Fm!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe36c604e-eb4f-493f-a660-5795684c69da_1025x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!c6Fm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe36c604e-eb4f-493f-a660-5795684c69da_1025x1280.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>This gives us something to watch as models improve on FrontierMath</h2><p>Epoch recently published a report on <a href="https://epoch.ai/blog/what-will-ai-look-like-in-2030">how we expect AI to look in 2030</a>, which included the following projection of FrontierMath performance.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!q-TH!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0a123725-77c5-406c-b684-e901e39701e9_1025x1280.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!q-TH!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0a123725-77c5-406c-b684-e901e39701e9_1025x1280.png 424w, https://substackcdn.com/image/fetch/$s_!q-TH!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0a123725-77c5-406c-b684-e901e39701e9_1025x1280.png 848w, https://substackcdn.com/image/fetch/$s_!q-TH!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0a123725-77c5-406c-b684-e901e39701e9_1025x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!q-TH!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0a123725-77c5-406c-b684-e901e39701e9_1025x1280.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!q-TH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0a123725-77c5-406c-b684-e901e39701e9_1025x1280.png" width="1025" height="1280" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0a123725-77c5-406c-b684-e901e39701e9_1025x1280.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1280,&quot;width&quot;:1025,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:68798,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://epochai.substack.com/i/176416854?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0a123725-77c5-406c-b684-e901e39701e9_1025x1280.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!q-TH!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0a123725-77c5-406c-b684-e901e39701e9_1025x1280.png 424w, https://substackcdn.com/image/fetch/$s_!q-TH!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0a123725-77c5-406c-b684-e901e39701e9_1025x1280.png 848w, https://substackcdn.com/image/fetch/$s_!q-TH!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0a123725-77c5-406c-b684-e901e39701e9_1025x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!q-TH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0a123725-77c5-406c-b684-e901e39701e9_1025x1280.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>While a performance of 70% seems high compared to today&#8217;s SOTA, our projections show it arriving in the first half of next year. The investigations in this post make this seem believable: we have already observed <em>some</em> model solving almost that many problems. If additional training can simply shore up the performance observed in our pass@the-kitchen-sink, it will already be most of the way there.</p><p>By the same token, this gives us something more precise to look at if and when a model does hit this performance level. In particular, we can see how much of the gain comes from problems that we have already observed being solved at least once, vs. problems that haven&#8217;t been solved at all before.</p><p>In the extremes, the interpretation of progress is very different. If gains come entirely from previously solved problems, then this represents purely improved reliability. If, however, much of the gain comes from problems that no model has solved before, then we can interpret it as a meaningful advance in capability.</p><p><em>Thanks to OpenAI for API credits used for the GPT-5 experiments. Thanks to Daniel Litt for feedback on this post.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://epochai.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading! Subscribe to receive the latest from Epoch AI.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div data-component-name="FragmentNodeToDOM"><p></p></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>In this post, by FrontierMath, I always mean the private problems in Tiers 1&#8211;3.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>This reasoning is only valid for benchmarks where random guessing scores &#8776;0%, but FrontierMath is such a benchmark: answers are typically large integers or complicated symbolic real numbers, so we can sample repeatedly without worrying that models are getting correct answers just by luck.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>These runs were conducted with a 10x token budget compared to our usual scaffold. In the run currently shown on our <a href="https://epoch.ai/benchmarks/frontiermath">benchmarking hub</a>, GPT-5 scored 25% but exceeded our scaffold&#8217;s 100K token budget on 3% of problems. I tried giving it a 10x token budget, and it scored 29% while never exceeding the budget. For our experiments here, I used our standard scaffold with 10x the standard token budget. We&#8217;ll be increasing the general token budget in our scaffold going forward.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>I intended to conduct these runs with GPT-5 (high), but due to a bug in our evaluations infrastructure, they were in fact conducted using GPT-5 (medium). On our standard scaffold, GPT-5 (high)&#8217;s now-correct score on FrontierMath Tiers 1-3 is 27%, compared to 25% for GPT-5 (medium). I expect the overall conclusions of this article would not be changed if GPT-5 (high) was used instead. In particular, the fact (discussed later) that all 32 runs of GPT-5 (medium) only solved <em>one</em> problem that no other model run has solved suggests that there just isn&#8217;t that much juice for GPT-5 to squeeze. Still, it is entirely possible that GPT-5 (high)&#8217;s asymptote is a few percentage points higher than GPT-5 (medium)&#8217;s.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>That said, we can&#8217;t rule out that this is simply a very slow-growing function that is nonetheless unbounded, e.g. c*log(N) for small c, or log(log(N)).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>Of course, these experiments in general lack the power to tell the difference between p(solve)=0 and p(solve)=small. If we crudely model a problem as requiring k steps, and that GPT-5 has a probability of p of performing each step correctly, then p(solve)=p^k. For a problem with k=10 and p=&#189;, GPT has a &#8776;50% chance of solving it in 709 samples. I would still want to count this problem as &#8220;within reach&#8221;.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>These problems were also chosen from among those that no <em>other</em> model run solved. These runs are discussed further on in the post.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-8" href="#footnote-anchor-8" class="footnote-number" contenteditable="false" target="_self">8</a><div class="footnote-content"><p>We plan to experiment with adding this.</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[OpenAI is projecting unprecedented revenue growth]]></title><description><![CDATA[No company has gone from $10B to $100B as fast as OpenAI projects to do]]></description><link>https://epochai.substack.com/p/openai-is-projecting-unprecedented</link><guid isPermaLink="false">https://epochai.substack.com/p/openai-is-projecting-unprecedented</guid><dc:creator><![CDATA[Greg Burnham]]></dc:creator><pubDate>Wed, 15 Oct 2025 16:05:54 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!JegC!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F140f9e18-7326-4efb-af0a-48df34bd911a_1025x1280.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Epoch&#8217;s new <a href="https://epoch.ai/data/ai-companies">AI companies database</a> shows the remarkable level and pace of growth of OpenAI&#8217;s revenue. It first exceeded $1 billion in 2023 and will exceed $10 billion in 2025. This is impressive, but not unprecedented &#8212; a few other companies have matched this growth rate historically.</p><p>OpenAI&#8217;s projections, however, are a different story. According to <em>The Information</em>, in Q3 2025 <a href="https://www.theinformation.com/articles/openai-says-business-will-burn-115-billion-2029">OpenAI projected</a> its 2028 revenue to be $100 billion<em>.</em> I couldn&#8217;t find any examples of a company growing its revenue from around $10 billion to $100 billion in such a short period of time.</p><p>What happens if OpenAI falls short of these projections? At a minimum, it would likely have to scale back its <a href="https://www.theinformation.com/articles/openais-350-billion-computing-cost-problem">plans</a> for large compute build-outs. The recently-announced deals with Nvidia, AMD, and Broadcom imply expenditures of roughly $1.3 trillion within the next decade, and some of this is presumably expected to be financed by revenue or debt raised against revenue.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a></p><p>But the second-order effects of a miss could be larger. This is because investors and other companies are <a href="https://peterwildeford.substack.com/i/174301735/the-economy-is-increasingly-a-leveraged-bet-on-agi">increasingly betting big</a> on OpenAI being highly valuable. As <a href="https://www.noahpinion.blog/p/americas-future-could-hinge-on-whether">Noah Smith recently wrote</a>, these bets depend not only on this value being realized, but on it being realized <em>fast enough</em> to cover the debt used to finance the bets. Failing to deliver value as fast as investors expected is all it took to turn several historical technology booms into busts.</p><p>We don&#8217;t know how sensitive the AI financing system is to OpenAI&#8217;s specific projections. It&#8217;s possible that there is a big margin of error and the tenor of the deals is, &#8220;Even if we miss our revenue projections, we all still get rich.&#8221; But it seems just as possible that more is at stake.</p><p>One thing we can say: OpenAI hitting its projections would be a historic achievement.</p><h2>OpenAI&#8217;s revenue grew very quickly from $1B to $10B</h2><p>I found four instances of US companies in the past fifty years growing their revenue from less than $1 billion to over $10 billion over the course of three years.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> It&#8217;s a somewhat eclectic group.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!NJ45!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F760a639b-0d0d-4e12-86f1-4007cfb4fc04_1024x1280.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!NJ45!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F760a639b-0d0d-4e12-86f1-4007cfb4fc04_1024x1280.png 424w, https://substackcdn.com/image/fetch/$s_!NJ45!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F760a639b-0d0d-4e12-86f1-4007cfb4fc04_1024x1280.png 848w, https://substackcdn.com/image/fetch/$s_!NJ45!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F760a639b-0d0d-4e12-86f1-4007cfb4fc04_1024x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!NJ45!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F760a639b-0d0d-4e12-86f1-4007cfb4fc04_1024x1280.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!NJ45!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F760a639b-0d0d-4e12-86f1-4007cfb4fc04_1024x1280.png" width="1024" height="1280" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/760a639b-0d0d-4e12-86f1-4007cfb4fc04_1024x1280.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1280,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:89296,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://epochai.substack.com/i/176240335?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F760a639b-0d0d-4e12-86f1-4007cfb4fc04_1024x1280.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!NJ45!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F760a639b-0d0d-4e12-86f1-4007cfb4fc04_1024x1280.png 424w, https://substackcdn.com/image/fetch/$s_!NJ45!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F760a639b-0d0d-4e12-86f1-4007cfb4fc04_1024x1280.png 848w, https://substackcdn.com/image/fetch/$s_!NJ45!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F760a639b-0d0d-4e12-86f1-4007cfb4fc04_1024x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!NJ45!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F760a639b-0d0d-4e12-86f1-4007cfb4fc04_1024x1280.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The one company to clearly cross these thresholds faster than OpenAI was the vaccine maker Moderna, whose revenue spiked in 2021 due to the pandemic. Another demand-driven entry is Cheniere Energy (2015&#8211;2018), which rode the US natural gas boom. Uber (2014&#8211;2017) and Google (2002-2005) are familiar tech startup success stories.</p><p>Being a pure software company, Google may be the most apt comparison to OpenAI. Of these four, only Google went on to hit $100 billion revenue. The others have yet to pass $40 billion.</p><p>OpenAI&#8217;s recent growth does look fast compared to the others. It is currently growing around 3&#215; per year, compared to Google&#8217;s 2&#215; per year in the year when it first hit $10 billion. So, apart from Moderna, OpenAI may be the <em>fastest</em>-growing company to hit $10 billion. Can it keep up the pace?</p><h2>No company has grown its revenue from $10 billion to $100 billion in three years</h2><p>OpenAI expects to make <a href="https://finance.yahoo.com/news/openai-cfo-we-will-more-than-triple-our-revenue-this-year-213816957.html">$13 billion</a> in revenue in 2025 and <a href="https://www.theinformation.com/articles/openai-says-business-will-burn-115-billion-2029">projects</a> to grow its revenue by 2.3&#215; in 2026, 2&#215; in 2027, and 1.6&#215; in 2028. I couldn&#8217;t find any historical examples of companies growing at such a fast pace from such a high level. The following chart shows the seven US companies I could find in the past fifty years that grew their revenue from $10B to $100B in less than a decade.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!_qnj!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8924e6e-1773-48c6-8a16-1ca9ad00d6eb_1025x1280.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!_qnj!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8924e6e-1773-48c6-8a16-1ca9ad00d6eb_1025x1280.png 424w, https://substackcdn.com/image/fetch/$s_!_qnj!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8924e6e-1773-48c6-8a16-1ca9ad00d6eb_1025x1280.png 848w, https://substackcdn.com/image/fetch/$s_!_qnj!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8924e6e-1773-48c6-8a16-1ca9ad00d6eb_1025x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!_qnj!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8924e6e-1773-48c6-8a16-1ca9ad00d6eb_1025x1280.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!_qnj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8924e6e-1773-48c6-8a16-1ca9ad00d6eb_1025x1280.png" width="1025" height="1280" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e8924e6e-1773-48c6-8a16-1ca9ad00d6eb_1025x1280.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1280,&quot;width&quot;:1025,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:104987,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://epochai.substack.com/i/176240335?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8924e6e-1773-48c6-8a16-1ca9ad00d6eb_1025x1280.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!_qnj!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8924e6e-1773-48c6-8a16-1ca9ad00d6eb_1025x1280.png 424w, https://substackcdn.com/image/fetch/$s_!_qnj!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8924e6e-1773-48c6-8a16-1ca9ad00d6eb_1025x1280.png 848w, https://substackcdn.com/image/fetch/$s_!_qnj!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8924e6e-1773-48c6-8a16-1ca9ad00d6eb_1025x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!_qnj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe8924e6e-1773-48c6-8a16-1ca9ad00d6eb_1025x1280.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The fastest companies to cross these thresholds were Tesla and Meta in seven years, followed by Nvidia, Apple, and Amazon in eight, Walmart in nine, and Google in ten. The median annual revenue growth rate of these companies during this high-growth period is 1.3&#215;. From this perspective, OpenAI is targeting an unprecedented performance.</p><p>However, the chart also shows the arbitrariness of the thresholds. Yes it took Nvidia eight years to go from $10 billion to $100 billion, but almost all of that growth came in just two banner years toward the end of the period, both a bit over 2&#215; per year, starting from a base of about $28 billion. This is the most optimistic reference point for OpenAI.</p><p>I also looked for any company with revenue over $10 billion (in 2024 USD) that more than doubled its revenue in a single year. As just mentioned, Nvidia did this twice. I only found two other examples. Gilead Sciences went from $11 billion in 2013 to $25 billion in 2014, driven by the success of a new hepatitis C drug. Cheniere Energy grew its revenue from $16 billion in 2021 to $33 billion in 2022, riding another natural gas price spike. Both companies&#8217; revenues plateaued and declined over the subsequent two years.</p><p>I focused on recent US companies to isolate factors like market access, political economy, and the availability of capital. Looking beyond the US, the only stand-out example of rapid revenue growth that I could find was ByteDance, the parent company of TikTok. In 2024 USD, its revenue first crossed $10B in 2019 and first crossed $100B in 2024, five years later.</p><h2>Can OpenAI do it?</h2><p>The outside view says: probably not. Companies usually don&#8217;t do this. Does a closer look reveal anything different?</p><p>The strongest argument in OpenAI&#8217;s favor is probably, as mentioned above, that its revenue is growing very quickly right now: 3&#215; per year. All data points since 2024, even recent ones, are consistent with this trend. Given that, its growth could slow somewhat while still hitting the target.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!00bw!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe472328f-732e-40ca-a75a-e3ce7318cce9_1025x1280.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!00bw!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe472328f-732e-40ca-a75a-e3ce7318cce9_1025x1280.png 424w, https://substackcdn.com/image/fetch/$s_!00bw!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe472328f-732e-40ca-a75a-e3ce7318cce9_1025x1280.png 848w, https://substackcdn.com/image/fetch/$s_!00bw!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe472328f-732e-40ca-a75a-e3ce7318cce9_1025x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!00bw!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe472328f-732e-40ca-a75a-e3ce7318cce9_1025x1280.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!00bw!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe472328f-732e-40ca-a75a-e3ce7318cce9_1025x1280.png" width="1025" height="1280" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e472328f-732e-40ca-a75a-e3ce7318cce9_1025x1280.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1280,&quot;width&quot;:1025,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:65238,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://epochai.substack.com/i/176240335?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe472328f-732e-40ca-a75a-e3ce7318cce9_1025x1280.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!00bw!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe472328f-732e-40ca-a75a-e3ce7318cce9_1025x1280.png 424w, https://substackcdn.com/image/fetch/$s_!00bw!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe472328f-732e-40ca-a75a-e3ce7318cce9_1025x1280.png 848w, https://substackcdn.com/image/fetch/$s_!00bw!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe472328f-732e-40ca-a75a-e3ce7318cce9_1025x1280.png 1272w, https://substackcdn.com/image/fetch/$s_!00bw!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe472328f-732e-40ca-a75a-e3ce7318cce9_1025x1280.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>There also appears to be a trend toward companies hitting $100 billion in revenue faster. Of our seven examples above, six hit $100 billion in the last fifteen years. If there is some general dynamic by which this happens faster over time, it could be a tailwind for OpenAI.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a></p><p>Where might this revenue come from? It probably won&#8217;t be ChatGPT subscriptions alone, which currently account for <a href="https://www.ft.com/content/b81d5fb6-26e9-417a-a0cc-6b6689b70c98">about 75%</a> of OpenAI&#8217;s revenue. OpenAI projects only $50 billion of its 2028 revenue to come from ChatGPT directly. As a reference point, that is the cost equivalent of 210 million of today&#8217;s &#8220;Plus&#8221; subscriptions. The most recent figures, from April 2025, show <a href="https://finance.yahoo.com/news/chatgpt-crosses-20m-paid-users-140519166.html">20 million</a> paid subscriptions of all kinds.</p><p>OpenAI also seems likely to make a play for the advertising and shopping revenue that currently flows through Google, Amazon, and Meta. Their main tool here will be the much larger user base of ChatGPT&#8217;s free tier. Winning market share in this relatively mature sector is at least plausible.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a></p><p>Or maybe that&#8217;s thinking small. The promise of AI, after all, is broad productivity. For a very rough estimate, let&#8217;s imagine that:</p><ul><li><p>OpenAI&#8217;s current and near-term products could provide an average 10% productivity uplift for all remote-compatible work tasks, and</p></li><li><p>OpenAI can capture 10% of the resulting value.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a></p></li></ul><p>We previously estimated that <a href="https://epoch.ai/gradient-updates/consequences-of-automating-remote-work">about a third of US work tasks</a> are remote-compatible, and that is probably lower globally. So, assuming these tasks account for around 20% of the roughly $110 trillion of global GDP, the total potential revenue from this near-term productivity uplift might be around $220 billion.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-8" href="#footnote-8" target="_self">8</a></p><p>There will be competition for this revenue, but perhaps OpenAI can maintain its current relatively high market share. Maybe AI itself can lower the hurdles to adoption, helping people quickly figure out how to use it to be more productive.</p><p>Still, reaching such heights in just three years would be an extraordinary achievement. History suggests that it doesn&#8217;t happen quite so fast.</p><h2>What if not?</h2><p>Maybe if OpenAI &#8220;merely&#8221; gets to $50 billion in 2028&#8212;an impressive feat in its own right&#8212;it will only have had to slow its data center build-outs, and will otherwise remain financially healthy. But the more stakeholders take big bets on OpenAI, the riskier such misses become. We&#8217;ll be paying close attention to OpenAI&#8217;s next few revenue prints.</p><p><em>Thanks to Steven Kryger and Shanu Mathew for feedback on this post.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://epochai.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading! Subscribe to receive the latest from Epoch AI.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>The deals total 26 GW of chips, and a 1 GW data center costs roughly $50 billion to build, all-in. Precise timeframes of the deals are not available, but all plans discussed publicly cite dates within the next decade.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>I don&#8217;t have access to a comprehensive database of company revenue. My methodology for this post relied in large part on queries to GPT-5 Pro, Grok 4 Heavy, and Gemini 2.5 Deep Think, combined with manual spot checks. If you know of examples of companies achieving faster revenue growth off larger bases than what I discuss, please tell me! <a href="mailto:greg@epoch.ai">greg@epoch.ai</a> or <a href="https://x.com/GregHBurnham">@GregHBurnham</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>All analysis in this post excludes cases of revenue growth driven primarily by mergers and acquisitions.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>The first year shown for each company is either the first year when the company&#8217;s revenue exceeded $10 billion, or the year prior to that, if the company&#8217;s revenue in this prior year was closer to $10 billion than the company&#8217;s revenue in the subsequent year. The range of revenues in this &#8220;Year 0&#8221; is $9-13 billion, with OpenAI&#8217;s toward the upper end of the range. Shifting the rest of the cohort one year earlier doesn&#8217;t change the main observation.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>Perhaps the better way to think about the $100 billion milestone is as a percent of GDP: in that case, as GDP grows, the milestone gets easier to achieve.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>Of course, if the tech giants face fierce competition in their core businesses, they may have to scale back their own AI investments. That could cause its own set of problems, depending on what other companies are levered to those anticipated expenditures.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>These numbers are speculative. <a href="https://economics.mit.edu/sites/default/files/inline-files/draft_copilot_experiments.pdf">Uplift estimates vary widely</a> and include <a href="https://metr.org/blog/2025-07-10-early-2025-ai-experienced-os-dev-study/">negative effects</a>. One estimate of the <a href="https://www.wsj.com/opinion/ais-overlooked-97-billion-contribution-to-the-economy-users-service-da6e8f55">consumer surplus of AI</a> puts it at about 10&#215; revenue.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-8" href="#footnote-anchor-8" class="footnote-number" contenteditable="false" target="_self">8</a><div class="footnote-content"><p>The eventual total addressable market for AI is much larger. The question here is rather how much can be captured in the next three years, given current AI capabilities and diffusion.</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[How many digital workers could OpenAI deploy?]]></title><description><![CDATA[OpenAI has the inference compute to deploy millions of digital workers, but only on a narrow set of tasks &#8211; for now.]]></description><link>https://epochai.substack.com/p/how-many-digital-workers-could-openai</link><guid isPermaLink="false">https://epochai.substack.com/p/how-many-digital-workers-could-openai</guid><dc:creator><![CDATA[JSD]]></dc:creator><pubDate>Sat, 04 Oct 2025 17:50:03 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!sVsF!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd88cdcde-9b31-4206-96c7-664d554cad04_1768x1554.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The core argument for how AI could drive <a href="https://epoch.ai/blog/explosive-growth-from-ai-a-review-of-the-arguments">explosive economic growth</a> is that you can dramatically scale up the number of AI &#8220;digital workers&#8221;. The idea is that growth is constrained by labor, so rapidly expanding the workforce would hugely accelerate growth rates.</p><p>This is where AI comes in. While you can&#8217;t double the human population each year, you can double the number of AI chips &#8211; as we&#8217;ve seen with <a href="https://epoch.ai/data-insights/nvidia-chip-production">NVIDIA</a> and OpenAI.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> So if AI can fully substitute for human workers, the workforce could grow many times faster than today. As a result, the economy could grow many times faster too.</p><p>If this framing is right, then to know how far we are from explosive growth, we need to answer three questions. First, how many AI &#8220;digital workers&#8221; can be deployed today? Second, how far is AI from fully substituting for human workers? And third, how are both of these changing over time?</p><p>In this post, we&#8217;ll take a stab at the first question: On the tasks that AIs are able to perform today, how many &#8220;human-equivalent digital workers&#8221; could frontier AI labs deploy to work on them?</p><p>Based on a speculative back-of-the-envelope calculation, we estimate that companies like OpenAI have the hardware to deploy on the order of 7 million digital workers, with a wide 90% confidence interval of 400,000 to around 300 million.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> This doesn&#8217;t mean that OpenAI could do the jobs of 7 million human employees <em>today</em>, because AIs can&#8217;t fully substitute for humans. But as AI progress continues, AIs will be able to perform an increasing fraction of the tasks that humans currently do.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!wSkL!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F330b8137-045c-44cf-ae9d-3f7ae1b11f4e_1268x404.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!wSkL!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F330b8137-045c-44cf-ae9d-3f7ae1b11f4e_1268x404.png 424w, https://substackcdn.com/image/fetch/$s_!wSkL!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F330b8137-045c-44cf-ae9d-3f7ae1b11f4e_1268x404.png 848w, https://substackcdn.com/image/fetch/$s_!wSkL!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F330b8137-045c-44cf-ae9d-3f7ae1b11f4e_1268x404.png 1272w, https://substackcdn.com/image/fetch/$s_!wSkL!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F330b8137-045c-44cf-ae9d-3f7ae1b11f4e_1268x404.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!wSkL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F330b8137-045c-44cf-ae9d-3f7ae1b11f4e_1268x404.png" width="1268" height="404" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/330b8137-045c-44cf-ae9d-3f7ae1b11f4e_1268x404.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:404,&quot;width&quot;:1268,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!wSkL!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F330b8137-045c-44cf-ae9d-3f7ae1b11f4e_1268x404.png 424w, https://substackcdn.com/image/fetch/$s_!wSkL!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F330b8137-045c-44cf-ae9d-3f7ae1b11f4e_1268x404.png 848w, https://substackcdn.com/image/fetch/$s_!wSkL!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F330b8137-045c-44cf-ae9d-3f7ae1b11f4e_1268x404.png 1272w, https://substackcdn.com/image/fetch/$s_!wSkL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F330b8137-045c-44cf-ae9d-3f7ae1b11f4e_1268x404.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>Estimating the number of digital workers that frontier labs can deploy</h2><p>But where does this 7 million number come from? There&#8217;s no &#8220;AI employee count&#8221; webpage from the Bureau of Labor Statistics. So what we need to do is somehow figure out how much &#8220;work&#8221; is being done by running AIs at frontier labs.</p><p>Our approach is to compare the number of tokens that AIs process daily to the number of effective tokens that a human worker &#8220;processes&#8221; over a day when they read, think, and write. Then when you take the ratio of these two, you get the number of digital workers <em>on the tasks that AIs can currently solve.</em></p><p>The first step is thus to figure out how many tokens frontier labs produce each day, and for concreteness let&#8217;s consider OpenAI&#8217;s GPT-5.</p><p>One approach is to look at OpenAI&#8217;s inference compute budget. Based on reporting from <a href="https://www.theinformation.com/articles/openai-projections-imply-losses-tripling-to-14-billion-in-2026?rc=9mzoog">The</a> <a href="https://www.theinformation.com/articles/openai-discusses-building-first-data-center-storage?rc=9mzoog">Information</a>, we think they currently have around 480,000 H100 equivalents for running models,<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> which is enough for around 10<sup>25</sup> FLOP each day.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a> It probably takes on the order of 100 to 600 billion FLOP for GPT-5 to generate each token,<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a> so that works out to around 10 to 100 trillion tokens per day.</p><p>For another approach, we also look directly at usage statistics: ChatGPT likely sends around 4 billion <a href="https://cdn.openai.com/pdf/a253471f-8260-40c6-a2cc-aa93fe9f142e/economic-research-chatgpt-usage-paper.pdf#page=37.10">messages each day</a>.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a> If API costs contribute a quarter of the tokens of ChatGPT, and the mean message length is four thousand tokens long, then that implies GPT-5 generates around 20 trillion tokens per day.</p><p>For our final calculations we simply take a mixture of the predictions of these two approaches (with confidence intervals to capture uncertainty), with a median of around 19 trillion tokens a day. This is comparable to the 35 trillion tokens per day generated by Google&#8217;s AI models, so we think this is probably in the right ballpark.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a></p><p>Now that we know the number of tokens, the next step is to convert these token numbers to &#8220;human-equivalent digital workers&#8221;. Essentially, how many effective tokens does a human employee &#8220;process&#8221; in a day?</p><p>We could look at the number of words spoken or the number of words written. But this would likely be an underestimate, since humans do a lot more thinking than they do speaking or writing. So we instead anchor to human thinking speeds of around <a href="https://bounded-regret.ghost.io/what-will-gpt-2030-look-like/#3-throughput-and-parallel-copies">380 words per minute</a>. If we convert this to tokens, and assume that each human works for 8 hours a day, that implies that each human worker processes around 240 thousand tokens a day.</p><p>Given this number, it&#8217;s really tempting to say &#8220;GPT-5 outputs 19 trillion tokens a day and each human processes around 240 thousand tokens, so there are 19 trillion / 240 thousand = 80 million digital workers&#8221;. But this assumes that each token is worth the same as each GPT-5 token, which might not be true.</p><p>So we can consider an alternative approach, that relies on a <a href="https://metr.org/blog/2025-03-19-measuring-ai-ability-to-complete-long-tasks/">study by METR</a> &#8211; this looks at whether AIs can perform a (software-related) task that takes humans a specified amount of time to perform. For example, one task might involve debugging a particular program, which typically takes human coders an hour to complete. The idea is then to look at how many tokens GPT-5 needs to do the same task and multiply it by the number of hours in a work day, representing the tokens needed to match a human over a day. In the case of the METR tasks, GPT-5 typically uses 100 thousand to 1 million tokens, so over an eight hour workday that&#8217;s 800 thousand to 8 million tokens. And if we divide 19 trillion by this, we end up with 2.4 million to 24 million digital workers.</p><p>Combining these two approaches suggests a median estimate of 7 million digital workers, but with very wide confidence intervals.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-8" href="#footnote-8" target="_self">8</a></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!sVsF!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd88cdcde-9b31-4206-96c7-664d554cad04_1768x1554.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!sVsF!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd88cdcde-9b31-4206-96c7-664d554cad04_1768x1554.png 424w, https://substackcdn.com/image/fetch/$s_!sVsF!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd88cdcde-9b31-4206-96c7-664d554cad04_1768x1554.png 848w, https://substackcdn.com/image/fetch/$s_!sVsF!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd88cdcde-9b31-4206-96c7-664d554cad04_1768x1554.png 1272w, https://substackcdn.com/image/fetch/$s_!sVsF!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd88cdcde-9b31-4206-96c7-664d554cad04_1768x1554.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!sVsF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd88cdcde-9b31-4206-96c7-664d554cad04_1768x1554.png" width="1456" height="1280" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d88cdcde-9b31-4206-96c7-664d554cad04_1768x1554.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1280,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!sVsF!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd88cdcde-9b31-4206-96c7-664d554cad04_1768x1554.png 424w, https://substackcdn.com/image/fetch/$s_!sVsF!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd88cdcde-9b31-4206-96c7-664d554cad04_1768x1554.png 848w, https://substackcdn.com/image/fetch/$s_!sVsF!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd88cdcde-9b31-4206-96c7-664d554cad04_1768x1554.png 1272w, https://substackcdn.com/image/fetch/$s_!sVsF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd88cdcde-9b31-4206-96c7-664d554cad04_1768x1554.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>What do these numbers tell us?</h2><p>7 million digital workers is a lot &#8211; for comparison, the world&#8217;s largest employer has <a href="https://en.wikipedia.org/wiki/List_of_largest_employers">2.1 million employees</a>. This estimate is quite different compared to prior approaches. For instance, <a href="https://www.openphilanthropy.org/research/what-a-compute-centric-framework-says-about-takeoff-speeds/">some</a> <a href="https://epoch.ai/gate">works</a> estimate the number of digital workers by dividing a compute stock by the <a href="https://www.openphilanthropy.org/research/how-much-computational-power-does-it-take-to-match-the-human-brain/">computational power of a human brain</a>.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-9" href="#footnote-9" target="_self">9</a> Given that OpenAI has enough compute to do 10<sup>25</sup> FLOP per day, and the human brain needs 10<sup>15</sup> FLOP/s to run, then that amounts to around 100,000 workers &#8211; around 500&#215; smaller than our estimate!<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-10" href="#footnote-10" target="_self">10</a></p><p>Of course, as we alluded to earlier, it&#8217;d be wrong to conclude that AIs can now do the jobs of 7 million people, because current AI systems aren&#8217;t able to do all the tasks that humans can.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-11" href="#footnote-11" target="_self">11</a> Humans and AIs have different skill profiles, and currently humans are able to do many tasks that AIs can&#8217;t do regardless of how much inference compute they use. But as AI progress continues, AIs will substitute humans for a wider and wider range of tasks, and our calculation approach becomes more directly meaningful - at least as a lower bound on AI&#8217;s economic impact, since AIs will also be capable of tasks that <a href="https://tecunningham.github.io/posts/2025-09-19-transformative-AI-notes.html#we-dont-have-a-standard-model-of-ai:~:text=wages%20high.%E2%86%A9%EF%B8%8E-,AI%20scientists%20will%20be%20unlike%20human%20scientists,-Will%20efficiency%20curves">no human could perform</a>.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-12" href="#footnote-12" target="_self">12</a></p><p>In particular, suppose AIs became good enough to automate a decent chunk of the tasks in the world economy, including the jobs of software engineers and remote workers. Even if we were stuck with the same number of digital workers, the impacts could be pretty notable. On the tasks that AIs can perform, they can typically <a href="https://arxiv.org/pdf/2503.14499#page=21.44">perform them much cheaper than humans</a>. For example, suppose a software engineer takes an hour to debug some code &#8211; then they might be paid $50 to $100. But if GPT-5 is able to do the same task, it would instead cost $1 to $10.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-13" href="#footnote-13" target="_self">13</a> Such a substantial cost difference would provide a strong incentive to use the AIs instead of humans on the automated tasks, potentially impacting the jobs of tens of millions of workers.</p><p>Moreover, if millions of digital workers can be deployed in parallel, learn from their real-world interactions and <a href="https://bounded-regret.ghost.io/what-will-gpt-2030-look-like/#:~:text=4.%20Knowledge%20Sharing">share the resulting knowledge</a>, this would allow AI systems to accumulate knowledge much more quickly than humans. So there&#8217;s the potential for AIs to learn missing skills very quickly.</p><p>On the other hand, one could argue that these 7 million workers are still peanuts compared to the overall economy &#8211; the world&#8217;s working population is still two orders of magnitude larger, so this wouldn&#8217;t be enough to drive explosive growth. With current compute stocks and automation costs, this is true even if we had AI that could fully substitute for humans.</p><p>But <a href="https://www.forethought.org/research/will-ai-r-and-d-automation-cause-a-software-intelligence-explosion">some</a> <a href="https://www.forethought.org/research/how-quick-and-big-would-a-software-intelligence-explosion-be">have</a> also raised the possibility of a more extreme scenario &#8211; if these digital workers were primarily concentrated into AI R&amp;D, that could substantially accelerate AI progress, further increasing the number of digital AI researchers, and so on. While 7 million workers is small on a global scale, it&#8217;s likely still substantially higher than the world&#8217;s current population of <em>AI researchers</em>.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-14" href="#footnote-14" target="_self">14</a></p><p>So the implications of our estimate could change pretty substantially depending on several factors. We need to know how quickly AI automates more tasks. We need to know how much more compute there&#8217;ll be to run digital workers over the next few years, and whether researchers can use this compute more efficiently at inference. And we also need to know how likely we are to see dynamics like explosive economic growth and a software intelligence explosion. These are questions that we&#8217;ll investigate in forthcoming issues.</p><p><em>We would like to thank Lynette Bye for her feedback on this post.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://epochai.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading! Subscribe to receive the latest from Epoch AI.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>For example, OpenAI&#8217;s compute spending grew from <a href="https://www.theinformation.com/articles/openai-projections-imply-losses-tripling-to-14-billion-in-2026?rc=spkbjw">$6 billion</a> mid-2024 to <a href="https://www.theinformation.com/articles/openai-forecast-shows-shift-from-microsoft-to-softbank?rc=spkbjw">$13 billion</a> mid-2025 &#8211; but this growth rate might also <a href="https://epoch.ai/gradient-updates/compute-scaling-will-slow-down-due-to-increasing-lead-times">slow down</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>You can find the full code for our analysis <a href="https://squigglehub.org/models/jsd/ai-simulated-workers">here</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>Specifically, based on reporting from <a href="https://www.theinformation.com/articles/openai-projections-imply-losses-tripling-to-14-billion-in-2026?rc=9mzoog">The</a> <a href="https://www.theinformation.com/articles/openai-discusses-building-first-data-center-storage?rc=9mzoog">Information</a> we estimate that they currently have around 1.1 million H100 equivalents in total compute stock, and we place a 90% credible interval on this from around 800 thousand to 1.4 million H100 equivalents. Around <a href="https://epoch.ai/data/ai-companies">44%</a> of this goes to inference compute, as opposed to training or experiments.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>This depends on several assumptions. First, we assume that inference runs on FP4 on hardware that supports it (e.g. GB200s) and on FP8 otherwise (e.g. H100s). Second, we assume an inference utilization of around 5% to 30%, in part based on our estimates of Llama 3.3&#8217;s and DeepSeek-R1&#8217;s utilizations. The full calculation can be found in the <a href="https://squigglehub.org/models/jsd/ai-simulated-workers">code</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>We estimate that GPT-5 has between 50 billion and 300 billion active parameters, so at 2 FLOP per active parameter, we get 100 billion to 600 billion FLOP per token.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>In June 2025, <a href="https://cdn.openai.com/pdf/a253471f-8260-40c6-a2cc-aa93fe9f142e/economic-research-chatgpt-usage-paper.pdf#page=2.10">ChatGPT daily message counts</a> were 2.627 billion, compared to 451 million one year earlier. This corresponds to a growth rate of 2627 / 451 = 5.8&#215; per year, such that we expect around 4 billion daily messages by September 2025.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>In July 2025, Google&#8217;s AI models <a href="https://blog.google/inside-google/message-ceo/alphabet-earnings-q2-2025/#search">reportedly processed 980 trillion tokens</a> per month, or around 35 trillion tokens per day.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-8" href="#footnote-anchor-8" class="footnote-number" contenteditable="false" target="_self">8</a><div class="footnote-content"><p>We combine these two approaches in part because we&#8217;re uncertain about how many &#8220;human tokens&#8221; equals one GPT-5 token. This can vary a lot depending on the task &#8211; for example, GPT-5 does translation with far fewer tokens than a human would need based on thinking time. So we don&#8217;t want to anchor too much on a 1:1 ratio or tasks like those from the METR study.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-9" href="#footnote-anchor-9" class="footnote-number" contenteditable="false" target="_self">9</a><div class="footnote-content"><p>This is true at a coarse level, where some inference compute budget is divided by some runtime compute cost. But in practice things can be more complicated, e.g. in the <a href="https://epoch.ai/gate">GATE model</a> the runtime compute requirements can change substantially depending on the task.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-10" href="#footnote-anchor-10" class="footnote-number" contenteditable="false" target="_self">10</a><div class="footnote-content"><p>Of course, both our approach and this compute-based approach are highly uncertain &#8211; e.g. because it&#8217;s hard to pin down just how many FLOP/s the brain is doing.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-11" href="#footnote-anchor-11" class="footnote-number" contenteditable="false" target="_self">11</a><div class="footnote-content"><p>We&#8217;ve also glossed over details like how fast AI systems are run.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-12" href="#footnote-anchor-12" class="footnote-number" contenteditable="false" target="_self">12</a><div class="footnote-content"><p>One dimension of these capability improvements is on <a href="https://epochai.substack.com/p/the-huge-potential-implications-of">long-context tasks</a>, which could impact our estimate of the number of digital workers. For example, long context decoding (but not prefill) is typically bottlenecked by memory bandwidth. In our analysis, this would translate into lower compute utilizations, but a more sophisticated approach could also directly model labs&#8217; total memory bandwidth stock, as <a href="https://ai-2027.com/research/compute-forecast">AI Futures attempted in the AI 2027 report</a>. This may be the case even with <a href="https://epoch.ai/data-insights/llm-inference-price-trends">improvements in inference efficiency</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-13" href="#footnote-anchor-13" class="footnote-number" contenteditable="false" target="_self">13</a><div class="footnote-content"><p>This is based on the METR task example we gave earlier. If an GPT-5 takes 100,000 to 1 million tokens to perform a task that takes humans an hour, then at <a href="https://web.archive.org/web/20250920151257/https://openai.com/api/pricing/">$10 per million output tokens</a> this would add up to $1 to $10.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-14" href="#footnote-anchor-14" class="footnote-number" contenteditable="false" target="_self">14</a><div class="footnote-content"><p>This is especially true if we restrict ourselves to research that occurs on frontier models. For example, if we sum the <a href="https://epoch.ai/data/ai-companies">number of employees at frontier AI labs</a>, we likely get on the order of 10,000 employees. This is already multiple orders of magnitude lower than our estimate, not to mention that the actual number is likely lower, since not all employees are researchers.</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[Why GPT-5 used less training compute than GPT-4.5 (but GPT-6 probably won’t)]]></title><description><![CDATA[OpenAI focused on scaling post-training on a smaller model]]></description><link>https://epochai.substack.com/p/why-gpt-5-used-less-training-compute</link><guid isPermaLink="false">https://epochai.substack.com/p/why-gpt-5-used-less-training-compute</guid><dc:creator><![CDATA[Yafah Edelman]]></dc:creator><pubDate>Fri, 26 Sep 2025 20:31:01 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!qAwq!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F74ce5a28-866a-4830-89ec-15fbc60893ab_2048x1536.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Out of all the GPT models, GPT-5 is the odd one out. Unlike all previous versions of GPT, it was likely <a href="https://x.com/EpochAIResearch/status/1953883611389702169">trained on less compute</a> than its immediate predecessor, GPT-4.5.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!qAwq!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F74ce5a28-866a-4830-89ec-15fbc60893ab_2048x1536.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!qAwq!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F74ce5a28-866a-4830-89ec-15fbc60893ab_2048x1536.png 424w, https://substackcdn.com/image/fetch/$s_!qAwq!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F74ce5a28-866a-4830-89ec-15fbc60893ab_2048x1536.png 848w, https://substackcdn.com/image/fetch/$s_!qAwq!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F74ce5a28-866a-4830-89ec-15fbc60893ab_2048x1536.png 1272w, https://substackcdn.com/image/fetch/$s_!qAwq!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F74ce5a28-866a-4830-89ec-15fbc60893ab_2048x1536.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!qAwq!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F74ce5a28-866a-4830-89ec-15fbc60893ab_2048x1536.png" width="1456" height="1092" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/74ce5a28-866a-4830-89ec-15fbc60893ab_2048x1536.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1092,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!qAwq!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F74ce5a28-866a-4830-89ec-15fbc60893ab_2048x1536.png 424w, https://substackcdn.com/image/fetch/$s_!qAwq!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F74ce5a28-866a-4830-89ec-15fbc60893ab_2048x1536.png 848w, https://substackcdn.com/image/fetch/$s_!qAwq!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F74ce5a28-866a-4830-89ec-15fbc60893ab_2048x1536.png 1272w, https://substackcdn.com/image/fetch/$s_!qAwq!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F74ce5a28-866a-4830-89ec-15fbc60893ab_2048x1536.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>While the exact numbers are uncertain, GPT-4.5 very likely used more training compute than GPT-5.</em></figcaption></figure></div><p>But this leads to a puzzle: Models trained with more compute tend to be better, so why did OpenAI train GPT-5 with less compute than GPT-4.5? And what will this mean for future OpenAI models?</p><p>In this post, we&#8217;ll argue that the answers to these questions are the following:</p><ul><li><p><strong>GPT-5 used less training compute than GPT-4.5 because OpenAI focused on scaling post-training. </strong>New post-training techniques made it possible to outperform GPT-4.5 with less training compute, but these methods likely weren&#8217;t yet mature enough to be applied at GPT-4.5&#8217;s compute scale. Doing so would&#8217;ve taken more time (and compute), which OpenAI likely chose not to do due to strong market pressures.</p></li><li><p><strong>OpenAI&#8217;s next flagship model (&#8220;GPT-6&#8221;) will probably be trained on more compute than GPT-4.5: </strong>When OpenAI figures out how to productively scale post-training, they&#8217;ll likely shift back using more training compute.</p></li></ul><p>Importantly, when we say &#8220;training compute&#8221;, we&#8217;re focusing on the compute to perform the final training run of a model. It&#8217;s likely that the total compute for developing GPT-5 was higher than for GPT-4.5, if we also account for the compute for running experiments. This is because OpenAI&#8217;s (projected) R&amp;D compute spend has grown from <a href="https://www.theinformation.com/articles/openai-projections-imply-losses-tripling-to-14-billion-in-2026?rc=spkbjw">~$5 billion</a> in 2024 to <a href="https://www.theinformation.com/articles/openai-spend-100-billion-backup-servers-ai-breakthroughs?rc=spkbjw">~$9 billion in 2025</a>.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a></p><p>Now let&#8217;s consider the arguments in turn.</p><h2>GPT-5 used less training compute than GPT-4.5 because OpenAI focused on scaling post-training</h2><p>Why did GPT-5 use less training compute than GPT-4.5? We believe this is a combination of two factors. First, OpenAI decided to prioritize scaling post-training, which had better returns on the margin. Second, they couldn&#8217;t readily scale post-training compute to GPT-4.5 levels at the time. And if they tried to scale post-training on a model with as much pre-training as GPT-4.5, they would&#8217;ve run into timing and experimental compute constraints.</p><p>Until recently, most LLMs were trained with <a href="https://epoch.ai/blog/ai-capabilities-can-be-significantly-improved-without-expensive-retraining">100&#215; more pre-training than post-training compute</a>. However, around September 2024, researchers developed novel techniques used in &#8220;reasoning models&#8221; that help scale post-training compute effectively. Researchers could now triple post-training compute in a way that was at least as useful as tripling pre-training compute. In fact, these reasoning techniques make it possible to reduce pre-training compute by <a href="https://epoch.ai/gradient-updates/quantifying-the-algorithmic-improvement-from-reasoning-models">roughly 10&#215;</a> while getting the same performance!<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a></p><p>This means that, rather than spending around $200 million on pre-training and $2 million on post-training GPT-4.5,<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a> new post-training techniques made it possible for that $2 million in post-training to achieve the same overall performance with only $20 million in pre-training. That&#8217;s roughly a ten-fold decrease in training costs, though this doesn&#8217;t imply that total model development costs were lower, due to increases in the compute needed to run experiments. The upshot is that OpenAI was likely able to train a model with less compute than GPT-4.5, while still <a href="https://openai.com/index/introducing-gpt-4-5/">outperforming</a> it on many useful tasks like <a href="https://www.nber.org/system/files/working_papers/w34255/w34255.pdf">coding and search</a>.</p><p>However, while this shows that OpenAI could&#8217;ve outperformed GPT-4.5 with less training compute, it doesn&#8217;t fully explain why they chose this strategy in practice. For example, why not just post-train GPT-4.5? And why not post-train a smaller model on enough data to reach GPT-4.5&#8217;s level of training compute?</p><p>The core reason is that scaling post-training in this way is challenging. It requires lots of testing and experimentation, which takes time and compute, especially when performed on larger, newer models.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a> It also requires a significant amount of high-quality post-training data, which takes time to design and collect.</p><p>Crucially, OpenAI faced major time constraints due to market pressures. This came in the form of fierce competition from rival AI labs, which would hurt their <a href="https://epoch.ai/data-insights/ai-companies-revenue">revenue</a> &#8211; e.g. Anthropic&#8217;s models had been <a href="https://epoch.ai/benchmarks/swe-bench-verified">consistently outperforming</a> OpenAI&#8217;s models at coding. And there was added pressure because many had expected OpenAI to release a model called &#8220;GPT-5&#8221; as early as <a href="https://www.ft.com/content/dd9ba2f6-f509-42f0-8e97-4271c7b84ded">November 2023</a>.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a></p><p>Given these constraints, we believe that OpenAI scaled post-training on a smaller model as much as they could. Scaling further would&#8217;ve either required more experiments than they had the compute or time for, or post-training data that they didn&#8217;t have. Post-training a GPT-4.5-sized model, let alone starting a larger <a href="https://epoch.ai/data-insights/longest-training-run">multi-month</a> pre-training run and doing post-training on top, would&#8217;ve taken too much time or too much experiment compute.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a></p><p>The result of these efforts in scaling post-training was GPT-5, a new state-of-the-art model that OpenAI was able to release by August.</p><h2>GPT-6 will probably be trained on more compute than GPT-4.5</h2><p>What does this mean for training compute trends moving forward? Our best guess is that future iterations of GPT will be trained on more compute.</p><p>To see why, consider the bigger picture. Training GPT-5 with less compute than GPT-4.5 is part of a broader trend, where the training compute of state-of-the-art models has grown more slowly than one might&#8217;ve expected a year ago.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-8" href="#footnote-8" target="_self">8</a> Since post-training was just a small portion of training compute and scaling it yielded huge returns, AI labs focused their limited training compute on scaling it rather than pre-training.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-9" href="#footnote-9" target="_self">9</a></p><p>In fact, these reasoning post-training techniques have scaled <a href="https://epoch.ai/gradient-updates/how-far-can-reasoning-models-scale">much faster than pre-training compute</a>.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-10" href="#footnote-10" target="_self">10</a> At this rate, tripling post-training compute will soon be akin to tripling the entire compute budget &#8211; so current growth rates likely <a href="https://epoch.ai/gradient-updates/how-far-can-reasoning-models-scale">can&#8217;t be sustained</a> for much more than a year.</p><p>That means that this broader trend is likely to end &#8211; we may see a reversion to the original trend of training compute growth. If this is right, GPT-6 is likely to need much more training compute than GPT-5, and probably more than GPT-4.5. Not to mention, OpenAI plans to significantly expand their compute stock, with many more <a href="https://x.com/sama/status/1947057625780396512?t=OwQjfzYJTdYzy8vRCXaJHg">GPUs brought online by the end of the year</a>,<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-11" href="#footnote-11" target="_self">11</a> and major clusters like <a href="https://openai.com/index/five-new-stargate-sites/">Stargate Abilene</a> coming out in phases.</p><p>On the other hand, it&#8217;s possible that the bottlenecks to scaling training compute are harder than we anticipate, e.g. due to limits in the <a href="https://www.theverge.com/2024/12/13/24320811/what-ilya-sutskever-sees-openai-model-data-training">availability of</a> <a href="https://epoch.ai/blog/will-we-run-out-of-data-limits-of-llm-scaling-based-on-human-generated-data">high-quality pre-training data</a> and <a href="https://www.mechanize.work/blog/cheap-rl-tasks-will-waste-compute/">post-training RL environments</a>. It&#8217;s also just hard to pin down <a href="https://epoch.ai/gradient-updates/three-issues-undermining-compute-based-ai-policies">how to measure &#8220;training compute&#8221;</a>. For instance, it&#8217;s plausible that a larger model could be used to generate synthetic data for a smaller one &#8211; so should the total &#8220;training compute&#8221; for the smaller model include the compute for training the larger model?</p><p>Overall, we think we&#8217;ll likely revert back to the trend of training compute growth &#8211; <a href="https://epoch.ai/gradient-updates/compute-scaling-will-slow-down-due-to-increasing-lead-times">though perhaps not for long</a>.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-12" href="#footnote-12" target="_self">12</a> Moreover, while training compute scaling has slowed, infrastructure buildout has continued. This means more compute availability relative to the size of training runs &#8211; so training compute scaling could come back with a vengeance.</p><p><em>We&#8217;d like to thank Josh You, Lynette Bye, and David Owen for their feedback.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://epochai.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading! Subscribe to receive the latest from Epoch AI.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>OpenAI has historically scaled up training compute by around 100&#215; with each (integer) generation of GPT, and <a href="https://x.com/EpochAIResearch/status/1953883611389702169">GPT-5 is an exception to this trend</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>There are several things to note about these numbers. First, R&amp;D compute accounts for both running experiments and the final training run. Second, these are estimates based on The Information&#8217;s projections of OpenAI&#8217;s spending, and are thus quite uncertain. Third, the 2024 R&amp;D compute cost is reported as two separate numbers: $3 billion in the &#8220;compute to train models&#8221;, and $1 billion in &#8220;research compute amortization&#8221;. It&#8217;s not entirely clear what this means, but GPUs are typically amortized over several years. For simplicity we assume amortization over two years, leading to $2 billion + $3 billion = $5 billion in total R&amp;D compute, but our point stands even if we assume amortization over four years.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>The exact number here should be taken as a rough estimate to illustrate the magnitude of the savings, in part because they were estimated using limited data, and in part because the effect size depends on the specific benchmark.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>For illustration, the cost of training Grok 4 was <a href="https://epoch.ai/data-insights/grok-4-training-resources">around $500 million</a>, and we expect GPT-4.5 to be in a similar ballpark (but slightly lower).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>One reason this is harder for larger models is that more time and compute is needed to run the model in post-training environments, a crucial part of post-training.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>GPT-4.5 was likely also an attempt to develop a model called &#8220;GPT-5&#8221;.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>One more reason that OpenAI likely chose to focus on scaling post-training without increasing pre-training compute for GPT-5 is to reduce inference costs. But it&#8217;s not clear how much of a role this played &#8211; for example, if OpenAI were worried about inference speeds or costs, they could&#8217;ve distilled a larger model into a smaller one, making it both faster and cheaper to run.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-8" href="#footnote-anchor-8" class="footnote-number" contenteditable="false" target="_self">8</a><div class="footnote-content"><p>For instance, in the 2.5 years since GPT-4 was released, most state-of-the-art models used less than 3&#215; more training compute than it, such as Claude 3.7 Sonnet and o3. This seems slow given other trends &#8211; the training compute for the most compute-intensive models has been expanding at 5&#215; year. The reason for this discrepancy is that models trained with the most compute aren&#8217;t necessarily state-of-the-art in performance. For example, while models like Llama 4 Behemoth exceeded this 3&#215; threshold, they didn&#8217;t clearly outperform the best models from competing labs on benchmarks.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-9" href="#footnote-anchor-9" class="footnote-number" contenteditable="false" target="_self">9</a><div class="footnote-content"><p>There are also other practical factors that we&#8217;ve alluded to, which also contribute to this trend. For instance, it&#8217;s also easier to scale novel reasoning techniques on smaller models, such that labs have temporarily focused on releasing relatively small models compared to further pre-training scaling.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-10" href="#footnote-anchor-10" class="footnote-number" contenteditable="false" target="_self">10</a><div class="footnote-content"><p>One estimate is that this post-training compute can be <a href="https://epoch.ai/gradient-updates/how-far-can-reasoning-models-scale">scaled up 10&#215; every 4 months</a>. This probably doesn&#8217;t go faster still because of scaling bottlenecks, such as in how quickly the RL environments can be built for RL post-training.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-11" href="#footnote-anchor-11" class="footnote-number" contenteditable="false" target="_self">11</a><div class="footnote-content"><p>While Sam Altman <a href="https://x.com/sama/status/1947057625780396512?t=OwQjfzYJTdYzy8vRCXaJHg">tweeted</a> that OpenAI would have &#8220;well over 1 million GPUs brought online&#8221; by the end of the year, this is unlikely to be purely top-end GPUs. For example, it plausibly includes a mix of A100s, H100s, and Blackwell GPUs.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-12" href="#footnote-anchor-12" class="footnote-number" contenteditable="false" target="_self">12</a><div class="footnote-content"><p>For example, this could happen through <a href="https://tmychow.substack.com/p/pre-training-isnt-dead-its-just-resting">further pre-training scaling</a> or <a href="https://epoch.ai/gradient-updates/how-far-can-reasoning-models-scale">RL post-training scaling</a>.</p><p></p></div></div>]]></content:encoded></item></channel></rss>