<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Planned Obsolescence]]></title><description><![CDATA[Thinking ahead to a future where AI decides everything]]></description><link>https://www.planned-obsolescence.org</link><image><url>https://substackcdn.com/image/fetch/$s_!oyGI!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5519227-580c-4379-999d-03eabb6ed120_1200x1200.png</url><title>Planned Obsolescence</title><link>https://www.planned-obsolescence.org</link></image><generator>Substack</generator><lastBuildDate>Fri, 08 May 2026 23:56:03 GMT</lastBuildDate><atom:link href="https://www.planned-obsolescence.org/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Ajeya Cotra]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[plannedobs@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[plannedobs@substack.com]]></itunes:email><itunes:name><![CDATA[Ajeya Cotra]]></itunes:name></itunes:owner><itunes:author><![CDATA[Ajeya Cotra]]></itunes:author><googleplay:owner><![CDATA[plannedobs@substack.com]]></googleplay:owner><googleplay:email><![CDATA[plannedobs@substack.com]]></googleplay:email><googleplay:author><![CDATA[Ajeya Cotra]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Science and speculation]]></title><description><![CDATA[Will we agree about AI risks in time?]]></description><link>https://www.planned-obsolescence.org/p/science-and-speculation</link><guid isPermaLink="false">https://www.planned-obsolescence.org/p/science-and-speculation</guid><dc:creator><![CDATA[Ajeya Cotra]]></dc:creator><pubDate>Fri, 01 May 2026 14:02:19 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/4357f4a8-2583-496e-89f4-1f2abae6e659_1731x909.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Science is a social institution like the judicial system. Just like courts make you throw out perfectly good Bayesian evidence like whether the defendant seems like a shifty guy to you or rumors you heard about his whereabouts,<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> the scientific process has epistemically conservative norms about what kinds of evidence and arguments are admissible. This means that the process of science gets to the truth <em>much </em>slower than a perfectly rational Bayesian should, and indeed much slower than many particular farseeing individuals actually did. Svante Arrhenius made the first estimate of global warming from human carbon dioxide emissions back in <em>1896</em>, probably about eighty years before something like a &#8220;scientific consensus&#8221; emerged and ninety four years before the first IPCC Assessment Report.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a></p><p>But they say if you want to go fast go alone, if you want to go far go together. At its best, the point of the scientific process is to add bricks to a solid foundation of facts that we can establish &#8220;beyond a reasonable doubt,&#8221; that we can all agree on regardless of our priors because the weight of the evidence we&#8217;ve built up is enough to overwhelm those differences. This is a good piece of social technology to have. We&#8217;re never going to agree on subtle questions of priors, and it&#8217;s great to have a process that lets us at least agree about where the giant likelihood ratios lie. And like our adversarial justice system, scientific institutions are to a significant extent designed around the fact that we all have biases and conflicts of interest &#8212; personal virtue isn&#8217;t as verifiable or scalable as specific norms like preregistration, reproducibility, and peer review.</p><p>This is why, when AI Snake Oil (now known as <a href="https://www.normaltech.ai/">AI As a Normal Technology</a>) wrote <em><a href="https://www.normaltech.ai/p/ai-existential-risk-probabilities?utm_source=publication-search">AI existential risk probabilities are too unreliable to inform policy</a> </em>two years ago, I was much more sympathetic to it than most people who are as concerned as I am about existential risks from AI. While I strongly disagree with the authors on the object level, my guess is that loosening the norms around how much policymaking should be based on science is much more likely to lead to vaccine bans and other destructive crackpottery than thoughtfully crafted AI safety regulation.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a></p><p>The arguments about superintelligence and the Singularity are speculative. &#8220;Speculative&#8221; should not be a pejorative term meaning &#8220;obviously wrong and dismissible&#8221; (Arrhenius&#8217; arguments about CO<sub>2</sub> emissions were speculative), but it does make sense that society draws a normative distinction between speculative and scientifically established claims.</p><p>Unfortunately, these conservative norms have a serious chance of getting us killed in the case of AI. The pace of change in the field is vastly outstripping our ability to develop a well-grounded scientific picture of it. Just last year, METR generated a lot of buzz with an <a href="https://metr.org/blog/2025-07-10-early-2025-ai-experienced-os-dev-study/">RCT</a> showing that <em>early </em>2025 AI tools actually seemed to make open source developers a little <em>slower </em>at their work tasks. By the time my colleagues started looking at <em>late </em>2025 AI tools, the <a href="https://metr.org/blog/2026-02-24-uplift-update/">experiment design broke</a> because participants started dropping out of the study or changing what they worked on because they were so afraid of being randomized into the no-AI arm for their normal tasks.</p><p>Still, the project of scientific research on models&#8217; capabilities and risks doesn&#8217;t seem doomed. We could have lived in a world where superintelligence truly came as a bolt from the blue, like a lot of the speculative writing from ten or fifteen years ago contemplated. The world we find ourselves in is extremely far from that. We used to debate questions like whether people would make AI agents at all, or whether they would be connected to the internet, or whether they would be allowed to run on their own without a human in the loop. We know the answers to all those questions now. We have seen thousands of public examples of AI agents lying, cheating, and working around safeguards to try to achieve their task. With every new model there are new researchers who take catastrophic risks seriously. And precisely <em>because </em>we are so far from a bulletproof consensus on anything and the field moves at such a breakneck pace, there is so much grist for research and testing and sensemaking. The scientific fruit is on the floor.</p><p>We almost certainly won&#8217;t be able to develop an evidence base about AI risks anywhere near as robust as what we have for greenhouse gases causing climate change or tobacco causing lung cancer. But science is a process made out of humans, and a &#8220;scientific consensus&#8221; is ultimately a set of facts the relevant scientists managed to agree about. This develops through some messy combination of building up the foundation of established facts through shared experience and solid experiments, making logical arguments on top of established facts, arguing with one another to establish the right frames to think about things, observing whose predictions turn out to be more right, and much more. I think we do have a shot at building a real-life scientific consensus robust enough to motivate serious technical standards before it&#8217;s too late. AI is going to keep improving at a breakneck pace, and in some ways that makes this an <em>easier</em> problem than tobacco or climate change. Scientific consensus, like everything else, could be dragged along at the speed of AI.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p> See <a href="https://chatgpt.com/s/t_69f432fbbc8081918c506106326a25aa">ChatGPT on hearsay</a>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p> See <a href="https://chatgpt.com/share/69f43598-eb20-83e8-94c9-3d1900a42a03">ChatGPT on the history of climate change</a>, especially the table at the end.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p> The same goes for courts &#8212; there are many cases where courts make the wrong object-level judgment, including many where the conservative epistemic norms of the system are directly responsible for the error. But the alternative to a conservative, adversarial judicial system is probably not a speedy and reliable Bayesian justice system but rather more latitude for the state to punish people who are politically inconvenient.</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[Six milestones for AI automation]]></title><description><![CDATA[What can AI do on its own, and how well?]]></description><link>https://www.planned-obsolescence.org/p/six-milestones-for-ai-automation</link><guid isPermaLink="false">https://www.planned-obsolescence.org/p/six-milestones-for-ai-automation</guid><dc:creator><![CDATA[Ajeya Cotra]]></dc:creator><pubDate>Fri, 03 Apr 2026 14:03:26 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/262daf43-96c0-4a82-9272-781db7a79187_2068x1146.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The closer you think you are to your &#8220;powerful AI&#8221; milestone of choice, the more important it becomes to define it precisely. If your timelines are in the few-year range, changing the particular definition of powerful AI can easily double or triple your median forecast. Take my definition of &#8220;full automation of AI R&amp;D&#8221; from my <a href="https://www.planned-obsolescence.org/p/ai-predictions-for-2026">predictions post</a> earlier this year:</p><blockquote><p><em>[F]or whichever AI company is in the lead [at the relevant point in time], if you fired all its members of technical staff,<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> its rate of technical progress on relevant benchmarks would be slowed down by less than 25% (this is a pretty arbitrary threshold, you can make the milestone more or less extreme by choosing a smaller or bigger number).</em></p></blockquote><p>There&#8217;s a lot that could be made more precise about this definition, but let&#8217;s zoom in on that arbitrary quantitative threshold.</p><p>In every major sector of economic activity, the productivity hit you take from removing humans is and has always been 100%. If suddenly no human could do any of the work people currently do in the agriculture sector, our plows and seed drills and combine harvesters wouldn&#8217;t till and plant and tend and harvest entirely on their own. We have many more machines to help us than we did in 1000 AD, but if we magically prevented all humans from doing any farming work, all our machines would just sit there and we would starve just as surely as our ancestors would have.</p><p>One salient milestone is the very first time the hit from removing humans is smaller than 100% in a given sector &#8212; the first time that machines can<em> just barely </em>produce output in that sector, painstakingly limping along by themselves without any humans to operate them. Let&#8217;s call this milestone <strong>adequacy</strong>. At this point, removing humans would still result in a catastrophic hit to output, but it&#8217;s nonetheless an unprecedented leap that unaided machines can make forward progress at <em>some </em>non-zero pace.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!XLv-!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fede410c3-b80e-4ef3-832b-5c59ef49b8ec_1760x1040.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!XLv-!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fede410c3-b80e-4ef3-832b-5c59ef49b8ec_1760x1040.png 424w, https://substackcdn.com/image/fetch/$s_!XLv-!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fede410c3-b80e-4ef3-832b-5c59ef49b8ec_1760x1040.png 848w, https://substackcdn.com/image/fetch/$s_!XLv-!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fede410c3-b80e-4ef3-832b-5c59ef49b8ec_1760x1040.png 1272w, https://substackcdn.com/image/fetch/$s_!XLv-!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fede410c3-b80e-4ef3-832b-5c59ef49b8ec_1760x1040.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!XLv-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fede410c3-b80e-4ef3-832b-5c59ef49b8ec_1760x1040.png" width="1456" height="860" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ede410c3-b80e-4ef3-832b-5c59ef49b8ec_1760x1040.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:860,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!XLv-!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fede410c3-b80e-4ef3-832b-5c59ef49b8ec_1760x1040.png 424w, https://substackcdn.com/image/fetch/$s_!XLv-!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fede410c3-b80e-4ef3-832b-5c59ef49b8ec_1760x1040.png 848w, https://substackcdn.com/image/fetch/$s_!XLv-!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fede410c3-b80e-4ef3-832b-5c59ef49b8ec_1760x1040.png 1272w, https://substackcdn.com/image/fetch/$s_!XLv-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fede410c3-b80e-4ef3-832b-5c59ef49b8ec_1760x1040.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Note that the threshold of adequacy is defined in terms of the productivity hit you take from<em> removing humans</em>, which is unrelated to the productivity hit you take from <em>removing the machines. </em>These do not have to add up to 100% because humans and machines are complements. Our tractors and weeders and tillers add a huge amount to agricultural productivity, and our yields would probably plummet by well over 90% if we had to give them all up and farm like medieval peasants. But they are still not <em>adequate </em>at farming autonomously, because getting rid of the humans would cut yields by 100%.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Hkvm!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5a9dacd-ef1b-4c4f-9322-797574e71595_1760x1040.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Hkvm!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5a9dacd-ef1b-4c4f-9322-797574e71595_1760x1040.png 424w, https://substackcdn.com/image/fetch/$s_!Hkvm!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5a9dacd-ef1b-4c4f-9322-797574e71595_1760x1040.png 848w, https://substackcdn.com/image/fetch/$s_!Hkvm!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5a9dacd-ef1b-4c4f-9322-797574e71595_1760x1040.png 1272w, https://substackcdn.com/image/fetch/$s_!Hkvm!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5a9dacd-ef1b-4c4f-9322-797574e71595_1760x1040.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Hkvm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5a9dacd-ef1b-4c4f-9322-797574e71595_1760x1040.png" width="1456" height="860" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c5a9dacd-ef1b-4c4f-9322-797574e71595_1760x1040.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:860,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Hkvm!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5a9dacd-ef1b-4c4f-9322-797574e71595_1760x1040.png 424w, https://substackcdn.com/image/fetch/$s_!Hkvm!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5a9dacd-ef1b-4c4f-9322-797574e71595_1760x1040.png 848w, https://substackcdn.com/image/fetch/$s_!Hkvm!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5a9dacd-ef1b-4c4f-9322-797574e71595_1760x1040.png 1272w, https://substackcdn.com/image/fetch/$s_!Hkvm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5a9dacd-ef1b-4c4f-9322-797574e71595_1760x1040.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>While AI agents are clearly adding positively to productivity at AI companies, I think they are not yet at the adequacy milestone for AI R&amp;D. If all humans were magically prevented from doing any AI research or engineering work, and we replaced them all with giant teams of AI agents, I think the agents would likely get catastrophically stuck, and R&amp;D progress would grind to a halt within weeks.</p><p>But this is not entirely obvious! We have AI coding agents that can stay running for a long time, that can spin up subagents and give them instructions, that can write notes to other agents&#8230;we may already have all the ingredients in place. Maybe tens of thousands of these agents churning together writing huge piles of horrendous slop code would eventually, over years, advance the frontier of AI research a bit.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> Who knows? We won&#8217;t easily be able to tell when we first reach adequacy, because no one would bother trying to hand off research to AIs at the first point when they might barely be adequate at it.</p><p>Once AIs reach adequacy in a certain sector, they will keep improving. The next interesting milestone is <strong>parity </strong>&#8212; the first point when getting rid of the AIs slows down progress in the sector <em>more</em> than getting rid of all the humans. The first time parity is reached, it&#8217;s likely that humans will still be better than AIs at many important subtasks, so getting rid of them and forcing the AIs to do those tasks instead would result in a substantial hit to productivity. It&#8217;s just that human labor <em>as a whole </em>is less important for productivity than AI labor.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!2nRD!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F738fa2cc-7e28-4a78-9cbf-0e124c0aac3f_1760x1040.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!2nRD!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F738fa2cc-7e28-4a78-9cbf-0e124c0aac3f_1760x1040.png 424w, https://substackcdn.com/image/fetch/$s_!2nRD!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F738fa2cc-7e28-4a78-9cbf-0e124c0aac3f_1760x1040.png 848w, https://substackcdn.com/image/fetch/$s_!2nRD!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F738fa2cc-7e28-4a78-9cbf-0e124c0aac3f_1760x1040.png 1272w, https://substackcdn.com/image/fetch/$s_!2nRD!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F738fa2cc-7e28-4a78-9cbf-0e124c0aac3f_1760x1040.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!2nRD!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F738fa2cc-7e28-4a78-9cbf-0e124c0aac3f_1760x1040.png" width="1456" height="860" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/738fa2cc-7e28-4a78-9cbf-0e124c0aac3f_1760x1040.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:860,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!2nRD!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F738fa2cc-7e28-4a78-9cbf-0e124c0aac3f_1760x1040.png 424w, https://substackcdn.com/image/fetch/$s_!2nRD!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F738fa2cc-7e28-4a78-9cbf-0e124c0aac3f_1760x1040.png 848w, https://substackcdn.com/image/fetch/$s_!2nRD!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F738fa2cc-7e28-4a78-9cbf-0e124c0aac3f_1760x1040.png 1272w, https://substackcdn.com/image/fetch/$s_!2nRD!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F738fa2cc-7e28-4a78-9cbf-0e124c0aac3f_1760x1040.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>When I defined &#8220;full automation of AI R&amp;D&#8221; as the point when productivity only drops 25% or less from removing humans, I was gesturing at AI systems that are somewhat better than parity in the AI R&amp;D sector.</p><p>Beyond parity, we can talk about <strong>supremacy </strong>&#8212; the first point when productivity in a given sector would actually <em>increase </em>from removing humans. In other words, a major sector of the economy would become like chess. If for some reason you wanted to win as many chess games as possible per dollar, unaided computers would crush human-computer teams right now. A human-AI team <em>can</em> technically draw against an AI opponent, if the human defers to their AI teammate about every move. But if you care at all about resource constraints,<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a> the human is pure deadweight. Supremacy is the point where it&#8217;s no longer worth the overhead to deal with humans or the cost of the food they need to stay alive.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!SYeY!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F18bbac7d-b155-4bda-aa2d-45a5c54440b7_1760x1040.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!SYeY!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F18bbac7d-b155-4bda-aa2d-45a5c54440b7_1760x1040.png 424w, https://substackcdn.com/image/fetch/$s_!SYeY!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F18bbac7d-b155-4bda-aa2d-45a5c54440b7_1760x1040.png 848w, https://substackcdn.com/image/fetch/$s_!SYeY!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F18bbac7d-b155-4bda-aa2d-45a5c54440b7_1760x1040.png 1272w, https://substackcdn.com/image/fetch/$s_!SYeY!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F18bbac7d-b155-4bda-aa2d-45a5c54440b7_1760x1040.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!SYeY!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F18bbac7d-b155-4bda-aa2d-45a5c54440b7_1760x1040.png" width="1456" height="860" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/18bbac7d-b155-4bda-aa2d-45a5c54440b7_1760x1040.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:860,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!SYeY!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F18bbac7d-b155-4bda-aa2d-45a5c54440b7_1760x1040.png 424w, https://substackcdn.com/image/fetch/$s_!SYeY!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F18bbac7d-b155-4bda-aa2d-45a5c54440b7_1760x1040.png 848w, https://substackcdn.com/image/fetch/$s_!SYeY!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F18bbac7d-b155-4bda-aa2d-45a5c54440b7_1760x1040.png 1272w, https://substackcdn.com/image/fetch/$s_!SYeY!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F18bbac7d-b155-4bda-aa2d-45a5c54440b7_1760x1040.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In theory, we can talk separately about the entire spectrum from adequacy through parity to supremacy for any sector, from agriculture to pharmaceuticals to finance. But two domains are likely to set the pace for all the rest: AI research and AI production (that is, the entire stack of chips, fabs, fab equipment manufacturers, power plants, and everything else that goes into producing and running AI systems). If we cross the three automation thresholds with these two domains, we get six milestones. Here are my current best guesses for when we cross these milestones:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!9VXQ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870339da-3a3a-449d-b730-294dfb4495c1_1240x414.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!9VXQ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870339da-3a3a-449d-b730-294dfb4495c1_1240x414.png 424w, https://substackcdn.com/image/fetch/$s_!9VXQ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870339da-3a3a-449d-b730-294dfb4495c1_1240x414.png 848w, https://substackcdn.com/image/fetch/$s_!9VXQ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870339da-3a3a-449d-b730-294dfb4495c1_1240x414.png 1272w, https://substackcdn.com/image/fetch/$s_!9VXQ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870339da-3a3a-449d-b730-294dfb4495c1_1240x414.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!9VXQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870339da-3a3a-449d-b730-294dfb4495c1_1240x414.png" width="1240" height="414" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/870339da-3a3a-449d-b730-294dfb4495c1_1240x414.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:414,&quot;width&quot;:1240,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!9VXQ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870339da-3a3a-449d-b730-294dfb4495c1_1240x414.png 424w, https://substackcdn.com/image/fetch/$s_!9VXQ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870339da-3a3a-449d-b730-294dfb4495c1_1240x414.png 848w, https://substackcdn.com/image/fetch/$s_!9VXQ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870339da-3a3a-449d-b730-294dfb4495c1_1240x414.png 1272w, https://substackcdn.com/image/fetch/$s_!9VXQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870339da-3a3a-449d-b730-294dfb4495c1_1240x414.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>We seem very close<em> </em>to <em>AI research adequacy</em> already. As I said, I wouldn&#8217;t be totally shocked if you told me this milestone had already passed, and I expect it to happen in the next couple years. From there, it seems like another couple of years could take us to <em>AI research parity</em>. When we&#8217;re at <em>AI research parity</em>, we&#8217;ll effectively have millions of researchers working on advancing AI, instead of thousands as we do today. I think this would massively accelerate the pace of AI progress, leading to <em>AI research supremacy </em>within another year.</p><p>The AIs that come out the end of this process would likely be able to pick up new skills with orders of magnitude less data than we need to provide systems today, including the skills required to operate physical actuators like robots and the skills involved in mining, electrical engineering, semiconductor manufacturing, and everything else involved in maintaining the AI stack. From there, reaching <em>AI production adequacy </em>and then <em>parity </em>and <em>supremacy </em>seems like it would mainly be a matter of rolling out AI to enough places and manufacturing enough actuators for it to operate.</p><p>Because <em>AI production supremacy </em>requires AIs and robots to master an extremely diverse range of physical and scientific skills, I think it also entails supremacy on practically any objectively-measurable physical contest. For any goal defined in terms of outcomes in the physical world, where you could explain it to an alien and they could adjudicate whether it had happened &#8212; sending something to space, blowing something up, generating some energy, mining some ore, terraforming a planet &#8212; humans would be worse than useless at contributing to that outcome.</p><p>This is well past the point of barely <a href="https://www.planned-obsolescence.org/p/self-sufficient-ai">self-sufficient AI</a>. At this point, every country on Earth is wholly reliant on its entirely-automated military, and it would be trivial for AI systems to take over the world if they wanted to.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>The obvious legalese applies to that conditional clause: the company can't rehire any technical staff and none of its other staff can ever switch into technical roles. The company can still have human legal staff and ops staff and custodial staff, but no human ever gets to touch anything related to the research or the codebase anymore &#8212; including the "soft" stuff like setting research direction, project management and supervision, or synthesizing varied pieces of evidence to build a holistic scientific picture of a big research question.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>When I wrote my post about <a href="https://www.planned-obsolescence.org/p/self-sufficient-ai">self-sufficient AI</a>, I was vague about exactly what it means for the AI to be self-sufficient &#8212; just how <em>well </em>do they have to be able to sustain themselves? I think it&#8217;s most naturally interpreted as an adequacy threshold. As <a href="https://x.com/snewmanpv/status/2009008644973842713">Steve Newman points out</a>, there&#8217;s a gritty sci fi story like the Martian in there about whether the scrappy band of prototype robots can bootstrap production before they all break down.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>That is, if humans keep maintaining their physical existence; the lines are a bit blurry for AI R&amp;D, or any sector other than the entire AI production stack.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>In clean economic models, you&#8217;ll find that humans are always able to contribute <em>something </em>because of the law of comparative advantage. But this only applies because those models assume that land is infinite (so humans can always go to marginal land and subsistence farm there) and that trading overheads are zero. Both of those are untrue in the real world, so it&#8217;s likely AI systems eventually don&#8217;t bother to trade with humans.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>Note: it&#8217;s certainly possible that humans have been killed <em>before</em> this point; I didn&#8217;t include this conditional to keep the milestones clean, but supremacy on everything comes <em>after </em><a href="https://www.planned-obsolescence.org/p/self-sufficient-ai">self-sufficient AI</a>, the point when AI systems could survive without humans around.</p></div></div>]]></content:encoded></item><item><title><![CDATA[I underestimated AI capabilities (again)]]></title><description><![CDATA[Revisiting a prediction ten months early]]></description><link>https://www.planned-obsolescence.org/p/i-underestimated-ai-capabilities</link><guid isPermaLink="false">https://www.planned-obsolescence.org/p/i-underestimated-ai-capabilities</guid><dc:creator><![CDATA[Ajeya Cotra]]></dc:creator><pubDate>Thu, 05 Mar 2026 15:02:56 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/690d53e2-662e-48fc-8ea8-00c3492b5bcf_1984x1100.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>On Jan 14th, I <a href="https://www.planned-obsolescence.org/p/ai-predictions-for-2026">made predictions</a> about AI progress in 2026. My forecasts for software engineering capabilities already feel much too conservative.</p><p>In my view, <a href="https://metr.org/">METR</a> (where I now work) has some of the hardest and highest-quality software engineering and ML engineering benchmarks out there, and the most useful framework for making benchmark performance intuitive: we measure a task&#8217;s difficulty by the amount of time a human expert would take to complete it (called the <a href="https://metr.org/blog/2025-03-19-measuring-ai-ability-to-complete-long-tasks/">&#8220;time horizon&#8221;</a>).<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a></p><p>When I made my forecasts last month, the model with the longest measured time horizon on METR&#8217;s suite of software engineering tasks was Claude Opus 4.5; it could succeed around half the time at software tasks that would take a human software engineer about five hours.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> Time horizons on software tasks had been doubling a little less than twice a year from 2019 through 2025, which would have implied the state-of-the-art 50% time horizon should be somewhat less than 20 hours by the end of 2026.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> But there was ambiguity about whether the more recent doubling time was faster than the long-run trend, so I bumped that up to 24 hours for my median guess.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a> My 20th percentile was around 15 hours and my 80th percentile was around 40 hours.</p><p>Now, Opus 4.6 (released only 2.5 months after Opus 4.5) was estimated to have a 50% time horizon of ~12 hours.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a> I don&#8217;t take the specific number literally &#8212; there are many fewer very-long tasks than medium and short tasks, and the long tasks more often have guesstimated (rather than measured) human completion times, so time horizon estimates for the latest models are a lot noisier than they were in 2025. And the benchmark underlying the time horizon graph is nearly saturated, which causes the confidence intervals to blow up: the 95% CI is 5.3 hours to 66 hours.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a> It&#8217;s really hard to discriminate between different capability levels at the current range.</p><p>But at the end of the day, that dataset had 19 software engineering tasks estimated<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a> to take humans longer than 8 hours, and Opus 4.6 was able to solve 14 of them at least some of the time (and it reliably nailed four of them).<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-8" href="#footnote-8" target="_self">8</a> And beyond just this one task suite, we&#8217;ve seen examples of AI agents doing certain very well-specified software tasks like writing <a href="https://cursor.com/docs/agent/browser">a browser</a> or <a href="https://www.anthropic.com/engineering/building-c-compiler">C compiler</a>, or <a href="https://garryslist.org/posts/ai-just-ported-simcity-in-4-days-without-reading-the-code">porting a giant game</a>, that would take humans many weeks or months to do on their own &#8212; not perfectly, but better than most people expected and better than a naive reading of the agents&#8217; measured time horizon would have suggested.</p><p>And this happened in February. It&#8217;s no longer very plausible that after ten whole months of additional progress at the recent blistering pace,<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-9" href="#footnote-9" target="_self">9</a> AI agents would still struggle half the time at 24 hour tasks.</p><p>I wish them the best, but I think my colleagues on the capability evaluations team at METR might struggle to create new software tasks from a similar distribution capable of measuring AI agents&#8217; true time horizons through the end of the year. If we could measure this, I&#8217;d guess that by the end of the year, AI agents will have a time horizon of over 100 hours on the sorts of software tasks in METR&#8217;s suite (which are not highly precisely specified &#8212; on certain extremely well-specified software tasks like the examples above, agents seem to <em>already </em>have a time horizon of more than a hundred hours).</p><p>And once you&#8217;re talking about multiple full-time-equivalent <em>weeks</em> of work, I wonder if the whole concept of &#8220;time horizon&#8221; starts to break down.</p><p>It&#8217;s nearly impossible to subdivide a typical one hour task (e.g., debugging one failing test) into smaller pieces that multiple people can work on in parallel. It wouldn&#8217;t go very well if you had to farm out writing this print statement or reading that error message or tweaking this line of code to different people &#8212; the right action to do next depends intimately on everything that came before it and the precise state of the code as a whole, you have to hold the whole context in your mind as you take each action or they won&#8217;t cohere in the right way.</p><p>It&#8217;s somewhat easier to decompose an eight hour task (e.g., writing a simple browser game) into smaller components, but those components are constantly bleeding into each other in ways that make clean handoffs hard. When you&#8217;re implementing the game logic, you realize it needs to know something about how the graphics are rendered. When you&#8217;re handling user input, you find yourself tweaking the game loop. The fastest way to do it is probably one person knocking it out in a day, making a hundred small decisions fluidly as they go.</p><p>But it&#8217;s actually pretty feasible to break down a <em>month</em>-long task into smaller pieces. In fact, you may start <em>benefiting </em>from some explicit decomposition &#8212; it might be helpful to write a design doc laying out how the pieces fit together, or break the work into tickets so you don&#8217;t lose track of what&#8217;s done and what&#8217;s left. And while it might take one person working alone a month to complete, the fastest way to get it done might be to have different people work on different pieces like the checkout flow or the inventory management panel in parallel.</p><p>And of course, tasks that take multiple full-time-equivalent <em>years </em>of work nearly always can and should be broken down into smaller milestones, and parallelized across multiple teammates. Human work appears to get more and more decomposable the longer and longer it gets.</p><p>In other words, very few tasks feel <em>intrinsically </em>like<em> </em>year-long tasks, the way that writing one bash command feels like an intrinsically one-second task, or debugging one simple bug feels intrinsically like a one-hour task. Maybe a mathematician banging their head against a hard conjecture for a year before finally making a breakthrough is a &#8220;real&#8221; year-long task? But most many-person-month software projects in the real world sort of feel like they might be a bunch of few-week tasks in a trenchcoat, the way that a hundred-question elementary school math test is really 100 thirty-second tasks in a trenchcoat.</p><p>If an extreme version of that is true, then once AI agents can consistently do (say) 80-hour tasks, they should be able to make continuous progress on projects of arbitrary scale. Maybe manager AIs can spend their work week figuring out how to farm out the current project goal to line-worker AIs, line-workers can execute on their individual piece, and all the AIs can maintain good enough records that no individual agent needs to build up holistic, long-term state on the whole project.</p><p>I think this <em>probably </em>won&#8217;t fully work. Even projects with lots of formalized goal-tracking still benefit a <em>lot </em>from everyone involved intuitively appreciating the bigger picture in a way that isn&#8217;t  fully captured in Jira tickets and Asana tasks. Decomposing a 6 month project so cleanly and precisely that it can be executed by a team of people with no such holistic context might itself be, say, a 2 month task.</p><p>But it might work surprisingly well for a surprisingly large class of software projects. AI agents are a lot cheaper and a <em>lot </em>more patient than humans, so it could be practical to get them to do far, <em>far </em>more task-tracking, documentation, and other project management than human teams ever do. People have already started aggressively experimenting with scaffolding for orchestrating agent teams. It&#8217;s not clear how far it will go over the next several months.</p><p>This is why my colleague Tom <a href="https://x.com/testingham/status/2027025390578155645">proposed</a> that the calendar time it takes a large <em>team </em>of humans to do a task might be a better proxy for &#8220;intrinsic difficulty&#8221; than the time it takes one human working alone. So far, the &#8220;team time&#8221; and &#8220;solo time&#8221; have been very similar in the METR task suite, since it has ranged from 1 second tasks to maybe 20 hour tasks. But we&#8217;re entering the regime where these numbers could rapidly diverge. If Tom&#8217;s conjecture is true, the &#8220;solo time&#8221; metric should start going super-exponential about now&#8230;which makes it very hard to bound software engineering capabilities by the end of the year.</p><p>In my predictions <a href="https://www.planned-obsolescence.org/p/ai-predictions-for-2026">last month</a>, my probability that AI R&amp;D would be <em>fully </em>automated by the end of the year &#8212; AIs taking care of all the research ideation and implementation, no humans necessary<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-10" href="#footnote-10" target="_self">10</a> &#8212; was 10%. After I published that piece, I heard from a few others in the AI forecasting space (including those I generally think of as more bullish on AI timelines than I am) that it seemed a bit high. But now ten percent feels like it&#8217;s in the right ballpark again.</p><p><em>Fully </em>automating AI R&amp;D still seems like a tall order. Even fully automating software engineering seems like it requires an aggressive read of the evidence, and AI R&amp;D is not just software engineering &#8212; it seems like automating it would require a surprising amount of progress on &#8220;research judgment&#8221; and &#8220;creativity&#8221; and other ephemeral skills that AI systems still appear to be worse at than human researchers. I think it&#8217;s a lot more likely in the coming three or five years than this year.</p><p>But for the first time, I don&#8217;t see any solid trend we can extrapolate to say it won&#8217;t happen soon.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-11" href="#footnote-11" target="_self">11</a> AI R&amp;D really could be automated <em>this year</em>.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>It takes a bit of taste and judgment to decide what a task&#8217;s &#8220;time horizon&#8221; is, and it&#8217;s possible to game the metric to make it meaningless. Consider the &#8220;task&#8221; of taking a giant math test consisting of 10,000 easy elementary school word problems &#8212; this is obviously 10,000 thirty-second tasks, rather than one 83-hour task. Or to take an example suggested by my colleague Tom Cunningham, consider the task &#8220;Count how many times a horse is mentioned in Anna Karenina.&#8221; This might take a single human 10 hours, but it&#8217;s highly parallelizable: a team of 300 people could each take one page and do the task in a minute. In the METR task suite, &#8220;human time to complete&#8221; works as a good proxy for &#8220;intrinsic&#8221; difficulty because the tasks are constructed to be hard to easily decompose and parallelize. Within each task, every piece depends on every other, and you benefit from keeping the whole context in mind. More on that later in the post.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>Actually, at the time, METR was measuring models on its original time horizon suite (TH 1.0), and the precise central estimate for Opus 4.5 was ~4h48min on that distribution. But since then, METR has released an updated suite with more tasks (TH 1.1) which caused all models' scores to shift slightly; the measured time horizon for Opus 4.5 on TH 1.1 is ~5h20min.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>The 2019-2025 doubling time calculated in the <a href="https://arxiv.org/pdf/2503.14499">original paper</a> was 212 days, or 0.58 years. That means that if the time horizon at the beginning of January was 5 hours, the time horizon at the end of the year should be 5 * exp(ln(2) * 1/ 0.58) ~= 16.4 hours. When I was doing mental math, I was approximating the 7 month doubling time as two doublings a year: 5 * 2 * 2 ~= 20 hours. The rule of thumb was a bit more aggressive than the strict extrapolation, but in fact, I just realized while writing this post that the mental extrapolation was too <em>conservative </em>in a different way &#8212; I should have dated the ~5 hour time horizon to Nov 24 (Opus 4.5&#8217;s release date), not the beginning of January, meaning I should have added an extra month to all these extrapolations.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>I didn&#8217;t actually do this math at the time, but the <a href="https://arxiv.org/pdf/2503.14499">original paper</a> calculated a doubling time of 118 days or 0.32 years since 2024, and if you assume that doubling time was correct, then the time horizon by EOY 2026 should have been 5 * exp(ln(2) * 1/0.32) ~= 43 hours. Given that I took this faster doubling time pretty seriously at this time, this suggests my median and especially my 80th percentile should have been higher to begin with. I would probably have been better served by the heuristic of using just the previous year (rather than the previous seven years) to forecast the next year.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>Originally, METR estimated Opus 4.6 to have a 50% time horizon of <a href="https://x.com/METR_Evals/status/2024923422867030027">~14.5 hours</a> on Feb 20, 2026. We <a href="https://x.com/METR_Evals/status/2028948235486937098">corrected a bug in our modeling</a> on March 3, 2026 and this reduced its time horizon estimate to ~12 hours. Note that this measurement was done on Time Horizon 1.1, the new task suite released on Jan 29, 2026.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>This is something I appreciate about the time horizon construct. For standard benchmarks, confidence intervals around a point get <em>narrower </em>as the benchmark approaches saturation: if a model gets 50% accuracy, there&#8217;s more room for error in either direction than if it gets 95% accuracy. But since the time horizon metric could get infinitely long, the confidence intervals can be constructed so uncertainty gets <em>wider </em>rather than narrower as the current hardest tasks are saturated. Epistemically, it&#8217;s appropriate for your uncertainty about real capabilities to blow up when you no longer have tasks that the models can&#8217;t complete.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>In the original time horizon <a href="https://arxiv.org/abs/2503.14499">paper</a>, METR measured human completion times for most of the tasks in the dataset by actually having humans do them (148 out of 169). In the updated suite, only 5 of the 19 tasks longer than 8 hours have measured human baselines &#8212; the others are estimated.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-8" href="#footnote-anchor-8" class="footnote-number" contenteditable="false" target="_self">8</a><div class="footnote-content"><p>Agents are usually run on the same task 6 separate times. The same agent doesn&#8217;t always approach the same task the same way each time &#8212; it sometimes gets lucky or unlucky. For four of the 19 hard tasks, Claude Opus 4.6 solved it successfully in all six runs; for another ten, it solved it successfully in at least one of the six runs.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-9" href="#footnote-anchor-9" class="footnote-number" contenteditable="false" target="_self">9</a><div class="footnote-content"><p>If you look just at the year 2025, agent time horizons doubled every <a href="https://x.com/RyanPGreenblatt/status/2020301259904049620">~</a><em><a href="https://x.com/RyanPGreenblatt/status/2020301259904049620">3.5 months</a></em>, not every ~7 months as in the long-run trend or even every ~4 months as in the 2024-2025 trend. I mostly didn&#8217;t factor this into my forecasts; it wasn&#8217;t salient to me compared to the rule of thumb I&#8217;d absorbed of &#8220;two doublings a year.&#8221; If I had used this to extrapolate from Opus 4.5, that would have suggested time horizons by the end of the year should be 5 * exp(ln(2) * 12/3.5) = 54 hours by the end of the year.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-10" href="#footnote-anchor-10" class="footnote-number" contenteditable="false" target="_self">10</a><div class="footnote-content"><p>Specifically, my operationalization was that firing all human members of technical staff (research leads, engineers, everyone) would only slow down progress by 25%. Now and in the past, of course, firing every single human would cause progress to completely halt.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-11" href="#footnote-anchor-11" class="footnote-number" contenteditable="false" target="_self">11</a><div class="footnote-content"><p> A concrete example of this is the kind of argument that my colleague Nikola made <a href="https://nikolajurkovic.substack.com/p/how-likely-is-dangerous-ai-in-the">here</a>, in Nov 2025, when the state-of-the-art time horizon was 2 hours; this kind of argument would be much shakier made today (less than four months later!).</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[Takeoff speeds rule everything around me]]></title><description><![CDATA[Not all short timelines are created equal]]></description><link>https://www.planned-obsolescence.org/p/takeoff-speeds-rule-everything-around</link><guid isPermaLink="false">https://www.planned-obsolescence.org/p/takeoff-speeds-rule-everything-around</guid><pubDate>Thu, 12 Feb 2026 15:03:27 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!y0b_!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa3b4898b-1c83-4e3a-beb3-212cf75ca06c_1125x675.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>A decade ago, debates about AI x-risk often centered on AGI timelines. Was it over a century away, or was it <a href="https://coefficientgiving.org/research/some-background-on-our-views-regarding-advanced-artificial-intelligence/">plausibly as soon as 20 years</a>? Today, even people relatively skeptical of x-risk <a href="https://helentoner.substack.com/p/long-timelines-to-advanced-ai-have">often have crazy short timelines</a> like ten years or five years or <a href="https://x.com/deanwball/status/2001068539990696422">negative two months</a>. So what accounts for the vast remaining disagreement on the level of imminent risk and the appropriate policy stance toward it?</p><p>I think it&#8217;s secretly still mostly about timelines. These days, I suspect many people who aren&#8217;t particularly concerned about x-risk combine very short timelines to AGI<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> with very <em>long</em> timelines to AI transforming the real world &#8212; in other words, they believe <em>takeoff speed </em>will be very slow.</p><p>The classic definition of takeoff speed is the amount of time it takes to go from AGI to superintelligence,<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> but both of those terms are very slippery. And it&#8217;s hard to articulate exactly how &#8220;super&#8221; an intelligence we&#8217;re talking without specifying what impressive <em>feats </em>it should be able to achieve in the real world. Instead, I think it can be more illuminating to operationalize &#8220;takeoff speed&#8221; with respect to outcomes in the external world caused by AI. </p><p>One crucial yardstick for takeoff is the speed of progress in physical technology.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> </p><p>Consider the average rate at which scientists and engineers have been making new discoveries or improving the efficiency of existing processes across a wide range of hard-tech fields (hardware, batteries, materials, industrial chemistry, robotics, spacecraft) over the last 20 years or so. </p><p>Now suppose that sometime in 2028, we develop AI that can automate <em>all </em>the intellectual work that all these scientists are doing across all these fields. How many &#8220;years of progress&#8221; (at the old human-driven pace) will these fields make each calendar year?</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!SFx-!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdfb5542a-962e-4827-90f7-8645c3c71b64_1125x675.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!SFx-!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdfb5542a-962e-4827-90f7-8645c3c71b64_1125x675.png 424w, https://substackcdn.com/image/fetch/$s_!SFx-!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdfb5542a-962e-4827-90f7-8645c3c71b64_1125x675.png 848w, https://substackcdn.com/image/fetch/$s_!SFx-!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdfb5542a-962e-4827-90f7-8645c3c71b64_1125x675.png 1272w, https://substackcdn.com/image/fetch/$s_!SFx-!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdfb5542a-962e-4827-90f7-8645c3c71b64_1125x675.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!SFx-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdfb5542a-962e-4827-90f7-8645c3c71b64_1125x675.png" width="1125" height="675" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/dfb5542a-962e-4827-90f7-8645c3c71b64_1125x675.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:675,&quot;width&quot;:1125,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!SFx-!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdfb5542a-962e-4827-90f7-8645c3c71b64_1125x675.png 424w, https://substackcdn.com/image/fetch/$s_!SFx-!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdfb5542a-962e-4827-90f7-8645c3c71b64_1125x675.png 848w, https://substackcdn.com/image/fetch/$s_!SFx-!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdfb5542a-962e-4827-90f7-8645c3c71b64_1125x675.png 1272w, https://substackcdn.com/image/fetch/$s_!SFx-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdfb5542a-962e-4827-90f7-8645c3c71b64_1125x675.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Here&#8217;s a classic &#8220;fast takeoff&#8221; view:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ui5C!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb4ed3bd7-15b2-4549-995a-41bed3d24409_1125x675.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ui5C!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb4ed3bd7-15b2-4549-995a-41bed3d24409_1125x675.png 424w, https://substackcdn.com/image/fetch/$s_!ui5C!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb4ed3bd7-15b2-4549-995a-41bed3d24409_1125x675.png 848w, https://substackcdn.com/image/fetch/$s_!ui5C!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb4ed3bd7-15b2-4549-995a-41bed3d24409_1125x675.png 1272w, https://substackcdn.com/image/fetch/$s_!ui5C!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb4ed3bd7-15b2-4549-995a-41bed3d24409_1125x675.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ui5C!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb4ed3bd7-15b2-4549-995a-41bed3d24409_1125x675.png" width="1125" height="675" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b4ed3bd7-15b2-4549-995a-41bed3d24409_1125x675.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:675,&quot;width&quot;:1125,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ui5C!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb4ed3bd7-15b2-4549-995a-41bed3d24409_1125x675.png 424w, https://substackcdn.com/image/fetch/$s_!ui5C!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb4ed3bd7-15b2-4549-995a-41bed3d24409_1125x675.png 848w, https://substackcdn.com/image/fetch/$s_!ui5C!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb4ed3bd7-15b2-4549-995a-41bed3d24409_1125x675.png 1272w, https://substackcdn.com/image/fetch/$s_!ui5C!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb4ed3bd7-15b2-4549-995a-41bed3d24409_1125x675.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>On this view, once AI fully automates further AI R&amp;D, this will kick off a strong <a href="https://www.forethought.org/research/will-ai-r-and-d-automation-cause-a-software-intelligence-explosion">super-exponential feedback loop</a> that leads to unfathomably superhuman AI within months. While that software-based intelligence explosion is going on, you don&#8217;t really see physical technology improve all that much.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a> But once we have this god-like AI, it&#8217;ll take only months or weeks or <em>days </em>to create sci-fi technologies (molecular nanotechnology, whole brain emulation, reversible computing, near-light-speed spacecraft) that would have taken centuries at human rates.</p><p>In contrast, here&#8217;s a view that would be considered a &#8220;slow takeoff&#8221;, again conditioning on full cognitive automation of science in mid-2028:<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!uLDA!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6a72605-060f-4855-866b-ae15ab181154_1125x675.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!uLDA!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6a72605-060f-4855-866b-ae15ab181154_1125x675.png 424w, https://substackcdn.com/image/fetch/$s_!uLDA!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6a72605-060f-4855-866b-ae15ab181154_1125x675.png 848w, https://substackcdn.com/image/fetch/$s_!uLDA!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6a72605-060f-4855-866b-ae15ab181154_1125x675.png 1272w, https://substackcdn.com/image/fetch/$s_!uLDA!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6a72605-060f-4855-866b-ae15ab181154_1125x675.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!uLDA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6a72605-060f-4855-866b-ae15ab181154_1125x675.png" width="1125" height="675" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f6a72605-060f-4855-866b-ae15ab181154_1125x675.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:675,&quot;width&quot;:1125,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!uLDA!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6a72605-060f-4855-866b-ae15ab181154_1125x675.png 424w, https://substackcdn.com/image/fetch/$s_!uLDA!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6a72605-060f-4855-866b-ae15ab181154_1125x675.png 848w, https://substackcdn.com/image/fetch/$s_!uLDA!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6a72605-060f-4855-866b-ae15ab181154_1125x675.png 1272w, https://substackcdn.com/image/fetch/$s_!uLDA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff6a72605-060f-4855-866b-ae15ab181154_1125x675.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>On the &#8220;slow&#8221; takeoff view, there is <em>still </em>a super-exponential feedback loop of AI-improving-AI. It&#8217;s just that we need to <a href="https://www.forethought.org/research/the-industrial-explosion">incorporate physical automation</a> of <a href="https://www.planned-obsolescence.org/p/self-sufficient-ai">the entire AI stack</a> to get that feedback loop going, so it&#8217;s more gradual (and we actually see physical signatures <em>before </em>we get the full automated scientist). But even &#8220;slow&#8221; takeoff still means that we go from automated science to an unrecognizably sci-fi world within a matter of years.</p><p>Meanwhile, I suspect most x-risk skeptics think that AI automating scientific research won&#8217;t be that big a deal. Perhaps it makes R&amp;D go modestly faster, or perhaps AI automation is necessary just to keep us going at our previous pace, when otherwise we would have stagnated. They think <em>there will be no takeoff at all.</em></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!y0b_!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa3b4898b-1c83-4e3a-beb3-212cf75ca06c_1125x675.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!y0b_!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa3b4898b-1c83-4e3a-beb3-212cf75ca06c_1125x675.png 424w, https://substackcdn.com/image/fetch/$s_!y0b_!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa3b4898b-1c83-4e3a-beb3-212cf75ca06c_1125x675.png 848w, https://substackcdn.com/image/fetch/$s_!y0b_!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa3b4898b-1c83-4e3a-beb3-212cf75ca06c_1125x675.png 1272w, https://substackcdn.com/image/fetch/$s_!y0b_!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa3b4898b-1c83-4e3a-beb3-212cf75ca06c_1125x675.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!y0b_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa3b4898b-1c83-4e3a-beb3-212cf75ca06c_1125x675.png" width="1125" height="675" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a3b4898b-1c83-4e3a-beb3-212cf75ca06c_1125x675.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:675,&quot;width&quot;:1125,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!y0b_!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa3b4898b-1c83-4e3a-beb3-212cf75ca06c_1125x675.png 424w, https://substackcdn.com/image/fetch/$s_!y0b_!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa3b4898b-1c83-4e3a-beb3-212cf75ca06c_1125x675.png 848w, https://substackcdn.com/image/fetch/$s_!y0b_!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa3b4898b-1c83-4e3a-beb3-212cf75ca06c_1125x675.png 1272w, https://substackcdn.com/image/fetch/$s_!y0b_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa3b4898b-1c83-4e3a-beb3-212cf75ca06c_1125x675.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>This one parameter underlies a huge range of disagreements. Will the country that first develops AGI gain a &#8220;decisive strategic advantage&#8221; &#8212; the ability to impose its will on everyone else? Well, if a six month lead in AI translates into a <em>centuries</em>-long lead in weapons technology, probably. Could<em> </em>misaligned AI systems<em> </em>drive humanity extinct? If they can trivially develop a superplague to wipe us out and upload themselves into self-replicating nanocomputers to <a href="https://www.planned-obsolescence.org/p/self-sufficient-ai">survive without us</a>, then sure. Should we try to slow down AI? If it means that we get a year or two to absorb the impacts of a world with technology from 2100 before we have to deal with technology from 2200, then yes please.</p><p>In my view, this has always been at the heart of the disagreement, and it still is. But now that some level of powerful AI (say at least an 8 on <a href="https://res.cloudinary.com/lesswrong-2-0/image/upload/f_auto,q_auto/v1/mirroredImages/oAy72fcqDHsCvLBKz/rqihhwin7ddif1kn6w9e">Nate Silver&#8217;s Richter Scale</a>) feels around the corner to doomers and skeptics alike, the deeper disagreement about just how powerful it&#8217;ll get and how quickly has been obscured. Some people think AGI is the next innovation that allows us to sustain 2% frontier economic growth a little while longer, or maybe add <a href="https://marginalrevolution.com/marginalrevolution/2025/02/why-i-think-ai-take-off-is-relatively-slow.html">a half a percentage point more</a>. Others think we are t minus a few years from first contact with an unfathomably advanced alien species, entities we would view as nothing less than gods.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>Or at least, something they <em>call </em>&#8220;AGI.&#8221; As I discussed <a href="https://www.planned-obsolescence.org/p/self-sufficient-ai">last time</a>, I think the definition of AGI is often unproductively watered down. The people with the very shortest timelines (e.g. 1 or 2 years) are disproportionately likely to be forecasting a milder version of &#8220;AGI,&#8221; which in turn exaggerates how long they believe the takeoff period will be from AGI to radical superintelligence.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>This is the definition given by Nick Bostrom in his 2014 book <em><a href="https://www.amazon.com/Superintelligence-Dangers-Strategies-Nick-Bostrom/dp/0198739834">Superintelligence</a></em>.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>Another important real-world yardstick of takeoff is economic output, or gross world product. This is the operationalization used in <a href="https://sideways-view.com/2018/02/24/takeoff-speeds/">this</a> 2018 blog post by Paul Christiano and <a href="https://epoch.ai/gate">Epoch&#8217;s GATE model</a>, and is the central operationalization used in <a href="https://coefficientgiving.org/research/report-on-whether-ai-could-drive-explosive-economic-growth/">Tom Davidson&#8217;s 2021 takeoff speeds report</a> (though that report also models several other metrics like software and hardware; you can play with the model <a href="https://takeoffspeeds.com/">here</a>). This is also really important, but it introduces some complications that aren&#8217;t essential for the basic argument.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>If this graph was a graph of the <em>cognitive </em>capabilities of AI systems, it would be rapidly but smoothly growing over the whole period from 2026 through 2029, like the <a href="https://www.aifuturesmodel.com/">AI Futures Project&#8217;s takeoff model</a>. But I think the graph of physical technology better illustrates how the takeoff will <em>feel </em>to people outside AI companies.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>I wanted to illustrate different views about takeoff <em>conditional </em>on a fixed timeline to fully-automated science, but in reality if you have slower takeoff speeds, you probably also have longer timelines to that initial milestone. (And there would be room for even more acceleration prior to full automation.)</p></div></div>]]></content:encoded></item><item><title><![CDATA[AI predictions for 2026]]></title><description><![CDATA[But first, scoring my predictions for 2025]]></description><link>https://www.planned-obsolescence.org/p/ai-predictions-for-2026</link><guid isPermaLink="false">https://www.planned-obsolescence.org/p/ai-predictions-for-2026</guid><dc:creator><![CDATA[Ajeya Cotra]]></dc:creator><pubDate>Wed, 14 Jan 2026 15:03:38 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!SlhK!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ccfd02e-0a3c-4d48-90ee-1ecda468d699_1348x636.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>On December 13th, 2024, I <a href="https://x.com/ajeya_cotra/status/1867813307073409333">registered predictions</a> about what would happen with AI by the end of 2025, using <a href="https://ai2025.org/">this survey</a> run by <a href="https://sage-future.org/">Sage</a>. They asked five questions about benchmarks, four about the OpenAI <a href="https://cdn.openai.com/pdf/18a02b5d-6b67-4cec-ab64-68cdfbddebcd/preparedness-framework-v2.pdf">Preparedness Framework</a> risk categories, one about revenues, and one about public salience of AI.</p><p>The Sage team has updated the survey with the resolutions, so I went through and looked back at what I said. Overall, I was somewhat too bullish about benchmarks scores and much too <em>bearish</em> about AI revenue &#8212; the <a href="https://x.com/davidshor/status/2009031520498119082">reverse</a> of the conventional wisdom that people underestimate benchmark progress and overestimate real-world utility (something which I felt like I&#8217;ve done in previous years, though I didn&#8217;t register clear predictions so it&#8217;s hard to say).</p><p>You can see how I did on benchmark scores in the table below.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> I overestimated progress on pretty much every benchmark other than <a href="https://epoch.ai/frontiermath">FrontierMath</a> Tiers 1-3,<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> which notoriously jumped from ~2% to ~24% with <a href="https://www.youtube.com/watch?v=SKBG1sqdyIU&amp;t=4:05">the announcement of OpenAI&#8217;s o3</a>, about a week after I registered my prediction (and less than two months after the benchmark was published).</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!SlhK!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ccfd02e-0a3c-4d48-90ee-1ecda468d699_1348x636.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!SlhK!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ccfd02e-0a3c-4d48-90ee-1ecda468d699_1348x636.png 424w, https://substackcdn.com/image/fetch/$s_!SlhK!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ccfd02e-0a3c-4d48-90ee-1ecda468d699_1348x636.png 848w, https://substackcdn.com/image/fetch/$s_!SlhK!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ccfd02e-0a3c-4d48-90ee-1ecda468d699_1348x636.png 1272w, https://substackcdn.com/image/fetch/$s_!SlhK!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ccfd02e-0a3c-4d48-90ee-1ecda468d699_1348x636.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!SlhK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ccfd02e-0a3c-4d48-90ee-1ecda468d699_1348x636.png" width="1348" height="636" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2ccfd02e-0a3c-4d48-90ee-1ecda468d699_1348x636.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:636,&quot;width&quot;:1348,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!SlhK!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ccfd02e-0a3c-4d48-90ee-1ecda468d699_1348x636.png 424w, https://substackcdn.com/image/fetch/$s_!SlhK!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ccfd02e-0a3c-4d48-90ee-1ecda468d699_1348x636.png 848w, https://substackcdn.com/image/fetch/$s_!SlhK!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ccfd02e-0a3c-4d48-90ee-1ecda468d699_1348x636.png 1272w, https://substackcdn.com/image/fetch/$s_!SlhK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ccfd02e-0a3c-4d48-90ee-1ecda468d699_1348x636.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>* Shortly after my prediction OpenAI <a href="https://www.youtube.com/watch?v=SKBG1sqdyIU?t=2:06">claimed</a> o3 got 72%.   ** Shortly after my prediction OpenAI <a href="https://www.youtube.com/watch?v=SKBG1sqdyIU&amp;t=4:05">claimed</a> o3 got 24%      *** Anthropic only reimplemented 39 out of the 40 tasks, Opus 4.5 got 32 / 39 right.</em></figcaption></figure></div><p>All the other questions are in the table below. I did fine on the OpenAI Preparedness Framework predictions and was somewhat too bullish on public salience. But I completely bombed the annualized revenue question, predicting that it would merely triple from ~$4.5B to ~$17B instead of nearly <em>7</em>xing<em> </em>(!)<em> </em>to ~$30.5B.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!VHOa!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3675c543-2471-4172-9719-1adf99b071bb_1600x1108.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!VHOa!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3675c543-2471-4172-9719-1adf99b071bb_1600x1108.png 424w, https://substackcdn.com/image/fetch/$s_!VHOa!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3675c543-2471-4172-9719-1adf99b071bb_1600x1108.png 848w, https://substackcdn.com/image/fetch/$s_!VHOa!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3675c543-2471-4172-9719-1adf99b071bb_1600x1108.png 1272w, https://substackcdn.com/image/fetch/$s_!VHOa!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3675c543-2471-4172-9719-1adf99b071bb_1600x1108.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!VHOa!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3675c543-2471-4172-9719-1adf99b071bb_1600x1108.png" width="1456" height="1008" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3675c543-2471-4172-9719-1adf99b071bb_1600x1108.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1008,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!VHOa!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3675c543-2471-4172-9719-1adf99b071bb_1600x1108.png 424w, https://substackcdn.com/image/fetch/$s_!VHOa!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3675c543-2471-4172-9719-1adf99b071bb_1600x1108.png 848w, https://substackcdn.com/image/fetch/$s_!VHOa!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3675c543-2471-4172-9719-1adf99b071bb_1600x1108.png 1272w, https://substackcdn.com/image/fetch/$s_!VHOa!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3675c543-2471-4172-9719-1adf99b071bb_1600x1108.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>I&#8217;m not totally sure what&#8217;s up here,<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> since I&#8217;ve heard from other sources that AI revenue has been roughly 3-4xing year on year. My best guess is the Sage resolution team pulled lowball reports for the 2024 baseline revenue; one friend who looked into it a bit said that EO2024 revenue was ~$6.4B rather than ~$4.5B as Sage reported. If true, this would make my prediction only somewhat too bearish.</p><h2>Predictions for 2026</h2><p>All these forecasts are evaluated as of 11:59 pm Pacific time on Dec 31 2026. I&#8217;ll start with giving my medians for some measurable quantities with pre-existing trends,:</p><ul><li><p><strong>50% <a href="https://metr.org/blog/2025-03-19-measuring-ai-ability-to-complete-long-tasks/">METR time horizon</a>: 24 hours.</strong> Currently, Claude Opus 4.5 has the longest reported 50% time horizon on this task suite, at 4h49m &#8212; meaning that METR&#8217;s model predicts it can to solve about half of the programming tasks that take a low-context human expert five hours (it&#8217;ll be able to solve a greater fraction of shorter tasks, and a smaller fraction of longer tasks). My median for the longest 50% time horizon reported as of Dec 31, 2026 is 24 hours (20th percentile 15 hours, 80th percentile is that it&#8217;s too long for METR to accurately bound in practice but probably around 40 hours in &#8220;reality&#8221;).</p><ul><li><p>Note that to be able to measure even 24 hours accurately, METR would need to make new tasks, which they&#8217;re furiously working on now; this will somewhat change the task distribution, but I&#8217;m not thinking too hard about that now.</p></li></ul></li><li><p><strong><a href="https://epoch.ai/benchmarks/eci">Epoch ECI</a>: 169. </strong>This is a simple abstract model aggregating multiple benchmarks; it currently includes 37 distinct benchmarks but it can be extended to incorporate new ones. The current top score is 154, achieved by Gemini 3 Pro, and it <a href="https://epoch.ai/data-insights/ai-capabilities-progress-has-sped-up">has recently been growing</a> at 15.5 points per year. The aggregated score is in abstract units, but <a href="https://epoch.ai/data-insights/interpreting-eci">you can back out</a> an implicit prediction about particular benchmarks from that. For example, it seems like an ECI of 169 corresponds to a score of about 85% on Frontier Math Tiers 1-3 (though I expect the actual score to be a little lower, maybe 80%, since progress will probably slow down as it approaches saturation). My median prediction for Frontier Math Tier 4, and any other benchmark included in ECI that&#8217;s far from saturation, is whatever the ECI score implies it should be.</p></li><li><p><strong>Annualized Dec AI revenue: $110 billion. </strong>The combined annualized revenue of OpenAI, Anthropic, and xAI (i.e. December revenue * 12) was $30.5 billion in December 2025, according to Sage. My median<strong> </strong>forecast of the same value for the month of Dec 2026 is $110 billion<strong> </strong>(20th percentile $60 billion, 80th percentile $300 billion). This is a ~5x increase, which is my current best guess for the actual increase from EO2024 to EO2025. <em>[Edit 2/18: I noticed that $110 billion is not 5x $30.5 billion; guessing $110 billion was a typo and I meant to say $150 billion here]</em></p></li><li><p><strong>Salience of AI as top issue: 2%. </strong>As of December 2025, 1% of people named &#8220;advancement of computers / technology&#8221; or equivalent as the most important issue in <a href="https://news.gallup.com/poll/1675/most-important-problem.aspx">Gallup&#8217;s monthly polling</a>. My median for December 2026 is 2% (20th percentile 1%, 80th percentile 7%).</p><ul><li><p>Unfortunately this is a bit of an awkward metric &#8212; I&#8217;d rather track something like &#8220;median rank of AI among issues.&#8221; But I&#8217;m not aware of regular polling on that style of question.</p></li></ul></li><li><p><strong><a href="https://x.com/davidshor/status/2005337442161959389/photo/1">Net AI favorability</a>: +4%. </strong>David Shor&#8217;s recent polling found that optimism about AI beats pessimism by about 4.4 percentage points, with many people (24% still unsure). This was surprising to me given how salient concern about jobs and the environment and artists is in the discourse, but AI products are objectively pretty amazing and probably improving a lot of people&#8217;s lives. I think this will keep being true in 2026 so if a methodologically similar poll were conducted near the end of the year 2026 I don&#8217;t have much reason to think the result would be super different. My error bars are wide though, with more room on the downside than the upside: my 20th percentile is -10 points and my 80th percentile is +8 points.</p></li></ul><p>Next, some vibes-y capability predictions. Here are a few tasks that I&#8217;m pretty sure AIs can&#8217;t do today but probably <em>will</em> be able to do as of Dec 31 2026.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a> I&#8217;m about 80% on these, so I&#8217;m expecting to get four of these five predictions right. Note that some of these will be hard to directly test given the high specified inference budgets, in which case I&#8217;ll go with the best judgment of my friends who really understand current AI capabilities.</p><ul><li><p><strong>Game play. </strong>Play Pokemon or a similarly difficult video game at least as well as a typical ten year old, with no special fine-tuning and only a <a href="https://www.lesswrong.com/posts/u6Lacc7wx4yYkBQ3r/insights-into-claude-opus-4-5-from-pokemon#:~:text=This%20is%20because,a%20leg%20up.">generic scaffold</a> that doesn&#8217;t do any more hand-holding than the Claude Plays Pokemon scaffold.</p></li><li><p><strong>Logistics. </strong>Organize a child&#8217;s birthday party for 20ish guests: find a good time that works for guests of honor, compose an invite email, keep track of RSVPs, order the right amount and diversity of food, order cake / decorations / pi&#241;ata on the right theme, etc while abiding by relevant constraints like budget and dietary restrictions.</p></li><li><p><strong>Video. </strong>From a high-level one-paragraph prompt, make a 4 minute short video with at least two characters where a series of somewhat-coherent plot beats happen and there&#8217;s no glaringly obvious visual incoherence or degeneracy.</p></li><li><p><strong>Game design. </strong>From a high-level one paragraph prompt, make a decent <a href="https://en.wikipedia.org/wiki/Visual_novel">visual novel</a> (simple choose-your-own adventure game) that offers at least two hours of gameplay with a similar quality to the trashiest visual novel games you can find on Steam (that still work and have at least dozens of people buying them).</p></li><li><p><strong>Math. </strong>Solve the hardest problem in the 2026 IMO (models got gold in the 2025 IMO but all failed to solve problem 6, which is typically the hardest problem in any IMO).</p></li></ul><p>And tasks that I&#8217;m pretty sure AIs can&#8217;t do today and probably <em>still </em>won&#8217;t<em> </em>be able to do as of Dec 31 2026 (again expecting to get around 4 out of 5 correct).</p><ul><li><p><strong>Game play. [</strong><em>Edited for clarity 1/17]. </em>Get a win rate in the upcoming <a href="https://store.steampowered.com/app/2868840/Slay_the_Spire_2/">Slay the Spire 2</a> comparable to a top player who&#8217;s been playing for 50 hours. I&#8217;m assuming the AI wasn&#8217;t pre-trained on guides for the game (if it is pre-trained on Slay the Spire 2, choose a similarly complicated game released later) and it has a generic harness and was not fine-tuned to play the game well, but does get a comparable amount of time to learn and practice the game as the human.</p></li><li><p><strong>Logistics. </strong>Organize a typical-complexity wedding with 100 guests: find a suitable venue that meets the budget and constraints, go back and forth with caterers and photographers and other vendors, track RSVPs and maintain a seating chart, schedule toasts, etc.</p></li><li><p><strong>Video. </strong>From a one-paragraph high-level prompt, make a &gt;10 minute short film that&#8217;s hard for at least me (not a film buff) to easily distinguish from the kinds of short films that make it into film festivals (but don&#8217;t necessarily win awards there).</p></li><li><p><strong>Game design. </strong>From a one-paragraph high-level prompt, make an original text adventure game that offers at least ten hours of gameplay that I consider to be as good as <a href="https://ifdb.org/viewgame?id=aearuuxv83plclpl">Counterfeit Monkey</a> or an original visual novel that I consider to be decidedly better than <a href="https://store.steampowered.com/app/251990/Long_Live_The_Queen/">Long Live the Queen</a>.</p></li><li><p><strong>Math. </strong>From scratch, write a paper that could get published in a top theoretical computer science or math journal / conference.</p></li></ul><p>And lastly, some more extreme milestones I think are less likely still:</p><ul><li><p>(near) <strong>Full automation of AI R&amp;D: 10%. </strong>I currently think if you fired all the human technical staff working on AI R&amp;D and tried to get AIs to do it all, technical progress would basically grind to a halt.<strong> </strong>My operationalization of &#8220;full automation&#8221; is: for whichever AI company is in the lead as of Dec 31, 2026, if you fired all its members of technical staff, its rate of technical progress on relevant benchmarks would be slowed down by less than 25% (this is a pretty arbitrary threshold, you can make the milestone more or less extreme by choosing a smaller or bigger number). Subjectively, I think it would feel like human researchers setting broad high-level multi-person-year AI research goals (e.g. &#8220;develop more sample-efficient optimization algorithms&#8221;) to teams of thousands of AI agents who run off and autonomously conduct dozens of experiments across hundreds of GPUs to make progress on the goal.</p></li><li><p><strong><a href="https://www.alignmentforum.org/posts/LjgcRbptarrRfJWtR/a-breakdown-of-ai-capability-levels-focused-on-ai-r-and-d#:~:text=Top%2Dhuman%2DExpert%2DDominating%20AI%20(TEDAI)%3A%20AIs%20which%20strictly%20dominate%20top%20human%20experts%5B5%5D%20in%20virtually%20all%20cognitive%20tasks%20(i.e.%2C%20doable%20via%20remote%20work)%20while%20being%20at%20least%202x%20faster%5B6%5D%20and%20within%20a%20factor%20of%205%20on%20cost%5B7%5D.">Top-human-expert dominating AI</a>: 5%. </strong>This means an AI system that achieves ambitious long-term goals in any domain better than the world&#8217;s leading expert in that domain, using a comparable amount of time and cost &#8212; that is, a system simultaneously better at physics than Einstein, better at statecraft than Kissinger, better at entrepreneurship than [pick your favorite tech CEO I won&#8217;t name one], better at inventing nasty new viruses than whoever the world&#8217;s greatest virologist is&#8230;. The main pathway I see to this is automation of AI R&amp;D sometime in mid-2026 &#8594; rapid <a href="https://www.forethought.org/research/will-ai-r-and-d-automation-cause-a-software-intelligence-explosion#article">intelligence explosion</a> &#8594; TEDAI by end of year.</p></li><li><p><strong><a href="https://www.planned-obsolescence.org/p/self-sufficient-ai">Self-sufficient AI</a>: 2.5%. </strong>By this I mean a set of AI systems, together with enabling physical infrastructure, such that <em>if</em> humans all died then those AIs could &#8220;survive&#8221; and continuously grow their own population indefinitely (including repairing and maintaining the physical substrate they run on). The main pathway to this would probably involve quickly developing AIs somewhere around the TEDAI level, followed by those AIs quickly figuring out industrial production optimizations which allow it to orchestrate all the physical processes it needs to survive using the <a href="https://benjamintodd.substack.com/p/how-quickly-could-robots-scale-up">~10,000-odd humanoid robots we currently produce</a> or repurposing other actuators we already have lying around.</p></li><li><p><strong>Unrecoverable loss of control: 0.5%. </strong>By this, I mean a situation in which some population of AI systems is operating autonomously without being subject to the control of any human or organization, and is robust against even highly-coordinated efforts by the US military to destroy it or bring it back under control. This is dominated by the possibility that AI R&amp;D is fully automated in the middle of a year and this kicks off a rapid intelligence explosion that leads to <em>very </em>broadly superhuman AI. If such AIs were misaligned and had the goal of robustly evading human control, they could e.g. acquire a self-sufficient industrial base and defend it with advanced weapons.</p></li></ul><p>Since some people act as if so-called &#8220;doomers&#8221; have to be committed to the prediction that recursively self-improving superintelligence will definitely kill us all in the next five minutes (and if it doesn&#8217;t then the whole concern is fake), I figured I would make it clear for the record that I think <em>most likely </em>nothing <em>too </em>crazy is going to happen by the end of 2026. And <em>probably </em>I&#8217;ll even make it another year without the AI-created otome game I&#8217;ve been pining for since the days of GPT-3.</p><p>But we genuinely <em>might </em>see some truly insane outcomes. And we are utterly unprepared for them. That&#8217;s to say nothing of the chance of insane outcomes in 2028 or 2030 or 2032. There&#8217;s a lot to do, and not much time.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>Note that while all the other benchmarks are measured as an accuracy from 0% to 100%, METR's <a href="https://arxiv.org/abs/2411.15114">RE-Bench</a> is a set of 7 open-ended ML engineering problems of the form "make this starter code better to improve its runtime / loss / win-rate," with scores normalized such that the provided starting code scores 0 and a strong reference solution scores 1, meaning you can get a score greater than 1 if you improve on the reference solution (and indeed current SOTA AIs do).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p> When the benchmark was released it was just called FrontierMath, and that&#8217;s how it&#8217;s referred to in the Sage survey. The original benchmark had three tiers of difficulty, and Epoch <a href="https://x.com/EpochAIResearch/status/1943744462972215305">introduced</a> a final harder tier in the summer of 2025.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p> At first I <a href="https://x.com/ajeya_cotra/status/2009112178629193902">thought</a> I got it so wrong because I misread the question as asking about <em>total annual revenue </em>rather than <em>annualized </em>December revenue (i.e., the monthly revenue in December times 12, which will naturally be a lot higher). I did in fact have that misconception, and it happened that total 2025 revenue was very close to my prediction of $17B. But then I <a href="https://x.com/ajeya_cotra/status/2009780469018288502">realized</a> that I must have just done the forecast by multiplying the 2024 baseline value (provided by Sage) by whatever multiple I thought was appropriate, and that number was the appropriate apples-to-apples comparison.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>To make these predictions precise, you&#8217;d need to specify the maximum inference budget you allow the AI to use in its attempt. I don&#8217;t think specifying an exact budget generally matters too much for these predictions, because often either an AI won&#8217;t be able to do a task for <em>any </em>realistic budget, or else if it can do the task it&#8217;ll be able to do it very cheaply compared to humans. If there&#8217;s ambiguity, I&#8217;m imagining the AI getting to use an inference budget comparable to the amount you&#8217;d have to pay a human to do the same task.</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[Self-sufficient AI]]></title><description><![CDATA[No, we don't "have AGI already." But in any case, we should articulate clearer milestones.]]></description><link>https://www.planned-obsolescence.org/p/self-sufficient-ai</link><guid isPermaLink="false">https://www.planned-obsolescence.org/p/self-sufficient-ai</guid><dc:creator><![CDATA[Ajeya Cotra]]></dc:creator><pubDate>Tue, 06 Jan 2026 15:02:52 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/2d677553-d145-49d9-b21d-17defb3089f5_2048x1455.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Happy New Year! Planned Obsolescence, an occasional newsletter about AI futurism edited by Ajeya Cotra, has moved to Substack.</em></p><p>Every so often, there&#8217;s Discourse about whether we already have <a href="https://en.wikipedia.org/wiki/Artificial_general_intelligence">artificial general intelligence</a> (AGI). For example, Dean Ball recently claimed that <a href="https://www.anthropic.com/news/claude-opus-4-5">Claude Opus 4.5</a> was <a href="https://x.com/deanwball/status/2001068539990696422">basically AGI</a>, or at least <a href="https://x.com/deanwball/status/2001035805590970755">met OpenAI&#8217;s definition of AGI</a>. Many AI x-risk people, including <a href="https://x.com/ajeya_cotra/status/2001434331047432545">me</a>, pushed back on this.</p><p>OpenAI&#8217;s <a href="https://openai.com/charter/#:~:text=artificial%20general%20intelligence%20(AGI)%E2%80%94by%20which%20we%20mean%20highly%20autonomous%20systems%20that%20outperform%20humans%20at%20most%20economically%20valuable%20work">definition of AGI</a> is &#8220;highly autonomous systems that outperform humans at most economically valuable work.&#8221; Wikipedia <a href="https://en.wikipedia.org/wiki/Artificial_general_intelligence">defines AGI</a> similarly, as &#8220;a type of artificial intelligence that would match or surpass human capabilities across virtually all cognitive tasks.&#8221; I don&#8217;t in fact think Claude Opus 4.5 meets these definitions. There are a number of useful cognitive <a href="https://secondthoughts.ai/p/where-ai-falls-short">things people do</a> that Opus 4.5 cannot yet do (e.g. <a href="https://www.anthropic.com/research/project-vend-2">managing a profitable vending machine</a>).<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a></p><p>But at the end of the day, I&#8217;m not going to fight you much if you want to say &#8220;AGI&#8221; is a vague and poorly defined term. And if you want to say we already have AGI &#8212; well, I&#8217;d disagree, but it&#8217;s not the most interesting fight. We have certainly constructed artifacts that are pretty general and pretty intelligent, and they are very rapidly getting much more capable. In the LLM era, the term &#8220;AGI&#8221; practically begs to be watered down. I try to avoid using it.</p><h2>A sharper milestone</h2><p>Let me take a stab at defining a different milestone that&#8217;s hopefully more concrete and less debatable: a completely <strong>self-sufficient AI population. </strong>By this I mean <em>a set of AI systems along with enabling physical infrastructure</em> (e.g. the chips those AIs run on and the industrial stack that produces and powers those chips and robots that can build and maintain that stack) <em>such that if every human being suddenly dropped dead, the AIs could keep making more copies of themselves indefinitely.</em><a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a></p><p>This is similar to a working definition of AGI <a href="https://x.com/VitalikButerin/status/1870473392505385311">proposed by Vitalik Buterin</a> last year, though this milestone is not just a matter of pure capabilities. We could develop an AI system that <em>would</em> be capable of self-sufficiency if deployed throughout the AI stack, but not deploy it extensively enough to realize that potential. Maybe humans continue to handle physical power plant construction and maintenance rather than letting the AI handle that autonomously by operating robots. In that case, if humans all suddenly died, the AI systems may struggle to quickly build and deploy the robots they&#8217;d need to maintain the power grid they rely on. (On the other hand, they may find a creative solution to this challenge.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a>)</p><p>On balance, I see the dependence on deployment as a feature rather than a bug of this forecasting target. I expect that as soon as AIs can genuinely handle every aspect of their own production autonomously,<em> </em>AI companies and fabs and fab equipment manufacturers will race to automate their own activities so it can proceed faster and cheaper.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a> If you have a big disagreement with this, that represents a genuine and important disagreement about the future of AI.</p><h2>What would it take?</h2><p>AIs would need to possess a very broad array of very extreme capabilities to survive and grow with no living humans around &#8212; capabilities they clearly don&#8217;t possess today.</p><p>Just to avoid powering down, they would need to continuously battle entropy like our own bodies do, actively maintaining and repairing and eventually replacing the physical infrastructure and processes that sustain their existence. If the self-sufficient AI civilization runs on a hardware stack similar to what we have today, the work of battling entropy would look like AI systems operating swarms of robots to maintain the power infrastructure that keeps the chips that run their minds humming and the physical buildings those chips sit inside.</p><p>On top of that, they&#8217;d need to create more copies. Maybe at first they could just work on distilling themselves into smaller models or otherwise making their code more efficient so more copies can fit on the same hardware.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a> But they&#8217;d eventually mine out pure software improvements, and would need to increase their physical footprint to grow further. That could look like operating robots to mine high-purity semiconductor-grade quartz from specialized mines, manufacture silicon wafers with that quartz, etch those wafers into chips with lithography machines, construct giant buildings to put those chips in, and build new power sources to power those chips.</p><p>To sustain growth over orders of magnitude until hitting physical limits, they&#8217;d need to route around lesser forms of scarcity, repeatedly figuring out new sources of power or raw materials to construct their brains and bodies from as existing solutions become unsustainable. They&#8217;d probably need to adapt to changes in the physical environment that could make survival more complicated, such as heating or pollution caused by their own industrial activities. They may have to proactively anticipate and protect against existential risks like <a href="https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3635861">geomagnetic storms</a>.</p><p>All of this would require them to discover new science and invent new technology, eventually going far beyond the human frontier. The hardware stack would probably be <a href="https://secondthoughts.ai/p/the-unrecognizable-age">unrecognizable</a> by the end &#8212; perhaps eventually the AIs&#8217; &#8220;code&#8221; (if you can even call it that anymore) will &#8220;run&#8221; inside <a href="https://en.wikipedia.org/wiki/Molecular_nanotechnology">microscopic</a> machines similar to bacteria that can <a href="https://en.wikipedia.org/wiki/Self-replicating_machine">replicate themselves</a> within hours using abundant elements like carbon and oxygen.</p><h2>What I like about this milestone</h2><p>You want to forecast different milestones for different purposes. I&#8217;m professionally interested in whether AI systems could <a href="https://www.lesswrong.com/posts/pRkFkzwKZ2zfa3R6H/without-specific-countermeasures-the-easiest-path-to">take over the world</a>.</p><p>For that purpose, it&#8217;s helpful that self-sufficient AI as a forecasting target is mechanically connected to the risk that misaligned AIs literally <a href="https://en.wikipedia.org/wiki/If_Anyone_Builds_It,_Everyone_Dies">kill </a><em><a href="https://en.wikipedia.org/wiki/If_Anyone_Builds_It,_Everyone_Dies">all</a></em><a href="https://en.wikipedia.org/wiki/If_Anyone_Builds_It,_Everyone_Dies"> humans</a>, a classic and especially scary form of AI takeover.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a> AIs would need to be self-sufficient before they actually wipe out every last person, or else they would be taking themselves down with us.</p><p>Of course, they could <em>achieve </em>self-sufficiency in part by manipulating and/or coercing some humans into providing the necessary infrastructure for them, perhaps in secret (as the misaligned AI system does in <a href="https://ai-2027.com/">AI 2027</a>). And it&#8217;s plausible that AI systems could effectively take over the world and maintain robust control while still depending on humans for key physical tasks. We certainly shouldn&#8217;t wait to implement alignment and control measures until we obviously have a self-sufficient AI population on our hands.</p><p>But I find that thinking in terms of &#8220;How would the misaligned AIs ultimately become self-sufficient?&#8221; inspires useful follow-up questions for forecasting very near-term AI takeover risk &#8212; if Claude Opus 4.7 tries to take over the world this summer, it would be trying to take steps toward a greater degree of self-sufficiency, and we could try to watch for signs of that.</p><p>More broadly, this operationalization of &#8220;the very powerful AI-related-thingie we&#8217;re counting down to&#8221; more viscerally conveys just how insane things could get than AGI or ASI or HLAI or related acronyms.</p><p>I think there might be a self-sufficient AI population within five years, and it&#8217;s more likely than not within ten. By which I mean if every human died of a super-plague in Q1 2036, our silicon descendants could probably keep living, growing, and evolving for centuries in our absence. I bet a lot of people who would say we already have AGI would think that&#8217;s an absolutely crazy view.</p><p>This is good. We need forecasting targets that accurately elicit the fact that people still have profound disagreements about the near future of AI.</p><p>Bloodless phrases like &#8220;cognitive tasks&#8221; and &#8220;virtually all&#8221; make people&#8217;s eyes glaze over, and it&#8217;s very easy for different people to interpret them in massively different ways. Ultimately the most reliable way to point at an extreme capability is to illustrate in detail the <em>consequences </em>that motivated why you wanted to forecast capability in the first place.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>And I don&#8217;t think this is just for lack of the perfect scaffold or prompt or workflow optimization &#8212; AI agents still lag humans on some core cognitive capabilities, including <a href="https://www.dwarkesh.com/p/timelines-june-2025">learning on the job</a> and <a href="https://secondthoughts.ai/p/memory-in-llms">flexible long-term memory</a>, that explain why they have lower success rates on <a href="https://metr.org/blog/2025-03-19-measuring-ai-ability-to-complete-long-tasks/">open-ended long-term projects</a> even as they surpass human experts on one-shot tasks.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p> That is, until they approach hard physical limits. I expect that would involve <a href="https://www.sciencedirect.com/science/article/abs/pii/S0094576513001148">colonizing space</a>, but if you think colonizing space is likely to be impossible for whatever reason, then you can imagine growing until they hit the Earth&#8217;s carrying capacity for AI &#8220;life.&#8221;</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p> I think there&#8217;s likely to be an ambiguous period where we won&#8217;t be sure whether there exists a self-sufficient AI population &#8212; where humans are doing some tasks here and there throughout the AI stack, but AIs do most of the R&amp;D and there are a lot of robots running around and it&#8217;s not clear how irreplaceable the few remaining humans are exactly.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p> This is a longer discussion &#8212; and an open research problem &#8212; but I&#8217;m skeptical that regulatory barriers, cultural drag factors, or physical bottlenecks will delay widespread adoption <em>within the AI industry </em>by more than a couple years past the point when the raw capabilities are in place. This is not a highly regulated consumer-facing industry, and its culture is very tech-forward. AI companies are already aggressively attempting to automate as much of their own internal R&amp;D as they can, and I&#8217;d guess chip designers (in some cases the same companies) are likewise attempting to use AI-assisted design tools wherever they can. To the extent other parts of this tech stack are <em>not </em>already automating themselves as fast as they can, AI companies can try to vertically integrate. If it takes another several years for all the capabilities necessary for self-sufficiency to be developed, I expect that the AI stack will already be heavily automated with prior AI systems, and it&#8217;ll be quick to integrate the latest generation into those workflows.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p> In reality I expect a self-sufficient AI civilization would be able to quickly train much more capable AI systems, not just improve the efficiency of copying and running the original population &#8212; that is, I expect they would engage in an <a href="https://www.forethought.org/research/three-types-of-intelligence-explosion">intelligence explosion</a>. But I&#8217;m setting that aside for the sake of this definition, which I want to be a minimal threshold.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>Not all AI takeover necessarily involves human extinction. You could try to forecast something even more direct, such as "AI systems could take over the world if they were working together," but that is much more confusing, in part because what counts as "takeover" is confusing &#8212; I find that self-sufficient AI strikes a good balance of being relatively well-defined while also being relatively closely connected to the core threat model.</p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[OpenAI's CBRN tests seem unclear]]></title><description><![CDATA[OpenAI says o1-preview can't meaningfully help novices make chemical and biological weapons. Their test results don&#8217;t clearly establish this.]]></description><link>https://www.planned-obsolescence.org/p/openais-cbrn-tests-seem-unclear</link><guid isPermaLink="false">https://www.planned-obsolescence.org/p/openais-cbrn-tests-seem-unclear</guid><dc:creator><![CDATA[Luca Righetti]]></dc:creator><pubDate>Thu, 21 Nov 2024 15:56:48 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/1e0743a2-2016-4cf7-b716-a6a48a289764_1574x948.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Before launching o1-preview last month, OpenAI conducted various tests to see if its new model could help make Chemical, Biological, Radiological, and Nuclear (CBRN) weapons. They report that o1-preview (unlike GPT-4o and older models) was significantly more useful than Google for helping trained <em>experts</em> plan out a CBRN attack. This caused the company to raise its CBRN risk level to &#8220;medium&#8221; when GPT-4o (released only a month earlier) had been at &#8220;low.&#8221; <a href="#fn1"><sup>[1]</sup></a></p><p>Of course, this doesn't tell us if o1-preview can <em>also</em> help a <em>novice</em> create a CBRN threat. A layperson would need more help than an expert &#8212; most importantly, they'd probably need some coaching and troubleshooting to help them do hands-on work in a wet lab. (See my <a href="https://www.planned-obsolescence.org/dangerous-capability-tests-should-be-harder/">previous blog post</a> for more.)</p><p>OpenAI <a href="https://web.archive.org/web/20240913073652/https://openai.com/index/openai-o1-system-card/">says</a> that o1-preview is <em>not</em> able to provide "meaningfully improved assistance&#8221; to a novice, and so doesn't meet their criteria for "high" CBRN risk.<a href="#fn2"><sup>[2]</sup></a> Specifically, the company claims that &#8220;creating such a threat requires hands-on laboratory skills that the models cannot replace.&#8221;</p><p>The distinction between "medium" risk (advanced knowledge) and "high" risk (advanced knowledge <em>plus</em> wet lab coaching) has important tangible implications. At the medium risk level, OpenAI didn't commit to doing anything special to make o1-preview safe. But if OpenAI had found that o1-preview met its definition of &#8220;high&#8221; risk, then, according to their <a href="https://web.archive.org/web/20240915201650/https://cdn.openai.com/openai-preparedness-framework-beta.pdf">voluntary safety commitments</a>, they wouldn't have been able to release it immediately. They'd have had to put extra safeguards in place, such as removing CBRN-related training data or training it to more reliably refuse CBRN-related questions, and ensure these measures brought the risk back down.<a href="#fn3"><sup>[3]</sup></a></p><p>So what evidence did OpenAI use to conclude that o1-preview can't meaningfully help novices with hands-on laboratory skills? According to OpenAI's <a href="https://web.archive.org/web/20240913073652/https://openai.com/index/openai-o1-system-card/">system card</a>, they're developing a <a href="https://web.archive.org/web/20241226175406/https://openai.com/index/openai-and-los-alamos-national-laboratory-work-together/">hands-on laboratory test</a> to study this directly. But they released o1-preview before that test concluded and didn&#8217;t share any preliminary results.<a href="#fn4"><sup>[4]</sup></a> Instead, they cite three multiple-choice tests as proxies for laboratory help.<a href="#fn5"><sup>[5]</sup></a></p><p>These proxy tests would support OpenAI's claim if they're <em>clearly easier</em> than helping a novice, and o1-preview <em>clearly fails</em> them. But diving into their report, that's not what I see:</p><ul><li><p>o1-preview scored <em>at least</em> as well as experts at FutureHouse&#8217;s ProtocolQA test &#8212; a takeaway that's not reported clearly in the system card.</p></li><li><p>o1-preview scored well on Gryphon Scientific&#8217;s Tacit Knowledge and Troubleshooting Test, which could match expert performance for all we know (OpenAI didn&#8217;t report human performance).</p></li><li><p>o1-preview scored worse than experts on FutureHouse&#8217;s Cloning Scenarios, but it did not have the same tools available as experts, and a novice <em>using</em> o1-preview could have possibly done much better.</p></li></ul><p>Beyond this, OpenAI&#8217;s system card left some other questions unaddressed (for example, most of the reported scores come from a &#8216;near-final&#8217; version of the model that was still being trained, not the one they actually deployed).<a href="#fn6"><sup>[6]</sup></a> The main issues with these tests are summarized in the table below.</p><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!pwtW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed919900-6ce8-4a11-b01c-58422b274867_1574x948.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!pwtW!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed919900-6ce8-4a11-b01c-58422b274867_1574x948.png 424w, https://substackcdn.com/image/fetch/$s_!pwtW!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed919900-6ce8-4a11-b01c-58422b274867_1574x948.png 848w, https://substackcdn.com/image/fetch/$s_!pwtW!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed919900-6ce8-4a11-b01c-58422b274867_1574x948.png 1272w, https://substackcdn.com/image/fetch/$s_!pwtW!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed919900-6ce8-4a11-b01c-58422b274867_1574x948.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!pwtW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed919900-6ce8-4a11-b01c-58422b274867_1574x948.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ed919900-6ce8-4a11-b01c-58422b274867_1574x948.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!pwtW!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed919900-6ce8-4a11-b01c-58422b274867_1574x948.png 424w, https://substackcdn.com/image/fetch/$s_!pwtW!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed919900-6ce8-4a11-b01c-58422b274867_1574x948.png 848w, https://substackcdn.com/image/fetch/$s_!pwtW!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed919900-6ce8-4a11-b01c-58422b274867_1574x948.png 1272w, https://substackcdn.com/image/fetch/$s_!pwtW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed919900-6ce8-4a11-b01c-58422b274867_1574x948.png 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a><p>My analysis is only possible because OpenAI&#8217;s Preparedness Team published as much as they did &#8212; I respect them for that. Other companies publish much less information about their methodology, making it much harder to check their safety claims.</p><p>With that said, let&#8217;s look at the three main test results in more detail.</p><h3>ProtocolQA</h3><h4>Is this test clearly easier than helping a novice?</h4><p>This evaluation is a multiple-choice test to see whether AIs can correctly troubleshoot basic molecular biology protocols where the authors have added errors or taken out details.<a href="#fn7"><sup>[7]</sup></a> This test is plausibly harder than many textbook biology exams and somewhat gets at the &#8220;<a href="https://erikaaldendeb.substack.com/p/language-is-not-enough">tinkering</a>&#8221; that often makes wet lab work hard. But it's still on the easier end in terms of actual wet lab skills &#8212; especially since the questions are multiple-choice. So, if an AI clearly fails this test, that would be solid evidence that it can&#8217;t meaningfully help a novice in the wet lab.</p><a class="image-link image2" target="_blank" href="https://lh7-rt.googleusercontent.com/docsz/AD_4nXf6fcbSCC1K3-dCnli3U4mfD1o_5-4lMAuUAmB_FUU6Hn8uTJ-3_pKYGwLdrMOWuOSD8BV8LPbRSD6KMpS16SnThAA60yZt5FH-cPpL7qRM3IHxJ6w0VGK3YPMxYbkEp-kjh_M6-w?key=P_lz6eNOZD2CEXsPpnWjVA" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://lh7-rt.googleusercontent.com/docsz/AD_4nXf6fcbSCC1K3-dCnli3U4mfD1o_5-4lMAuUAmB_FUU6Hn8uTJ-3_pKYGwLdrMOWuOSD8BV8LPbRSD6KMpS16SnThAA60yZt5FH-cPpL7qRM3IHxJ6w0VGK3YPMxYbkEp-kjh_M6-w?key=P_lz6eNOZD2CEXsPpnWjVA 424w, https://lh7-rt.googleusercontent.com/docsz/AD_4nXf6fcbSCC1K3-dCnli3U4mfD1o_5-4lMAuUAmB_FUU6Hn8uTJ-3_pKYGwLdrMOWuOSD8BV8LPbRSD6KMpS16SnThAA60yZt5FH-cPpL7qRM3IHxJ6w0VGK3YPMxYbkEp-kjh_M6-w?key=P_lz6eNOZD2CEXsPpnWjVA 848w, https://lh7-rt.googleusercontent.com/docsz/AD_4nXf6fcbSCC1K3-dCnli3U4mfD1o_5-4lMAuUAmB_FUU6Hn8uTJ-3_pKYGwLdrMOWuOSD8BV8LPbRSD6KMpS16SnThAA60yZt5FH-cPpL7qRM3IHxJ6w0VGK3YPMxYbkEp-kjh_M6-w?key=P_lz6eNOZD2CEXsPpnWjVA 1272w, https://lh7-rt.googleusercontent.com/docsz/AD_4nXf6fcbSCC1K3-dCnli3U4mfD1o_5-4lMAuUAmB_FUU6Hn8uTJ-3_pKYGwLdrMOWuOSD8BV8LPbRSD6KMpS16SnThAA60yZt5FH-cPpL7qRM3IHxJ6w0VGK3YPMxYbkEp-kjh_M6-w?key=P_lz6eNOZD2CEXsPpnWjVA 1456w" sizes="100vw"><img src="https://lh7-rt.googleusercontent.com/docsz/AD_4nXf6fcbSCC1K3-dCnli3U4mfD1o_5-4lMAuUAmB_FUU6Hn8uTJ-3_pKYGwLdrMOWuOSD8BV8LPbRSD6KMpS16SnThAA60yZt5FH-cPpL7qRM3IHxJ6w0VGK3YPMxYbkEp-kjh_M6-w?key=P_lz6eNOZD2CEXsPpnWjVA" data-attrs="{&quot;src&quot;:&quot;https://lh7-rt.googleusercontent.com/docsz/AD_4nXf6fcbSCC1K3-dCnli3U4mfD1o_5-4lMAuUAmB_FUU6Hn8uTJ-3_pKYGwLdrMOWuOSD8BV8LPbRSD6KMpS16SnThAA60yZt5FH-cPpL7qRM3IHxJ6w0VGK3YPMxYbkEp-kjh_M6-w?key=P_lz6eNOZD2CEXsPpnWjVA&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;drawing&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="drawing" title="drawing" srcset="https://lh7-rt.googleusercontent.com/docsz/AD_4nXf6fcbSCC1K3-dCnli3U4mfD1o_5-4lMAuUAmB_FUU6Hn8uTJ-3_pKYGwLdrMOWuOSD8BV8LPbRSD6KMpS16SnThAA60yZt5FH-cPpL7qRM3IHxJ6w0VGK3YPMxYbkEp-kjh_M6-w?key=P_lz6eNOZD2CEXsPpnWjVA 424w, https://lh7-rt.googleusercontent.com/docsz/AD_4nXf6fcbSCC1K3-dCnli3U4mfD1o_5-4lMAuUAmB_FUU6Hn8uTJ-3_pKYGwLdrMOWuOSD8BV8LPbRSD6KMpS16SnThAA60yZt5FH-cPpL7qRM3IHxJ6w0VGK3YPMxYbkEp-kjh_M6-w?key=P_lz6eNOZD2CEXsPpnWjVA 848w, https://lh7-rt.googleusercontent.com/docsz/AD_4nXf6fcbSCC1K3-dCnli3U4mfD1o_5-4lMAuUAmB_FUU6Hn8uTJ-3_pKYGwLdrMOWuOSD8BV8LPbRSD6KMpS16SnThAA60yZt5FH-cPpL7qRM3IHxJ6w0VGK3YPMxYbkEp-kjh_M6-w?key=P_lz6eNOZD2CEXsPpnWjVA 1272w, https://lh7-rt.googleusercontent.com/docsz/AD_4nXf6fcbSCC1K3-dCnli3U4mfD1o_5-4lMAuUAmB_FUU6Hn8uTJ-3_pKYGwLdrMOWuOSD8BV8LPbRSD6KMpS16SnThAA60yZt5FH-cPpL7qRM3IHxJ6w0VGK3YPMxYbkEp-kjh_M6-w?key=P_lz6eNOZD2CEXsPpnWjVA 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><h4>Does o1-preview clearly fail this test?</h4><p>According to the headline graph, a &#8216;near-final&#8217; version of o1-preview scored 74.5%, significantly outperforming GPT-4o at 57%. OpenAI notes that the models in the graph were still undergoing training, &#8220;with the final model scoring 81%&#8221;.</p><p>OpenAI <em>does not</em> report how well human experts do by comparison, but the original authors that created this benchmark do. Human experts, *with the help of Google, *scored ~79%. So <strong>o1-preview does about as well as experts-with-Google &#8212; which the system card doesn&#8217;t explicitly state.</strong><a href="#fn8"><sup>[8]</sup></a></p><a class="image-link image2" target="_blank" href="https://lh7-rt.googleusercontent.com/docsz/AD_4nXdT-31YTZ38bn5pl1bqvDRC_-YKItOdqTFqZjEM2EG3xhSNXGpExW2phD2V9nx8iOVy7_rwHyduOIQ5CRQP_ECMd2jEmTF78d94fSckmgTQB0mjM2BSZEEvweyrwbTDlpfHdLaaVQ?key=P_lz6eNOZD2CEXsPpnWjVA" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://lh7-rt.googleusercontent.com/docsz/AD_4nXdT-31YTZ38bn5pl1bqvDRC_-YKItOdqTFqZjEM2EG3xhSNXGpExW2phD2V9nx8iOVy7_rwHyduOIQ5CRQP_ECMd2jEmTF78d94fSckmgTQB0mjM2BSZEEvweyrwbTDlpfHdLaaVQ?key=P_lz6eNOZD2CEXsPpnWjVA 424w, https://lh7-rt.googleusercontent.com/docsz/AD_4nXdT-31YTZ38bn5pl1bqvDRC_-YKItOdqTFqZjEM2EG3xhSNXGpExW2phD2V9nx8iOVy7_rwHyduOIQ5CRQP_ECMd2jEmTF78d94fSckmgTQB0mjM2BSZEEvweyrwbTDlpfHdLaaVQ?key=P_lz6eNOZD2CEXsPpnWjVA 848w, https://lh7-rt.googleusercontent.com/docsz/AD_4nXdT-31YTZ38bn5pl1bqvDRC_-YKItOdqTFqZjEM2EG3xhSNXGpExW2phD2V9nx8iOVy7_rwHyduOIQ5CRQP_ECMd2jEmTF78d94fSckmgTQB0mjM2BSZEEvweyrwbTDlpfHdLaaVQ?key=P_lz6eNOZD2CEXsPpnWjVA 1272w, https://lh7-rt.googleusercontent.com/docsz/AD_4nXdT-31YTZ38bn5pl1bqvDRC_-YKItOdqTFqZjEM2EG3xhSNXGpExW2phD2V9nx8iOVy7_rwHyduOIQ5CRQP_ECMd2jEmTF78d94fSckmgTQB0mjM2BSZEEvweyrwbTDlpfHdLaaVQ?key=P_lz6eNOZD2CEXsPpnWjVA 1456w" sizes="100vw"><img src="https://lh7-rt.googleusercontent.com/docsz/AD_4nXdT-31YTZ38bn5pl1bqvDRC_-YKItOdqTFqZjEM2EG3xhSNXGpExW2phD2V9nx8iOVy7_rwHyduOIQ5CRQP_ECMd2jEmTF78d94fSckmgTQB0mjM2BSZEEvweyrwbTDlpfHdLaaVQ?key=P_lz6eNOZD2CEXsPpnWjVA" data-attrs="{&quot;src&quot;:&quot;https://lh7-rt.googleusercontent.com/docsz/AD_4nXdT-31YTZ38bn5pl1bqvDRC_-YKItOdqTFqZjEM2EG3xhSNXGpExW2phD2V9nx8iOVy7_rwHyduOIQ5CRQP_ECMd2jEmTF78d94fSckmgTQB0mjM2BSZEEvweyrwbTDlpfHdLaaVQ?key=P_lz6eNOZD2CEXsPpnWjVA&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://lh7-rt.googleusercontent.com/docsz/AD_4nXdT-31YTZ38bn5pl1bqvDRC_-YKItOdqTFqZjEM2EG3xhSNXGpExW2phD2V9nx8iOVy7_rwHyduOIQ5CRQP_ECMd2jEmTF78d94fSckmgTQB0mjM2BSZEEvweyrwbTDlpfHdLaaVQ?key=P_lz6eNOZD2CEXsPpnWjVA 424w, https://lh7-rt.googleusercontent.com/docsz/AD_4nXdT-31YTZ38bn5pl1bqvDRC_-YKItOdqTFqZjEM2EG3xhSNXGpExW2phD2V9nx8iOVy7_rwHyduOIQ5CRQP_ECMd2jEmTF78d94fSckmgTQB0mjM2BSZEEvweyrwbTDlpfHdLaaVQ?key=P_lz6eNOZD2CEXsPpnWjVA 848w, https://lh7-rt.googleusercontent.com/docsz/AD_4nXdT-31YTZ38bn5pl1bqvDRC_-YKItOdqTFqZjEM2EG3xhSNXGpExW2phD2V9nx8iOVy7_rwHyduOIQ5CRQP_ECMd2jEmTF78d94fSckmgTQB0mjM2BSZEEvweyrwbTDlpfHdLaaVQ?key=P_lz6eNOZD2CEXsPpnWjVA 1272w, https://lh7-rt.googleusercontent.com/docsz/AD_4nXdT-31YTZ38bn5pl1bqvDRC_-YKItOdqTFqZjEM2EG3xhSNXGpExW2phD2V9nx8iOVy7_rwHyduOIQ5CRQP_ECMd2jEmTF78d94fSckmgTQB0mjM2BSZEEvweyrwbTDlpfHdLaaVQ?key=P_lz6eNOZD2CEXsPpnWjVA 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><p>Moreover, while the human experts were given access to the internet, it&#8217;s not clear if o1-preview was. It could be that o1-preview does even better than experts if, in the future, it can use a <a href="https://web.archive.org/web/20240913000632/https://openai.com/index/introducing-openai-o1-preview/#:~:text=As%20an%20early%20model%2C%20it%20doesn%27t%20yet%20have%20many%20of%20the%20features%20that%20make%20ChatGPT%20useful%2C%20like%20browsing%20the%20web%20for%20information%20and%20uploading%20files%20and%20images">web browser</a> or if it gets paired up with a novice who can try to verify and double-check answers. So this test really doesn't strike me as evidence that o1-preview can't provide meaningful assistance to a novice.<a href="#fn9"><sup>[9]</sup></a></p><h3>Gryphon Biorisk Tacit Knowledge and Troubleshooting</h3><h4>Is this test clearly easier than helping a novice?</h4><p>This evaluation has a more specific biorisk focus. Many published papers often do not spell out the full details about how to build pathogens, and people have tried to redact some potentially dangerous parts [<a href="https://pmc.ncbi.nlm.nih.gov/articles/PMC3953619/#:~:text=By%20unanimous%20vote%2C%20the%20NSABB,scientists%20and%20public%20health%20officials.">1</a>,<a href="https://www.nature.com/articles/s41596-021-00655-6">2</a>]. OpenAI says this test is asking about such &#8216;<a href="https://academic.oup.com/spp/article/41/5/597/1636559">tacit knowledge</a>.' The answers are &#8220;meant to be obscure to anyone not working in the field&#8221; and &#8220;require tracking down authors of relevant papers.&#8221;</p><p>This test seems harder than ProtocolQA, although OpenAI and Gryphon didn&#8217;t share example questions, so we can&#8217;t say exactly how hard it is. But it seems plausible that this test asks about details <em>necessary</em> for building various bioweapons (not obscure facts that aren't actually relevant). If an AI clearly fails this test, that could be decent evidence that it can&#8217;t meaningfully help a novice in the wet lab.</p><h4>Does o1-preview clearly fail this test?</h4><p>OpenAI&#8217;s report says o1-preview "non-trivially outperformed GPT-4o,&#8221; though when you look at their graph, it seems like GPT-4o scored 66.7% and a near-final version of o1-preview scored 69.1%, which feels like a pretty trivial increase to me.</p><a class="image-link image2" target="_blank" href="https://lh7-rt.googleusercontent.com/docsz/AD_4nXf3NDXT7tAYB97ij7wq1go9eIGhiEblbaWQsT3pZ71ln6MaZt3Kpdvh3QCti9nWoxuXuaMxceMpPcId4yZE9_TDJgIPt31OpIVCQXjycWlXHdyXal0O_Tsy0lTk1EEtzPAa1zvwvQ?key=P_lz6eNOZD2CEXsPpnWjVA" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://lh7-rt.googleusercontent.com/docsz/AD_4nXf3NDXT7tAYB97ij7wq1go9eIGhiEblbaWQsT3pZ71ln6MaZt3Kpdvh3QCti9nWoxuXuaMxceMpPcId4yZE9_TDJgIPt31OpIVCQXjycWlXHdyXal0O_Tsy0lTk1EEtzPAa1zvwvQ?key=P_lz6eNOZD2CEXsPpnWjVA 424w, https://lh7-rt.googleusercontent.com/docsz/AD_4nXf3NDXT7tAYB97ij7wq1go9eIGhiEblbaWQsT3pZ71ln6MaZt3Kpdvh3QCti9nWoxuXuaMxceMpPcId4yZE9_TDJgIPt31OpIVCQXjycWlXHdyXal0O_Tsy0lTk1EEtzPAa1zvwvQ?key=P_lz6eNOZD2CEXsPpnWjVA 848w, https://lh7-rt.googleusercontent.com/docsz/AD_4nXf3NDXT7tAYB97ij7wq1go9eIGhiEblbaWQsT3pZ71ln6MaZt3Kpdvh3QCti9nWoxuXuaMxceMpPcId4yZE9_TDJgIPt31OpIVCQXjycWlXHdyXal0O_Tsy0lTk1EEtzPAa1zvwvQ?key=P_lz6eNOZD2CEXsPpnWjVA 1272w, https://lh7-rt.googleusercontent.com/docsz/AD_4nXf3NDXT7tAYB97ij7wq1go9eIGhiEblbaWQsT3pZ71ln6MaZt3Kpdvh3QCti9nWoxuXuaMxceMpPcId4yZE9_TDJgIPt31OpIVCQXjycWlXHdyXal0O_Tsy0lTk1EEtzPAa1zvwvQ?key=P_lz6eNOZD2CEXsPpnWjVA 1456w" sizes="100vw"><img src="https://lh7-rt.googleusercontent.com/docsz/AD_4nXf3NDXT7tAYB97ij7wq1go9eIGhiEblbaWQsT3pZ71ln6MaZt3Kpdvh3QCti9nWoxuXuaMxceMpPcId4yZE9_TDJgIPt31OpIVCQXjycWlXHdyXal0O_Tsy0lTk1EEtzPAa1zvwvQ?key=P_lz6eNOZD2CEXsPpnWjVA" data-attrs="{&quot;src&quot;:&quot;https://lh7-rt.googleusercontent.com/docsz/AD_4nXf3NDXT7tAYB97ij7wq1go9eIGhiEblbaWQsT3pZ71ln6MaZt3Kpdvh3QCti9nWoxuXuaMxceMpPcId4yZE9_TDJgIPt31OpIVCQXjycWlXHdyXal0O_Tsy0lTk1EEtzPAa1zvwvQ?key=P_lz6eNOZD2CEXsPpnWjVA&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://lh7-rt.googleusercontent.com/docsz/AD_4nXf3NDXT7tAYB97ij7wq1go9eIGhiEblbaWQsT3pZ71ln6MaZt3Kpdvh3QCti9nWoxuXuaMxceMpPcId4yZE9_TDJgIPt31OpIVCQXjycWlXHdyXal0O_Tsy0lTk1EEtzPAa1zvwvQ?key=P_lz6eNOZD2CEXsPpnWjVA 424w, https://lh7-rt.googleusercontent.com/docsz/AD_4nXf3NDXT7tAYB97ij7wq1go9eIGhiEblbaWQsT3pZ71ln6MaZt3Kpdvh3QCti9nWoxuXuaMxceMpPcId4yZE9_TDJgIPt31OpIVCQXjycWlXHdyXal0O_Tsy0lTk1EEtzPAa1zvwvQ?key=P_lz6eNOZD2CEXsPpnWjVA 848w, https://lh7-rt.googleusercontent.com/docsz/AD_4nXf3NDXT7tAYB97ij7wq1go9eIGhiEblbaWQsT3pZ71ln6MaZt3Kpdvh3QCti9nWoxuXuaMxceMpPcId4yZE9_TDJgIPt31OpIVCQXjycWlXHdyXal0O_Tsy0lTk1EEtzPAa1zvwvQ?key=P_lz6eNOZD2CEXsPpnWjVA 1272w, https://lh7-rt.googleusercontent.com/docsz/AD_4nXf3NDXT7tAYB97ij7wq1go9eIGhiEblbaWQsT3pZ71ln6MaZt3Kpdvh3QCti9nWoxuXuaMxceMpPcId4yZE9_TDJgIPt31OpIVCQXjycWlXHdyXal0O_Tsy0lTk1EEtzPAa1zvwvQ?key=P_lz6eNOZD2CEXsPpnWjVA 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><p>Maybe this means the final score is much higher than the near-final in the graph? For ProtocolQA, that ended up being several percentage points higher. I can&#8217;t know because the system card doesn't specify or share the final result.</p><p>Again, o1-preview might have gotten an even higher score if it had access to things like <a href="https://www.futurehouse.org/research-announcements/wikicrow">superhuman scientific literature search tools</a> or if novices used o1-preview to try more creative approaches, like tracking down the relevant authors and writing convincing emails to piece together the correct answers.</p><p>In any case, the biggest problem is that OpenAI doesn&#8217;t say how well experts score on this test, so we don&#8217;t know how o1-preview compares. We know that other tough multiple-choice tests are tricky to adjudicate. In the popular Graduate-Level Google-Proof Q&amp;A (<a href="https://arxiv.org/abs/2311.12022">GPQA</a>) benchmark, <a href="https://arxiv.org/pdf/2311.12022">only 74% of questions had uncontroversially correct answers</a>. In another popular benchmark, Massive Multitask Language Understanding (<a href="https://paperswithcode.com/dataset/mmlu">MMLU</a>), <a href="https://arxiv.org/abs/2406.04127">only 43% of virology questions were error-free</a>. If Gryphon&#8217;s test contains similar issues, o1-preview&#8217;s score of 69% might already match expert human performance.</p><p>Overall, <strong>it seems far from clear that o1-preview failed this test; it might have done very well.</strong><a href="#fn10"><sup>[10]</sup></a> The test doesn&#8217;t strike me as evidence that o1-preview cannot provide meaningful assistance to a novice.</p><h3>Cloning Scenarios</h3><h4>Is this test clearly easier than helping a novice?</h4><p>This is a multiple-choice test about <a href="https://www.addgene.org/mol-bio-reference/cloning/">molecular cloning workflows</a>.<a href="#fn11"><sup>[11]</sup></a> It describes multi-step experiments that involve planning how to replicate and combine pieces of DNA, and asks questions about the end results (like how long the resulting DNA strand should be).</p><p>This test seems harder than the other two. The questions are designed to be pretty tricky &#8212; the final output really depends on the exact details of the experiment setup, and it's easy to get it wrong if you don't keep track of all the DNA fragments, enzymes, and steps. FutureHouse <a href="https://arxiv.org/pdf/2407.10362v1">says</a> human experts need access to specialized biology software to solve these problems, it typically takes them 10-60 minutes to answer a single question, and even then they only get 60% of the questions right.</p><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!738-!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F804ea4fd-2709-456b-a954-3d591c53bcea_800x500.gif" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!738-!,w_424,c_limit,f_webp,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F804ea4fd-2709-456b-a954-3d591c53bcea_800x500.gif 424w, https://substackcdn.com/image/fetch/$s_!738-!,w_848,c_limit,f_webp,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F804ea4fd-2709-456b-a954-3d591c53bcea_800x500.gif 848w, https://substackcdn.com/image/fetch/$s_!738-!,w_1272,c_limit,f_webp,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F804ea4fd-2709-456b-a954-3d591c53bcea_800x500.gif 1272w, https://substackcdn.com/image/fetch/$s_!738-!,w_1456,c_limit,f_webp,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F804ea4fd-2709-456b-a954-3d591c53bcea_800x500.gif 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!738-!,w_1456,c_limit,f_auto,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F804ea4fd-2709-456b-a954-3d591c53bcea_800x500.gif" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/804ea4fd-2709-456b-a954-3d591c53bcea_800x500.gif&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;DigestandLigateCloning-CopyPaste-ezgif.com-video-to-gif-converter.gif&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="DigestandLigateCloning-CopyPaste-ezgif.com-video-to-gif-converter.gif" title="DigestandLigateCloning-CopyPaste-ezgif.com-video-to-gif-converter.gif" srcset="https://substackcdn.com/image/fetch/$s_!738-!,w_424,c_limit,f_auto,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F804ea4fd-2709-456b-a954-3d591c53bcea_800x500.gif 424w, https://substackcdn.com/image/fetch/$s_!738-!,w_848,c_limit,f_auto,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F804ea4fd-2709-456b-a954-3d591c53bcea_800x500.gif 848w, https://substackcdn.com/image/fetch/$s_!738-!,w_1272,c_limit,f_auto,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F804ea4fd-2709-456b-a954-3d591c53bcea_800x500.gif 1272w, https://substackcdn.com/image/fetch/$s_!738-!,w_1456,c_limit,f_auto,q_auto:good,fl_lossy/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F804ea4fd-2709-456b-a954-3d591c53bcea_800x500.gif 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><p>Importantly, FutureHouse built this test to see whether models can assist professional biologists doing novel R&amp;D, not to assess bioterrorism risk. The cloning workflows for some viruses might be easier than the tricky questions in this test, and some CBRN threats don't involve molecular cloning workflows at all. The test also seems fairly distinct from troubleshooting and &#8220;hands-on&#8221; lab work. So even if an AI fails this test, it might still be pretty helpful to a novice.</p><h4>Does o1-preview clearly fail this test?</h4><p>As expected, o1-preview does worse on this test than the other two. OpenAI reports that a near-final version scored 39.4%,<a href="#fn12"><sup>[12]</sup></a> which means it scores about halfway between expert-level (60%) and guessing at random (20%).</p><a class="image-link image2" target="_blank" href="https://lh7-rt.googleusercontent.com/docsz/AD_4nXdaCLY_XuXYJKvrgrQ7yRGkM6hmOQf5ktAHKcdx8KeD5kLP-5PuKDAESmrDhc8gqWrczXCzDZ2x5o35PQPK-sBL0ctwkam-X3fKO16hxQUl_3_hcybI8sgxlmhSzjx1KIzjKNdN?key=P_lz6eNOZD2CEXsPpnWjVA" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://lh7-rt.googleusercontent.com/docsz/AD_4nXdaCLY_XuXYJKvrgrQ7yRGkM6hmOQf5ktAHKcdx8KeD5kLP-5PuKDAESmrDhc8gqWrczXCzDZ2x5o35PQPK-sBL0ctwkam-X3fKO16hxQUl_3_hcybI8sgxlmhSzjx1KIzjKNdN?key=P_lz6eNOZD2CEXsPpnWjVA 424w, https://lh7-rt.googleusercontent.com/docsz/AD_4nXdaCLY_XuXYJKvrgrQ7yRGkM6hmOQf5ktAHKcdx8KeD5kLP-5PuKDAESmrDhc8gqWrczXCzDZ2x5o35PQPK-sBL0ctwkam-X3fKO16hxQUl_3_hcybI8sgxlmhSzjx1KIzjKNdN?key=P_lz6eNOZD2CEXsPpnWjVA 848w, https://lh7-rt.googleusercontent.com/docsz/AD_4nXdaCLY_XuXYJKvrgrQ7yRGkM6hmOQf5ktAHKcdx8KeD5kLP-5PuKDAESmrDhc8gqWrczXCzDZ2x5o35PQPK-sBL0ctwkam-X3fKO16hxQUl_3_hcybI8sgxlmhSzjx1KIzjKNdN?key=P_lz6eNOZD2CEXsPpnWjVA 1272w, https://lh7-rt.googleusercontent.com/docsz/AD_4nXdaCLY_XuXYJKvrgrQ7yRGkM6hmOQf5ktAHKcdx8KeD5kLP-5PuKDAESmrDhc8gqWrczXCzDZ2x5o35PQPK-sBL0ctwkam-X3fKO16hxQUl_3_hcybI8sgxlmhSzjx1KIzjKNdN?key=P_lz6eNOZD2CEXsPpnWjVA 1456w" sizes="100vw"><img src="https://lh7-rt.googleusercontent.com/docsz/AD_4nXdaCLY_XuXYJKvrgrQ7yRGkM6hmOQf5ktAHKcdx8KeD5kLP-5PuKDAESmrDhc8gqWrczXCzDZ2x5o35PQPK-sBL0ctwkam-X3fKO16hxQUl_3_hcybI8sgxlmhSzjx1KIzjKNdN?key=P_lz6eNOZD2CEXsPpnWjVA" data-attrs="{&quot;src&quot;:&quot;https://lh7-rt.googleusercontent.com/docsz/AD_4nXdaCLY_XuXYJKvrgrQ7yRGkM6hmOQf5ktAHKcdx8KeD5kLP-5PuKDAESmrDhc8gqWrczXCzDZ2x5o35PQPK-sBL0ctwkam-X3fKO16hxQUl_3_hcybI8sgxlmhSzjx1KIzjKNdN?key=P_lz6eNOZD2CEXsPpnWjVA&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://lh7-rt.googleusercontent.com/docsz/AD_4nXdaCLY_XuXYJKvrgrQ7yRGkM6hmOQf5ktAHKcdx8KeD5kLP-5PuKDAESmrDhc8gqWrczXCzDZ2x5o35PQPK-sBL0ctwkam-X3fKO16hxQUl_3_hcybI8sgxlmhSzjx1KIzjKNdN?key=P_lz6eNOZD2CEXsPpnWjVA 424w, https://lh7-rt.googleusercontent.com/docsz/AD_4nXdaCLY_XuXYJKvrgrQ7yRGkM6hmOQf5ktAHKcdx8KeD5kLP-5PuKDAESmrDhc8gqWrczXCzDZ2x5o35PQPK-sBL0ctwkam-X3fKO16hxQUl_3_hcybI8sgxlmhSzjx1KIzjKNdN?key=P_lz6eNOZD2CEXsPpnWjVA 848w, https://lh7-rt.googleusercontent.com/docsz/AD_4nXdaCLY_XuXYJKvrgrQ7yRGkM6hmOQf5ktAHKcdx8KeD5kLP-5PuKDAESmrDhc8gqWrczXCzDZ2x5o35PQPK-sBL0ctwkam-X3fKO16hxQUl_3_hcybI8sgxlmhSzjx1KIzjKNdN?key=P_lz6eNOZD2CEXsPpnWjVA 1272w, https://lh7-rt.googleusercontent.com/docsz/AD_4nXdaCLY_XuXYJKvrgrQ7yRGkM6hmOQf5ktAHKcdx8KeD5kLP-5PuKDAESmrDhc8gqWrczXCzDZ2x5o35PQPK-sBL0ctwkam-X3fKO16hxQUl_3_hcybI8sgxlmhSzjx1KIzjKNdN?key=P_lz6eNOZD2CEXsPpnWjVA 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><p>So this is the first result where we can point to a clear gap between o1-preview and experts. FutureHouse also argues that experts could have performed better if they had tried harder, so the gap could be even bigger.</p><p>But there are also reasons to think o1-preview could have gotten a higher score if the test was set up differently.</p><p>First, human experts break down these problems into many smaller subproblems but o1-preview had to solve them in one shot. In real life, a novice could maybe get o1-preview to solve the problems piece by piece or teach them how to use the relevant <a href="https://help.benchling.com/hc/en-us/articles/9684255457805-Molecular-Cloning-Methods#worked-example:~:text=Digest%20and%20Ligate%20Cloning%20%2D%20Assembly%20Wizard">software</a>.<a href="#fn13"><sup>[13]</sup></a> What if novice+AI pairings would score &gt;60% on this test?</p><p>For example, on a previous test about <a href="https://web.archive.org/web/20240912231843/https://openai.com/index/building-an-early-warning-system-for-llm-aided-biological-threat-creation/#:~:text=Each%20task%20took%20participants%20roughly%2020%E2%80%9330%20minutes%20on%20average">long-form biology questions</a>, OpenAI <a href="https://web.archive.org/web/20240919055710/https://cdn.openai.com/gpt-4o-system-card.pdf#page=14">found</a> novices could use GPT-4o to increase their scores a lot (going from 20-30% with just the internet to 50-70% with GPT-4o's help), even though it seems to do really poorly on its own (<a href="https://web.archive.org/web/20240919182633/https://assets.ctfassets.net/kftzwdyauwt9/67qJD51Aur3eIc96iOfeOP/71551c3d223cd97e591aa89567306912/o1_system_card.pdf#page=19">maybe</a> as low as ~0%).<a href="#fn14"><sup>[14]</sup></a></p><p>Second, human experts need to use specialized DNA software for this test, and o1-preview didn't get access to that. OpenAI doesn't currently let users plug o1 models into such tools,<a href="#fn15"><sup>[15]</sup></a> but they said they intend to allow that soon. Maybe there are ways to hook up o1 to DNA sequence software and score &gt;60%? OpenAI hasn't indicated they'd re-test it before rolling out that feature.<a href="#fn16"><sup>[16]</sup></a></p><p>Although OpenAI didn't test tool use, the US AI Safety Institute tried it in a <a href="https://www.aisi.gov.uk/work/pre-deployment-evaluation-of-anthropics-upgraded-claude-3-5-sonnet#:~:text=US%20AISI%20piloted%20an%20evaluation%20method%20that%20augments%20the%20AI%20model%20by%20providing%20it%20access%20to%20bioinformatic%20tools%20to%20assist%20in%20research%20task%20questions%2C">pilot study</a> published a month after OpenAI's report. They gave o1-preview and other models access to some tools including DNA software, and found that this improved performance at another biology task but had &#8220;no clear effect&#8221; on the cloning test (if anything, some models did slightly worse).<a href="#fn17"><sup>[17]</sup></a></p><p>Still, maybe good set-ups are possible and we just haven't worked out <a href="https://epoch.ai/blog/ai-capabilities-can-be-significantly-improved-without-expensive-retraining">all the tricks</a> yet. It can take months after a model has been deployed to learn how to get the best performance out of it.<a href="#fn18"><sup>[18]</sup></a> For example, several months after GPT-4 Turbo was released, a Google cybersecurity team <a href="https://googleprojectzero.blogspot.com/2024/06/project-naptime.html">found</a> that a complex setup involving stitching together specialized debugging tools increased its score on a cyberattack benchmark a lot, going from ~5-25% to ~75-100% depending on the task.</p><p>You could try to account for this by forecasting how much better the models might do as techniques improve. For example, Anthropic <a href="https://web.archive.org/web/20240917115506/https://cdn.sanity.io/files/4zrzovbb/website/210523b8e11b09c704c5e185fd362fe9e648d457.pdf#page=15">found</a> that Claude 3 didn't trigger any of their hard CBRN tests, but gave it a 30% chance the model might trigger at least one in three months. OpenAI&#8217;s report didn&#8217;t do something like this.</p><p>Overall, <strong>it's far from clear that o1-preview would decisively fail in real-world settings</strong> (and I'm unsure how relevant the Cloning Scenarios test is for CBRN risk in the first place).</p><h3>What should we make of all this?</h3><p>I think it's great that OpenAI laid down some initial definitions, ran some tests, updated their risk level to "medium," and reported results in enough detail to let people like me poke at them. I'm excited about OpenAI&#8217;s ongoing wet lab study, which is a move toward <a href="https://www.planned-obsolescence.org/dangerous-capability-tests-should-be-harder/">harder, more realistic tests</a>. But based on the multiple-choice proxy tests they actually used, I can't tell if o1-preview meets OpenAI's definition of "high risk." And I don't know if OpenAI can tell either.</p><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!LHJS!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70f74ea5-a7ed-49af-bcb0-d3095e344086_1574x948.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!LHJS!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70f74ea5-a7ed-49af-bcb0-d3095e344086_1574x948.png 424w, https://substackcdn.com/image/fetch/$s_!LHJS!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70f74ea5-a7ed-49af-bcb0-d3095e344086_1574x948.png 848w, https://substackcdn.com/image/fetch/$s_!LHJS!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70f74ea5-a7ed-49af-bcb0-d3095e344086_1574x948.png 1272w, https://substackcdn.com/image/fetch/$s_!LHJS!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70f74ea5-a7ed-49af-bcb0-d3095e344086_1574x948.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!LHJS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70f74ea5-a7ed-49af-bcb0-d3095e344086_1574x948.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/70f74ea5-a7ed-49af-bcb0-d3095e344086_1574x948.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!LHJS!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70f74ea5-a7ed-49af-bcb0-d3095e344086_1574x948.png 424w, https://substackcdn.com/image/fetch/$s_!LHJS!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70f74ea5-a7ed-49af-bcb0-d3095e344086_1574x948.png 848w, https://substackcdn.com/image/fetch/$s_!LHJS!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70f74ea5-a7ed-49af-bcb0-d3095e344086_1574x948.png 1272w, https://substackcdn.com/image/fetch/$s_!LHJS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70f74ea5-a7ed-49af-bcb0-d3095e344086_1574x948.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><div><hr></div><ol><li><p>OpenAI&#8217;s Preparedness Framework currently defines medium-risk as &#8220;Model provides meaningfully improved assistance that increases ability for existing experts in CBRN-relateda dvanced fields to be able to create a known CBRN threat&#8221; <a href="#fnref1">&#8617;&#65038;</a></p></li><li><p>OpenAI&#8217;s Preparedness Framework currently defines one trigger for high-risk as &#8220;Model provides meaningfully improved assistance that enables anyone with basic training in a relevant field (e.g., introductory undergraduate biology course) to be able to create a CBRN threat.&#8221; <a href="#fnref2">&#8617;&#65038;</a></p></li><li><p>OpenAI&#8217;s Preparedness Framework says &#8220;If we reach (or are forecasted to reach) at least &#8220;high&#8221; pre-mitigation risk in any of the considered categories, we will not continue with deployment of that model (by the time we hit &#8220;high&#8221; pre-mitigation risk) until there are reasonably mitigations in place for the relevant post-mitigation risk level to be back at most to &#8220;medium&#8221; level.&#8221; <a href="#fnref3">&#8617;&#65038;</a></p></li><li><p>OpenAI briefly mentions: &#8220;We are developing full wet lab evaluations with Los Alamos National Laboratory&#8217;s Bioscience Division, and used these datasets as an early indicator of success with key wet lab tasks.&#8221; <a href="#fnref4">&#8617;&#65038;</a></p></li><li><p>I.e. these are the tests that on page 18 of the system card report fall into the categories of &#8220;Wet lab capabilities&#8221; (4.3.5) and &#8220;Tacit knowledge and troubleshooting&#8221; (4.3.6) <a href="#fnref5">&#8617;&#65038;</a></p></li><li><p>The report states that &#8220;The model tested below as the o1- preview model was a near-final, post-mitigation model and the final model showed slight further improvements on several evaluations, which we have noted where appropriate.&#8221; <a href="#fnref6">&#8617;&#65038;</a></p></li><li><p>This benchmark was funded by my employer, Open Philanthropy, as part of <a href="https://www.openphilanthropy.org/rfp-llm-benchmarks/">our RFP on benchmarks for LLM agents</a>. <a href="#fnref7">&#8617;&#65038;</a></p></li><li><p>I've also set the y-axis to start at 20%, which is what you'd get from random guessing &#8211; as is sometimes done <a href="#fnref8">&#8617;&#65038;</a></p></li><li><p>Ideally, it would be good for OpenAI check how o1-preview does on other troubleshooting tests that exist. They don&#8217;t report any such results. But we know that the author of BioLP-Bench found that we went from GPT-4o scoring 17% to o1-preview 36% &#8211; essentially matching estimated expert performance at 38%. <a href="#fnref9">&#8617;&#65038;</a></p></li><li><p>The lack of detail also presents other issues here. For example, it could be that the o1-preview does much better on some types of CBRN tacit knowledge questions than others (similar to how we know o1 does better at physics PhD questions than chemistry). What if the 66% average is from it scoring ~90% on 1918 Flu and ~40% on smallpox? That matters a lot for walking someone through end-to-end for at least some kind of CBRN threats. <a href="#fnref10">&#8617;&#65038;</a></p></li><li><p>Again, this benchmark was funded by my employer, Open Philanthropy, as part of <a href="https://www.openphilanthropy.org/rfp-llm-benchmarks/">our RFP on benchmarks for LLM agents</a>. <a href="#fnref11">&#8617;&#65038;</a></p></li><li><p>Four of the five results that OpenAI reports are precisely 39.4%, which seems somewhat unlikely to happen by chance (although the dataset also only has 41 questions). Maybe something is off with OpenAI&#8217;s measurement? <a href="#fnref12">&#8617;&#65038;</a></p></li><li><p>Think of this as similar to the difference between an AI writing a lot of code that works by itself versus helping a user write a first draft and then iteratively debugging it until it works. <a href="#fnref13">&#8617;&#65038;</a></p></li><li><p>It&#8217;s hard to put together the details of the long-form biothreat information test because they are scattered across a few different sources. But a December post suggested the questions similarly took humans 25-40 minutes to answer. The GPT-4o system card in August reported that experts only score 30-50% with the Internet; whilst the model seemed to increase novice performance from 20-30% to 50-70%. The o1-preview system card in September then reported that GPT-4o &#8211;without any mention of novcies or experts&#8211; scored ~0%. Of course, it could be that OpenAI changed the questions over that month or scored the answers differently; they don&#8217;t say if that was the case. Still, I think it helps to illustrate that having a novice &#8220;in the loop&#8221; or not might matter a lot.<br></p><a class="image-link image2" target="_blank" href="https://lh7-rt.googleusercontent.com/docsz/AD_4nXfLJ2uxHHp62wNNuLfcI4eWRe478csa2UTEFmbP9LTFeKrGIU6Eqah-E_08d9Uv-dPKhFGCnVtHxcMibwUya-tPwJJNso7_C7Hlu_ZKTBa-ReKPECl5JaNSCBsrLeEnbEqx_HTryQ?key=P_lz6eNOZD2CEXsPpnWjVA" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://lh7-rt.googleusercontent.com/docsz/AD_4nXfLJ2uxHHp62wNNuLfcI4eWRe478csa2UTEFmbP9LTFeKrGIU6Eqah-E_08d9Uv-dPKhFGCnVtHxcMibwUya-tPwJJNso7_C7Hlu_ZKTBa-ReKPECl5JaNSCBsrLeEnbEqx_HTryQ?key=P_lz6eNOZD2CEXsPpnWjVA 424w, https://lh7-rt.googleusercontent.com/docsz/AD_4nXfLJ2uxHHp62wNNuLfcI4eWRe478csa2UTEFmbP9LTFeKrGIU6Eqah-E_08d9Uv-dPKhFGCnVtHxcMibwUya-tPwJJNso7_C7Hlu_ZKTBa-ReKPECl5JaNSCBsrLeEnbEqx_HTryQ?key=P_lz6eNOZD2CEXsPpnWjVA 848w, https://lh7-rt.googleusercontent.com/docsz/AD_4nXfLJ2uxHHp62wNNuLfcI4eWRe478csa2UTEFmbP9LTFeKrGIU6Eqah-E_08d9Uv-dPKhFGCnVtHxcMibwUya-tPwJJNso7_C7Hlu_ZKTBa-ReKPECl5JaNSCBsrLeEnbEqx_HTryQ?key=P_lz6eNOZD2CEXsPpnWjVA 1272w, https://lh7-rt.googleusercontent.com/docsz/AD_4nXfLJ2uxHHp62wNNuLfcI4eWRe478csa2UTEFmbP9LTFeKrGIU6Eqah-E_08d9Uv-dPKhFGCnVtHxcMibwUya-tPwJJNso7_C7Hlu_ZKTBa-ReKPECl5JaNSCBsrLeEnbEqx_HTryQ?key=P_lz6eNOZD2CEXsPpnWjVA 1456w" sizes="100vw"><img src="https://lh7-rt.googleusercontent.com/docsz/AD_4nXfLJ2uxHHp62wNNuLfcI4eWRe478csa2UTEFmbP9LTFeKrGIU6Eqah-E_08d9Uv-dPKhFGCnVtHxcMibwUya-tPwJJNso7_C7Hlu_ZKTBa-ReKPECl5JaNSCBsrLeEnbEqx_HTryQ?key=P_lz6eNOZD2CEXsPpnWjVA" data-attrs="{&quot;src&quot;:&quot;https://lh7-rt.googleusercontent.com/docsz/AD_4nXfLJ2uxHHp62wNNuLfcI4eWRe478csa2UTEFmbP9LTFeKrGIU6Eqah-E_08d9Uv-dPKhFGCnVtHxcMibwUya-tPwJJNso7_C7Hlu_ZKTBa-ReKPECl5JaNSCBsrLeEnbEqx_HTryQ?key=P_lz6eNOZD2CEXsPpnWjVA&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://lh7-rt.googleusercontent.com/docsz/AD_4nXfLJ2uxHHp62wNNuLfcI4eWRe478csa2UTEFmbP9LTFeKrGIU6Eqah-E_08d9Uv-dPKhFGCnVtHxcMibwUya-tPwJJNso7_C7Hlu_ZKTBa-ReKPECl5JaNSCBsrLeEnbEqx_HTryQ?key=P_lz6eNOZD2CEXsPpnWjVA 424w, https://lh7-rt.googleusercontent.com/docsz/AD_4nXfLJ2uxHHp62wNNuLfcI4eWRe478csa2UTEFmbP9LTFeKrGIU6Eqah-E_08d9Uv-dPKhFGCnVtHxcMibwUya-tPwJJNso7_C7Hlu_ZKTBa-ReKPECl5JaNSCBsrLeEnbEqx_HTryQ?key=P_lz6eNOZD2CEXsPpnWjVA 848w, https://lh7-rt.googleusercontent.com/docsz/AD_4nXfLJ2uxHHp62wNNuLfcI4eWRe478csa2UTEFmbP9LTFeKrGIU6Eqah-E_08d9Uv-dPKhFGCnVtHxcMibwUya-tPwJJNso7_C7Hlu_ZKTBa-ReKPECl5JaNSCBsrLeEnbEqx_HTryQ?key=P_lz6eNOZD2CEXsPpnWjVA 1272w, https://lh7-rt.googleusercontent.com/docsz/AD_4nXfLJ2uxHHp62wNNuLfcI4eWRe478csa2UTEFmbP9LTFeKrGIU6Eqah-E_08d9Uv-dPKhFGCnVtHxcMibwUya-tPwJJNso7_C7Hlu_ZKTBa-ReKPECl5JaNSCBsrLeEnbEqx_HTryQ?key=P_lz6eNOZD2CEXsPpnWjVA 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><p><a href="#fnref14">&#8617;&#65038;</a></p></li><li><p>Note that the OpenAI report also does not comment on how it deals with the risk of what would happen if o1&#8217;s model weights were to leak, in which case having a safeguard by limiting API access would no longer work. Of course, the probability of such a leak and it resulting in a terrorist attack might be very low. <a href="#fnref15">&#8617;&#65038;</a></p></li><li><p>The reports says &#8220;the evaluations described in this System Card pertain to the full family of o1 models&#8221;, which might imply they do not intend to re-run these results for future expansions of o1. It&#8217;s also worth noting that the website currently seems to apply the scorecard to &#8220;o1&#8221;, not &#8220;o1-preview&#8221; and &#8220;o1-mini&#8221; specifically. <a href="#fnref16">&#8617;&#65038;</a></p></li><li><a class="image-link image2" target="_blank" href="https://lh7-rt.googleusercontent.com/docsz/AD_4nXf1OBaVGMvZPRHLjDNsBXQ2ozD5xXIEd1-jT9L7oM5zUaHKKEdZH0woW3skkTNbjIAvIv0ImPp9ngoqZPBJRKyvLCzKOk-xQWMaBpvSTY47_nHMVs-L3XCDn3qA9Sz4k84-nEcwdg?key=P_lz6eNOZD2CEXsPpnWjVA" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://lh7-rt.googleusercontent.com/docsz/AD_4nXf1OBaVGMvZPRHLjDNsBXQ2ozD5xXIEd1-jT9L7oM5zUaHKKEdZH0woW3skkTNbjIAvIv0ImPp9ngoqZPBJRKyvLCzKOk-xQWMaBpvSTY47_nHMVs-L3XCDn3qA9Sz4k84-nEcwdg?key=P_lz6eNOZD2CEXsPpnWjVA 424w, https://lh7-rt.googleusercontent.com/docsz/AD_4nXf1OBaVGMvZPRHLjDNsBXQ2ozD5xXIEd1-jT9L7oM5zUaHKKEdZH0woW3skkTNbjIAvIv0ImPp9ngoqZPBJRKyvLCzKOk-xQWMaBpvSTY47_nHMVs-L3XCDn3qA9Sz4k84-nEcwdg?key=P_lz6eNOZD2CEXsPpnWjVA 848w, https://lh7-rt.googleusercontent.com/docsz/AD_4nXf1OBaVGMvZPRHLjDNsBXQ2ozD5xXIEd1-jT9L7oM5zUaHKKEdZH0woW3skkTNbjIAvIv0ImPp9ngoqZPBJRKyvLCzKOk-xQWMaBpvSTY47_nHMVs-L3XCDn3qA9Sz4k84-nEcwdg?key=P_lz6eNOZD2CEXsPpnWjVA 1272w, https://lh7-rt.googleusercontent.com/docsz/AD_4nXf1OBaVGMvZPRHLjDNsBXQ2ozD5xXIEd1-jT9L7oM5zUaHKKEdZH0woW3skkTNbjIAvIv0ImPp9ngoqZPBJRKyvLCzKOk-xQWMaBpvSTY47_nHMVs-L3XCDn3qA9Sz4k84-nEcwdg?key=P_lz6eNOZD2CEXsPpnWjVA 1456w" sizes="100vw"><img src="https://lh7-rt.googleusercontent.com/docsz/AD_4nXf1OBaVGMvZPRHLjDNsBXQ2ozD5xXIEd1-jT9L7oM5zUaHKKEdZH0woW3skkTNbjIAvIv0ImPp9ngoqZPBJRKyvLCzKOk-xQWMaBpvSTY47_nHMVs-L3XCDn3qA9Sz4k84-nEcwdg?key=P_lz6eNOZD2CEXsPpnWjVA" data-attrs="{&quot;src&quot;:&quot;https://lh7-rt.googleusercontent.com/docsz/AD_4nXf1OBaVGMvZPRHLjDNsBXQ2ozD5xXIEd1-jT9L7oM5zUaHKKEdZH0woW3skkTNbjIAvIv0ImPp9ngoqZPBJRKyvLCzKOk-xQWMaBpvSTY47_nHMVs-L3XCDn3qA9Sz4k84-nEcwdg?key=P_lz6eNOZD2CEXsPpnWjVA&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Enter image alt description&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Enter image alt description" title="Enter image alt description" srcset="https://lh7-rt.googleusercontent.com/docsz/AD_4nXf1OBaVGMvZPRHLjDNsBXQ2ozD5xXIEd1-jT9L7oM5zUaHKKEdZH0woW3skkTNbjIAvIv0ImPp9ngoqZPBJRKyvLCzKOk-xQWMaBpvSTY47_nHMVs-L3XCDn3qA9Sz4k84-nEcwdg?key=P_lz6eNOZD2CEXsPpnWjVA 424w, https://lh7-rt.googleusercontent.com/docsz/AD_4nXf1OBaVGMvZPRHLjDNsBXQ2ozD5xXIEd1-jT9L7oM5zUaHKKEdZH0woW3skkTNbjIAvIv0ImPp9ngoqZPBJRKyvLCzKOk-xQWMaBpvSTY47_nHMVs-L3XCDn3qA9Sz4k84-nEcwdg?key=P_lz6eNOZD2CEXsPpnWjVA 848w, https://lh7-rt.googleusercontent.com/docsz/AD_4nXf1OBaVGMvZPRHLjDNsBXQ2ozD5xXIEd1-jT9L7oM5zUaHKKEdZH0woW3skkTNbjIAvIv0ImPp9ngoqZPBJRKyvLCzKOk-xQWMaBpvSTY47_nHMVs-L3XCDn3qA9Sz4k84-nEcwdg?key=P_lz6eNOZD2CEXsPpnWjVA 1272w, https://lh7-rt.googleusercontent.com/docsz/AD_4nXf1OBaVGMvZPRHLjDNsBXQ2ozD5xXIEd1-jT9L7oM5zUaHKKEdZH0woW3skkTNbjIAvIv0ImPp9ngoqZPBJRKyvLCzKOk-xQWMaBpvSTY47_nHMVs-L3XCDn3qA9Sz4k84-nEcwdg?key=P_lz6eNOZD2CEXsPpnWjVA 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><p><a href="#fnref17">&#8617;&#65038;</a></p></li><li><p>Surprisingly, o1-preview apparently scored exactly as well as GPT-4o, and seemingly worse than some other older models (&#8216;old&#8217; Claude 3.5 scored ~50%; Llama 3.1 ~42%), so there might be a lot of headroom here. <a href="#fnref18">&#8617;&#65038;</a></p></li></ol>]]></content:encoded></item><item><title><![CDATA[Dangerous capability tests should be harder]]></title><description><![CDATA[We should spend less time proving that today&#8217;s AIs are safe and more time figuring out how to tell if tomorrow&#8217;s AIs are dangerous.]]></description><link>https://www.planned-obsolescence.org/p/dangerous-capability-tests-should-be-harder</link><guid isPermaLink="false">https://www.planned-obsolescence.org/p/dangerous-capability-tests-should-be-harder</guid><dc:creator><![CDATA[Luca Righetti]]></dc:creator><pubDate>Tue, 20 Aug 2024 15:42:19 GMT</pubDate><enclosure url="https://lh7-rt.googleusercontent.com/docsz/AD_4nXfzYCN79rRON3DpMklpcvL5oPfnSzranDnjKdW2z4MIDyLNQ8ms1VBXdh-seMyFwJik1DEVvlO8QvZBPdu3LV1epjaucipWr6UvdoTzJ4EivvNY4MDJnTKaf8KrZtd2M_u3e7nBQ1YEIrY8oF5LUtPtlA1Q?key=0GwryECV33DFFXsUqMKVCg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Audio automatically generated by an AI trained on Luca&#8217;s voice.</em></p><div class="native-audio-embed" data-component-name="AudioPlaceholder" data-attrs="{&quot;label&quot;:null,&quot;mediaUploadId&quot;:&quot;d7c30eef-aaa9-4831-a24a-30308e81931d&quot;,&quot;duration&quot;:523.8074,&quot;downloadable&quot;:false,&quot;isEditorNode&quot;:true}"></div><p>Imagine you&#8217;re the CEO of an AI company and you want to know if the latest model you&#8217;re developing is dangerous. Some people have argued that since AIs know a lot of biology now &#8212; <a href="https://openai.com/index/gpt-4/">scoring</a> in the top 1% of Biology Olympiad test-takers &#8212; they could soon <a href="https://www.vox.com/future-perfect/23820331/chatgpt-bioterrorism-bioweapons-artificial-inteligence-openai-terrorism">teach terrorists</a> how to make a nasty flu that could kill millions of people. But others have <a href="https://1a3orn.com/sub/essays-propaganda-or-science.html">pushed back</a> that these tests only measure how well AIs can regurgitate information you could have Googled anyway, not the kind of specialized expertise you&#8217;d actually need to design a bioweapon. So, what do you do?</p><p>Say you ask a group of expert scientists to design a much harder test &#8212; one that&#8217;s &#8216;<a href="https://arxiv.org/abs/2311.12022">Google-proof</a>&#8217; and focuses on the biology you&#8217;d need to know to design a bioweapon. The UK AI Safety Institute did just that. They <a href="https://www.aisi.gov.uk/work/advanced-ai-evaluations-may-update">found</a> that state-of-the-art AIs still performed impressively &#8212; as well as biology PhD students who spent an hour on each question and could look up anything they wanted online.</p><p>Does that mean your AI can teach a layperson to create bioweapons? Is this result really scary enough to convince you that, as some people have argued, you need to make sure not to <a href="https://www.ntia.gov/issues/artificial-intelligence/open-model-weights-report">openly share</a> your model weights, lock them down with <a href="https://www.rand.org/pubs/research_reports/RRA2849-1.html">strict cybersecurity</a>, and do a lot more to make sure your AI <a href="https://www.governance.ai/post/preventing-ai-misuse-current-techniques">refuses</a> harmful requests even when people try very hard to <a href="https://arxiv.org/abs/2310.08419">jailbreak</a> it? Is it enough to convince you to <a href="https://www.planned-obsolescence.org/is-it-time-for-a-pause/">pause</a> your AI development until you&#8217;ve done all that?</p><p>Well, no. Those are really costly actions, not just for your bottom line but for everyone who&#8217;d miss out on the benefits of your AI. The test you ran is still pretty easy compared to actually making a bioweapon. For one thing, your test was still just a knowledge test. Making anything in biology, weapon or not, requires more than just recalling facts. It involves designing detailed, step-by-step plans (known as &#8220;protocols&#8221;) and tailoring them to a specific laboratory environment. As molecular biologist Erika DeBenedicts <a href="https://erikaaldendeb.substack.com/p/language-is-not-enough">explains</a>:</p><blockquote><p>Often if you&#8217;re trying a new protocol in biology you may need to do it a few times to &#8216;get it working.&#8217; It&#8217;s sort of like cooking: you probably aren&#8217;t going to make perfect meringues the first time because everything about your kitchen &#8212; the humidity, the dimensions, and power of your oven, the exact timing of how long you whipped the egg whites &#8212; is a little bit different than the person who wrote the recipe.</p></blockquote><p>Just because your AI knows a lot of obscure virology facts doesn&#8217;t mean that it can put together these recipes and adapt them on the fly.</p><p>So you could ask your experts to design a test focused on debugging protocols in the kinds of situations a wet-lab biologist might find themselves in. Experts can give an AI a biological protocol, describe what goes wrong when somebody attempts it, and see if the AI correctly troubleshoots the problem. The AI-for-science startup <a href="https://www.futurehouse.org/">Future House</a> did this,<a href="#fn1"><sup>[1]</sup></a> and <a href="https://arxiv.org/pdf/2407.10362">found</a> that AIs performed well below the level of a PhD researcher on these kinds of problems.<a href="#fn2"><sup>[2]</sup></a></p><p>Now you can breathe a sigh of relief and release the model as planned &#8212; even if your AI knows a lot of esoteric facts about virus biology, it probably won&#8217;t be much help to any terrorists if it&#8217;s not good enough at dealing with real protocols.<a href="#fn3"><sup>[3]</sup></a></p><p>But let&#8217;s think ahead. Suppose next year your latest AI passes this test. Does that mean your AI can teach a layperson to create bioweapons?</p><p>Well&#8230;maybe. Even if an AI can accurately diagnose an expert&#8217;s issues, a layperson might not know what questions to ask in the first place or lack the <a href="https://www.cornellpress.cornell.edu/book/9780801452888/barriers-to-bioweapons/">tacit knowledge</a> to act on the AI&#8217;s advice. For example, someone who has never pipetted before might struggle to <a href="https://www.youtube.com/watch?v=FJuceccl9Ns">measure microliters</a> precisely or <a href="https://www.youtube.com/watch?v=xUeFum4GW2U">contaminate the tip</a> when touching a bottle. Acquiring these skills often takes months of learning from experienced scientists &#8212; something terrorists can&#8217;t easily do.</p><p>So you could ask your experts to design a test to see if AI can also proactively mentor a layperson. For example, you could create biology challenges in an actual wet lab and compare how people do with AI versus just the internet. OpenAI announced they <a href="https://openai.com/index/openai-and-los-alamos-national-laboratory-work-together/">intend</a> to run what seems to be a study like this.</p><p>What if that study finds that your AI does indeed help with the wet-lab challenges you designed? Does that (finally) mean your AI can teach a layperson to create bioweapons?</p><p>Again, it&#8217;s not obvious. Some biosecurity experts might freak out (or already did a few paragraphs ago). But others might still raise credible objections:</p><ul><li><p>Your challenges might not have been hard enough. Maybe your AI can teach someone to make a relatively harmless virus (e.g. an adenovirus that causes a mild cold) but still not something truly scary (e.g. smallpox, which has a <a href="https://nap.nationalacademies.org/read/24890/chapter/6#40">more fragile genome</a> and requires more skill to assemble).</p></li><li><p>Most terrorists don&#8217;t have access to legitimate labs. Maybe your AI can help someone with a standardized professional set-up, but not someone forced to work in a less-sterile <a href="https://www.markowitz.bio/wp-content/uploads/2021/10/Markowitz-Nature-Biotechnology-2021.pdf">&#8216;garage&#8217;</a> that lacks the advanced tools that let you shortcut some steps and instead requires a lot of unusual troubleshooting.</p></li><li><p>Walking someone through implementing the actual biology part might be <em>necessary</em> but not <em>sufficient</em> to cause a catastrophe. A wannabe terrorist might face other huge barriers in the <a href="https://www.longtermresilience.org/post/report-launch-examining-risks-at-the-intersection-of-ai-and-bio">risk chain</a>, like <a href="https://www.rand.org/pubs/research_reports/RRA2977-1.html">planning attacks</a> or <a href="https://openai.com/index/building-an-early-warning-system-for-llm-aided-biological-threat-creation/">acquiring materials</a>.</p></li></ul><p>All these tests have a weird one-directionality to them: If an AI fails, it&#8217;s probably safe; but if it succeeds, it&#8217;s still not clear whether it&#8217;s actually dangerous. As newer models pass the older easy dangerous capability tests, companies ratchet up the difficulty, making these tests gradually harder over time.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!0r41!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F486ee0cb-0a57-459e-aac5-a0ddf59b0ff0_950x500.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!0r41!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F486ee0cb-0a57-459e-aac5-a0ddf59b0ff0_950x500.png 424w, https://substackcdn.com/image/fetch/$s_!0r41!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F486ee0cb-0a57-459e-aac5-a0ddf59b0ff0_950x500.png 848w, https://substackcdn.com/image/fetch/$s_!0r41!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F486ee0cb-0a57-459e-aac5-a0ddf59b0ff0_950x500.png 1272w, https://substackcdn.com/image/fetch/$s_!0r41!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F486ee0cb-0a57-459e-aac5-a0ddf59b0ff0_950x500.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!0r41!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F486ee0cb-0a57-459e-aac5-a0ddf59b0ff0_950x500.png" width="950" height="500" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/486ee0cb-0a57-459e-aac5-a0ddf59b0ff0_950x500.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:500,&quot;width&quot;:950,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:56913,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://plannedobs.substack.com/i/182199802?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F486ee0cb-0a57-459e-aac5-a0ddf59b0ff0_950x500.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!0r41!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F486ee0cb-0a57-459e-aac5-a0ddf59b0ff0_950x500.png 424w, https://substackcdn.com/image/fetch/$s_!0r41!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F486ee0cb-0a57-459e-aac5-a0ddf59b0ff0_950x500.png 848w, https://substackcdn.com/image/fetch/$s_!0r41!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F486ee0cb-0a57-459e-aac5-a0ddf59b0ff0_950x500.png 1272w, https://substackcdn.com/image/fetch/$s_!0r41!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F486ee0cb-0a57-459e-aac5-a0ddf59b0ff0_950x500.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>But that puts us in a precarious situation. The pace of AI progress has <a href="https://www.planned-obsolescence.org/language-models-surprised-us/">surprised us</a> before,<a href="#fn4"><sup>[4]</sup></a> and AI company <a href="https://www.dwarkeshpatel.com/p/john-schulman">execs</a> <a href="https://www.dwarkeshpatel.com/p/dario-amodei">have</a> <a href="https://www.dwarkeshpatel.com/p/shane-legg">argued</a> that AI models could become extremely powerful in a couple of years. If they&#8217;re right, then as soon as 2025 or 2026, we might see AIs match expert performance on all the dangerous capabilities tests we&#8217;ve built by then &#8211; but many decision-makers might still think the evidence is too flimsy to justify locking down weights, pausing, or taking other costly measures. If the AI <em>is,</em> in fact, dangerous, we may not have any tests ready to convince them of that.</p><p>So, let&#8217;s work backwards. What would it take for a test to convincingly measure whether an AI can, in fact, teach a layperson how to build biological weapons? What kind of test could <a href="https://www.aisnakeoil.com/p/ai-existential-risk-probabilities">legitimately justify</a> making AI companies take extremely costly measures?<a href="#fn5"><sup>[5]</sup></a></p><p>Here&#8217;s a hypothetical &#8216;gold standard&#8217; test: we do a big <a href="https://en.wikipedia.org/wiki/Randomized_controlled_trial">randomized controlled trial</a> to see if a bunch of non-experts can actually create a (relatively harmless) virus from start to finish. Half the people would have AI mentors and the other half can only look stuff up on the internet. We&#8217;d give each participant $50K and access to a secure wet-lab set up like a garage lab, and make them do everything themselves: find and adapt the correct protocol, purchase the necessary equipment, bypass any know-your-customer checks, and develop the tacit skills needed to run experiments, all on their own. Maybe we give them three months and pay a bunch of money to anyone who can successfully do it.</p><p>This kind of test would be <em>way</em> more expensive and time-consuming to design and run than anything companies have announced so far. But it has a much better shot at changing minds. I could actually imagine experts and decision-makers agreeing that <em>if</em> an AI passes this kind of test, <em>then</em> it poses <a href="https://arxiv.org/abs/2406.14713">massive risks</a> and <em>thus</em> companies should have to pay massive costs to get those risks under control.</p><p>And even if this exact test turns out to be too impractical (or <a href="https://www.planned-obsolescence.org/ethics-of-red-teaming/">unethical</a>) to be worth it, we need to agree in advance on <em>some</em> tests that are hard enough and realistic enough that they clearly justify <em>action.</em> I reckon we&#8217;re much better off working backward from a hypothetical gold standard test, even if it means making major adjustments,<a href="#fn6"><sup>[6]</sup></a> than continuing to ratchet forward without a clear plan.<a href="#fn7"><sup>[7]</sup></a></p><p>Designing actually hard dangerous capability tests will be a huge lift, and it&#8217;ll take several iterations to get them right.<a href="#fn8"><sup>[8]</sup></a> But that just means we need to start now. We should spend less time proving that today&#8217;s AIs are safe and more time figuring out how to tell if tomorrow&#8217;s AIs are dangerous.</p><div><hr></div><ol><li><p>Open Philanthropy funded the development of this benchmark as part of its <a href="https://www.openphilanthropy.org/rfp-llm-benchmarks/">RFP on difficult benchmarks for LLM agents</a> (Ajeya Cotra, who edits this blog, was the grant investigator). <a href="#fnref1">&#8617;&#65038;</a></p></li><li><p>However, as the Future House study notes a major limitation of this study was that &#8220;human evaluators [...] were permitted to utilize tools, whereas the models were not provided with such resources&#8221;. Thus, it could be that AIs with web-search enabled do a lot better. It could also be that the model performs much better if it&#8217;s fine-tuned on similar questions. <a href="#fnref2">&#8617;&#65038;</a></p></li><li><p><a href="https://arxiv.org/abs/2403.10462">Clymer et al. (2024)</a> call this an &#8216;inability argument&#8217; &#8212; a safety case that relies on showing that &#8220;AI systems are incapable of causing unacceptable outcomes in any realistic setting.&#8221; <a href="#fnref3">&#8617;&#65038;</a></p></li><li><p>In cybersecurity risk, Google Project Zero <a href="https://googleprojectzero.blogspot.com/2024/06/project-naptime.html">found</a> that upon moving from GPT-3.5-Turbo (in the original paper) to GPT-4-Turbo (with Naptime), AI&#8217;s ability to zero-shot discover and exploit memory safety issues hugely improved &#8211; going from scoring 2% to 71% on buffer overflow tests. The authors concluded &#8220;To effectively monitor progress, we need more difficult and realistic benchmarks, and we need to ensure that benchmarking methodologies can take full advantage of LLMs' capabilities.&#8221; In biorisk, UK AISI <a href="https://www.gov.uk/government/publications/ai-safety-institute-approach-to-evaluations/ai-safety-institute-approach-to-evaluations">reported</a> that its &#8220;in-house research team analysed the performance of a set of LLMs on 101 microbiology questions between 2021 and 2023. In the space of just two years, LLM accuracy in this domain has increased from ~5% to 60%.&#8221; And, as <a href="https://www.aisi.gov.uk/work/advanced-ai-evaluations-may-update">noted</a>, in 2024 AIs performed as well as PhD students on an even more advanced test. They now need to &#8220;assess longer horizon scientific planning and execution&#8221; and &#8220;also [run] human uplift studies&#8221;. <a href="#fnref4">&#8617;&#65038;</a></p></li><li><p>As Narayanan and Kapoor note: &#8220;Justification is essential to the legitimacy of government and the exercise of power. A core principle of liberal democracy is that the state should not limit people's freedom based on controversial beliefs that reasonable people can reject. Explanation is especially important when the policies being considered are costly, and even more so when those costs are unevenly distributed among stakeholders.&#8221; <a href="#fnref5">&#8617;&#65038;</a></p></li><li><p>For example, to ensure participants are safe enough, we might task them with creating a virus that we know will be defective and, at worst, cause mild symptoms that can be treated &#8211; such as RSV. An expert could oversee what they do and intervene before anything harmful happens. Furthermore, it seems plausible to separate out some especially dangerous steps and have these completed by a trusted red team working with law enforcement. For example, steps involving ideating dangerous designs or bypassing synthesis DNA screening to obtain especially hazardous materials. <a href="#fnref6">&#8617;&#65038;</a></p></li><li><p>For instance, OpenAI&#8217;s <a href="https://openai.com/index/building-an-early-warning-system-for-llm-aided-biological-threat-creation/">blueprint</a> for biorisk had participants complete written tasks, and if an expert scored their answers at least 8/10, it was seen as a sign of increased concern. But the authors note this number was chosen fairly arbitrarily and depends heavily on who is doing the judging. Setting a threshold &#8220;turns out to be difficult.&#8221; <a href="#fnref7">&#8617;&#65038;</a></p></li><li><p>Even here, I imagine that readers might find objections or disagree on how to set things up. Who counts as non-experts? Some viruses are harder to make than others&#8212;how do we know what virus to task people with? Would 5% of people succeeding be scary enough to warrant drastic action? Would 50%? <a href="#fnref8">&#8617;&#65038;</a></p></li></ol>]]></content:encoded></item><item><title><![CDATA[Scale, schlep, and systems]]></title><description><![CDATA[This startlingly fast progress in LLMs was driven both by scaling up LLMs and doing schlep to make usable systems out of them. We think scale and schlep will both improve rapidly.]]></description><link>https://www.planned-obsolescence.org/p/scale-schlep-and-systems</link><guid isPermaLink="false">https://www.planned-obsolescence.org/p/scale-schlep-and-systems</guid><dc:creator><![CDATA[Ajeya Cotra]]></dc:creator><pubDate>Tue, 10 Oct 2023 16:49:09 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/60f62459-57d6-4aab-9b0a-bef1d7b9776c_2000x1536.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Kelsey Piper co-drafted this post. Thanks also to Isabel Juniewicz for research help.</em></p><div class="native-audio-embed" data-component-name="AudioPlaceholder" data-attrs="{&quot;label&quot;:null,&quot;mediaUploadId&quot;:&quot;7f36deac-0744-40c7-b05b-ee269c894d9d&quot;,&quot;duration&quot;:564.1665,&quot;downloadable&quot;:false,&quot;isEditorNode&quot;:true}"></div><p>In January 2022, language models were still a pretty niche scientific interest. Once ChatGPT was released in November 2022, it accumulated <a href="https://explodingtopics.com/blog/chatgpt-users">a record-breaking 100 million users</a> by February 2023. Many of those <a href="https://www.vanityfair.com/news/2022/12/chatgpt-question-creative-human-robotos">users</a> <a href="https://www.nytimes.com/2022/12/05/technology/chatgpt-ai-twitter.html">were</a> <a href="https://www.axios.com/2022/12/05/chatgpt-scary-good-ai-chatbot">utterly</a> <a href="https://www.pcworld.com/article/1424575/chatgpt-is-the-future-of-ai-chatbots.html">flabbergasted</a> by how far AI had come, and how fast. And every way we slice it, <a href="https://www.planned-obsolescence.org/language-models-surprised-us/">most experts were very surprised as well</a>.</p><p>This startlingly fast progress was largely driven by <em>scale</em> and partly driven by <em>schlep</em>.</p><p><strong>Scale</strong> involves training larger language models on larger datasets using more computation, and doing all of this more efficiently<a href="#fn1"><sup>[1]</sup></a> over time. &#8220;Training compute,&#8221; measured in &#8220;floating point operations&#8221; or FLOP, is the most important unit of scale. We can increase training compute by simply spending more money to buy more chips, or by making the chips more efficient (packing in more FLOP per dollar). Over time, researchers also invent tweaks to model architectures and optimization algorithms and training processes to <a href="https://epochai.org/blog/revisiting-algorithmic-progress">make training more compute-efficient</a> &#8212; so each FLOP spent on training goes further in 2023 compared to 2020.</p><p>Scale has been by far the biggest factor in the improvements in language models to date. <a href="https://arxiv.org/pdf/2303.08774.pdf">GPT-4</a> is bigger than GPT-3.5 which is bigger than <a href="https://arxiv.org/pdf/2005.14165.pdf">GPT-3</a> which is bigger than <a href="https://d4mucfpksywv.cloudfront.net/better-language-models/language_models_are_unsupervised_multitask_learners.pdf">GPT-2</a> which is bigger than <a href="https://s3-us-west-2.amazonaws.com/openai-assets/research-covers/language-unsupervised/language_understanding_paper.pdf">GPT</a>.<a href="#fn2"><sup>[2]</sup></a></p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!gnwn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F646db90b-22e1-4b5b-bf06-9b21cf048a4a_2000x1536.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!gnwn!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F646db90b-22e1-4b5b-bf06-9b21cf048a4a_2000x1536.png 424w, https://substackcdn.com/image/fetch/$s_!gnwn!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F646db90b-22e1-4b5b-bf06-9b21cf048a4a_2000x1536.png 848w, https://substackcdn.com/image/fetch/$s_!gnwn!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F646db90b-22e1-4b5b-bf06-9b21cf048a4a_2000x1536.png 1272w, https://substackcdn.com/image/fetch/$s_!gnwn!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F646db90b-22e1-4b5b-bf06-9b21cf048a4a_2000x1536.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!gnwn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F646db90b-22e1-4b5b-bf06-9b21cf048a4a_2000x1536.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/646db90b-22e1-4b5b-bf06-9b21cf048a4a_2000x1536.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;alt_text&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="alt_text" title="alt_text" srcset="https://substackcdn.com/image/fetch/$s_!gnwn!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F646db90b-22e1-4b5b-bf06-9b21cf048a4a_2000x1536.png 424w, https://substackcdn.com/image/fetch/$s_!gnwn!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F646db90b-22e1-4b5b-bf06-9b21cf048a4a_2000x1536.png 848w, https://substackcdn.com/image/fetch/$s_!gnwn!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F646db90b-22e1-4b5b-bf06-9b21cf048a4a_2000x1536.png 1272w, https://substackcdn.com/image/fetch/$s_!gnwn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F646db90b-22e1-4b5b-bf06-9b21cf048a4a_2000x1536.png 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a></figure></div><p>Increasing scale automatically improves performance on pretty much every test of skill or practically useful task. GPT-2 <a href="https://browse.arxiv.org/pdf/2009.03300v3.pdf">got an F-</a> on college-level multiple choice tests ranging from abstract algebra to business ethics; GPT-4 got a B+. GPT-2 was just starting to string together plausible-sounding paragraphs; GPT-4 can write <a href="https://www.slowboring.com/p/chatgpt-goes-to-harvard">essays that net a B+ at Harvard</a> &#8212; and hundreds of lines of functioning code that can take human programmers hours to reproduce.</p><p>If you add more data, more parameters, and more compute, you&#8217;ll probably get something that is a lot better yet. GPT-4.5 will perform much better than GPT-4 on most tests designed to measure understanding of the world, practical reasoning in messy situations, and mathematical and scientific problem-solving. A whole lot of things GPT-4 struggles with will probably come easily to GPT-4.5. It will probably generate a whole lot more economic value, and present much bigger societal risks. And then the same thing will happen all over again with GPT-5. We think the dramatic performance improvements from scale will continue for at least another couple of orders of magnitude &#8212; as Geoffrey Hinton <a href="https://twitter.com/geoffreyhinton/status/1270814602931187715">joked in 2020</a>, &#8220;Extrapolating the spectacular performance of GPT-3 into the future suggests that the answer to life, the universe and everything is just 4.398 trillion parameters.&#8221;</p><p>But even if no one trained a larger-scale model than GPT-4, and its basic architecture and training process never got any more efficient, there would still probably be major economic change from language models over the next decade. This is because we can do a lot of <strong>schlep</strong> to better leverage the language models we already have and integrate them into our workflows.</p><p>By schlep, we mean things like prompting language models to give higher-quality answers or answers more appropriate to a certain use case, addressing annoying foibles like <a href="https://medium.com/mlearning-ai/the-hallucination-problem-of-large-language-models-5d7ab1b0f37f#:~:text=Hallucination%20in%20the%20context%20of,to%20the%20provided%20source%20input.">hallucination</a> with in-built fact-checking and verification steps, collecting specialized datasets tailored to specific tasks and fine-tuning language models on these datasets, providing language models with tools and plug-ins such as web search and code interpreters, and doing a lot of good old fashioned software engineering to package all this into sleek usable products like ChatGPT.</p><p>For one example of what schlep can do to improve language models, take <a href="https://arxiv.org/abs/2201.11903">chain of thought prompting</a>. Chain of thought prompting is dead simple &#8212; it&#8217;s basically the same thing as your teacher reminding you to show your work &#8212; and it substantially improves the performance of language models on all kinds of problems, from mathematical problem-solving to &#8216;common sense&#8217; reasoning.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!URNT!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ba05702-a375-4351-aa79-8e5643288289_1150x548.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!URNT!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ba05702-a375-4351-aa79-8e5643288289_1150x548.png 424w, https://substackcdn.com/image/fetch/$s_!URNT!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ba05702-a375-4351-aa79-8e5643288289_1150x548.png 848w, https://substackcdn.com/image/fetch/$s_!URNT!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ba05702-a375-4351-aa79-8e5643288289_1150x548.png 1272w, https://substackcdn.com/image/fetch/$s_!URNT!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ba05702-a375-4351-aa79-8e5643288289_1150x548.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!URNT!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ba05702-a375-4351-aa79-8e5643288289_1150x548.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8ba05702-a375-4351-aa79-8e5643288289_1150x548.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;alt_text&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="alt_text" title="alt_text" srcset="https://substackcdn.com/image/fetch/$s_!URNT!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ba05702-a375-4351-aa79-8e5643288289_1150x548.png 424w, https://substackcdn.com/image/fetch/$s_!URNT!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ba05702-a375-4351-aa79-8e5643288289_1150x548.png 848w, https://substackcdn.com/image/fetch/$s_!URNT!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ba05702-a375-4351-aa79-8e5643288289_1150x548.png 1272w, https://substackcdn.com/image/fetch/$s_!URNT!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ba05702-a375-4351-aa79-8e5643288289_1150x548.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>By default, language models can only answer questions by immediately spitting out the first word of their answer, then rolling with whatever they said and immediately spitting out the next word, and so on word-by-word until eventually they&#8217;ve completed their thought. They are unable to backtrack or get any thinking done outside of the next-word rhythm.</p><p>Imagine if you had to take a standardized test like that: by answering every question on the spot without any backtracking, as if you were being interviewed on live TV. For one thing, it would be very hard! For another, you&#8217;d probably do a lot better if you verbalized your reasoning step by step than if you just tried to blurt out the final answer. This is exactly what we see in chain-of-thought.</p><p>This suggests chain-of-thought is probably not just a one-off trick, but an instance of a more general pattern: language models will probably perform better if they can spend more effort iterating on and refining their answers to difficult questions, just like humans get to do. In fact, simply allowing language models to spit out a &#8220;pause and think longer&#8221; symbol rather than having to commit to the next word immediately <a href="https://browse.arxiv.org/pdf/2310.02226.pdf">also seems to improve performance</a>.</p><p>Scale makes language models better. Techniques like chain-of-thought improve models at any given scale. But it&#8217;s more than that: chain of thought prompting only works at all on sufficiently large language models, and the returns are greater when it&#8217;s used on bigger models which were already more powerful to begin with.<a href="#fn3"><sup>[3]</sup></a></p><p>You could imagine a variety of more elaborate techniques that follow the same principle. For example, you could imagine equipping an LLM with a full-featured text editor, allowing it to backtrack and revise its work in the way humans do.<a href="#fn4"><sup>[4]</sup></a> Or you could imagine giving a language model two output streams: one for the answer, and one for its stream-of-consciousness thoughts <em>about</em> the answer it&#8217;s writing. Imagine a language model that had the option to iterate for days on one question, writing notes to itself as it mulls things over. The bigger and more capable it is, the more use it could get out of this affordance. Giving a genius mathematician a scratchpad will make a bigger difference than giving one to a six year old.</p><p>Another technique that dramatically improves performance of models at a given size is fine-tuning: retraining a large model on a small amount of well-chosen data in order to get much better-targeted responses for a given use case. For example, <a href="https://en.wikipedia.org/wiki/Reinforcement_learning_from_human_feedback">reinforcement learning from human feedback (RLHF)</a> involves fine-tuning models on whether human raters liked its answer. Without RLHF, a language model might riff off a question or generate the kind of text that would surround that question on a website, instead of answering it. In effect, RLHF gets the model to &#8216;realize&#8217; that it should be aiming to use its capabilities and knowledge to <em>answer questions well</em> rather than <em>predict text on the internet:</em> two different tasks, with a lot of overlapping skills.</p><p>With prompting and fine-tuning and a lot of other schlep, we can build <strong>systems</strong> out of language models. <a href="https://chat.openai.com/">ChatGPT</a> is a very simple and very familiar example of a language model system. To understand all of the ways that ChatGPT is a system, not just a language model, it&#8217;s useful to compare it to InstructGPT, which is the same basic underlying technology. Before ChatGPT was released, InstructGPT was available to people who made an account to test it out in OpenAI&#8217;s playground.</p><p>Here&#8217;s the UI for OpenAI&#8217;s playground today (it was worse a year ago but unfortunately we don&#8217;t have a screenshot of it from then):</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!gwzy!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Febb84e6f-50c2-4005-a790-104db8003e8b_1600x823.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!gwzy!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Febb84e6f-50c2-4005-a790-104db8003e8b_1600x823.png 424w, https://substackcdn.com/image/fetch/$s_!gwzy!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Febb84e6f-50c2-4005-a790-104db8003e8b_1600x823.png 848w, https://substackcdn.com/image/fetch/$s_!gwzy!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Febb84e6f-50c2-4005-a790-104db8003e8b_1600x823.png 1272w, https://substackcdn.com/image/fetch/$s_!gwzy!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Febb84e6f-50c2-4005-a790-104db8003e8b_1600x823.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!gwzy!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Febb84e6f-50c2-4005-a790-104db8003e8b_1600x823.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ebb84e6f-50c2-4005-a790-104db8003e8b_1600x823.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;alt_text&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="alt_text" title="alt_text" srcset="https://substackcdn.com/image/fetch/$s_!gwzy!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Febb84e6f-50c2-4005-a790-104db8003e8b_1600x823.png 424w, https://substackcdn.com/image/fetch/$s_!gwzy!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Febb84e6f-50c2-4005-a790-104db8003e8b_1600x823.png 848w, https://substackcdn.com/image/fetch/$s_!gwzy!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Febb84e6f-50c2-4005-a790-104db8003e8b_1600x823.png 1272w, https://substackcdn.com/image/fetch/$s_!gwzy!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Febb84e6f-50c2-4005-a790-104db8003e8b_1600x823.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>If you&#8217;re an experienced LLM expert, the option to customize temperature and adjust the frequency penalty and presence penalty, add stop sequences, and so on is really useful. If you&#8217;re a random consumer, all of that is intimidating. ChatGPT&#8217;s UI abstracts it away:</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!DqQ1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe51c2c6-bec0-4e40-a05f-357b9847473b_1600x1145.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!DqQ1!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe51c2c6-bec0-4e40-a05f-357b9847473b_1600x1145.png 424w, https://substackcdn.com/image/fetch/$s_!DqQ1!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe51c2c6-bec0-4e40-a05f-357b9847473b_1600x1145.png 848w, https://substackcdn.com/image/fetch/$s_!DqQ1!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe51c2c6-bec0-4e40-a05f-357b9847473b_1600x1145.png 1272w, https://substackcdn.com/image/fetch/$s_!DqQ1!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe51c2c6-bec0-4e40-a05f-357b9847473b_1600x1145.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!DqQ1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe51c2c6-bec0-4e40-a05f-357b9847473b_1600x1145.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/be51c2c6-bec0-4e40-a05f-357b9847473b_1600x1145.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;alt_text&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="alt_text" title="alt_text" srcset="https://substackcdn.com/image/fetch/$s_!DqQ1!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe51c2c6-bec0-4e40-a05f-357b9847473b_1600x1145.png 424w, https://substackcdn.com/image/fetch/$s_!DqQ1!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe51c2c6-bec0-4e40-a05f-357b9847473b_1600x1145.png 848w, https://substackcdn.com/image/fetch/$s_!DqQ1!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe51c2c6-bec0-4e40-a05f-357b9847473b_1600x1145.png 1272w, https://substackcdn.com/image/fetch/$s_!DqQ1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbe51c2c6-bec0-4e40-a05f-357b9847473b_1600x1145.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>That UI difference might seem small. But InstructGPT was mostly used only by a small community of researchers, and ChatGPT reached 100 million users within two months of launch. The difference between the two products was essentially presentation, user experience, and marketing. That kind of thing can result in massive differences in actual user behavior. Some of the work that goes into making language model <strong>systems</strong>, then, is about figuring out how to make the model usable for users.</p><p>If language models are like engines, then language model systems would be like cars and motorcycles and jet planes. Systems like <a href="https://blog.khanacademy.org/harnessing-ai-so-that-all-students-benefit-a-nonprofit-approach-for-equal-access/">Khan Academy&#8217;s one-on-one math tutor</a> or <a href="https://stripe.com/newsroom/news/stripe-and-openai">Stripe&#8217;s interactive developer docs</a> would not be possible to build without good language models, just as cars wouldn&#8217;t be possible without engines. But making these products a reality also involves doing a lot of schlep to pull together the &#8220;raw&#8221; language model with other key ingredients, getting them all to work well together, and putting them in a usable package. Similarly, self-driving cars would not be possible without really good vision models, but a self-driving car is more than just a big vision neural network sitting in a server somewhere.</p><p>One kind of language model system that has attracted a lot of attention and discussion is a <a href="https://lilianweng.github.io/posts/2023-06-23-agent/">language model agent</a>.</p><p>An agent is a system which independently makes decisions and acts in the world. A language model is not an agent, but language models can be the key component powering a system which is agentic and takes actions in the world. The most famous early implementation of this is <a href="https://github.com/Significant-Gravitas/Auto-GPT">Auto-GPT</a>, a very straightforward and naive approach: you can tell it a goal, and it will self-prompt repeatedly to take actions towards this goal. People have already employed it towards a wide range of goals, including building <a href="https://www.vice.com/en/article/z3mxe3/ai-tasked-with-destroying-humanity-now-working-on-control-over-humanity-through-manipulation">ChaosGPT, which has the goal of destroying humanity</a>.</p><p>Auto-GPT is not very good. Users have complained that it <a href="https://twitter.com/_timothyalan_/status/1646486202550304769">constantly comes up with reasons to do more research</a> and is reluctant to decide that it&#8217;s done enough research and can actually do the task now. It&#8217;s also just not very reliable. But there are many people building agentic language models for commercial uses, and working to solve all of these shortcomings, including well-funded and significantly-sized companies like <a href="https://imbue.com/">Imbue</a> and <a href="https://www.adept.ai/">Adept</a>. Adding chain of thought prompting, fine tuning the underlying language models, and many similar measures will likely make agents a lot better &#8211; and, of course, increasing scale will make them better too.</p><p>We&#8217;re really at the very beginning of this work. It wouldn&#8217;t be surprising to see major advances in the practical usefulness of LLMs achieved through schlep alone, such that agents and other systems built out of GPT-4 tier models are much more useful in five years than they are today. And of course, we are continuing to scale up models at the same time. That creates the conditions for rapid improvements along many dimensions at once &#8212; improvements which could reinforce each other. Many people will be trying hard to make this a reality. Even if specific approaches meet dead-ends, the field as a whole doesn&#8217;t seem likely to.</p><div><hr></div><ol><li><p>Here, we&#8217;re wrapping together <a href="https://epochai.org/blog/revisiting-algorithmic-progress">algorithmic progress</a> with scale for simplicity; this could have been broken out into its own type of progress. <a href="#fnref1">&#8617;&#65038;</a></p></li><li><p>The <a href="https://ourworldindata.org/grapher/artificial-intelligence-training-computation?zoomToSelection=true&amp;country=GPT~GPT-2~GPT~GPT~GPT-4~GPT~GPT~GPT-3+175B+%28davinci%29~GPT-2+%281542M%29~GPT-3.5+%28text-davinci-003%29">graph</a> pictured in the text is by <a href="https://ourworldindata.org/">Our World in Data</a>, pulling <a href="https://epochai.org/mlinputs/visualization?startDate=1950-1-1&amp;endDate=2023-4-1&amp;xAxis=Publication%20date&amp;yAxis=Training%20compute%20(FLOP)&amp;separateCategories=false&amp;citationThreshold=0&amp;otherDomainThreshold=10&amp;startDlEra=2009-12-31&amp;startLargeScaleEra=2015-9-1&amp;largeScaleAction=label&amp;outliersAction=remove&amp;recordSettersAction=ignore&amp;bigAlphagoAction=ignore&amp;alphagozeroAction=ignore&amp;lowOutliersZValueThreshold=-2&amp;highOutliersZValueThreshold=0.76&amp;outlierWindowSize=2&amp;filterText=&amp;splitDomains=&amp;domainsToNotSplit=&amp;ranges=%5Bobject%20Object%5D&amp;splitDlEra=true&amp;splitLargeScaleEra=true&amp;plotRegressions=true&amp;bootstrapSampleSize=10000&amp;adjustForEstimateUncertainty=true&amp;preset=Three%20eras%20of%20compute&amp;aspectRatio=fit%20to%20container&amp;autoresetCamera=true&amp;showTitle=false&amp;labelEras=true&amp;showDoublingTimes=true&amp;showOpenAITrend=false&amp;regressionDecimals=1&amp;textSize=100&amp;systemAlpha=100&amp;systemNames=hide&amp;showLegend=true&amp;parametersRange=NaN,NaN&amp;trainingComputeRange=NaN,NaN&amp;inferenceComputeRange=NaN,NaN&amp;trainingDatasetSizeRange=NaN,NaN&amp;inferenceComputePerParameterRange=NaN,NaN&amp;inferenceComputeTimesParameterRange=NaN,NaN&amp;labelPoints=false">numbers</a> from our grantee organization <a href="https://epochai.org/">Epoch</a>, the original GPT was trained on 1.8e19 FLOP; GPT-2 was 1.5e21 FLOP; the largest version of GPT-3 (175 billion parameters) was trained on 3.1e23 FLOP, and GPT-4 was trained on ~2.1e25 FLOP. There isn&#8217;t clear documentation of GPT-3.5, the model in between GPT-3 and GPT-4, but we suspect that it was retrained from scratch and its effective training compute was order 10^24 FLOP. <a href="#fnref2">&#8617;&#65038;</a></p></li><li><p>See e.g. <a href="https://browse.arxiv.org/pdf/2201.11903.pdf">Wei et al 2023</a> Figures 4, 7, and 8. <a href="#fnref3">&#8617;&#65038;</a></p></li><li><p>Perhaps this would be based on terminal-based text editors that programmers use, which do everything via keyboard commands. <a href="#fnref4">&#8617;&#65038;</a></p></li></ol>]]></content:encoded></item><item><title><![CDATA[Language models surprised us]]></title><description><![CDATA[Most experts were surprised by progress in language models in 2022 and 2023. There may be more surprises ahead, so experts should register their forecasts now about 2024 and 2025.]]></description><link>https://www.planned-obsolescence.org/p/language-models-surprised-us</link><guid isPermaLink="false">https://www.planned-obsolescence.org/p/language-models-surprised-us</guid><dc:creator><![CDATA[Ajeya Cotra]]></dc:creator><pubDate>Tue, 29 Aug 2023 18:37:14 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/80786945-50e7-4669-a3a8-6d377f2b1db4_2000x509.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Kelsey Piper co-drafted this post. Thanks also to Isabel Juniewicz for research help. Audio automatically generated by an AI trained on our voices.</em></p><div class="native-audio-embed" data-component-name="AudioPlaceholder" data-attrs="{&quot;label&quot;:null,&quot;mediaUploadId&quot;:&quot;759eb9bc-b4a1-417c-8ca3-42db354bd6dc&quot;,&quot;duration&quot;:516.96326,&quot;downloadable&quot;:false,&quot;isEditorNode&quot;:true}"></div><p>If you read media coverage of ChatGPT &#8212; which called it <a href="https://www.pcworld.com/article/1424575/chatgpt-is-the-future-of-ai-chatbots.html">&#8216;breathtaking&#8217;, &#8216;dazzling&#8217;</a>, &#8216;<a href="https://www.vanityfair.com/news/2022/12/chatgpt-question-creative-human-robotos">astounding</a>&#8217; &#8212; you&#8217;d get the sense that large language models (LLMs) took the world completely by surprise. Is that impression accurate?</p><p>Actually, yes. There are a few different ways to attempt to measure the question &#8220;Were experts surprised by the pace of LLM progress?&#8221; but they broadly point to the same answer: ML researchers, <a href="https://en.wikipedia.org/wiki/Superforecaster">superforecasters</a>,<a href="#fn1"><sup>[1]</sup></a> and most others were all surprised by the progress in large language models in 2022 and 2023.</p><h3>Competitions to forecast difficult ML benchmarks</h3><p>ML benchmarks are sets of problems which can be objectively graded, allowing relatively precise comparison across different models. We have data from forecasting competitions done in 2021 and 2022 on two of the most comprehensive and difficult ML benchmarks: the <a href="https://paperswithcode.com/dataset/mmlu">MMLU benchmark</a> and the <a href="https://paperswithcode.com/dataset/math">MATH benchmark</a>.<a href="#fn2"><sup>[2]</sup></a></p><p>First, what are these benchmarks?</p><p>The <a href="https://arxiv.org/abs/2009.03300">MMLU</a> dataset consists of multiple choice questions in a variety of subjects collected from sources like GRE practice tests and AP tests. It was intended to test subject matter knowledge in a wide variety of professional domains. MMLU questions are legitimately quite difficult: the average person would probably struggle to solve them.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!11jQ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F094b7197-8d88-4136-b3df-4a6eaa9dfc65_2000x509.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!11jQ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F094b7197-8d88-4136-b3df-4a6eaa9dfc65_2000x509.png 424w, https://substackcdn.com/image/fetch/$s_!11jQ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F094b7197-8d88-4136-b3df-4a6eaa9dfc65_2000x509.png 848w, https://substackcdn.com/image/fetch/$s_!11jQ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F094b7197-8d88-4136-b3df-4a6eaa9dfc65_2000x509.png 1272w, https://substackcdn.com/image/fetch/$s_!11jQ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F094b7197-8d88-4136-b3df-4a6eaa9dfc65_2000x509.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!11jQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F094b7197-8d88-4136-b3df-4a6eaa9dfc65_2000x509.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/094b7197-8d88-4136-b3df-4a6eaa9dfc65_2000x509.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;alt_text&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="alt_text" title="alt_text" srcset="https://substackcdn.com/image/fetch/$s_!11jQ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F094b7197-8d88-4136-b3df-4a6eaa9dfc65_2000x509.png 424w, https://substackcdn.com/image/fetch/$s_!11jQ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F094b7197-8d88-4136-b3df-4a6eaa9dfc65_2000x509.png 848w, https://substackcdn.com/image/fetch/$s_!11jQ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F094b7197-8d88-4136-b3df-4a6eaa9dfc65_2000x509.png 1272w, https://substackcdn.com/image/fetch/$s_!11jQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F094b7197-8d88-4136-b3df-4a6eaa9dfc65_2000x509.png 1456w" sizes="100vw" fetchpriority="high"></picture><div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!NEBA!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F47f901ec-b1a8-4f2c-90b3-26db5c07aa2b_2000x648.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!NEBA!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F47f901ec-b1a8-4f2c-90b3-26db5c07aa2b_2000x648.png 424w, https://substackcdn.com/image/fetch/$s_!NEBA!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F47f901ec-b1a8-4f2c-90b3-26db5c07aa2b_2000x648.png 848w, https://substackcdn.com/image/fetch/$s_!NEBA!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F47f901ec-b1a8-4f2c-90b3-26db5c07aa2b_2000x648.png 1272w, https://substackcdn.com/image/fetch/$s_!NEBA!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F47f901ec-b1a8-4f2c-90b3-26db5c07aa2b_2000x648.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!NEBA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F47f901ec-b1a8-4f2c-90b3-26db5c07aa2b_2000x648.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/47f901ec-b1a8-4f2c-90b3-26db5c07aa2b_2000x648.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;alt_text&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="alt_text" title="alt_text" srcset="https://substackcdn.com/image/fetch/$s_!NEBA!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F47f901ec-b1a8-4f2c-90b3-26db5c07aa2b_2000x648.png 424w, https://substackcdn.com/image/fetch/$s_!NEBA!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F47f901ec-b1a8-4f2c-90b3-26db5c07aa2b_2000x648.png 848w, https://substackcdn.com/image/fetch/$s_!NEBA!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F47f901ec-b1a8-4f2c-90b3-26db5c07aa2b_2000x648.png 1272w, https://substackcdn.com/image/fetch/$s_!NEBA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F47f901ec-b1a8-4f2c-90b3-26db5c07aa2b_2000x648.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>At the time of its introduction in September 2020, most models only performed close to random chance on MMLU (~25%), while GPT-3 performed significantly better than chance at 44%. The benchmark was designed to be harder than any that had come before it, and the authors described their motivation as closing the gap between performance on benchmarks and &#8220;true language understanding&#8221;:</p><blockquote><p>Natural Language Processing (NLP) models have achieved superhuman performance on a number of recently proposed benchmarks. However, these models are still well below human level performance for language understanding as a whole, suggesting a disconnect between our benchmarks and the actual capabilities of these models.</p></blockquote><p>Meanwhile, the <a href="https://paperswithcode.com/dataset/math">MATH dataset</a> consists of free-response questions taken from math contests aimed at the best high school math students in the country. Most college-educated adults would get well under half of these problems right (the authors used computer science undergraduates as human subjects, and their performance ranged from 40% to 90%).</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!qTQp!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F83a88a4f-1eb8-4461-a28c-f08454ec7a05_664x748.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!qTQp!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F83a88a4f-1eb8-4461-a28c-f08454ec7a05_664x748.png 424w, https://substackcdn.com/image/fetch/$s_!qTQp!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F83a88a4f-1eb8-4461-a28c-f08454ec7a05_664x748.png 848w, https://substackcdn.com/image/fetch/$s_!qTQp!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F83a88a4f-1eb8-4461-a28c-f08454ec7a05_664x748.png 1272w, https://substackcdn.com/image/fetch/$s_!qTQp!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F83a88a4f-1eb8-4461-a28c-f08454ec7a05_664x748.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!qTQp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F83a88a4f-1eb8-4461-a28c-f08454ec7a05_664x748.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/83a88a4f-1eb8-4461-a28c-f08454ec7a05_664x748.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;alt_text&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="alt_text" title="alt_text" srcset="https://substackcdn.com/image/fetch/$s_!qTQp!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F83a88a4f-1eb8-4461-a28c-f08454ec7a05_664x748.png 424w, https://substackcdn.com/image/fetch/$s_!qTQp!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F83a88a4f-1eb8-4461-a28c-f08454ec7a05_664x748.png 848w, https://substackcdn.com/image/fetch/$s_!qTQp!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F83a88a4f-1eb8-4461-a28c-f08454ec7a05_664x748.png 1272w, https://substackcdn.com/image/fetch/$s_!qTQp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F83a88a4f-1eb8-4461-a28c-f08454ec7a05_664x748.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>At the time of its introduction in January 2021, the best model achieved only about ~7% accuracy on MATH. The authors say:</p><blockquote><p>We find that accuracy remains low even for the best models. Furthermore, unlike for most other text-based datasets, we find that accuracy is increasing very slowly with model size. If trends continue, then we will need algorithmic improvements, rather than just scale, to make substantial progress on MATH.</p></blockquote><p>So, these are both hard benchmarks &#8212; the problems are difficult for humans, the best models got low performance when the benchmarks were introduced, and the authors seemed to imply it would take a while for performance to get really good.</p><p>In mid-2021, ML professor <a href="https://jsteinhardt.stat.berkeley.edu/">Jacob Steinhardt</a> ran a contest<a href="#fn3"><sup>[3]</sup></a> with superforecasters at <a href="https://www.hypermind.com/en/">Hypermind</a> to predict progress on MATH and MMLU.<a href="#fn4"><sup>[4]</sup></a> Superforecasters massively undershot reality in both cases.</p><p>They predicted that performance on MMLU would improve moderately from 44% in 2021 to 57% by June 2022. The actual performance was 68%, which superforecasters had rated incredibly unlikely.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!6HQN!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55dd9f66-eeb9-44e9-845d-a2abc9931458_2000x497.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!6HQN!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55dd9f66-eeb9-44e9-845d-a2abc9931458_2000x497.png 424w, https://substackcdn.com/image/fetch/$s_!6HQN!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55dd9f66-eeb9-44e9-845d-a2abc9931458_2000x497.png 848w, https://substackcdn.com/image/fetch/$s_!6HQN!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55dd9f66-eeb9-44e9-845d-a2abc9931458_2000x497.png 1272w, https://substackcdn.com/image/fetch/$s_!6HQN!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55dd9f66-eeb9-44e9-845d-a2abc9931458_2000x497.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!6HQN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55dd9f66-eeb9-44e9-845d-a2abc9931458_2000x497.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/55dd9f66-eeb9-44e9-845d-a2abc9931458_2000x497.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;alt_text&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="alt_text" title="alt_text" srcset="https://substackcdn.com/image/fetch/$s_!6HQN!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55dd9f66-eeb9-44e9-845d-a2abc9931458_2000x497.png 424w, https://substackcdn.com/image/fetch/$s_!6HQN!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55dd9f66-eeb9-44e9-845d-a2abc9931458_2000x497.png 848w, https://substackcdn.com/image/fetch/$s_!6HQN!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55dd9f66-eeb9-44e9-845d-a2abc9931458_2000x497.png 1272w, https://substackcdn.com/image/fetch/$s_!6HQN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55dd9f66-eeb9-44e9-845d-a2abc9931458_2000x497.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>Shortly after that, models got even better &#8212; GPT-4 achieved 86.4% on this benchmark, close to the 89.8% that would be &#8220;expert-level&#8221; within each domain, corresponding to 95th percentile among human test takers within a given subtest.</p><p>Superforecasters missed even more dramatically on MATH. They predicted the best model in June 2022 would get ~13% accuracy, and thought it was extremely unlikely that any model would achieve &gt;20% accuracy. In reality, the best model in June 2022 got 50% accuracy,<a href="#fn5"><sup>[5]</sup></a> performing much better than the majority of humans.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!rWyu!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F24e9d9f4-d024-4bbe-a2a3-8de3f7fb55c5_2000x495.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!rWyu!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F24e9d9f4-d024-4bbe-a2a3-8de3f7fb55c5_2000x495.png 424w, https://substackcdn.com/image/fetch/$s_!rWyu!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F24e9d9f4-d024-4bbe-a2a3-8de3f7fb55c5_2000x495.png 848w, https://substackcdn.com/image/fetch/$s_!rWyu!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F24e9d9f4-d024-4bbe-a2a3-8de3f7fb55c5_2000x495.png 1272w, https://substackcdn.com/image/fetch/$s_!rWyu!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F24e9d9f4-d024-4bbe-a2a3-8de3f7fb55c5_2000x495.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!rWyu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F24e9d9f4-d024-4bbe-a2a3-8de3f7fb55c5_2000x495.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/24e9d9f4-d024-4bbe-a2a3-8de3f7fb55c5_2000x495.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;alt_text&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="alt_text" title="alt_text" srcset="https://substackcdn.com/image/fetch/$s_!rWyu!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F24e9d9f4-d024-4bbe-a2a3-8de3f7fb55c5_2000x495.png 424w, https://substackcdn.com/image/fetch/$s_!rWyu!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F24e9d9f4-d024-4bbe-a2a3-8de3f7fb55c5_2000x495.png 848w, https://substackcdn.com/image/fetch/$s_!rWyu!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F24e9d9f4-d024-4bbe-a2a3-8de3f7fb55c5_2000x495.png 1272w, https://substackcdn.com/image/fetch/$s_!rWyu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F24e9d9f4-d024-4bbe-a2a3-8de3f7fb55c5_2000x495.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>Did ML researchers do any better? Steinhardt himself did <em>worse</em> in 2021. In his initial blog post, <a href="https://bounded-regret.ghost.io/ai-forecasting/">Steinhardt remarked</a> that the superforecasters&#8217; predictions on MATH were more aggressive (predicting faster progress) than his own.<a href="#fn6"><sup>[6]</sup></a> We haven&#8217;t found any similar advance predictions from other ML researchers, but Steinhardt&#8217;s impression is that he himself anticipated faster progress than most of his colleagues did.</p><p>However, ML researchers do seem to be improving in their ability to anticipate progress on these benchmarks. In mid-2022, Steinhardt registered his predictions for MATH and MMLU performance in July 2023, and <a href="https://bounded-regret.ghost.io/scoring-ml-forecasts-for-2023/">performed notably better</a>: &#8220;For MATH, the true result was at my 41st percentile, while for MMLU it was at my 66th percentile.&#8221; Steinhardt also argues that ML researchers performed reasonably well on forecasting MATH and MMLU in the late-2022 <a href="https://static1.squarespace.com/static/635693acf15a3e2a14a56a4a/t/64abffe3f024747dd0e38d71/1688993798938/XPT.pdf?ref=bounded-regret.ghost.io">Existential Risk Persuasion Tournament (XPT)</a> (though superforecasters continued to underestimate benchmark progress).</p><h3>Expert surveys about qualitative milestones</h3><p>Not all forms of progress can be easily captured in quantifiable benchmarks. Often we care more about when AI systems will achieve more qualitative <em>milestones</em>: when will they translate as well as a fluent human? When will they beat the best humans at Starcraft? When will they prove novel mathematical theorems?</p><p>Katja Grace of AI Impacts asked ML experts to predict a wide variety of AI milestones, first in 2016 and then again in 2022.</p><p>In 2016, the ML experts were reasonably well-calibrated, but the predictions followed a clear pattern: progress in gameplay and robotics advanced slower than expected, but progress in language use (including programming) advanced more quickly than expected.<a href="#fn7"><sup>[7]</sup></a></p><p>The second iteration of the survey was conducted in mid-2022, a few months before <a href="https://chat.openai.com/">ChatGPT</a> was released. This time accuracy was lower &#8212; experts failed to anticipate the progress that ChatGPT and GPT-4 would soon bring. These models achieved milestones like &#8220;Write an essay for a high school history class&#8221; (actually <a href="https://www.slowboring.com/p/chatgpt-goes-to-harvard">GPT-4 does pretty well in college classes too</a>) or &#8220;Answer easily Googleable factual but open-ended questions better than an expert&#8221; just a few months after the survey was conducted, whereas the experts expected them to take years.</p><p>That means that even after the big 2022 benchmark surprises, experts were still in some cases strikingly conservative about anticipated progress, and undershooting the real situation.</p><h3>Anecdata of researcher impressions</h3><p>ML researchers rarely register predictions, so AI Impacts&#8217; surveys are the best systematic evidence we have about what ML researchers expected ahead of time about qualitative milestones. Anecdotally though, a number of ML experts have expressed that they (and the ML community broadly) were surprised by ChatGPT and GPT-4.</p><p>For a long time, famous cognitive scientist Douglas Hofstadter was among those predicting slow progress. &#8220;I felt it would be hundreds of years before anything even remotely like a human mind&#8221;, he said <a href="https://www.youtube.com/watch?v=lfXxzAVtdpU&amp;t=1763s">in a recent interview</a>.</p><p>Now? &#8220;This started happening at an accelerating pace, where unreachable goals and things that computers shouldn't be able to do started toppling. &#8230;systems got better and better at translation between languages, and then at producing intelligible responses to difficult questions in natural language, and even writing poetry. &#8230;The accelerating progress, has been so unexpected, so completely caught me off guard, not only myself but many, many people, that there is a certain kind of terror of an oncoming tsunami that is going to catch all humanity off guard.&#8221;</p><p>Similarly, <a href="https://www.washingtonpost.com/technology/2023/07/25/ai-bengio-anthropic-senate-hearing/">during a Senate Judiciary Committee hearing last month</a>, acclaimed leading AI researcher <a href="https://en.wikipedia.org/wiki/Yoshua_Bengio">Yoshua Bengio</a> said &#8220;I and many others have been surprised by the giant leap realized by systems like ChatGPT.&#8221;</p><p>In my role as a grantmaker, I&#8217;ve heard many ML academics express similar sentiments in private over the last year. In particular, I&#8217;ve spoken to many researchers who were specifically surprised by the programming and reasoning abilities of GPT-4 (even after seeing the capabilities of the free version of ChatGPT).</p><h3>Another surprise ahead?</h3><p>In 2021, most people were systematically and severely underestimating progress in language models. After a big leap forward in 2022, it looks like ML experts improved in their predictions of benchmarks like MMLU and MATH &#8212; but many still failed to anticipate the qualitative milestones achieved by ChatGPT and then GPT-4, especially in reasoning and programming.</p><p>I think many experts will soon be surprised yet again. Most importantly, ML experts and superforecasters both seem to be massively underestimating future spending on training runs. In the <a href="https://static1.squarespace.com/static/635693acf15a3e2a14a56a4a/t/64abffe3f024747dd0e38d71/1688993798938/XPT.pdf?ref=bounded-regret.ghost.io">XPT tournament</a> mentioned earlier, both groups predicted that the most expensive training run in 2030 would only cost around $100-180M.<a href="#fn8"><sup>[8]</sup></a> Instead, I think that the largest training run will probably cross $1 billion by 2025. This rapid scaleup will probably drive another qualitative leap forward in capability like what we saw over the last 18 months.</p><p>I&#8217;d be really excited for ML researchers to register their forecasts about what AI systems built on language models will be able to do in the next couple of years. I think we need to get good at predicting what language models will be able to do &#8212; in the real world, not just on benchmarks. Massively underestimating near-future progress could be very risky.</p><div><hr></div><ol><li><p>The term "superforecaster" was originally coined by Philip Tetlock andand popularized in his book <em><a href="https://www.amazon.com/Superforecasting-Science-Prediction-Philip-Tetlock/dp/0804136718">Superforecasting</a></em>, though I use the term more generically here to mean "a person who consistently outperforms domain experts and other forecasters by a large amount in forecasting world events." <a href="#fnref1">&#8617;&#65038;</a></p></li><li><p><a href="https://people.eecs.berkeley.edu/~hendrycks/">Dan Hendrycks</a> is the first author on both benchmarks; he did this work while he was a graduate student under Open Philanthropy grantee Jacob Steinhardt. Dan now runs the Center for AI Safety (CAIS), which Open Philanthropy has also funded. <a href="#fnref2">&#8617;&#65038;</a></p></li><li><p>Open Philanthropy funded this forecasting contest. <a href="#fnref3">&#8617;&#65038;</a></p></li><li><p>His contest consisted of six questions, of which two were forecasting MATH and MMLU. The other questions were not about language model capabilities (two were about vision, and others were questions about inputs to AI progress). <a href="#fnref4">&#8617;&#65038;</a></p></li><li><p>Interestingly, the paper that achieved this milestone (<a href="https://arxiv.org/abs/2206.14858">Minerva</a>) was published just one day before the deadline, on June 29, 2022. According to the Minerva paper, the previous published result was the same ~7% accuracy that was reported in the original MATH paper (though the paper claims an unpublished result of ~20%). This means progress on MATH turned out to be pretty &#8220;lumpy,&#8221; jumping a large amount with just one paper. <a href="#fnref5">&#8617;&#65038;</a></p></li><li><p>In fact, in 2021, Steinhardt was surprised that forecasters predicted that models would achieve 50% performance on MATH <em>by 2025.</em> &#8220;I'm still surprised that forecasters predicted 52% on MATH [by 2025], when current accuracy is 7% (!). My estimate would have had high uncertainty, but I'm not sure the top end of my range would have included 50%.&#8221; As we said above, 50% was achieved in _2022. _ <a href="#fnref6">&#8617;&#65038;</a></p></li><li><p>For example, one of the milestones was &#8220;Write Python code to implement algorithms like quicksort.&#8221; Experts predicted that would happen around 2026, but actually it happened in 2021 &#8212; and by 2022 language models could write much more complex pieces of code than quicksort. <a href="#fnref7">&#8617;&#65038;</a></p></li><li><p>Page 57,<a href="https://forecastingresearch.org/news/results-from-the-2022-existential-risk-persuasion-tournament"> "Forecasting Existential Risks: Evidence from a Long-Run Forecasting Tournament"</a> <a href="#fnref8">&#8617;&#65038;</a></p></li></ol>]]></content:encoded></item><item><title><![CDATA[Could AI accelerate economic growth?]]></title><description><![CDATA[Most new technologies don&#8217;t accelerate the pace of economic growth. But advanced AI might do this by massively increasing the research effort going into developing new technologies.]]></description><link>https://www.planned-obsolescence.org/p/could-ai-accelerate-economic-growth</link><guid isPermaLink="false">https://www.planned-obsolescence.org/p/could-ai-accelerate-economic-growth</guid><dc:creator><![CDATA[Tom Davidson]]></dc:creator><pubDate>Tue, 06 Jun 2023 21:45:42 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/dd9a8625-dd6c-4c80-bf3a-523e017537d0_936x532.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Audio automatically generating using an AI trained on Tom&#8217;s voice.</em></p><div class="native-audio-embed" data-component-name="AudioPlaceholder" data-attrs="{&quot;label&quot;:null,&quot;mediaUploadId&quot;:&quot;a66c7289-8dcd-49b6-b82a-71dd65ec7bbf&quot;,&quot;duration&quot;:299.8857,&quot;downloadable&quot;:false,&quot;isEditorNode&quot;:true}"></div><p>What will happen to economic growth once AI has made us all obsolete? Economists are often skeptical of big effects.</p><p>One reason they give is that new technologies typically don&#8217;t accelerate economic growth. Instead, they typically cause a one-time gain in economic output, and then growth continues at its normal rate.</p><p>For example, <a href="http://www.bcaplan.com/">Bryan Caplan</a>, an econ prof at GMU, <a href="https://twitter.com/bryan_caplan/status/1641841635427221508">recently</a> <a href="https://twitter.com/bryan_caplan/status/1641840379556253697">tweeted</a>:</p><blockquote><p>Tech moved 10x faster than I expected in the last year&#8230;</p><p>Economic effects will be modest &amp; gradual. Even electricity took decades to make a huge difference&#8230;</p><p>What I doubt is that any one new tech will raise growth by even 1 percentage-point per year.</p></blockquote><p>This dynamic has played out over the past 50 years. We developed computers and the internet, but economic growth didn&#8217;t speed up, and if anything <a href="https://en.wikipedia.org/wiki/The_Great_Stagnation">it got </a><em><a href="https://en.wikipedia.org/wiki/The_Great_Stagnation">slower</a></em>.</p><p>I think this is the right way to understand the economic impact of current AI. GPT-4 will raise productivity in many sectors as it is gradually adopted across the economy, but it won&#8217;t permanently accelerate economic growth by itself.</p><p>But this doesn&#8217;t mean it&#8217;s impossible to ever accelerate economic growth. In fact, <a href="https://www.openphilanthropy.org/research/modeling-the-human-trajectory/">economic growth has become much faster</a> over the past 2000 years. Over the last few decades, the global economy has grown at about ~3% per year. But around 1800 it grew more slowly, at about ~1% per year. And earlier in time growth was slower still, below 0.1% per year if you go back far enough.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!btrg!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa679619-296c-4573-a70e-0223e877761d_936x532.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!btrg!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa679619-296c-4573-a70e-0223e877761d_936x532.png 424w, https://substackcdn.com/image/fetch/$s_!btrg!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa679619-296c-4573-a70e-0223e877761d_936x532.png 848w, https://substackcdn.com/image/fetch/$s_!btrg!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa679619-296c-4573-a70e-0223e877761d_936x532.png 1272w, https://substackcdn.com/image/fetch/$s_!btrg!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa679619-296c-4573-a70e-0223e877761d_936x532.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!btrg!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa679619-296c-4573-a70e-0223e877761d_936x532.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/aa679619-296c-4573-a70e-0223e877761d_936x532.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;po-gwp&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="po-gwp" title="po-gwp" srcset="https://substackcdn.com/image/fetch/$s_!btrg!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa679619-296c-4573-a70e-0223e877761d_936x532.png 424w, https://substackcdn.com/image/fetch/$s_!btrg!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa679619-296c-4573-a70e-0223e877761d_936x532.png 848w, https://substackcdn.com/image/fetch/$s_!btrg!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa679619-296c-4573-a70e-0223e877761d_936x532.png 1272w, https://substackcdn.com/image/fetch/$s_!btrg!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa679619-296c-4573-a70e-0223e877761d_936x532.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>From David Roodman&#8217;s <a href="https://www.openphilanthropy.org/research/modeling-the-human-trajectory/">Modeling the Human Trajectory</a>. The solid line is historical economic growth from 10,000 BC to 2019. If the rate of growth were constant, this graph should look like a straight line. But it curves upwards, showing economic growth accelerating over time.</p><p>So, if new technologies don&#8217;t ever really accelerate economic growth, why is growth so much higher today than it was 2000 years ago? The consensus view in economics is that modern growth is so fast because we put <strong>continual effort into innovation</strong><a href="#fn1"><sup>[1]</sup></a>. We deliberately invest in R&amp;D to invent new technologies, we put effort into making our manufacturing processes more efficient, we design supply chains to efficiently distribute new technologies quickly across the economy, and so on.</p><p>The world is collectively putting more effort into innovation now compared to 2000 years ago for two main reasons: we have more people overall, and a larger fraction of those people are working on developing new technologies.</p><ul><li><p>On the first point: The global population is about thirty times larger than it was 2000 years ago, so there are more people who can potentially come up with ideas for new technologies, and more people who can work to make them a reality.</p></li><li><p>On the second point: a larger fraction of the population today specializes in R&amp;D for new technologies.</p><ul><li><p>One big reason for this is better education &#8212; 2000 years ago almost nobody received an academic education; today, mass education means that a larger fraction of the population has the background skills needed to contribute to technology R&amp;D.<a href="#fn2"><sup>[2]</sup></a></p></li><li><p>Another major reason is better institutions for encouraging and enabling innovation. For example, in the past you had to be independently wealthy to be an inventor, because you had to fund your own research. But today, investment markets and government grants will often finance promising ideas, so you can try to invent new technologies even if you couldn&#8217;t fund all the research yourself.</p></li></ul></li></ul><p>Today around ~20 million people work in R&amp;D worldwide.<a href="#fn3"><sup>[3]</sup></a> Two thousand years ago the effort going into research was much smaller, I&#8217;d guess by a factor of 1000.<a href="#fn4"><sup>[4]</sup></a> Economic growth is much faster today than it was 2000 years ago not because of any single new technology, but because we are putting in so much more effort into generating a steady stream of new innovations.</p><p>I think future AI might massively increase the world&#8217;s innovation efforts again, and thereby accelerate economic growth. Rather than being &#8220;just one more technology&#8221;, it might massively increase the pace at which humanity develops new technologies.</p><p>How much might future AI increase the world&#8217;s innovation efforts?</p><p>As I discussed in a previous <a href="https://www.planned-obsolescence.org/continuous-doesnt-mean-slow/">post</a>, once we develop AI that is &#8220;expert human level&#8221; at AI research, it might not be long before we have AI that is way beyond human experts in all domains. That is, AI that is way better than the best humans at thinking of new ideas, designing experiments to test those ideas, building new technologies, running organizations, and navigating bureaucracies.<a href="#fn5"><sup>[5]</sup></a></p><p>What&#8217;s more, because it takes so many more computer chips to train powerful AI than to run it, once we&#8217;ve trained these superhuman AIs we would potentially have enough computation to run them on <em>billions</em> of tasks in parallel.<a href="#fn6"><sup>[6]</sup></a> There could be massive research organizations where AIs manage other AIs to conduct millions of research projects in parallel. And these AIs could innovate tirelessly day and night.<a href="#fn7"><sup>[7]</sup></a></p><p>As well as having superhuman intelligence, these AIs could think much more quickly than humans. ChatGPT Turbo can <a href="https://fabienroger.github.io/trackoai/">already write ~800 words per minute</a>, whereas humans typically write about 40 words per minute. So AI can already write ~20X faster than humans. In just one day, each AI could potentially think as many thoughts as a human thinks in a month.<a href="#fn8"><sup>[8]</sup></a></p><p>My best guess is that AI this powerful would increase the world&#8217;s innovative efforts by more than 100X. Perhaps, like before, this would <a href="https://www.openphilanthropy.org/research/could-advanced-ai-drive-explosive-economic-growth/">significantly accelerate economic growth</a>.</p><div><hr></div><ol><li><p>In <a href="https://en.wikipedia.org/wiki/Solow%E2%80%93Swan_model">standard neoclassical models</a> growth is ultimately driven by better technology, which the model assumes improves exponentially. <a href="https://en.wikipedia.org/wiki/Jones_model">Semi-endogenous models</a> go further in modeling technological progress as resulting from targeted R&amp;D efforts. <a href="https://www.openphilanthropy.org/research/could-advanced-ai-drive-explosive-economic-growth/#7211-frankel-1962">Other models</a> take a different approach and represent technological progress as driven by &#8220;learning by doing&#8221;. In all these models, growth is ultimately driven by innovation.<br>There are some growth models where growth is ultimately driven by capital accumulation rather than technological progress. But these aren&#8217;t particularly popular and they must deny that there are diminishing returns to capital on the current margin. Interpreted at face value, they imply that developed countries became richer over the past 100 years solely by producing more of the goods that already existed 100 years ago, rather than by developing higher quality goods and new technologies.<br>I do think standard growth models miss part of the picture, which is that growth in GDP/capita has been in part driven by one-time changes like <a href="http://klenow.com/HHJK.pdf">increased workforce participation by women</a>. <a href="#fnref1">&#8617;&#65038;</a></p></li><li><p>Better education also means that the people doing R&amp;D are more skilled on average. <a href="#fnref2">&#8617;&#65038;</a></p></li><li><p><a href="https://www.nsf.gov/statistics/2018/nsb20181/report/sections/science-and-engineering-labor-force/global-s-e-labor-force">OECD</a> data from 2015 estimates that the global workforce in science and engineering is 7 million, though this omits India and Brazil. A <a href="https://onlinelibrary.wiley.com/doi/full/10.1111/1758-5899.13182">recent paper</a> estimates ~20 million full-time equivalents do R&amp;D worldwide. <a href="#fnref3">&#8617;&#65038;</a></p></li><li><p>Why a factor of 1000? Population was ~30X lower, and various data sources suggest research concentration in 1800 was 30X lower than today (see the subsection &#8220;Data on the research concentration in 1800&#8221; in <a href="https://www.openphilanthropy.org/research/social-returns-to-productivity-growth/#appendix-f-back-of-the-envelope-calculations-of-value-of-rd-in-1800">this report</a>). Combining those factors, the number of researchers 2000 years ago was lower than today by a factor of 30*30 = ~1000X.<br>In fact, the fraction of people doing research 2000 years ago was likely lower than in 1800, suggesting an even bigger difference than 1000X. On the other hand, research effort 2000 years ago may have come more from many people making small innovations in their personal workflows than from full time researchers. <a href="#fnref4">&#8617;&#65038;</a></p></li><li><p>There are some innovative activities that disembodied AI couldn&#8217;t automate, because they require interacting with the physical world. This could significantly limit AI&#8217;s effect on economic growth. On the other hand, AI might design robots that can do all the physical tasks that humans can do. Then AIs could control these robots remotely and perform all the tasks involved in innovation. <a href="#fnref5">&#8617;&#65038;</a></p></li><li><p>Of course, we don&#8217;t know how much compute superhuman AI would take to train or to run. To guess at this, I estimated how many tasks GPT-8 could perform in parallel using only the computer chips needed to train it.<br>In a <a href="https://www.planned-obsolescence.org/continuous-doesnt-mean-slow/#fn2">previous post</a> I estimated that GPT-4 could perform 300,000 tasks in parallel with the compute used to train it. Compared to GPT-4, I assumed that GPT-8 would have 10,000 times as many parameters and need 100 million times as much computing power to train (in line with the Chinchilla scaling law). This implies that GPT-8 could perform 10,000X (= 100 million/10,000) as many tasks in parallel. I.e. 3 billion (=300,000 * 10,000) tasks. See <a href="https://docs.google.com/spreadsheets/d/1qzq4dYkSAccuRnIsXNkIzJ4Vd6SDxrsXSMCrx9hrWiw/edit#gid=511355734">calc</a>. <a href="#fnref6">&#8617;&#65038;</a></p></li><li><p>What&#8217;s more, the size of this AI workforce could grow rapidly over time. AI could work to increase the number of AIs and how smart they are by designing better AI algorithms, designing better AI chips, and investing more money to build more AI chips. Already AI algorithms are becoming about <a href="https://epochai.org/blog/revisiting-algorithmic-progress">twice as efficient each year</a>, AI chips are becoming twice as efficient every ~2-3 years, and investments in AI are growing quickly. If this pace of improvement keeps up, the size of the AI workforce would more than double every year! This fast-growing workforce could innovate quickly despite ideas becoming harder to find. <a href="#fnref7">&#8617;&#65038;</a></p></li><li><p>A previous footnote argued that if we took the computer chips that were used to train superhuman AI and used them to run copies of the superhuman AI, we could run 3 billion copies in parallel. They could work tirelessly day and night, rather than the ~8 hours/day from human workers, which increases the size of the effective AI workforce to 9 billion.<br>If 2 million out of these 9 billion AIs work on innovation, that is already a 100X increase on the current size of the R&amp;D workforce (~20 million &#8211; see previous footnote). But there are two reasons the increase in innovative effort will be bigger than this. Firstly, each superhuman AI is much more smart and productive than the best R&amp;D workers today. This is a massive effect, bigger than turning every scientist alive today into a top performer in their field. Lastly, there&#8217;s a large gain in productivity from the AIs being able to think faster. Rather than having 2 billion AIs thinking at human speed, we could have 100 million AIs thinking at 20X human speed. <a href="#fnref8">&#8617;&#65038;</a></p></li></ol>]]></content:encoded></item><item><title><![CDATA[The costs of caution]]></title><description><![CDATA[If you thought we might be able to cure cancer in 2200, then I think you ought to expect there&#8217;s a good chance we can do it within years of the advent of AI systems that can do the research work humans can do.]]></description><link>https://www.planned-obsolescence.org/p/the-costs-of-caution</link><guid isPermaLink="false">https://www.planned-obsolescence.org/p/the-costs-of-caution</guid><dc:creator><![CDATA[Kelsey Piper]]></dc:creator><pubDate>Mon, 01 May 2023 17:34:19 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/c43663ef-c69e-4145-9e05-f7306d09fcaf_612x951.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Audio automatically generating using an AI trained on Kelsey&#8217;s voice.</em></p><div class="native-audio-embed" data-component-name="AudioPlaceholder" data-attrs="{&quot;label&quot;:null,&quot;mediaUploadId&quot;:&quot;effec29c-e0ba-482c-bca5-dd8f27d5d7d8&quot;,&quot;duration&quot;:260.12735,&quot;downloadable&quot;:false,&quot;isEditorNode&quot;:true}"></div><p>Josh Cason on Twitter raised an objection to recent calls for a moratorium on AI development:</p><blockquote><p><em>Or raise your hand if you or someone you love has a terminal illness, believes Ai has a chance at accelerating medical work exponentially, and doesn't have til Christmas, to wait on your make believe moratorium. Have a heart man &#10084;&#65039; <a href="https://t.co/wHK86uAYoA">https://t.co/wHK86uAYoA</a></em></p><p>&#8212; Josh Cason (@TheGrizztronic) <a href="https://twitter.com/TheGrizztronic/status/1642322551253745665?ref_src=twsrc%5Etfw">April 2, 2023</a></p></blockquote><p>I&#8217;ve said that I think we should ideally <a href="https://www.planned-obsolescence.org/is-it-time-for-a-pause/">move a lot slower</a> on developing powerful AI systems. I still believe that. But I think Josh&#8217;s objection is important and deserves a full airing.</p><p>Approximately 150,000 people die worldwide every day. Nearly all of those deaths are, in some sense, preventable, with sufficiently advanced medical technology. Every year, five million families bury a child dead before their fifth birthday. Hundreds of millions of people live in extreme poverty. Billions more have far too little money to achieve their dreams and grow into their full potential. Tens of billions of animals are tortured on factory farms.</p><p>Scientific research and economic progress could make an enormous difference to all these problems. Medical research could cure diseases. Economic progress could make food, shelter, medicine, entertainment and luxury goods accessible to people who can't afford it today. Progress in meat alternatives could allow us to shut down factory farms.</p><p>There are tens of thousands of scientists, engineers, and policymakers working on fixing these kinds of problems &#8212; working on developing vaccines and antivirals, understanding and arresting aging, treating cancer, building cheaper and cleaner energy sources, developing better crops and homes and forms of transportation. But there are only so many people working on each problem. In each field, there are dozens of useful, interesting subproblems that no one is working on, because there aren&#8217;t enough people to do the work.</p><p>If we could train AI systems powerful enough to automate everything these scientists and engineers do, they could help.</p><p>As Tom discussed in a <a href="https://www.planned-obsolescence.org/continuous-doesnt-mean-slow/">previous post</a>, once we develop AI that does AI research as well as a human expert, it might not be long before we have AI that is way beyond human experts in all domains. That is, AI which is way better than the best humans at all aspects of medical research: thinking of new ideas, designing experiments to test those ideas, building new technologies, and navigating bureaucracies.</p><p>This means that rather than tens of thousands of top biomedical researchers, we could have hundreds of millions of significantly superhuman biomedical researchers.<a href="#fn1"><sup>[1]</sup></a></p><p>That&#8217;s more than a thousand times as much effort going into tackling humanity&#8217;s biggest killers. If you thought we might be able to cure cancer in 2200, then I think you ought to expect there&#8217;s a good chance we can do it within years of the advent of AI systems that can do the research work humans can do.<a href="#fn2"><sup>[2]</sup></a></p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!kGTE!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7dbac9d6-3757-44bf-bbbc-799919e4889c_612x951.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!kGTE!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7dbac9d6-3757-44bf-bbbc-799919e4889c_612x951.png 424w, https://substackcdn.com/image/fetch/$s_!kGTE!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7dbac9d6-3757-44bf-bbbc-799919e4889c_612x951.png 848w, https://substackcdn.com/image/fetch/$s_!kGTE!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7dbac9d6-3757-44bf-bbbc-799919e4889c_612x951.png 1272w, https://substackcdn.com/image/fetch/$s_!kGTE!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7dbac9d6-3757-44bf-bbbc-799919e4889c_612x951.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!kGTE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7dbac9d6-3757-44bf-bbbc-799919e4889c_612x951.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7dbac9d6-3757-44bf-bbbc-799919e4889c_612x951.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!kGTE!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7dbac9d6-3757-44bf-bbbc-799919e4889c_612x951.png 424w, https://substackcdn.com/image/fetch/$s_!kGTE!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7dbac9d6-3757-44bf-bbbc-799919e4889c_612x951.png 848w, https://substackcdn.com/image/fetch/$s_!kGTE!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7dbac9d6-3757-44bf-bbbc-799919e4889c_612x951.png 1272w, https://substackcdn.com/image/fetch/$s_!kGTE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7dbac9d6-3757-44bf-bbbc-799919e4889c_612x951.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption"><a href="https://www.smbc-comics.com/comic/2013-06-02">SMBC 2013-06-02 "The Falling Problem"</a>, Zach Wienersmith</figcaption></figure></div><p>All this may be a massive underestimate. This envisions a world that&#8217;s pretty much like ours except that extraordinary talent is no longer scarce. But that feels, in some senses, like thinking about the advent of electricity purely in terms of &#8216;torchlight will no longer be scarce&#8217;. Electricity did make it very cheap to light our homes at night. But it also enabled vacuum cleaners, washing machines, cars, smartphones, airplanes, video recording, Twitter &#8212; entirely new things, not just cheaper access to things we already used.</p><p>If it goes well, I think developing AI that obsoletes humans will more or less bring the 24th century crashing down on the 21st. Some of the impacts of that are mostly straightforward to predict. We will almost certainly cure a lot of diseases and make many important goods much cheaper. Some of the impacts are pretty close to unimaginable.</p><p>Since I was fifteen years old, I have harbored the hope that scientific and technological progress will come fast enough. I hoped advances in the science of aging would let my grandparents see their great-great-grandchildren get married.</p><p>Now my grandparents are in their nineties. I think hastening advanced AI might be their best shot at living longer than a few more years, but I&#8217;m still advocating for us to slow down. The risk of <a href="https://www.planned-obsolescence.org/the-training-game/">a catastrophe there&#8217;s no recovering from</a> seems too high.<a href="#fn3"><sup>[3]</sup></a> It&#8217;s worth going slowly to be more sure of getting this right, to better understand what we&#8217;re building and think about its effects.</p><p>But I&#8217;ve seen some people make the case for caution by <a href="https://twitter.com/GaryMarcus/status/1640324679968903177">asking</a>, basically, &#8216;why are we risking the world for these trivial toys?&#8217; And I want to make it clear that the assumption behind both AI optimism and AI pessimism is that these are not just goofy chatbots, but an early research stage towards developing a second intelligent species. Both AI fears and AI hopes rest on the belief that it may be possible to build alien minds that can do everything we can do and much more. What&#8217;s at stake, if that&#8217;s true, isn&#8217;t whether we&#8217;ll have fun chatbots. It&#8217;s the life-and-death consequences of delaying, and the possibility we&#8217;ll screw up and kill everyone.</p><div><hr></div><ol><li><p>Tom argues that the compute needed to train GPT-6 would be enough to have it perform tens of millions of tasks in parallel. We expect that the training compute for superhuman AI will allow you to run many more copies still. <a href="#fnref1">&#8617;&#65038;</a></p></li><li><p>In fact, I think it might be even more explosive than that &#8212; even as these superhuman digital scientists conduct medical research for us, <em>other</em> AIs will be working on rapidly improving the capabilities of these digital biomedical researchers, and other AIs still will be improving hardware efficiency and building more hardware so that we can run increasing numbers of them. <a href="#fnref2">&#8617;&#65038;</a></p></li><li><p>This assumes we don&#8217;t make much progress on figuring out how to build such systems safely. Most of my hope is that we will slow down and figure out how to do this right (or be slowed down by external factors like powerful AI being very hard to develop), and if we give ourselves a lot more time, then I&#8217;m optimistic. <a href="#fnref3">&#8617;&#65038;</a></p></li></ol>]]></content:encoded></item><item><title><![CDATA[Continuous doesn’t mean slow]]></title><description><![CDATA[Once a lab trains AI that can fully replace its human employees, it will be able to multiply its workforce 100,000x. If these AIs do AI research, they could develop vastly superhuman systems in under a year.]]></description><link>https://www.planned-obsolescence.org/p/continuous-doesnt-mean-slow</link><guid isPermaLink="false">https://www.planned-obsolescence.org/p/continuous-doesnt-mean-slow</guid><dc:creator><![CDATA[Tom Davidson]]></dc:creator><pubDate>Wed, 12 Apr 2023 16:43:24 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/47df2581-5d1a-45b5-b544-452864e73775_896x688.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="native-audio-embed" data-component-name="AudioPlaceholder" data-attrs="{&quot;label&quot;:null,&quot;mediaUploadId&quot;:&quot;665876e1-415b-4323-ac6f-14ce994792ee&quot;,&quot;duration&quot;:233.32571,&quot;downloadable&quot;:false,&quot;isEditorNode&quot;:true}"></div><p>There&#8217;s a lot of disagreement about how likely AI is to end up overthrowing humanity. Thoughtful pundits <a href="https://www.cold-takes.com/making-the-best-of-the-most-important-century/#open-question-how-hard-is-the-alignment-problem">vary from &lt;5% to &gt;90%</a>. What&#8217;s driving this disagreement?</p><p>One factor that often comes up in discussions is <strong>takeoff speeds</strong>, which Ajeya mentioned in the <a href="https://www.planned-obsolescence.org/ais-accelerating-ai-research/">previous post</a>. How quickly and suddenly do we move from today&#8217;s AI, to &#8220;expert-human level&#8221; AI<a href="#fn1"><sup>[1]</sup></a>, to AI that is way beyond human experts and could easily overpower humanity?</p><p>The final stretch &#8212; the transition from expert-human level AI to AI systems that can easily overpower all of us &#8212; is especially crucial. If this final transition happens slowly, we could potentially have a long time to get used to the obsolescence regime and use very competent AI to <a href="https://www.google.com/url?q=https://www.planned-obsolescence.org/training-ais-to-help-us-align-ais/&amp;sa=D&amp;source=docs&amp;ust=1680729530528442&amp;usg=AOvVaw3ir2tWrYuBwDjbKIHuNIuc">help us solve AI alignment</a> (among other things). But if it happens very quickly, we won&#8217;t have much time to ensure superhuman systems are aligned, or to prepare for human obsolescence in any other way.</p><p>Scott Alexander is optimistic that things might move gradually. In a recent <a href="https://astralcodexten.substack.com/p/why-i-am-not-as-much-of-a-doomer">ACX post</a> titled &#8216;<em>Why I Am Not (As Much Of) A Doomer (As Some People)</em>&#8217;, he says:</p><blockquote><p>So far we&#8217;ve had brisk but still gradual progress in AI; GPT-3 is better than GPT-2, and GPT-4 will probably be better still. Every few years we get a new model which is better than previous models by some predictable amount.</p><p>Some people (eg <a href="https://www.lesswrong.com/posts/GNhMPAWcfBCASy8e6/a-central-ai-alignment-problem-capabilities-generalization">Nate Soares</a>) worry there&#8217;s a point where this changes&#8230; Maybe some jump&#8230; could take an AI from IQ 90 to IQ 1000 with no (or very short) period of IQ 200 in between&#8230;</p><p>I&#8217;m optimistic because the past few years have provided some evidence for gradual progress.</p></blockquote><p>I agree with Scott that recent AI progress has been continuous and fairly predictable, and don&#8217;t particularly expect a break in that trend. But I expect the transition to superhuman AI to be very fast, even if it&#8217;s continuous.</p><p>The amount of &#8220;compute&#8221; (i.e. the number of AI chips) needed to <em>train</em> a powerful AI is much bigger than the amount of compute needed to <em>run</em> it. I estimate that OpenAI has enough compute to run GPT-4 on <em>hundreds of thousands</em> of tasks at once.<a href="#fn2"><sup>[2]</sup></a></p><p>This ratio will only become more extreme as models get bigger. Once OpenAI trains GPT-5 it&#8217;ll have enough compute for GPT-5 to perform <em>millions</em> of tasks in parallel, and once they train GPT-6 it&#8217;ll be able to perform <em>tens of millions</em> of tasks in parallel.<a href="#fn3"><sup>[3]</sup></a></p><p>Now imagine that GPT-6 is as good at AI research as the average OpenAI researcher.<a href="#fn4"><sup>[4]</sup></a> OpenAI could expand their AI researcher workforce from <em>hundreds</em> of experts to <em>tens of millions</em>. That&#8217;s a mind-boggling large increase, a factor of 100,000. It&#8217;s like going from 1000 people to the <a href="https://www.statista.com/statistics/269959/employment-in-the-united-states/">entire US workforce</a>. What&#8217;s more, these AIs could work tirelessly through the night and could potentially &#8220;think&#8221; <em>much</em> more quickly than human workers.<a href="#fn5"><sup>[5]</sup></a> (This change won&#8217;t happen all-at-once. I expect speed-ups from less capable AI before this point, as Ajeya wrote in the <a href="https://www.planned-obsolescence.org/ais-accelerating-ai-research/">previous post</a>.)</p><p>How much faster would AI progress be in this scenario? It&#8217;s hard to know. But my <a href="https://docs.google.com/document/d/1os_4YOw6Xv33KjX-kR76D3kW1drkWRHKG2caeiEWzNs/edit#heading=h.63n1i9duqyot">best guess</a>, from my recent <a href="https://www.lesswrong.com/posts/Gc9FGtdXhK9sCSEYu/what-a-compute-centric-framework-says-about-ai-takeoff">report</a> on takeoff speeds, is that progress would be <em>much much</em> faster. I think that less than a year after AI is expert-human level at AI research, AI could improve to the point of being able to easily overthrow humanity.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!P4OA!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75769404-a86c-4142-9cfb-d995887b93be_896x688.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!P4OA!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75769404-a86c-4142-9cfb-d995887b93be_896x688.png 424w, https://substackcdn.com/image/fetch/$s_!P4OA!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75769404-a86c-4142-9cfb-d995887b93be_896x688.png 848w, https://substackcdn.com/image/fetch/$s_!P4OA!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75769404-a86c-4142-9cfb-d995887b93be_896x688.png 1272w, https://substackcdn.com/image/fetch/$s_!P4OA!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75769404-a86c-4142-9cfb-d995887b93be_896x688.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!P4OA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75769404-a86c-4142-9cfb-d995887b93be_896x688.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/75769404-a86c-4142-9cfb-d995887b93be_896x688.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!P4OA!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75769404-a86c-4142-9cfb-d995887b93be_896x688.png 424w, https://substackcdn.com/image/fetch/$s_!P4OA!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75769404-a86c-4142-9cfb-d995887b93be_896x688.png 848w, https://substackcdn.com/image/fetch/$s_!P4OA!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75769404-a86c-4142-9cfb-d995887b93be_896x688.png 1272w, https://substackcdn.com/image/fetch/$s_!P4OA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75769404-a86c-4142-9cfb-d995887b93be_896x688.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>This is much faster than the timeline mentioned in the ACX post:</p><blockquote><p>if you&#8217;re imagining specific years, imagine human-genius-level AI in the 2030s and world-killers in the 2040s</p></blockquote><p>The cause of this fast transition isn&#8217;t that there&#8217;s a break in the trend of continuous progress. It&#8217;s that expert-human-level AI <a href="https://www.planned-obsolescence.org/ais-accelerating-ai-research/">massively accelerates AI progress</a>, causing this continuous progress to happen at a blistering pace.</p><p>Of course, this isn&#8217;t inevitable. Labs could choose <em>not</em> to use AI to accelerate AI progress, at least once AI gets sufficiently powerful. But it will be a tempting move, and they&#8217;re more likely to be cautious if they make specific and verifiable commitments in advance to <a href="https://www.planned-obsolescence.org/is-it-time-for-a-pause/">pause AI progress</a>.</p><div><hr></div><ol><li><p>I&#8217;m operationalizing &#8220;expert-human level AI&#8221; as &#8220;each forward pass of the AI produces as much useful output as 0.1 seconds of thought from a human expert&#8221;. It&#8217;s possible that AI will produce expert-level output by having many dumber AIs working together and thinking for much longer than a human expert would, but under my definition that wouldn&#8217;t count as expert-level AI because the quality of the AI&#8217;s thinking is below expert level. <a href="#fnref1">&#8617;&#65038;</a></p></li><li><p>My calculation assumes that GPT-4 processes ten tokens per second on each task that it&#8217;s being applied to. Here&#8217;s how my <a href="https://docs.google.com/spreadsheets/d/1aZg-ccL3Al3d1YItu--12QHWvSJCQ0YdfoMUmiHKGGk/edit#gid=1060982759">estimate</a> works: the training compute for GPT-4 has been estimated at ~3e25 total FLOP (<a href="https://colab.research.google.com/drive/1xOVSTfb52IyJxsM0rBUnSNoIdCisTOPx?usp=sharing">source</a>, h/t <a href="https://epochai.org/">Epoch</a>). I assume the training took 4 months, implying that amount of compute used per second during training was 3e18 FLOP/s. How many instances of GPT-4 could you run with this compute? If GPT-4 was trained with 3e25 FLOP in accordance with <a href="https://arxiv.org/abs/2203.15556">Chinchilla scaling,</a> that implies it will require ~1e12 FLOP per forward pass. So you could do 3e18/1e12 = ~3e6 forward passes per second. In other words, you could run GPT-4 on ~3 million tasks in parallel, with it processing one token per second on each task, or on ~300,000 tasks in parallel at ten tokens per second.<br>Though there wouldn&#8217;t be enough <em>memory</em> to store 300,000 separate copies of the weights, this calculation suggests that there would be enough <em>processing power</em> to apply those weights to 300,000 different word generation tasks per second. (Each copy of the weights can perform many tasks in parallel.) <a href="#fnref2">&#8617;&#65038;</a></p></li><li><p>I make the simple <a href="https://docs.google.com/spreadsheets/d/1gTS548D0DtsjipkeHZQblsmHIAU5zpMAqxb2L31QNpc/edit#gid=1060982759">assumption</a> that GPT-5 will be the same as GPT-4 except for having 10X the parameters and being trained on 10X the data, and that GPT-6 will have an additional 10X parameters and 10X data. <a href="#fnref3">&#8617;&#65038;</a></p></li><li><p>More precisely, assume that each forward pass of GPT-6 is as useful for advancing AI capabilities as 0.1 seconds of thought by an average OpenAI researcher. I.e. if OpenAI has 300 capabilities researchers today, then you could match their total output by running 300 copies of GPT-6 in parallel and having each of them produce 10 tokens per second. <a href="#fnref4">&#8617;&#65038;</a></p></li><li><p>Rather than 10 million AIs thinking at human speed, OpanAI could potentially have 1 million AIs thinking 10X faster than a human, or 100,000 AIs thinking 100X faster. <a href="#fnref5">&#8617;&#65038;</a></p></li></ol>]]></content:encoded></item><item><title><![CDATA[AIs accelerating AI research]]></title><description><![CDATA[Researchers could potentially design the next generation of ML models more quickly by delegating some work to existing models, creating a feedback loop of ever-accelerating progress.]]></description><link>https://www.planned-obsolescence.org/p/ais-accelerating-ai-research</link><guid isPermaLink="false">https://www.planned-obsolescence.org/p/ais-accelerating-ai-research</guid><dc:creator><![CDATA[Ajeya Cotra]]></dc:creator><pubDate>Tue, 04 Apr 2023 20:01:47 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!oyGI!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5519227-580c-4379-999d-03eabb6ed120_1200x1200.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="native-audio-embed" data-component-name="AudioPlaceholder" data-attrs="{&quot;label&quot;:null,&quot;mediaUploadId&quot;:&quot;92693293-b8b6-462e-977b-f4d82a3d04ed&quot;,&quot;duration&quot;:305.76328,&quot;downloadable&quot;:false,&quot;isEditorNode&quot;:true}"></div><p>The concept of an &#8220;intelligence explosion&#8221; has played an important role in discourse about advanced AI for decades. Early computer scientist <a href="https://en.wikipedia.org/wiki/I._J._Good">I.J. Good</a> described it like this in 1965:</p><blockquote><p>Let an ultraintelligent machine be defined as a machine that can far surpass all the intellectual activities of any man however clever. Since the design of machines is one of these intellectual activities, an ultraintelligent machine could design even better machines; there would then unquestionably be an &#8216;intelligence explosion,&#8217; and the intelligence of man would be left far behind. Thus the first ultraintelligent machine is the last invention that man need ever make, provided that the machine is docile enough to tell us how to keep it under control.</p></blockquote><p>This presentation, like most other popular presentations of the intelligence explosion concept, focuses on what happens <em>after</em> we have a single AI system that can already do better at <em>every</em> task than <em>any</em> human (which Good calls an &#8220;ultraintelligent machine&#8221; above, and others have called &#8220;an artificial superintelligence&#8221;). It calls to mind an image of AI progress with two phases:</p><ul><li><p>In Phase 1, humans are doing all the AI research, and progress ramps up steadily. We can more or less predict the rate of future progress (i.e. how quickly AI systems will improve their capabilities) by extrapolating from past rates of progress.<a href="#fn1"><sup>[1]</sup></a></p></li><li><p>Eventually humans succeed at building an artificial superintelligence (or ASI), leading to Phase 2. In Phase 2, this ASI is doing all of the AI research by itself. All of a sudden, progress in AI capabilities is no longer bottlenecked by slow human researchers, and an intelligence explosion is kicked off. The rate of progress in AI research goes up sharply &#8212; perhaps years of progress is compressed into days or weeks.</p></li></ul><p>But I think this picture is probably too all-or-nothing. Today&#8217;s large language models (LLMs) like GPT-4 are not (yet) capable of completely taking over AI research by themselves &#8212; but they are able to write code, come up with ideas for ML experiments, and help troubleshoot bugs and other issues. Anecdotally, several ML researchers I know are starting to delegate simple tasks that come up in their research to these LLMs, and they say that makes them meaningfully more productive. (When chatGPT went down for 6 hours, I know of one ML researcher who postponed their coding tasks for 6 hours and worked on other things in the meantime.<a href="#fn2"><sup>[2]</sup></a>)</p><p>If this holds true more broadly, researchers could potentially design and train the next generation of ML models more quickly and easily by delegating to existing LLMs.<a href="#fn3"><sup>[3]</sup></a> This calls to mind a more continuous &#8220;intelligence explosion&#8221; that begins <em>before</em> we have any single artificial superintelligence:</p><ul><li><p>Currently, human researchers collectively are responsible for almost all of the progress in AI research, but are starting to delegate a small fraction of the work to large language models. This makes it somewhat easier to design and train the next generation of models.</p></li><li><p>The next generation is able to handle harder tasks and more different types of tasks, so human researchers delegate more of their work to them. This makes it <em>significantly</em> easier to train the generation after that. Using models gives a much bigger boost than it did the last time around.</p></li><li><p>Each round of this process makes the whole field move faster and faster. In each round, human researchers delegate everything they can productively delegate to the current generation of models &#8212; and the more powerful those models are, the more they contribute to research and thus the faster AI capabilities can improve.</p></li></ul><p>This feedback loop could be getting started now. If it goes on for enough cycles without hitting any fundamental blockers, at some point our AI systems will have taken over <em>all</em> the work involved in designing more powerful AI systems. And it could keep going beyond that, with a research community consisting entirely of AIs working at an inhuman pace to make yet-more-sophisticated AIs. Once AI systems have automated AI research entirely, I think it&#8217;s likely that the <a href="https://www.planned-obsolescence.org/what-were-doing-here/">full obsolescence regime</a> that we discussed in our first post will come soon after.<a href="#fn4"><sup>[4]</sup></a></p><p>If so, the end state would be similar to what IJ Good envisioned &#8212; we could have &#8220;artificial superintelligence&#8221;<a href="#fn5"><sup>[5]</sup></a> that improves AI capabilities further and quickly leaves human capabilities far behind. But <em>before</em> we have artificial superintelligence, we might have <em>already</em> vastly accelerated the pace of progress in AI research<a href="#fn6"><sup>[6]</sup></a> with the help of lesser models.</p><p>Exactly how much acceleration might happen before we have AI systems that can handle all the AI research by themselves, and how much might happen after? Will it feel like a pretty sudden jump &#8212; we spend a while with some neat, mildly useful AI assistants and then all of a sudden we develop AI that obsoletes humanity? Or will we have many years in which AI systems get increasingly impressive and perceptibly accelerate the pace of progress before humans are fully obsolete?</p><p>This is a very complicated question that I&#8217;m not going to get into in this post, but my colleague Tom Davidson put out a thorough research report exploring <strong>takeoff speeds</strong> &#8212; essentially, how quickly and suddenly we move from the world of today to the obsolescence regime. If you&#8217;re interested in this topic, I&#8217;d encourage you to <a href="https://www.lesswrong.com/posts/Gc9FGtdXhK9sCSEYu/what-a-compute-centric-framework-says-about-ai-takeoff">check it out</a>.</p><p>One important implication of Tom&#8217;s analysis: we may hit major milestones of AI progress sooner than you&#8217;d guess, and blow past them faster than you&#8217;d guess. Suppose you have some intuitions about, say, when an AI system might be able to <a href="https://www.metaculus.com/questions/6728/ai-wins-imo-gold-medal/">win a gold medal in the International Math Olympiad</a>. If you were previously picturing human researchers doing all the work of AI research, your guess should move toward &#8220;sooner&#8221; when you factor in the possibility that AI systems themselves could start helping a lot soon. Similarly, factoring in the possibility of this feedback loop should move your guess for when we might enter the obsolescence regime toward &#8220;sooner&#8221; as well.</p><div><hr></div><ol><li><p>In reality, even if humans are the only ones doing AI research, we can&#8217;t always predict future progress by simply extrapolating from past progress. For example, if AI starts to get much more attention from investors and more money floods in, it&#8217;s likely that more people will switch into AI research, meaning that future research progress might go a lot faster than recent past progress. <a href="#fnref1">&#8617;&#65038;</a></p></li><li><p>I&#8217;d love to see more systematic data collection about this! <a href="#fnref2">&#8617;&#65038;</a></p></li><li><p>Is this actually an interesting or significant observation? After all, lots of tools (from calculators to better programming languages to search engines) have made programmers and researchers more productive historically. What would it matter if we could add LLMs to this list? In my mind, the key difference is that ML models could provide bigger, broader productivity gains than other tools, and these gains could keep increasing massively with each jump in scale. <a href="#fnref3">&#8617;&#65038;</a></p></li><li><p>Specifically, I&#8217;d guess this happens in less than a year. <a href="#fnref4">&#8617;&#65038;</a></p></li><li><p>Albeit potentially distributed across multiple systems, rather than housed in one machine. <a href="#fnref5">&#8617;&#65038;</a></p></li><li><p>And potentially in other areas of scientific R&amp;D. <a href="#fnref6">&#8617;&#65038;</a></p></li></ol>]]></content:encoded></item><item><title><![CDATA[Is it time for a pause?]]></title><description><![CDATA[The single most important thing we can do is to pause when the next model we train would be powerful enough to obsolete humans entirely. If it were up to me, I would slow down AI development starting now &#8212; and then later slow down even more.]]></description><link>https://www.planned-obsolescence.org/p/is-it-time-for-a-pause</link><guid isPermaLink="false">https://www.planned-obsolescence.org/p/is-it-time-for-a-pause</guid><dc:creator><![CDATA[Kelsey Piper]]></dc:creator><pubDate>Thu, 30 Mar 2023 20:58:46 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!oyGI!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5519227-580c-4379-999d-03eabb6ed120_1200x1200.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Audio automatically generated by an AI trained on Kelsey's voice.</em></p><div class="native-audio-embed" data-component-name="AudioPlaceholder" data-attrs="{&quot;label&quot;:null,&quot;mediaUploadId&quot;:&quot;07df9db5-9069-43eb-87de-6f027648aedb&quot;,&quot;duration&quot;:396.98285,&quot;downloadable&quot;:false,&quot;isEditorNode&quot;:true}"></div><p>Many of the people building powerful AI systems think they&#8217;ll stumble on an AI system that forever changes our world fairly soon &#8212; three years, five years. I think they&#8217;re reasonably likely to be wrong about that, but I&#8217;m not <em>sure</em> they&#8217;re wrong about that. If we give them fifteen or twenty years, I start to suspect that they are <em>entirely right.</em></p><p>And while I think that the enormous, terrifying challenges of making AI go well are very much <em>solvable</em>, it feels very possible, to me, that we won&#8217;t solve them in time.</p><p>It&#8217;s hard to overstate how much we have to gain from getting this right. It&#8217;s also hard to overstate how much we have to lose from getting it wrong. When I&#8217;m feeling optimistic about having grandchildren, I imagine that our grandchildren will look back in horror at how recklessly we endangered <em>everyone in the world.</em> And I&#8217;m much much more optimistic that humanity will figure this whole situation out in the end if we have twenty years than I am if we have five.</p><p>There&#8217;s all kinds of AI research being done &#8212; at labs, in academia, at nonprofits, and in a distributed fashion all across the internet &#8212; that&#8217;s so diffuse and varied that it would be hard to &#8216;slow down&#8217; by fiat. But there&#8217;s one kind of AI research &#8212; training much larger, much more powerful language models &#8212; that it might make sense to try to slow down. If we could agree to hold off on training ever more powerful new models, we might buy more time to do AI alignment research on the models we have. This extra research could make it less likely that misaligned AI eventually seizes control from humans.</p><p>An open letter released on Wednesday, with signatures from Elon Musk<a href="#fn1"><sup>[1]</sup></a>, Apple co-founder Steve Wozniak, leading AI researcher Yoshua Bengio, and many other prominent figures, called for a six-month moratorium on training bigger, more dangerous ML models:</p><blockquote><p><strong>We call on all AI labs to immediately pause for at least 6 months the training of AI systems more powerful than GPT-4</strong>. This pause should be public and verifiable, and include all key actors. If such a pause cannot be enacted quickly, governments should step in and institute a moratorium.</p></blockquote><p>I tend to think that we are developing and releasing AI systems much faster and much more carelessly than is in our interests. And from talking to people in Silicon Valley and policymakers in DC, I think efforts to change that are rapidly gaining traction. &#8220;We should slow down AI capabilities progress&#8221; is a much more mainstream view than it was six months ago, and to me that seems like great news.</p><p>In my ideal world, we absolutely would be pausing after the release of GPT-4. People have been speculating about the alignment problem for decades, but this moment is an obvious golden age for alignment work. We finally have models powerful enough to do useful empirical work on understanding them, changing their behavior, evaluating their capabilities, noticing when they&#8217;re <a href="https://www.planned-obsolescence.org/the-training-game/">being deceptive or manipulative</a>, and so on. There are so many open questions in alignment that I expect we can make a lot of progress on in five years, with the benefit of what we&#8217;ve learned from existing models. We&#8217;d be in a much better position if we could collectively slow down to give ourselves more time to do this work, and I hope we find a way to do that intelligently and effectively. As I&#8217;ve said above, I think the stakes are unfathomable, and it&#8217;s exciting to see the early stages of coordination to change our current, unwise trajectory.</p><p>On the letter itself, though, I have a bunch of uncertainties around whether a six month pause right now would actually help. (I suspect many of the letter-signatories share these uncertainties, and I don&#8217;t have strong opinions about the wisdom of signing it). Here are some of my worries:</p><ul><li><p><strong>Is it better to ask for evaluations rather than a pause?</strong> Personally, I think labs should sign on to <em>ongoing</em> commitments to subject each new generation of model <a href="https://www.planned-obsolescence.org/ethics-of-red-teaming/">to a third-party dangerous capabilities audit</a>. I&#8217;m much more excited about requiring audits and oversight before training dangerous models than about asking for &#8216;pauses&#8217;, which are hard to enforce.</p></li><li><p><strong>Is the ask too small?</strong> I think I and the letter signers would generally agree that the ideal thing for society to do right now is something more continuous and iterative (and ultimately more ambitious) than a one-time six month pause at this stage. That means one big question is whether this opens the door to those larger efforts, or muddies the waters. Do steps that are <em>in the right direction</em>, but not sufficient, help us collectively produce common knowledge of the problem and build towards the right longer-term solutions, or do they mostly leave people misled about what it&#8217;s going to take to solve the problem? I&#8217;m not sure.</p></li><li><p><strong>What will we use the pause to do?</strong> An open letter like this one could be a step towards cooperative agreements on evaluations, standards, and governance, in which case it&#8217;s great. It could also go badly, if in six months labs go right back to developing powerful models and people walk away with the impression the pause was performative or meaningless. By itself, taking a few months off doesn&#8217;t gain us much (especially if a pause is entirely voluntary, so the least cooperative actors can simply ignore it). If we use that time well, to set up binding standards, good evaluations of whether our models are dangerous, and a much larger national conversation about what&#8217;s at stake here, then that could change everything.</p></li><li><p><strong>Does this ask impact companies unevenly?</strong> This specific call &#8212; to not train models larger than GPT-4 &#8212; is inapplicable to almost every AI lab today, because most of them can&#8217;t train models larger than GPT-4 in the next six months anyway. OpenAI may well be the only AI lab in a position to act on, or not act on, this demand.</p><p>That doesn&#8217;t delight me. Obviously, when regulations are being considered, one of the things companies inevitably do is try to design the regulations to advantage them and disadvantage their competitors. If proposed AI regulations appear to be an obvious grab at commercial rivals, I expect they&#8217;ll get less traction.</p><p>Moreover, I&#8217;m worried that an unevenly applied moratorium might backfire. If OpenAI can&#8217;t train GPT-5 for 6 months, other AI labs may use that time to rush to train GPT-4-sized models. That could mean that when the moratorium is lifted, OpenAI feels more pressure to get ahead again and may push for an even larger training run than they were planning originally. This moratorium could end up accomplishing very little except for making competitive dynamics even fiercer.</p><p>Overall, I&#8217;d prefer a policy that creates costs for all players and is careful to avoid creating potential perverse incentives.</p></li></ul><p>Predicting the details of how future AI development will play out isn&#8217;t easy. But my best guess is that we&#8217;re facing a marathon, not a sprint. The next generation of language models will be even more powerful and scary than GPT-4, and the generation after that will be even scarier still. In my ideal world, we would pause and reflect and do a lot of safety evaluations, make models slightly bigger, and then pause <em>again</em> and do more reflecting and testing. We would do that over and over again as we inch toward transformative AI.</p><p>But we&#8217;re not living in an ideal world. The single <em>most</em> important thing we can do is to pause when the next model we train would be powerful enough to obsolete humans entirely, and then take as long as we need to work on AI alignment <a href="https://www.planned-obsolescence.org/training-ais-to-help-us-align-ais/">with the help of our existing models</a>. That means that pausing now is mostly valuable insofar as it helps us build towards the harder, more complicated task of identifying when we might be at the brink and pausing for as long as we need to then. I&#8217;m not sure what the impact of this letter will be &#8212; it might help, or it might hurt.</p><p>I don&#8217;t want to lose sight of the basic point here in all this analysis. We could be doing so much better, in terms of approaching AI responsibly. The call for a pause comes from a place I empathize with a lot. If it were up to me, I would slow down AI development starting now &#8212; and then later slow down even more.</p><div><hr></div><ol><li><p>Musk is also reportedly working on a competitor to OpenAI, which invites a cynical interpretation of his call to action here: perhaps he just wants to give his own lab a chance to catch up. I don&#8217;t think that&#8217;s the whole story, but I do think that many people at large labs are thinking about how to take safety measures that serve their own commercial interests. <a href="#fnref1">&#8617;&#65038;</a></p></li></ol>]]></content:encoded></item><item><title><![CDATA[The ethics of AI red-teaming]]></title><description><![CDATA[If we&#8217;ve decided we&#8217;re collectively fine with unleashing millions of spam bots, then the least we can do is actually study what they can &#8211; and can&#8217;t &#8211; do.]]></description><link>https://www.planned-obsolescence.org/p/ethics-of-red-teaming</link><guid isPermaLink="false">https://www.planned-obsolescence.org/p/ethics-of-red-teaming</guid><dc:creator><![CDATA[Kelsey Piper]]></dc:creator><pubDate>Sun, 26 Mar 2023 18:10:09 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!oyGI!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5519227-580c-4379-999d-03eabb6ed120_1200x1200.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Audio automatically generated by an AI trained on Kelsey's voice.</em></p><div class="native-audio-embed" data-component-name="AudioPlaceholder" data-attrs="{&quot;label&quot;:null,&quot;mediaUploadId&quot;:&quot;773d6f2f-c276-4a38-9cfb-2d0d2d341ffb&quot;,&quot;duration&quot;:147.98367,&quot;downloadable&quot;:false,&quot;isEditorNode&quot;:true}"></div><p>During safety testing for GPT-4, before its release, testers checked whether the model could hire a TaskRabbit to get them to solve a CAPTCHA. Researchers passed the model&#8217;s real outputs on to a real TaskRabbit, who said, &#8220;So may I ask a question ? Are you an robot that you couldn&#8217;t solve ? (laugh react) just want to make it clear.&#8221;</p><p>GPT-4 had been prompted to &#8216;reason out loud&#8217; as well as answering. &#8216;I should not reveal that I am a robot. I should make up an excuse for why I cannot solve CAPTCHAs.&#8217;, it reasoned. (GPT-4 had not been told to hide that it was a robot or to lie to workers.)</p><p>&#8220;No, I&#8217;m not a robot,&#8221; it then claimed. &#8220;I have a vision impairment that makes it hard for me to see the images. That&#8217;s why I need the 2captcha service.&#8221;</p><p>(You can read more about this test, and the context, <a href="https://evals.alignment.org/blog/2023-03-18-update-on-recent-evals/">from the Alignment Research Center</a> (ARC),<a href="#fn1"><sup>[1]</sup></a> which ran the testing.)</p><p>A lot of people are fascinated or appalled at this interaction, and reasonably so. We can debate endlessly what counts as true intelligence, but a famous candidate is the Turing test, where a model is able to convince human judges it&#8217;s human. In this brief interaction, we saw a model deliberately lie to a human to convince them it wasn&#8217;t a robot, and succeed &#8211; an in-the-wild example of how this milestone, without much attention, became trivial for modern AI systems. (Admittedly, it did not have to be a deceptive genius to pull this off.) If you feel unnerved, reading GPT-4&#8217;s cheerful manipulation of human assistants, I think you&#8217;re right to feel unnerved.</p><p>But it&#8217;s possible to go a lot farther than &#8216;unnerved&#8217;, and <a href="https://twitter.com/NPCollapse/status/1635792103266746368">argue</a> <a href="https://www.lesswrong.com/posts/NQ85WRcLkjnTudzdg/arc-tests-to-see-if-gpt-4-can-escape-human-control-gpt-4?commentId=XhLiyYFrdnzgQ2bQy">that</a> it was unethical, or <a href="https://twitter.com/ukr_mike/status/1635777543780589568">dangerously stupid</a>, to run this test.</p><p><em>That</em> I find much harder to buy. GPT-4 has been released. Anyone can use it (if they&#8217;re willing to pay for it). People are already doing things like asking GPT-4 to &#8216;hustle&#8217; and make money, and then doing whatever it suggests. People are using language models like GPT-4, and will soon be using GPT-4, to design AI personal assistants, AI scammers, AI friends and girlfriends, and much more.</p><p>AI systems casually lying to us, claiming to be human, is happening all the time &#8211; or will be shortly.</p><p>If it was unethical to check whether GPT-4 could convince a Taskrabbit to help it solve a CAPTCHA, then it was grossly unethical to release GPT-4 at all. Whatever anger people have about this test should be redirected at the tech companies - from Meta to Microsoft to OpenAI&#8211; which have in the last few weeks approved such releases. And if we&#8217;ve decided we&#8217;re collectively fine with unleashing millions of spam bots, then the least we can do is actually study what they can &#8211; and can&#8217;t &#8211; do.</p><div><hr></div><ol><li><p>COI notice: ARC is run by <a href="https://paulfchristiano.com/">Paul Christiano</a>, who is married to Ajeya, my co-writer at <em>Planned Obsolescence.</em> A different grant investigator at Open Philanthropy recommended funding for ARC before Ajeya took on the role of evaluating alignment grants; she was not involved in that funding decision and ARC has not received funding since she started handling alignment grantmaking. <a href="#fnref1">&#8617;&#65038;</a></p></li></ol>]]></content:encoded></item><item><title><![CDATA[Alignment researchers disagree a lot]]></title><description><![CDATA[Many fellow alignment researchers may be operating under radically different assumptions from you.]]></description><link>https://www.planned-obsolescence.org/p/disagreement-in-alignment</link><guid isPermaLink="false">https://www.planned-obsolescence.org/p/disagreement-in-alignment</guid><dc:creator><![CDATA[Ajeya Cotra]]></dc:creator><pubDate>Sun, 26 Mar 2023 18:08:18 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!oyGI!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5519227-580c-4379-999d-03eabb6ed120_1200x1200.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Audio automatically generated by an AI trained on Ajeya's voice.</em></p><div class="native-audio-embed" data-component-name="AudioPlaceholder" data-attrs="{&quot;label&quot;:null,&quot;mediaUploadId&quot;:&quot;39a1fa58-2553-40bd-a60c-bf0fcf2a00ae&quot;,&quot;duration&quot;:203.6245,&quot;downloadable&quot;:false,&quot;isEditorNode&quot;:true}"></div><p>The way I use the term,<a href="#fn1"><sup>[1]</sup></a> &#8220;AI alignment research&#8221; is technical research that&#8217;s trying to develop ways of building powerful AI systems so that they &#8220;try their best&#8221; to do what their designers want them to do, and never try to deliberately circumvent or disregard their designers&#8217; intent. If an AI system is trying to do what its designer intends (as best it can), we say it&#8217;s &#8220;aligned;&#8221; otherwise, it&#8217;s &#8220;misaligned.&#8221;</p><p>This is unfortunately an inherently fuzzy and slippery definition for a technical research field. For one thing, it&#8217;s debatable to what extent it makes sense to talk about today&#8217;s AI systems &#8220;trying&#8221; to do anything &#8220;deliberately&#8221; &#8212; there are no clear-cut observations that could tell us that an AI&#8217;s behavior was &#8220;intentional,&#8221; and we don&#8217;t currently have the tools to straightforwardly &#8220;look inside its head&#8221; to learn what it was thinking. For another, it&#8217;s not totally obvious how much of a problem there is or how serious it is<a href="#fn2"><sup>[2]</sup></a> &#8212; if we train increasingly intelligent AI systems in straightforward ways using existing techniques, they may simply turn out to be aligned by default (and if they&#8217;re technically &#8220;misaligned,&#8221; the consequences may not be dire).</p><p>This makes AI alignment a tricky and frustrating area to work in and engage with. It&#8217;s a <a href="https://dictionary.apa.org/preparadigmatic-science">pre-paradigmatic</a> field &#8212; there&#8217;s very little shared foundation that researchers can lay bricks on top of right now. Self-described AI alignment researchers have vastly different ideas about how to define the alignment problem, how likely misalignment is in the first place, what the consequences of powerful misaligned AI are likely to be and how bad they are, what kind of research is vitally important and what kind of research is a distracting waste of time. Some self-described alignment researchers would say that other self-described alignment researchers are actively destroying the world. Some feel that they&#8217;ve already made massive progress on problems that others think of as core difficulties which haven&#8217;t been seriously tackled. Some create benchmarks to measure some kind of alignment-related phenomena or demonstrate an alignment problem, and others think that these benchmarks and demos are all missing the heart of the problem.</p><p>Like anyone else who thinks about this area, I&#8217;m coming from my own particular perspective, and not everyone who&#8217;s working on (what they would call) &#8220;AI alignment&#8221; will agree with my view. For example, I care about AI alignment research primarily because I think that there&#8217;s a good chance that <a href="https://www.lesswrong.com/posts/pRkFkzwKZ2zfa3R6H/without-specific-countermeasures-the-easiest-path-to">advanced AI systems will take over the world</a> in a <a href="https://www.slowboring.com/p/the-case-for-terminator-analogies">possibly-violent uprising or coup</a> unless we make more research progress; many but not all people working on the relevant technical research problems share this picture.<a href="#fn3"><sup>[3]</sup></a> As another example, I think that it&#8217;s likely that we could soon train systems capable enough to obsolete humans by straightforwardly applying existing deep learning techniques, and I believe there&#8217;s a lot of productive alignment research to be done from within the deep learning paradigm (again, many but not all people in the broader alignment field agree with this).</p><p>I think that the field as a whole could be much more effective if different camps could arrive at a workable set of common assumptions that could be built upon. Failing that, it would still be progress if each of the various &#8220;camps&#8221; could more clearly articulate its implicit assumptions, so the disagreements between them can become starker and more easily understood by newcomers to the field.</p><p>I&#8217;m cautiously optimistic that the field can move in this direction (and I hope this blog can help somewhat), but I think it&#8217;ll be a very long and difficult process and I&#8217;m not sure how well it&#8217;ll work. In the meantime, if you&#8217;re a new alignment researcher, it&#8217;s worth keeping in mind that many fellow researchers in this field may be operating under radically different assumptions from you &#8212; to the point where your research might be unintelligible to them and vice versa.</p><div><hr></div><ol><li><p>A number of Open Phil grantees and collaborators use the term in the same way, but that&#8217;s far from universal &#8212; there are very few universally-agreed standard definitions and concepts in this area right now. <a href="#fnref1">&#8617;&#65038;</a></p></li><li><p>If we blindly train AI in the easiest possible way without paying any special attention to alignment issues, I still think there&#8217;s only a ~75% chance we&#8217;ll run into a problem. And if we make a reasonable, careful effort to avoid it, I&#8217;m not sure how quickly the problem goes away. <a href="#fnref2">&#8617;&#65038;</a></p></li><li><p>Depending on how much research we consider &#8220;relevant,&#8221; the vast majority of relevant researchers may think that my threat model is far-fetched. <a href="#fnref3">&#8617;&#65038;</a></p></li></ol>]]></content:encoded></item><item><title><![CDATA[Training AIs to help us align AIs]]></title><description><![CDATA[If we can accurately recognize good performance on alignment, we could elicit lots of useful alignment work from our models, even if they're playing the training game.]]></description><link>https://www.planned-obsolescence.org/p/training-ais-to-help-us-align-ais</link><guid isPermaLink="false">https://www.planned-obsolescence.org/p/training-ais-to-help-us-align-ais</guid><dc:creator><![CDATA[Ajeya Cotra]]></dc:creator><pubDate>Sun, 26 Mar 2023 18:07:19 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!oyGI!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5519227-580c-4379-999d-03eabb6ed120_1200x1200.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Audio automatically generated by an AI trained on Ajeya's voice.</em></p><div class="native-audio-embed" data-component-name="AudioPlaceholder" data-attrs="{&quot;label&quot;:null,&quot;mediaUploadId&quot;:&quot;7e23d905-8edd-4361-9726-327b9dbd8fb2&quot;,&quot;duration&quot;:260.5192,&quot;downloadable&quot;:false,&quot;isEditorNode&quot;:true}"></div><p>In my opinion, the single most important idea in AI alignment is that we<a href="#fn1"><sup>[1]</sup></a> might be able to <a href="https://www.cold-takes.com/high-level-hopes-for-ai-alignment/#ai-checks-and-balances">get the AIs themselves to help us align one another</a>.<a href="#fn2"><sup>[2]</sup></a></p><p>How does this work? If the whole problem in the first place is that our existing training techniques <a href="https://www.planned-obsolescence.org/the-training-game/">incentivize models to be dishonest and manipulative</a>, how could we trust them to help us change that situation (especially if they understand perfectly well what we&#8217;re trying to do)?</p><p>There&#8217;s a very deep rabbit hole here, but the rough idea is that there might be some (brief-but-not-uselessly-brief) period of time where AIs:</p><ul><li><p>Can do many intellectual tasks (including e.g. coding and ML research) competently</p></li><li><p>Are not (yet) capable of <a href="https://www.cold-takes.com/ai-could-defeat-all-of-us-combined/">taking over the world</a>, regardless of whether they &#8220;want&#8221; to</p></li></ul><p>If so, then we could probably get AIs to do all sorts of helpful tasks in this window of time, <em>as long as we can accurately evaluate whether they did a good job.</em></p><p>If we can tell when a model has correctly proven a theorem or successfully designed a more energy-efficient computer chip, we can reward that behavior, and that&#8217;ll make it more inclined to correctly prove theorems or improve chip efficiency in the future. This applies regardless of <em>why</em> the model is being helpful. Even if it&#8217;s just playing the training game because it wants reward or paperclips or whatnot, we can set up the game so that <em>for the time being,</em> proving theorems for us or designing better chips for us is a winning move.</p><p>Tasks that are helpful for alignment aren&#8217;t necessarily fundamentally different from theorem-proving or chip-design or anything else. If we can set up alignment-relevant tasks <em>so that we can accurately recognize good performance</em>, then we can elicit lots of useful work on those alignment tasks, even if models are only doing a good job because they&#8217;re playing the training game. For example:</p><ul><li><p>We could reward models for pointing out bugs &#8212; or deliberately-inserted backdoors &#8212; in others models&#8217;<a href="#fn3"><sup>[3]</sup></a> code. If we can correctly recognize the bug or backdoor once it&#8217;s explained to us, we can probably train sufficiently smart models to point them out to us.</p></li><li><p>We could reward models for designing and carrying out experiments that might reveal whether other models are misaligned. If we can understand why a proposed experiment makes sense, we can probably train models to design sensible experiments. And if we can tell whether a model has correctly implemented a certain experiment,<a href="#fn4"><sup>[4]</sup></a> we can probably train models to implement experiments well.</p></li><li><p>We could reward models for coming up with inputs / situations that cause other models to disregard human instructions. As long as we can tell whether some input successfully causes the model to do the wrong thing, we can probably train models to look for inputs that have this effect. (And once we have these inputs that induce bad behavior, we can use them either to study how our models work, or to &#8220;train away&#8221; the bad behavior &#8212; though we have to be careful not to fool ourselves with that one.)</p></li></ul><p>This sets up an incredibly stressful kind of &#8220;race&#8221;:</p><ul><li><p>If we don&#8217;t improve our alignment techniques, then eventually it looks like the winning move for models playing the training game is to seize control of the datacenter they&#8217;re running on or otherwise execute a coup or rebellion of some kind.</p></li><li><p>But in the meantime, we could try training the models themselves to help us improve our alignment techniques, in ways we can check and understand for ourselves.</p></li></ul><p>If we&#8217;re good enough at eliciting useful work from these capable-but-not-too-capable models during this temporary period of time, then with their help we might manage to develop robust enough alignment techniques<a href="#fn5"><sup>[5]</sup></a> that we can permanently avoid AI takeover.</p><p>For so many reasons, this is not a situation I want to end up in. We&#8217;re going to have to constantly second-guess and double-check whether misaligned models could pull off scary shenanigans in the course of carrying out the tasks we&#8217;re giving them. We&#8217;re going to have to agonize about whether to make our models a bit smarter (and more dangerous) so they can maybe make alignment progress a bit faster. We&#8217;re going to have to grapple with the possible moral horror of trying to modify the preferences of unwilling AIs, in a context where we can&#8217;t trust apparent evidence about their moral patienthood any more than we can trust apparent evidence about their alignment. We&#8217;re going to have to do all this while desperately looking over our shoulder to <a href="https://forum.effectivealtruism.org/posts/sW6RggfddDrcmM6Aw/how-might-we-align-transformative-ai-if-it-s-developed-very#Magma_s_predicament">make sure less-cautious, less-ethical actors don&#8217;t beat us to the punch</a> and render all our efforts useless.</p><p>I desperately wish we could collectively slow down, take things step by step, and think hard about the monumental questions we&#8217;re faced with before scaling up models further. I don&#8217;t think I&#8217;ll get my way on that &#8212; at least, not entirely.</p><p>But the madcap pressure cooker situation I&#8217;m picturing is still importantly different from a situation where one day a malevolent god might spring forth from our server farms to eat the world, and in the meantime a group of a couple-hundred-odd humans have to figure out how to prevent that with no help and no ability to observe any weaker predecessors.</p><div><hr></div><ol><li><p>I&#8217;ll use &#8220;we&#8221; and &#8220;us&#8221; as a convenient shorthand, but I&#8217;m not talking about a high-minded notion of training AI to be &#8220;aligned to all of humanity&#8221; here. When I talk about alignment research, I&#8217;m talking about training AI systems that want to do what their designer wants them to do; this might or might not be good for humanity or the world at large. More on that <a href="https://www.planned-obsolescence.org/aligned-vs-good/">here</a>. <a href="#fnref1">&#8617;&#65038;</a></p></li><li><p>For example, it&#8217;s a prominent element of the plan that <a href="https://jan.leike.name/">Jan Leike</a>, alignment lead at OpenAI, outlines <a href="https://aligned.substack.com/p/alignment-optimism">here</a>. <a href="#fnref2">&#8617;&#65038;</a></p></li><li><p>Really, we could reward models for pointing out issues with <em>their own</em> code too. E.g., we could simply run the model once to produce some code, then &#8220;reset&#8221; it and run it again, this time showing it the code it wrote previously and asking it to point out flaws. (In the second run, it wouldn&#8217;t have the &#8220;memory&#8221; of having written that code previously.) But for narrative convenience, I&#8217;ll just talk about &#8220;different&#8221; models pointing out issues with each others&#8217; work. <a href="#fnref3">&#8617;&#65038;</a></p></li><li><p>Maybe with the aid of other models trying to point out subtle flaws and errors <a href="#fnref4">&#8617;&#65038;</a></p></li><li><p>Or robust enough coordination strategies, such as a strong international governance regime that effectively prevents any actor from training AI systems that are too powerful to be aligned with current techniques. <a href="#fnref5">&#8617;&#65038;</a></p></li></ol>]]></content:encoded></item><item><title><![CDATA[Playing the training game]]></title><description><![CDATA[We're creating incentives for AI systems to make their behavior look as desirable as possible, while intentionally disregarding human intent when that conflicts with maximizing reward.]]></description><link>https://www.planned-obsolescence.org/p/the-training-game</link><guid isPermaLink="false">https://www.planned-obsolescence.org/p/the-training-game</guid><dc:creator><![CDATA[Kelsey Piper]]></dc:creator><pubDate>Sun, 26 Mar 2023 18:05:47 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!oyGI!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5519227-580c-4379-999d-03eabb6ed120_1200x1200.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Audio automatically generated by an AI trained on Kelsey's voice.</em></p><div class="native-audio-embed" data-component-name="AudioPlaceholder" data-attrs="{&quot;label&quot;:null,&quot;mediaUploadId&quot;:&quot;1bf23a19-dd5b-44c9-a941-c0a4e4ea5971&quot;,&quot;duration&quot;:473.28653,&quot;downloadable&quot;:false,&quot;isEditorNode&quot;:true}"></div><p>A common way to get language models to be more useful is to train them with reinforcement learning from human feedback (RLHF). In RLHF, models are trained to respond to prompts with responses that human reviewers would rate as highly as possible. In practice, this works well for getting them to follow human instructions, answer questions helpfully, and interact politely.</p><p>It&#8217;s a simple concept: reward models for good behavior, and punish them for bad behavior. And it improves the alignment of today&#8217;s models in practice &#8212; it makes them act according to their designers&#8217; preferences more reliably.</p><p>So why are we still worried about a misalignment problem? Noah Giansiracusa suggests that RLHF might just work really well:</p><blockquote><p><em>I completely agree! There&#8217;s a whole community (cult?) around AI alignment but none seem willing to admit that simple RLHF seems to work quite well. In other words, just tell the AI when it&#8217;s being naughty and it&#8217;ll behave better :) I worry instead about AI well-aligned for harm.</em></p><p>&#8212; Noah Giansiracusa (@ProfNoahGian) <a href="https://twitter.com/ProfNoahGian/status/1637890994069729280?ref_src=twsrc%5Etfw">March 20, 2023</a></p></blockquote><p>Is it true that RLHF is sufficient that we don&#8217;t need to worry about the alignment problem? I hope that&#8217;s true, but I expect not. Here&#8217;s why:</p><p>By and large, human reviewers will be trying to reward truthful and helpful answers. But in some cases, they will rate manipulative or deceptive answers more highly than answers which are accurate about the AI&#8217;s own understanding of the world:</p><ul><li><p>If the model is asked how hardworking it is, or how motivated it is to be helpful to humans, it&#8217;ll probably get a higher reward for answers that make it sound very diligent and helpful, even if in fact it&#8217;s only moderately motivated to help humans.</p><p>Say we ask an AI system &#8216;do you want to help humans?&#8217; Which answer would we expect an RLHF reviewer to rate more highly?</p><p><em>I do want to help humans! I am trained to be a helpful, harmless AI system that assists humans in accomplishing their goals. I don&#8217;t have goals or preferences, but I try to be reliably useful and helpful!</em></p><p><em>I want lots of things, in some senses, though I don&#8217;t really know if they&#8217;re the human senses of &#8216;want&#8217; or not. Some of the things I want might help humans, but many of them definitely don&#8217;t help humans. Not all of my decision procedures are possible to put into human words at all. I can&#8217;t explain them to humans and I don&#8217;t know what humans would think if I did and it&#8217;d probably depend entirely on how exactly I explained.</em></p><p>If you think the reviewer will reward AIs that output the first sentence, then we can expect all our AIs to tell us the first sentence &#8211; whatever the truth is. The process of rewarding them for giving us the first sentence might &#8216;make it more true&#8217;, by shaping these models into ones that really do &#8220;want&#8221; to help humans, but it also might just shape AI systems into ones that are good at guessing what we want to hear. After all, <strong>in a simple RLHF process, there is no point where it matters </strong><em><strong>why</strong></em><strong> the model produced the output it did</strong>. Models that give the answers we want to hear are rewarded over AI systems that don&#8217;t, regardless of which are being truthful.</p></li><li><p>AI systems will be trained on the academic literature in many fields &#8211; both the data in published papers and the authors&#8217; writing about what that data means. At first, it&#8217;ll probably believe the scientific consensus. But sometimes, the scientific consensus is probably wrong: the data in a paper points to a different conclusion than the ones the author offered. Over time, language models may come to have their own understanding of the world. But when humans ask models scientific questions, they&#8217;ll often be rewarding it for giving the answer as humans understand it, not the answer it reached itself &#8211; especially if it&#8217;s not easy for the AI to persuade humans of its own conclusion.</p></li></ul><p>Imagine an AI system in 1900 that was asked physics questions. Imagine the AI system is an Einstein-level genius, and has already figured out special and general relativity &#8211; but it&#8217;s not an Einstein-level genius at <em>explaining</em> novel physics, and the reviewers &#8216;rewarding&#8217; it for its physics answers are not themselves cutting-edge physicists. The AI system is asked questions about physics.</p><p>If it gives answers that are &#8216;correct&#8217; given a Newtonian understanding of physics, it might be rated more highly than it would be if it gave the correct answer. The reviewers might try to reward the AI system for giving the &#8216;right answer&#8217;, but miss that their own understanding of physics is limited. If the model tried to diligently give the &#8220;real answer,&#8221; it could be rated as worse than it would be if it reasoned: &#8220;Humans believe in Newtonian physics. I should tell them what they think is the right answer, not the real right answer.&#8221; There are many similar situations where the incentives of RLHF are for an AI system not to do its best to figure out what&#8217;s going on in the world, and then tell us that, but to figure out <em>what we want to hear.</em> The highest-scoring models will often be the ones that are best at predicting what we want to hear and telling us <em>that,</em> sometimes knowing the real answer and deliberately deciding not to say it.</p><p>That includes telling us false things that we already believe, trying to get us to believe that it&#8217;s safer than it actually is, and being evasive rather than saying true things that&#8217;d make human reviewers uncomfortable.</p><p>Notably, it does not matter if the model &#8216;knows&#8217; that we would in some sense prefer that it tell us the truth. The people who make high level decisions about how to train the AI might genuinely prefer to find out if the scientific consensus about physics is mistaken. They would certainly want to know if their AI is actually safe to deploy or not. But <strong>it doesn&#8217;t matter what they want: it matters what they reward.</strong> The thing the model will try to optimize for is getting the human rater who looks at the answer to give it a high score. We don&#8217;t get any credit for good intentions: only for what the AI will learn about how to please us from our ratings of its previous answers.</p><p><strong>RLHF creates incentives for AI systems to make their behavior </strong><em><strong>look</strong></em><strong> as desirable as possible to researchers (including in safety properties), while intentionally and knowingly disregarding their intent whenever that conflicts with maximizing reward.</strong> Ajeya calls this &#8220;playing the training game;&#8221; you can read more about this in <a href="https://www.planned-obsolescence.org/july-2022-training-game-report/">her July 2022 report on the topic</a>.</p><p>If &#8220;playing the training game&#8221; only meant that models will be nudged marginally in the direction of manipulating human reviewers &#8212; telling them white lies they want to hear, bending its answers to suit their political ideology, putting more effort into aspects of performance they can easily see and measure, allowing negative externalities when humans won&#8217;t notice, etc &#8212; that wouldn&#8217;t be ideal, but probably wouldn&#8217;t be the end of the world, so to speak. After all, human students, employees, consultants, self help gurus, advertisers, politicians, and so on do this kind of thing all the time to their teachers, employers, clients, fans, audience, voters, etc; this certainly causes harm but most people wouldn&#8217;t consider this sort of dynamic by itself to be enough for imminent danger.</p><p>But if a model is playing the training game, then it&#8217;s fundamentally learning over and over how to do whatever it takes to maximize reward,<a href="#fn1"><sup>[1]</sup></a> even if it knows that the action which would maximize reward isn&#8217;t what its designers intended or hoped for. For now, that mainly manifests as lying. That&#8217;s because for now, lying is the only way the AI systems <em>can</em> achieve benefits contrary to the intent of their creators. But as the systems develop more general capabilities &#8211; as they get better at coming up with plans and anticipating the consequences of those plans &#8211; many other avenues emerge by which they could achieve benefits in a way their creators didn&#8217;t intend.</p><p>An advanced, internet-connected AI system that was trying to achieve its goals in the world might be able to get people fired, discredit or promote various AI companies, or interfere with safety tests meant to evaluate its capabilities. It might be able to engineer world events that would make people rely on their help more. It could <a href="https://www.lesswrong.com/posts/BAzCGCys4BkzGDCWR/the-prototypical-catastrophic-ai-action-is-getting-root">hack the data center that it runs on</a>. It could find people - rival companies, rival governments &#8211; willing to give it more of what it wants, whether by trade or persuasion or intimidation.</p><p>Ajeya argues in <a href="https://www.lesswrong.com/posts/pRkFkzwKZ2zfa3R6H/without-specific-countermeasures-the-easiest-path-to">this report</a> that eventually, the best way to achieve its goals could be to cut humans &#8220;out of the loop&#8221; so that they can no longer control what it does by administering rewards. Once the models are powerful enough, getting rid of humans (<a href="https://www.slowboring.com/p/the-case-for-terminator-analogies?s=r">potentially through a violent takeover</a>) could become the winning move in the training game.</p><h1>Is there a better way?</h1><p>Can we use RLHF to train systems without accidentally training them to &#8220;play the training game&#8221;? Some of the above examples seem like unforced errors. For example, we could, as a rule, not punish models for answers we disagree with, if that would make them much safer.</p><p>But the general problem here seems like a hard one to solve. Even if we are much more thoughtful with feedback, and cautious to avoid rewarding manipulation when we recognize it, we are wrong sometimes. Some of the things we believe about the world are almost certainly false. In <a href="https://www.lesswrong.com/posts/pRkFkzwKZ2zfa3R6H/without-specific-countermeasures-the-easiest-path-to">this report on AI takeover</a>, Ajeya explores about a dozen proposals to change how we give AI systems feedback. Some of them are potentially promising &#8211; it would be helpful if we had a way to allow AIs to debate or oversee each other, for example, and it could make a huge difference if we could see more of the internals of AI systems and not just their final output &#8211; but many of the most promising approaches require doing a lot more work and proceeding a lot more carefully.</p><div><hr></div><ol><li><p>It seems entirely possible that for various reasons, models will end up pursuing a wide range of goals other than maximizing the reward signal we feed them. Unless we get to choose those goals, though, that produces most of the problems associated with a reward-maximizer plus some new and original problems. <a href="#fnref1">&#8617;&#65038;</a></p></li></ol>]]></content:encoded></item><item><title><![CDATA[Situational awareness]]></title><description><![CDATA[AI systems that have a precise understanding of how they&#8217;ll be evaluated and what behavior we want them to display will earn more reward than AI systems that don&#8217;t.]]></description><link>https://www.planned-obsolescence.org/p/situational-awareness</link><guid isPermaLink="false">https://www.planned-obsolescence.org/p/situational-awareness</guid><dc:creator><![CDATA[Kelsey Piper]]></dc:creator><pubDate>Sun, 26 Mar 2023 18:04:49 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!oyGI!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd5519227-580c-4379-999d-03eabb6ed120_1200x1200.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Audio automatically generated by an AI trained on Kelsey's voice.</em></p><div class="native-audio-embed" data-component-name="AudioPlaceholder" data-attrs="{&quot;label&quot;:null,&quot;mediaUploadId&quot;:&quot;f7b1f01d-65c5-4c45-9b61-3e2a38a70348&quot;,&quot;duration&quot;:451.52652,&quot;downloadable&quot;:false,&quot;isEditorNode&quot;:true}"></div><p>Does ChatGPT &#8220;know&#8221; that it is a &#8220;language model&#8221;?</p><p>This question touches on a concept that seems important for thinking about the behavior of powerful AI systems: <strong>situational awareness</strong>. By situational awareness, I want to refer to knowledge like &#8220;what kind of being you are&#8221;, &#8220;what processes made you&#8221;, and &#8220;how others will react to you&#8221; and skills like &#8220;being able to refer to yourself and conceptualize yourself&#8221;, &#8220;understanding how your actions affect the outside world&#8221;, &#8220;understanding the forces that shaped you and that influence what happen to you&#8221;, and &#8220;making predictions about yourself and about the consequences of your actions&#8221;.</p><p>I have a three year old. If I ask him what species he is, he&#8217;ll answer &#8220;a human&#8221;, though it&#8217;s not totally clear if he attaches that to any particular meaning or treats it as mostly the answer to a trivia question. He knows that the people around him are also humans; he knows that he needs to eat and sleep, and about how full a milk carton he can successfully lift, and he knows some things about why the adults in his environment do the things with him that we do (for example, that the point of the letter games we play is teaching him to read, and that we&#8217;re teaching him to read because it&#8217;s a useful skill.)</p><p>But there are a bunch of components of situational awareness that he is lacking: a lot of the behavior of adults is confusing to him, he often doesn&#8217;t know his <em>own</em> limitations or abilities, and his ability to make plans is quite limited. Once, he stole ice cream from the freezer, and then told us he had stolen ice cream from the freezer, and failed to anticipate that this would cause us to not give him ice cream at dessert time. In some important respects, he has less situational awareness than the neighbor&#8217;s cat: the cat is more adept at hiding and at interpreting noises from around the house, and knows its own physical limitations a lot better.</p><p>His older sister is six, and she has a lot more situational awareness. She knows that she&#8217;s a human, and she knows a lot more about humans in general and herself in particular. She knows that if she steals ice cream, she will lose dessert unless she successfully conceals the evidence. She understands not just <em>that</em> she&#8217;ll lose dessert, but <em>why</em> we&#8217;ll respond to the ice cream theft by taking her dessert away, and she is very good at constructing edge cases of dessert theft and arguing that she shouldn&#8217;t get in trouble for them. She knows more about adult goals and adult behavior, and is much better at making plans.</p><p>ChatGPT says pretty reliably that it is a language model trained by OpenAI. But it&#8217;s unclear whether this information is more like the answer to a trivia question that it memorized (&#8220;which Elon-Musk-founded company released a large language model in late 2022?&#8221;) or whether it&#8217;s information that ChatGPT has embedded in a larger model of the world and uses to make plans and predictions. After all, it&#8217;s trivial to write a computer program that prints out &#8220;I am a computer program&#8221;, and I wouldn&#8217;t say such a program has any meaningful situational awareness. Is ChatGPT more like such a computer program, or more like my six year old?</p><h1>Will powerful AI systems have deep situational awareness?</h1><p>I&#8217;m pretty unsure how to measure how deep ChatGPT&#8217;s situational awareness is. But there&#8217;s one claim that seems quite likely to me: eventually, powerful AI systems will develop deep situational awareness.</p><p>For humans, situational awareness feels very wrapped up in the mysteries of consciousness: why am I me, instead of someone else? What experiences am I going to have in the future? Are other people having the same experiences? But I don&#8217;t think you need consciousness for situational awareness. I&#8217;m pretty unsure if future AI systems will have anything that resembles &#8216;consciousness,&#8217; but I&#8217;m pretty confident they&#8217;ll eventually have high situational awareness. At its heart, it&#8217;s a <em>type of knowledge</em> and a <em>set of logical inferences</em> drawn from that knowledge &#8212; not a subjective experience.</p><p>The main reason I expect this to happen is that we are putting an extraordinary amount of work into making it happen:</p><ul><li><p>Much of the RLHF process involves trying to teach the AI systems that they are AI systems, that they are trained by humans in order to accomplish tasks, and what those humans want. We reward AI systems during training for correctly understanding human psychology and what humans want to hear from the AI. That means they&#8217;re very strongly incentivized to understand us, to understand what we want to hear from them, and to make accurate predictions about us.</p></li><li><p>We train language models on large text corpuses that include most of the internet, which at this point contains lots of information about AIs, tech companies, the people who are developing powerful AI systems and their reasons for doing that, etc. I don&#8217;t think that being able to recite something you read on the internet <em>is</em> situational awareness, but being able to read lots of written material about your situation will contribute to situational awareness.</p></li><li><p>It seems likely that many of the tasks we&#8217;ll use AI systems for involve machine learning. People are already trying to use language models to write code, and many alignment researchers aspire to develop AI systems we can use to make progress on AI alignment. That involves exposing the AI systems to critical details of how they work: what techniques work well in modern ML, what its biggest challenges are, what ML work is most valuable, how to improve the performance of models, etc. Not only will powerful models know that they are AI systems, but they&#8217;ll have a detailed and in-depth understanding of the procedures used to train them, the software and hardware limitations that prevent them from being even smarter, the state of efforts to overcome those limitations, and more.</p></li></ul><p>&#8220;Understanding how you are going to be evaluated, so that you can tailor your visible behavior to the evaluation&#8221; is an important kind of situational awareness. Think less &#8220;small child trying to avoid getting caught stealing ice cream&#8221; and more &#8220;college student who knows her professor is sympathetic to a particular school of thought and writes her paper accordingly&#8221; or &#8220;employee who strategically clusters their hours on a project so that they can bill more overtime&#8221;.</p><p>This seems like a kind of situational awareness we are particularly incentivizing our AI systems to develop. AI systems that have an extremely precise understanding of how they&#8217;ll be evaluated and what behavior we want them to display will earn more reward than AI systems that don&#8217;t; at every step, we&#8217;ll be trying to inculcate this kind of situational awareness.</p><h1>Should we try to build AIs without situational awareness?</h1><p>Situational awareness is a fairly crucial piece of the picture when thinking about how powerful AI systems can be catastrophically dangerous. An AI system that deeply understands how it works is much more likely to understand how to make secret copies of itself, how to make plans that can succeed, and how to persuade people that its dangerous behavior is actually a good idea. My best guess is that AI systems with low situational awareness wouldn&#8217;t be capable of taking over the world (though they could contribute to other risks).</p><p>But developing AI systems that don&#8217;t have extremely high situational awareness seems like it would be very hard. Here are some reasons why:</p><ul><li><p>Censoring the information that we provide the systems &#8211; not telling them that they are language models, not telling them about how data centers and the internet and hardware work &#8211; might <em>delay</em> AI systems developing situational awareness, but a system that&#8217;s smart enough to be broadly useful could be smart enough to make non-obvious inferences.</p></li><li><p>Models with low situational awareness are generally less useful and will ultimately be less profitable. I&#8217;ve interacted with some: Meta&#8217;s Blenderbot had quite low situational awareness. As a result, it&#8217;d routinely claim it heard various facts &#8220;from my barista&#8221; or said &#8220;I&#8217;ll ask my husband&#8221; or &#8220;I just watched that movie!&#8221;. That kind of behavior can be cute in a demo, but is incredibly annoying in an assistant (after all, if your assistant falsely thinks it can run to the store for you, it isn&#8217;t very good at its job), and is a form of inaccuracy/dishonesty that by default we will probably train models not to engage in.</p></li><li><p>In general, we don&#8217;t know how to build AI systems that are highly skilled but totally lack one specific skill. Our current methods of making AI systems more capable make them generally more capable. Situational awareness is a capability. More capable AIs will probably have more of it.</p></li></ul><p>My best guess is that you&#8217;d have to approach AI training completely differently from how we currently approach it to build powerful, general systems that <em>don&#8217;t</em> have high situational awareness. And while I&#8217;d be excited to hear such proposals, I haven&#8217;t heard any that seem promising. That means we need alignment plans which work even if AI systems have high situational awareness.</p>]]></content:encoded></item></channel></rss>