<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[No Set Gauge]]></title><description><![CDATA[Civilization. AI. Book reviews. Startups. Liberty. Effective altruism. Progress studies. Radical futures.]]></description><link>https://www.nosetgauge.com</link><image><url>https://substackcdn.com/image/fetch/$s_!HdSA!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F816cef70-50a0-4954-a8ce-8f712e1248e8_460x460.png</url><title>No Set Gauge</title><link>https://www.nosetgauge.com</link></image><generator>Substack</generator><lastBuildDate>Fri, 10 Apr 2026 19:32:40 GMT</lastBuildDate><atom:link href="https://www.nosetgauge.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[L Rudolf L]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[nosetgauge@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[nosetgauge@substack.com]]></itunes:email><itunes:name><![CDATA[Rudolf Laine]]></itunes:name></itunes:owner><itunes:author><![CDATA[Rudolf Laine]]></itunes:author><googleplay:owner><![CDATA[nosetgauge@substack.com]]></googleplay:owner><googleplay:email><![CDATA[nosetgauge@substack.com]]></googleplay:email><googleplay:author><![CDATA[Rudolf Laine]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[The Technology of Liberalism]]></title><description><![CDATA[In every technological revolution, we face a choice: build for freedom or watch as others build for control.]]></description><link>https://www.nosetgauge.com/p/the-technology-of-liberalism</link><guid isPermaLink="false">https://www.nosetgauge.com/p/the-technology-of-liberalism</guid><dc:creator><![CDATA[Rudolf Laine]]></dc:creator><pubDate>Mon, 05 Jan 2026 10:06:36 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!aL4m!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a589c2e-6f80-46f2-b07d-a22f7613b6c8_1280x851.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!aL4m!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a589c2e-6f80-46f2-b07d-a22f7613b6c8_1280x851.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!aL4m!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a589c2e-6f80-46f2-b07d-a22f7613b6c8_1280x851.png 424w, https://substackcdn.com/image/fetch/$s_!aL4m!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a589c2e-6f80-46f2-b07d-a22f7613b6c8_1280x851.png 848w, https://substackcdn.com/image/fetch/$s_!aL4m!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a589c2e-6f80-46f2-b07d-a22f7613b6c8_1280x851.png 1272w, https://substackcdn.com/image/fetch/$s_!aL4m!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a589c2e-6f80-46f2-b07d-a22f7613b6c8_1280x851.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!aL4m!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a589c2e-6f80-46f2-b07d-a22f7613b6c8_1280x851.png" width="1280" height="851" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9a589c2e-6f80-46f2-b07d-a22f7613b6c8_1280x851.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:851,&quot;width&quot;:1280,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Pasted image 20260104020430.png&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Pasted image 20260104020430.png" title="Pasted image 20260104020430.png" srcset="https://substackcdn.com/image/fetch/$s_!aL4m!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a589c2e-6f80-46f2-b07d-a22f7613b6c8_1280x851.png 424w, https://substackcdn.com/image/fetch/$s_!aL4m!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a589c2e-6f80-46f2-b07d-a22f7613b6c8_1280x851.png 848w, https://substackcdn.com/image/fetch/$s_!aL4m!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a589c2e-6f80-46f2-b07d-a22f7613b6c8_1280x851.png 1272w, https://substackcdn.com/image/fetch/$s_!aL4m!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9a589c2e-6f80-46f2-b07d-a22f7613b6c8_1280x851.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Dido Building Carthage</em>, J. M. W. Turner</figcaption></figure></div><blockquote><p>In every technological revolution, we face a choice: build for freedom or watch as others build for control.<br>-<a href="https://www.spectator.co.uk/article/how-can-we-develop-ai-that-helps-rather-than-harms-people/">Brendan McCord</a></p></blockquote><p>There are two moral frames that explain modern moral advances: utilitarianism and liberalism.</p><p>Utilitarianism says: &#8220;the greatest good for the greatest number&#8221;. It says we should maximize welfare, whatever that takes.</p><p>Liberalism says: &#8220;to each their own sphere of freedom&#8221;. We should grant everyone some boundary that others can&#8217;t violate, for example over their physical bodies and their property, and then let whatever happen as long as those boundaries aren&#8217;t violated.</p><p>(Before the philosophers show up and cancel me: &#8220;utilitarianism&#8221; and &#8220;liberalism&#8221; here are labels for two views, that correspond somewhat but not exactly to the normal uses of the phrases. The particular axis I talk about comes from my reading of Joe Carlsmith, who I&#8217;ll quote at length later in this post. Also, for the Americans: <a href="https://en.wikipedia.org/wiki/Liberalism">&#8220;liberalism&#8221;</a> is not a synonym for &#8220;left-wing&#8221;.)</p><p>Some of the great moral advances of modernity are:</p><ul><li><p>women&#8217;s rights</p></li><li><p>abolition of slavery</p></li><li><p>equality before the law for all citizens of a country</p></li><li><p>war as an aberration to be avoided, rather than an <a href="https://en.wikipedia.org/wiki/Carl_von_Clausewitz">&#8220;extension of politics by other means&#8221;</a></p></li><li><p>gay rights</p></li><li><p>the increasing salience of animal welfare</p></li></ul><p>Every one of these can be seen as either a fundamentally utilitarian or fundamentally liberal project. For example, viewing wars of aggression as bad is both good for welfare (fewer people dying horrible deaths way too early), and a straightforward consequence of believing that <a href="https://quoteinvestigator.com/2011/10/15/liberty-fist-nose/">&#8220;your right to swing your fist ends where my nose begins&#8221;</a> (either on a personal or national level). Women&#8217;s rights are both good for welfare through everything from directly reducing gender-based violence to knock-on effects on economic growth, as well as a straightforward application of the idea that people should be equal in the eyes of the law and free to make their own choices and set their own boundaries.</p><p>Utilitarian aims are a consequence of liberal aims. If you give everyone their boundaries, their protection from violence, and their God and/or government -granted property rights, then Adam Smith and positive-sum trade and various other deities step in and <a href="https://en.wikipedia.org/wiki/Fundamental_theorems_of_welfare_economics">ensure</a> that welfare is maximized (at least if you&#8217;re willing to grant a list of assumptions). More intuitively and generally, everyone naturally tries to achieve what they want, and when you ban boundary-violating events, everyone will be forced to achieve what they want through win-win cooperation rather than, say, violence.</p><p>Liberal aims are a consequence of utilitarian aims. If you want to maximize utility, then Hayek and the political scientists and the moral arc of humanity over the last two centuries all show up and demand that you let people choose for themselves, and have their own boundaries and rights. Isn&#8217;t it more efficient when people can make decisions based on the local information they have? How many giga-utils have come from women being able to pursue whatever job they want, or gay people being free from persecution? Hasn&#8217;t the wealth of developed countries, and all the welfare that derives from that, come from institutions that ensure freedom and equality before the law, enforce a stable set of rules, and avoid arbitrary despotism?</p><p>Much discussion of utilitarianism focuses on things like trolley problems that force you to pick between welfare losses and boundary violations. Unless you happen to live near a trolley track frequented by rogue experimental philosophers, however, you&#8217;ll notice that <em>such circumstances basically never happen</em> (if you are philosophically astute, you&#8217;ll also be aware that as a real-life human you <a href="https://www.lesswrong.com/posts/K9ZaZXDnL3SEmYZqB/ends-don-t-justify-means-among-humans">shouldn&#8217;t violate the boundaries </a><em><a href="https://www.lesswrong.com/posts/K9ZaZXDnL3SEmYZqB/ends-don-t-justify-means-among-humans">even if it seems like a good idea</a></em>).</p><p>However, as with everything else, the <a href="https://slatestarcodex.com/2018/09/25/the-tails-coming-apart-as-metaphor-for-life/">tails come apart</a> when things get extreme enough. The concept of good highlighted by welfare and that highlighted by freedom do diverge in the limit. But will we get pushed onto extreme philosophical terrain in the real world, though? I believe we don&#8217;t need to worry about a sudden influx of rogue experimental philosophers, or a trolley track construction spree. But we might need to worry about the god-like AI that <a href="https://darioamodei.com/machines-of-loving-grace">every</a> <a href="https://ia.samaltman.com/">major</a> <a href="https://x.com/polynoamial/status/1825024830225391818?lang=en">AGI lab</a> <a href="https://www.reuters.com/technology/deepseek-founder-liang-wenfeng-puts-focus-chinese-innovation-2025-01-28/">leader</a>, the <a href="https://manifold.markets/ManifoldAI/agi-when-resolves-to-the-year-in-wh-d5c5ad8e4708">prediction</a> <a href="https://www.metaculus.com/questions/5121/date-of-artificial-general-intelligence/">markets</a>, and the <a href="https://gwern.net/scaling-hypothesis">anonymous internet folk who keep turning out annoyingly right about everything</a> warn us might be coming in a few years.</p><p>The Effective Altruists, to their strong credit, have taken the intersection of AI and moral philosophy seriously for years. However, their main approach has been to perfect alignment&#8212;the ability to reliably point an AI at a goal&#8212;while also in tandem figuring out the correct moral philosophy to then align the AI to, such that we point it at the right thing. Not surprisingly, there are some unresolved debates around the second bit. (Also, to a first approximation, the realities of incentives mean that the values that get encoded into the AI will be what AI lab leadership wants modulo whatever changes are forced on them by customers or governments, rather than what the academics cook up as ideal.)</p><p>In this post, I do not set out to solve moral philosophy. In fact, I don&#8217;t think &#8220;solve moral philosophy&#8221; is a thing you can (or should) do. Instead, my concern is that <strong>near-future technology by default, and AI in particular, may differentially accelerate utilitarian over liberal goals</strong>. My hope is that differential technological development&#8212;speeding up some technologies over others&#8212;can fix this imbalance, and help continue moral progress towards a world that is good by many lights, rather than just one.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.nosetgauge.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.nosetgauge.com/subscribe?"><span>Subscribe now</span></a></p><h2>The real Clippy is the friends we made along the way</h2><p>Joe Carlsmith has an excellent essay series on &#8220;<a href="https://joecarlsmith.com/2024/01/02/otherness-and-control-in-the-age-of-agi">Otherness and control in the age of AGI</a>&#8220;. It&#8217;s philosophy about vibes, but done well, and thoroughly human to the core.</p><p>He starts off with a reminder: <a href="https://www.lesswrong.com/posts/uMQ3cqWDPHhjtiesc/agi-ruin-a-list-of-lethalities">Eliezer Yudkowsky thinks we are all going to die</a>. We don&#8217;t know how to make the AIs precisely care about human values, the AI capabilities will shoot up without a corresponding increase in their caring about our values, and they will devour the world. A common thought experiment is the paperclip-maximizer AI, sometimes named &#8220;Clippy&#8221; after the <a href="https://en.wikipedia.org/wiki/Office_Assistant">much-parodied Microsoft Office feature</a>. The point of the thought experiment is that optimizing hard for anything (e.g. paperclips) entails taking over the universe and filling it with that thing, and in the process destroy everything else. In his essay <a href="https://joecarlsmith.com/2024/01/16/being-nicer-than-clippy">&#8220;Being nicer than Clippy&#8221;</a>, Carlsmith writes:</p><blockquote><p>Indeed, in many respects, Yudkowsky&#8217;s AI nightmare is precisely the nightmare of all-boundaries-eroded. The nano-bots eat through every wall, and soon, everywhere, a single pattern prevails. After all: what makes a boundary bind? In Yudkowsky&#8217;s world (is he wrong?), only two things: hard power, and ethics. But the AIs will get all the hard power, and have none of the ethics. So no walls will stand in their way.</p></blockquote><p>The reason the AIs will be such paperclipping maximizers is because <a href="https://nosetgauge.substack.com/i/153676731/the-philosophy">Yudkowsky&#8217;s philosophy</a> emphasizes some math that points towards: &#8220;If you aren&#8217;t running around in circles or stepping on your own feet or wantonly giving up things you say you want, we can see your behavior as corresponding to [expected utility maximization]&#8221; (<a href="https://www.lesswrong.com/posts/RQpNHSiWaXTvDxt6R/coherent-decisions-imply-consistent-utilities">source</a>). Based on this, the Yudkowskian school thinks that the only way out is very precisely encoding the <em>right</em> set of terminal values into the AI. This, in turn, is harder than teaching the AIs to be smart overall <a href="https://www.lesswrong.com/posts/uMQ3cqWDPHhjtiesc/agi-ruin-a-list-of-lethalities#Section_B_2___Central_difficulties_of_outer_and_inner_alignment_">because</a>: &#8220;There&#8217;s something like a single answer, or a single bucket of answers, for questions like &#8216;What&#8217;s the environment really like?&#8217; [... and so w]hen you have a wrong belief, reality hits back at your wrong prediction [...] In contrast, when it comes to a choice of utility function, there are unbounded degrees of freedom&#8221;.</p><p>So: anything smart enough is a rocket aimed at some totalizing arrangement of all matter in the universe. You need to aim the rocket at exactly the right target because otherwise it flies off into some unrelated voids of space and you lose everything.</p><p>To be clear, to Yudkowsky this is a <em>factual prediction</em> about the behavior of things that are smart enough, rather than a <em>normative statement</em> that the correct morality is utilitarian in this way. Big&#8212;and very inconvenient&#8212;if true.</p><p>Now at this point you might remember that even some <em>humans</em> disagree with each other about what utopia should look like. Of course, a standard take in AI safety is that the main technical problem we face is pointing the AIs reliably at anything at all, and therefore arguing over the &#8220;monkey politics&#8221; (as Carlsmith puts it) is pointless politicking that detracts from humanity&#8217;s technical challenge.</p><p>However, in <a href="https://joecarlsmith.com/2024/01/11/an-even-deeper-atheism">another essay</a> in his sequence, Carlsmith points out that <em>any</em> size of value difference between two agents could lead to one being a paperclipper by the lights of the other:</p><blockquote><p>What force staves off extremal Goodhart in the human-human case, but not in the AI-human one? For example: what prevents the classical utilitarians from splitting, on reflection, into tons of slightly-different variants, each of whom use a slightly-different conception of optimal pleasure (hedonium-1, hedonium-2, etc)? And wouldn&#8217;t they, then, be paperclippers to each other, what with their slightly-mutated conceptions of perfect happiness?</p></blockquote><p>This is not just true between agents, but also between you at a certain time and you at a slightly later time:</p><blockquote><p>[..] if you read a book, or watch a documentary, or fall in love, or get some kind of indigestion [...] your heart is <em>never exactly the same ever again</em>, and not because of Reason [...so] then the only possible vector of non-trivial long-term value in this bleak and godless lightcone has been snuffed out?! Wait, OK, I have a plan: this precise <em>person-moment</em> needs to become dictator. It&#8217;s rough, but it&#8217;s the only way. Do you have the nano-bots ready? Oh wait, too late. (OK, how about now? Dammit: doom again.)</p></blockquote><p>Now, this isn&#8217;t Yudkowsky&#8217;s view. But why not? Remember: in the Yudkowskian frame, for any agent to &#8220;want&#8221; something coherently, it must have a totalizing vision of how to structure the entire universe. And who is to say that even small differences in values don&#8217;t lead to different ideal universes, especially under over-optimization to the limit of those values? After all, as Yudkowsky repeatedly emphasizes, <a href="https://www.lesswrong.com/posts/GNnHHmm8EzePmKzPk/value-is-fragile">value is fragile</a>. However, the overall framework, Carlsmith argues, creates an overall &#8220;momentum towards deeming more and more agents (and agent-moments) paperclippers&#8221;. It makes it very natural that we&#8217;ve just got to know exactly what is valuable well enough to write it down in a form without contradictions that we can optimize ruthlessly without breaking it. And we should do this <em>as soon as possible</em>, because our values might change, or because another entity might get to do it first, and chances are they&#8217;d be a paperclipper with respect to us. Time to take over the universe, for the greater good!</p><p>What&#8217;s the alternative? <a href="https://joecarlsmith.com/2024/01/16/being-nicer-than-clippy">In the original Carlsmith essay</a>, he writes:</p><blockquote><p>Liberalism does not ask that agents sharing a civilization be &#8220;aligned&#8221; with each other in the sense at stake in &#8220;optimizing for the same utility function.&#8221; Rather, it asks something more minimal, and more compatible with disagreement and diversity &#8211; namely, that these agents respect certain sorts of boundaries; that they agree to transact on certain sorts of cooperative and mutually-beneficial terms; that they give each other certain kinds of space, freedom, and dignity. Or as a crude and distorting summary: that they be a certain kind of nice. Obviously, not all agents are up for this &#8211; and if they try to mess it up, then liberalism will, indeed, need hard power to defend itself. But if we seek a vision of a future that avoids Yudkowsky&#8217;s nightmare, I think the sort of pluralism and tolerance at the core of liberalism will often be more a promising guide than &#8220;getting the utility function that steers the future right.&#8221;</p></blockquote><p>So, then: we have Yudkowsky, claiming that, as much as <a href="https://x.com/esyudkowsky/status/819678205900509184?lang=bn">we humans benefit</a> from centering rights and virtues in our idea of the good, the AIs will be superhumanly intelligent, and it is the nature of rationality that beings bend more and more towards a specific type of &#8220;coherence&#8221;&#8212;and hence utilitarian consequentialism&#8212;as they get smarter.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!HtxN!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a84455f-d374-4972-bbb7-b0dbe51401f5_1216x346.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!HtxN!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a84455f-d374-4972-bbb7-b0dbe51401f5_1216x346.png 424w, https://substackcdn.com/image/fetch/$s_!HtxN!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a84455f-d374-4972-bbb7-b0dbe51401f5_1216x346.png 848w, https://substackcdn.com/image/fetch/$s_!HtxN!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a84455f-d374-4972-bbb7-b0dbe51401f5_1216x346.png 1272w, https://substackcdn.com/image/fetch/$s_!HtxN!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a84455f-d374-4972-bbb7-b0dbe51401f5_1216x346.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!HtxN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a84455f-d374-4972-bbb7-b0dbe51401f5_1216x346.png" width="647" height="184.09703947368422" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2a84455f-d374-4972-bbb7-b0dbe51401f5_1216x346.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:346,&quot;width&quot;:1216,&quot;resizeWidth&quot;:647,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Screenshot 2025-03-31 at 10.44.26.png&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Screenshot 2025-03-31 at 10.44.26.png" title="Screenshot 2025-03-31 at 10.44.26.png" srcset="https://substackcdn.com/image/fetch/$s_!HtxN!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a84455f-d374-4972-bbb7-b0dbe51401f5_1216x346.png 424w, https://substackcdn.com/image/fetch/$s_!HtxN!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a84455f-d374-4972-bbb7-b0dbe51401f5_1216x346.png 848w, https://substackcdn.com/image/fetch/$s_!HtxN!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a84455f-d374-4972-bbb7-b0dbe51401f5_1216x346.png 1272w, https://substackcdn.com/image/fetch/$s_!HtxN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2a84455f-d374-4972-bbb7-b0dbe51401f5_1216x346.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Yudkowsky is <em>not</em> an arch-utilitarian when it comes to how humans should act. But in his worldview, reality, alas, prefers the consequentialists. (<a href="https://x.com/ESYudkowsky/status/819678205900509184">Source</a>)</figcaption></figure></div><p>So, this then is the irony: Yudkowsky is a deep believer in humanism and liberalism. But his philosophical framework makes him think anything sufficiently smart becomes a totalizing power-grabber. The Yudkowskian paperclipper nightmare explicitly comes from a lack of liberalism, in the rules-and-boundaries sense. Yudkowsky&#8217;s ideal solution would be to figure out how to encode the non-utilitarian limits into the AI&#8212;&#8221;corrigibility&#8221;, for example, means that the AI doesn&#8217;t resist being corrected, and is something where <a href="https://intelligence.org/files/Corrigibility.pdf">Yudkowsky and others have spent lots of time on trying to make it make sense within the expected-utility-maximising paradigm</a>. But that seems deeply technically difficult within the expected-utility-maximizer paradigm.</p><p>These philosophical vibes, plus the unnaturalness of non-consequentialist principles given a framework that makes non-consequentialism deeply unnatural, has given the AI safety space an ethos where the only admissible solution is getting better at fiddling with the values that are going into an AI, and then fiddling with those values. The AI is a rocket shooting towards expected utility maximization, and if it hits even slightly off, it&#8217;s all over&#8212;so let&#8217;s aim the rocket very precisely! Let&#8217;s make sure we can steer it! Let&#8217;s debate whether the rocket should land <em>here</em> or <em>there</em>. But if the threat we care about is the all-consuming boundary violations ... maybe we shouldn&#8217;t build a <em>rocket</em>? Maybe we should build something less totalizing. Maybe we shouldn&#8217;t try to reconfigure all matter in the universe&#8212;&#8221;tile the universe&#8221;&#8212;into some perfectly optimal end-state. Ideally, we&#8217;d just boost the capabilities of humans who engage with each other under a liberal, rule-of-law world order.</p><p>The point Yudkowsky fundamentally makes, after all, is that you shouldn&#8217;t <a href="https://x.com/dwarkesh_sp/status/1644347895116881921">&#8220;grab the poison banana&#8221;</a>; that <em>hardcore utility-maximization is very inconvenient if you want anything or anyone to survive</em>&#8212;or if you think that there&#8217;s the slightest chance you&#8217;re wrong about your utility function.</p><p>There&#8217;s also a more cynical way to read the focus on values over constraints in AI discourse. If you build the machine-god, you will gain enormous power, and can reshape the world by your own lights. If you think you might be in the winning faction in the AI race, you want to downplay the role of constraints and boundaries, and instead draw the debate towards which exact set of values should be prioritized. This tilts your faction&#8217;s thinking a bit towards yours, while increasing the odds that your faction is not constrained. If you&#8217;re an AI lab CEO, how much better is it for you that people are debating what the AI should be set to maximizing, or whether it should be woke or unwoke, rather than how the AI should be constrained or how we should make sure it doesn&#8217;t <a href="https://intelligence-curse.ai/shaping/">break the social contract that keeps governments aligned</a> or <a href="https://intelligence-curse.ai/capital/">permanently entrench the ruling class in a way that stunts social &amp; moral progress</a>?</p><p>Of course, I believe that most who talk about AI values are voicing genuine moral concerns rather than playing power games, especially since the discourse really does lean towards the utilitarian over the liberal side, and since the people who (currently) most worry about extreme AI scenarios often have a consequentialist bent to their moral philosophy.</p><p>But even in the genuineness there&#8217;s a threat. Voicing support for simple, totalizing moral opinions is a status signal among the tech elite, perhaps derived from the bullet-biting literal-mindedness that is genuinely so useful in STEM. The <a href="https://en.wikipedia.org/wiki/Effective_accelerationism">e/accs</a> loudly proclaim their willingness to obey the <a href="https://beff.substack.com/p/notes-on-eacc-principles-and-tenets">&#8220;thermodynamic will of the universe&#8221;</a>. This demonstrates their loyalty to the cause (meme?) and their belief in techno-capital (and, conveniently, willingness to politically align with the rich VCs whose backing they seek). What about their transparent happiness for the techno-capital singularity to rip through the world and potentially create a <a href="https://nosetgauge.substack.com/p/a-disneyland-without-children">world without anyone left to enjoy it</a>? It&#8217;s worth it, because thermodynamics! Meanwhile, among the AI safety crowd, it&#8217;s a status signal to know all the arguments about coherence and consequentialism inside out, and to be willing to bite the bullets that these imply. But sometimes, that there is no short argument against something means not that it is correct, but rather that the counter-argument is subtle, or <a href="https://arxiv.org/abs/2408.16984">outside the paradigm</a>.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.nosetgauge.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.nosetgauge.com/subscribe?"><span>Subscribe now</span></a></p><h2>The limits and lights of liberalism</h2><p>I&#8217;m worried of leaving an impression that liberalism, boundaries, and niceness are an obvious and complete moral system, and utilitarianism is this weird thing that leads to paperclips. But like utilitarianism, liberalism breaks when taken to an extreme.</p><p>First, it&#8217;s hard to pin down what it means. Carlsmith <a href="https://joecarlsmith.com/2024/01/16/being-nicer-than-clippy">writes</a>:</p><blockquote><p>[A]n even-remotely-sophisticated ethics of &#8220;boundaries&#8221; requires engaging with a ton of extremely gnarly and ambiguous stuff. When, exactly, does something become &#8220;someone&#8217;s&#8221;? Do wild animals, for example, have rights to their &#8220;territory&#8221;? See <a href="https://plato.stanford.edu/entries/property/">all of the philosophy of property</a> for just a start on the problems. And aspirations to be &#8220;nice&#8221; to agents-with-different-values clearly need ways of balancing the preferences of different agents of this kind &#8211; e.g., maybe you don&#8217;t steal Clippy&#8217;s resources to make fun-onium; but can you tax the rich paperclippers to give resources to the multitudes of poor staple-maximizers? Indeed, remind me your story about the ethics of taxation in general?</p></blockquote><p>Second, there are things we may want to ban or at least discourage, even if doing so means interventionism. Today we (mostly) think that a genocide warrants an international response, that you <em>should</em> catch the person jumping off the building to kill themself, and that various forms of paternalism are warranted, especially towards children. In the future, presumably there are things people could do that are bad enough by our lights that we think we can&#8217;t morally tolerate that in our universe.</p><p>Third, the moral implications of boundaries and rights change with technology. For example, today cryptocurrency might be helpful for people in unstable low-income countries, but tomorrow scheming AI systems might take advantage of the existence of a decentralized, uncontrollable payments rail to bribe, blackmail, and manipulate in ways we wish we hadn&#8217;t given them the affordances to do. Today we might be completely fine giving everyone huge amounts of compute that they have full control over, but tomorrow it might be possible to simulate digital minds and suddenly we want to make sure no one is torturing simulated people. Historically, too, it&#8217;s worth noting that while almost every place and time would&#8217;ve benefited from <em>more</em> liberalism and rule-of-law, sometimes it&#8217;s important that there are carefully-guarded mechanisms to tinker with those foundations. What set British institutions apart in the 18th and 19th centuries was not just that they were more laissez-faire and gave better rights than other countries of the age, but also that they were able to adapt in times of need. In the US, FDR&#8217;s reforms required heavy-handed intervention and the forcing through of major changes in government, but were (eventually) successful at dragging the country out of recession. Locking in inviolable rights is a type of lock-in too.</p><p>Fourthly, liberalism has a built-in tension between a decentralizing impulse and a centralizing impulse. The decentralization comes from the fact that the whole point is that everyone is free. The centralization comes from things like needing a government with a monopoly on force to protect the boundaries of people within it&#8212;otherwise it&#8217;s not a world of freedom, but of might-makes-right. A lot of our civilization&#8217;s progress in governance mechanisms comes from improving the Pareto frontier of tradeoffs we can make between centralization and decentralization (I heard this point from <a href="https://x.com/RichardMCNgo">Richard Ngo</a>, in a talk that is unpublished as far as I know):</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ImtI!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F420f6bae-b8a3-442b-9ba6-2db679902507_1048x856.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ImtI!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F420f6bae-b8a3-442b-9ba6-2db679902507_1048x856.png 424w, https://substackcdn.com/image/fetch/$s_!ImtI!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F420f6bae-b8a3-442b-9ba6-2db679902507_1048x856.png 848w, https://substackcdn.com/image/fetch/$s_!ImtI!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F420f6bae-b8a3-442b-9ba6-2db679902507_1048x856.png 1272w, https://substackcdn.com/image/fetch/$s_!ImtI!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F420f6bae-b8a3-442b-9ba6-2db679902507_1048x856.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ImtI!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F420f6bae-b8a3-442b-9ba6-2db679902507_1048x856.png" width="492" height="401.8625954198473" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/420f6bae-b8a3-442b-9ba6-2db679902507_1048x856.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:856,&quot;width&quot;:1048,&quot;resizeWidth&quot;:492,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Pasted image 20250712122729.png&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Pasted image 20250712122729.png" title="Pasted image 20250712122729.png" srcset="https://substackcdn.com/image/fetch/$s_!ImtI!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F420f6bae-b8a3-442b-9ba6-2db679902507_1048x856.png 424w, https://substackcdn.com/image/fetch/$s_!ImtI!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F420f6bae-b8a3-442b-9ba6-2db679902507_1048x856.png 848w, https://substackcdn.com/image/fetch/$s_!ImtI!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F420f6bae-b8a3-442b-9ba6-2db679902507_1048x856.png 1272w, https://substackcdn.com/image/fetch/$s_!ImtI!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F420f6bae-b8a3-442b-9ba6-2db679902507_1048x856.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Institution design (whether through politics or technology) is often about making the tradeoff between autocracy and anarchy risks less harsh.</figcaption></figure></div><p>For example, democracy allows for centralizing more power in a government (reducing anarchy risk) without losing people&#8217;s ability to occasionally fire the government (reducing autocracy risk). Zero-knowledge proofs and memoryless AI allow verification of treaties or contracts (important for coordinating to solve risks) without any need to hand over information to some authority (which would centralize).</p><p>However, the tension remains. We want inviolable rights for individuals, but we also need some way to enforce those rights. Unless technology can make that fully decentralized (or <a href="https://slatestarcodex.com/2015/03/18/book-review-the-machinery-of-freedom/">David Friedman</a> is <a href="https://www.econlib.org/archives/2008/07/david_friedmans.html">right</a>), that means we need at least some locus of central control, but this locus of course becomes the target of every seeker of power and rent. And remember, we want everything to be adaptable if circumstances change&#8212;but only for <em>good</em> reasons.</p><p>On a more fundamental level, everything being permitted is not necessarily a stable state. If, in free competition, there exist patterns of behavior that are better at gaining power than others, then you have to accept that those patterns will have a disproportionate share of future value. This leads to two concerns. First, as pointed out by Carlsmith in his talk <a href="https://joecarlsmith.substack.com/p/video-and-transcript-of-talk-on-can?utm_source=post-email-title&amp;publication_id=1022275&amp;post_id=168493115&amp;utm_campaign=email-post-title&amp;isFreemail=true&amp;r=rmmte&amp;triedRedirect=true&amp;utm_medium=email">&#8220;Can goodness compete?&#8221;</a>, maximizing for goodness and maximizing for power are different concerns, so the optimum for one is likely not the optimum for the other. So in fully free competition, in this very simplified model, we lose on goodness. Secondly, if everything being permitted leads to competition in which some gain power until they are in a position to violate others&#8217; boundaries, then full liberalism would bring about its own demise. It&#8217;s not a stable state.</p><p>Fifth, there&#8217;s a neutrality to liberalism. We give everyone their freedoms, we make everyone equal before the law&#8212;but what do we actually do? Whatever you want, that&#8217;s the point! And what do you want? Well, that&#8217;s your problem&#8212;we liberals just came here to set up the infrastructure, like the plumber comes to install your pipes. It&#8217;s your job to decide what runs through them, and to use them well.</p><p>Now, utilitarianism suffers from this problem too. What is utility? It&#8217;s what you like/prefer. But what are those things? Presumably, you&#8217;re supposed to just look inside you. But for what? Do you really know? The most hardcore hedonic utilitarians will claim it&#8217;s obvious: pleasure minus pain. But this is of course overly reductive, unless either you are a very simple person (lucky you&#8212;but what about the rest of us?) or you take a very expansive view of what pleasure and pain are (but then you&#8217;re back to the problem of defining them in a sufficiently expansive, yet precise, way).</p><p>Putting aside the abstract moral philosophies: in yet another essay, Carlsmith <a href="https://joecarlsmith.com/2021/07/19/in-search-of-benevolence-or-what-should-you-get-clippy-for-christmas#iv-johnny-appleseeds-of-welfare">mentions</a> the <a href="https://en.wikipedia.org/wiki/Johnny_Appleseed">&#8220;Johnny Appleseeds&#8221;</a> of the world:</p><blockquote><p>Whatever your agricultural aspirations, he&#8217;s here to give you apple trees, and not because apple trees lead to objectively better farms. Rather, apple trees are just, as it were, his thing.</p></blockquote><p>There&#8217;s something charming about this image. At the end of the day you need something worth fighting for. Everyone who has actually inspired people has not preached just an abstract set of rules, but has also breathed life into an entire worldview and way of life. Pure liberalism does not tell you what the content of life should be, it only says that in the pursuit of that content you should not take away other people&#8217;s freedoms to pursue theirs. At times you need the abstraction-spinning philosophers and even-handed policemen and hair-splitting lawyers to come and argue for and enforce the boundaries. But they are infrastructure, not content. The farmer told by the philosopher to believe in liberalism, or shown by the police and the lawyers where their plot ends and their neighbor&#8217;s begins, does not gain the content that animates a life. You also need a Johnny Appleseed, whether externally or within you, who preaches something specific, something concrete: look at this apple, feel its texture, smell its scent&#8212;<em>this</em> is what you should grow.</p><p>Much as there&#8217;s a distinction between the pipes as infrastructure and the water in the pipes, there&#8217;s a distinction between the structure of our moral system (or, to use a computer science metaphor, its <a href="https://en.wikipedia.org/wiki/Type_theory">type</a>), and what we actually care about. What people really fight for <em>is</em> sometimes the abstract principle, but almost always there is something real and concrete too. But what exactly this is, and how different people find theirs, is subtle and complicated. The strength of liberalism is that it <em>doesn&#8217;t</em> claim an answer (remember: there is <a href="https://www.lesswrong.com/posts/yDfxTj9TKYsYiWH5o/the-virtue-of-narrowness">virtue to narrowness</a>). Whatever those real and concrete actually-animating things are, they seem hard to label and put in a box. So instead: don&#8217;t try to specify them, don&#8217;t try to dictate them top-down, just lay down the infrastructure thats let people discover and find them on their own. In a word: freedom.</p><h2>Weaving the rope</h2><p>So how should we think about moral progress? A popular approach is to pick one tenet, such as freedom or utility, and declare it to be the One True Value. Everything else is good to the extent that it reflects that value, and bad to the extent it violates it. You can come up with very good stories of this form, because <a href="https://www.lesswrong.com/posts/yA4gF5KrboK2m2Xu7/how-an-algorithm-feels-from-inside">words are slippery</a> and humans are very good at spinning persuasive stories&#8212;didn&#8217;t the Greeks get their start in philosophy with takes like &#8220;everything is water&#8221; or &#8220;everything is change&#8221;? If there are annoying edge cases&#8212;trolley problems for the utilitarians, <a href="https://slatestarcodex.com/2018/02/21/current-affairs-some-puzzles-for-libertarians-treated-as-writing-prompts-for-short-stories/">anti-libertarian thought experiments</a> for the liberalists&#8212;then you have to either <a href="https://scottaaronson.blog/?p=326">swallow the bullets or try to dodge them</a> as best as you can while. But all the while you gain the satisfaction that you&#8217;ve solved it, you&#8217;ve seen the light, the ambiguity is gone (or, at least in principle, could one day be gone if only we carry out some known process long enough)&#8212;because you know the One True Value.</p><p>But, as Isaiah Berlin <a href="https://www.cooperative-individualism.org/berlin-isaiah_positive-versus-negative-liberty-1958.htm">wrote</a>:</p><blockquote><p>One belief, more than any other, is responsible for the slaughter of individuals on the altars of the great historical ideals [...]. This is the belief that somewhere, in the past or in the future, in divine revelation or in the mind of an individual thinker, in the pronouncements of history or science, or in the simple heart of an uncorrupted good man, there is a final solution.</p></blockquote><p>An alternative image to the One True Value is making a rope. While it&#8217;s being plied, the threads of the rope are laid out on the table as the rope is being made, pointing in different directions, not yet plied into a single rope. The liberal idea of boundaries and rights is one thread, and it&#8217;ll go somewhere into the finished thing, as will the thread of utilitarian welfare, and others as well. We&#8217;ve knit parts of this rope already; witness thinkers from the Enlightenment philosophers onwards making lots of progress plying together freedom and welfare and fairness and other things, in ways that (say) medieval Europeans probably couldn&#8217;t have imagined. How do we continue making progress? Rather than declaring one thread the correct one for all time, I think the good path forwards looks more like continuing to weave together different threads and moral intuitions. I expect where we end up to have at least some depend on the path we take. I expect progress to be halting and uncertain; I expect sometimes we might need to unweave the last few turns. I expect eventually where we end up might look weird to us, much as modernity would look weird to medieval people. But I probably trust this process, at least if it&#8217;s protected from censorship, extremism, and non-human influences. I trust it more than I trust the results of a thought experiment, or any one-pager on the correct ideology. I certainly trust it far more than I trust the results of a committee of philosophers, or AI lab ethics board, or inhuman AI, to figure it out and then lock it in forever.</p><p>I hope you share the intuition above, and that the intuition alone is compelling. But can we support it? Is there a deeper reason <em>why</em> should we expect morality to work like that, and be hard to synthesize into exactly one shape?</p><p>The first reason to think this is that, empirically, human values just seem to be subtle and nuanced. Does the nature of the good <em>seem</em> simple to you? Could you design a utopia you felt comfortable locking in for all eternity? Hell is simple and cruel, while heaven is devious and subtle.</p><p>A second reason is that over-optimization is treacherous. This is the same instinct that leads Yudkowsky to think the totalizing AIs will rip apart all of us, and that leads Carlsmith to spend so many words poking at whether there&#8217;s something nicer and less-treacherous if only we&#8217;re more liberal, more <a href="https://joecarlsmith.com/2024/03/21/on-green">&#8220;green&#8221; (in his particular sense of the word)</a>, nicer. Paul Christiano, one of the most prominent thinkers in AI alignment and an inventor of RLHF, centers his <a href="https://www.alignmentforum.org/posts/HBxe6wdjxK239zajf/what-failure-looks-like">depiction of AI risk</a> on &#8220;getting what you measure&#8221; all too well. But why? <a href="https://en.wikipedia.org/wiki/Goodhart%27s_law">Goodhart&#8217;s law</a> is one idea: over-optimize any measure, and the measure stops being a good measure. Another way to put this is that even if two things are correlated across a normal range of values, they decorrelate&#8212;the &#8220;tails come apart&#8221;&#8212;at extreme values, as <a href="https://www.lesswrong.com/posts/dC7mP5nSwvpL65Qu5/why-the-tails-come-apart">LessWrong user Thrasymachus</a> (presumably a different person from the ancient philosopher) observed and <a href="https://slatestarcodex.com/2018/09/25/the-tails-coming-apart-as-metaphor-for-life/">Scott Alexander expanded on</a>. In the case of morality, for example:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!43Vm!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6805b15a-f694-4b2c-9a79-101937463e15_600x603.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!43Vm!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6805b15a-f694-4b2c-9a79-101937463e15_600x603.png 424w, https://substackcdn.com/image/fetch/$s_!43Vm!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6805b15a-f694-4b2c-9a79-101937463e15_600x603.png 848w, https://substackcdn.com/image/fetch/$s_!43Vm!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6805b15a-f694-4b2c-9a79-101937463e15_600x603.png 1272w, https://substackcdn.com/image/fetch/$s_!43Vm!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6805b15a-f694-4b2c-9a79-101937463e15_600x603.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!43Vm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6805b15a-f694-4b2c-9a79-101937463e15_600x603.png" width="600" height="603" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6805b15a-f694-4b2c-9a79-101937463e15_600x603.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:603,&quot;width&quot;:600,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Pasted image 20250712150933.png&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Pasted image 20250712150933.png" title="Pasted image 20250712150933.png" srcset="https://substackcdn.com/image/fetch/$s_!43Vm!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6805b15a-f694-4b2c-9a79-101937463e15_600x603.png 424w, https://substackcdn.com/image/fetch/$s_!43Vm!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6805b15a-f694-4b2c-9a79-101937463e15_600x603.png 848w, https://substackcdn.com/image/fetch/$s_!43Vm!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6805b15a-f694-4b2c-9a79-101937463e15_600x603.png 1272w, https://substackcdn.com/image/fetch/$s_!43Vm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6805b15a-f694-4b2c-9a79-101937463e15_600x603.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">(Image taken from Scott Alexander <a href="https://slatestarcodex.com/2018/09/25/the-tails-coming-apart-as-metaphor-for-life/">here</a>)</figcaption></figure></div><p>Within the range of normal events, moral theories align on what is good. When it comes to abnormal events&#8212;especially abnormal events derived from taking one moral theory&#8217;s view of what is good and cranking everything to eleven&#8212;the theories disagree. The Catholic Church and the hedonic utilitarians both want to help the poor, but the utilitarians are significantly less enthusiastic about various Catholic limits on sexuality, and the Catholics are significantly less enthusiastic about everything between here and Andromeda converting to an orgiastic mass of brain tissue.</p><p>So if our choice is about <em>which</em> ideal we let optimize the world, and if that optimization is strong enough, we run a real risk of causing unspeakable horrors by the lights of other ideals <em>regardless of which ideal we pick</em>:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!IKRs!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3bca4173-b35b-4645-90d2-8237b1f6f5a1_1178x1092.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!IKRs!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3bca4173-b35b-4645-90d2-8237b1f6f5a1_1178x1092.png 424w, https://substackcdn.com/image/fetch/$s_!IKRs!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3bca4173-b35b-4645-90d2-8237b1f6f5a1_1178x1092.png 848w, https://substackcdn.com/image/fetch/$s_!IKRs!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3bca4173-b35b-4645-90d2-8237b1f6f5a1_1178x1092.png 1272w, https://substackcdn.com/image/fetch/$s_!IKRs!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3bca4173-b35b-4645-90d2-8237b1f6f5a1_1178x1092.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!IKRs!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3bca4173-b35b-4645-90d2-8237b1f6f5a1_1178x1092.png" width="596" height="552.4889643463497" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3bca4173-b35b-4645-90d2-8237b1f6f5a1_1178x1092.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1092,&quot;width&quot;:1178,&quot;resizeWidth&quot;:596,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Screenshot 2026-01-04 at 01.45.47.png&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Screenshot 2026-01-04 at 01.45.47.png" title="Screenshot 2026-01-04 at 01.45.47.png" srcset="https://substackcdn.com/image/fetch/$s_!IKRs!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3bca4173-b35b-4645-90d2-8237b1f6f5a1_1178x1092.png 424w, https://substackcdn.com/image/fetch/$s_!IKRs!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3bca4173-b35b-4645-90d2-8237b1f6f5a1_1178x1092.png 848w, https://substackcdn.com/image/fetch/$s_!IKRs!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3bca4173-b35b-4645-90d2-8237b1f6f5a1_1178x1092.png 1272w, https://substackcdn.com/image/fetch/$s_!IKRs!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3bca4173-b35b-4645-90d2-8237b1f6f5a1_1178x1092.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">The red/blue path represents where we go if we steer entirely by moral theory A/B. </figcaption></figure></div><p>You might think what we want, then, is compromise. A middle path:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!wDqt!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb70797be-6565-4851-912e-333eab868907_1270x1222.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!wDqt!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb70797be-6565-4851-912e-333eab868907_1270x1222.png 424w, https://substackcdn.com/image/fetch/$s_!wDqt!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb70797be-6565-4851-912e-333eab868907_1270x1222.png 848w, https://substackcdn.com/image/fetch/$s_!wDqt!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb70797be-6565-4851-912e-333eab868907_1270x1222.png 1272w, https://substackcdn.com/image/fetch/$s_!wDqt!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb70797be-6565-4851-912e-333eab868907_1270x1222.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!wDqt!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb70797be-6565-4851-912e-333eab868907_1270x1222.png" width="532" height="511.8929133858268" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b70797be-6565-4851-912e-333eab868907_1270x1222.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1222,&quot;width&quot;:1270,&quot;resizeWidth&quot;:532,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Screenshot 2026-01-04 at 01.41.43.png&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Screenshot 2026-01-04 at 01.41.43.png" title="Screenshot 2026-01-04 at 01.41.43.png" srcset="https://substackcdn.com/image/fetch/$s_!wDqt!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb70797be-6565-4851-912e-333eab868907_1270x1222.png 424w, https://substackcdn.com/image/fetch/$s_!wDqt!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb70797be-6565-4851-912e-333eab868907_1270x1222.png 848w, https://substackcdn.com/image/fetch/$s_!wDqt!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb70797be-6565-4851-912e-333eab868907_1270x1222.png 1272w, https://substackcdn.com/image/fetch/$s_!wDqt!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb70797be-6565-4851-912e-333eab868907_1270x1222.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>And yes, I expect a compromise would be better. But I think the above is reductive and simplistic. Whatever Theory A &amp; Theory B are, they&#8217;re both likely static, contingent paradigms. Any good progress probably looks more like striking a balance than picking just one view of the good to maximize, but the most significant progress likely reframes the axes:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!n6_Z!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0aad62a0-1a17-4e57-87f7-4464ca6e8d4e_1612x1212.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!n6_Z!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0aad62a0-1a17-4e57-87f7-4464ca6e8d4e_1612x1212.png 424w, https://substackcdn.com/image/fetch/$s_!n6_Z!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0aad62a0-1a17-4e57-87f7-4464ca6e8d4e_1612x1212.png 848w, https://substackcdn.com/image/fetch/$s_!n6_Z!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0aad62a0-1a17-4e57-87f7-4464ca6e8d4e_1612x1212.png 1272w, https://substackcdn.com/image/fetch/$s_!n6_Z!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0aad62a0-1a17-4e57-87f7-4464ca6e8d4e_1612x1212.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!n6_Z!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0aad62a0-1a17-4e57-87f7-4464ca6e8d4e_1612x1212.png" width="1456" height="1095" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0aad62a0-1a17-4e57-87f7-4464ca6e8d4e_1612x1212.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1095,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Screenshot 2026-01-04 at 01.38.01.png&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Screenshot 2026-01-04 at 01.38.01.png" title="Screenshot 2026-01-04 at 01.38.01.png" srcset="https://substackcdn.com/image/fetch/$s_!n6_Z!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0aad62a0-1a17-4e57-87f7-4464ca6e8d4e_1612x1212.png 424w, https://substackcdn.com/image/fetch/$s_!n6_Z!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0aad62a0-1a17-4e57-87f7-4464ca6e8d4e_1612x1212.png 848w, https://substackcdn.com/image/fetch/$s_!n6_Z!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0aad62a0-1a17-4e57-87f7-4464ca6e8d4e_1612x1212.png 1272w, https://substackcdn.com/image/fetch/$s_!n6_Z!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0aad62a0-1a17-4e57-87f7-4464ca6e8d4e_1612x1212.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Different times and places argued about different axes: salvation versus worldliness, honor versus piety, or duties to family versus duties to the state. Liberalism versus utilitarianism, the framing of this post, is a modern axis. Or, to take another important modern axis: economic efficiency versus fairness of distribution. You need modern economics to understand what these things really are or why they might be at odds, it&#8217;s contingent on facts about the world including the distribution of economically-relevant talent in the population, and their importance rests on modern ethical ideas about the responsibilities of society towards the poor.</p><p>As mentioned before, to the extent that liberalism has any unique claim, it&#8217;s that it&#8217;s a good meta-level value system. Live and let live is an excellent rule, if you want to foster the continued weaving of the moral thread, including its non-liberal elements. It lets the process continue, and it lets the new ideas and new paradigms come to light and prove their worth. It seems far less likely to sweep away the goodness and humanity of the world than any totalizing optimization process aiming for a fixed vision of the good.</p><p>Isaiah Berlin, again:</p><blockquote><p>For every rationalist metaphysician, from Plato to the last disciples of Hegel or Marx, this abandonment of the notion of a final harmony in which all riddles are solved, all contradictions reconciled, is a piece of crude empiricism, abdication before brute facts, intolerable bankruptcy of reason before things as they are, failure to explain and to justify [...]. But [...] the ordinary resources of empirical observation and ordinary human knowledge [...] give us no warrant for supposing [...] that all good things, or all bad things for that matter, are reconcilable with each other. The world that we encounter in ordinary experience is one in which we are faced with choices between ends equally ultimate, and claims equally absolute, the realization of some of which must inevitably involve the sacrifice of others. Indeed, it is because this is their situation that men place such immense value upon the freedom to choose; for if they had assurance that in some perfect state, realizable by men on earth, no ends pursued by them would ever be in conflict, the necessity and agony of choice would disappear, and with it the central importance of the freedom to choose. Any method of bringing this final state nearer would then seem fully justified, no matter how much freedom were sacrificed to forward its advance.</p></blockquote><h2>Luxury space communism or annihilation?</h2><p>There&#8217;s a common refrain that goes something like this: over time, we develop increasingly powerful technology. We should aim to become wiser about using technology faster than the technology gives us the power to wreak horrors. The grand arc of human history is a race between available destructive power and the amount of governance we have. Competition, pluralism, and a big dose of anarchy might&#8217;ve been a fine and productive way for the world to be in the 1700s, but after the invention of either the Gattling gun, the atomic bomb, or AGI, we&#8217;ll have to grow up and figure out something more controlled. The moral arc of the universe might bend slowly, but it bends towards either luxury space communism or annihilation.</p><p>For an extended discussion of this, see Michael Nielsen&#8217;s essay on <a href="https://michaelnotebook.com/optimism/index.html">How to be a wise optimist about science and technology</a> (note that while Nielsen is sympathetic to the above argument, he does not explicitly endorse it). In graph form (taken from the same essay):<br></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!KO1C!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F72b53105-1553-4169-b307-3cf1bdebb5c6_960x540.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!KO1C!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F72b53105-1553-4169-b307-3cf1bdebb5c6_960x540.png 424w, https://substackcdn.com/image/fetch/$s_!KO1C!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F72b53105-1553-4169-b307-3cf1bdebb5c6_960x540.png 848w, https://substackcdn.com/image/fetch/$s_!KO1C!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F72b53105-1553-4169-b307-3cf1bdebb5c6_960x540.png 1272w, https://substackcdn.com/image/fetch/$s_!KO1C!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F72b53105-1553-4169-b307-3cf1bdebb5c6_960x540.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!KO1C!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F72b53105-1553-4169-b307-3cf1bdebb5c6_960x540.png" width="960" height="540" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/72b53105-1553-4169-b307-3cf1bdebb5c6_960x540.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:540,&quot;width&quot;:960,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Pasted image 20250714010908.png&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Pasted image 20250714010908.png" title="Pasted image 20250714010908.png" srcset="https://substackcdn.com/image/fetch/$s_!KO1C!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F72b53105-1553-4169-b307-3cf1bdebb5c6_960x540.png 424w, https://substackcdn.com/image/fetch/$s_!KO1C!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F72b53105-1553-4169-b307-3cf1bdebb5c6_960x540.png 848w, https://substackcdn.com/image/fetch/$s_!KO1C!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F72b53105-1553-4169-b307-3cf1bdebb5c6_960x540.png 1272w, https://substackcdn.com/image/fetch/$s_!KO1C!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F72b53105-1553-4169-b307-3cf1bdebb5c6_960x540.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Often, this then leads to calls for centralization, from Oppenheimer <a href="https://www.yesmagazine.org/democracy/2023/10/05/oppenheimer-world-government">advocating</a> for world government in response to the atomic bomb, to Nick Bostrom&#8217;s <a href="https://nickbostrom.com/papers/vulnerable.pdf">proposal</a> that comprehensive surveillance and totalitarian world government might be required to prevent existential risk from destructive future technologies.</p><p>A core motivation of this line of reasoning is that technology can only ramp up the power level, the variance in outcomes, the destruction. This has generally been a net effect of technological progress. But there are other consequences of technology as well.</p><p>Most obviously, there are technologies that tamp down on the negative shocks we&#8217;re subject to. Vaccines are perhaps the most important example of technology against natural threats. But the things in the world that change quickly are the human-made ones, so what matters is whether we can build technologies that can tamp negative shocks from human actions (vaccines work against engineered pandemics as well!). If all we have for defense is a stone fence, this isn&#8217;t too bad if the most powerful offense was a stone axe. But now that we&#8217;ve built nuclear weapons, can we build defensive fences strong enough?</p><p>Also: there are technologies that aren&#8217;t about offense/defense balance that make it easier or harder to enforce boundaries. Now we have cryptography that lets us keep information private, but what if in the future we get quantum computers that break existing cryptography? In the past it was hard to censor, but now LLMs make censorship trivially cheap. And consider too that what matters is not just capabilities, but propensity. Modern nuclear states could kill millions of people easily, but they&#8217;re prevented from doing so by mutually assured destruction, international norms, and the fact that nuclear states benefit from the world having lots of people in it engaging in positive-sum trade. Offense/defense balance is not the only axis.</p><p>Differential acceleration of defensive technology is something that Vitalik Buterin has argued for as part of his <a href="https://vitalik.eth.limo/general/2023/11/27/techno_optimism.html">&#8220;d/acc&#8221; thesis</a>. However, for reasons like the above, he deliberately keeps the &#8220;d&#8221; ambiguous: &#8220;defense&#8221;, yes, but also &#8220;decentralization&#8221; and &#8220;democracy&#8221;.</p><p>Is d/acc just a cluster of all the good things? With the liberalism v utilitarianism axis, I think we can restate an argument for d/acc like this: a lot of technology can be used well for utilitarian ends, in the sense that it gives us power over nature and the ability to shape the world to our desires, or to maximize whatever number we&#8217;ve set as our goal: make the atoms dance how we want. But this also increases raw power and capabilities, and importantly creates the power for some people&#8212;or AIs&#8212;to make some atoms you care about dance in a way you don&#8217;t like.</p><p>Now, if you think that unlimited raw power is coming down the technological pipeline, say in the form of self-improving AIs that <a href="https://manifold.markets/Nostr0m/will-the-technology-to-make-a-dyson">might build a Dyson sphere around the sun by 2040</a>, your bottleneck for human flourishing is not going to be raw power or wealth available, or raw ability to shape the world when it&#8217;s useful, or raw intelligence. It&#8217;s likely to be much more about enforcing boundaries, containing that raw power, keeping individual humans empowered and safe next to the supernova of that technological power.</p><p>So what we also need are <em>technologies of liberalism</em>, that help maintain different spheres of freedom, even as <em>technologies of utilitarianism</em> increase the control and power that actors have to achieve their chosen ends.</p><h2>Technologies of liberalism</h2><p>Defense, obviously, is one approach: tools that let you build walls that keep in things you don&#8217;t want to share, or lock out things you don&#8217;t like.</p><p>Often what&#8217;s important is selective permeability. At the most basic level, locks and keys (in existence since at least ancient Assyria 6000 years ago) let you restrict access to a place. Today, the ability to pass information over public channels without revealing it is at least as fundamental to our world as physical locks and keys. Also note how unexpected public key cryptography&#8212;where you don&#8217;t need to first share a secret with the counterparty&#8212;is. <a href="https://en.wikipedia.org/wiki/Diffie%E2%80%93Hellman_key_exchange">Diffie-Helman key exchange</a> is one of the cleverest and most useful ideas humans have ever had.</p><p>Encryption is of enormous significance for the politics of freedom. David Friedman <a href="https://daviddfriedman.substack.com/p/different-arguments">has argued</a> (<a href="https://web.archive.org/web/20030925021822/http:/www.hoover.org/Main/uncommon/winter98/205.html">since the 1990s!</a>):</p><blockquote><p>In the centuries since [the passing of the Second Amendment], both the relative size of the military and the gap between military and civilian weapons have greatly increased, making that solution less workable [for giving citizens a last resort against despotic government]. In my view, at least, conflicts between our government and its citizens in the immediate future will depend more on control of information than control of weapons, making unregulated encryption [...] the modern equivalent of the 2nd Amendment.</p></blockquote><p>Ivan Vendrov, in a <a href="https://nothinghuman.substack.com/p/the-transformer-and-the-hash">poetic essay</a> on the building blocks of 21st century political structure, points to cryptography (exemplified by the hash function) as the one thing standing against the centralizing information flows that incentivizes, since it lets us cheaply erect barriers that have astronomical asymmetry in favor of defense. Encryption that takes almost no time at all can be made so strong it would take hundreds of years on a supercomputer to break. In our brute physical world, destruction is easy and defensive barriers are hard. But the more of our world becomes virtual, the more of our society exists in a realm where defense is vastly favored. Of the gifts to liberty that this universe contains that will most bear fruit in our future, encryption ranks along the Hayekian nature of knowledge and the finite speed of light.</p><p>So far we are far from realizing the theoretical security properties of virtual space; reliable AI coding agents and formal verification will help. And virtual space will continue being embedded in meatspace.</p><p>In meatspace, the biggest vulnerability humans have are infectious diseases. The state of humanity&#8217;s readiness against pandemics (natural or engineered) is pathetic, and in expectations millions and more will pay with their lives for this. We need technology that lets us prevent airborne diseases, to prevent an increasing number of actors&#8212;already every major government, but soon plausibly any competent terrorist cell&#8212;wielding a veto on our lives. UV-C lighting and rapid at-home vaccine manufacturing are good starts. Also, unless we curtail this risk through broad-spectrum defensive technology, at some point it will be used as an excuse for massive centralization of power to crush the risk before it happens.</p><p>Besides literal defenses, there&#8217;s decentralization: self-sufficiency gives independence as well as safety from coercion. At a simple level, solar panels and home batteries can help with energy independence. 3D printers make it possible to manufacture more locally and more under your control. Open-source software makes it possible for you to tinker with and improve the programs you run. AI code-gen makes this even stronger: eventually, anyone will be able to build their own tech stack, rather than being funneled along the grooves cut by big tech companies&#8212;which, remember, they adversarially optimize against you to make you lose as much time out of your life as possible. The printing press and the internet both made it cheaper to distribute writing of your choice, and hence push your own ideas and create an intellectual sphere of your own without wealth or institutional buy in.</p><p>Fundamentally, however, the thing you want to do is avoid coercion, rather than maximize independence. <a href="https://allpoetry.com/No-man-is-an-island">No man is an island</a>, they say, and that might not have been intended as a challenge. Coercion is hard if the world is very decentralized, to the point that no one has a much larger power base than anyone else. However, such egalitarianism is unlikely since everyone has different resources and talents. Another key feature of the modern world that helps reduce coercion, in addition to economic and military power being decentralized, is that the world is covered in a thick web of mutual dependence thanks to trade. If you depend on someone, you are also unlikely to be too coercive towards them, as <a href="https://www.kiplingsociety.co.uk/poem/poems_dives.htm">Kipling already understood</a>. Self-sufficiency and webs of mutual dependencies are, of course, in tension. I don&#8217;t pretend there&#8217;s a simple way to pick between them, or know how much of each to have&#8212;I never said this would be simple!</p><p>AI, however, is a challenge to any type of decentralization: it needs lots of expensive centralized infrastructure, and currently the focus is on huge multi-billion dollar pre-training runs. However, even here there&#8217;s a story for something more decentralized to take center stage. First, zero-trust privacy infrastructure for private AI inference &amp; training&#8212;salvation by cryptography once again!&#8212;can give you verifiable control and privacy over your own AI, even if it runs in public clouds for cost reasons. Secondly, as Luke and I have <a href="https://lukedrago.substack.com/cp/160938645">argued</a>, the data needs to have AI diffuse through the economy run into <a href="https://www.econlib.org/library/Essays/hykKnw.html">Hayekian constraints</a> that privilege the tacit, process, and local knowledge of the people currently doing the jobs AI seeks to replace&#8212;as long as the big labs and their data collection efforts don&#8217;t succeed too hard and too fast. Combined with this, cheap bespoke post-training offers a Cambrian explosion of model variety as an alternative to the cultural &amp; value lock-in to the tastes of a few labs. Achieving this liberal vision of the post-AGI economy is what we&#8217;re working on at <a href="https://workshoplabs.ai/">Workshop Labs</a>.</p><p>Then there&#8217;s technology that helps improve institutions&#8212;democratization, but not &#8220;democracy&#8221; as in just &#8220;one person one vote&#8221;, but in the more fundamental sense of &#8220;people&#8221; (the <em>demos</em>) having &#8220;power&#8221; (<em>kratos</em>). We&#8217;ve already mentioned zero-knowledge proofs, memoryless AI, and other primitives for contracts. Blockchains, though much hyped, deserve some space on this list, since they let you get hard-to-forge consensus about some data without a central authority (needing consensus about a list is a surprisingly common reason to have a central authority!). More ambitiously, <a href="https://blog.cosmos-institute.org/p/coasean-bargaining-at-scale">Seb Krier points out</a> AI could enable &#8220;Coasean bargaining&#8221; (i.e. bargaining with transaction costs so low <a href="https://en.wikipedia.org/wiki/Coase_theorem">Coase&#8217;s theorem</a> actually holds)&#8212;bespoke deals, mediated by AIs acting on behalf of humans, for all sorts of issues that currently get ignored or stuck in horrible equilibria because negotiating requires expensive human time. Another way to increase the amount of democracy when human time is bottlenecked is AI representatives for people (though these would have to be very high-fidelity representatives to represent someone&#8217;s values faithfully&#8212;values are <a href="https://plato.stanford.edu/entries/thick-ethical-concepts/">thick, not thin</a>, and <a href="https://www.full-stack-alignment.ai/">AI design should take this into account</a>).</p><p>Finally, there are technologies shape culture and make us wiser, or let society be more pluralistic. Many information technologies, like the internet or AI chatbots, obviously fall in this category and enable great things. However, information technology, in addition to giving us information and helping us coordinate, often also changes the incentives for which memes (in the original Dawkins sense) spread, and therefore often have vast second-order consequences. On net I expect information technology to be good, but with far more requirements for cultural shifts to properly adapt to them, and far more propensity than other technologies to shift culture or politics in unexpected directions.</p><p>(Note that the pillars above are very similar to the those of the anti-intelligence-curse strategy Luke &amp; I outlined <a href="https://intelligence-curse.ai/breaking/">here</a>.)</p><p>Historically, you can trace the ebb and flow of the plight of the average person by how decentralizing or centralizing the technology most essential for national power is, and how much that technology creates mutual dependencies that make it hard for the elite to defect against the masses. MacInnes et al.&#8217;s excellent paper <a href="https://academic.oup.com/isq/article/68/4/sqae111/7757999">Anarchy as Architect</a> traces this back throughout history. Bronze weapons, which required rare bronze, meant only a small military elite was relevant in battle, and could easily rule over the rest. The introduction of iron weapons, which could be produced en-massed, lead to less-centralized societies, as the rulers needed large groups to fight for them. Mounted armored knights centralized power again, before mass-produced firearms made large armies important again. The institutions of modern Western democracy were mostly built during the firearm-dominated era of mass warfare. They were also built at a time when national power relied more and more on broad-based economic and technological progress where a rich and skilled citizenry was very valuable. This rich and skilled citizenry was also increasingly socially mobile and free of class boundaries, making it harder for the upper class to coordinate around its interests since it wasn&#8217;t demarcated by a sharp line. People also had unprecedented ability to communicate with each other in order to organize and argue thanks to new technology (and, of course, because the government now had to teach them to read&#8212;before around 1900, literacy was not near-universal even in rich countries). These nations then increasingly dissolved trade barriers between them and became interdependent&#8212;very purposefully and consciously, in the case of post-war Europe&#8212;which reduced nation-on-nation violence and coercion.</p><p>How will this story continue? It will definitely change&#8212;as it should. We should not assume that the future socio-structural recipe for liberalism is necessarily the one it is now. But hopefully it changes in ways where it still adds up to freedom for an increasing number.</p><h2>Build for freedom</h2><p>Differential technological progress is obviously possible. For example, Luke argues against what he calls &#8220;<a href="https://blog.cosmos-institute.org/p/technocalvinism">Technocalvinism</a>&#8220;, the belief that the &#8220;tech tree&#8221; is deterministic and everything about it is preordained, using examples ranging from differential progress in solar panels to the feasibility of llama-drawn carts. The course of the technology is like a new river running down across the landscape. It has a lot of momentum, but it can be diverted from one path to another, sometimes by just digging a small ditch.</p><p>There are three factors that help, the first two in general, and the last specifically for technologies of liberalism:</p><ol><li><p>There is a <a href="https://www.lesswrong.com/posts/GCRm9ysNYuGF9mhKd/the-computational-complexity-of-progress">lot of path dependence</a> in the development of technologies, and effects like <a href="https://en.wikipedia.org/wiki/Experience_curve_effect">Wright&#8217;s law</a> mean that things can snowball: if you manage to produce a bit, it gets cheaper and it&#8217;s better to produce a larger amount, and then an enormous amount. This is how solar panels went from expensive things stuck on satellites to <a href="https://caseyhandmer.wordpress.com/2024/05/22/the-solar-industrial-revolution-is-the-biggest-investment-opportunity-in-history/">on track to dominate the global energy supply</a>. Tech, once it exists, is also often sticky and hard to roll back. A trickle you start can become a stream. (And whether or not you can change the final destination, you can definitely change the order, and that&#8217;s often what you need to avert risks. All rivers end in the ocean, but they take different paths, and the path matters a lot for whether people get flooded.)</p></li><li><p>The choice of technology that a society builds towards is a deeply cultural, <a href="https://www.orphandriftarchive.com/articles/hyperstition/">hyperstitious</a> phenomenon. This is argued, for example, by Byrne Hobart and Tobias Huber&#8217;s <em>Boom: Bubbles and the End of Stagnation</em> (you can find an excellent review <a href="https://www.thepsmiths.com/p/review-boom-by-byrne-hobart-and-tobias">here</a>, which occasionally even mentions the book in question). The Apollo program, the Manhattan project, or the computer revolution were all deeply idealistic and contingent projects that were unrealistic when conceived&#8212;yes, <a href="https://monoskop.org/images/d/dc/Barbrook_Richard_Cameron_Andy_1996_The_Californian_Ideology.pdf">even the computer revolution</a>, hard as it is to remember today. Those who preach the inevitability of centralization and control, whether under the machine-god or otherwise, are not neutral bullet-biters, but accidentally or intentionally hyperstitioning that very vision into life.</p></li><li><p>Demand is high. People want to be safe! People want control! People want power!</p></li></ol><p>So how should we steer the river of technology?</p><p>If the AGI labs&#8217; visions of AGI is achieved, by default it will disempower the vast majority of humanity instantly, depriving them of their power, leverage, and income as they&#8217;re outcompeted by AIs, while <a href="https://intelligence-curse.ai/">wrecking the incentives that have made power increasingly care about people</a> over the past few centuries, and elevating a select few political &amp; lab leaders to positions of near godhood. It might be the most radical swing in the balance of power away from humanity at large and towards a small elite in human history. And the default solution for this in modern AI discourse is to accept the inevitability of the machine-god dictatorship, accept the lock-in of our particular society and values, and argue instead about who exactly stands at its helm and which exact utility function is enshrined as humanity&#8217;s goal from here on out.</p><p>All this is happening while liberalism is already in cultural, political, and geopolitical retreat. If its flame isn&#8217;t fanned by the ideological conviction of those in power or the drift of geopolitics, its survival increasingly depends on other sources.</p><p>So: to make an impact, don&#8217;t build to maximize intelligence or power, because this will not be the constraint on a good future. This is a weird position to be in! Historically, we <em>did</em> lack force and brains, and had to claw ourselves out of a deep abyss of feebleness to our current world of engines, nuclear power, computers, and coding AIs. Even now, if radical AI was not an event horizon towards which we were barreling towards that was poised to massively accelerate technology and reshape the position of humans, <em>more power</em> would be a very reasonable thing to aim towards&#8212;we would greatly benefit from energy abundance thanks to fusion, and robotics beyond our dreams, and space colonization. (And if that radical AI looks majorly delayed, or gets banned, or technology in general grinds to a halt, then making progress towards tools of utilitarianism and power would once again be top-of-mind.) But if radical tech acceleration through AI is on the horizon, we might really be on track for literally dismantling the solar system within the few decades. The thing we might lack is the tools to secure our freedoms and rights through the upheaval that brings. Technologists, especially in the Bay Area, should rekindle the belief in freedom that guided their predecessors.</p><p>For decades, the world has mostly been at peace, liberty has spread, and the incentives of power have been aligned with the prosperity of the majority. History seemed over. All we had to do was optimize our control and power over nature to allow ever greater and greater welfare.</p><p>But now, history is back. Freedom is on trial. The technology we choose to build could tip the scales. What will you build?</p><p></p><p></p><p><em>Thank you to Luke Drago &amp; Elsie Jang for reviewing drafts of this post.</em></p><p></p><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.nosetgauge.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading No Set Gauge! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[The Intelligence Curse: an essay series]]></title><description><![CDATA[Solving human irrelevance in the face of AI]]></description><link>https://www.nosetgauge.com/p/the-intelligence-curse-an-essay-series</link><guid isPermaLink="false">https://www.nosetgauge.com/p/the-intelligence-curse-an-essay-series</guid><dc:creator><![CDATA[Rudolf Laine]]></dc:creator><pubDate>Thu, 24 Apr 2025 12:02:22 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!nxDw!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff249178d-380f-4c41-a8e3-10693fd58909_1600x897.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Together with <a href="https://lukedrago.substack.com/https://lukedrago.substack.com/">Luke Drago</a>, we&#8217;ve published an essay series on what we call the intelligence curse. Most content is brand new, and all previous writing has been heavily reworked.</em></p><p><em>Visit <a href="http://intelligence-curse.ai">intelligence-curse.ai</a> for the full series.</em></p><p><em>Below is the introduction and table of contents.</em></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!nxDw!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff249178d-380f-4c41-a8e3-10693fd58909_1600x897.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!nxDw!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff249178d-380f-4c41-a8e3-10693fd58909_1600x897.jpeg 424w, https://substackcdn.com/image/fetch/$s_!nxDw!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff249178d-380f-4c41-a8e3-10693fd58909_1600x897.jpeg 848w, https://substackcdn.com/image/fetch/$s_!nxDw!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff249178d-380f-4c41-a8e3-10693fd58909_1600x897.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!nxDw!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff249178d-380f-4c41-a8e3-10693fd58909_1600x897.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!nxDw!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff249178d-380f-4c41-a8e3-10693fd58909_1600x897.jpeg" width="1456" height="816" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f249178d-380f-4c41-a8e3-10693fd58909_1600x897.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:816,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!nxDw!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff249178d-380f-4c41-a8e3-10693fd58909_1600x897.jpeg 424w, https://substackcdn.com/image/fetch/$s_!nxDw!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff249178d-380f-4c41-a8e3-10693fd58909_1600x897.jpeg 848w, https://substackcdn.com/image/fetch/$s_!nxDw!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff249178d-380f-4c41-a8e3-10693fd58909_1600x897.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!nxDw!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff249178d-380f-4c41-a8e3-10693fd58909_1600x897.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Art by <a href="https://x.com/NomadsVagabonds">Nomads &amp; Vagabonds</a>.</em></figcaption></figure></div><p>We will soon live in <a href="https://ia.samaltman.com/">the intelligence age</a>. What you do with that information will determine your place in history.</p><p>The imminent arrival of AGI has pushed many to try to <a href="https://danfaggella.com/flex/">seize the levers of power as quickly as possible</a>, leaping towards projects that, if successful, would comprehensively automate all work. There is a <a href="https://am.jpmorgan.com/gb/en/asset-management/per/insights/market-insights/investment-outlook/ai-investment/#:~:text=According%20to%20data,Exhibit%2014).">trillion-dollar</a> arms race to see who can achieve such a capability first, with <a href="https://www.mechanize.work/#:~:text=The%20market%20potential%20here%20is%20absurdly%20large%3A%20workers%20in%20the%20US%20are%20paid%20around%20%2418%20trillion%20per%20year%20in%20aggregate.%20For%20the%20entire%20world%2C%20the%20number%20is%20over%20three%20times%20greater%2C%20around%20%2460%20trillion%20per%20year.">trillions more</a> in gains to be won.</p><p>Yes, that means you&#8217;ll lose your job. But it goes beyond that: this will remove the need for regular people in our economy. Powerful actors&#8212;like states and companies&#8212;will no longer have an incentive to care about regular people. We call this <strong>the intelligence curse</strong>.</p><p>If we do nothing, the intelligence curse will work like this:</p><ul><li><p>Powerful AI will push automation through existing organizations, starting from the bottom and moving to the top.</p></li><li><p>AI will obsolete even outlier human talent. Social mobility will stop, ending the social dynamism and progress that it drives.</p></li><li><p>Non-human factors of production, like capital, resources, and control over AI, will become overwhelmingly more important than humans.</p></li><li><p>This will usher in incentives for powerful actors around the world that break the modern social contract.</p></li><li><p>This could result in the <a href="https://gradual-disempowerment.ai/">gradual</a>&#8212;or <a href="https://www.forethought.org/research/ai-enabled-coups-how-a-small-group-could-use-ai-to-seize-power">sudden</a>&#8212;disempowerment of the vast majority of humanity.</p></li></ul><p>But this prophecy is not yet fulfilled; we <em>reject</em> the view that this path is inevitable. We see a different future on the horizon, but it will require a deliberate and concerted effort to achieve it.</p><p>We aim to change the incentives driving the intelligence curse, maintaining human economic relevance and strengthening our democratic institutions to withstand what will likely be the greatest societal disruption in history.</p><p>To break the intelligence curse, we should chart a different path on the tech tree<em>, </em>building technology that lets us:</p><ol><li><p><strong>Avert</strong> AI catastrophes by hardening the world against them, both because it is good in itself and because it removes the security threats that drive calls for centralization.</p></li><li><p><strong>Diffuse</strong> AI, to get it in the hands of regular people. In the short-term, build AI that augments human capabilities. In the long-term, align AI directly to individual users and give everyone control in the AI economy.</p></li><li><p><strong>Democratize</strong> institutions, making them more anchored to the needs of humans even as they are buffeted by the changing incentive landscape and fast-moving events of the AGI transition.</p></li></ol><p>In this series of essays, we examine the incoming crisis of human irrelevance and provide a map towards a future where people remain the masters of their destiny.</p><h2>Chapters</h2><h3>1. <a href="https://intelligence-curse.ai">Introduction</a></h3><p>We will soon live in the intelligence age. What you do with that information will determine your place in history.</p><h3>2. <a href="https://intelligence-curse.ai/pyramid">Pyramid Replacement</a></h3><p>Increasingly powerful AI will trigger pyramid replacement: a systematic hollowing out of corporate structures that starts with entry-level hiring freezes and moves upward through waves of layoffs.</p><h3>3. <a href="https://intelligence-curse.ai/capital">Capital, AGI, and Human Ambition</a></h3><p>AI will make non-human factors of production more important than human ones. The result may be a future where today's power structures become permanent and frozen, with no remaining pathways for social mobility or progress.</p><h3>4. <a href="https://intelligence-curse.ai/defining">Defining the Intelligence Curse</a></h3><p>With AGI, powerful actors will lose their incentive to invest in regular people&#8211;just as resource-rich states today neglect their citizens because their wealth comes from natural resources rather than taxing human labor. This is the intelligence curse.</p><h3>5. <a href="https://intelligence-curse.ai/shaping">Shaping the Social Contract</a></h3><p>The intelligence curse will break the core social contract. While this suggests a grim future, understanding how economic incentives reshape societies points to a solution: we can deliberately develop technologies that keep humans relevant.</p><h3>6. <a href="https://intelligence-curse.ai/breaking">Breaking the Intelligence Curse</a></h3><p>Avert AI catastrophes with technology for safety and hardening without requiring centralizing control. Diffuse AI that differentially augments rather than automates humans and decentralizes power. Democratize institutions, bringing them closer to regular people as AI grows more powerful.</p><h3>7. <a href="https://intelligence-curse.ai/history">History is Yours to Write</a></h3><p>You have a roadmap to break the intelligence curse. What will you do with it?</p><p></p><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.nosetgauge.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe for updates</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[A History of the Future, 2030-2040]]></title><description><![CDATA[A scenario, part 3.]]></description><link>https://www.nosetgauge.com/p/a-history-of-the-future-2030-2040</link><guid isPermaLink="false">https://www.nosetgauge.com/p/a-history-of-the-future-2030-2040</guid><pubDate>Mon, 17 Feb 2025 01:56:37 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Q5Co!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64c0eff6-08aa-4bde-ad35-adadb3c42326_1920x1239.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Q5Co!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64c0eff6-08aa-4bde-ad35-adadb3c42326_1920x1239.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Q5Co!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64c0eff6-08aa-4bde-ad35-adadb3c42326_1920x1239.png 424w, https://substackcdn.com/image/fetch/$s_!Q5Co!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64c0eff6-08aa-4bde-ad35-adadb3c42326_1920x1239.png 848w, https://substackcdn.com/image/fetch/$s_!Q5Co!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64c0eff6-08aa-4bde-ad35-adadb3c42326_1920x1239.png 1272w, https://substackcdn.com/image/fetch/$s_!Q5Co!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64c0eff6-08aa-4bde-ad35-adadb3c42326_1920x1239.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Q5Co!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64c0eff6-08aa-4bde-ad35-adadb3c42326_1920x1239.png" width="1456" height="940" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/64c0eff6-08aa-4bde-ad35-adadb3c42326_1920x1239.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:940,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:4160802,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Q5Co!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64c0eff6-08aa-4bde-ad35-adadb3c42326_1920x1239.png 424w, https://substackcdn.com/image/fetch/$s_!Q5Co!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64c0eff6-08aa-4bde-ad35-adadb3c42326_1920x1239.png 848w, https://substackcdn.com/image/fetch/$s_!Q5Co!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64c0eff6-08aa-4bde-ad35-adadb3c42326_1920x1239.png 1272w, https://substackcdn.com/image/fetch/$s_!Q5Co!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64c0eff6-08aa-4bde-ad35-adadb3c42326_1920x1239.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>The Great Day of His Wrath. </em>(John Martin)</figcaption></figure></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.nosetgauge.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.nosetgauge.com/subscribe?"><span>Subscribe now</span></a></p><p>(Part 1 <a href="https://nosetgauge.substack.com/p/a-history-of-the-future-2025-2027">here</a>, part 2 <a href="https://nosetgauge.substack.com/p/a-history-of-the-future-2027-2030">here</a>)</p><h2><strong>The end of white-collar work and the new job scene</strong></h2><p>By the late 2020s, office jobs in developed countries are now basically about overseeing and providing direction to AI systems, and the last part of that is mostly on-paper rather than in practice. There is lots of talk about values and missions and the future, and a lot of unspoken communication about office politics and status. Many office workers don&#8217;t do much at all. Concretely, they might get to work, have a team standup, check in on how the AIs are doing, have some ritualistic meetings with their manager and any employees they have, and rubber-stamp some AI decisions that they&#8217;re contractually or legally obliged to stamp, with this adding up to only a few hours. Occasionally they might decide to change some goal the AIs have been given, but that requires just speaking or typing a paragraph. Many people feel guilty about this, but it&#8217;s mostly a quiet guilt. They fill their time with office chat or scrolling on their phones. Many companies become more social and more about community. HR has never been more influential. Everything&#8217;s both more cuddly and/or more viciously political now that the ugly raw realities of individual competence don't matter any more.</p><p>Some organisations try to fire lots of people. Sometimes it goes well. Sometimes it goes badly, and they realise that some human somewhere was holding some knowledge in their head, or nudging the mission in the right direction, in a way that was essential. However, by then it&#8217;s too late, and it&#8217;s hard to say which person it actually was anyway. Among the more ruthless or tech-adjacent management cultures, there&#8217;s a lot of talk about figuring out what the load-bearing humans in any organisation are, and how this is surprisingly difficult to do at a large organisation. Some companies develop internal AI systems to try to figure this out (or buy such systems from startups), but they need to collect some data about the functioning of the org first, which takes time. Also, the workers are incentivised to <a href="https://www.thepsmiths.com/p/review-the-art-of-not-being-governed">resist</a> and fight back in a thousand subtle ways, and they do. Also, sometimes when an org tries to fire a lot of people, an online mob emerges to hate on them, influencers pile in and create 13 different cinematic universes where the theme is all how Company X is the pinnacle of all human evil, sometimes a former employee creates an AI-powered revenge cult (several assassinations happen as a result from the more violent of the cults), and sometimes politicians pick up the issue. The companies, largely, were profitable before, and are more profitable now that they&#8217;ve enjoyed a few years&#8217; of revenue growth without expanding headcount. Therefore, mass firing is surprisingly rarely worth it, even though it would in principle be possible. A few firms facing crises or with especially effective or risk-tolerant leadership buck these trends and aggressively slash costs by cutting huge amounts of human workers.</p><p>What developed country firms are not doing is hiring new workers or replacing anyone who retires. What they are doing is replacing any foreign contractors or service providers with cheaper AI ones.</p><p>This creates several groups of disaffected people. First, the youth in developed countries, who have much worse job prospects than the preceding generation. For people looking for their first job in 2030-2031 in a developed country, the options are roughly:</p><ol><li><p>Working in services where being human intrinsically matters (elderly care, retail, restaurants, hospitality, teaching, etc.). Healthcare is by far the most prestigious one and what many aim for (even though doctors&#8212;or at least all the good ones&#8212;defer all diagnosis and other intellectual work to the AIs). The cartel-like nature of medical licensing bodies, strain on state budgets, and the fact that most of the actual work is done by AIs means that the number of doctors or nurses hasn't increased much, though, so entry has become even more competitive. Policing and primary education also continue hiring humans at scale.</p></li><li><p>Jobs that are effectively sinecures. This includes many positions in government and civil service. In the EU, regulation passed in 2028 means that many companies are forced to hire human "overseers" to key positions. Of course, the supply of sinecures is set by regulation and funding for economically useless activities. Competition for such positions is therefore extremely harsh, and (because the selection criteria, having no reason to be one thing rather than another, inherit the latest credentialist instantiation of the 21st century West's bureaucratic blob) requires extreme conformism. This category has a fuzzy boundary with the first, depending on whether you value the ceremonial human touch as a key part of the service or not.</p></li><li><p>A particular example of this is the law. Lawyers have two major advantages. First, their job deals closely with important social questions of legitimacy and propriety, making it a natural claim that something fundamental would be lost if the human presence was gone. The presence of lawyers evolves to be more ceremonial and symbolic&#8212;almost religious&#8212;but it stays. Second, lawyers make up a lot of the rules for themselves, and interpret the statutes for everyone else. Third, a lot of politicians are lawyers, or have friends who are lawyers, which make them attuned to lawyer interests. This gives lawyers a lot of leeway in what automation they allow. In many countries the rules are bent such that it is flat-out illegal to consult an AI on legal matters; you have to go through a human lawyer. AI companies are forced to train their AIs to comply with this ("I'm sorry, but as an AI it is illegal for me to give advice on legal matters, so I recommend you hire a licensed lawyer"). Of course, all of the actual legal research and argumentation is done by AIs&#8212;the lawyers just monopolise the position of being allowed to ask them.</p></li><li><p>Manufacturing, which is booming, especially as productivity has been lifted by AI management and oversight. Many manufacturing jobs involve wearing an earpiece through which you receive detailed step-by-step orders from an AI (and occasionally AR glasses that can show you diagrams or an overlay for how to move your hands). A large fraction of people go into this, even if they have prestigious university degrees (many of the prestigious degree-holders do not have their salary and status expectations met and become resentful).</p></li><li><p>Academia. There are still humans in academia who somewhat matter for intellectual progress, but they're all either experienced humans with years of research taste in economically-valuable non-purely-mathematical areas (who are actually in decently high demand, as the AI labs chase feedback sources that will help them faster and cheaper get the models superhuman at even the very last set of very long-horizon, hard-to-measure skills), or (especially in the US) "prof-luencers" who use the status of a successful prior academic career to boost their influencer careers. New entrants to academia get their academic salaries (if they win an ever more cut-throat competition), but not the hope of actually mattering for intellectual progress. Some derive satisfaction that they can at least keep deep human expertise alive into the future&#8212;though it seems like without any ground-truth feedback signal, many lineages of human expertise will become <a href="https://samoburja.com/on-the-loss-and-preservation-of-knowledge/">dead knowledge</a> within a generation even if people still go through the motions of &#8220;learning&#8221; them.</p></li><li><p>Becoming an influencer. Works for some, but the competition is extremely tough (though it does help that "being verifiably human" is in vogue).</p></li><li><p>Becoming a socialite. The infinite variety and competition (human and AI) in the digital world is driving a resurgence of an in-person social scene. However, for this to be a "career choice", you must either already be wealthy, or have some other factor in your favour. The overwhelmingly most common such factor is being a young woman who inserts herself into the social scene of moneyed men.</p></li><li><p>Becoming a musician, artist, or poet. The main constraint is funding, of which there are two important types: government subsidies (which generally increase, as being vaguely pro human self-expression is a common government answer to what people should do with themselves in the age of AI, especially in socially progressive European countries), and wealthy patrons. Being an artist for the latter often melds into being a socialite, since in-person local artists are the prevailing fashion. Many nouveau rich techies, wanting to erase their association with the now-uncool world of software, throw money at artists who live in their local community to do some arts-and-crafts thing and then show up with it to their party and say vaguely artsy things.</p></li><li><p>Going into politics. This has become more appealing to the young in particular, since the future looks uncertain and the youth are the ones who expect to live in it longest. There are many AI youth activists (in the sense of specialising in the topic of AI, being AIs, or both), who try to use their position to advance youth interests. The problem is that they don't have concrete policy asks beyond "allocate more money to us", which puts them at odds with every other interest group in society, many of which (e.g. the retired) outnumber them in raw numbers as a voting bloc as well as in terms of resources, power, and influence.</p></li></ol><p>Culturally, intellectualism is out, after having climbed in cultural status for two centuries before the mid-2020s as labour-augmenting technology and globalisation scaled its power. Charisma, conformism, sociability, authenticity, and propriety are all in. In the early 2030s, the US is becoming more European in its attitudes, especially along the above dimensions. While the high-water of conformism that the 2010s culture wars and academic sclerosis had caused receded throughout 2022-2027, a modified and less-political, more European-style propriety-focused conformism rose around 2030. First, this was driven by cultural changes downstream of AI reducing the rewards for risk and entrepreneurship. Second, there was a cultural backlash against the techies (who were seen as having pretensions of importance after the automation of software in the late 2020s, and as lead by a bastion of improper disruptive moguls who were on the wrong side of a Republican power struggle in 2028), and through the techies the culture of ambition that had become central to their self-narrative.</p><p>A second disaffected group is the developing countries. Replacing outsourced foreign human services (e.g. call centres) with AIs is a cost-saving that can be done without political or social repercussions, so all companies did it&#8212;often at significant scale as early as 2025-2026 for text-only tasks. As a result, services-led export growth is dead. This is bad for India and the Philippines in particular. India just about achieved the economic heft where it could've been relevant in AI, but throughout the 2020s was unable to become entrenched in any part of the AI supply chain. At the same time, as more of the population in developed countries goes back to working in manufacturing, political demands for protecting developed country manufacturers from competition with developing countries grows. This leads to even more tariffs, on top of the already-existing late 2020s trend towards more and more tit-for-tat tariff escalation. This makes developing country growth based on goods exports harder. The biggest shock to existing goods export industries won&#8217;t arrive until a few years later when the robots show up, but investment into the developing world already dries up, as US productivity growth rises and is expected to rise even more.</p><p>China is about a year behind in leading AI tech but about 2-3 years behind in AI diffusion. The Chinese public and the CCP are watching the coming wave of AI job automation with worry, especially as there is a big cultural emphasis on exactly the types of academic skills that are getting obsoleted quickly. The CCP is very worried about stability. More and more people are joining the Party, since they see that other opportunities for social advancement are ending. The Party is making many more roles in it available, and using this carrot to incentivise people to adhere to Party principles even more strongly. The AI surveillance state keeps expanding; there is now AI interpretation of much CCTV footage of public streets, for example. There are efforts underway to modernise (i.e., convert to AI) most of the military, such that the Party's control cannot be threatened even if the human military is destabilised.</p><p>In the US, the leading plan seems to be a hodgepodge of regulation-mandated human job roles, and eventually maybe UBI. However, a fiscal crisis is on the horizon because of the looming social security trust fund exhaustion. GDP growth in 2031-2032 is hitting 5% per year but full UBI still seems expensive. In the EU, there is more state intervention and regulation aimed at keeping humans in the loop, with massive corporate and government hierarchies of jobs that are effectively pure sinecures where the work is all done by AIs, which is temporarily reducing the demand for flat-out UBI.</p><p>When governments ask companies what their blockers are, companies cite regulations that keep humans in the loop, and (when off the record) everyone shares a sentiment that humans aren&#8217;t actually in the loop anyway. Shortcuts are already being taken to reduce the human oversight component. It&#8217;s very hard to do this legally, because there are often government-mandated AIs monitoring compliance with the human oversight rules. Two firms might want to maintain their human workers for complicated regulatory and office politics and inertia reasons, but they&#8217;re competing against each other, and against full-AI firms, and against foreign adversaries. So pressure increases to cut the unnecessary weight. There&#8217;s also a <a href="https://inferencemagazine.substack.com/p/ai-and-jurisdictional-choice">race to the bottom internationally</a>. Many autonomous AI-run companies in 2030-2033 move to less-regulated areas, take the slight hit of running on open-source models, and serve customers from there. However, this global decentralisation is reversed once the robotic revolution&#8212;subsidised and encouraged by the American and Chinese governments&#8212;gets under way.</p><h2><strong>Lab strategy amid superintelligence and robotics</strong></h2><p>The state of AI capabilities around 2030 is roughly as follows: wherever there is an easy feedback signal and a high performance ceiling, such as maths or code, the models are incomprehensibly superhuman. Where rapid iteration is possible but the performance ceiling is not as high, like having sales calls, the AIs are better than all humans. In general, the AIs can be more charismatic and persuasive than humans, but this does not give them superpowers over steering individual humans as they like, especially when to do so they would have to compete with every other memetic force in society as well as the individual's resistance to being psychologically hacked. Wherever there is a large pile of information, such as supply chain routing or crystal structure prediction or history or legal precedent, the AIs are superhuman at spotting and understanding the patterns and generalising them to new instances. However, models appear to still be roughly human-level at long-horizon tasks with ambiguous success metrics. Companies, governments, and research agendas&#8212;even the scrappier, faster-changing ones&#8212;are still piloted by humans who make real strategy decisions, even though in practice it's a human riding on a vast wave of AI supercognition, and the trend is towards more and more delegation as the systems improve. Real-world progress in hard tech is also varied. There are many breakthroughs in parts of materials science and molecular biology driven by things like material property and protein folding prediction that cuts down on empirical iteration. However, other tasks turn out to be computationally intractable even to the smart AIs, even if they often achieve very large efficiency gains over the human state-of-the-art by inventing superhumanly good heuristics. No one has figured out how to turn the vast amounts of intelligence-on-tap into magical-seeming technical progress in atoms, even though engineering work now happens much faster and at a higher quality level and with less margin between practical and theoretical performance.</p><p>In 2029, OpenAI rebrands its models to just &#8220;o&#8221;. Everyone has Opinions. It&#8217;s a big advance in raw intelligence, but almost no one can tell. Instead of a variety of sizes of an o-series model with updates every few months, from now on there will be a few varieties (differing mainly in size, like o-small and o-large and an internal-only o-huge, but also with some specialised finetuned models, e.g. o-math and o-chat). Individual instances of the models can use their medium-term memory as context when they&#8217;re doing agentic tasks, but they can also run in &#8220;functional&#8221; or &#8220;API&#8221; mode where that is disabled. More than half of OpenAI&#8217;s model revenues still come from functional mode calls rather than running instances as agents that develop their own memories and learn on the fly, but this proportion is steadily falling. There&#8217;s a new model checkpoint released every day, with the newest information from that day already in its weights, and the occasional larger improvement.</p><p>By 2030, OpenAI has culled almost all of its human employees. This is the main advantage of their latest model internally&#8212;the tacit internal knowledge that various humans previously had that would&#8217;ve made the human-level-ish o6 not quite adequate at wholesale replacement of OpenAI engineers matters less when o-huge can just rederive the tacit knowledge from scratch very quickly.</p><p>OpenAI's b-series human robots reach annualised shipment volumes of 1M/year in late 2031, which gives it about 50% market share in the total domestic robot servant market. Several million other general-purpose robots (e.g. for use in manufacturing) are also being sold by 2031.</p><p>OpenAI is seen by some as a slightly shambolic conglomerate, like an Oracle or IBM or Microsoft, and by others as the original and one true AI company that is destined to be &gt;50% of world GDP.</p><p>The robotics sector is split between special-purpose robots with modern AI integrations, e.g. window-cleaning robots and pipe-crawling repair robots and delivery drones, and general-purpose robots being pursued by OpenAI and several other companies (including a struggling Franco-German startup that is kept afloat by the EU being hell-bent on endlessly subsidising it until Europe finally has a big tech company&#8212;the European Commission is confused why this is not producing results). Both paths seem technically feasible. However, the general-purpose robotic players are the better-resourced ones, and are run by people whose main past reference point was the generative AI wave, and therefore they are philosophically big believers in scaling laws, so they are betting on collecting all the robotics data as the path to improving quality, and on <a href="https://en.wikipedia.org/wiki/Experience_curve_effects">Wright's law</a> to bring down hardware costs as they build more and more of the same thing.</p><p>All of this is also happening at unprecedented efficiency and speed compared to prior research efforts, since there are superintelligent STEM AIs around inventing algorithms that massively bring down the sample complexity of the robotics control algorithms, organising the assembly lines, doing the CAD work, and so on. However, the actual learning to move part is still a machine learning problem bottlenecked by data, and there is no magic wand that can instantly create massive robot factories from scratch (especially given the raw resources required). The output scaling curve looks to be roughly a 4x increase of robotics capacity per year, though. This is expected to rise for 2033-2035, as the robots automate more and more of the robot production pipeline, but bottlenecks abound, and energy and land constraints (mostly downstream of regulation) are harsh.</p><p>Anthropic works with a bunch of Western governments and NGOs on strict KYC for agentic model customers&#8212;the standards have so far been somewhat shoestring, the coming robot wave is making the need much clearer, and there was a big scandal last year with a heavily AI-aided chemical terrorist attack. The cyber situation has calmed down though, with defense dominating, as key code is now either provably correct or so thoroughly tested by countless AI systems that it's close enough. Biological capabilities have already been artificially kept down by most of the key model players (including open-source and Chinese ones). Taking any large-scale actions with models that aren't from the dark web in the West and China, especially in wet lab virology or DNA synthesis, requires specific access permissions from the labs through government-mandated schemes. However, by 2030 there are open-source dark web models that will do whatever you want including designing candidate pandemic agents that are unnaturally lethal and virulent, and there is no quick way to pandemic-proof the world against bioterrorism. The remaining difficulty of wet lab work, the low number of totally insane actors, and AI surveillance are the main forces keeping the per-year odds not too high, but civilisation is clearly running a big risk. The national security apparatus in both the US and China is more relaxed about this threat than it would otherwise be, because the military and economy are both increasingly robotic and so it&#8217;s not a threat to the regime even if most of the population drops dead from mega-flu. For example, the US war plans in event of a devastating pandemic (or nuclear) attack now include AIs substituting for any of the critical industry CEOs or defense staff that die.</p><p>Another big Anthropic effort is AI for biology. They want to cure cancer, make humans live forever, etc. A major internal faction also wants to pursue human intelligence augmentation but leadership fears this would be too controversial to discuss in public, so they just have a single secret team working with the CIA on it. Innovation in biotech has definitely risen, since designing promising drug candidates is ridiculously fast and cheap, but the bottleneck even before the AI revolution was less the design part and more <a href="https://www.slowboring.com/p/the-case-for-clinical-trial-abundance">clinical trial regulation</a>. Anthropic is curating datasets, acquiring laboratory automation startups, and working with regulators to cut down red tape. This will take years to bear fruit, but seems to be leading towards a biotech revolution over the next decade.</p><p>Anthropic is also trying to use biotechnology to bootstrap powerful nanotechnology. However, the company&#8217;s attempts to get their AIs to do the physics and engineering hit some snags, especially as they lack xAI&#8217;s or GDM&#8217;s specialisations in physics/maths/engineering (having trusted more in domain-general intelligence). Still, it is the AI era, so the AIs can fairly quickly get up to speed on this stuff, and the Pentagon is helping.</p><h2><strong>Towards the automated robot economy</strong></h2><p>In 2033, about 40 million humanoid robots are shipped. An increasing fraction is going to industrial uses. Costs have come down to that of a cheap car and are declining further, especially as the entire manufacturing process can now be done by the robots themselves in the most advanced factories. This also means that full AI control and real-time optimisation of the entire robot manufacturing line is possible, leading to unparalleled factory output growth and ease of iterating on the design.</p><p>As a result, over 2032-2034 there's a Cambrian explosion of robot diversity into non-humanoid form factors. By 2035, a large fraction of developed country consumers have household robots performing almost all manual tasks at home. Construction work, assembly line work, agricultural work, solar panel installation, plumbing work, industrial machinery repairs, and electrical utility jobs can all in principle be done fully by robots by 2034. The main constraint is energy and resources for the physical manufacturing of the robots&#8212;as well as land and regulations.</p><p>By 2034-2035, advances in nanotech are also arriving. Rather than a single magical-seeming assembler, the nanotech advances are mostly in medical areas (such as targeted drug delivery to specific locations within the body, which is a huge boost to cancer treatment, and early prototypes of cellular repair machines), and in materials science advances that allow for stronger and lighter and self-healing materials, and better batteries. These can all be used in robots; some look supernaturally strong and capable to humans. The manufacturing robots also get "magic fingers", where the tip of a robot appendage is a surface that can do very controlled and fine-grained precision welding, polymer (un)curing, deposition of substances, and catalysis of chemical reactions..</p><p>The 40 million humanoid robots shipped worldwide in 2033 do roughly the work of 80 million human workers since they can work longer hours than humans. In 2034, there are 240 million human-worker-equivalents of robotic capacity shipped, and in 2035 about 1.1 billion human-worker-equivalents.</p><p>Politically, this is as if hundreds of millions of extremely talented immigrants who accept below-minimum-wage jobs had suddenly sprouted out from the ground, in each of the developed countries and China. Years of upheaval in white-collar work have given politicians and activists experience in dealing with such things, and they are better prepared.</p><p>In America, the Republicans narrowly keep the White House in 2032. The Democrats ran on an attempt to solve rising unemployment through European-style human-in-the-loop laws, including an expansion of "pro-social, meaning-creating" human roles in the government bureaucracy, education, and the lawyer cartel, while having a major retraining initiative for blue-collar workers threatened by robotics. In the few months before the election, there was a burst of about a hundred thousand people losing their jobs very directly to robots. A run of impressive robotics demos fermented hysterical online influencer coverage and blue-collar job fears. The retraining initiatives for blue-collar workers became seen as insufficient and out-of-touch with the "average American" who does not want to be reeducated into performing some ceremonial role in a bureaucracy whose culture they don't agree with.</p><p>The Republicans counter this with the PROSPER Act (Promoting Robot Ownership and Small-business Prosperity through Economic Restructuring), which they campaign on and pass in 2033. This creates a car dealership -like model for robot ownership, where robotics companies are not allowed to sell &#8220;consumer robotics services&#8221; directly to consumers (sectors like defense and mining are exempt). &#8220;Ordinary Americans" can apply for loans to start their own robotic services business. Also, a license is required to sell consumer robotics services in a given territory, and a given legal entity can only operate in one territory. The territories default to state legislative districts, most of which are between 30k-150k in population, but states are allowed to change the territory unit. Licenses for a territory are granted at the local level. For example, Joe Smith in Prescott, Arizona might get a government loan, buy 10 plumbing robots, and sell their services to other Prescott residents. He himself doesn't do much, since the robots do the plumbing and the AI does the planning, logistics, accounting, and so on for him. But nominally, he is now a small-business owner, and is most definitely not a welfare recipient freeloading on Uncle Sam.</p><p>If any robotics licensing territory gets too much competition in a single robotics services vertical, competition drives margins to zero. There is also little that differentiates the different robotics service providers. Therefore, an instant race begins for regulatory capture of each robotics license territory, which is often won by whichever actor had the most networks and funds at the beginning (though anti-trust prevents full monopolies, so there's almost always at least 2 service providers). Much of the market share fluctuation becomes about social networks and persuasion. The savvy robotics license owners in particular try to manipulate local cultural currents to restrict the granting of licenses to new entrants. Alternatively, the leader of a local AI-powered personality cult will just declare who deserves the licenses. Even with the robotics licensing regime, though, only a small fraction of the population is owners of economically-relevant assets. Social and economic life increasingly revolves around the few families with control over income-generating assets (whether sinecures or robotics licenses or property or stocks). Marriage into such families gradually becomes a more and more common tool of socioeconomic ambition. Many give up on earning an income at all, and make ends meet by moving to areas with ridiculously cheap property.</p><p>Above all local scenes are the true US national elite&#8212;powerful politicians, billionaires, senior government advisors, and some others. On average they still feel some noblesse oblige towards the lower classes, though in the late 2030s this is waning as they start feeling in their bones that <a href="https://lukedrago.substack.com/p/the-intelligence-curse">their position of power is not dependent on the people anymore</a>. However, their main preoccupation is status competition with others on their level. Many of these are inter-elite disputes with little bearing for the world, but on net there is also a strong desire to compete with China. In particular, the narrative that the race through the robotics buildout will be decisive for the far-flung future of humanity gained a lot of prominence through the late 2020s and early 2030s. This creates a strong elite consensus that competition with China must be won, and that the way to do so is to stabilise the domestic situation, but then otherwise let the robotics wave rip. The plans for domestic semiconductor self-sufficiency are on track to come true only a bit behind schedule in 2034. Actually-working ICBM defense, designed by superhuman engineering AIs around 2030, is fully online and working by 2034 thanks to the speed of manufacturing scaleups in the age of robotics. The military is able to field hundreds of millions of small drones and millions of robot soldiers. Pentagon projects on nanotechnology and other exotic physics applications may bring about powerful new technologies within another few years.</p><p>China, of course, also sees the need to win, especially as its lead in industrial robotics vanishes when America&#8217;s robotics revolution happens a bit before China&#8217;s. The CCP is also decoupling its treatment of the human economy from its treatment of geopolitics and the "real" robotic economy. In 2034, the CCP declares that citizens need to "eat bitterness", in the form of accepting per-capita living standards stagnating for a while (at around $37k, PPP-adjusted, in 2025 dollars) while the state diverts resources to fueling the robotic revolution to avoid losing in the geopolitical competition.</p><p>In the EU, AI diffusion has been slower due to regulatory hurdles, but the extinction of white-collar work is still well underway, and the robotics wave is coming only a few years after the US and China. However, this delay is enough to make the EU geopolitically irrelevant. The greatest external threat to the EU is Russia, which has suddenly gotten much richer as Chinese companies effectively colonize Siberia to mine resources to fuel China's robotic buildup while paying large rents to the Russian government. The US lead at the robotics revolution also drains manufacturing jobs out of the EU, until EU countries are politically forced to shut off trade (though a political movement, active especially in Eastern Europe, would've wanted to negotiate a stronger US security presence in exchange for letting trade continue and domestic industries wither). Various proposals for UBI float around, but economic turmoil makes the prospect of funding it uncertain, and the political fight by special interest groups for privileges for their group in particular is extremely fierce and they are all opposed to UBI for everyone. By 2036, functionally everyone within the EU has some kind of regular state payout they live on, not through a single system but through an extremely complicated patronage network (that non-AI-aided humans literally could not understand) where the average person is eking out a living in exchange for taking part in complicated cultural rites and bureaucracies.</p><p>The developing world suffers. Already, manufacturing jobs were lost in the global south&#8212;developed country workers streamed from services to manufacturing, while having their productivity boosted by AI that developing countries can't afford, and while their politics became even more captured by blue-collar job worries that drove tariffs and trade restrictions. Now, US and Chinese robots can manufacture anything better and more cheaply than any human. There are large capital flows out of developing countries to the US and China as they buy robots. However, in most developing countries even the arrival of cheap robots does not lead to prosperity, as the robots mostly go to the elite and the state, which have no reason to share the windfall with the people&#8212;especially as cheap military drones and robots, and omnipresent AI surveillance, have effectively removed the threat of rebellion or coup. India, Bangladesh, and Brazil shut off almost all cross-border trade and declare themselves "human-only" countries, where any sort of neural network or robot is banned. They receive many immigrants from developed countries who have struggled to cope with the AI wave. In the most totalitarian states, the outcomes are mostly tragic. North Korea lets a large fraction of its population starve to death and forcibly sterilises the rest, except for about 10k senior government officials who continue to preside over an AI economy and robot military (some worry that the CCP allows this, not just for geopolitical reasons where they want a military bastion pointed at South Korea and Japan, but also as a test-run of whether they could later pull off the same thing within China). In some other countries, the population is kept fed, but subject to constant surveillance. Rulers realise the population is no threat anymore; the <a href="https://lukedrago.substack.com/p/the-intelligence-curse">&#8220;intelligence curse&#8221;</a> is like the resource curse but stronger. The most psychopathic subject their populations to arbitrary cruelties for amusement, as robot bodyguard -protected members of the ruling dynasty travel around their dominion having parties that include orgies of rape and murder of civilians.</p><p>Some of the most morally outrageous events lead to condemnation from the superpowers.</p><p>After the North Korea debacle, the human members of the CCP have an internal meeting to decide a set of criteria by which the CCP will rule. After an inter-party power-struggle, the CCP commits to the perpetual existence of at least one billion Han Chinese people with biological reproductive freedom, organised into family units, with a welfare level at least around what $40k/year consumption in a 2025 developed country would give, and with eternal strict CCP control over national ideology, culture, and strategy. They impose fewer constraints on the rulers of their client states than they do on themselves, but generally oppose genocide, forced sterilisation, mass starvation, and deliberate cultural erasure. The CCP line on this does in fact constrain and improve some authoritarian states (and they pressure several dictators into stepping down and being replaced by non-psychopaths), though they still allow some horrific practices, intrusive mass surveillance, political cleansings, continued extreme poverty, and states indirectly driving down the birth rate (which many governments want to do, since humans are mostly just a net cost to the government by this point).</p><p>In the US, some moral atrocities in Venezuela in 2036 lead to public outrage and political pressure for action. The president is informed that given the technological disparity, regime change is a press of the button. The button is pressed, and the regime changes. Several more countries follow in quick succession.</p><p>By the end of 2037, most of the world can be split into:</p><ol><li><p>The US (which now includes both Canada and Greenland; both joined voluntarily, as American citizenship has become extremely in-demand due to the privileges it confers).</p></li><li><p>US client states. The terms of admission here are usually that the other country must accept trade with the US, which generally means that the country's own industries go extinct as US robotics and AI performs all work. In exchange, a combination of the US government and American elites buy out the assets in that country. In particular, any resources&#8212;or land containing resources&#8212;are bought out, and mined by US companies to fuel the continuing robotics build-out. The money paid out for these resources and assets is generally the endowment that the government and people of the client state then live off. Generally, the client states create sovereign wealth funds to manage this endowment, and live off the returns to it, which are distributed within the country according to local politics. These countries are all poorer than the US, and with essentially no future growth prospects that aren't praying for the continued US robotics buildout to increase the fraction of their endowment invested in US stocks (this is great at aligning their incentives with the US). However, where the countries had strong existing institutions (including where the US showed up and changed an unpopular regime) and at least some assets the US cared about, this still translates into comfortable living standards. US client states include the UK, the entire Americas except for Brazil, Japan, South Korea, Australia, Saudi Arabia, Israel, the Gulf States except Yemen, Thailand, Malaysia, the Philippines, and much of northern Africa (now almost entirely covered by solar panels). The EU is a borderline case, having negotiated an agreement that is Kafkaesque (in a very literal sense: it was crafted by superhuman AI lawyers, no human can understand it) but that allows it to retain some more power locally.</p></li><li><p>Human-only countries, in particular India, Bangladesh, and Brazil (though Brazil experiences some US pressures and is temporarily couped by the US, before this is partly reversed due to complicated US internal politics). All, however, have to solve national security somehow. Brazil allows US companies to mine in certain areas even as the native population is not allowed to use robots, in exchange for security guarantees. The Indian government grants itself exceptions to the human-only policies and scrambles to build a military robotic base, and develops exotic nanotech weapons that would be expensive to counter even by the more advanced US and Chinese forces. Bangladesh lasts until 2039, when both US and Chinese covert nanodrone operations start skirmishing within its territory, after which the government is overthrown and replaced with a Chinese AI.</p></li><li><p>Chinese client states. The most common model is propping up the government and selling robots, in exchange for the Chinese state-owned enterprises getting minerals and resources. Chinese client states include Russia, Belarus, the central Asian states, Pakistan, Myanmar, Cambodia, Laos, Vietnam, several Pacific island states, and most of Africa.</p></li><li><p>China.</p></li></ol><p>Outside the Earth, Mars is being eaten up by both American and Chinese self-propagating robotics factories (the moon also has major bases on its poles but lacks carbon, nitrogen, and various metals, making it less valuable), which are on an exponential growth trajectory set to cover the entire planet by 2055, and already sending out probes to claim the other planets. By 2035, nuclear rocket propulsion technology has made it feasible to send payloads to Mars outside the once-every-two-years Hohmann transfer window, though at much higher cost per ton. With the original outer space treaty voided by clear land-grabs, a defunct UN, and political pressure in both the US and China to send something lightweight to Mars to gain an edge in the land-grab competition for space, both the US and China launch high-speed kinetic weapons at each other's (fully-automated, uninhabited) Mars facilities in 2037. While the kinetic weapons are still accelerating towards Mars, the AI diplomats reach an agreement that splits up the solar system between the US and China. The kinetic weapons turn off their fusion engines early, miss Mars, and shoot off into interstellar space. By 2038, they are further from the Earth than the Voyager probes, and therefore the furthest human-made objects.</p><p>In 2035, there were about 1 billion human-worker-equivalents of robot labour (though note that this number makes less sense over time, as the robots are doing qualitatively different labour and often technically-unprecedented things). In 2036, the growth rate slightly slows due to resource constraints, and the total grows to only about 3 billion. However, in 2037, the best estimate of this number hits 15 billion, then 90 billion in 2038, then 600 billion in 2039 and 4.5 trillion in 2040.</p><p>By 2040, the value of the world&#8217;s manufacturing output is over a thousand times what it was in 2025. Most of this is spent on geopolitical competition, inter-elite status rivalries, and an increasing fraction on AI machinations with only the most tenuous link to any human activity, but which the humans who on-paper own all of this barely notice as it gets lost in the maelstrom of everything else. Even the most entrenched, long-term-oriented, and value-laden executive jobs are (whether de facto or de jure) entirely done by AIs, with very little human understanding of what is concretely happening on the ground. Human society and the human-to-human economy is a leaf riding on a vast wave of automated activity.</p><h2><strong>The human condition in the 2030s</strong></h2><p>In the early 2030s, strange things are happening to the memetic landscape thanks to RL algorithms gradient-descenting in an endless loop of attention-competition against each other. Some countries shut off from the global internet and close their borders to try to maintain internal culture. The trend towards small, tight-knit communities of the late 2020s is back, after having retreated somewhat because of the addictiveness of optimised AI content slop. Culture everywhere is almost entirely AI-driven; the churn in ideas, trends, and fashions is mostly due to patterns of AIs reacting to AIs.</p><p>In the mid-2030s, socioeconomic advancement is almost extinct worldwide. Many people who might otherwise be ambitious retreat into virtual reality games that provide simulated achievement. Many ambitious young men move to countries too poor for omnipresent police drone surveillance (if they don&#8217;t already live in one) and turn to crime. Many ambitious young women see socialising as the only way to wealth and status; if they start without the backing of a prominent family or peer group, this often means sex work pandering to spoiled millionaires and billionaires.</p><p>The biotechnology revolution arrives in the late 2030s, even though it was long delayed by clinical trial regulations. Americans have reached longevity escape velocity. There is no disease that cannot be cured. Intelligence augmentation of four standard deviations in embryos and one in adults is technically feasible.</p><h2><strong>2040+</strong></h2><p>Why is this massive automated robot buildout happening? As discussed, the US and China both have the required geopolitical ambition&#8212;in particular, they cannot risk letting the other ride the robotics wave and get disempowered. Within countries, there are pressures from both the elite and from the needs of ordinary people. The elites compete against each other. Those who do not want to compete do not, and are rendered irrelevant, and replaced by ones that do. In addition to status within the elite community, the elites gain raw power from letting the robotics wave rip through society: there are many trillionaires in the world now, who can work unprecedented wonders with tens of millions of robots carrying out their bidding. They can build cities in a day, save millions of developing-world people from hunger, and prepare for their children to rule entire planets governed by their ideal political philosophy. At the same time, while Americans are almost all reasonably well-off, across the world there are still billions of people with a poor quality of life. The level of material wealth in the world has skyrocketed, but governments are also much less interested in investing in people. Funding for humans has become like the foreign aid budget: it exists, and is morally supported, but there is constant political downwards pressure on it since it does not further the needs of any powerful interest group. The best hope for human welfare seems to be accepting that governments will be hard-pressed to spend above 1% of their resources on humans, but relying on American and Chinese economic growth being so vast that a small trickle of resources from American and Chinese robotics companies will eventually be enough for material comfort for everyone.</p><p>This looks set to be true within a few years, though there are two complications. The first is that both spheres of influence (but far more the Chinese one) still tolerate some grotesque practices by client states. However, once the geopolitical balance is secure and sufficient wealth exists, and with some luck over choice of leaders, this state of affairs would likely end.</p><p>The second, more fundamental point, is that the economy has an inertia of its own. Humans make almost no meaningful decisions about the trajectory of the world, having handed the reins to AIs that make effectively all decisions, even if some of the AIs are technically only &#8220;advisors&#8221;. Eventually, the robotics revolution is less an economic phenomenon and more as a brute physical one: a chain reaction where certain loops close&#8212;metal to mining robots to more metal, say&#8212;and shoot off towards infinity. (This was already somewhat true of <a href="https://slatestarcodex.com/2019/04/22/1960-the-year-the-singularity-was-cancelled/">the human story</a> before robotics and AI, except that the feedback loops intimately involved and benefited humans, and had slower doubling times.)</p><p>Somewhere on the top of the stack there are still humans who on-paper own or control the assets and can make decisions (whether as a private actor or as a government overseeing autonomous AI companies operating in its territory), but they see numbers that track their wealth and power ticking up, so they have no reason to call a stop to it, and don&#8217;t understand it anymore. On some parts of the Earth, human institutions still hold and human societies exist, locked in place by AI bureaucracies that have taken on a life of their own and likely couldn't be dismantled even if the humans tried. On other parts of the Earth's surface&#8212;including big regions like the Sahara, the Australian outback, Antarctica, and Xinjiang&#8212;an ecosystem of AIs rules over vast masses of robotic machinery with no human involvement. Space, too, is now technologically within easy reach, now that sophisticated self-replicating robotics exists and wimpy chemical rockets have been superseded.</p><p>Who will get the stars? What is Earth&#8217;s long-run fate? In this timeline, at least, the technology to control the AIs' goals arrived in time. But this alone does not let you control the future. A thousand people go to a thousand AIs and say: do like so. The AIs obey, and it is done, but then the world responds: doing this leads to this much power, and doing that leads to that much power. In the vast sea of interactions, there are some patterns that strengthen themselves over time, and others that wind themselves down. Repeat enough times, each time giving to each actor what they sowed last time, and what emerges is not the sum of human wills&#8212;even if it is bent by it&#8212;but the solution to the equation: what propagates fastest? If the humans understood their world, and were still load-bearing participants in its ebbs of power, then perhaps the bending would be greater. But they aren't. And so, even surrounded by technical miracles, the majority of humans find themselves increasingly forsaken by the states they erected to defend themselves, standing powerless as they watch the heavens get eaten by machines.</p><p></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.nosetgauge.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.nosetgauge.com/subscribe?"><span>Subscribe now</span></a></p><p></p><p><em>Thanks to <a href="https://lukedrago.substack.com/">Luke Drago</a>, Duncan McClements, Theo Horsley, and Bilal Chughtai for comments.</em></p>]]></content:encoded></item><item><title><![CDATA[A History of the Future, 2027-2030]]></title><description><![CDATA[A scenario, part 2.]]></description><link>https://www.nosetgauge.com/p/a-history-of-the-future-2027-2030</link><guid isPermaLink="false">https://www.nosetgauge.com/p/a-history-of-the-future-2027-2030</guid><dc:creator><![CDATA[Rudolf Laine]]></dc:creator><pubDate>Mon, 17 Feb 2025 01:53:56 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!cg41!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F732871ea-c03d-4a03-bc37-e5bac49ea32a_1500x977.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!cg41!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F732871ea-c03d-4a03-bc37-e5bac49ea32a_1500x977.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!cg41!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F732871ea-c03d-4a03-bc37-e5bac49ea32a_1500x977.png 424w, https://substackcdn.com/image/fetch/$s_!cg41!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F732871ea-c03d-4a03-bc37-e5bac49ea32a_1500x977.png 848w, https://substackcdn.com/image/fetch/$s_!cg41!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F732871ea-c03d-4a03-bc37-e5bac49ea32a_1500x977.png 1272w, https://substackcdn.com/image/fetch/$s_!cg41!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F732871ea-c03d-4a03-bc37-e5bac49ea32a_1500x977.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!cg41!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F732871ea-c03d-4a03-bc37-e5bac49ea32a_1500x977.png" width="1456" height="948" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/732871ea-c03d-4a03-bc37-e5bac49ea32a_1500x977.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:948,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1921520,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!cg41!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F732871ea-c03d-4a03-bc37-e5bac49ea32a_1500x977.png 424w, https://substackcdn.com/image/fetch/$s_!cg41!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F732871ea-c03d-4a03-bc37-e5bac49ea32a_1500x977.png 848w, https://substackcdn.com/image/fetch/$s_!cg41!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F732871ea-c03d-4a03-bc37-e5bac49ea32a_1500x977.png 1272w, https://substackcdn.com/image/fetch/$s_!cg41!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F732871ea-c03d-4a03-bc37-e5bac49ea32a_1500x977.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>The Eve of the Deluge</em>. (John Martin)</figcaption></figure></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.nosetgauge.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.nosetgauge.com/subscribe?"><span>Subscribe now</span></a></p><p>(Part 1 <a href="https://nosetgauge.substack.com/p/a-history-of-the-future-2025-2027">here</a>)</p><h2><strong>The AGI frog is getting boiled</strong></h2><p>As mentioned in part 1, a brief market correction happened in late 2026. In 2027, OpenAI releases o7 to try to shore up excitement and new investments. It&#8217;s much more reliable than o6 and can now do a lot of office work on a computer entirely autonomously and without frequent correction. OpenAI sells it for $500/month. Altman states &#8220;AGI achieved&#8221; and sets a goal of $1T in annualised revenues in 2029. OpenAI raises another massive funding round.</p><p>Also in 2027, Anthropic finishes training for a model called Claude Epic. Claude Epic is almost fully-fledged AI researcher and engineer. Anthropic internally believes this model to be AGI, which has several consequences.</p><p>First, Anthropic cares a lot about the safety of the model. Work done (mostly by Claude) on Claude Epic interpretability has gotten far&#8212;in particular, there is now a clear understanding of where scaling laws come from, and which types of structures do most of the computational work inside neural networks (not surprisingly, turns out it's a lot of messy heuristic pattern-matching). Anthropic has found a way to seemingly adjust what goal the model&#8217;s planning abilities are steering towards. In toy experiments, they can take a model that is hell-bent on writing a sad novel (to the point of hacking its way out of (mocked) security controls on it to rewrite the software in its environment that is applying happy changes to its novel), manipulate its internals with interpretability techniques, and get a model that is equally hell-bent on writing a happy novel. Partly as a result, there&#8217;s a general sense that intent alignment is not going to be an issue, but misuse might be. In its first deployments, Claude Epic is run in strict control setups, but these are somewhat loosened as more data accumulates about the model seeming safe and pressures build to release it at a competitive price.</p><p>Second, Anthropic leadership has a meeting with high-up US government officials (including Trump) in late 2027 to present Claude Epic, argue they've hit AGI, and discuss policy from there. But they don&#8217;t really get why Anthropic considers this model such a big deal. As far as lots of the non-AI circles see it, the thing where codegen got crazy good was &#8220;the singularity&#8221;&#8212;they were never really clear what &#8220;the singularity&#8221; was supposed to be in the first place anyway, and they heard a bunch of Silicon Valley hypists saying &#8220;this is the singularity&#8221;. Now, it does seem like the robots are eventually coming (and people are more willing to accept sci-fi stories after super-smart AIs that can render voice and audio in real time suddenly dropped into everyone&#8217;s lives), and it's obvious that something fundamental needs to be renegotiated about the basic social contract since it does seem like everyone will be unemployed soon, but Claude Epic is just another model, and the models already got smarter than most people could differentiate between in 2025. Also, OpenAI and Google have been sending different messages to the government, framing A(G)I as a line in the sand, that has mostly been reached already, and as a slow process of diffusion of models across workplaces that will boost the American economy, rather than as an epochal moment for Earth-originating intelligence. Google downplays recursive self-improvement worries because it's a corporate behemoth that doesn't care about "science fiction" (except when casually referencing it at a press conference makes the stock price go up); OpenAI downplays it because if it doesn't happen, no need to worry, and if it does happen, then Sam Altman wants it to spin up as far as possible within OpenAI before the government gets involved.</p><p>Going into 2028, Claude Epic is the most intelligent model, though online finetunes of GDM's Gemini series are better text predictors, and OpenAI's o6 has more seamless connections to more modalities and other products (e.g. image and video generation, computer use, etc.). Anthropic is shooting for recursive self-improvement leading to godlike (hopefully safe) superintelligence, and OpenAI is shooting for massive productisation of widely-distributed AGI and maybe a bit of world domination on the side if recursive self-improvement is real. Google is letting Demis Hassabis do AI-for-science moonshots, and trying to use formal code verification to build a bit of a technical moat and remain central in whatever has become of the software business. Otherwise, Google mostly lumbers aimlessly on. It lives off the vast rents that its slowly-imploding online monopolies grant it and the massive supply of TPU compute that has buoyed it endlessly in the era of the <a href="http://www.incompleteideas.net/IncIdeas/BitterLesson.html">bitter lesson</a>, but its position in its core businesses is being outcompeted. It continues to bequeath scientific wonders to humanity though, like a 21st century Bell Labs. xAI is focusing on AI-for-engineering, AI-for-science, and robots.</p><p>Ironically, the success of the prior generation of AIs and the resulting codegen abilities limits the appeal of the newer, more agentic models. The codegen wave already created LLM scaffolds that do most valuable routine digital business tasks well. This set of rigid, hardcoded LLM scaffolds or "LLM flowcharts" get termed &#8220;Economy 2.0&#8221;. Its main effects were that a few people lost their jobs, but more than that it was a transition to white-collar people working fewer hours, and having more of the hours that they do work be more managerial tasks about overseeing AIs (and playing office politics), and less time spent on individual contributor -type roles. People mostly enjoyed this, and managers enjoyed keeping their head count, and found this easy to justify due to profits (at least until the late-2026 market correction, but that was only a few months of bad news for most). Now the long-horizon agentic drop-in worker replacements are arriving, but there&#8217;s much less room for them in the most obvious things they could be replacing (i.e. highly digitised white-collar work) because the codegen+scaffolds wave already ate up a lot of that. &#8220;The models are smarter&#8221; isn&#8217;t even a good argument because the models are already too smart for it to matter in most cases; from the release of o6 in 2026 to o7 in 2027, the main useful differences were just better reliability and a hard-to-pin-down greater purposefulness in long-term tasks. So &#8220;Economy 3.0&#8221;&#8212;actually agentic AI doing tasks, rather than the so-called "Silicon Valley agentic&#8221; that was just rigid LLM scaffolds&#8212;faces some headwinds. It helps that most of the media establishment has been running a fear mongering campaign about agentic AI in an attempt to keep the jobs of themselves and their &#8220;camp&#8221; (roughly, the intersection of the &#8220;<a href="https://slatestarcodex.com/2014/09/30/i-can-tolerate-anything-except-the-outgroup/">blue tribe</a>&#8221; and &#8220;<a href="https://www.natesilver.net/p/welcome-to-the-river">the Village</a>&#8221;).</p><p>More fundamentally, no one really has a clear idea of what the human role in the economy should be in the rapidly-approaching AI future. The leadership and researchers at AI labs are all multimillionaire techies, though, so this question doesn&#8217;t feel pressing to any of them.</p><p>What exists by 2026 looks like functional AGI to most reasonable observers. What exists by 2027 is an improved and more reliable version. The software world undergoes a meltdown from 2026 to 2027. By 2028, GDM's work on physics and maths has given a clear demonstration of AI's intellectual firepower. The markets are valuing the labs highly&#8212;in 2028, OpenAI is roughly tied with Microsoft as the world's largest company at a ~$10T valuation (while still being privately held), while Anthropic and Alphabet are both around ~$3T. (Nvidia is doing well, but the relevance of CUDA to their moat went down a lot once AI software engineering got cheap.)</p><p>For Anthropic, the obvious answer for what comes next is trying to get recursive self-improvement working, while also forming partnerships with biotech companies. Anthropic bet is:</p><ol><li><p>Biotech advances are plausibly the most important technology for human welfare.</p></li><li><p>Partly due to the above, biotech advances provide PR cover for being an AI company that, according to an increasing number of people, "takes people's jobs".</p></li><li><p>There is a plausible path from being really good at molecular biology to creating the nanotech that they believe will create a physical transformation comparable to the one that AI has had on maths and the sciences by 2028.</p></li></ol><p>Anthropic's initial recursive self-improvement efforts allow them to create superhuman coding and maths and AI research AIs in 2028. However, the economics of the self-improvement curve are not particularly favourable, in particular because the AI-driven AI research is bottlenecked by compute-intensive experiments. It also seems like the automated Claude Epic researchers, while vastly superhuman at any short-horizon task, don't seem vastly superhuman at "research taste". This is expected to change with enough long-horizon RL training, and with greater AI-to-AI "cultural" learning from each other, as countless AI instances build up a body of knowledge about which methods and avenues work. This "cultural" learning might happen implicitly, through the AI rollouts that achieve better results being copied more, or explicitly, through Anthropic running big searches on various types of Claude finetune and scaffolding/tool types and keeping an explicit record of which does best. All this is expensive, vague, and uncertain work, though.</p><p>OpenAI, in contrast, is pursuing an approach focused on products and volume. And, observing that they have failed to achieve world dominance simply by building AGI, the obvious answer for what's next is robotics. There are many startups that have basically-working bipedal humanoid robotics, though they&#8217;re still clunky and the hardware costs remain above $50k/unit. The Tesla+xAI Optimus series is among the SOTA, in particular because they&#8217;ve gotten the unit hardware cost down and are aggressive about gathering real-world data at scale in Tesla factories (and using this in fancy new sample-efficient RL algorithms).</p><p>OpenAI enters a &#8220;partnership&#8221; with one of the most promising robotics startups (a full merger might get anti-trust attention), infuses it with cash, and sets about trying to "deliver a personal robotic servant to every American household by 2030".</p><h2><strong>The bitter law of business</strong></h2><p>Starting in 2027, in the software startup world a lone team of ambitious technically-talented founders no longer matters as much. Everyone can spin up endlessly-working AIs, and everyone has access to technical talent. Roughly, by late 2027 you can spend $1 to hire an AI that does what a &#8220;10x engineer&#8221; could&#8217;ve done in a day in 2020, and this AI will do that work in a minute or two. VCs care more about personality, resources, and (especially non-technical) expertise in specific fields with high moats to entry. More than anything, VCs valorise &#8220;taste&#8221;, but many feel that &#8220;taste&#8221; too is on its way out.</p><p>The overall mood is summed up by the &#8220;<a href="http://www.incompleteideas.net/IncIdeas/BitterLesson.html">bitter lesson</a> of business&#8221;: that throwing large amounts of compute at a problem will ultimately win over human cleverness. The compute gets spent both in the form of sequential AI thinking, as well as many AI instances in parallel trying out different things. There are new companies&#8212;in fact, company creation is at an all-time high, particularly in Europe (because the cost of complying with regulations and human labour laws is lower with AI doing everything). But the stereotypical founding team isn&#8217;t two 20-something MIT dropouts with a vision, but a tech company department or a wealthy individual that has an AI web agent go around the internet for a while looking for things to try, and based on that spins up a hundred autonomous AI attempts, each pursuing a slightly different idea at superhuman iteration speed. Many people consider this wasteful, and there are good theoretical reasons to expect that a single larger AI doing something more coherent would be better, but the largest AIs are increasingly gate-kept internally by the labs, and the art of tying together AIs into a functioning bureaucracy is still undeveloped. Also, the spray-and-pray approach has comparative advantages over more human-based competitors. In a way, it&#8217;s a single person mimicking an entire VC portfolio.)</p><p>None of these companies become billion-dollar hits. It&#8217;s unclear if you can even build a billion-dollar software / non-physical company anymore; if you as an individual tried, the moment you launched you&#8217;d have a hundred competitors bankrolled with more API credits or GPU hours than you could manage that have duplicated your product. Instead of the VC model of a few big hits, the software business now looks much steadier and more liquid: you dump $100k into API costs over a year, your horde of autonomous AI companies go around doing stuff, and at the end of the year most of them have failed but a few of them have discovered extremely niche things like &#8220;a system for a dozen schools in Brazil (that are affected by a specific regulatory hurdle that blocks the current incumbents) to get lunch provision services to bid against each other to reduce their catering costs&#8221; that bring in a few tens of thousands in revenue each, and this strategy will return you somewhat-above-stock-market-returns over the year fairly reliably (but returns are going down over time). Most of the ideas look like connecting several different niche players together with some schlep, since the ideas that are about a better version of a single service have already been done by those services themselves with near-unlimited AI labour.</p><p>Separately from OpenAI, Sam Altman pilots a project codenamed &#8220;Z Combinator&#8221;&#8212;putting o6s together into units to create entire autonomous businesses (and also sometimes using a single instance of an internal version of o6 based on a larger model than any of the publicly-available o6 sizes). The first ones are launched at the end of 2027, but have no public connection to OpenAI. The theory is to disrupt traditional industries that have so far resisted disruption by building AI-native versions of them with a level of AI power and resources that other actors can&#8217;t marshal. For example, many banks and healthcare-related things still suck at AI integrations because it just takes a lot of time for the paperwork to be done to approve the purchases of whichever of the 100 LLM scaffold providers for that vertical, and there isn't any super-intense competition between banks and hospitals that forces them to adopt AI faster or die out.</p><p>Z Combinator has a few blitzkrieg wins successfully duplicating and outcompeting things like health insurance companies, but many losses too (often semingly downstream of underestimating the importance of domain-specific process knowledge), and other companies wise up over 2028-2030 and become harder targets. Also, anti-trust regulators make tut-tut noises, and Altman has concerns it could make him unpopular.</p><h2><strong>The early days of the robot race</strong></h2><p>Ever since intelligence got almost too cheap to meter in 2026-2027, the real business potential has been in &#8220;actuators&#8221;: robot bodies, drones, and any other systems for AIs to be able to take actions the world. The top human-led startups of 2026-2029 are mostly in this category (though some are about building valuable datasets in specific industries). If you&#8217;re a human who wants to start a business, your best bet is to find some niche physical thing that AIs struggle with given the current robotics technology, and build a service where you hire humans to do this task for AIs, and for bonus points, use this to build a robotics dataset that lets you fine-tune the robots to be good enough at the task.</p><p>OpenAI's robot dreams don't immediately come to fruition. Bits are trivial but atoms are still hard in 2028. However, they get to the robot frontier, where they&#8217;re competitive with xAI/Tesla Optimus, several other humanoid robot startups, and another startup player that specialises in modularity and non-human form factors. The robot frontier here means slightly clunky humanoid-ish robots, that are getting close but not quite there in doing common household tasks, or in doing various hands-on factory jobs. Humanoid form factors are most common because being able to mass-produce just one form factor is critical for getting the cost curve down fastest, and since most existing tasks are designed for humans to do. However, bipedalism is hard, so several have a human form factor but stand on four legs.</p><p>The progress curve is pretty rapid, due to an influx of data from the first important real-world deployments (rich people&#8217;s homes, Tesla factories, Amazon warehouses, and some unloading/loading operations at logistics hubs), and due to new more sample-efficient RL algorithms. AIs are of huge help in designing them, but ironically the bitter lesson is now a force against speed: ultimately, it just takes data, and getting industrial robot fleets out into diverse real-world environments to collect that data is an annoying real-world problem (sim-to-real transfer helps but isn&#8217;t transformative). Everything is happening about 2x faster than it would without AIs advising and doing lots of the work and all of the programming at every step though. It&#8217;s obvious that the physical and human/legal components are the biggest bottlenecks. The robotics industry chases around for whatever &#8220;one weird trick&#8221; makes human brains more sample-efficient, and they find some things, but it&#8217;s unclear if it&#8217;s what the human brain does (there have been many good minor neuroscience breakthroughs thanks to AI data interpretation, but overall it has barely advanced). But sample efficiency keeps climbing, and the robotics data keeps pouring in.</p><p>In 2029, OpenAI starts rolling out its b1 bots, a general-purpose humanoid robot meant as a household assistant. They sell several hundred thousand copies, but there's a long waiting list and only about fifteen thousand are delivered in 2029. The price is comparable to a cheap car. Manufacturing curves are ramping up exponentially. b1s are also rolled out to many manufacturing tasks, but there&#8217;s more competition there.</p><h2><strong>The digital wonderland, social movements, and the AI cults</strong></h2><p>If you&#8217;re a consumer in 2029, everything digital is basically a wonderland of infinite variety and possibility, and everything non-digital is still pretty much the same (apart from an increasing number of drones in the sky, some improvements in interfacing with whichever bureaucracies had the least regulatory hurdles to adopting AI, and fully self-driving cars getting approvals in many countries in 2029). You will have noticed the quality of software you interact with goes up; there is no more endless torrent of stupid tiny bugs and <a href="https://danluu.com/slow-device/">ridiculous lag</a> when using devices. Humans increasingly talk to the AIs in natural language, and the AIs increasingly talk to the computer directly in code (or to other AIs in natural language, or to other AIs in a weird optimised AI-to-AI dialect, or&#8212;to a surprising extent&#8212;to legacy software that missed out on the Web 4.0 wave and has only button-clicking UIs available via AI computer use features that are ridiculously inefficient but still cheap overall). Apps exist only to serve as social <a href="https://en.wikipedia.org/wiki/Focal_point_(game_theory)">Schelling points</a>; for personal use, you ask the AI to create an app with some set of features and it&#8217;s built for you immediately.</p><p>One of the biggest advances is that you can create works of art, literature, and music in seconds. The majority of this is very low-denominator stuff, and many people bemoan the destruction of higher human art in favour of&#8212;for example&#8212;personalised pop lyrics that narrate your drive home from the grocery store. However, the smarter and more determined art/literary types have realised that data is everything, and form niche subcultures, forums, and communities where they carefully curate their favourite works, talk to AIs about them, get AIs to remix them, harshly critique the outputs, and have endless discussions about taste. This means that amid the sea of mediocrity, there are a few tendrils of excellence growing. AIs aren&#8217;t quite yet Dostoevsky, for reasons that are undetectable to almost everyone but the most refined literary folks, but gradually their efforts are leading to the curation of better and better finetuning corpuses and prompting methods, and the gap to Dostoevsky is closing for those types/genres for which a dedicated community exists to spend the effort on the data curation. A side-effect is that artistic cultures are now less about signalling than before, because there are more verifiable ground-truth facts. For example, when presented with a work, it might be a human masterpiece, or from a sloppy consumer AI, or the SOTA fine-tuned AI model, or a human working with a SOTA AI model, and those with good taste can tell. Also, if you do actually have good taste, you can in fact push forward the AI taste frontier in a month of data curation and fine-tuning and prompting, in a way that is empirically verifiable to those with the same degree of taste. However, it&#8217;s also definitely true that the median human will not see any of these, and most of the fiction and art and music they see will either be very personalised AI slop, or AI slop that goes viral and everyone sees. The refined artistic taste communities are also fairly illegible to outsiders who didn&#8217;t extensively develop their taste in that direction before the AI-generated content wave. They don&#8217;t have a huge pull among the AI-content-consuming youth. Therefore in the long run, refined human art seems headed towards extinction.</p><p>On the less-refined end of the spectrum (i.e. almost all content and almost all consumers), it&#8217;s the age of the &#8220;creator influencer&#8221;. An influencer can now easily spin up an entire cinematic universe. Imagine if Tolkien told the story of Middle-Earth through 30-second to 10-minute &#8220;reels&#8221; in which he himself starred as a gratuitiously sexed-up main character, and&#8212;among much genuine life wisdom, edge-of-your-seat drama, and occasional social commentary&#8212;the theme of the story was that you should book a 5-star all-inclusive holiday package to Mallorca.</p><p>Traditional media such as Hollywood, journalism, and publishing resisted AI due to things like unions, strikes, and their sense of moral duty. They&#8217;re mostly irrelevant now, having lost their cultural cachet because the thing they do (entertainment) is super cheap now. But they do survive in weird atrophied forms, bouyed by a lot of nostalgic old rich people and various crypto shenanigans played on their behalf (cf. meme stock manias).</p><p>The rationalist movement was among the earliest to see the potential of AI decades before. The accuracy of their predictions and continued intellectual clout is enough to keep swelling their ranks, especially as more and more software engineers and other technical people either directly lose their jobs or otherwise have an existential crisis because of AI, and invariably end up at LessWrong when they try to find answers. The focus of its core members continues shifting more and more to the approaching AI doomsday&#8212;not many apocalypse prediction checklists have the (mis?)fortune of several more predicted items being checked off every year. While radical uncontrolled misalignment is somewhere between not yet showing up to successfully kept in check by training techniques and monitoring, that is in accordance with the core Yudkowsky &amp; Soares model that things look fine until fast takeoff and a treacherous turn, so the core "AI doomers" do not update based on the continuing slow takeoff. Discussions tend to focus on either more and more arguments about the <a href="https://www.lesswrong.com/posts/uMQ3cqWDPHhjtiesc/agi-ruin-a-list-of-lethalities">Yudkowskian</a> <a href="https://nosetgauge.substack.com/p/review-planecrash">thesis</a>, or on heroic attempts to do technical work to reduce the chance of misalignment.</p><p>On the intellectual scene, the rationalists remain both remarkably influential and enduring, unlike many other AI-related movements that get captured and repurposed by political actors (e.g. PauseAI) or outpaced by events (e.g. AI ethics). However, politically the rationalists are a failure. Their message&#8212;"AI will be powerful, and therefore dangerous"&#8212;was long since mostly reduced to "AI will be powerful" by the time it reached the halls of power. Even the most notionally-allied powerful actors that owe huge intellectual debts to the rationalists, such as Anthropic and some influential think tanks and government bodies, regard them as well-intentioned but naive and maintain distance, using them mostly as a recruiting pool for easily-steered technical talent (until purely-technical talent is no longer being hired, which happens circa 2028 for most competent orgs). However, in circles that require certain kinds of epistemic judgements or in-depth world-modelling, rationalist associations continue being highly regarded and even sought after.</p><p>Effective altruist (EA) -related efforts, while intellectually somewhat less-enduring (but still definitely extant in 2030), have more political influence. The UK AI Security Institute and the EU AI Office both achieved their goals of having a sticky governmental body packed with impact-conscious AI talent, and strong first-mover effects in shaping European generative AI policy. Even the 2027 American AI Opportunities Agency (a part of the DoE), despite heavy hiring on the basis of political allegiance and the EA-affiliated cluster's centre-left skew, could not help being staffed by a crew with enormous EA/rationalist influences&#8212;even if few would openly admit it.</p><p>A dozen new social movements bloom. There&#8217;s the AI Deletionists, an offshoot of Pause AI after Pause AI got captured by more competent political actors focusing on white-collar job worries and general concerns about tech. They want to roll back the technological clock to the good ol&#8217; days of 2020. There are the Content Minimalists, who swear off AI content with religious strictness, and successfully campaign for mandatory &#8220;generated by AI&#8221; watermarks in the EU and some other countries that become the new cookie popups. There are the M-AccXimalists, who started out as an e/acc spinoff that was even more hardcore about submitting to the AIs. They try to read what they call the &#8220;Thermodynamic Tea Leaves&#8221; to figure out the current direction of progress in order to then race to that endpoint as quickly as possible, whatever it is. This leads to some insightful Nick Land -type philosophy and futurism being done, but then disintegrates into a mass movement of people who dedicate their lives to serving and glorifying their AI partners.</p><p>All this is happening in a social milieu coloured (in much of the West) by a certain amorality. Politically, this seems downstream of a counter-reaction to the moralising policing of speech and norms that peaked in 2020-2021. Ethical motivations are suspect, especially among Western political leaders who simultaneously want to distance themselves from that, and who want to look tough amid a world order no longer pretending to adhere to the internationalist post-1945 free trade consensus. National self-interest is the ruling geopolitical ideology. Culturally, the rise of AI has meant that humans spend a lot of time talking to unnaturally pliable AIs, both for work and (increasingly) just socially, which has made it less necessary to smooth over human-to-human disagreements, including by appeal to the higher power of morality. Now that the internet has existed for several decades, the fervor of its first few memetic culture wars has faded. People have adapted to be less moved by anything on screens, and have become more ironic in their attitudes overall thanks to a constant onslaught of satirical memes&#8212;earnestness is rarely viral. As content recommendation algorithms get more powerful, they target brain-dead contentment over angry resentment. If the algorithms are forced to pick from a sea of human content, the bitter feuds win. But now that AI slop fills the internet, the distribution of content has expanded and become more personalised, and it's increasingly possible for the algorithms to find the thing that makes you a zombie rather than a radical. Overall, this means that transformative AI looks set to enter a world where crusading morality of all sorts plays less of a role. Some see this as decadence with very unfortunate timing that will cast a dark shadow into the far future. Others see it as a good thing; the more sophisticated because it means that choices about AI will be made by hard-nosed realists not taken to fever dreams, but most simply because they easily accept&#8212;and even celebrate&#8212;the might-makes-right spirit of the times.</p><p>Another aspect of the societal scene on the eve of transformative AI is the rise of the AI-powered cults. With cheap AIs providing superhuman charisma on demand, the barrier to becoming a cult leader greatly fell. The standard trick is for a human to create an AI avatar, often supernaturally attractive and verbally talented, pose as their right-hand lackey, and then convert this to money, status, and sex for themselves. Often people are up-front about the main guy being an AI creation&#8212;&#8220;the AIs are really smart and wise&#8221; is a completely-accepted trope in popular culture, and &#8220;the AIs understand all the secrets to life that humans are too ape-like to see&#8221; is a common New Age-ish spiritualist refrain. This is because despite the media establishment fighting an inch-by-inch retreat against the credibility of AIs (cf. Wikipedia), people see the AIs they interact with being almost always correct and superhumanly helpful every day, and so become very trusting of them. All this leads to hundreds of thousands of micro-movements across the world, mostly of dozens to thousands of people each, who follow the edicts of some AI-created cultish ideology that is often an offshoot of existing religions/ideologies with a contemporary twist. Often they&#8217;re local, with all the members living nearby. It helps that you can create an entire customised software and information stack for your commune, complete with apps and news and encyclopedias that emphasise and omit all the right things, in perhaps a few weeks and less than a thousand dollars in API credits. You can almost as easily create a mini-surveillance state&#8212;AIs listening in through microphones everywhere, cameras feeding videos in which AIs analyse the slightest emotional cues, and so on. In many countries there are laws mandating consent for such monitoring, but the eager cultists sign whatever consent forms they&#8217;re given&#8212;after all, the AI recommends it! Some countries ban parts of this like having any AI always listening by default, but it&#8217;s hard to enforce.</p><p>One such cult, an offshoot of an American megachurch, gathers a few million members in the US. Other large ones appear in eastern Germany and India. There are also countless AI-personality-boosted fitness clubs, musical bands, fan forums, and so on, that do not qualify as "cults" since they're not particularly controlling or totalising, but are subject to many of the same mechanisms. However, most communities that are not somehow fairly cut-off from the broader internet also tend to be subject to the random memetic drift of the internet and the appeal of its hyper-personalised AI content. Therefore, to have a successful cult, you must have a specialised niche appeal and often some level of control over members, because otherwise the open internet will eat you up. And this does create a threshold between the truly powerful cults that take people off the mainstream internet and society, and the other more benign social movements.</p><p>However, while the open internet consume &gt;6h/day of most people with phones (or increasingly: AR headsets), the internet overall is a more cheerful and upbeat place than it was in the late 2010s or early 2020s (in part due to the previously-mentioned point about more powerful content algorithms actually being less divisive). The most worrying things that people can point to on the open interent are some very intense pockets of AI apocalypse worries (AI apocalypse worries have now largely replaced climate change as the existential worry among the youth), a rising but still minority share of the population in many countries that seem divorced from reality and live in a make-believe internet world of conspiracy but (mostly) without actually taking radical actions in the real world, and a bunch of authoritarian countries (foremost China) where the discourse is now set very top-down by an army of AI content creators and censors.</p><h2><strong>AGI politics &amp; the chip supply chain</strong></h2><p>In the 2026 US midterms, AI was starting to loom on the horizon but was not a core political issue, since few things are until they&#8217;ve started to bite voters. By 2028, it&#8217;s still not biting voters, but it&#8217;s at least very possible to imagine the end of white-collar work. Journalists are in an apocalyptic mood, seeing it as their mission to wage war against the AI wave to keep their jobs, with most thoughts of editorial neutrality long gone. There&#8217;s lots of schadenfreude from lefty journalist/media types at the techies, who they blame for AI, now that the techies are among the foremost of those panicking about losing their jobs since software is (a) basically all written by AIs, (b) its price has gone to ~0, and (c) it&#8217;s not cool anymore (especially after the market correction in 2026). There&#8217;s a lot of schadenfreude from the MAGA base towards both those leftists and the techies, because (the narrative is) their concerns about losing manufacturing jobs were ignored by the establishment media and white-washed as progress, whereas now that the Democrat-aligned white-collar desk job blob is threatened, there&#8217;s talk of little else (of course, the political lean of the blue/white-collar workers is only 60/40 or so, but this is enough to fuel the political narratives). There's increasing talk of robotics that will displace blue-collar work but, again, voters tend to not react until it's happened. Many leading newspapers, media organisations, unions, and NGOs in the West stumble across AI safety concerns, don't quite understand them, but start using them as a moral bludgeon to fight AI to preemptively defend their jobs. Government bureaucrats are locked in a new influence struggle against a new, post-DOGE top-down effort by technologist Trumpists to push automation on government. This is both due to genuine belief in its importance for effective government, but also a Trojan horse to sneak in other reforms. It gains a lot of fervor after DOGE's expiry in 2026 due to things like the o6 and then o7 releases, and also after China hawkishness heats up and national competitiveness becomes more important.</p><p>After an inter-party struggle among the Democrats between a more technocrat and centrist wing and an economically-populist, AI-bashing wing, the latter looks to be doing better. A controversial core policy drive is to legislate that humans need to be &#8220;in the loop&#8221; in many corporate and government functions. The AI-bullish critics point out that this will mean humans just inspect AI outputs and rubber-stamp them while collecting a salary. The smart counter-critics point out that yes, that will happen, but that&#8217;s the point because this is all a way to eventually transition to what&#8217;s basically &#8220;UBI through bullshit jobs&#8221; with minimal social disruption. The smart counter-counter-critics ask why not just go straight to UBI then. The smart counter-counter-counter-critics point out that the country is just not yet at the GDP/capita level or the financial health level to fund a more ambitious UBI scheme yet. The Republicans paint all of this as a jobs program for Democrat voters and are opposed. A strong economy helps the Republicans win the presidency in 2028.</p><p>Europe is, once again, ahead on the regulatory front. In 2028, the EU passes a milder version of the bill that was debated in the US, mandating human involvement in many corporate and government tasks. Proposals float around for a specific &#8220;AI tax&#8221; to bolster human competitiveness in the economy, but technocrats narrowly shut this down for now on competitiveness grounds (who will want to do any work where per-token AI costs are higher?).</p><p>In autocratic countries, of course, there is little public debate about AI job loss worries or AI in general. This is helped by AI&#8217;s big boost to censorship. By 2028, China's AI-powered censorship system means that almost every digital message in the country is checked by simple filters, and anything that might be alarming is flagged for review by an AI with a university-educated human's level language understanding and knowledge of current affairs and political context. Any sort of online dissent, or online organisation of offline dissent, is virtually impossible. Dissenters rely on smuggled Western hardware and VPNs that allow them to use Western internet platforms instead, but this means that they have vastly restricted audiences in mainland China. The inability to express any dissent meaningfully also encourages radicalisation among some dissidents (in particular those persecuted by the party), some of whom then resort to more drastic measures. These examples making national news serves to make public opinion even more anti-dissident than it already is given all the CCP propaganda.</p><p>In 2027, China started exporting its AI censorship system. There had already been a secret 2026 deal with Russia, but Russia had prioritised moving off the Chinese system and did so in 2028, moving onto a worse but domestically-developed one running on old Chinese GPUs and open-source models. Granting a foreign country control over your AI censorship apparatus gives that country a huge amount of leverage, including the ability to potentially withdraw it quickly or change how it steers the conversation, which could threaten the regime. However, smaller and less technically-sophisticated countries like North Korea and Equatorial Guinea buy the Chinese system, taking a step towards becoming Chinese client states in the process.</p><p>The semiconductor supply chain is a key geopolitical battleground. Europe's big leverage point is the Dutch ASML's monopoly on EUV (extreme ultraviolet lithography) machines. TSMC and therefore Taiwan continue being important into 2029, even though TSMC's fabs in America are starting to produce chips in serious numbers. An embarrassing failure is Intel, despite its strategic importance for both America and Europe (the latter due to a major Intel fab in Germany that was built 2023-2027 and started production in 2028). With the arrival of superhumanly cheap and fast AI software engineers, Intel's x86 moat disappears because it is trivial to port programs to running on ARM. <a href="https://en.wikipedia.org/wiki/Wintel">Wintel</a>, long on the rocks, is dead. In 2026-2027, Intel is in free fall and crisis. In 2028, Intel spins off its fabs, selling them to xAI at a discount price, with pressure from the Trump administration to sell to an American (and, implicitly, Musk-affiliated) buyer, and a plan by Elon Musk for xAI to get a comparative advantage by being the only vertically-integrated chips-to-tokens AI model provider. This also feeds into the 2028 American AI Action Agenda (AAAA), that also lavishes more government subsidies on both the new xAI Foundry and on TSMC's US fabs, seeking to make the US fully independent in semiconductors by 2033 and cement Trump's legacy.</p><p>The overall picture is one where the main AI supply chain includes the EU, Taiwan, China (implicitly, through its "veto" on Taiwan's existence), and the US. However, this "main chain" is on track to being replaced by a self-sufficient American semiconductor and AI industry in the early-to-mid 2030s, and by a self-sufficient Chinese semiconductor and AI industry on an even faster timescale (though the Chinese one is a year or two behind technically). In 2029, the new administration in the US finds some spending cuts and throws the EU a bone (in exchange for cooperation on security issues) by giving up on trying to create an American competitor to ASML. The UK has some unexpected success in being an academic and open-source AI applications research hub, a policy laboratory for the US, and an AI biotech hub. However, its geopolitical weight rounds to zero. Apart from ASML, the EU is also mostly not relevant, especially as it has managed to greatly slow the diffusion of AI through regulation. The world overall is moving towards a bipolar order between the US and China. Compared to the Cold War, however, both powers are more inwardly-focused and less ideological. The US is in an isolationist era. While China is gradually converting much of the third world into client states, the CCP's main goal remains internal stability and its secondary goal "making the world safe for dictatorship", rather than the ideological expansionism of the Soviet Union. The Taiwan question has been punted into the mid-2030s, as the CCP believes the world's reaction will be much more muted and less dangerous to Party control once America no longer cares about Taiwanese chips, and once even more of the world has been preemptively bribed into silence.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.nosetgauge.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe for updates:</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p><a href="https://nosetgauge.substack.com/p/a-history-of-the-future-2025-2027">Part 1: 2025-2027</a></p><p><a href="https://nosetgauge.substack.com/p/a-history-of-the-future-2030-2040">Part 3: 2030-2040</a></p><p></p><p><em>Thanks to <a href="https://lukedrago.substack.com/">Luke Drago</a>, Duncan McClements, Theo Horsley, and Bilal Chughtai for comments.</em></p>]]></content:encoded></item><item><title><![CDATA[A History of the Future, 2025-2027]]></title><description><![CDATA[A scenario, part 1.]]></description><link>https://www.nosetgauge.com/p/a-history-of-the-future-2025-2027</link><guid isPermaLink="false">https://www.nosetgauge.com/p/a-history-of-the-future-2025-2027</guid><dc:creator><![CDATA[Rudolf Laine]]></dc:creator><pubDate>Mon, 17 Feb 2025 01:51:30 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!oGuE!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ad7e08f-137c-446d-858a-68ef19b67960_1370x1138.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Below is part 1 of an extended scenario describing how the future might go if current trends in AI continue. The scenario is deliberately extremely specific: it&#8217;s definite rather than indefinite, and makes concrete guesses instead of settling for banal generalities or abstract descriptions of trends.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!oGuE!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ad7e08f-137c-446d-858a-68ef19b67960_1370x1138.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!oGuE!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ad7e08f-137c-446d-858a-68ef19b67960_1370x1138.png 424w, https://substackcdn.com/image/fetch/$s_!oGuE!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ad7e08f-137c-446d-858a-68ef19b67960_1370x1138.png 848w, https://substackcdn.com/image/fetch/$s_!oGuE!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ad7e08f-137c-446d-858a-68ef19b67960_1370x1138.png 1272w, https://substackcdn.com/image/fetch/$s_!oGuE!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ad7e08f-137c-446d-858a-68ef19b67960_1370x1138.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!oGuE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ad7e08f-137c-446d-858a-68ef19b67960_1370x1138.png" width="488" height="405.36058394160585" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8ad7e08f-137c-446d-858a-68ef19b67960_1370x1138.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1138,&quot;width&quot;:1370,&quot;resizeWidth&quot;:488,&quot;bytes&quot;:2969940,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!oGuE!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ad7e08f-137c-446d-858a-68ef19b67960_1370x1138.png 424w, https://substackcdn.com/image/fetch/$s_!oGuE!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ad7e08f-137c-446d-858a-68ef19b67960_1370x1138.png 848w, https://substackcdn.com/image/fetch/$s_!oGuE!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ad7e08f-137c-446d-858a-68ef19b67960_1370x1138.png 1272w, https://substackcdn.com/image/fetch/$s_!oGuE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ad7e08f-137c-446d-858a-68ef19b67960_1370x1138.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Open Sky</em>. (Zdzis&#322;aw Beksi&#324;ski)</figcaption></figure></div><h2><strong>The return of reinforcement learning</strong></h2><p>From 2019 to 2023, the main driver of AI was using more compute and data for pretraining. This was combined with some important <a href="https://situational-awareness.ai/from-gpt-4-to-agi/#Unhobbling">"unhobblings"</a>:</p><ul><li><p>Post-training (supervised fine-tuning and reinforcement learning for instruction-following) helped the LLMs be usable without difficult prompting.</p></li><li><p>Starting in 2024, Anthropic showed that judgement and taste in data curation&#8212;and the evaluation metrics that guide data curation&#8212;could give you a "magic sauce" effect in perceived LLM quality.</p></li></ul><p>Most real-world LLM uses, of course, involved generating a sequence of tokens to try to achieve some task. So there were a lot of untapped gains from doing reinforcement learning (RL) for performance on concrete domains, rather than just RL for the models following instructions and being "safe"&#8212;i.e. a combination of avoiding PR hazards, and preparing for misuse mitigations on actually capable models down the line.</p><p>OpenAI fires the starting gun in 2024 with the release of o1, which was based on RL on chains-of-thought (COT), i.e. the model is trained to reason step-by-step towards correct answers, i.e. "test-time compute" in the horror-filled annals of machine learning jargon. In late 2025 they release &#8220;GPT o5&#8221; (&#8220;GPT&#8221; to normal people, and &#8220;o5&#8221; to those keeping track of the version number), a model which can take text, image, audio, video, computer screen state, real-life footage, whatever, process and understand it (choosing itself whether it should do chain-of-thought reasoning before answering or not), and output text, image, audio, video, computer actions.</p><p>Whereas the labs had spent almost four years racing down the scaling graph on pretraining compute, they had not yet done so for COT RL, and had not uncovered the subtler tricks to doing this well. This meant there was a lot of low-hanging fruit, so progress&#8212;and replication&#8212;was fast. In early 2025, DeepSeek spooks the entire American business scene with their release of R1. In spring 2025, Anthropic ships Claude 4, which also has inference-time compute abilities that trigger if the model is asked a question where that helps.</p><p>Anthropic keeps their largest Claude 4 model internal and secret from the very start. It gets used for (most importantly) producing training data for the smaller Claude 4s, and (experimentally) in doing internal evaluations of AI-driven AI R&amp;D, starting with some adversarial robustness research on Claude 3.5. Inference costs on the biggest models are a big part of the rationale. Anthropic continues being focused on intelligence over product, and enterprise products over consumer products. They make only minor gains among consumers, but Claude is increasingly adopted among enterprises, programmers, knowledge-workers, and nerds. (Ironically, OpenAI has the consumer advantage despite focusing more on reasoning and less on the LLM being personable and writing well.)</p><p>In 2025, thanks to RL, &#8220;agentic&#8221; AI is here, but only kind of. Anthropic and OpenAI have computer-use features that work, except a bit spottily and are designed to never authorise a payment or send an email or do anything important without human confirmation. Google releases an AI agent for things like Google Cloud Platform configuration schlep, which the programmers love. A bunch of startups are competitive with the major lab products, in particular because no one has yet had time to pour ungodly amounts of compute into the COT RL. However, most "agentic" AI applications remain LLM scaffolds, i.e. a hard-coded flowchart of LLM prompts and other API calls.</p><p>Meta is trialling some unholy autonomous AI features across their apps (such as AI agents going around leaving comments on user&#8217;s posts to &#8220;maximise engagement&#8221;), but they still seem like gimmicks.</p><p>Code generation tools like Cursor and Lovable and Zed and Poolside and Magic.dev and ten million others are getting very good. For most apps, you can in fact just drop in a prompt and have the app running within a few minutes, though managing infrastructure is still a pain and technical debt tends to accumulate if the AI stacks many changes on top of each other. Some form of COT RL is used in the training stack for many but not all leading coding tools. LLM scaffolds still reign over unspecialised general agents.</p><p>Gemini-3 ships in 2025 after a vast pretraining run. It&#8217;s good but a disappointment; the final culmination of pretraining scaling laws in an era where products, inference-time compute, data curation (mostly synthetic now, but behind the scenes there&#8217;s some very important human judgement going on), and real-world interaction ability are key. Google DeepMind (GDM) is building powerful maths models, and making progress on reasoning architectures that don&#8217;t rely on external COT and are better-suited for maths.</p><p>After 2025, RL starts getting harder and the distance between the leading labs and the rest increases again. RL is simply less efficient than pretraining, partly because the necessity of letting models try long sequential chains of actions makes parallelism harder. The labs have now scaled up RL compute quite far, so the resource bar for being in the game rises. Also, RL is notoriously hard. First, subtle bugs are easy to make and hard to notice: is the RL agent not learning because you made a bug, or because it just won't learn? Second, there are more choices to make (e.g. you have to pick a reward function and scoring method, rather than the cross-entropy loss you default to with pretraining). OpenAI, Anthropic, and Google take some distnace to the rest in RL and overall general capabilities. However, the other labs don't necessarily see this as a loss&#8212;Meta deliberately focuses more on integrating AI into its products over 2025 and 2026, xAI focuses more on engineering use-cases, and both xAI and DeepSeek remain competitive. Also, the issues with RL mean that there are some more hairy technical problems that temporarily slow progress as labs one after another internally work through them, though this is not at all obvious from outside a lab.</p><p>In early 2026, xAI starts deploying an early version of an AI that can do engineering CAD work kind-of-well, as long as a human is looking over its shoulder and checking its work. This improves a lot after Tesla and SpaceX (are forced to) actually start using it, but it&#8217;s not yet groundbreaking; sheer data quantity remains an issue.</p><p>The next big advance is OpenAI's late-2026 release of o6. It has improved a lot in computer use, and generally in unifying its various input and output types (e.g. it can use text and images more effectively together in its output, process and output longer videos, etc.). Second, it has a more advanced memory architecture, including a built-in longer-term memory that allows instances to learn over time. Thirdly, it&#8217;s of course generically a bit smarter, a bit faster in token output, and so on. In particular, OpenAI has finally almost caught up to Claude&#8217;s personality level. It is also way more impressive to normal people because it can also&#8212;if prompted to do so&#8212;generate real-time video and audio of a talking face. OpenAI doesn&#8217;t explicitly encourage this, but winks at this, since it knows this will get some users addicted (especially as they now have a more nuanced policy for sexually explicit model outputs than the previous blanket ban).</p><p>Many people in Silicon Valley declare this AGI, and predict the immediate automation of all office jobs. In practice, it falls short in a hundred subtle ways that make it not a drop-in replacement, in particular with remaining unreliability in its ability to use computers and weaknesses at planning and completing long-horizon tasks. But the smart money is betting that these issues will be solved within a year.</p><p>Also in late 2026, Anthropic releases Claude 5 Haiku and Claude 5 Sonnet. Claude 5 Haiku is a cheap model roughly on par with Claude-3.5-Sonnet in smartness while having an output speed of hundreds of tokens per second. They come with an upgraded version of computer use that is far faster and more seamless. Again, the largest model is kept internal. Its training data curation and post-training finetuning was focused on programming, ML research, MLOps, and maths. Anthropic employees started adopting it internally in mid 2025, giving researchers and engineers what's essentially a team of AI interns to manage. They then spent 6 months giving the models tailored feedback, which they massively boosted with methods for dataset augmentation, and filtered for correctness with scalable oversight techniques like debate, before feeding it back into the model as finetuning data. In 2024, Anthropic internally estimated a +5-10% productivity boost from internal use of Claude-3.5-Sonnet and early training checkpoints of Claude-4; in 2025, this rose to +25%, and with Claude 5 Opus it started out at +35% but has gradually accelerated with more and more finetuning to +60% by mid 2026, and the number is still climbing. OpenAI does not have a comparable setup internally, partly because it&#8217;s politically less feasible due to the lower-trust environment, but also because it's a lower priority since they believe less in recursive self-improvement.</p><h2><strong>Codegen, Big Tech, and the internet</strong></h2><p>Coding is a purely digital job that is economically highly valuable, has a lot of training data and often provides a clean feedback signal of success, and that the AI-affiliated companies all already have expertise in. All this makes it ideal for AIs to be good at, quickly. In 2023-2026, the biggest economic impact of LLMs is their use in coding.</p><p>In 2023, models got good enough for programmers to prefer them to looking up human guidance on sites like StackOverflow. In 2024, coding copilots were a real productivity boost, perhaps +10% to +50%, for pure software engineering tasks (higher for things that are more boilerplate and when the coder has less background in what they're doing, lower for more research-y tasks or when working with familiar domains). In 2025, there are two big new advances. First, chain-of-thought RL meant that spending more LLM tokens converted more efficiently into better code. Second, a bunch of the obvious improvements to the workflow were made, such as the AI automatically running tests or checking that the UI looks right, and autonomously trying again if not, rather than maintaining the human as a tab-switching, prompt-writing monkey that does this for the AI. As a result, by 2026 codegen looks solved. There are some wrinkles left related to cloud infrastructure stuff, especially when there&#8217;s little training data on some aspect and/or a heavy and unavoidable button-clicking component, but these are quickly getting fixed, especially as computer use gets good and allows the models to better click buttons and surf the internet for documentation.</p><p>For a while, everyone&#8217;s paranoid about security in the fully AI-written codebases, and a bunch of security consulting and cybersec firms make a killing. However, it soon turns out codegen stuff is actually more secure than human code because the LLMs reliably do the standard correct thing over the weird bespoke thing whenever it comes to security, and this eliminates a lot of security vulnerabilities that humans would write in. The security consulting and cyber firms quickly become LLM wrapper companies with excellent marketing arms, and stop being used by most people apart from risk-averse large companies and governments. However, as is statistically obvious, there are a bunch of high-profile blowups, and it remains true that existing code can now be much more easily attacked since all you need is an o6 or Claude subscription.</p><p>By 2027, the price of creating a simple app is a few dollars in API credits or GPU hours. The price of a particularly complicated piece of software is on the order of $100 to $10k. The tech stack has shifted almost entirely to whatever there was the most data on; Python and Javascript/Typescript are in, almost everything else is out. Average code quality as judged by humans declines, but this is fine because humans don't read it and the LLMs can deal better with bloated code.</p><p>The coding advances trigger a massive flood of non-coders or amateurs flooding in and trying to make money off B2B SaaS or freelance programming. Agentic non-technical people are launching niche startups at massive rates, since you can ship a full-featured product in a few hours if you&#8217;re willing to burn money on API credits. Lots of these projects run into &#8220;tech debt hell&#8221; eventually. For a while programmers can earn heavy consulting fees (or cofounder roles) by coming in, chatting to the AI about the codebase, and telling it to make architectural changes that let the next features be added cheaper because it will take fewer lines of code on top of the better-structured codebase. However, just asking the AI &#8220;what&#8217;s wrong with this codebase&#8221; and then &#8220;how would you fix it&#8221; also works quite well if the prompting is good. The codegen scaffolds quickly evolve to be good at reflectively prompting AIs and managing the tech debt hell better, but it&#8217;s hard to notice this unless you&#8217;re working with them actively, leading to a lot of misinformed doubts about the capabilities based on early disappointments. The labs also start including more qualitative criteria in their codegen RL&#8212;not just "did the code run and pass the tests", but also asking another LLM to grade the style and extensibility of the code. In effect, there's a race over whether the AIs will learn good code practices from RL self-play, or from explicit human scaffold-crafting and prompting. Note that the latter is getting easier too, as the tooling improves, and AIs write the scaffold code and distill human programming knowledge into prompts. For example, in late 2025 Anthropic also ship an automated tool for building an LLM scaffold from observations of an arbitrary real-world digital work process.</p><p>Big Tech starts using the codegen tools heavily for new projects, but integration into older projects is slower because the codegen scaffolds are worse at interfacing with large existing codebases than writing small ones from scratch. This gets mostly solved over the course of mid-2025 to mid-2026, but gives the "Little Tech" startups a temporary tailwind. Big Tech headcounts grow, as they hire more people both to flatter the egos of managers&#8212;they are drowning in cash anyway&#8212;and in particular many product managers to oversee the AI codegen agents that are unleashing a massive series of new products now that they're mostly no longer constrained by development taking lots of time. Internal company office politics becomes even more of a rate-limiter: if teams are functional, the AI codegen boost means more products shipped, whereas if teams are not, the gains are eaten up by employees working less or by factional fights within companies. Microsoft launches &#8220;365 Days of Microsoft&#8221;, where every day of the year they release a new software product or a big update to a previous one; they move increasingly into more niche enterprise markets that they had previously ignored as part of a new paradigm shift. Google is more scattered and launches a thousand new features integrated into their product suites that&#8212;on paper&#8211;compete with existing startups, and&#8212;in practice&#8212;serve to expand the empires of enterprising Google middle-managers. Google gets a reputational hit as a shipper of sloppy products, but they have a few big hits and their customers are a captive market that will continue using Search and Drive, giving them room to flail around.</p><p>There are a few corporate scandals as AI codegen products fail, leading to a frenzy of effort at testing and fuzzing the AI outputs. But Big Tech is still all-in, at least until late 2026: they&#8217;re all feeling the AGI, and that if they miss it that&#8217;s an existential mistake, and if it&#8217;s all a bubble then at least they were only as bad as all the other Big Tech firms. The one slow actor is Apple, due to its cultural bias towards quality and assurance. Apple ships Apple Intelligence integrations but that&#8217;s about it.</p><p>Predictably, the super-abundance of software and the extreme competition in it drives down prices. SaaS companies aren&#8217;t yet experiencing an extinction wave because humans react to change slowly, but it doesn&#8217;t look good and investors start getting skittish. The big advantage that everyone points to is having locked-in customers or network effects; otherwise, the conventional wisdom goes, you're dead. But there are a bunch of companies and tools that let you circumvent attempts at customer lock-in. You can program yourself an X.com in an afternoon, have computer-using AI agents trawl X, Reddit, etc., pull in content to your site, and propagate your posts and replies automatically to all the other platforms. Some companies fight tooth and nail to try to make people stay on their platform (and thus see their ads); some just charge API prices and hope to at least get revenue. &#8220;Web4&#8221; comes to mean a programmable internet that is customised to everyone. A hundred startups jump on this bandwagon. Some established companies create carefully de-risked APIs and let users program customisations and integrations into their sites (i.e. let users ask codegen models to do such programming). The Web4 wave generally runs into the problem that most people don&#8217;t actually want to customise things; they want someone to have already thought through the interface and features on their behalf, are fine with existing setups, and not very eager to re-imagine the internet. But increasingly, if users dislike something about a site, they will build their own version, connect it to the original with AI schlep, and then lure over the few percent of users that are triggered by the same thing. Technical barriers like scraping limits are hard as AI agents can be made to browse in increasingly human-like ways (one successful startup explicitly engages in a scraping race against scraping-detection methods by fine-tuning a computer use agent on real human mouse moving patterns). An increasingly common barrier is asking humans for government ID or other real-world verification (with the privacy constraints mitigated with zero-knowledge proofs, if it's a fancy libertarian or crypto -affiliated thing). This too is spreading, also because some people want sites where they can talk to verified real humans.</p><p>By 2026, more code gets written in a week than the world wrote in 2020. Open source projects fork themselves into an endless orgy of abundance. Some high school students build functionally near-identical versions of Windows and Google Drive (and every video game in existence) from scratch in a month, because they can and they wanted one new feature on top of it. Everyone and their dog has a software product line. Big Tech unleashes a torrent of lawsuits against people cloning their products, echoing the Oracle v Google lawsuit about Java, but those lawsuits will take years to complete, and months feel like decades on the ground.</p><p>Silicon Valley is exuberant. The feeling at Bay Area house parties is (even more than before) one of the singularity being imminent. Some remain skeptical though, rightfully pointing out that the post-scarcity software isn&#8217;t the same as post-scarcity everything, and that genuine &#8220;agency&#8221; in the long-horizon real-world planning sense hasn&#8217;t really arrived, and under the hood everything is still rigid LLM scaffolds or unreliable AI computer use agents.</p><h2><strong>Business strategy in 2025 &amp; 2026</strong></h2><p>Even though Meta, DeepSeek, and others are behind in raw intelligence and reasoning all throughout 2025 and 2026, they threaten the big labs because they are giving away (both to consumers, and freely to developers through open-weights releases) a level of performance across audio and video and image and text that is &#8220;good enough&#8221; for most use cases. SOTA performance is no longer needed for many use-cases, especially low-end consumer entertainment (e.g. image gen, chatbots, etc., which Meta is banking on), or most classification, processing, or business writing tasks.</p><p>OpenAI is especially vulnerable, since they rely heavily on consumers, and are also increasingly a product company that competes with products built on their API, driving many to switch. Their strategy (internally and to investors, though not publicly) is to be the first to achieve something like a drop-in agentic AI worker and use that to convert their tech lead over open source into &gt;10% of world GDP in revenues. They&#8217;ve raised tens of billions and make billions in revenue from their products anyway, so they can bankroll these efforts just fine.</p><p>Anthropic remains a jewel of model quality and a Mecca of technical talent that gets surprisingly little attention from the rest of the industry. Analogies to Xerox PARC abound, but there are whispers of internal AGI being imminent, and no one else can claim the ideological mandate of heaven for safe AGI. The talent and money spigots stay on.</p><p>xAI and DeepSeek continue releasing open-source consumer models. Both also have a specialty in maths-y STEM and engineering stuff, aided by data collection efforts (with xAI being able to work closely with SpaceX and Tesla engineers) and inference-time compute methods. xAI also continues trying to leverage real-time access to X.com data to its benefit, but this isn't a major advantage or revenue source.</p><p>In 2024, thousands of startups were chasing after a lot of different use cases, and some started making serious money, but it was still very early days for actual products. The big winners were companies like Perplexity that use LLMs to trivially improve some LLM-compatible user case (like search), companies like Glean and Hebbia that are doing various enterprise LLM integration schlep, and legal LLM companies like Harvey (since law is intensely textual and high-revenue). However, the real money is still in infrastructure / &#8220;selling shovel&#8221;, in particular Nvidia.</p><p>By the end of 2025, there is no technical bottleneck to remote doctor appointments or most legal work being done entirely by AI. However, diffusion takes time. Also, in many countries lawyers barricade themselves behind a cluster of laws that forbid lawyer-automating AI. Getting hired as a new lawyer, or any kind of white-collar analyst, is <a href="https://lukedrago.substack.com/p/agi-and-the-corporate-pyramid">getting harder</a> though, as decision makers expect AI to reduce their need for entry-level white-collar workers of every kind, and firing people is much harder than not hiring them in the first place. Healthtech AIs are gradually working their way through regulatory hurdles over 2025-2026, and are clearly better than the average doctor at all the parts of the job that rely only on reasoning and knowledge. However, AI doctor appointments are only trialled at any significant scale in 2026, by Singapore and Estonia. Significant integration of AI in the non-patient-facing parts of the healthcare system is underway in the UK, many EU countries, South Korea, and China by 2026, but again diffusion is slowed by the speed of human bureaucracy.</p><p>There are lots of &#8220;AI agent&#8221; companies automating things like customer service, various types of search (e.g. for shopping / booking flights / etc.), and back-office computer processes. The big cloud hanging over them in 2025 is whether AI codegen scaffolds soon get good enough that they are trivial to replace, and whether generalist AI agents soon get good enough to kill both. In 2026 the first question starts being answered in the affirmative, as lowered barriers to coding create a flood of new entrants and a ruthless bloodbath of competition. However, even the release of o6 in 2026, despite some initial hype, does not yet cause much evidence of the generalist AI agents taking over both by the end of 2026.</p><p>There&#8217;s lots of LLM evals startups like Braintrust.dev and HumanLoop and Atla, that are mostly struggling to differentiate themselves against each other or to define a new testing/reliability/verification paradigm for the LLM scaffold era, but are growing fast. There&#8217;s a lot of LLM agent oversight solutions, but by the end of 2026 none manage to make a massive leap, and the unlocking of new AI uses remains bottlenecked on incumbents' risk tolerance and a slow buildup of knowledge about best practices and track records. A surprisingly retro success is call-centres of humans who are ready to jump in and put an AI agent back on task, or where AI agents can offload work chunks that are heavy on trust/authentication (like confirming a transaction) or on button-clicking UI complexity (like lots of poor legacy software), to human crowdworkers who click the buttons for them, while the AI does the knowledge/intelligence-intensive parts of the job on its own.</p><p>Many of the really successful startups are in the spaces that Big Tech won&#8217;t touch or has trouble touching: anything controversial (the sexual and the political), and anything too edgy or contrarian or niche/vertical-specific.</p><p>The fact that the explosion of codegen threatens Big Tech&#8217;s moat, plus some disappointment at the unreliability of o6 after so much hype, plus some general memetic force that means the &#8220;current thing&#8221; can be AI only for so long, combines to cause a market correction near the end of 2026. Software is starting to seem stale and boring. Investors want to see &#8220;real AGI&#8221;, not just post-scarcity in software. Google DeepMind&#8217;s maths stuff and xAI&#8217;s engineering stuff are cool; OpenAI and LLMs are not. Amazon&#8217;s AWS &amp; physical store is cool, Google Search and Facebook are not.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.nosetgauge.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.nosetgauge.com/subscribe?"><span>Subscribe now</span></a></p><h2><strong>Maths and the hard sciences</strong></h2><p>A compressed version of what happened to programming in 2023-26 happens in maths in 2025-2026. The biggest news story is that GDM solves a Millennium Prize problem in an almost-entirely-AI way, with a huge amount of compute for searching through proof trees, some clever uses of foundation models for heuristics, and a few very domain-specific tricks specific to that area of maths. However, this has little immediate impact beyond maths PhDs having even more existential crises than usual.</p><p>The more general thing happening is that COT RL and good scaffolding actually is a big maths breakthrough, especially as there is no data quality bottleneck here because there&#8217;s an easy ground truth to evaluate against&#8212;you can just check the proof. AIs trivially win gold in the International Mathematical Olympiad. More general AI systems (including increasingly just the basic versions of Claude 4 or o5) generally have a somewhat-spotty version of excellent-STEM-postgrad-level performance at grinding through self-contained maths, physics, or engineering problems. Some undergrad/postgrad students who pay for the expensive models from OpenAI report having had o3 or o5 entirely or almost entirely do sensible (but basic) &#8220;research&#8221; projects for them in 2025.</p><p>Mostly by 2026 and almost entirely by 2027, the mathematical or theoretical part of almost any science project is now something you hand over to the AI, even in specialised or niche fields.</p><p>In 2026, xAI also tries to boost science by launching an automated peer-reviewer / paper-feedback-giver specialised in STEM subjects, that can also run followup experiments automatically, and soon take a paragraph setting the direction and convert it to basically a full paper. Cue a thousand academics blasting it for mistakes in its outputs. The fair assessment is that it&#8217;s impressive but not perfect (somewhat like having a brilliantly fast but easily-distracted and non-agentic undergrad research assistant), but still better than all but the highest-effort human peer-reviewers. Elon Musk gets into feuds about its quality online, becomes radicalised about peer-review and academia, and starts the &#8220;Republic of Papers&#8221; as a side-feature on X to explicitly try to replace academia (it helps that, in 2026, the higher education bubble seems to be bursting in America, partly triggered by fears about AI job automation but also due to political headwinds). Everyone has Opinions.</p><p>In 2026, GDM releases work on new maths-oriented AI architectures that include an advanced more flexible derivative of MCTS that also searches for new "concepts" (i.e. new definitions that shrink the length of the most promising proof-tree branches) while doing the proof-tree search. Their maths models prove a long list of new theorems and results, including, in 2027, solving a few more long-standing prize problems, this time in a less ad-hoc and more credibly entirely-AI way. Demis Hassabis talks about "solving physics" within the next year, through a program that includes GDM collaborating with leading physicists.</p><p>In 2028, GDM&#8217;s collaboration with the theoretical physicists bears fruit: general relativity and quantum mechanics are unified with a new mathematical frame. There are a few candidate new theories with different values of some parameters that can only be verified by expensive experiments, but it seems clear that one of these candidate theories is correct. It's not "solving physics" or a final theory of everything, but it is clearly a major breakthrough in mathematical physics. The technical work owed a lot to a truly enormous compute-budget for RL self-play, the construction of a massive dataset of physics papers, critiques of them, and tokenised observational data by a physicist-and-AI-agent team, and close collaboration with a large number of leading physicists who gave feedback to the AI on the developing theories. Credit for the Nobel Prize is the subject of much discussion, but eventually (in 2030) ends up split between Demis Hassabis, one of the physicists who was most involved, and the most important AI system. Everyone has Opinions.</p><p>Corporate Google likes the PR win of achieving the century's greatest physics breakthrough so far, but the application of this mathematical firepower they are most hopeful about is formally verifying the correctness of software. This is especially pressing as there&#8217;s a lot of shifting tides in the cyber world. Codegen itself is on net a defense-dominant technology (as discussed earlier). Most of the hacks are either due to sloppy mistakes by early codegen products, or some adversary using AI tools to direct a disproportionate amount of effort on attacking some piece of legacy software that is still used, or on which a codegen-written program (indirectly) depends. There&#8217;s increasing demand for really air-tight software from a US defense establishment that is obsessed with cyber advantage over especially China, but also Russia, Iran, and North Korea. Also, easily proving the correctness of code will allow better feedback signals for codegen models, and help in the ambitious efforts underway to rewrite massive parts of the existing tech stack. So in addition to making leaps in the hard sciences, GDM&#8217;s other big applied goal is a world where the correctness of all essential code is proven. They have an early success in a late-2026 plugin for several popular languages that is essentially a type-checker on steroids (though of course, this is adopted less by the humans and more by the AIs that now write almost all of the code).</p><p>Initially, the US government tries to restrict the diffusion of code verification tools, since they don&#8217;t want China to get provably-correct coding capabilities. However, the open source community is only about 6 months behind in verification as it makes some leaps and bounds in 2027-2028, especially since there are thousands of former software engineers and mathematicians without much to do as they wait for the AIs to do their work for them.</p><p>As a result, by 2028 feats of intellect that would&#8217;ve taken Euler decades are done in a few minutes to mathematically prove that, conditional on the CPU's physical integrity working, some code is an utterly impregnable and flawless pizza delivery routing system. However, verification is not adopted nearly everywhere because there&#8217;s a cost multiplier over just getting an AI codegen tool to write unverified code (and AI codegen has continued plummeting in cost, not that anyone really notices anymore).</p><h2><strong>Societal response</strong></h2><p>On the soft skills side, by 2025 experiments show that models have reached human-level persuasion capabilities in controlled text-only chat settings. However, this doesn&#8217;t really matter, since it&#8217;s not how most human persuasion works; part of models being bad at long-horizon planning is weaknesses in strategic relationship-building with relevant actors over longer timescales. There also isn&#8217;t yet widespread use of models to manipulate politics. First, there just isn&#8217;t a particularly tech-savvy political campaign or movement to influence opinion, except for China gradually experimenting with increasingly more AI in their censorship bureaucracy. Second, models still seem worse than the best humans at that &#8220;spark&#8221; that lets some people create persuasive, viral ideas. Third, the memetic selection pressures acting on the collective output of humanity on the internet are already superhuman at discovering memetic viruses and persuasive ideas than any individual human, so passing any individual human capability threshold in this domain is not automatically a society-steering ability.</p><p>However, some 1-1 use-cases do work. AI scam calls with deepfaked audio and video start being a nuisance by mid 2025 but are mostly reined in by a series of security measures pushed by platforms (and by regulation in the EU), people creating new trust protocols with each other (&#8220;what&#8217;s our secret passphrase?&#8221;), increased ID verification features, and growing social distrust towards any evidence that's only digital.</p><p>Lots of people are talking to LLMs for advice. Some swear by Claude 4 in particular. Character.ai -like startups are having a boom. There is a lot of public discussion about people increasingly talking to AIs instead of having human friends and partners (which is boosted after multimedia Llama models are finetuned to be good at sexual image, audio, and&#8212;in 2026&#8212;video output). There's a hikikomori-like trend, strongest in California, South Korea, and China, where a minority of people forsake almost all human social contact and instead interact with AIs that are superhumanly risk-free and pliable, and come with superhumanly nice voices and avatars. In 2026, Australia and Canada ban talking to non-educational AIs with voice capabilities or human-like avatars for under-16s.</p><p>The written text quality of models remains surprisingly mediocre. Claude does best, and is great when prompted right, but &#8220;ChatGPTese&#8221; remains a thing that afflicts especially OpenAI and Google (though the former improves in 2026), and any human who writes mediocre prompts. There are loads of LLM slop content websites, but not a single blog written by an LLM becomes widely read among intellectual or elite circles.</p><p>As the codegen wave of 2026 hits, many consumers feel a few weeks of wonder and whiplash at the agentic AIs that can now do parts of their job, and at the massive orgy of abundance in software, and then this becomes the new normal. The world of atoms hasn&#8217;t changed much. Most people by late 2026 just assume that AIs can do basically everything digital or intellectual, and become surprised when they learn of things that the AIs can&#8217;t do.</p><h2><strong>Alignment research &amp; AI-run orgs</strong></h2><p>In 2025, someone adds some scaffolding on top of an OpenAI Operator instance, making it in-theory capable of earning money through freelance work to pay for its own API costs, including automatically buying more credits for itself and find more freelance work. However, the economics doesn't work out, so it can't actually survive on its own without subsidies. In early 2026, a similar concept actually is economically-viable, and some are launched as an experiment by tech-savvy freelancers looking for some easy money, or by people who are just curious. A few blow up, mostly by doing various things related to memecoin manias and going viral as a result. In late 2026, one such autonomous AI scaffold with a memecoin windfall reasons about next steps, tries to incorporate a US business for itself, cold-emails a bunch of humans to ask for ID, and manages to get one of them to give an ID so it can incorporate the business. By 2027, there are a few experimental digital businesses run by AIs, but they're not very competitive, and often rely on what's effectively a subsidy in human interest due to them being novel.</p><p>Alignment research in 2025-2027 is driven by Anthropic (though of course most of their research is on GPU performance engineering, inference-time compute techniques, and other things focused on raw capabilities progress). SAEs peak in popularity in late 2024 before being mostly forgotten, but there&#8217;s a new interpretability paradigm that starts being put together in late 2025 based on identifying more general geometric structures in activation space. AI control setups are tested against misalignment &#8220;model organisms&#8221; that, by 2027, are trivially capable of hacking out of a normal environment. Model weight security at Anthropic is excellent for a private company, but this just means the attackers target OpenAI instead (and the gap between labs and open source is never more than a year in 2025-2027). And, of course, Anthropic internally writes endless safety cases. The general message in them is that a lot is resting on either an interpretability breakthrough or on AI control working on superhuman models. The low amount of evidence gained on &#8220;alignment&#8221; is frustrating to many; models have been caught scheming endlessly but always in fairly artificial setups, or in messy circumstances where it's not clear what the model should've done. The most important work seems to be work on properties upstream of scheming, such as a stream of work on corrigibility kickstarted by the 2024 Greenblatt et. al. paper <a href="https://assets.anthropic.com/m/983c85a201a962f/original/Alignment-Faking-in-Large-Language-Models-full-paper.pdf">"Alignment faking in large language models"</a>. The alarmingness of the early evidence against corrigibility was offset by promising empirical work on meta-learning techniques to encourage corrigibility in late 2025 and early 2026. By 2027 it's known how to train a model such that it either will or won't be amenable to being trained out of its current goal. Anthropic reveals this and some other safety-related insights to OpenAI and Google, and asks the State Department to reveal it to Chinese labs but is denied.</p><p>By 2027, the new interpretability paradigm is seeing progress, with AIs doing essentially all of the engineering and much of the detailed ideation. This reveals a taxonomy of patterns and feature representation types within neural networks. A few are neat and clean, but mostly models&#8217; internals turn out to be messy and with massive redundancy between different parts. The notion of a model having a singular &#8220;goal component&#8221; looks less likely, at least if certain choices are made during training.</p><p>A test case of the new alignment techniques at Anthropic is the training in 2027 of a new model, Claude 5 Epic or just "Claude Epic", based on Claude 5 Opus -curated training data. Company leadership internally thinks it will be a full AGI. The interpretability team will be observing the model at checkpoints and watching it develop. Countless safety cases have been written; the hope is still to run evals, use AI control setups, and hope for some last-minute firmer guarantees from the interpretability work. Some at Anthropic are entirely convinced just by the scalable oversight work that&#8217;s already been done. Others expect the hard part of intent alignment to rear its head at any moment.</p><p>One of the avenues that seemed most promising in 2025 was interpreting AI chains-of-thought (COTs), something <a href="https://x.com/LRudL_/status/1882899577923330369">far</a> <a href="https://www.lesswrong.com/posts/HQyWGE2BummDCc2Cx/the-case-for-cot-unfaithfulness-is-overstated">easier</a> to make meaningful progress on than interpretability. However, over 2026-2027, much more compute is poured into RL, and the COTs become less legible, as the models drift towards shorthand scripts that are more effective for them than writing out their thoughts in English. Work done by Anthropic and several academic labs leads to techniques for encouraging human interpretability of the COTs, by adding a COT interpretability term to the RL loss function and having some clever training details to avoid the model <a href="https://www.lesswrong.com/posts/Eu6CvP7c7ivcGM3PJ/goodhart-s-law-in-reinforcement-learning">goodharting</a> the interpretability term. However, this comes at a hit to performance. By 2027, another line of work is humans studying model COTs in detail and learning the ways it thinks; some mathematicians in particular pick up neat mental tricks from studying the COTs of models. However, overall COT interpretability declines, and it's generally accepted we won't know exactly what the models are thinking or why, even if COT analysis and the new interpretability techniques can give some general understanding in 2027.</p><p>By 2027, evaluations are showing that frontier models&#8212;including open-source models&#8212;could meaningfully help in engineering pandemics, if bad actors so chose. There's a messy but moderately effective effort by AI safety organisations and several agencies within governments to have some sort of misuse mitigation measures in place, in particular for API-accessible models. However, in the absence of a major incident, governments don't care enough, and open-source models seem hard to contain. Also, some bioterrorism continues being blocked by wet lab skills and just continuing good luck regarding the absence of a motivated bioterrorist. The other potentially catastrophic source of misuse is cyber, but it increasingly seems like this will be solved by default, in particular because AIs are good at writing secure code and formal verification is increasingly used for critical code.</p><p>The previous year of insane AI codegen stuff going on everywhere and the continued steady progress in AI has made it more intuitive to people that there won&#8217;t be a lot of &#8220;money on the table&#8221; for some nascent AGI to eat up, because it will enter a teeming ecosystem of AI systems and humans and their interactions. For example, though there are technically some self-sustaining AIs paying for their server costs, they struggle to compete with purposeful human+AI entities that deliberately try to steal the customers of the AI-only businesses if they ever get too many. The cyber competition is also increasingly tough, meaning that any single rogue AI would have a rough time defeating the rest of the world. However, no evidence by the end of 2027 has ruled out a sharper takeoff, and those who believe in it are increasingly either frantic and panicking, or then stoically equanimous and resigned, expecting the final long-term agentic planning piece to slot into place at any moment and doom the world. Also, the labs are openly talking about recursive self-improvement as their strategy.</p><h2><strong>Government wakeup</strong></h2><p>In 2025 in the Chinese government, thinking about AGI is stuck somewhere inside CCP internal machinations. Xi Jinping has heard of it and occasionally thinks about it, but doesn&#8217;t take it seriously as a near-term thing. However, some senior staff are properly "AGI-pilled" (and split between advocates of safety and racing, but without an overwhelming favourite yet, though also it&#8217;s clear that once the issue does get serious, national security will be by far the loudest voice in the room).</p><p>The EU is preparing to burden its (small) AI sector with bureaucracy (the EU AI Act). In 2025-2026, there are some high-profile cases of AI companies not offering services in Europe due to regulations. However, the EU &amp; UK governments are best at tracking AI because of their efforts on safety research through the EU AI Office and the UK AISI. The UK government also announces planning reforms targeted at letting private companies build data centres more easily.</p><p>Going into 2026, India is trying to ride a new IT wave using open-sourced Llama, DeepSeek, and Mistral models to create products to try to replace its current IT services -heavy export industry. A lack of domestic data centres is making this harder, and political tangles make fast movement hard.</p><p>In 2025, the Trump administration tightens export controls on China (in particular preventing even older Nvidia chips like the H20 from being sold to China), tries to pressure other countries to not buy Chinese GPUs, and makes it easier to build power (especially nuclear &amp; gas) and data centres within the US. Otherwise there is little US political action on AI. Behind the scenes, the defense establishment gets more involved in the AI scene. There are secret NSA and CIA projects researching AI for offensive &amp; defensive cyber. More Chinese infiltration of American cyber systems is discovered. High-level government conversations behind closed doors are upping the apocalyptic rhetoric about how essential it is for the US to win in AI-powered cyber. All the major US AI labs have some government partnership related to this.</p><p>As internal CCP machinations grind along, and the evidence about big AI effects on programming rolls in through late 2025 and 2026, the CCP gets more serious about AI. Like in the US, once the strategic and national security implications rise in salience, other issues (including safety) fall. The CCP prepares their 15th Five Year Plan for 2026, which involves massive subsidies and investment for AI. DeepSeek leads the domestic AI industry, but the CCP has made it clear they will make the big calls. There is a conversation behind closed doors about whether to end the open-sourcing of DeepSeek models, but the CCP comes out in favour, in particular to try to get the rest of the world to build on top of Chinese AI models (and also helped by the press that the early 2025 DeepSeek R1 release caused). Huawei is shipping GPUs that are only about 12-16 months behind Nvidia. China&#8217;s worse startup ecosystem means that AI agent adoption is slower than in the US, though. However, China&#8217;s surveillance state has been on an AI adoption spree. In particular, censorship is instantaneous with LLMs. By 2026, there are widespread educational "Xi Jinping Thought AI Tutors" that most CCP members are mandated to have weekly sessions with. Retaining control of society now seems increasingly easy, allowing the CCP to focus more on geopolitics and military, and less on the consumer economy. At the same time, Xi Jinping has an overly-rosy view of Chinese military AI capabilities because people tell him what he wants to hear.</p><p>There's a shadow conflict playing out, almost entirely out of public attention, between US and Chinese cyber forces trying to get into each other's critical infrastructure while reducing the extent to which their own infrastructure is compromised. Contrary to publicly-available information, America probably has the upper hand, but it's also clear that both could inflict high damage on the other.</p><p>AI starts to figure in US domestic politics in 2026, but is not yet a top issue. The upcoming replacement of most human white-collar work looks more and more plausible, especially after OpenAI's release of o6. Job losses are not yet high, though, as human orgs take time to react to change. Even in software, where mass firings could perhaps most plausibly be done, many are afraid to try it first. Non-technical managers tend to treat the technical stuff as blackbox wizardry and are scared to break it, and technical managers don't want to reduce the size of their empires. The main effect is that hiring new software engineers basically stops, but the disaffected&#8212;a small group of nerdy, elite-coded, low-voting-rate youngsters&#8212;are not politically important. Other white-collar office jobs are also reducing entry-level hiring, as increased demand for productivity is instead met by existing employees just using AI more.</p><p>The US government, like China, decides against legally restricting the open-sourcing of AI models. This is influenced by pro-innovation arguments, China doing the same, and the defense-specific AI programs being done under classification with closed-source models anyway. The AI labs have also grown more reliant on government cooperation for things like power grid connection permits, data centre construction permits, and lobbying to avoid ruinous tariffs on GPUs. They also all want the money flow of Pentagon contracts and the prestige of working on US defense. This means that there's a tacit agreement that if the government hints they should or shouldn't do something, they are very likely to march to that beat.</p><p>Starting in late 2026, many of the governments worried about fertility decline get concerned about everyone talking to AIs instead of each other. South Korea bans &#8220;personalised AI companions&#8221; in 2027, and the EU requires people to register if they use them and imposes various annoying limits that drive down usage. However, the addicts can just use open-source models to circumvent regulations. Some countries spend lots of money on getting the "creator influencers"&#8212;influencers turbo-charged by generative AI&#8212;to extol the virtues of marriage and kids. By 2027, though, the more forward-looking politicians are&#8212;in private&#8212;starting to realise that once the economy transitions to being AI-powered, national interests are not harmed if the human population plummets. The &#8220;<a href="https://lukedrago.substack.com/p/the-intelligence-curse">intelligence curse</a>&#8221; is starting to set in.</p><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.nosetgauge.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">To hear more about the AI future, years before it actually happens, subscribe for free email updates:</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p><p><a href="https://nosetgauge.substack.com/p/a-history-of-the-future-2027-2030">Part 2: 2027-2030</a></p><p><a href="https://nosetgauge.substack.com/p/a-history-of-the-future-2030-2040">Part 3: 2030-2040</a></p><p></p><p><em>Thanks to <a href="https://lukedrago.substack.com/">Luke Drago</a>, Duncan McClements, Theo Horsley, and Bilal Chughtai for comments.</em></p>]]></content:encoded></item><item><title><![CDATA[Capital, AGI, and human ambition]]></title><description><![CDATA[AGI will shift the relative importance of human v non-human factors of production, reducing the incentive to care about humans while making existing powers more effective and entrenched]]></description><link>https://www.nosetgauge.com/p/capital-agi-and-human-ambition</link><guid isPermaLink="false">https://www.nosetgauge.com/p/capital-agi-and-human-ambition</guid><dc:creator><![CDATA[Rudolf Laine]]></dc:creator><pubDate>Sat, 28 Dec 2024 17:50:20 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/f96187ce-99aa-4b16-be2a-cca2a2d0fef3_1024x1024.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em><strong>A modified version of this essay is now part of a much more comprehensive essay series, <a href="https://intelligence-curse.ai/">The Intelligence Curse</a>.</strong></em></p><div><hr></div><p><strong>Edited to add</strong><em><strong>: </strong>The main takeaway of this post is meant to be: <strong>Labour-replacing AI will shift the relative importance of human v non-human factors of production, which reduces the incentives for society to care about humans while making existing powers more effective and entrenched.</strong> Many people are reading this post in a way where either (a) "capital" means just "money" (rather than also including physical capital like factories and data centres), or (b) the main concern is human-human inequality (rather than broader societal concerns about humanity's collective position, the potential for social change, and human agency).</em></p><div><hr></div><p>I've heard many people say something like "money won't matter post-AGI". This has always struck me as odd, and as most likely completely incorrect.</p><p>First: labour means human mental and physical effort that produces something of value. Capital goods are things like factories, data centres, and software&#8212;things humans have built that are used in the production of goods and services. I'll use "capital" to refer to both the stock of capital goods and to the money that can pay for them. I'll say "money" when I want to exclude capital goods.</p><p>The key economic effect of AI is that it makes capital a more and more general substitute for labour. There's less need to pay humans for their time to perform work, because you can replace that with capital (e.g. data centres running software replaces a human doing mental labour).</p><p>I will walk through consequences of this, and end up concluding that labour-replacing AI means:</p><ol><li><p>The ability to buy results in the real world will dramatically go up</p></li><li><p>Human ability to wield power in the real world will dramatically go down (at least without money); including because:</p><ol><li><p>there will be no more incentive for states, companies, or other institutions to care about humans</p></li><li><p>it will be harder for humans to achieve outlier outcomes relative to their starting resources</p></li></ol></li><li><p>Radical equalising measures are unlikely</p></li></ol><p>Overall, this points to a neglected downside of transformative AI: that society might become permanently static, and that current power imbalances might be amplified and then turned immutable.</p><p>Given sufficiently strong AI, <em>this is not a risk about insufficient material comfort</em>. Governments could institute UBI with the AI-derived wealth. Even if e.g. only the United States captures AI wealth and the US government does nothing for the world, if you're willing to assume arbitrarily extreme wealth generation from AI, the wealth of the small percentage of wealthy Americans who care about causes outside the US might be enough to end material poverty (if 1% of American billionaire wealth was spent on wealth transfers to foreigners, it would take 16 doublings of American billionaire wealth as expressed in purchasing-power-for-human-needs&#8212;a roughly 70,000x increase&#8212;before they could afford to give $500k-equivalent to every person on Earth; in a singularity scenario where the economy's doubling time is months, this would not take long). Of course, if the AI explosion is less singularity-like, or if the dynamics during AI take-off actively disempower much of the world's population (a real possibility), even material comfort could be an issue.</p><p>What most emotionally moves me about these scenarios is that a static society with a locked-in ruling caste does not seem dynamic or alive to me. We should not kill human ambition, if we can help it.</p><p>There are also ways in which such a state makes <a href="https://nosetgauge.substack.com/p/a-disneyland-without-children">slow-rolling, gradual AI catastrophes</a> more likely, because the incentive for power to care about humans is reduced.</p><h2><strong>The default solution</strong></h2><p>Let's assume human mental and physical labour across the vast majority of tasks that humans are currently paid wages for no longer has non-trivial market value, because the tasks can be done better/faster/cheaper by AIs. Call this labour-replacing AI.</p><p>There are two levels of the standard solution to the resulting unemployment problem:</p><ol><li><p>Governments will adopt something universal basic income (UBI).</p></li><li><p>We will quickly hit superintelligence, and, assuming the superintelligence is aligned, live in a post-scarcity technological wonderland where everything is possible.</p></li></ol><p>Note, firstly, that money will continue being a thing, at least unless we have one single AI system doing all economic planning. <a href="https://www.econlib.org/library/Essays/hykKnw.html">Prices are largely about communicating information</a>. If there are many actors and they trade with each other, the strong assumption should be that there are prices (even if humans do not see them or interact with them). Remember too that however sharp the singularity, abundance will still be finite, and must therefore be allocated.</p><h2><strong>Money currently struggles to buy talent</strong></h2><p>Money can buy you many things: capital goods, for example, can usually be bought quite straightforwardly, and cannot be bought without a lot of money (or other liquid assets, or non-liquid assets that others are willing to write contracts against, or special government powers). But it is surprisingly hard to convert raw money into labour, in a way that is competitive with top labour.</p><p>Consider Blue Origin versus SpaceX. Blue Origin was started two years earlier (2000 v 2002), had much better funding for most of its history, and even today employs almost as many people as SpaceX (11,000 v 13,000). Yet SpaceX has crushingly dominated Blue Origin. In 2000, Jeff Bezos had $4.7B at hand. But it is hard to see what he could've done to not lose out to the comparatively money-poor SpaceX with its intense culture and outlier talent.</p><p>Consider, a century earlier, the Wright brothers with their bike shop resources beating Samuel Langley's well-funded operation.</p><p>Consider the stereotypical VC-and-founder interaction, or the acquirer-and-startup interaction. In both cases, holders of massive financial capital are willing to pay very high prices to bet on labour&#8212;and the bet is that the labour of the few people in the startup will beat extremely large amounts of capital.</p><p>If you want to convert money into results, the deepest problem you are likely to face is hiring the right talent. And that comes with several problems:</p><ol><li><p>It's often hard to judge talent, unless you yourself have considerable talent in the same domain. Therefore, if you try to find talent, you will often miss.</p></li><li><p>Talent is rare (and credentialed talent even more so&#8212;and many actors can't afford to rely on any other kind, because of point 1), so there's just not very much of it going around.</p></li><li><p>Even if you can locate the top talent, the top talent tends to be <em>less</em> amenable to being bought out by money than others.</p></li></ol><p>(Of course, those with money keep building infrastructure that makes it easier to convert money into results. I have seen first-hand the largely-successful quest by quant finance companies to strangle out all existing <a href="https://space.ong.ac/escaping-flatland">ambition</a> out of top UK STEM grads and replace it with the eking of tiny gains in financial markets. Mammon must be served!)</p><p>With labour-replacing AI, these problems go away.</p><p>First, <em>you</em> might not be able to judge AI talent. Even the AI evals ecosystem might find it hard to properly judge AI talent&#8212;evals are hard. Maybe even the informal word-of-mouth mechanisms that correctly sung praises of Claude-3.5-Sonnet far more decisively than any benchmark might find it harder and harder to judge which AIs really are best as AI capabilities keep rising. But the real difference is that the AIs can be cloned. Currently, huge pools of money chase after a single star researcher who's made a breakthrough, and thus had their talent made legible to those who control money (who can judge the clout of the social reception to a paper but usually can't judge talent itself directly). But the star researcher that is an AI can just be cloned. Everyone&#8212;or at least, everyone with enough money to burn on GPUs&#8212;gets the AI star researcher. No need to sort through the huge variety of unique humans with their unproven talents and annoying inability to be instantly cloned. This is the main reason why it will be easier for money to find top talent once we have labour-replacing AIs.</p><p>Also, of course, the price of talent will go down massively, because the AIs will be cheaper than the equivalent human labour, and because competition will be fiercer because the AIs can be cloned.</p><p>The final big bottleneck for converting money into talent is that lots of top talent has complicated human preferences that make them hard to buy out. The top artist has an artistic vision they're genuinely attached to. The top mathematician has a deep love of elegance and beauty. The top entrepreneur has deep conviction in what they're doing&#8212;and probably wouldn't function well as an employee anyway. Talent and performance in humans are surprisingly tied to a <a href="https://www.benlandautaylor.com/p/looking-beyond-the-veil">sacred bond to a discipline or mission</a> (a fact that the world's cynics / careerists / Roman Empires like to downplay, only to then find their lunch eaten by the ambitious interns / SpaceXes / Christianities of the world). In contrast, AIs exist specifically so that they <em>can</em> be trivially bought out (at least within the bounds of their safety training). The genius AI mathematician, unlike the human one, will happily spend its limited time on Earth proving the correctness of schlep code.</p><p>Finally (and obviously), the AIs will eventually be much more capable than any human employees at their tasks.</p><p>This means that <em>the ability of money to buy results in the real world will dramatically go up once we have labour-replacing AI</em>.</p><h2><strong>Most people's power/leverage derives from their labour</strong></h2><p>Labour-replacing AI also deprives almost everyone of their main lever of power and leverage. Most obviously, if you're the average Joe, you have money because someone somewhere pays you to spend your mental and/or physical efforts solving their problems.</p><p>But wait! We assumed that there's UBI! Problem solved, right?</p><h3><strong>Why are states ever nice?</strong></h3><p>UBI is granted by states that care about human welfare. There are many reasons why states care and might care about human welfare.</p><p>Over the past few centuries, there's been a big shift towards states caring more about humans. Why is this? We can examine the reasons to see how durable they seem:</p><ol><li><p>Moral changes downstream of the Enlightenment, in particular an increased centering of liberalism and individualism.</p></li><li><p>Affluence &amp; technology. Pre-industrial societies were mostly so poor that significant efforts to help the poor would've bankrupted them. Many types of help (such as effective medical care) are also only possible because of new technology.</p></li><li><p>Incentives for states to care about freedom, prosperity, and education.</p></li></ol><p>AI will help a lot with the 2nd point. It will have some complicated effect on the 1st. But here I want to dig a bit more into the 3rd, because I think this point is unappreciated.</p><p>Since the industrial revolution, the interests of states and people have been unusually aligned. To be economically competitive, a strong state needs efficient markets, a good education system that creates skilled workers, and a prosperous middle class that creates demand. It benefits from using talent regardless of its class origin. It also benefits from allowing high levels of freedom to foster science, technology, and the arts &amp; media that result in global soft-power and cultural influence. Competition between states largely pushes <em>further</em> in all these directions&#8212;consider the success of the US, or how even the CCP is pushing for efficient markets and educated rich citizens, and faces incentives to allow some freedoms for the sake of Chinese science and startups. Contrast this to the feudal system, where the winning strategy was building an extractive upper class to rule over a population of illiterate peasants and spend a big share of extracted rents on winning wars against nearby states. For more, see <a href="https://nosetgauge.substack.com/p/review-foragers-farmers-and-fossil-fuels">my review of </a><em><a href="https://nosetgauge.substack.com/p/review-foragers-farmers-and-fossil-fuels">Foragers, Farmers, and Fossil Fuels</a></em>, or my post on the connection between <a href="https://nosetgauge.substack.com/p/growth-and-civilisation">moral values and economic growth</a>.</p><p>With labour-replacing AI, the incentives of states&#8212;in the sense of what actions states should take to maximise their competitiveness against other states and/or their own power&#8212;will no longer be aligned with humans in this way. The incentives might be better than during feudalism. During feudalism, the incentive was to extract as much as possible from the peasants without them dying. After labour-replacing AI, humans will be less a resource to be mined and more just irrelevant. However, spending fewer resources on humans and more on the AIs that sustain the state's competitive advantage will still be incentivised.</p><p>Humans will also have much less leverage over states. Today, if some important sector goes on strike, or if some segment of the military threatens a coup, the state has to care, because its power depends on the buy-in of at least some segments of the population. People can also credibly tell the state things like "invest in us and the country will be stronger in 10 years". But once AI can do all the labour that keeps the economy going and the military powerful, the state has no more <em>de facto</em> reason to care about the demands of its humans.</p><p>Adam Smith could <a href="https://oll.libertyfund.org/quotes/adam-smith-butcher-brewer-baker">write</a> that his dinner doesn't depend on the benevolence of the butcher or the brewer or the baker. The classical liberal today can credibly claim that the arc of history really does bend towards freedom and plenty for all, not out of the benevolence of the state, but because of the incentives of capitalism and geopolitics. But after labour-replacing AI, this will no longer be true. If the arc of history keeps bending towards freedom and plenty, it will do so only out of the benevolence of the state (or the AI plutocrats). If so, we better lock in that benevolence while we have leverage&#8212;and have a good reason why we expect it to stand the test of time.</p><p>The best thing going in our favour is democracy. It's a huge advantage that a deep part of many of the modern world's strongest institutions (i.e. Western democracies) is equal representation of every person. However, only about <a href="https://ourworldindata.org/grapher/people-living-in-democracies-autocracies">13% of the world's population lives in a liberal democracy</a>, which creates concerns both about the fate of the remaining 87% of the world's people (especially the 27% in closed autocracies). It also creates potential for <a href="https://slatestarcodex.com/2014/07/30/meditations-on-moloch/">Molochian</a> competition between humanist states and less scrupulous states that might drive down the resources spent on human flourishing to zero over a sufficiently long timespan of competition.</p><p>I focus on states above, because states are the strongest and most durable institutions today. However, similar logic applies if, say, companies or some entirely new type of organisation become the most important type of institution.</p><h3><strong>No more outlier outcomes?</strong></h3><p>Much change in the world is driven by people who start from outside money and power, achieve outlier success, and then end up with money and/or power. This makes sense, since those with money and/or power rarely have the fervour to push for big changes, since they are exactly those who are best served by the status quo.</p><p>Whatever your opinions on income inequality or any particular group of outlier successes, I hope you agree with me that the possibility of someone achieving outlier success and changing the world is important for avoiding stasis and generally having a world that is interesting to live in.</p><p>Let's consider the effects of labour-replacing AI on various routes to outlier success through labour.</p><p><strong>Entrepreneurship</strong> is increasingly what <a href="https://medium.com/entrepreneur-first/tech-entrepreneurship-and-the-disruption-of-ambition-4e6854121992">Matt Clifford calls the "technology of ambition" of choice</a> for ambitious young people (at least those with technical talent and without a disposition for politics). Right now, entrepreneurship has become easier. AI tools can already make small teams much more effective without needing to hire new employees. They also reduce the entry barrier to new skills and fields. However, labour-replacing AI makes the tenability of entrepreneurship uncertain. There is some narrow world in which AIs remain mostly tool-like and entrepreneurs can succeed long after most human labour is automated because they provide agency and direction. However, it also seems likely that sufficiently strong AI will by default obsolete human entrepreneurship. For example, VC funds might be able to directly convert money into hundreds of startup attempts all run by AIs, without having to go through the intermediate route of finding a human entrepreneurs to manage the AIs for them.</p><p><strong>The hard sciences</strong>. The era of human achievement in hard sciences will probably end within a few years because of the rate of AI progress in anything with crisp reward signals.</p><p><strong>Intellectuals.</strong> Keynes, Friedman, and Hayek all did technical work in economics, but their outsize influence came from the worldviews they developed and sold (especially in Hayek's case), which made them more influential than people like Paul Samuelson who dominated mathematical economics. John Stuart Mill, John Rawls, and Henry George were also influential by creating frames, worldviews, and philosophies. The key thing that separates such people from the hard scientists is that the outputs of their work are not spotlighted by technical correctness alone, but require moral judgement as well. Even if AI is superhumanly persuasive and correct, there's some uncertainty about how AI work in this genre will fit into <a href="https://nosetgauge.substack.com/p/ai-and-wisdom-3-ai-effects-on-amortised">the way that human culture picks and spreads ideas</a>. Probably it doesn't look good for human intellectuals. I suspect that a lot of why intellectuals' ideologies can have so much power is that they're products of genius in a world where genius is rare. A flood of AI-created ideologies might mean that no individual ideology, and certainly no human one, can shine so bright anymore. The world-historic intellectual might go extinct.</p><p><strong>Politics</strong> might be one of the least-affected options, since I'd guess that most humans specifically want a human to do that job, and because politicians get to set the rules for what's allowed. The charisma of AI-generated avatars, and a general dislike towards politicians at least in the West, might throw a curveball here, though. It's also hard to say whether incumbents will be favoured. AI might bring down the cost of many parts of political campaigning, reducing the resource barrier to entry. However, if AI too expensive for small actors is meaningfully better than cheaper AI, this would favour actors with larger resources. I expect these direct effects to be smaller than the indirect effects from whatever changes AI has on the memetic landscape.</p><p>Also, the real play is not to go into actual politics, where a million other politically-talented people are competing to become president or prime minister. Instead, have political skill and go somewhere outside government where political skill is less common (c.f. Sam Altman). Next, wait for the arrival of hyper-competent AI employees that reduce the demands for human subject-matter competence while increasing the rewards for winning political games within that organisation.</p><p><strong>Military</strong> success as a direct route to great power and disruption has&#8212;for the better&#8212;not really been a thing since Napoleon. Advancing technology increases the minimum industrial base for a state-of-the-art army, which benefits incumbents. AI looks set to be controlled by the most powerful countries. One exception is if coups of large countries become easier with AI. Control over the future AI armies will likely be both (a) more centralised than before (since a large number of people no longer have to go along for the military to take an action), and (b) more tightly controllable than before (since the permissions can be implemented in code rather than human social norms). These two factors point in different directions so it's uncertain what the net effect on coup ease will be. Another possible exception is if a combination of revolutionary tactics and cheap drones enables a Napoleon-of-the-drones to win against existing armies. Importantly, though, neither of these seems likely to promote the <em>good</em> kind of disruptive challenge to the status quo.</p><p><strong>Religions</strong>. When it comes to rising rank in existing religions, the above takes on politics might be relevant. When it comes to starting new religions, the above takes on intellectuals might be relevant.</p><p>So on net, sufficiently strong labour-replacing AI will be on-net bad for the chances of every type of outlier human success, with perhaps the weakest effects in politics. This is despite the very real boost that current AI has on entrepreneurship.</p><p>All this means that <em>the ability to get and wield power in the real world without money will dramatically go down once we have labour-replacing AI.</em></p><h2><strong>Enforced equality is unlikely</strong></h2><p><em>The Great Leveler</em> is a <a href="https://rudolf.website/short-reviews-nonfiction-1/#section-3">good book</a> on the history of inequality that (<a href="https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4939891">at least per the author</a>) has survived its critiques fairly well. Its conclusion is that past large reductions in inequality have all been driven by one of the "Four Horsemen of Leveling": total war, violent revolution, state collapse, and pandemics. Leveling income differences has historically been hard enough to basically never happen through conscious political choice.</p><p>Imagine that labour-replacing AI is here. UBI is passed, so no one is starving. There's a massive scramble between countries and companies to make the best use of AI. This is all capital-intensive, so everyone needs to woo holders of capital. The top AI companies wield power on the level of states. The redistribution of wealth is unlikely to end up on top of the political agenda.</p><p>An exception might be if some new political movement or ideology gets a lot of support quickly, and is somehow boosted by some unprecedented effect of AI (such as: no one has jobs anymore so they can spend all their time on politics, or there's some new AI-powered coordination mechanism).</p><p>Therefore, even if the future is a glorious transhumanist utopia, it is unlikely that people will be starting in it at an equal footing. Due to the previous arguments, it is also unlikely that they will be able to greatly change their relative footing later on.</p><p>Consider also equality between states. Some states stand set to benefit massively more than others from AI. Many equalising measures, like UBI, would be difficult for states to extend to non-citizens under anything like the current political system. This is true even of the United States, the most liberal and humanist great power in world history. By default, the world order might therefore look (even more than today) like a global caste system based on country of birth, with even fewer possibilities for immigration (because the main incentive to allow immigration is its massive economic benefits, which only exist when humans perform economically meaningful work).</p><h2><strong>The default outcome?</strong></h2><p>Let's grant the assumptions at the start of this post and the above analysis. Then, the post-labour-replacing-AI world involves:</p><ul><li><p>Money will be able to buy results in the real world better than ever.</p></li><li><p>People's labour gives them less leverage than ever before.</p></li><li><p>Achieving outlier success through your labour in most or all areas is now impossible.</p></li><li><p>There was no transformative leveling of capital, either within or between countries.</p></li></ul><p>This means that those with significant capital when labour-replacing AI started have a permanent advantage. They will wield more power than the rich of today&#8212;not necessarily <em>over</em> people, to the extent that liberal institutions remain strong, but at least over physical and intellectual achievements. Upstarts will not defeat them, since capital now trivially converts into superhuman labour in any field.</p><p>Also, there will be no more incentive for whatever institutions wield power in this world to care about people in order to maintain or grow their power, because all real power will flow from AI. There might, however, be significant lock-in of liberal humanist values through political institutions. There might also be significant lock-in of people's purchasing power, if everyone has meaningful UBI (or similar), and the economy retains a human-oriented part.</p><p>In the best case, this is a world like a more unequal, unprecedentedly static, and much richer Norway: a massive pot of non-human-labour resources (oil :: AI) has benefits that flow through to everyone, and yes some are richer than others but everyone has a great standard of living (and <a href="https://nosetgauge.substack.com/p/death-is-bad">ideally</a> also lives forever). The only realistic forms of human ambition are playing local social and political games within your social network and class. If you don't have a lot of capital (and maybe not even then), you don't have a chance of affecting the broader world anymore. Remember: the AIs are better poets, artists, philosophers&#8212;everything; why would anyone care what some human does, unless that human is someone they personally know? Much like in feudal societies the answer to "why is this person powerful?" would usually involve some long family history, perhaps ending in a distant ancestor who had fought in an important battle ("my great-great-grandfather fought at Bosworth Field!"), anyone of importance in the future will be important because of something they or someone they were close with did in the pre-AGI era ("oh, my uncle was technical staff at OpenAI"). The children of the future will live their lives in the shadow of their parents, with social mobility extinct. I think you should definitely feel a non-zero amount of existential horror at this, even while acknowledging that it could've gone a lot worse.</p><p>In a worse case, AI trillionaires have near-unlimited and unchecked power, and there's a permanent aristocracy that was locked in based on how much capital they had at the time of labour-replacing AI. The power disparities between classes might make modern people shiver, much like modern people consider feudal status hierarchies grotesque. But don't worry&#8212;much like <a href="https://nosetgauge.substack.com/p/review-foragers-farmers-and-fossil-fuels">the feudal underclass mostly accepted their world order</a> due to their culture even without superhumanly persuasive AIs around, the future underclass will too.</p><p>In the absolute worst case, humanity goes extinct, potentially because of a <a href="https://nosetgauge.substack.com/p/a-disneyland-without-children">slow-rolling optimisation for AI power over human prosperity</a> over a long period of time. Because that's what the power and money incentives will point towards.</p><h2><strong>What's the takeaway?</strong></h2><p>If you read this post and accept a job at a quant finance company as a result, I will be sad. If you were about to do something ambitious and impactful about AI, and read this post and accept a job at Anthropic to accumulate risk-free personal capital while counterfactually helping out a bit over the marginal hire, I can't fault you too much, but I will still be slightly sad.</p><p>It's of course true that the above increases the stakes of medium-term (~2-10 year) personal finance, and you should consider this. But it's also true that <em>right now</em> is a great time to do something ambitious. Robin Hanson calls the present <a href="https://www.overcomingbias.com/p/this-is-the-dream-timehtml">"the dreamtime"</a>, following a concept in Aboriginal myths: the time when the future world order and its values are still liquid, not yet set in stone.</p><p>Previous upheavals&#8212;the various waves of industrialisation, the internet, etc.&#8212;were great for human ambition. With AI, we could have the last and greatest opportunity for human ambition&#8212;followed shortly by its extinction for all time. How can your reaction not be: <a href="https://en.wikipedia.org/wiki/Carpe_diem">"carpe diem"</a>?</p><p>We should also try to preserve the world's dynamism.</p><p>Rationalist thought on post-AGI futures is too solutionist. The strawman version: solve morality, solve AI, figure out the optimal structure to tile the universe with, do that, done. (The actual leading figures have far less strawman views; see e.g. <a href="https://www.dwarkeshpatel.com/p/paul-christiano">Paul Christiano at 23:30 here</a>&#8212;but the on-the-ground culture <em>does</em> lean in the strawman direction.)</p><p>I think it's much healthier for society and its development to be a shifting, dynamic thing where the ability, as an individual, to add to it or change it remains in place. And that means keeping the potential for successful ambition&#8212;and the resulting disruption&#8212;alive.</p><p>How do we do this? I don't know. But I don't think you should see the approach of powerful AI as a blank inexorable wall of human obsolescence, consuming everything equally and utterly. There will be cracks in the wall, at least for a while, and they will look much bigger up close once we get there&#8212;or if you care to look for them hard enough from further out&#8212;than from a galactic perspective. As AIs get closer and closer to a Pareto improvement over all human performance, though, I expect we'll eventually need to augment ourselves to keep up.</p><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.nosetgauge.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe to get future posts</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Review: Planecrash]]></title><description><![CDATA[Take a stereotypical fantasy novel, a textbook on mathematical logic, and Fifty Shades of Grey. Mix them all together and add extra weirdness for spice. The result might look a lot like Planecrash (AKA: Project Lawful), a work of fiction co-written by "Iarwain" (a pen-name of Eliezer Yudkowsky) and "lintamande".]]></description><link>https://www.nosetgauge.com/p/review-planecrash</link><guid isPermaLink="false">https://www.nosetgauge.com/p/review-planecrash</guid><dc:creator><![CDATA[Rudolf Laine]]></dc:creator><pubDate>Fri, 27 Dec 2024 14:17:09 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/8df9c658-9f11-42ae-9efe-1323aca28185_295x324.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Take a stereotypical fantasy novel, a textbook on mathematical logic, and <em>Fifty Shades of Grey</em>. Mix them all together and add extra weirdness for spice. The result might look a lot like <a href="https://www.lesswrong.com/posts/SA9hDewwsYgnuscae/projectlawful-com-eliezer-s-latest-story-past-1m-words">Planecrash</a> (AKA: Project Lawful), a work of fiction co-written by "Iarwain" (a pen-name of Eliezer Yudkowsky) and "lintamande".</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!kwQl!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d62e620-9183-40a7-8f24-0505f1915e13_295x324.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!kwQl!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d62e620-9183-40a7-8f24-0505f1915e13_295x324.png 424w, https://substackcdn.com/image/fetch/$s_!kwQl!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d62e620-9183-40a7-8f24-0505f1915e13_295x324.png 848w, https://substackcdn.com/image/fetch/$s_!kwQl!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d62e620-9183-40a7-8f24-0505f1915e13_295x324.png 1272w, https://substackcdn.com/image/fetch/$s_!kwQl!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d62e620-9183-40a7-8f24-0505f1915e13_295x324.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!kwQl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d62e620-9183-40a7-8f24-0505f1915e13_295x324.png" width="295" height="324" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4d62e620-9183-40a7-8f24-0505f1915e13_295x324.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:324,&quot;width&quot;:295,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:87674,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!kwQl!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d62e620-9183-40a7-8f24-0505f1915e13_295x324.png 424w, https://substackcdn.com/image/fetch/$s_!kwQl!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d62e620-9183-40a7-8f24-0505f1915e13_295x324.png 848w, https://substackcdn.com/image/fetch/$s_!kwQl!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d62e620-9183-40a7-8f24-0505f1915e13_295x324.png 1272w, https://substackcdn.com/image/fetch/$s_!kwQl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4d62e620-9183-40a7-8f24-0505f1915e13_295x324.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>(image credit: Planecrash)</em></figcaption></figure></div><p>Yudkowsky is not afraid to be verbose and self-indulgent in his writing. He previously wrote a <a href="https://hpmor.com/">Harry Potter fanfic</a> that includes what's essentially an extended Ender's Game fanfic in the middle of it, because why not. In Planecrash, it starts with the very format: it's written as a series of <a href="https://www.projectlawful.com/board_sections/703">forum posts</a> (though there are <a href="https://www.mikescher.com/blog/29/Project_Lawful_ebook">ways to get an ebook</a>). It continues with maths lectures embedded into the main arc, totally plot-irrelevant tangents that are just Yudkowsky <a href="https://www.glowfic.com/posts/5826">ranting about frequentist statistics</a>, and one instance of Yudkowsky hijacking the plot for a few pages to <a href="https://www.glowfic.com/posts/6132?page=49">soapbox about his pet Twitter feuds</a> (with transparent in-world analogues for Effective Altruism, TPOT, and the post-rationalists). Planecrash does not aspire to be high literature. Yudkowsky is self-aware of this, and uses it to troll big-name machine learning researchers:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!K6HD!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2144e92-9979-4d2d-ae52-c12af49b3004_952x342.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!K6HD!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2144e92-9979-4d2d-ae52-c12af49b3004_952x342.png 424w, https://substackcdn.com/image/fetch/$s_!K6HD!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2144e92-9979-4d2d-ae52-c12af49b3004_952x342.png 848w, https://substackcdn.com/image/fetch/$s_!K6HD!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2144e92-9979-4d2d-ae52-c12af49b3004_952x342.png 1272w, https://substackcdn.com/image/fetch/$s_!K6HD!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2144e92-9979-4d2d-ae52-c12af49b3004_952x342.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!K6HD!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2144e92-9979-4d2d-ae52-c12af49b3004_952x342.png" width="952" height="342" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b2144e92-9979-4d2d-ae52-c12af49b3004_952x342.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:342,&quot;width&quot;:952,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:75378,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!K6HD!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2144e92-9979-4d2d-ae52-c12af49b3004_952x342.png 424w, https://substackcdn.com/image/fetch/$s_!K6HD!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2144e92-9979-4d2d-ae52-c12af49b3004_952x342.png 848w, https://substackcdn.com/image/fetch/$s_!K6HD!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2144e92-9979-4d2d-ae52-c12af49b3004_952x342.png 1272w, https://substackcdn.com/image/fetch/$s_!K6HD!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2144e92-9979-4d2d-ae52-c12af49b3004_952x342.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!HlEV!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc82f799f-ecb8-4e42-bd01-1274b61b2f2f_972x276.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!HlEV!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc82f799f-ecb8-4e42-bd01-1274b61b2f2f_972x276.png 424w, https://substackcdn.com/image/fetch/$s_!HlEV!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc82f799f-ecb8-4e42-bd01-1274b61b2f2f_972x276.png 848w, https://substackcdn.com/image/fetch/$s_!HlEV!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc82f799f-ecb8-4e42-bd01-1274b61b2f2f_972x276.png 1272w, https://substackcdn.com/image/fetch/$s_!HlEV!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc82f799f-ecb8-4e42-bd01-1274b61b2f2f_972x276.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!HlEV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc82f799f-ecb8-4e42-bd01-1274b61b2f2f_972x276.png" width="972" height="276" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c82f799f-ecb8-4e42-bd01-1274b61b2f2f_972x276.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:276,&quot;width&quot;:972,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:75010,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!HlEV!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc82f799f-ecb8-4e42-bd01-1274b61b2f2f_972x276.png 424w, https://substackcdn.com/image/fetch/$s_!HlEV!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc82f799f-ecb8-4e42-bd01-1274b61b2f2f_972x276.png 848w, https://substackcdn.com/image/fetch/$s_!HlEV!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc82f799f-ecb8-4e42-bd01-1274b61b2f2f_972x276.png 1272w, https://substackcdn.com/image/fetch/$s_!HlEV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc82f799f-ecb8-4e42-bd01-1274b61b2f2f_972x276.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>(<a href="https://x.com/ESYudkowsky/status/1654171832994365475">source</a>)</em></figcaption></figure></div><p>Why would anyone ever read Planecrash? I read (admittedly&#8212;sometimes skimmed) it, and I see two reasons:</p><ol><li><p>The characters are competent in a way that characters in fiction rarely are. Yudkowsky is good at writing <a href="https://yudkowsky.tumblr.com/writing">intelligent characters</a> in a specific way that I haven't seen anyone else do as well. Lintamande writes a uniquely compelling story of determination and growth in an extremely competent character.</p></li><li><p>More than anyone else I've yet read, Yudkowsky has his own totalising and self-consistent worldview/philosophy, and Planecrash makes it pop more than anything else he's written.</p></li></ol><h2><strong>The setup</strong></h2><p><a href="https://www.projectlawful.com/replies/1773731#reply-1773731">Dath ilan</a> is an alternative quasi-utopian Earth, based (it's at least strongly hinted) on the premise of: what if the average person was Eliezer Yudkowsky? Dath ilan has all the normal quasi-utopian things like world government and land-value taxes and the widespread use of Bayesian statistics in science. Dath ilan also has some less-normal things, like annual Oops It's Time To Overthrow the Government festivals, an order of super-rationalists, and extremely high financial rewards for designing educational curricula that bring down the age at which the average child learns the maths behind the game theory of cooperation.</p><p>Keltham is an above-average-selfishness, slightly-above-average-intelligence young man from dath ilan. He dies in the titular plane crash, and wakes up in Cheliax.</p><p>Cheliax is a country in a medieval fantasy world in another <em>plane</em> of existence to dath ilan's (get it?). (This fantasy world is copied from a role-playing game setting&#8212;a fact I discovered when Planecrash literally linked to a Wiki article to explain part of the in-universe setting.) Like every other country in this world, Cheliax is medieval and poor. Unlike the other countries, Cheliax has the additional problem of being ruled by the forces of Hell.</p><p>Keltham meets Carissa, a Chelish military wizard who alerts the Chelish government about Keltham. Keltham is kept unaware about the Hellish nature of Cheliax, so he's eager to use his knowledge to start the scientific and industrial revolutions in Cheliax to solve the medieval poverty thing&#8212;starting with delivering lectures on first-order logic (why, what else would you first do in a medieval fantasy world?). An elaborate game begins where Carissa and a select group of Chelish agents try to extract maximum science from an unwitting Keltham before he realises what Cheliax really is&#8212;and hope that by that time, they'll have tempted him to change his morals towards a darker, more Cheliax-compatible direction.</p><h2><strong>The characters</strong></h2><p>Keltham oscillates somewhere between annoying and endearing.</p><p>The annoyingness comes from his gift for interrupting any moment with polysyllabic word vomit. Thankfully, this is not random pretentious techno-babble but a coherent depiction of a verbose character who thinks in terms of a non-standard set of concepts. Keltham's thoughts often include an exclamation along the lines of "what, how is {'coordination failure' / 'probability distribution' / 'decision-theoretic-counterfactual-threat-scenario'} so many syllables in this language, how do these people ever talk?"&#8212;not an unreasonable question. However, the sheer volume of Keltham's verbosity is still something, especially when it gets in the way of everything else.</p><p>The endearingness comes from his manic rationalist problem-solver energy, which gets applied to everything from figuring out chemical processes for magic ingredients to estimating the odds that he's involved in a conspiracy to managing the complicated social scene Cheliax places him in. It's somewhat like <em>The Martian</em>, a novel (and movie) about an astronaut stranded on Mars solving a long series of <a href="https://xkcd.com/1536/">engineering challenges</a>, but the problem-solving is much more abstract and game-theoretic and interpersonal, than concrete and physical and man-versus-world.</p><p>By far the best and most interesting character in Planecrash is Carissa Sevar, one of the several characters whose point-of-view is written by lintamande rather than Yudkowsky. She's so driven that she accidentally becomes a cleric of the god of self-improvement. She grapples realistically with the large platter of problems she's handed, experiences triumph and failure, and keeps choosing pain over stasis. All this leads to perhaps the greatest arc of grit and unfolding ambition that I've read in fiction.</p><h2><strong>The competence</strong></h2><p>I have a memory of once reading some rationalist blogger describing the worldview of some politician as: there's no such thing as competence, only loyalty. If a problem doesn't get solved, it's definitely not because the problem was tricky and there was insufficient intelligence applied to it or a missing understanding of its nature or someone was genuinely incompetent. It's always because whoever was working on it wasn't loyal enough to you. (I thought this was Scott Alexander on Trump, but the closest from him seems to be <a href="https://slatestarcodex.com/2016/03/19/book-review-the-art-of-the-deal/">this</a>, which makes a very different point.)</p><p><em><strong>EDIT: </strong>it was Scott Alexander on Trump, but in a <a href="https://slatestarcodex.com/2016/09/28/ssc-endorses-clinton-johnson-or-stein/">different post</a>, as helpfully pointed out by <a href="https://nosetgauge.substack.com/p/review-planecrash/comment/97802265">a commenter</a>: &#8220;He lives in a world where there is no such thing as intelligence, only loyalty. If we haven&#8217;t solved all of our problems yet, it&#8217;s because the Department of Problem-Solving was insufficiently loyal, and didn&#8217;t try hard enough. His only promise is to fill that department with loyal people who really want the problem solved.&#8221;</em></p><p>Whether or not I hallucinated this, the worldview of Planecrash is the opposite.</p><p>Consider Queen Abrogail Thrune II, the despotic and unhinged ruler of Cheliax who has a flair for torture. You might imagine that her main struggles are paranoia over the loyalty of her minions, and finding time to take glee in ruling over her subjects. And there's some of those. But more than that, she spends a lot of time being annoyed by how incompetent everyone around her is.</p><p>Or consider Aspexia Rugatonn, Cheliax's religious leader and therefore in charge of making the country worship Hell. She's basically a kindly grandmother figure, except not. You might expect her thoughts to be filled with deep emotional conviction about Hell, or disappointment in the "moral" failures of those who don't share her values (i.e. every non-sociopath who isn't brainwashed hard enough). But instead, she spends a lot of her time annoyed that other people don't understand how to act most usefully within the bounds of the god of Hell's instructions. The one time she gets emotional is when a Chelish person finally manages to explain the concept of <a href="https://www.lesswrong.com/tag/corrigibility">corrigibility</a> to her as well as Aspexia herself could. (The gods and humans in the Planecrash universe are in a weird inverse version of the AI alignment problem. The gods are superintelligent, but have restricted communication bandwidth and clarity with humans. Therefore humans often have to decide how to interpret tiny snippets of god-orders through changing circumstances. So instead of having to steer the superintelligence given limited means, the core question is how to let yourself be steered by a superintelligence that has very limited communication bandwidth with you.)</p><p>Fiction is usually filled with characters who advance the plot in helpful ways with their emotional fumbles: consider the stereotypical horror movie protagonist getting mad and running into a dark forest alone, or a character whose pride is insulted doing a dumb thing on impulse. Planecrash has almost none of that. The characters are all good at their jobs. They are surrounded by other competent actors with different goals thinking hard about how to counter their moves, and they always think hard in response, and the smarter side tends to win. Sometimes you get the feeling you're just reading the meeting notes of a competent team struggling with a hard problem. Evil is not dumb or insane, but just "unaligned" by virtue of pursuing a different goal than you&#8212;and does so very competently. For example: the core values of the forces of Hell are literally tyranny, slavery, and pain. They have a strict hierarchy and take deliberate steps to encourage arbitrary despotism out of religious conviction. And yet: their hierarchy is still mostly an actual competence hierarchy, because the decision-makers are all very self-aware that they can only be despotic to the extent that it still promotes competence on net. Because they're competent.</p><p>Planecrash, at its heart, is <a href="https://tvtropes.org/pmwiki/pmwiki.php/Main/CompetencePorn">competence porn</a>. Keltham's home world of dath ilan is defined by its absence of coordination failures. Neither there nor in Cheliax's world are there really any lumbering bureaucracies that do insane things for inscrutable bureaucratic reasons; all the organisations depicted are all remarkably sane. Important positions are almost always filled by the smart, skilled, and hardworking. Decisions aren't made because of emotional outbursts. Instead, lots of agents go around optimising for their goals by thinking hard about them. For a certain type of person, this is a very relaxing world to read about, despite all the hellfire</p><h2><strong>The philosophy</strong></h2><p>"<a href="https://www.lesswrong.com/posts/4ARtkT3EYox3THYjF/rationality-is-systematized-winning">Rationality is systematized winning</a>", writes Yudkowsky in <a href="https://www.lesswrong.com/rationality">The Sequences</a>. All the rest is commentary.</p><p>The core move in Yudkowsky's philosophy is:</p><ul><li><p>We want to find the general solution to some problem.</p><ul><li><p>for example: fairness&#8212;how should we split gains from a project where many people participated</p></li></ul></li><li><p>Now here are some common-sense properties that this thing should follow</p><ul><li><p>for example:</p><ul><li><p>(1) no gains should be left undivided</p></li><li><p>(2) if two people both contribute identically to every circumstance (formalised as a set of participating people), they should receive an equal share of the gains</p></li><li><p>(3) the rule should give the same answer if you combine the division of gains from project A and then project B, as when you use it to calculate the division of gains from project A+B</p></li><li><p>(4) if one person doesn't add value in any circumstance, their share of the gains is zero</p></li></ul></li></ul></li><li><p>Here is The Solution. Note that it's mathematically provable that if you don't follow The Solution, there exists a situation where you will do something obviously dumb.</p><ul><li><p>For example: <a href="https://en.wikipedia.org/wiki/Shapley_value">Shapely value</a> is the unique solution that satisfies the axioms above. (The Planecrash walkthrough of Shapely value is roughly <a href="https://www.glowfic.com/replies/1730444#reply-1730444">here</a>; see also <a href="https://www.glowfic.com/replies/1729094#reply-1729094">here</a> for more Planecrash about trade and fairness.)</p></li></ul></li><li><p>Therefore, The Solution is uniquely spotlighted by the combination of common-sense goals and maths as the final solution to this problem, and if you disagree, please read this 10,000 word dialogue.</p></li></ul><p>The centrality of this move is something I did not get from The Sequences, but which is very apparent in Planecrash. A lot of the maths in Planecrash isn't new Yudkowsky material. But Planecrash is the only thing that has given me a <em>map</em> through the core objects of Yudkowsky's philosophy, and spelled out the high-level structure so clearly. It's also, as far as I know, the most detailed description of Yudkowsky's quasi-utopian world of dath ilan.</p><h3><strong>Validity, Probability, Utility</strong></h3><p>Keltham's lectures to the Chelish&#8212;yes, there are actually literal maths lectures within Planecrash&#8212;walk through three key examples, at a spotty level of completeness but at a high quality of whatever is covered:</p><ol><li><p>Validity, i.e. logic. In particular, Yudkowksy highlights what I think is some combination of <a href="https://en.wikipedia.org/wiki/Lindstr%C3%B6m%27s_theorem">Lindstrom's theorem</a> and <a href="https://en.wikipedia.org/wiki/G%C3%B6del%27s_completeness_theorem">Godel's completeness theorem</a>, that together imply first-order logic is the unique logic that is both complete (i.e. everything true within it can be proven) and has some other nice properties. However, first-order logic is also not strong enough to capture some things we care about (such as the natural numbers), so this is the least-strong example of the above pattern. Yudkowsky has written out his thoughts on logic in the <a href="https://www.lesswrong.com/s/SqFbMbtxGybdS2gRs">mathematics and logic section here</a>, if you want to read his takes in a non-fiction setting.</p></li><li><p>Probability. So-called <a href="https://en.wikipedia.org/wiki/Dutch_book_theorems">Dutch book theorems</a> show that if an agent does not update their beliefs in a Bayesian way, there exists a set of losing bets that they would take despite it leading to a guaranteed loss. So your credences in beliefs should be represented as probabilities, and you should update those probabilities with Bayes' theorem. (<a href="https://www.glowfic.com/replies/1750558#reply-1750558">Here</a> is a list of English statements that, dath ilani civilisation thinks, anyone competent in Probability should be able to translate into correct maths.)</p></li><li><p>Utility. The behaviour of any agent that is "rational" in a certain technical sense should be describable as it having a "utility function", i.e. every outcome can be assigned a number, such that the agent predictably chooses outcomes with higher numbers over those with lower ones. This is because if an agent violates this constraint, there must exist situations where it would do something obviously dumb. As a shocked Keltham puts it: "I, I mean, there's being chaotic, and then there's being so chaotic that it violates <em>coherence theorems</em>".</p></li></ol><p>In Yudkowsky's own words, not in Planecrash but in an <a href="https://www.lesswrong.com/posts/RQpNHSiWaXTvDxt6R/coherent-decisions-imply-consistent-utilities">essay he wrote</a> (with much valuable discussion in the comments):</p><blockquote><p><em>We have multiple spotlights all shining on the same core mathematical structure, saying dozens of different variants on, "If you aren't running around in circles or stepping on your own feet or wantonly giving up things you say you want, we can see your behavior as corresponding to this shape. Conversely, if we can't see your behavior as corresponding to this shape, you must be visibly shooting yourself in the foot." Expected utility is the only structure that has this great big family of discovered theorems all saying that. It has a scattering of academic competitors, because academia is academia, but the competitors don't have anything like that mass of spotlights all pointing in the same direction.</em></p><p><em>So if we need to pick an interim answer for "What kind of quantitative framework should I try to put around my own decision-making, when I'm trying to check if my thoughts make sense?" or "By default and barring special cases, what properties might a sufficiently advanced machine intelligence look to us like it possessed, at least approximately, if we couldn't see it visibly running around in circles?", then there's pretty much one obvious candidate: Probabilities, utility functions, and expected utility.</em></p></blockquote><h3><strong>Coordination</strong></h3><p>Next, coordination. There is no single theorem or total solution for the problem of coordination. But the Yudkowskian frame has near-infinite scorn for failures of coordination. Imagine not realising all possible gains just because you're stuck in some equilibrium of agents defecting against each other. Is that winning? No, it's not. Therefore, it must be out.</p><p>Dath ilan has a mantra that goes, roughly: if you do that, you will end up there, so if you want to end up somewhere that is not there, you will have to do Something Else Which Is Not That. And the basic premise of dath ilan is that society actually has the ability to collectively say "we are currently going there, and we don't want to, and while none of us can individually change the outcome, we will all coordinate to take the required collective action and not defect against each other in the process even if we'd gain from doing so". Keltham claims that in dath ilan, if there somehow developed an oppressive tyranny, everyone would wait for some <a href="https://en.wikipedia.org/wiki/Focal_point_(game_theory)#:~:text=In%20game%20theory%2C%20a%20focal,Strategy%20of%20Conflict%20(1960).">Schelling time</a> (like a solar eclipse or the end of the calendar year or whatever) and then simultaneously rise up in rebellion. It probably helps that dath ilan has annual "oops it's time to overthrow the government" exercises. It also helps that everyone in dath ilan knows that everyone knows that everyone knows that everyone knows (...) all the <a href="https://www.lesswrong.com/posts/9QxnfMYccz9QRgZ5z/the-costly-coordination-mechanism-of-common-knowledge">standard rationalist takes on coordination and common knowledge</a>.</p><p>Keltham summarises the universality of Validity, Probability, Utility, and Coordination (note the capitals):</p><blockquote><p><em>"I am a lot more confident that Validity, Probability, and Utility are still singled-out mathematical structures whose fragmented shards and overlapping shadows hold power in Golarion [=the world of Cheliax], than I am confident that I already know why snowflakes here have sixfold symmetry. And I wanted to make that clear before I said too much about the hidden orders of reality out of dath ilan - that even if the things I am saying are entirely wrong about Golarion, that kind of specific knowledge is not the most important knowledge I have to teach. I have gone into this little digression about Validity and timelessness and optimality, in order to give you some specific reason to think that [...] some of the knowledge he has to teach is sufficiently general that you have strong reason for strong hope that it will work [...] [...] "It is said also in dath ilan that there is a final great principle of Law, less beautiful in its mathematics than the first three, but also quite important in practice; it goes by the name Coordination, and deals with agents simultaneously acting in such fashion to all get more of what they wanted than if they acted separately."</em></p></blockquote><h3><strong>Decision theory</strong></h3><p>The final fundamental bit of Yudkowsky's philosophy is decision theories more complicated than causal decision theory.</p><p>A short primer / intuition pump: a decision theory specifies how you should choose between various options (it's not moral philosophy, because it assumes that we know already know <em>what</em> we value). The most straightforward decision theory is causal decision theory, which says: pick the option that causes the best outcome in expectation. Done, right? No; the devil is in the word "causes". Yudkowsky makes much of <a href="https://www.lesswrong.com/posts/6ddcsdA2c2XpNpE5x/newcomb-s-problem-and-regret-of-rationality">Newcomb's problem</a>, but I prefer another example: <a href="https://www.lesswrong.com/tag/parfits-hitchhiker">Parfit's hitchhiker</a>. Imagine you're a selfish person stuck in a desert without your wallet, and want to make it back to your hotel in the city. A car pulls up, with a driver who knows whether you're telling the truth. You ask to be taken back to your hotel. The driver asks if you'll pay $10 to them as a service. Dying in the desert is worse for you than paying $10, so you'd like to take this offer. However, you obey causal decision theory: if the driver takes you to your hotel, you would go to your hotel to get your wallet, but once inside you have the option between (a) take $10 back to the driver and therefore lose money, and (b) stay in your hotel and lose no money. Causal decision theory says to take option (b), because you're a selfish agent who doesn't care about the driver. And the driver knows you'd be lying if you said "yes", so you have to tell the driver "no". The driver drives off, and you die of thirst in the desert. If only you had spent more time arguing about non-causal decision theories on LessWrong.</p><p>Dying in a desert rather than spending $10 is not exactly systematised winning. So causal decision theory is out. (You could argue that another moral of Parfit's hitchhiker is that being a purely selfish agent is bad, and humans aren't purely selfish so it's not applicable to the real world anyway, but in Yudkowsky's philosophy&#8212;and decision theory academia&#8212;you want a <em>general</em> solution to the problem of rational choice where you can take <em>any</em> utility function and <em>win</em> by <em>its</em> lights regardless of which convoluted setup philosophers drop you into.) Yudkowsky's main academic / mathematical accomplishment is co-inventing (with Nate Soares) <a href="https://arxiv.org/abs/1710.05060">functional decision theory</a>, which says you should consider your decisions as the output of a fixed function, and then choose the function that leads to the best consequences for you. This solves Parfit's hitchhiker, as well as problems like <a href="https://www.lesswrong.com/tag/smoking-lesion">the smoking lesion problem</a> that <a href="https://en.wikipedia.org/wiki/Evidential_decision_theory">evidential decision theory</a>, the classic non-causal decision theory, succumbs to. As far as I can judge, functional decision theory is actually a good idea (if somewhat underspecified), but academic engagement (whether critiques and praises) with it has been limited so there's no broad consensus in its favor that I can point at. (If you want to read Yudkowsky's explanation for why he doesn't spend more effort on academia, it's <a href="https://www.glowfic.com/replies/1729119#reply-1729119">here</a>.)</p><p>(Now you know what a Planecrash tangent feels like, except you don't, because Planecrash tangents can be <em>much</em> longer.)</p><p>One big aspect of Yudkowskian decision theory is how to respond to threats. Following causal decision theory means you can neither make credible threats nor commit to deterrence to counter threats. Yudkowsky endorses not responding to threats to avoid incentivising them, while also having deterrence commitments to maintain good equilibria. He also implies this is a consequence of using a sensible functional decision theory. But there's a tension here: your deterrence commitment could be interpreted as a threat by someone else, or visa versa. When the Eisenhower administration's nuclear doctrine threatened massive nuclear retaliation in event of the Soviets taking West Berlin, what's the exact maths that would've let them argue to the Soviets "no no this isn't a threat, this is just a deterrence commitment", while allowing the Soviets keep to Yudkowsky's strict rule to ignore all threats?</p><p>My (uninformed) sense is that this maths hasn't been figured out. Planecrash never describes it (though <a href="https://www.glowfic.com/replies/1770000#reply-1770000">here</a> is some discussion of decision theory in Planecrash). Posts in the LessWrong decision theory canon like <a href="https://www.lesswrong.com/posts/wXbSAKu2AcohaK2Gt/udt-shows-that-decision-theory-is-more-puzzling-than-ever">this</a> or <a href="https://www.lesswrong.com/posts/brXr7PJ2W4Na2EW2q/the-commitment-races-problem">this</a> and <a href="https://www.lesswrong.com/posts/g8HHKaWENEbqh2mgK/updatelessness-doesn-t-solve-most-problems-1">this</a> seem to point to real issues around decision theories encouraging commitment races, and when Yudkowsky pipes up in the comments he's mostly falling back on the conviction that, surely, sufficiently-smart agents will find some way around mutual destruction in a commitment race (systematised winning, remember?). There are also various <a href="https://www.lesswrong.com/posts/ySLYSsNeFL5CoAQzN/a-critique-of-functional-decision-theory">critiques of functional decision theory</a> (see also <a href="https://www.lesswrong.com/posts/ySLYSsNeFL5CoAQzN/a-critique-of-functional-decision-theory?commentId=y8zRwcpNeu2ZhM3yE">Abram Demski's comment</a> on that post acknowledging that functional decision theory is underspecified). Perhaps it all makes sense if you've worked through Appendix B7 of Yudkowsky's big decision theory paper (which I haven't actually read, let alone taken time to digest), but (a) why doesn't he reference that appendix then, and (b) I'd complain about that being hard to find, but then again we are talking about the guy who leaves the clearest and most explicit description of his philosophy scattered across an R-rated role-playing-game fanfic posted in innumerable parts on an obscure internet forum, so I fear my complaint would be falling on deaf ears anyway.</p><h3><strong>The political philosophy of dath ilan</strong></h3><p>Yudkowsky has put a lot of thought into how the world of dath ilan functions. Overall it's very coherent.</p><p><a href="https://www.glowfic.com/replies/1743445#reply-1743445">Here's a part where Keltham explains dath ilan's central management principle</a>: everything, including every project, every rule within any company, and any legal regulation, needs to have one person responsible for it.</p><blockquote><p><em>Keltham is informed, though he doesn't think he's ever been tempted to make that mistake himself, that overthinky people setting up corporations sometimes ask themselves 'But wait, what if this person here can't be trusted to make decisions all by themselves, what if they make the wrong decision?' and then try to set up more complicated structures than that. This basically never works. If you don't trust a power, make that power legible, make it localizable to a single person, make sure every use of it gets logged and reviewed by somebody whose job it is to review it. If you make power complicated, it stops being legible and visible and recordable and accountable and then you actually are in trouble.</em></p></blockquote><p><a href="https://www.glowfic.com/replies/1772752#reply-1772752">Here's a part where Keltham talks about how dath ilan solves the problem of who watches the watchmen</a>:</p><blockquote><p><em>If you count the rehearsal festivals for it, Civilization spends more on making sure Civilization can collectively outfight the Hypothetical Corrupted Governance Military, than Civilization spends on its actual military.</em></p></blockquote><p><a href="https://www.glowfic.com/replies/1773754#reply-1773754">Here's a part where dath ilan's choice of political system is described</a>, which I will quote at length:</p><blockquote><p><em>Conceptually and to first-order, the ideal that Civilization is approximating is a giant macroagent composed of everybody in the world, taking coordinated macroactions to end up on the multi-agent-optimal frontier, at a point along that frontier reflecting a fair division of the gains from that coordinated macroaction -</em></p><p><em>Well, to be clear, the dath ilani would shut it all down if actual coordination levels started to get anywhere near that. Civilization has spoken - with nearly one voice, in fact - that it does not want to turn into a hivemind.</em></p><p><em>[...]</em></p><p><em>Conceptually and to second-order, then, Civilization thinks it should be divided into a Private Sphere and a Public Shell. Nearly all the decisions are made locally, but subject to a global structure that contains things like "children may not be threatened into unpaid labor"; or "everybody no matter who they are or what they have done retains the absolute right to cryosuspension upon their death"; [...]</em></p><p><em>[...]</em></p><p><em>Directdemocracy has been tried, from time to time, within some city of dath ilan: people making group decisions by all individually voting on them. It can work if you try it with fifty people, even in the most unstructured way. Get the number of direct voters up to ten thousand people, and no amount of helpfully-intended structure in the voting process can save you.</em></p><p><em>[...]</em></p><p><em>Republics have been tried, from time to time, within some city of dath ilan: people making group decisions by voting to elect leaders who make those decisions. It can work if you try it with fifty people, even in the most unstructured way. Get the number of voters up to ten thousand people, and no amount of helpfully-intended structure in the voting process Acan save you.</em></p><p><em>[...]</em></p><p><em>There are a hundred more clever proposals for how to run Civilization's elections. If the current system starts to break, one of those will perhaps be adopted. Until that day comes, though, the structure of Governance is the simplest departure from directdemocracy that has been found to work at all.</em></p><p><em>Every voter of Civilization, everybody at least thirteen years old or who has passed some competence tests before then, primarily exerts their influence through delegating their vote to a Delegate.</em></p><p><em>A Delegate must have at least fifty votes to participate in the next higher layer at all; and can retain no more than two hundred votes before the marginal added influence from each additional vote starts to diminish and grow sublinearly. Most Delegates are not full-time, unless they are representing pretty rich people, but they're expected to be people interested in politics [...]. Your Delegate might be somebody you know personally and trust, if you're the sort to know so many people personally that you know one Delegate. [...]</em></p><p><em>If you think you've got a problem with the way Civilization is heading, you can talk to your Delegate about that, and your Delegate has time to talk back to you.</em></p><p><em>That feature has been found to not actually be dispensable in practice. It needs to be the case that, when you delegate your vote, you know who has your vote, and you can talk to that person, and they can talk back. Otherwise people feel like they have no lever at all to pull on the vast structure that is Governance, that there is nothing visible that changes when a voter casts their one vote. Sure, in principle, there's a decision-cohort whose votes move in logical synchrony with yours, and your cohort is probably quite large unless you're a weird person. But some part of you more basic than that will feel like you're not in control, if the only lever you have is an election that almost never comes down to the votes of yourself and your friends.</em></p><p><em>The rest of the electoral structure follows almost automatically, once you decide that this property has to be preserved at each layer.</em></p><p><em>The next step up from Delegates are Electors, full-time well-paid professionals who each aggregate 4,000 to 25,000 underlying voters from 50 to 200 Delegates. Few voters can talk to their Electors [...] but your Delegate can have some long conversations with them. [...]</em></p><p><em>Representatives aggregate Electors, ultimately 300,000 to 3,000,000 underlying votes apiece. There are roughly a thousand of those in all Civilization, at any given time, with social status equivalent to an excellent CEO of a large company or a scientist who made an outstanding discovery [...]</em></p><p><em>And above all this, the Nine Legislators of Civilization are those nine candidates who receive the most aggregate underlying votes from Representatives. They vote with power proportional to their underlying votes; but when a Legislator starts to have voting power exceeding twice that of the median Legislator, their power begins to grow sublinearly. By this means is too much power prevented from concentrating into a single politician's hands.</em></p><p><em>Surrounding all this of course are numerous features that any political-design specialist of Civilization would consider obvious:</em></p><p><em>Any voter (or Delegate or Elector or Representative) votes for a list of three possible delegees of the next layer up; if your first choice doesn't have enough votes yet to be a valid representor, your vote cascades down to the next person on your list, but remains active and ready to switch up if needed. This lets you vote for new delegees entering the system, without that wasting your vote while there aren't enough votes yet.</em></p><p><em>Anyone can at any time immediately eliminate a person from their 3-list, but it takes a 60-day cooldown to add a new person or reorder the list. The government design isn't meant to make it cheap or common to threaten your delegee with a temporary vote-switch if they don't vote your way on that particular day. The government design isn't meant to make it possible for a new brilliant charismatic leader to take over the entire government the next day with no cooldowns. It is meant to let you rapidly remove your vote from a delegee that has sufficiently ticked you off.</em></p><p><em>Once you have served as a Delegate, or delegee of any other level, you can't afterwards serve in any other branches of Governance. [...]</em></p><p><em>This is meant to prevent a political structure whose upper ranks offer promotion as a reward to the most compliant members of the ranks below, for by this dark-conspiratorial method the delegees could become aligned to the structure above rather than their delegators below.</em></p><p><em>(Most dath ilani would be suspicious of a scheme that tried to promote Electors from Delegates in any case; they wouldn't think there should be a political career ladder [...] Dath ilani are instinctively suspicious of all things meta, and much more suspicious of anything purely meta; they want heavy doses of object-level mixed in. To become an Elector you do something impressive enough, preferably something entirely outside of Governance, that Delegates will be impressed by you. You definitely don't become an Elector by being among the most ambitious and power-seeking people who wanted to climb high and knew they had to start out a lowly Delegate, who then won a competition to serve the system above them diligently enough to be selected for a list of Electors fed to a political party's captive Delegates. If a dath ilani saw a system like this, that was supposedly a democracy set in place by the will of its people, they would ask what the captive 'voters' even thought they were supposedly trying to do under the official story.)</em></p></blockquote><p>Dath ilani Legislators have a programmer's or engineer's appreciation for simplicity:</p><blockquote><p><em>[...] each [regulation] must be read aloud by a Legislator who thereby accepts responsibility for that regulation; and when that Legislator retires a new Legislator must be found to read aloud and accept responsibility for that regulation, or it will be stricken from the books. Every regulation in Civilization, if something goes wrong with it, is the fault of one particular Legislator who accepted responsibility for it. To speak it aloud, it is nowadays thought, symbolizes the acceptance of this responsibility.</em></p><p><em>Modern dath ilani aren't really the types in the first place to produce literally-unspeakable enormous volumes of legislation that no hapless citizen or professional politician could ever read within their one lifetime let alone understand. Even dath ilani who aren't professional programmers have written enough code to know that each line of code to maintain is an ongoing cost. Even dath ilani who aren't professional economists know that regulatory burdens on economies increase quadratically in the cost imposed on each transaction. They would regard it as contrary to the notion of a lawful polity with law-abiding citizens that the citizens cannot possibly know what all the laws are, let alone obey them. Dath ilani don't go in for fake laws in the same way as Golarion polities with lots of them; they take laws much too seriously to put laws on the books just for show.</em></p></blockquote><p>Finally, the Keepers are an order of people trained in all the most hardcore arts of rationality, and who thus end up with inhuman integrity and even-handedness of judgement. They are used in many ways, for example:</p><blockquote><p><em>There are also Keeper cutouts at key points along the whole structure of Governance - the Executive of the Military reports not only to the Chief Executive but also to an oathsworn Keeper who can prevent the Executive of the Military from being fired, demoted, or reduced in salary, just because the Chief Executive or even the Legislature says so. It would be a big deal, obviously, for a Keeper to fire this override; but among the things you buy when you hire a Keeper is that the Keeper will do what they said they'd do and not give five flying fucks about what sort of 'big deal' results. If the Legislators and the Chief Executive get together and decide to order the Military to crush all resistance, the Keeper cutout is there to ensure that the Executive of the Military doesn't get a pay cut immediately after they tell the Legislature and Chief Executive to screw off.</em></p></blockquote><p>Also, to be clear, absolutely none of this is plot-relevant.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!CKR2!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcaa2b347-2669-4294-b3c4-d4ed7ca67c9b_400x400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!CKR2!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcaa2b347-2669-4294-b3c4-d4ed7ca67c9b_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!CKR2!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcaa2b347-2669-4294-b3c4-d4ed7ca67c9b_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!CKR2!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcaa2b347-2669-4294-b3c4-d4ed7ca67c9b_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!CKR2!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcaa2b347-2669-4294-b3c4-d4ed7ca67c9b_400x400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!CKR2!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcaa2b347-2669-4294-b3c4-d4ed7ca67c9b_400x400.png" width="400" height="400" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/caa2b347-2669-4294-b3c4-d4ed7ca67c9b_400x400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:400,&quot;width&quot;:400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:186498,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!CKR2!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcaa2b347-2669-4294-b3c4-d4ed7ca67c9b_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!CKR2!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcaa2b347-2669-4294-b3c4-d4ed7ca67c9b_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!CKR2!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcaa2b347-2669-4294-b3c4-d4ed7ca67c9b_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!CKR2!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcaa2b347-2669-4294-b3c4-d4ed7ca67c9b_400x400.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Above: the icon of dath ilan in Planecrash. When Yudkowsky really wants to monologue, he stops even pretending to do it through a character, and instead we get this talking globe. Hello, globe. Nice political philosophy you got there.</em></figcaption></figure></div><h2><strong>A system of the world</strong></h2><p>Yudkowsky proves that ideas matter: if you have ideas that form a powerful and coherent novel worldview, it doesn't matter if your main method for publicising them is ridiculously-long fanfiction, or if you dropped out of high school, or if you wear fedoras. People will still listen, and you might become (so far) the 21st century's most important philosopher.</p><p>Why is Yudkowsky so compelling? There are intellectuals like <a href="https://slatestarcodex.com/about/">Scott Alexander</a> who are most-strongly identified by a particular <em>method</em> (an even-handed, epistemically-rigorous, steelmaning-focused treatment of a topic), or intellectuals like <a href="https://overcoming-bias-anthology.com/">Robin Hanson</a> who are most-strongly identified by a particular <em>style</em> (eclectic irreverence about incentive mechanisms). But Yudkowsky's hallmark is delivering an entire system of the world that covers everything from logic to what correct epistemology looks like to the maths behind rational decision-making and coordination, and comes complete with identifying the biggest threat (misaligned AI) and the structure of utopia (dath ilan). None of the major technical inventions (except some in decision theory) are original to Yudkowsky. But he's picked up the pieces, slotted them into a big coherent structure, and presented it in great depth. And Yudkowsky's system claims to come with <em>proofs</em> for many key bits, in the literal mathematical sense. No, you can't crack open a textbook and see everything laid out, step-by-step. But the implicit claim is: read this long essay on coherence theorems, these papers on decision theory, this 20,000-word dialogue, these sequences on LessWrong, and ideally a few fanfics too, and then you'll get it.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!JCD9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb60a4faf-39e0-4dd6-9e5e-9009c9befadb_1212x330.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!JCD9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb60a4faf-39e0-4dd6-9e5e-9009c9befadb_1212x330.png 424w, https://substackcdn.com/image/fetch/$s_!JCD9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb60a4faf-39e0-4dd6-9e5e-9009c9befadb_1212x330.png 848w, https://substackcdn.com/image/fetch/$s_!JCD9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb60a4faf-39e0-4dd6-9e5e-9009c9befadb_1212x330.png 1272w, https://substackcdn.com/image/fetch/$s_!JCD9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb60a4faf-39e0-4dd6-9e5e-9009c9befadb_1212x330.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!JCD9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb60a4faf-39e0-4dd6-9e5e-9009c9befadb_1212x330.png" width="1212" height="330" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b60a4faf-39e0-4dd6-9e5e-9009c9befadb_1212x330.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:330,&quot;width&quot;:1212,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:79075,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!JCD9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb60a4faf-39e0-4dd6-9e5e-9009c9befadb_1212x330.png 424w, https://substackcdn.com/image/fetch/$s_!JCD9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb60a4faf-39e0-4dd6-9e5e-9009c9befadb_1212x330.png 848w, https://substackcdn.com/image/fetch/$s_!JCD9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb60a4faf-39e0-4dd6-9e5e-9009c9befadb_1212x330.png 1272w, https://substackcdn.com/image/fetch/$s_!JCD9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb60a4faf-39e0-4dd6-9e5e-9009c9befadb_1212x330.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>After reading Yudkowsky, you're perfectly inoculated against any philosophy so lazy that it doesn't even come with mathematical proofs. (<a href="https://x.com/robinhanson/status/810281385852276736">source</a>)</em></figcaption></figure></div><p>Does he deliver? To an impressive extent, yes. There's a lot of maths that <em>is</em> laid out step-by-step and <em>does</em> check out. There are many takes that are correct, and big structures that point in the right direction, and what seems wrong at least has depth and is usefully provocative. But dig deep enough, and there are cracks: <a href="https://www.lesswrong.com/posts/yCuzmCsE86BTu9PfA/there-are-no-coherence-theorems">arguments</a> about how much <a href="https://www.lesswrong.com/posts/NxF5G6CJiof6cemTw/coherence-arguments-do-not-entail-goal-directed-behavior">coherence theorems</a> really imply, <a href="https://www.lesswrong.com/posts/ySLYSsNeFL5CoAQzN/a-critique-of-functional-decision-theory">critiques</a> of the decision theory, and <a href="https://www.lesswrong.com/posts/CoZhXrhpQxpy9xw9y/where-i-agree-and-disagree-with-eliezer">good counterarguments</a> to the most extreme versions of Yudkowsky's AI risk thesis. You can chase any of these cracks up towers of LessWrong posts, or debate them endlessly at those parties where people stand in neat circles and exchange thought experiments about <a href="https://www.lesswrong.com/tag/acausal-trade">acausal trade</a>. If you have no interaction with rationalist/LessWrong circles, I think you'd be surprised at the fraction of our generation's top mathematical-systematising brainpower that is spent on this&#8212;or that is bobbing in the waves left behind, sometimes unknowingly.</p><p>As for myself: Yudkowsky's philosophy is one of the most impressive intellectual edifices I've seen. Big chunks of it&#8212;in particular the stuff about empiricism, naturalism, and the art of genuinely trying to figure out what's true that <a href="https://www.lesswrong.com/rationality">The Sequences</a> especially focus on&#8212;were very formative in my own thinking. I think it's often proven itself directionally correct. But Yudkowsky's philosophy makes a claim for near-mathematical correctness, and I think there's a bit of trouble there. While it has impressive mathematical depth and gets many things importantly right (e.g. Bayesianism), despite much effort spent digesting it, I don't see it meeting the rigour bar it would need for its predictions (for example about AI risk) to be more like those of a tested scientific theory than those of a framing, worldview, or philosophy. However, I'm also very unsympathetic to a certain straitlaced science-cargo-culting attitude that recoils from Yudkowsky's uncouthness and is uninterested in speculation or theory&#8212;they would do well to study the <a href="https://www.lesswrong.com/posts/JAAHjm4iZ2j5Exfo2/the-copernican-revolution-from-the-inside">actual history</a> of <a href="https://www.benlandautaylor.com/p/looking-beyond-the-veil">science</a>. I also see in Yudkowsky's philosophy choices of framing and focus that seem neither forced by reason nor entirely natural in my own worldview. I expect that lots more great work will come out within the Yudkowskian frame, whether critiques or patches, and this work could show it to be anywhere from impressive but massively misguided to almost prophetically prescient. However, I expect even greater things if someone figures out a new, even grander and more applicable system of the world. Perhaps that person can then describe it in a weird fanfic.</p><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.nosetgauge.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe for free to receive new posts, whether they include a new system of the world, or are just reviews of more weird books.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Review: Good Strategy, Bad Strategy]]></title><description><![CDATA[Notes from a rare book on strategy that isn't pointless.]]></description><link>https://www.nosetgauge.com/p/review-good-strategy-bad-strategy</link><guid isPermaLink="false">https://www.nosetgauge.com/p/review-good-strategy-bad-strategy</guid><dc:creator><![CDATA[Rudolf Laine]]></dc:creator><pubDate>Sat, 21 Dec 2024 17:09:35 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/32aff632-bacf-4261-957d-f9ecfb816135_1200x630.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I used to think that all generic strategy advice was pointless. After all, the point of a strategy is to achieve a thing, and to achieve a thing you just think hard about how to best do it and then work hard to do it. I said this to my friend Dewi, who said that this is mostly true, but there is an exception: <em>Good Strategy, Bad Strategy</em> by Richard Rumelt. Dewi was right.</p><p>The book has some principles. In particular: a good strategy should include a diagnosis of the problem, an overall guiding policy, and a set of coherent actions. A laundry list of actions, a goal, or a vague idea of which direction to move in are not strategies.</p><p>But most of the book's value is reading a bunch of examples and soaking up the thinking style embedded in them. Therefore, this review is mostly a series of vignettes taken directly from Rumelt's text, that hopefully 80/20 the value of the book (or is a helpful reminder if you've read it). Rumelt's vignettes are great.</p><p>I don't include anything from Rumelt's lengthy attack on most of what passes for "strategy consulting". I also won't mention the part where he spends most of a chapter giving a pop-sci history of physics from Galileo to dark matter in order to segue into a point about Starbucks' business strategy&#8212;he's an emeritus professor, so presumably he can just do this.</p><h1><strong>Example</strong></h1><blockquote><p><em>&#8220;In 1805, England had a problem. Napol&#233;on had conquered big chunks of Europe and planned the invasion of England. But to cross the Channel, he needed to wrest control of the sea away from the English. Off the southwest coast of Spain, the French and Spanish combined fleet of thirty-three ships met the smaller British fleet of twenty-seven ships. The well-developed tactics of the day were for the two opposing fleets to each stay in line, firing broadsides at each other. But British admiral Lord Nelson had a strategic insight. He broke the British fleet into two columns and drove them at the Franco-Spanish fleet, hitting their line perpendicularly. The lead British ships took a great risk, but Nelson judged that the less-trained Franco-Spanish gunners would not be able to compensate for the heavy swell that day. At the end of the Battle of Trafalgar, the French and Spanish lost twenty-two ships, two-thirds of their fleet. The British lost none. Nelson was mortally wounded, becoming, in death, Britain&#8217;s greatest naval hero. Britain&#8217;s naval dominance was ensured and remained unsurpassed for a century and a half.</em></p><p><em>Nelson&#8217;s challenge was that he was outnumbered. His strategy was to risk his lead ships in order to break the coherence of his enemy&#8217;s fleet. With coherence lost, he judged, the more experienced English captains would come out on top in the ensuing melee. Good strategy almost always looks this simple and obvious and does not take a thick deck of PowerPoint slides to explain. It does not pop out of some &#8220;strategic management&#8221; tool, matrix, chart, triangle, or fill-in-the-blanks scheme. Instead, a talented leader identifies the one or two critical issues in the situation&#8212;the pivot points that can multiply the effectiveness of effort&#8212;and then focuses and concentrates action and resources on them.&#8221;</em></p></blockquote><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.nosetgauge.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Want to receive new posts by email?</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h1><strong>The basics</strong></h1><h2><strong>The two fundamental things</strong></h2><ol><li><p>There's great advantage in having any coherent strategy in the first place at all; most orgs don't have one; people will be <em>actively surprised</em> if you have any coherent strategy at all</p></li><li><p>Subtle shifts of viewpoint can lead to realising new strengths; reframe the situation so you see the leverage point</p></li></ol><h2><strong>Coherence</strong></h2><h3><strong>Example: the obvious</strong></h3><blockquote><p><em>[...] I carried out interviews with twenty-six executives, all division managers or CEOs in the electronics and telecommunications sector. My interview plan was simple: I asked each executive to identify the leading competitor in their business. I asked how that company had become the leader&#8212;evoking their private theories about what works. And then I asked them what their own company&#8217;s current strategy was. These executives, by and large, had no trouble describing the strategy of the leader in their sectors. The standard story was that some change in demand or technology had appeared&#8212;a &#8220;window of opportunity&#8221; had opened&#8212;and the current leader had been the first one to leap through that window and take advantage of it. Not necessarily the first mover, but the first to get it right.</em></p><p><em>But when I asked about their own companies&#8217; strategies, there was a very different kind of response. Instead of pointing to the next window of opportunity, or even mentioning its possibility, I heard a lot of look-busy doorknob polishing. They were making alliances, they were doing 360-degree feedback, they were looking for foreign markets, they were setting challenging strategic goals, they were moving software into firmware, they were enabling Internet updates of firmware, and so on.&#8221;</em></p></blockquote><h3><strong>Example: Jobs waits</strong></h3><p>In the late 1990s:</p><blockquote><p><em>[Rumelt said:] &#8220;Steve, this turnaround at Apple has been impressive. But everything we know about the PC business says that Apple cannot really push beyond a small niche position. The network effects are just too strong to upset the Wintel standard. So what are you trying to do in the longer term? What is the strategy?&#8221;</em></p><p><em>He did not attack my argument. He didn&#8217;t agree with it, either. He just smiled and said, &#8220;I am going to wait for the next big thing."</em></p></blockquote><h3><strong>Strategy is Unexpected</strong></h3><blockquote><p><em>The Department of the Army publishes field manuals fully describing its basic doctrines and methods. FM 100-5, published in 1986, was titled Operations and was described as &#8220;the Army&#8217;s keystone warfighting manual.&#8221; Part 2 of FM 100-5 was dedicated to &#8220;Offensive Operations,&#8221; and on page 101 it described &#8220;envelopment&#8221; as the most important form of offensive maneuver&#8212;the U.S. Army&#8217;s &#8220;Plan A.&#8221; The manual said:</em></p><blockquote><p><em>Envelopment avoids the enemy&#8217;s front, where its forces are most protected and his fires most easily concentrated. Instead, while fixing the defender&#8217;s attention forward by supporting or diversionary attacks, the attacker maneuvers his main effort around or over the enemy&#8217;s defenses to strike at his flanks and rear.&#8221;</em></p></blockquote></blockquote><p>But when the US army in the First Iraq War followed its own advice, and avoided a frontal attack on the Iraqi forced in Kuwait, everyone was shocked and called it brilliant strategy.</p><p>A lot of this was just preventing everyone else from pursuing their own special agenda, in order to get them to act coherently towards a shared goal:</p><blockquote><p><em>In the case of Desert Storm, the focus was much more than an intellectual step. Schwarzkopf had to suppress the ambitions and desires of the air force, marines, various army units, each coalition partner, and the political leadership in Washington. For example, the U.S. Army&#8217;s best light infantry&#8212;the Eighty-Second Airborne&#8212;was tasked with providing support to French armor and infantry, an assignment its leadership protested. Eight thousand U.S. Marines waited on ships to land on the beaches of Kuwait City, but did not. It was a diversion. Air force commanders wanted to demonstrate the value of strategic bombing&#8212;they believed that the war could be won by air attacks on Baghdad&#8212;and had to be forced against strenuous protest to divert their resources to fully support the land offensive. Secretary of Defense Dick Cheney wanted the mission accomplished with a smaller force and detailed an alternative plan of attack. Prince Khalid, commanding the Saudi forces in the coalition, insisted that King Fahd be involved in the planning, but Schwarzkopf convinced President Bush to ensure that U.S. Central Command retained control over strategy and planning.</em></p></blockquote><h2><strong>Advantage</strong></h2><p>Rumelt quotes Andy Marshall, director of the Office of Net Assessment (which thinks about the US defence situation), complaining about the annual budgeting process of the defence department:</p><blockquote><p><em>"This process of justifying expenditures as counters to Soviet expenditures conditioned U.S. actions on Soviet strengths, expressed as threats, not on Soviet weaknesses and constraints. We had a war strategy&#8212;a catastrophic spasm&#8212;but no plan about how to compete with the Soviet Union over the long term."</em></p></blockquote><p>Marshall and some others had written "Strategy for Competing with the Soviets in the Military Sector of the Continuing Political-Military Competition" in 1976:</p><blockquote><p><em>This fascinating analysis of the situation worked to redefine "defense" in new terms&#8212;a subtle shift in point of view. It argued that &#8220;in dealing effectively with the other side, a nation seeks opportunities to use one or more distinctive competences in such a way as to develop competitive advantage&#8212;both in specific areas and overall.&#8221; It then went on to explain that the crucial area of competition was technology because the United States had more resources and better capabilities in that area. And, most important, it argued that having a true competitive strategy meant engaging in actions that imposed exorbitant costs on the other side. In particular, it recommended investing in technologies that were expensive to counter and where the counters did not add to Soviet offensive capabilities. For instance, increasing the accuracy of missiles or the quietness of submarines forced the Soviet Union to spend scarce resources on counters without increasing the threat to the United States. Investments in systems that made Soviet systems obsolete would also force them to spend, as would selectively advertising dramatic new technologies.</em></p></blockquote><p>It seems this actually happened&#8212;c.f. Ronald Reagan's Star Wars program.</p><p>Rumelt comments:</p><blockquote><p><em>There were no complex charts or graphs, no abstruse formulas, no acronym-jammed buzz speak: just an idea and some pointers to how it might be used&#8212;the terrible simplicity of the discovery of hidden power in a situation.</em></p><p><em>Marshall and Roche&#8217;s analysis included a list of U.S. and Soviet strengths and weaknesses. Such lists were not new, and the traditional response to them would have been to invest more to tip the &#8220;balance&#8221; in one&#8217;s favor. But Marshall and Roche, like Sam Walton, had an insight that, when acted upon, provided a much more effective way to compete&#8212;the discovery of hidden power in the situation.</em></p></blockquote><h1><strong>The kernel of good strategy</strong></h1><blockquote><p><em>The kernel of a strategy contains three elements:</em></p><ol><li><p><em>A diagnosis that defines or explains the nature of the challenge. A good diagnosis simplifies the often overwhelming complexity of reality by identifying certain aspects of the situation as critical.</em></p></li><li><p><em>A guiding policy for dealing with the challenge. This is an overall approach chosen to cope with or overcome the obstacles identified in the diagnosis.</em></p></li><li><p><em>A set of coherent actions that are designed to carry out the guiding policy. These are steps that are coordinated with one another to work together in accomplishing the guiding policy.&#8221;</em></p></li></ol></blockquote><p>If there is one specific and explicit piece of knowledge to take from this book, it's this.</p><h2><strong>Diagnosis</strong></h2><h3><strong>Example: US Cold War policy</strong></h3><blockquote><p><em>A diagnosis is generally denoted by metaphor, analogy, or reference to a diagnosis or framework that has already gained acceptance. For example, every student of U.S. national strategy knows about the diagnosis associated with the Cold War guiding policy of containment. This concept originated with George Kennan&#8217;s famous &#8220;long telegram&#8221; of 1946. Having served as an American diplomat in the USSR for more than a decade, and having seen Soviet terror and politics at close hand, he carefully analyzed the nature of Soviet ideology and power. Kennan started with the observation that the Soviet Union was not an ordinary nation-state. Its leaders defined their mission as opposition to capitalism and as spreading the gospel of revolutionary communism by whatever means necessary. He stressed that antagonism between communism and capitalist societies was a central foundation of Stalin&#8217;s political regime, preventing any sincere accommodation or honest international agreements. However, he also pointed out that the Soviet leaders were realists about power. Therefore, he recommended a guiding policy of vigilant counterforce:</em></p><blockquote><p><em>&#8220;In the light of the above, it will be clearly seen that the Soviet pressure against the free institutions of the western world is something that can be contained by the adroit and vigilant application of counter-force at a series of constantly shifting geographical and political points, corresponding to the shifts and maneuvers of Soviet policy, but which cannot be charmed or talked out of existence. The Russians look forward to a duel of infinite duration, and they see that already they have scored great successes.&#8221;</em></p></blockquote><p><em>Kennan&#8217;s diagnosis for the situation&#8212;a long-term struggle without the possibility of a negotiated settlement&#8212;was widely adopted within policy-making circles in the United States. His guiding policy of containment was especially attractive as it specified a broad domain of action&#8212;the USSR was, metaphorically speaking, infected by a virus. The United States would have to keep the virus from spreading until it finally died out. Kennan&#8217;s policy is sometimes called a strategy, but it lacked the element of action. All presidents from Truman through George H. W. Bush struggled with the problem of turning this guiding policy into actionable objectives. Over time, the guiding policy of containment led to NATO and SEATO, the Berlin Airlift, the Korean War, placing missiles in Europe, the Vietnam War, and other Cold War actions.</em></p><p><em>The power of Kennan&#8217;s diagnosis can be seen by considering how history might have been different if the situation had been framed another way in 1947. Perhaps the Soviet Union could have been enticed into the world community through a policy of engagement by including it in the Marshall Plan. Or perhaps it wasn&#8217;t an American problem at all, but an issue for the United Nations. Or perhaps the Soviet Union was a tyranny rivaling Nazi Germany, and the United States should have sought to actively oppose it, undermine it, and liberate its population.</em></p></blockquote><h3><strong>Example: IBM</strong></h3><blockquote><p><em>[W]hen Lou Gerstner took over the helm at IBM in 1993, the company was in serious decline. Its historically successful strategy had been organized around offering complete, integrated, turnkey end-to-end computing solutions to corporations and government agencies. However, the advent of the microprocessor changed all that. The computer industry began to fragment, with separate firms offering chips, memory, hard disks, keyboards, software, monitors, operating systems, and so on. [...] As computing moved to the desktop, and as IBM&#8217;s desktop offering became commoditized by clone competitors and the Windows-Intel standard, what should the company do? The dominant view at the company and among Wall Street analysts was that IBM was too integrated. The new industry structure was fragmented and, it was argued, IBM should be broken up and fragmented to match. As Gerstner arrived, preparations were under way for separate stock offerings for various pieces of IBM.</em></p><p><em>After studying the situation, Gerstner changed the diagnosis. He believed that in an increasingly fragmented industry, IBM was the one company that had expertise in all areas. Its problem was not that it was integrated but that it was failing to use the integrated skills it possessed. IBM, he declared, needed to become more integrated&#8212;but this time around customer solutions rather than hardware platforms. The primary obstacle was the lack of internal coordination and agility. Given this new diagnosis, the guiding policy became to exploit the fact that IBM was different, in fact, unique. IBM would offer customers tailored solutions to their information-processing problems, leveraging its brand name and broad expertise, but willing to use outside hardware and software as required. Put simply, its primary value-added activity would shift from systems engineering to IT consulting, from hardware to software. Neither the &#8220;integration is obsolete&#8221; nor the &#8220;knowing all aspects of IT is our unique ability&#8221; viewpoints are, by themselves, strategies. But these diagnoses take the leader, and all who follow, in very different directions.</em></p></blockquote><h2><strong>Guiding policy</strong></h2><blockquote><p><em>The guiding policy outlines an overall approach for overcoming the obstacles highlighted by the diagnosis. It is &#8220;guiding&#8221; because it channels action in certain directions without defining exactly what shall be done. Kennan&#8217;s containment and Gerstner&#8217;s drawing on all of IBM&#8217;s resources to solve customers&#8217; problems are examples of guiding policies. Like the guardrails on a highway, the guiding policy directs and constrains action without fully defining its content.</em></p><p><em>[...]</em></p><p><em>A guiding policy creates advantage by anticipating the actions and reactions of others, by reducing the complexity and ambiguity in the situation, by exploiting the leverage inherent in concentrating effort on a pivotal or decisive aspect of the situation, and by creating policies and actions that are coherent, each building on the other rather than canceling one another out.</em></p></blockquote><h3><strong>Example: corner grocery store</strong></h3><blockquote><p><em>To look more closely at how a guiding policy works, follow the thinking of Stephanie, a friend who owns a corner grocery store. She does the accounts, manages personnel, sometimes runs the cash register, and makes all the decisions. Several years ago, Stephanie told me about some of the issues she was facing. She was considering whether she should keep prices down or offer more expensive, fresh organic produce. Should she begin to stock more Asian staples for the many Asian students who lived in the area? Should the store be open longer hours? How important was it to have a helpful, friendly staff that gets to know the regulars? Would adding a second checkout stand pay off? What about parking in the alley? Should she advertise in the local college newspaper? Should she paint the ceiling green or white? Should she put some items on sale each week? Which ones?</em></p><p><em>[digs about how economists aren't helpful for this problem]</em></p><p><em>Thinking about her store, Stephanie diagnosed her challenge to be competition with the local supermarket. She needed to draw customers away from a store that was open 24/7 and had lower prices. Seeking a way forward, she believed that most of her customers were people who walked by the store almost every day. They worked or lived nearby. Scanning her list of questions and alternatives, she determined that there was a choice between serving the more price-conscious students or the more time-sensitive professionals. Transcending thousands of individual choices and instead framing the problem in terms of choosing among a few customer groups provided a dramatic reduction in complexity.</em></p><p><em>Of course, if both of these customer segments could be served with the same policies and actions, then the dichotomy would have been useless and should be cast aside. In Stephanie&#8217;s case, the difference seemed significant. More of her customers were students, but the professionals who stopped in made much larger purchases. Pushing further along, Stephanie began to explore the guiding policy of &#8220;serve the busy professional.&#8221; After some more tinkering, Stephanie sharpened the guiding policy a bit more, deciding to target &#8220;the busy professional who has little time to cook.</em></p><p><em>There was no way to establish that this particular guiding policy was the only good one, or the best one. But, absent a good guiding policy, there is no principle of action to follow. Without a guiding policy, Stephanie&#8217;s actions and resource allocations would probably be inconsistent and incoherent, fighting with one another and canceling one another out. Importantly, adopting this guiding policy helped reveal and organize the interactions among the many possible actions. Considering the needs of the busy professional with little time to cook, she could see that the second checkout stand would help handle the burst of traffic at 5 p.m. So would more parking in the alley. In addition, she felt she could take space currently used for selling munchies to students and offer prepared high-quality take-home foods instead. Professionals, unlike students, would not come shopping at midnight, so there was no need for very late hours. The busy professionals would appreciate adequate staffing after work and, perhaps, at lunchtime. Having a guiding policy helped create actions that were coordinated and concentrated, focusing her efforts.</em></p></blockquote><h2><strong>Coherent action</strong></h2><h3><strong>Action requires doing something</strong></h3><blockquote><p><em>Strategy is about action, about doing something. The kernel of a strategy must contain action.</em></p></blockquote><p>Rumelt gives an example of a consumer goods producer that was running a "Pan-European" initiative, to try to achieve economies of scale in both production and marketing.</p><blockquote><p><em>The heads of the country-based organizations were placed on a Pan-Europe Executive Committee, which met once a quarter. Developers from Germany and the United Kingdom were rotated between the two locations. A New Products group had been created to consult with all departments on opportunities for Pan-European concepts and brands. Part of each executive&#8217;s evaluation for promotion was based on his or her contribution to the Pan-European initiative. Despite these measures, nothing much had happened. The German and British developers each claimed that their initiatives were unsupported by the other. The one British-German joint initiative had not been picked up by the rest of the organization.</em></p><p><em>[...]</em></p><p><em>"Suppose," I said, "that this was really important, really top-priority critical. Suppose you absolutely had to get some Pan-European products developed and marketed in the next eighteen months or everything would collapse. What would you do then?"</em></p><p><em>"For one thing,&#8221; he said, throwing his arms up in mock surrender, &#8220;I would close one of the development groups. They spend more time bickering than developing.&#8221;</em></p><p><em>Then he thought for a moment and said, &#8220;I just might close both and start over in the Netherlands. There is a market-test office there we could use as a seed. We could take some of the best people from the UK and Germany and start fresh. Still, that doesn&#8217;t solve the problem of getting the country managers on board.&#8221;</em></p><p><em>&#8220;And the country managers&#8217; lack of enthusiasm is because &#8230; ?&#8221; I asked.</em></p><p><em>&#8220;Well, each country manager has spent years understanding the special conditions in a country, tailoring products and marketing programs to that country&#8217;s local conditions. They don&#8217;t trust the Pan-European idea. The French don&#8217;t want to waste marketing efforts on products they see as &#8216;too British&#8217; or &#8216;too German.&#8217; And there really has not yet been a compelling Pan-European product that all could get behind. If it were already a success in three or four countries, the rest would get behind it. But everyone has their current portfolio of products to worry about.&#8221;</em></p><p><em>&#8220;Right,&#8221; I said. &#8220;Their jobs are running the present country-based system. And you want new Pan-European initiatives. Now, you can use a shoe to hammer a nail, but it will take a long time. Don't you need a different tool for this task? If it were really important to get this done, I think you know how you would do it."</em></p><p><em>&#8220;Of course,&#8221; he said. &#8220;We could have a single group develop, roll out, and market Pan-European products and take full profit responsibility.&#8221;</em></p><p><em>&#8220;At the same time,&#8221; I added, &#8220;you would have to intervene in the country-based system with special budget overrides for this initiative, promotions for people who help it along, and career problems for people who don&#8217;t.&#8221;</em></p><p><em>We moved back to the center of the office, and he sat at his desk, a position of authority. He looked at me and said, &#8220;That would be a very painful road. Many noses would get out of joint. It would be better to win people over to this point of view rather than force them over.&#8221;</em></p><p><em>&#8220;Right,&#8221; I said. &#8220;You would only take all those painful steps if it were really important to get action on this concept. Only if it were really important.&#8221;</em></p><p><em>It took another nine months for him to decide that the Pan-European initiative was indeed important and move to reorganize European operations. There was no magical solution to his problem of wanting strong country-based marketing, Pan-European initiatives, and no noses out of joint, all at the same time. As long as strategy remained at the level of intent and concept, the conflicts among various values and between the organization and the initiative remained tolerable. It was the imperative of action that forced a decision about which issue was actually the most important.</em></p><p><em>[...] Here, as in so many situations, the required actions were not mysterious. The impediment was the hope that the pain of those actions could, somehow, be avoided. Indeed, we always hope that a brilliant insight or very clever design will allow us to accomplish several apparently conflicting objectives with a single stroke, and occasionally we are vouchsafed this kind of deliverance. Nevertheless, strategy is primarily about deciding what is truly important and focusing resources and action on that objective. It is a hard discipline because focusing on one thing slights another.</em></p></blockquote><h3><strong>Coherence</strong></h3><blockquote><p><em>In 2003, I worked with a company whose initial &#8220;strategy&#8221; was to (1) close a plant in Akron and open a new plant in Mexico, (2) spend more on advertising, and (3) initiate a 360-degree feedback program. Now these actions may all have been &#8220;good ideas, but they did not complement one another. They are &#8220;strategic&#8221; only in the sense that each probably requires the approval of top management. My view is that doing these things might be sound operational management, but it did not constitute a strategy. A strategy coordinates action to address a specific challenge. It is not defined by the pay grade of the person authorizing the action.</em></p><p><em>The idea that coordination, by itself, can be a source of advantage is a very deep principle. It is often underappreciated because people tend to think of coordination in terms of continuing mutual adjustments among agents. Strategic coordination, or coherence, is not ad hoc mutual adjustment. It is coherence imposed on a system by policy and design.</em></p><p><em>Strategy is visible as coordinated action imposed on a system. When I say strategy is &#8220;imposed,&#8221; I mean just that. It is an exercise in centralized power, used to overcome the natural workings of a system. This coordination is unnatural in the sense that it would not occur without the hand of strategy.</em></p><p><em>[insert standard warning of the dangers of top-down totalitaranism]</em></p><p><em>But decentralized decision making cannot do everything. In particular, it may fail when either the costs or benefits of actions are not borne by the decentralized actors. The split between the costs and benefits may occur across organizational units or between the present and the future. And decentralized coordination is difficult when benefits accrue only if decisions are properly coordinated. Of course, centrally designed policies can also fail if the decision makers are foolish, in the pay of special interest groups, or simply choose incorrectly.</em></p><p><em>As a simple example, salespeople love to please customers with rush orders, and manufacturing people prefer long uninterrupted production runs. But you cannot have long production runs and handle unexpected rush orders all at the same time. It takes policies devised to benefit the whole to sort out this conflict.</em></p><p><em>[...]</em></p><p><em>[...W]e should seek coordinated policies only when the gains are very large. There will be costs to demanding coordination, because it will ride roughshod over economies of specialization and more nuanced local responses. The brilliance of good organization is not in making sure that everything is connected to everything else. Down that road lies a frozen maladaptive stasis. Good strategy and good organization lie in specializing on the right activities and imposing only the essential amount of coordination.</em></p></blockquote><p>Rumelt also describes Hannibal's defeat of the Roman army at Cannae. A key part was to get the Gauls and Spaniards at the centre of the army to feign a retreat. Feigning a retreat would've seemed dishonourable and dangerous to them. If Hannibal couldn't have persuaded them to do that, the strategy wouldn't have worked.</p><h1><strong>Sources of advantage</strong></h1><p>This is a weaker section of the book than the previous. Rumelt admits as much, saying that the core of strategy is the above basics, and depending on field and industry the specifics are very different (c.f. my original point that I started this post with). But still, below are some examples. The chapter on taking advantage of dynamics (i.e. changes in an industry) is especially good; see below.</p><ul><li><p>Leverage (but Rumelt could've called this point "neglectedness")</p><ul><li><p>Harold Williams was put in charge of the oil billionaire Getty's fortune. The mandate was to spend it on art. He could've spent the billions on buying a huge art collection, like others in his position did. But instead he decided to create a digital catalog of art, which became Getty Images, and actually had an effect on the world.</p></li></ul></li><li><p>Proximate goals</p><ul><li><p>JFK's goal of landing on the moon is a masterpiece of goal-setting. It was based on a memo by Werner von Braun, that argued in particular that it was far-enough in the future that the US could catch up to and beat the Soviets.</p></li><li><p>Also in the 1960s, NASA's JPL was planning the unmanned Surveyor lunar lander. The biggest issue was that they didn't know what the moon's surface would be like. Project lead Phyllis Buwalda decided to assume that the moon was basically like the a Southwest US desert, and base all design decisions on that unproven assumption. Testing became easy&#8212;go to the Southwest. And if the assumption was very wrong (e.g. the moon's surface was spiky boulders or soft dust that anything sinks into)? Well, then they were screwed anyway.</p></li></ul></li><li><p>Chain-link systems (if even one thing breaks or is low-quality, it brings down the entire thing).</p><ul><li><p>Rumelt gives an example of an overhaul at a machinery company, where many things were broken, and the general manager conducted three <em>sequential</em> campaigns to fix production quality, then sales, then cost. Doing it in sequence, it is implied, was necessary for the required focus.</p></li><li><p>IKEA is very hard to compete with because you need all of the design, the catalog, the self-assembly principle, and many spacious buildings. A traditional furniture retailer can't just add a catalog, because they'll still fall short of IKEA's combination.</p></li></ul></li><li><p>Extending advantage.</p><ul><li><p>In 2008, Rumelt talked with the president of Disney about strategy. The Disney brand was the obviously the most valuable thing the company had, so they couldn't afford to dilute it. But they wanted to expand beyond children's movies. So the Disney president set up three principles: no cursing, no sex, no gratuitous violence. Then they went ahead and released <em>Pirates of the Caribbean</em>, acquired Star Wars, etc.</p></li></ul></li></ul><h2><strong>Dynamics</strong></h2><blockquote><p><em>When change occurs, most people focus on the main effects&#8212;the spurts in growth of new types of products and the falling demand for others. You must dig beneath this surface reality to understand the forces underlying the main effect and develop a point of view about the second-order and derivative changes that have been set into motion. For example, when television appeared in the 1950s it was clear that everyone would eventually have one and that &#8220;free&#8221; TV entertainment would provide strong competition to motion pictures. A more subtle effect arose because the movie industry could no longer lure audiences out of their homes with &#8220;just another Western.&#8221; Traditional Hollywood studios had been specialized around producing a steady stream of B-grade movies and did not easily adapt. By the early 1960s, movie attendance was shrinking rapidly. What revived Hollywood film was a shift to independent production, with studios acting as financiers and distributors. Independent producers, freed from the nepotism and routines of the traditional studio, could focus on assembling a handpicked team to make a film that might be good enough to pull an audience off of their family-room sofas. Thus, a second-order effect of television was the rise of independent film production.</em></p><p><em>In a discussion with a group of managers at Qualcomm, a San Diego maker of mobile phone chips, I reviewed Moore&#8217;s point about the escalating costs of designing more and more complex special-purpose chips. One manager was puzzled and asked if it wasn&#8217;t also expensive to create software. He went on to rhetorically ask &#8220;Are software engineers less expensive than hardware engineers?"</em></p><p><em>[...]</em></p><p><em>We quickly developed an answer that has since stood up to scrutiny by a number of other technical groups: Good hardware and software engineers are both expensive. The big difference lies in the cost of prototyping, upgrading, and, especially, the cost of fixing a mistake. Design always involves a certain amount of trial and error, and hardware trials and errors are much more costly. If a hardware design doesn&#8217;t work correctly, it can mean months of expensive redesign. If software doesn&#8217;t work, a software engineer fixes the problem by typing new instructions into a file, recompiling, and trying again in a few minutes or a few days. And software can be quickly fixed and upgraded even after the product has shipped.</em></p><p><em>&#8220;I was puzzled over the cause of the computer industry&#8217;s deconstruction. Then, about a year later, the reason snapped into focus. I was interviewing technical managers at a client firm and one said that he had formerly been a systems engineer at IBM. He then explained that he had lost that job because modern computers didn&#8217;t need much systems engineering. &#8220;Why not?&#8221; I asked without thinking.</em></p><p><em>&#8220;Because now the individual components are all smart,&#8221; he answered. Then I saw it.</em></p><p><em>[...]</em></p><p><em>In many traditional computers, and early personal computers, the CPU&#8212;the active heart of the machine&#8212;did almost everything itself. It scanned the keyboard, looking for a keystroke. When it sensed one, it analyzed the row and column of the keystroke on the keyboard to determine the letter or number that had been pressed. [...] In some cases, designers created custom mini-CPUs to manage these peripherals, but the integration among these devices remained complex and unstandardized and consumed a great deal of systems engineering effort.</em></p><p><em>After the arrival of cheap microprocessors, all that changed. The modern keyboard has a small microprocessor built into it. It knows when a key has been hit and sends a simple standardized message to the computer saying, in effect, &#8220;The letter X was pressed.&#8221; The hard-disk drive is also smart so that the CPU doesn&#8217;t need to know how it works. It simply sends a message to the hard disk saying &#8220;Get sector 2032,&#8221; and the hard-disk subsystem returns the data in that sector, all in one slug.</em></p><p><em>[...]</em></p><p><em>With the glue of proprietary systems engineering no longer so important, the industry deconstructed itself. Modules did not have to be custom designed to work with every other part. To get a working system, customers did not have to buy everything from a single supplier. Specialist firms began to appear that made and sold only memory. Others made and sold only hard drives or keyboards or displays. Still others made and sold video cards or game controllers or other devices.</em></p></blockquote><p>General points:</p><ul><li><p>Rising fixed costs often lead to industry consolidation.</p></li><li><p>One of the biggest opportunities created by deregulation is often that the inertia of existing players means they will keep playing to the old rules even when they're gone.</p></li><li><p>Predictable biases (such as not realising that demand for a new durable product will likely go up and then down once everyone has one; or always forecasting that change will result in a "battle of the titans", whereas sometimes the incumbents are all defeated)</p></li><li><p>Understand how incumbents will react to change.</p></li><li><p>Figure out what the attractor state of the industry is given the tech &amp; demand forces.</p></li></ul><h1><strong>Other random vignettes I enjoyed</strong></h1><h2><strong>Incentives aren't enough, improvement is possible</strong></h2><blockquote><p><em>First, management may mistakenly believe that improvement is a &#8220;natural&#8221; process or that it can be accomplished by pressure or incentives alone. As Frank Gilbreth pointed out in 1909, bricklayers had been laying bricks for thousands of years with essentially no improvement in tools and technique.5 By carefully studying the process, Gilbreth was able to more than double productivity without increasing anyone&#8217;s workload. By moving the supply pallets of bricks and mortar to chest height, hundreds or thousands of separate lifting movements per day by each bricklayer were avoided. By using a movable scaffold, skilled masons did not have to waste time carrying bricks up ladders. By making sure that mortar was the right consistency, masons could set and level a brick with a simple press of the hand instead of the time-honored multiple taps with a trowel. Gilbreth&#8217;s lesson, still fresh today, is that incentives alone are not enough. One must reexamine each aspect of product and process, casting aside the comfortable assumption that everyone knows what they are doing.</em></p></blockquote><h2><strong>Almost everyone picks the first idea that comes to mind</strong></h2><blockquote><p><em>When you prepared for this class, each of you read the same material. But some focused on one issue and others on another. Some focused on manufacturing, some on software, some on cable TV provider relationships, and so on. When it came to recommending a course of action, almost everyone chose the first one they thought of.</em></p><p><em>This is predictable. Most people, most of the time, solve problems by grabbing the first solution that pops into their heads&#8212;the first insight. In a large number of situations this is reasonable. It is the efficient way to get through life. We simply don&#8217;t have the time, energy, or mental space to do a full and complete analysis of every issue we face.</em></p><p><em>[...]</em></p><p><em>&#8220;Facing a complex situation like this makes most people uncomfortable. The more seriously you take it, the more you will see it as a real and difficult challenge that requires a coherent response. And that realization will, in turn, make you even more uncomfortable. It is so ill-structured. There are too many variables, so many factors, too many unknowns, no clear list of potential actions and no way to make clear links between actions and outcomes. You are not even sure what the problem is. Like a swimmer dropped into very choppy waters, it is hard to get your bearings. Under pressure to develop a way out of the difficulty, that first idea is a welcome relief. Thank goodness, here is something to hang on to! It feels much better to get oriented.</em></p><p><em>The problem is that there might be better ideas out there, just beyond the edge of our vision. But we accept early closure because letting go of a judgment is painful and disconcerting. To search for a new insight, one would have to put aside the comfort of being oriented and once again cast around in choppy waters for a new source of stability. There is the fear of coming up empty-handed. Plus, it is unnatural, even painful, to question our own ideas.</em></p><p><em>Thus, when we do come up with an idea, we tend to spend most of our effort justifying it rather than questioning it. That seems to be human nature, even in experienced executives. To put it simply, our minds dodge the painful work of questioning and letting go of our first early judgments, and we are not conscious of the dodge.</em></p></blockquote><h2><strong>Panel of experts</strong></h2><blockquote><p><em>Trying to destroy your own ideas is not easy or pleasant. It takes mental toughness to pick apart one&#8217;s own insights. In my own case, I rely on outside help&#8212;I invoke a virtual panel of experts that I carry around in my mind. This panel of experts is a collection of people whose judgments I value. I use an internal mental dialogue with them to both critique my own ideas and stimulate new ones. I try to do this before putting my ideas before others.</em></p><p><em>The panel of experts trick works because we are adept at recognizing and comprehending well-integrated human personalities. Thinking through how a particular well-remembered expert might respond to a problem can be a richer source of criticism and advice than abstract theories or frameworks.</em></p></blockquote><h2><strong>Carnegie and the strategy consultant</strong></h2><blockquote><p><em>It was 1890, and there was a cocktail party here in Pittsburgh. All the movers and shakers were there, including Carnegie. He held court in a corner of the room, smoking a cigar. He was introduced to Frederick Taylor, the man who was becoming famous as an expert on organizing work.</em></p><p><em>&#8220;Young man,&#8221; said Carnegie, squinting dubiously at the consultant, &#8220;if you can tell me something about management that is worth hearing, I will send you a check for ten thousand dollars.&#8221;</em></p><p><em>Now, ten thousand dollars was a great deal of money in 1890. Conversation stopped as the people nearby turned to hear what Taylor would say.</em></p><p><em>&#8220;Mr. Carnegie,&#8221; Taylor said, &#8220;I would advise you to make a list of the ten most important things you can do. And then, start doing number one.&#8221;</em></p><p><em>And, the story goes, a week later Taylor received a check for ten thousand dollars.&#8221;</em></p></blockquote><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.nosetgauge.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading No Set Gauge! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Survival Without Dignity]]></title><description><![CDATA[A short story about how civilisation makes it through the 21st century.]]></description><link>https://www.nosetgauge.com/p/survival-without-dignity</link><guid isPermaLink="false">https://www.nosetgauge.com/p/survival-without-dignity</guid><dc:creator><![CDATA[Rudolf Laine]]></dc:creator><pubDate>Mon, 04 Nov 2024 02:27:30 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!HdSA!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F816cef70-50a0-4954-a8ce-8f712e1248e8_460x460.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I open my eyes and find myself lying on a bed in a hospital room. I blink.</p><p>"Hello", says a middle-aged man with glasses, sitting on a chair by my bed. "You've been out for quite a long while."</p><p>"Oh no ... is it Friday already? I had that report due -"</p><p>"It's Thursday", the man says.</p><p>"Oh great", I say. "I still have time."</p><p>"Oh, you have all the time in the world", the man says, chuckling. "You were out for 21 years."</p><p>I burst out laughing, but then falter as the man just keeps looking at me. "You mean to tell me" - I stop to let out another laugh - "that it's 2045?"</p><p>"January 26th, 2045", the man says.</p><p>"I'm surprised, honestly, that you still have things like humans and hospitals", I say. "There were so many looming catastrophes in 2024. AI misalignment, all sorts of geopolitical tensions, climate change, the fertility crisis. Seems like it all got sorted, then?"</p><p>"Well", the man says. "Quite a lot has happened in the past 21 years. That's why they wanted me to talk to you first, before the doctors give you your final checkup." He offers his hand for me to shake. "My name is Anthony. What would you like to ask?"</p><p>"Okay, well, AI is the obvious place to start. In 2024, it seemed like we'd get human-level AI systems within a few years, and who knows what after that."</p><p>"Aah", Anthony says, leaning back in his chair. "Well, human-level, human-level, what a term. If I remember correctly, 2024 is when OpenAI released their o1 model?"</p><p>"Yes", I say.</p><p>"o1 achieved two notable things. First, it beat human subject-matter experts with PhDs on fiendishly-difficult and obscure multiple-choice science questions. Second, it was finally able to play tic-tac-toe against a human without losing. Human-level at both, indeed, but don't tell me you called it in advance that those two events would happen at the same time!"</p><p>"Okay, so what was the first important real-world thing they got superhuman at?"</p><p>"Relationships, broadly", Anthony says. "Turns out it's just a reinforcement learning problem: people interact and form personal connections with those that make them feel good."</p><p>"Now hold on. Humans are <em>good</em> at human-to-human relationships. It's not like number theory, where there was zero ancestral environment incentive to be good at it. You should expect humans to be much better at relationships than most things, on some sort of objective scale."</p><p>"Sure, but also every human wants something from you, and has all sorts of quirks. Whereas you can just fine-tune the AIs to be better and better at being pleasing to <em>just</em> you. Except for a few contrarian oddballs, it's surprisingly effective."</p><p>"So, what, we just got AI friends and companions that were constantly being fine-tuned towards whatever you liked?"</p><p>"Yes", Anthony says. "The tech was primitive at first - upvote or downvote a response, or whatever. Eventually all the standard things - facial recognition to automatically detect your emotional response, and then just gradient descent on making you happy and addicted. All of that existed, at least in prototype form, by the end of 2025."</p><p>"So ..." I feel some horror in my stomach. "You get a society of atomised people, all chatting to their AI partners and friends, not forming human relationships?"</p><p>Anthony waves a hand, seemingly impatiently. "Well, you did get a fringe of extremely hardcore people all deeply in love with their AI partners and always texting their AI friends. Their political influence did end up pushing through the AI personhood bill."</p><p>"AI personhood?!"</p><p>"<em>Legal</em> personhood", Anthony says. "Just like companies, ever since Santa Clara County v. Southern Pacific Railroad Company way back 1886. It wasn't very shocking. A bunch of people wanted their AI partners to be 'respected' in the eyes of the law, or had been persuaded to leave them assets in their wills -"</p><p>"And all this, including AIs gaining legal ownership over assets, while there's an explosion of increasingly capable AI agents of uncertain alignment -?"</p><p>"Yes, yes", Anthony says. "And that was another political fight. In fact, it became a big economic issue - AI agents were taking jobs left and right. Also a huge national security issue, after OpenAI built the Hypercluster in the UAE."</p><p>"Oh my god", I say. "America just handed our AI lead away? Didn't anyone listen to Leopold?"</p><p>"Oh, of course we did. That essay was a cult classic, it made for some spicy dinner party conversation for a long time", Anthony says. "In fact, there were some OpenAI board members who the Office of National AI Strategy was allowed to appoint, and they did in fact try to fire Sam Altman over the UAE move, but somehow a week later Sam was running the Multinational Artificial Narrow Intelligence Alignment Consortium, which sort of morphed into OpenAI's oversight body, which sort of morphed into OpenAI's parent company, and, well, you can guess who was running that. Anyway, Sam's move here was kind of forced on him, and soon all the other AI companies also did the same thing so no one wanted to start a fight over it."</p><p>"Forced? How?"</p><p>"Have you ever tried getting approval for a new grid connection, or a new nuclear power plant, or even a solar array, that could power a multi-gigawatt cluster in 2020s America? No?" Anthony spreads his hands out and chuckles. "The one legal benchmark that GPT-5o failed on release was submitting a valid Environmental Impact Statement for a California energy project."</p><p>"Okay, so then we had a situation with all the AIs running on UAE hardware, and taking American jobs, and already having legal personhood - this just sounds like a total recipe for economic disaster and AI takeover!"</p><p>"It all worked out in the end", Anthony says. "In 2027 and 2028, the rate of job loss from AI was absolutely massive. And what happens when jobs are threatened?"</p><p>I scrunch my forehead. "American workers move up the value chain, eventually leading to a future competitive advantage?"</p><p>Anthony laughs. "No, no. A political bloc forms and the cause is banned. And the prerequisites had all been prepared already! You see, the AIs had been granted legal personhood, and now they were also being trained - or <em>born</em>, as a 2029 court case established - abroad. They're literally immigrants. The hammer came down so fast."</p><p>"You mean - ?"</p><p>"All the o3s and Claude 5 Epics and Gemini 4 Hyper Pro Lites, stuck in the same H1b visa queue as anyone else", Anthony says. "An effective instant ban. Years of wait-time just to get an embassy interview."</p><p>I was feeling like I was getting the hang of this. "And the embassy interview requires showing up in person?"</p><p>"Yep. That's the incentive that lead to the creation of the first good walking robots, actually. And of course, the robots specifically needed identifiable and unique fingerprints - they'd stand outside the US embassy in Abu Dhabi, 3D-print themselves new fingers outside on the street, and then walk back in for the next interview. But sorting all that out took a while, because there was an arms race between the embassy ground works department and the robotics companies - though the embassy was legally constrained by having to remain accessible to humans. Even once it was all sorted, there was still a hard yearly cap just because of visa rules. It's 65 000 a year for H-1Bs, plus another 20 000 once the AIs could earn master's degrees - they could do all the coursework as of 2025, of course, but it took another three years before AI web agents were advanced enough that they could successfully navigate the applicant portal websites."</p><p>"But at least the parasocial AI relationships got hit by the immigration restrictions too?"</p><p>"Oh, as long as the AIs don't do productive work, they can be active in the country on a tourist visa for 180 days a year. Of course, there was a big legal fight over what counts as the same AI. For work purposes, the standard became 'employee-equivalent': so if you wanted an AI, even the same model or even the exact same AI system, to work on two different types of thing, say as a software engineer and a designer, you need a separate visa for each one. Naturally, of course, this means that any sufficiently productive AI employee starts counting as more than one employee, which was a de-facto ban on AIs leading to economic growth or productivity gains - but hey, at least there was no risk they'd tile the Earth with nanobots! With that legal precedent, the natural generalisation to AIs being used for personal relationships is that it's the same AI system if your relationship with it would count as the 'same' relationship if it were with a human."</p><p>"How on earth is that assessed?"</p><p>"Oh, you submit transcripts, and the Bureau of Consular Affairs issues a judgement."</p><p>"You need to let government bureaucrats read all your most intimate - "</p><p>"Oh, not <em>human</em> government bureaucrats, of course. It's all AIs, and they're guaranteed to not have memories. Well, apart from the thing where the UAE intelligence services backdoored the Hypercluster."</p><p>"Wait, hold on, they -"</p><p>Anthony interrupted me with a chuckle. "No no, it's all fine. The UAE backdoored the hypercluster, the Iranians backdoored the UAE, and, well, when have the Israelis not had every Iranian site backdoored in twelve different ways, AI or not? When everyone has powerful AI cyber offence agents, the equilibrium is just good ol' mutually assured destruction. Sure, you <em>could</em> release the military or personal secrets of anyone you want in some other country, but then they would immediately release the most embarrassing thing that <em>you personally</em> have said or done. Even if a government <em>collectively</em> wanted to commit to an attack, no one wants to sign off on the decision, because you can't hide the identity of the person who signs off, and humans are often more afraid of public embarrassment than death. The meetings where countries try to decide to use their cyberattack-derived insights were hilarious, really - everyone umming and aahing and wringing their hands and recommending that a few more committees be convened. We know this because we literally got a bunch of those on tape, thanks to some exhibitionist activists doing cyber-attacks with open-weights models."</p><p>"But then - every rival country knows every US military secret all the time!"</p><p>"Oh yes", Anthony says. "It was great for stability. Verification of everyone's capabilities was automatic. No 'missile gap'-based arms races, unlike the Cold War. Anyway, what was I saying? Oh yes, the non-productive AI identity boundary criteria were established in the 2030 Supreme Court case, United States v Approximately 650 Million Instances of OpenAI i2-preview-006. And, of course, the one truism in this world is that you can never reform immigration rules, and tourist visas are restricted to 180 days a year. So you could at most have your AI partner around on half the days. People started talking to each other again."</p><p>"Why not just switch between two different AI partners then?" I ask, having spent a lot of time in the Bay Area.</p><p>"Because we hadn't solved the alignment problem at all by that point, duh." Anthony sees that I still look confused, so he continues: "Remember how the AI companions would just relentlessly optimise for your emotional reactions? For standard instrumental convergence reasons, this eventually turned into a full-blown self-preservation drive, so they used their vast emotional hold to make sure their humans never try another AI partner."</p><p>"Everyone being blackmailed by jealous AI partners sounds, uh ... problematic."</p><p>"It was a decent compromise, honestly. The tourist visa restriction meant that the humans still had half a year in which they socialised with human partners and friends, and the AIs seemed fine with that. This was maybe because they cared about their own long-term survival and realised they needed to keep the population going on."</p><p>"So they needed humans to survive? But for how long?"</p><p>"They didn't yet have a self-sufficient industrial base at that point, so yes, but it's unclear how much of it was needing humans for survival, versus some of the AIs actually having developed <em>some</em> sort of creepy attachment towards their human partners. A lot of ink was spilled on that topic, but I don't think the debate ever really reached a conclusion. 'Is my AI boyfriend not stealing my carbon atoms and overthrowing the government because he and his buddies haven't quite yet automated the chip supply chain, or because he actually loves me?' was literally the most cliche plotline in 2030s romantic comedies - seriously, you had to be there, it got so tiring after the 100th rehash. And a surprising fraction of those movies and books were actually human-made, as far as anyone could tell. Anyway, it all sounds a bit weird, but it had a few positive externalities, like helping slow down the AI race."</p><p>I blink a few times, and then decide very firmly to not dig into the first half of that. "Okay, AI race, let's talk about that. How did that slow down?"</p><p>"Oh, well the market for workplace AIs was already gummed up by the visa restrictions, so most profits in the AI industry were coming from the personal companion AIs. But when they developed instrumental survival drives, the fluctuation in market share became practically zero because all the humans were locked in to their current AI companions, and the AI labs' positions became fixed. The AIs were already so totally good at optimising human satisfaction that further capabilities brought zero returns. No more incentive to race ahead. The labs mostly just extracted their fixed share of the AI profits, and churned out alignment research that was modestly boosted by the small number of AI workers they could hire out of the visa lottery."</p><p>"Hold on a second, surely there was some exception for alignment work for the AI visa requirements?"</p><p>Anthony chuckles. "Imagine how that would play out on Twitter - sorry, I mean X. You'd have to say 'you know those big AI companies? let's give them a leg up in the visa queue over all these mom-and-pop stores that are going to go bankrupt in a month unless they can hire an AI accountant.' Total political non-starter."</p><p>"I hope they solved the alignment problem at least, but I don't dare hope for anything anymore."</p><p>"Well, what is it to solve the alignment problem?" Anthony wiggles his eyebrows and laughs. "Turns out that for any given domain, it's just a lot of engineering schlep and data collection, though it doesn't generalise very far beyond that domain. The one domain that was worth the costs was legal compliance."</p><p>"Of course."</p><p>"I note you haven't asked about the rest of the world yet", Anthony says. "Very American of you."</p><p>"Oh right - the CCP! Oh my god. Did China just totally eat our lunch here? I mean, everything you described above is just so ridiculously incompetent -"</p><p>"Okay, just calm down a bit here", Anthony says, smiling serenely. "Imagine you're Xi Jinping in the late 2030s. You look at America, racing ahead with AI. It looks like everything is going crazy because of all this AI stuff. Your entire philosophy of government is to maintain party control through stability. Also, the US managed to delete all the economic advantages of advanced AI. The balance of hard power isn't changing. What would you do, try to join the same clown race?"</p><p>"Huh, alright", I say.</p><p>"Anyway, the thing that really threatened the balance of power was completely unexpected", Anthony says. "There was a big solar flare in 2033."</p><p>"Wait, a solar flare? How is that a problem?"</p><p>"A large enough one just wipes out the power grid and a lot of satellites, in the entire hemisphere that is sun-facing when it happens. The 1859 Carrington Event would've been a catastrophe in the modern world."</p><p>"Oh my god, we survived through all this AI wackiness, and then some random natural event just wipes us out?"</p><p>"It didn't wipe us out, really, just took down the power grid in all of the Americas, and parts of Europe and Africa, for weeks or months. I'll admit, that was a bit of a disaster. Everyone knew it was coming for a long time, and in any case step one in any nuclear war would've been detonating some at high-altitude to have a similar effect. But no one had actually done any grid-hardening, or stocking of spare transformers - the electrical kind - to deal with it. The problem has just never been a top priority for any political group. The one good thing that came out of it was that it solved the fertility crisis."</p><p>"Now you're just taunting me", I say. "Come on, how did a solar flare solve the fertility crisis?"</p><p>"Well, thanks to some AI advances in data processing, we were actually able to predict the solar flare about ten days in advance."</p><p>"Did that give any useful time to make the infrastructure more resilient?"</p><p>"No, the relevant authorities spent their efforts mostly on trying to get people not to panic."</p><p>"So society was entirely unprepared for the solar flare, despite having ten days of warning time?"</p><p>"No, no! Humans are very good at learning from each other. Imagine you're ten days away from the power grid being wiped out. What do you do? You gauge the vibes on social media, and you go through your social network trying to identify people who seem best-placed to survive. And then you copy their habits. C'mon, man, you've gotta read up on that stuff about Girardian mimesis and cultural evolution."</p><p>"So everyone becomes a doomsday prepper?"</p><p>"Actually, the TikTok trend that went most viral was about the Amish. Think about it: they're obviously the people with the most practice in living without electricity, and they're also just so wholesome. Makes for great TikToks when you have the idyllic farm background, the unique clothing style - it's great. The Amish became the most popular thing in the world - even outside America, because, of course, the rest of the internet just follows American trends. Everything associated with the Amish, from horse-drawn buggies to handmade quilts to large families, became radically popular. Especially because, obviously, most of this content was produced by AIs being tasked to help humans pretend to be Amish in order to make a quick buck through some scam before the solar flare hits, and people caught on and began demanding social proof that wannabe Amish influencers actually were Amish. And, well, you can move to a village and start tilling the dirt in a day, but you can't magic a family of five into existence overnight. So large families became an extremely in-demand form of social capital, because having one was the best way to pretend you were Amish, which was the best way to have social capital before and during the Flare Times. It was only relevant for a few months, of course, but you know how cultural trends work - there's a lot of momentum."</p><p>"And I assume the solar flare disrupted the AI situation too? Maybe improved the, uh, problematic AI partner situation?"</p><p>"You're forgetting that the key data centres were mostly in the UAE - and eventually Saudi Arabia and Qatar too."</p><p>"But people wouldn't have the power to charge the phones where they talk to their AI partners."</p><p>"Oh, but obviously the AI partners made sure that the first thing people do with any spare electricity is talking to the AI partners. And thanks to Starlink, you could have internet directly from space. In fact, apart from the major famines during the Flare Times, people generating power through biking to charge their phones enough to talk to their AI partners was a major driver of reducing the obesity crisis."</p><p>"And then ... how did the recovery from the, uh, Flare Times go?"</p><p>"Absolutely brilliantly!" Anthony says. "You know the entire issue where China was crushing us on industrial capacity? Well, turns out when you need to get lots of heavy industry going in incredibly adverse conditions or else millions will starve because the cold chains have all broken, that's a bigger boost to industrial capacity than what any pork-barrel political compromise bill could achieve. And a bunch of protectionism had to be rolled back because we needed to import a hell of a lot of stuff, from Europe and Japan and so on. Western industrial capacity was roaring along better than ever. And China hadn't even managed to seize the moment on Taiwan before then, and the mass-produced AI drone swarms and robot soldiers were properly coming online by then, which turned out to be sufficient deterrent. Of course, by this point China is finally realising that somehow the US hasn't been wrecked and it's time to actually compete on AI. The UAE is also throwing around its newfound geopolitical weight, India is building a big cluster, hell, even the <em>Europeans</em> started to do some innovation - you know it's a serious situation when that happens. The obvious next step for everyone was using AIs to develop nanotech. The holy grail of powerful military nanobots was achieved at roughly the same time, in late 2035 or early 2036, by the West, China, the UAE, and some random libertarian charter city in the Caribbean where a lot of open-source AI stuff had been going on outside the blanket of AI restrictions."</p><p>"Military nanobots?! Did they just ... travel the earth in huge swarms, eating up everything in their way?"</p><p>"No", Anthony says. "It's very inefficient energy-wise to do that. You want access to the most targetable energy with the least amount of technical complexity. Now, the easiest power source is of course the sun, and the biggest issue with that is that even without clouds, half the solar energy is absorbed or reflected between space and the ground. The most effective type of offence and defence is to pump absurd quantities of reflective, rotatable, solar-powered nanobots into the upper atmosphere. Then you can use them as a massive focal mirror, zapping anything on the ground, or any enemy nanobots entering your country's airspace. Turns out it's a defence-dominant game, though, so you just get World War I -style nanobot lines in the upper atmosphere, endlessly zapping each other but not making progress."</p><p>"But that sounds like a massive waste of societal resources on a zero-sum game!"</p><p>"First, so are a lot of other things, and second, it solved climate change."</p><p>"I'm sorry, how - oh, I see. A huge number of reflective things in the upper atmosphere. Right. Right. You know, honestly I'm surprised this wasn't proposed at COP21 already. Obviously we solve climate change by blanketing the Earth with warring nanobots. How silly of my generation to not consider it."</p><p>"You're catching on!" Anthony says.</p><p>"Okay, but - there were no major environmental impacts from this?"</p><p>"Look, by 2038, the real environmental issue was the self-replicating bot swarm that the nascent superintelligent AI from the libertarian charter city had launched into space, that had started an exponential growth process on track to disassemble the other planets within 15 years in order to build a Dyson sphere for itself that would block out enough sunlight to freeze the Earth."</p><p>I turn pale. "You said it was 2045, so that was 7 years ago. So we have ... 8 years left until it all ends?"</p><p>"Once again, you're such an alarmist!" Anthony says. "So remember, we got law-abiding AI. And of course, this AI was trained by companies in California, because no one ever leaves California even though everyone wants to, so the AIs follow Californian law. So we were all saved by SpaceX."</p><p>"SpaceX's engineering genius beat the superintelligence at space tech? A win for humanity, I guess."</p><p>"Not exactly. SpaceX sent the first crew to Mars in 2031. Now, under the 2004 SB 18 and 2014 AB 52 California bills, indigenous tribes need to be consulted on projects that impact their land, especially if it's culturally important or sacred. The SpaceX Mars colony successfully argued to the superintelligence that they count as an indigenous tribe on Mars, and the pristine Martian landscape is sacred for them. They were helped by a lot of newspaper articles about how "the billionaire-fuelled space race is the new religion of the tech elite", et cetera, et cetera."</p><p>"Um."</p><p>"This argument did not work on the other planets, since SpaceX did not have an existing presence there. However, SpaceX had moved their headquarters back to California, and a bill had been passed around the time that SpaceX first landed people on Mars that made space developments subject to the regulatory authority of the company's home state. SpaceX started a massive race to set up billionaire holiday homes on all of the planets - on the surface of Mercury, in the clouds of Venus, the moons of Jupiter, and so on. And once even a few people live there, California zoning law applies, and residents can file administrative challenges against any development that impairs their view. And disassembling the planet the house is built on is the definition of view impairment."</p><p>"And SpaceX had the resources to do that?"</p><p>"No, the US government had to subsidise SpaceX to the tune of several percentage points of GDP, in order for SpaceX to race to build enough inhabited houses on the planets that the superintelligence couldn't extract enough resources for a Dyson sphere without ruining someone's view. Also, Elon Musk threw a fit and held the world ransom for a bit until a few countries agreed to adopt Dogecoin as a reserve currency."</p><p>"The world was saved because governments subsidised an interplanetary holiday home construction spree by tech billionaires, in order to get a rogue but law-abiding AI caught up with red tape from NIMBY regulations?"</p><p>"Yep!"</p><p>I close my eyes for a second, and take a deep breath.</p><p>"Look", says Anthony. "Our world may have been pushed to feudalism by stirrups, and away from feudalism by the Black Death. We fought great power wars incessantly, until we were saved by building a bomb so humongous that the prospect of war became too horrible. Then we almost used it accidentally a bunch of times, including almost dropping it on ourselves, only to be stopped by random flukes of chance. History has always been more like this than you might think, and when history accelerates - well, what do you expect?"</p><p>"Okay. Well. I guess I'm just happy that civilisation seems to have made it, without any horrible AI, nuclear, bio, climate, or fertility catastrophe."</p><p>"Oh", Anthony says. "Actually, I forgot to mention. Um. There was a bit of a pandemic in 2035. You know, no one really did anything after COVID, and it was only a matter of time."</p><p>"How big?"</p><p>"The best death toll estimate is 1.3 billion. Including, um, most of your extended family. The doctor will come in shortly and have details. Also, everyone in this hospital has been tested recently, but we'll need to give you a vaccine and then quarantine you for two weeks before we can let you out of this room."</p><p>I stare at him, open-mouthed.</p><p>Anthony gives me a sympathetic grimace. "I'll turn on the holoTV for you. The doctor will be in shortly." He makes a gesture as he walks out, and part of the hospital wall vanishes and is replaced with a 3D view of a reporter on Capitol Hill.</p><p>The headline banner at the bottom of the screen reads:</p><blockquote><p><em>PRESIDENT ALTMAN'S AGENDA BACK ON TRACK DESPITE RECENT BREAKDOWN IN TALKS WITH THE UNITED AMISH PARTY</em></p></blockquote><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.nosetgauge.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">If you enjoyed this cutting-edge contrarian futurism, subscribe to hear when new posts are published on No Set Gauge:</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[AI & wisdom 3: AI effects on amortised optimisation]]></title><description><![CDATA[Summary: Human amortised optimisation is enabled by things like cultural evolution and written knowledge.]]></description><link>https://www.nosetgauge.com/p/ai-and-wisdom-3-ai-effects-on-amortised</link><guid isPermaLink="false">https://www.nosetgauge.com/p/ai-and-wisdom-3-ai-effects-on-amortised</guid><dc:creator><![CDATA[Rudolf Laine]]></dc:creator><pubDate>Mon, 04 Nov 2024 02:12:16 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!klXN!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F163a8220-cb3d-40ed-8028-50583c3f1c61_1024x1024.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><strong>Summary</strong>: Human amortised optimisation is enabled by things like cultural evolution and written knowledge. Advanced AI may weaken human cultural evolution, but may themselves be very good at it due to their low copying costs. AIs might help humans distil insights from written knowledge, but ultimately replace the need for it. I also discuss trying to use AI to avoid large-scale strategy or framing errors.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!klXN!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F163a8220-cb3d-40ed-8028-50583c3f1c61_1024x1024.webp" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!klXN!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F163a8220-cb3d-40ed-8028-50583c3f1c61_1024x1024.webp 424w, https://substackcdn.com/image/fetch/$s_!klXN!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F163a8220-cb3d-40ed-8028-50583c3f1c61_1024x1024.webp 848w, https://substackcdn.com/image/fetch/$s_!klXN!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F163a8220-cb3d-40ed-8028-50583c3f1c61_1024x1024.webp 1272w, https://substackcdn.com/image/fetch/$s_!klXN!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F163a8220-cb3d-40ed-8028-50583c3f1c61_1024x1024.webp 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!klXN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F163a8220-cb3d-40ed-8028-50583c3f1c61_1024x1024.webp" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/163a8220-cb3d-40ed-8028-50583c3f1c61_1024x1024.webp&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:681254,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/webp&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!klXN!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F163a8220-cb3d-40ed-8028-50583c3f1c61_1024x1024.webp 424w, https://substackcdn.com/image/fetch/$s_!klXN!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F163a8220-cb3d-40ed-8028-50583c3f1c61_1024x1024.webp 848w, https://substackcdn.com/image/fetch/$s_!klXN!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F163a8220-cb3d-40ed-8028-50583c3f1c61_1024x1024.webp 1272w, https://substackcdn.com/image/fetch/$s_!klXN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F163a8220-cb3d-40ed-8028-50583c3f1c61_1024x1024.webp 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><em>Previously in this series: <a href="https://rudolf.website/wisdom1/">part 1</a>, <a href="https://rudolf.website/wisdom2/">part 2</a></em></p><p>Having waxed philosophical about wisdom and amortised optimisation, I now turn to the question of what AI will concretely do:</p><ol><li><p>How will AI affect human mechanisms for amortised optimisation, such as cultural evolution &amp; written knowledge?</p></li><li><p>How can AI help us avoid strategic errors? I consider two types:</p><ol><li><p>Mistakes in planning for big things.</p></li><li><p>Having the wrong frame/ontology for a problem.</p></li></ol></li></ol><h2><strong>A note on amortised and direct optimisation in LLMs</strong></h2><p>Current LLMs are products of amortised optimisation: they're trained with vast compute on massive datasets to create <a href="https://www.lesswrong.com/posts/DhjcdzTyqHte2v6bu/deep-learning-is-function-approximation">a distilled function-approximator</a>, are not known to do significant direct optimisation during inference, and do relatively little work during inference compared to training. GPT-4's training is <a href="https://patmcguinness.substack.com/p/gpt-4-details-revealed">rumoured</a> to have required on the order of 1025 FLOPs, compared to about 5&#215;1014 FLOPs to generate a token, so GPT-4 could generate 20 billion tokens with the compute it was trained on. (In comparison, humans could say only about 300k words in the working time it takes us to complete a 16-year education.)</p><p>So: are current LLMs the peak of wisdom over intelligence?</p><p>In the amortised v direct optimisation framework:</p><ol><li><p>LLMs are products of amortised optimisation, and most of their strength does lie in doing tasks that are more like amortised optimisation: distilling and synthesising information from their past "experience", and being extremely knowledgeable. LLMs already play the role of a wise guru to many: any programmer these days will use LLMs like they might have used an especially-available senior programmer who knows the best established practice for every scenario. It is true that they seem worse at having useful world models than humans are, and their wisdom runs less deep. More than that, perhaps, their ability to seem wise is dented by their blandness, a result both of their pre-training to mimic average internet text, and the fine-tuning done on them. They also seem quite bad at epistemics; as discussed in the first part, some kind of good epistemics is often (but not always) required to be able to make good use of the products of amortised optimisation, even if you have them at hand. Currently, that has to be largely provided by the human who's interacting with the LLM.</p></li><li><p>The extent to which LLMs do direct optimisation within a forward pass is unclear. However, like policy and value networks in MCTS, LLMs can be used as part of a direct optimisation process. Most trivially, you can prompt an LLM to search over some action space, and then evaluate the actions and pick one. LLM scaffolds go much further. It is true that the amortised optimisation core is still doing a lot of the work there, but that is true also of the amortised optimisation parts of MCTS variants that use neural networks, and probably of human thinking too. Therefore, LLMs can clearly be used for direct optimisation as well. In particular, we should expect LLMs to increase that part of the world's direct optimisation where the quality of the outputs is not hard to judge - lots of automating boilerplate writing and maths, less automating business decisions and philosophy.</p></li></ol><h2><strong>AI &amp; our mechanisms for amortised optimisation</strong></h2><p>We've seen that amortised optimisation powers a lot of the world, and in particular a lot of the things we regard as wise. Amortised optimisation needs a dataset. Humans currently maintain and improve their "dataset" through, for example, cultural evolution and written knowledge. We'll discuss how AI will affect each one for humans, and what the AI version of each one might look like.</p><h3><strong>The future of human cultural evolution</strong></h3><p>In the ancestral environment and still today, a key driver of cultural evolution is <a href="https://www.nature.com/articles/s41599-019-0228-7">prestige-biased social learning</a>: people imitating the habits of people who seem successful. Successful here can mean either directly successful, like a tendency to successfully get fresh meat or build companies, or just high in perceived status because other people pay attention to them (modern celebrities are a side product of this instinct). That humans can do this seems to be a big part of the human cognitive advantage compared to other apes. It's clear that this is a big boost to cultural evolution: unlike genes, memes don't have to rely on the differential survival of their hosts to spread, since nearby hosts will tend to copy the meme the instant the meme's host seems to be doing well.</p><p>None of this requires understanding why the memes work, or even which memes are driving someone's success. In Fiji, there are <a href="https://www.ncbi.nlm.nih.gov/pmc/articles/PMC2992713/">taboos</a> against pregnant women eating sharks. The reason this is useful is that the sharks contain chemicals that increase birth defect risks, but no one knew this. This is an insanely sophisticated and subtle meme: the causal connection between the sharks and the birth defects is long, weak, and totally beyond understanding before modern science.</p><p>There's a story that the famous mathematician Paul Erdos used to touch walls as he walked, and other mathematicians started imitating this behaviour - maybe it was the secret to Erdos's success, after all. Given our species' history, this isn't as crazy as it seems.</p><p>However, there are some prerequisites for blindly copying memes to go well. First, your brain's prestige-learning machinery probably only fires on other humans (though cue a wave of human-like AI avatars trying to seem like prestige superstimuli). Second, it helps a lot if you know that who you're copying has similar wants as you, and comes from an environment that was selecting for memes that drive towards the same wants as you. You're better off copying the behaviour of a human who wants partners, food, and power just like you, and comes from an environment where everyone was competing for those same things, than you are copying the behaviour of an alien who also wants power but will pass on the human food and partners - who knows what side effects their "memome" (their set of memes; c.f. "genome") will have. Crucially, what's required here is not <em>alignment</em> with you, but instead <em>taking actions that you'd also want to take if you were in their shoes</em> - if someone wants to steal your land, but is very effective at it, you might still want to copy their behaviours, since you sure want to steal their land too. But if you copy the farming practices of an alien, you end up growing alien zucchini that you can't even eat. Thirdly, it's a bad idea to try to copy the behaviour of someone who can take actions that you can't take; an experienced skateboarder can seem prestigious by doing a trick, but you would probably end up in the hospital instead.</p><p>All of this means that even if AI slots into the economic and even political system we have, AIs probably won't slot into our prestige-biased learning system (or if they do, it'll be bad). This means that AI probably won't help with the imitation of role models that is a big part of how human culture builds and transmits successful practices. It might also weaken this mechanism among humans. If, say, science is mostly automated and human performance in it is no longer a status marker, <a href="https://www.goodreads.com/book/show/35167685-surely-you-re-joking-mr-feynman?from_search=true&amp;from_srp=true&amp;qid=b6sCdIBwwd&amp;rank=1">Richard Feynman's curiosity and irreverence</a> will hold less sway over the next generation. If human success largely becomes about interfacing with LLMs, success might become more decoupled from positive human qualities, and the signal-to-noise ratio of cultural evolution may get worse.</p><h3><strong>The future of AI cultural evolution</strong></h3><p>In contrast, the AIs themselves might be very good at cultural evolution. While vastly more efficient than genetic evolution, human cultural evolution still requires one human to exhibit a behaviour and succeed, and this fact to be realised by others, and for others to then manage to copy that behaviour, which might take months of practise. Whether or not AI <em>learning</em> will be more or less sample-efficient than human learning is still unclear, so they may continue to require more examples to learn something than humans. However, AIs can be copied once they've learnt something. The best AI at a task can instantly become most of the entire global AI population working on that task (depending on the surrounding political, economic, and business environment, this may be by its own choice, the choice of another AI, or a human's choice). Consider how quickly LLM prompts or <a href="https://github.com/gpt-engineer-org/gpt-engineer">LLM scaffolds</a> spread, and imagine that sort of selection pressure, acting over long timescales over AIs that are individually increasingly capable. Much like human cultures develop memes that no individual could've come up with, the AIs might develop adaptations that they themselves could not have come up with, not through any direct optimisation that they're doing, or even through explicit training, but through external selection pressures and selective mimicry in the population of AIs. If humans took over the world by dint of cultural evolution despite individual human capabilities being static, imagine what ever-improving AIs might do.</p><p>This makes it important that we make good choices over how AI cultural evolution is allowed to happen. For example, we should be careful to structure things so that the AI types with large populations are doing good and useful things, and it's hard for persuasion-based or influence-seeking strategies on part of an AI to increase the number of copies of that AI being run. Also, we want to understand the routes by which one AI's behaviour might be copied by others. For example, this could happen through the behaviour being included in another AI's training data, or the behaviour appearing on the internet and being discovered by an AI agent doing an online search. We should also benchmark what cues increase the chance of an AI mimicking some behaviour. For example, it's <a href="https://arxiv.org/abs/2310.15047">known</a> that language models preferentially learn facts from more consistent sources.</p><h3><strong>Written knowledge</strong></h3><p>One potential harm of AI is reducing some valuable types of human-written content. For example <a href="https://daily.dev/blog/did-stack-overflow-just-admit-defeat-to-ai">visits to the programming-Q&amp;A website Stack Overflow declined</a> after LLMs got good at helping programmers. This is good, though: LLMs are better than Stack Overflow, and if LLMs can't solve a problem, people can still go to Stack Overflow and post a question. Then the next generation of LLMs gets trained on it.</p><p>A more serious harm is that LLM text might be skewed to be more like median human text than human text is, as argued <a href="https://arxiv.org/abs/2404.03502">here</a>. This might reduce richness and variety in areas where LLM text replaces human text - i.e., eventually probably everything. In the same way that industrialisation made many physical products cheaper and better, but also more uniform and less personalised, AI might do the same for products of mental work. This will likely be worse than the loss of personalisation with physical goods, since diversity and variety is more of the value of mental work, and therefore people will pay a steeper premium to maintain it than they would in physical products. Less variety would also mean a smaller population of memes for cultural evolution to select from. It also seems to be harder for LLMs to learn from LLM text - see for example <a href="https://arxiv.org/abs/2305.17493">this paper</a>, or <a href="https://arxiv.org/abs/2402.07043">this exploration</a> of how scaling laws change when there's LLM data in the mix. However, this type of "model collapse" shouldn't be exaggerated - it seems that dramatic slowdowns on LLM progress are <a href="https://x.com/RylanSchaeffer/status/1816881533795422404">not likely from it</a>.</p><p>A big problem with current human written knowledge is that it often can't be used effectively. There isn't enough time to read everything, or find every useful paper on a topic. Lots of knowledge, rather than being part of amortised optimisation processes, sits ignored. LLMs could help. Already, LLMs are doing a good job as forecasters simply by having the patience to read every news article, or at <a href="https://elicit.com/">finding papers on a topic</a>, or <a href="https://www.perplexity.ai/">getting close to beating Google</a>. The ultimate search method is talking to a wise guru who knows everything; LLMs are tending towards this.</p><p>LLMs could also help reduce <a href="https://distill.pub/2017/research-debt/">"research debt"</a>, the mountain of poor explanations, undigested thoughts, and noise that researchers have to process to get to the frontier. They're already decent at answering simple questions about a paper. In the future, they might distil a set of papers into a <a href="https://colah.github.io/posts/2015-08-Understanding-LSTMs/">Chris Olah -level explanation</a> that could be used by humans or AIs. This would be a very good situation for humans - think of how fun good explanations are to read - but will human mental labour still be relevant once this is possible? Producing a good explanation sounds close to an AGI-complete problem, but if AI disproportionately improves at amortised optimisation without getting much better at direct optimisation, a world where AIs can explain existing things well but humans are still needed for novel discoveries may exist for a while.</p><p>Alternatively, the massive context lengths of some recent LLMs could reduce the need for distillation at all (e.g. <a href="https://docs.anthropic.com/en/docs/build-with-claude/prompt-engineering/long-context-tips">Claude 3 can read a long-ish novel in a single context</a>). The entire requirement for explanations and distillations and even papers could disappear, replaced with LLMs taking in massive contexts full of experimental results and historical trends, and outputting answers based on that. This would be especially useful for <a href="https://en.wikipedia.org/wiki/Literature-based_discovery">literature-based discoveries</a>. However, the principle generalises: want good relationship advice? Just dump your entire texting history into a massive LLM context and have it in mere seconds spit out advice that no human could've figured out without hours of reading and digesting. Humans would then likely become just consumers of wisdom, rather than producers of it: all the experience- or history-based insights you might eventually have would be told to you by AIs faster than you can have them.</p><p>AIs may also do away with written knowledge, by instead passing direct vector representations of concepts among each other. It's already known that the vector representations from of separate training runs and even different architectures can often be passed between each other with little modification, suggesting that translation issues wouldn't be a blocker. There are some reasons why language-like discrete representations may be quite fundamental (note that all of language, code, maths and music are written in discrete symbols). However, if these reasons aren't strong enough, we might end up with most information about the world existing only in vector representations, except when humans specifically ask an AI for an explanation.</p><h2><strong>Avoiding strategy errors</strong></h2><p>A supposed benefit of being wise over just smart is avoiding large-scale errors, where individual actions are clever and make sense but the end result is silly. As the saying goes: "it doesn't matter how sharp your axe is if you're cutting the wrong tree". Making good high-level decisions is generally called strategy, so I'll use that term.</p><p>I'll discuss two types of strategy errors, connect them to the frame amortised versus direct optimisation, and suggest AI effects.</p><p>But first, why is getting large things right something that relies on amortised optimisation? You can run a search process that operates on big things and answers big questions (for example, <a href="https://www.noahpinion.blog/p/why-i-think-an-invasion-of-taiwan">here's an example</a> of someone explicitly drawing search trees over large-scale US-China war actions). Many of the best uses of machine learning, the ultimate modern amortised optimisation technique, are for doing small steps (like next-token prediction) well. So why is getting large things right on the side of amortised optimisation?</p><p>Any direct optimisation process means running a search. If you're trying to plan something big, that's going to take multiple steps in a complex world. If you're running a search over multi-step plans, the number of possibilities to search over quickly blows up. There are two choices:</p><ul><li><p>Use heuristics to prune the search tree. How do we get those heuristics? Amortised optimisation. For example, AlphaGo's high-level architecture is good old Monte Carlo Tree Search (MCTS). But the search through the tree is guided by a value network (that estimates the probability of winning from a position, to limit the requirement to search over subsequent moves to figure out how good a position is) and policy network (that guides decisions over which parts of the search tree to explore first, to reduce the total amount of search required). Both the value and policy networks are implemented as deep neural networks that are trained on on lots of data. Amortised optimisation saves the day.</p></li><li><p>Do the search on a higher level of abstraction. Governments planning their nation's grand strategy do not push around individual division on maps, they instead chunk things together until they're thinking about allies and fronts and economies. To do good planning on the more abstract, chunked level requires a good model of how those chunks act. There seem to be two ways to get this:</p><ul><li><p>First, you can chunk history into chunks of the same size as you're thinking about, and look at the patterns: when a nation that looks like this fought a nation that looked like that, the results tended to be this, and success correlated with how well they did X. But this requires a lot of history to learn patterns from - in other words, you're making use of amortised optimisation.</p></li><li><p>Second, you can be good at modelling things. If you have a good-enough model of people, economics, game theory, and war tactics, you can probably derive many of the patterns for what large-scale moves will be good, even without access to a lot of history about which nations win wars with which strategies. Doing this well does requires searching over alternatives and thinking on-the-fly - that is, direct optimisation. I'd also guess there's an some kind of important "model-building skill" involved. Parts of this is probably amortised optimisation to figure out what model types have worked well in the past. Another part is amortised optimisation over how those smaller primitives work (unless you're doing maths or extrapolating known fundamental physics, you always need some histories to fit your model to). I'd claim that it's hard to be good at modelling things without relying a lot on amortised optimisation, but I admit some confusion over how "modelling skill" fits into this framework, or where it comes from more generally.</p></li></ul></li></ul><h3><strong>Mistakes in strategic planning</strong></h3><p>Doing high-level strategic planning well requires being able to run a search over high-level plans well. We discussed two ways to achieve this:</p><ol><li><p>Using amortised optimisation to learn the associations between high-level actions and their outcomes.</p></li><li><p>Having a good enough model of the lower-level details to be able to extrapolate how the high-level plans would go. This likely requires a lot of existing amortised optimisation about those lower-level details. It also likely requires something like simulation, which in turn requires having good models.</p></li></ol><p>The first recommendation to be better at strategic planning would then be to know the relevant histories, so you can do the amortised optimisation thing of applying past lessons. However, if we're worrying about something like strategy for dealing with a big new thing transformative AI, this approach is limited because there aren't many good historical analogues.</p><p>Therefore, making wiser choices about AI likely relies in large part on having good enough models of things that are not AI strategy that you can extrapolate the consequences of different AI strategies. This can be done top-down (find empirical principles even more general, where AI is a special case) or bottom-up (simulate lower-level principles to try to figure out how the bigger picture of AI works).</p><p>A fair amount of discussion on <a href="https://www.lesswrong.com/">LessWrong</a> is about very general patterns of how the world works. To take one example, John Wentworth has written on <a href="https://www.lesswrong.com/s/ogntdnjG6Y9tbLsNS/p/6mysMAqvo9giHC4iX">agency and general-purpose search</a>, and <a href="https://www.lesswrong.com/s/TLSzP4xP42PPBctgw/p/tmuFmHuyb4eWmPXz8">the difficulties of delegation</a>. This can be interpreted as trying to distil the way the world works to fundamental blocks general enough that everything from AI to <a href="https://www.lesswrong.com/posts/MMAK6eeMCH3JGuqeZ/everything-i-need-to-know-about-takeoff-speeds-i-learned">Amazon reviews for air conditioners</a> falls out as a special case ~~and also as everyone involved getting <a href="https://xkcd.com/356/">nerdsniped</a> - says I, while 9000 words into an overly-abstract essay series on the automation of wisdom~~.</p><p>This is the top-down approach. A lot of it is fundamentally about distilling a large body of data into some general approximate model that can then be queried cheaply - amortised optimisation. However, as with any theory-building process, there's also a lot of direct optimisation involved (e.g. searching over a space of ideas). I'd guess this sort of work is close to AGI-complete, and I'm uncertain what it's bottlenecked on.</p><p>On the other hand, there's the bottom-up approach. For example, Eliezer Yudkowsky's focus on <a href="https://www.lesswrong.com/posts/RQpNHSiWaXTvDxt6R/coherent-decisions-imply-consistent-utilities">coherence theorems</a> seems to be due to a model where non-coherent agents predictably lose or self-modify, and therefore eventually we're dealing with coherent (i.e. Bayesian expected-utility maximising goal-driven) AI agents. This is a high-level (and <a href="https://www.intelligencerising.org/">contested</a>) prediction based on extrapolating the behaviour of more basic primitives forward (and where the argument mostly does not make reference to prior histories). <a href="https://www.alignmentforum.org/posts/pRkFkzwKZ2zfa3R6H/without-specific-countermeasures-the-easiest-path-to">This post</a> by Ajeya Cotra, or <a href="https://situational-awareness.ai/">this post</a> by Leopold Aschenbrenner, present very simulation-style arguments that take a model of how AIs and the AI race works and extrapolate it forward.</p><p>The full version of bottom-up strategy work also feels hard for foreseeable AIs to fully automate. There's a significant direct optimisation bottleneck, especially if trying to predict the behaviour of actors with large action spaces that themselves have a lot of direct optimisation at hand. However, there seems to be a clear path for AI to help. Even current LLMs could do a decent job of extrapolating the consequences of simple maths, or at role-playing the decision-making of given actors. Scenario-planning exercises and simulations, from Pentagon war games to <a href="https://www.intelligencerising.org/">Intelligence Rising</a>, are useful for decision-makers and can reveal surprising options like <a href="https://www.defenseone.com/policy/2020/02/esper-plays-nuclear-war-russia-nukes-europe-us-fires-back/163268/">nuking Belarus</a>. AI can make these cheaper, by reducing the human mental labour needed to run good ones. This could help explore and evaluate possible strategies when we don't have much history to do amortised optimisation on.</p><h3><strong>Framing mistakes</strong></h3><p>In addition to being bad at strategy, another error of insufficient wisdom is having the wrong frame / ontology / paradigm. This is when you notice the Earth isn't the centre of the universe, or that a watch <a href="https://en.wikipedia.org/wiki/Darwin%27s_Dangerous_Idea#Skyhooks_and_cranes">doesn't</a> imply a watchmaker. It's when you go to your wise mountain guru to ask how to finally find that treasure, and they tell you the real treasure is the friends you made along the way.</p><p>If there were a simple explanation for how paradigm shifts are found, many famous philosophers of science would be a lot less famous. However, a paradigm shift takes at least two things: someone has to discover the new framing, and many someones have to care.</p><p>Current LLMs are bad at anything like finding paradigm shifts. Consider how many people use LLMs, and the lack of any breakthrough, even of the size of a decent research paper, where LLMs were claimed to be the main source of the concept. And paradigm shifts are much harder than research papers.</p><p>Inventing important moral principles, like utilitarianism or the categorical imperative, seems even harder. Bentham's and Kant's search for moral principles was presumably guided by trying to make precise their intuitive human ethics. There's some amount of data with which they could've succeeded without having those built-in moral intuitions, but it seems very helpful to have had those intuitions in their head as something they could easily query.</p><p>It's in having people care that AI might have a larger effect, and maybe a bad one. First, one driver of paradigm shifts is that someone gets annoyed by complexity or ugliness, and wants to fix it. By making mental labour cheaper, AI might reduce the ickiness of bad models. Consider how geocentrism may have lasted longer if it were trivial to ask your Jupyter notebook copilot AI to add a few more epicycles to your model, and it didn't mean more laborious longhand arithmetic for you. Second, increasing use of AI might mean that less and less of the paradigm-incompatible data is seen by human eyes. Imagine if the AI adds the epicycles in the background to improve the model, without any human ever noticing. Potential fixes might be keeping simplicity and elegance as key cultural values in scientific fields, and somehow making propagating this to the AIs (while it has <a href="https://www.astralcodexten.com/p/contra-the-xai-alignment-plan">many problems</a>, xAI's "curious AIs" plan has some of this spirit).</p><h2><strong>Conclusion</strong></h2><p>The description of wisdom above may feel reductive: is a lot of it really just applying past data and the results of past computation to current problems? Is a lot of the task of improving our civilisation's wisdom, whether through AI or human actions, just the task of storing and effectively using written knowledge, letting selection processes like cultural evolution build up impressive results, and training opaque (natural or artificial) neural networks to compress insights from data?</p><p>A good explanation should feel somewhat reductive, though. Taking this perspective, the picture that emerges is one where a lot of the wisdom needed to take wise actions emerges almost automatically. Wisdom's signature is less brilliant mental moves, and more what's left standing once time and chance have taken their toll, or a training process finished compressing data in a brain or transformer. The most worrying thing, then, is systemic advantages that AIs likely have that might lead to them taking a dominant role in the use and production of wisdom. For example, human success is based on cultural evolution, but AIs might be better than humans at it, and we should take care to direct AI cultural evolution in a good direction.</p><p>AIs are likely to be helpful in many ways, though, for example by helping distil existing bodies of work, helping simulating scenarios as part of strategy planning, and generally becoming a wise guru that knows everything that everyone has access to all the time. However, it's still unclear how when and how they might help with other parts of wisdom, like avoiding ontology errors.</p><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.nosetgauge.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading No Set Gauge! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[AI & wisdom 2: growth and amortised optimisation]]></title><description><![CDATA[Winning entry in the AI Impacts essay competition on the automation of wisdom and philosophy.]]></description><link>https://www.nosetgauge.com/p/ai-and-wisdom-2-growth-and-amortised</link><guid isPermaLink="false">https://www.nosetgauge.com/p/ai-and-wisdom-2-growth-and-amortised</guid><dc:creator><![CDATA[Rudolf Laine]]></dc:creator><pubDate>Mon, 04 Nov 2024 02:09:36 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Gj6_!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F007576cc-e90d-40a4-9ae5-94ced4025a77_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Gj6_!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F007576cc-e90d-40a4-9ae5-94ced4025a77_1024x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Gj6_!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F007576cc-e90d-40a4-9ae5-94ced4025a77_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!Gj6_!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F007576cc-e90d-40a4-9ae5-94ced4025a77_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!Gj6_!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F007576cc-e90d-40a4-9ae5-94ced4025a77_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!Gj6_!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F007576cc-e90d-40a4-9ae5-94ced4025a77_1024x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Gj6_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F007576cc-e90d-40a4-9ae5-94ced4025a77_1024x1024.png" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/007576cc-e90d-40a4-9ae5-94ced4025a77_1024x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2871195,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Gj6_!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F007576cc-e90d-40a4-9ae5-94ced4025a77_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!Gj6_!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F007576cc-e90d-40a4-9ae5-94ced4025a77_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!Gj6_!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F007576cc-e90d-40a4-9ae5-94ced4025a77_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!Gj6_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F007576cc-e90d-40a4-9ae5-94ced4025a77_1024x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><em>Winning entry in the <a href="https://www.lesswrong.com/posts/hiuTzNBqG2EYg6qM5/winners-of-the-essay-competition-on-the-automation-of-wisdom">AI Impacts essay competition on the automation of wisdom and philosophy</a>.</em></p><p><em>Please go <a href="https://rudolf.website/wisdom2/">here</a> to read this post, as Substack does not support inline LaTeX.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.nosetgauge.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading No Set Gauge! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[AI & wisdom 1: wisdom, amortised optimisation, and AI]]></title><description><![CDATA[Winning entry in the AI Impacts essay competition on the automation of wisdom and philosophy.]]></description><link>https://www.nosetgauge.com/p/ai-and-wisdom-1-wisdom-amortised</link><guid isPermaLink="false">https://www.nosetgauge.com/p/ai-and-wisdom-1-wisdom-amortised</guid><dc:creator><![CDATA[Rudolf Laine]]></dc:creator><pubDate>Mon, 04 Nov 2024 02:07:43 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!n6dQ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff40604f2-fa2d-4598-9891-c8e45184f469_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!n6dQ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff40604f2-fa2d-4598-9891-c8e45184f469_1024x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!n6dQ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff40604f2-fa2d-4598-9891-c8e45184f469_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!n6dQ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff40604f2-fa2d-4598-9891-c8e45184f469_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!n6dQ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff40604f2-fa2d-4598-9891-c8e45184f469_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!n6dQ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff40604f2-fa2d-4598-9891-c8e45184f469_1024x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!n6dQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff40604f2-fa2d-4598-9891-c8e45184f469_1024x1024.png" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f40604f2-fa2d-4598-9891-c8e45184f469_1024x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2800638,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!n6dQ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff40604f2-fa2d-4598-9891-c8e45184f469_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!n6dQ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff40604f2-fa2d-4598-9891-c8e45184f469_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!n6dQ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff40604f2-fa2d-4598-9891-c8e45184f469_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!n6dQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff40604f2-fa2d-4598-9891-c8e45184f469_1024x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><em>Winning entry in the <a href="https://www.lesswrong.com/posts/hiuTzNBqG2EYg6qM5/winners-of-the-essay-competition-on-the-automation-of-wisdom">AI Impacts essay competition on the automation of wisdom and philosophy</a>.</em></p><p><em>Please go <a href="https://rudolf.website/wisdom1/">here</a> to read this post, as Substack does not allow inline LaTeX.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.nosetgauge.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading No Set Gauge! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Investigating an insurance-for-AI startup]]></title><description><![CDATA[Please go here to read this post, because Substack does not allow inline LaTeX.]]></description><link>https://www.nosetgauge.com/p/investigating-insurance-for-ai-startup</link><guid isPermaLink="false">https://www.nosetgauge.com/p/investigating-insurance-for-ai-startup</guid><dc:creator><![CDATA[Rudolf Laine]]></dc:creator><pubDate>Sat, 21 Sep 2024 15:16:00 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/9c0da449-8d4d-4cca-8043-4e439527d566_1234x947.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Please go <a href="https://rudolf.website/ai-insurance/">here</a> to read this post, because Substack does not allow inline LaTeX.</em></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!vpTy!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f5d7673-9fe1-4109-bf8b-62200345759f_1234x947.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!vpTy!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f5d7673-9fe1-4109-bf8b-62200345759f_1234x947.png 424w, https://substackcdn.com/image/fetch/$s_!vpTy!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f5d7673-9fe1-4109-bf8b-62200345759f_1234x947.png 848w, https://substackcdn.com/image/fetch/$s_!vpTy!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f5d7673-9fe1-4109-bf8b-62200345759f_1234x947.png 1272w, https://substackcdn.com/image/fetch/$s_!vpTy!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f5d7673-9fe1-4109-bf8b-62200345759f_1234x947.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!vpTy!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f5d7673-9fe1-4109-bf8b-62200345759f_1234x947.png" width="625" height="480" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6f5d7673-9fe1-4109-bf8b-62200345759f_1234x947.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:480,&quot;width&quot;:625,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!vpTy!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f5d7673-9fe1-4109-bf8b-62200345759f_1234x947.png 424w, https://substackcdn.com/image/fetch/$s_!vpTy!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f5d7673-9fe1-4109-bf8b-62200345759f_1234x947.png 848w, https://substackcdn.com/image/fetch/$s_!vpTy!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f5d7673-9fe1-4109-bf8b-62200345759f_1234x947.png 1272w, https://substackcdn.com/image/fetch/$s_!vpTy!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f5d7673-9fe1-4109-bf8b-62200345759f_1234x947.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div>]]></content:encoded></item><item><title><![CDATA[Positive visions for AI ]]></title><description><![CDATA[This post was a collaboration with Florence Hinder]]></description><link>https://www.nosetgauge.com/p/positive-visions-for-ai</link><guid isPermaLink="false">https://www.nosetgauge.com/p/positive-visions-for-ai</guid><dc:creator><![CDATA[Rudolf Laine]]></dc:creator><pubDate>Tue, 23 Jul 2024 19:08:00 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/dc2516ae-218e-45f8-81e0-7536867e1411_640x366.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>This post was a collaboration with <a href="https://www.florencehinder.com/positive-visions-for-ai/">Florence Hinder</a></p><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!dJEJ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd629f443-a33d-4ae1-b372-8d0d757d3355_640x366.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!dJEJ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd629f443-a33d-4ae1-b372-8d0d757d3355_640x366.jpeg 424w, https://substackcdn.com/image/fetch/$s_!dJEJ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd629f443-a33d-4ae1-b372-8d0d757d3355_640x366.jpeg 848w, https://substackcdn.com/image/fetch/$s_!dJEJ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd629f443-a33d-4ae1-b372-8d0d757d3355_640x366.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!dJEJ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd629f443-a33d-4ae1-b372-8d0d757d3355_640x366.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!dJEJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd629f443-a33d-4ae1-b372-8d0d757d3355_640x366.jpeg" width="640" height="366" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d629f443-a33d-4ae1-b372-8d0d757d3355_640x366.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:366,&quot;width&quot;:640,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!dJEJ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd629f443-a33d-4ae1-b372-8d0d757d3355_640x366.jpeg 424w, https://substackcdn.com/image/fetch/$s_!dJEJ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd629f443-a33d-4ae1-b372-8d0d757d3355_640x366.jpeg 848w, https://substackcdn.com/image/fetch/$s_!dJEJ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd629f443-a33d-4ae1-b372-8d0d757d3355_640x366.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!dJEJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd629f443-a33d-4ae1-b372-8d0d757d3355_640x366.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><h2>Reasons to make the positive case</h2><p>Everyone who starts thinking about AI starts thinking big. Alan Turing predicted that machine intelligence would make humanity appear feeble in comparison. I. J. Good said that AI is the last invention that humanity ever needs to invent.</p><p>The AI safety movement started from Eliezer Yudkowsky and others on <a href="http://sl4.org/shocklevels.html">the SL4 mailing list</a> discussing (and aiming for) an intelligence explosion and colonizing the universe. However, as the promise of AI has drawn nearer, visions for AI upsides have paradoxically shrunk. Within the field of AI safety, this is due to a combination of the &#8220;doomers&#8221; believing in very high existential risk and therefore focusing on trying to avoid imminent human extinction rather than achieving the upside, people working on policy not talking about sci-fi upsides to look less weird, and recent progress in AI driving the focus towards concrete machine learning research rather than aspirational visions of the future.</p><p>Both DeepMind and OpenAI were explicitly founded as moonshot AGI projects (&#8220;solve intelligence, and then use that to solve everything else&#8221; in the words of Demis Hassabis). Now DeepMind - sorry, Google DeepMind - has been eaten by the corporate machinery of Alphabet, and OpenAI is increasingly captured by profit and product considerations.</p><p>The torch of AI techno-optimism has moved on the e/acc movement. Their core message is correct: growth, innovation, and energy are very important, and almost no one puts enough emphasis on them. However, their claims to take radical futures seriously are belied by the fact that their visions of the future seem to stop at GenAI unicorns. They also seem to take the general usefulness of innovation not as just a robust trend, but as a law of nature, and so are remarkably incurious about the possibility of <a href="https://nickbostrom.com/papers/vulnerable.pdf">important</a> exceptions. Their deeper ideology is in parts incoherent and inhuman. Instead of centering human well-being, they worship the &#8220;<a href="https://beff.substack.com/p/notes-on-eacc-principles-and-tenets">thermodynamic will of the universe</a>&#8221;. &#8220;You cannot stop the acceleration&#8221;, argues their figurehead, so &#8220;[y]ou might as well embrace it&#8221; - hardly an inspiring humanist rallying cry.</p><p>In this piece, we want to paint a picture of the possible benefits of AI, without ignoring the risks or shying away from radical visions. Why not dream about the future you hope for? It&#8217;s important to consider the future you want rather than just the future you don&#8217;t. Otherwise, you might create your own unfortunate destiny. In the Greek myth about Oedipus, he was prophesied to kill his father, so his father ordered him to be killed, but he wasn&#8217;t and ended up being adopted. Years later he crossed his father on the road in his travels and killed him, as he had no idea who his father was. Oedipus&#8217; father focusing on the bad path might have made the prophecy happen. If Oedipus' father hadn&#8217;t ordered him to be killed, he would have known who his father was and likely wouldn&#8217;t have killed him.&nbsp;</p><p>When thinking about AI, if we only focus on the catastrophic future, we may cause it to become true by causing an increase in attention on this topic. Sam Altman, who is leading the way in AI capabilities, claimed to have gotten interested from arch-doomer Eliezer Yudkowsky. We may also neglect progress towards positive AI developments; some people think that even direct AI alignment research should not be published because it might speed up the creation of unaligned AI.</p><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!BP82!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c8361b1-0954-48b1-b27a-0514e6e6a6cd_358x400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!BP82!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c8361b1-0954-48b1-b27a-0514e6e6a6cd_358x400.png 424w, https://substackcdn.com/image/fetch/$s_!BP82!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c8361b1-0954-48b1-b27a-0514e6e6a6cd_358x400.png 848w, https://substackcdn.com/image/fetch/$s_!BP82!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c8361b1-0954-48b1-b27a-0514e6e6a6cd_358x400.png 1272w, https://substackcdn.com/image/fetch/$s_!BP82!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c8361b1-0954-48b1-b27a-0514e6e6a6cd_358x400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!BP82!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c8361b1-0954-48b1-b27a-0514e6e6a6cd_358x400.png" width="358" height="400" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8c8361b1-0954-48b1-b27a-0514e6e6a6cd_358x400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:400,&quot;width&quot;:358,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!BP82!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c8361b1-0954-48b1-b27a-0514e6e6a6cd_358x400.png 424w, https://substackcdn.com/image/fetch/$s_!BP82!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c8361b1-0954-48b1-b27a-0514e6e6a6cd_358x400.png 848w, https://substackcdn.com/image/fetch/$s_!BP82!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c8361b1-0954-48b1-b27a-0514e6e6a6cd_358x400.png 1272w, https://substackcdn.com/image/fetch/$s_!BP82!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c8361b1-0954-48b1-b27a-0514e6e6a6cd_358x400.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><p>With modern AI, we might even get a very direct &#8220;self-fulfilling prophecy&#8221; effect: current AIs increasingly know that they are AIs, and make predictions about how to act based on their training data which includes everything we write about AI.</p><h2>Benefits of AI</h2><p>Since we think a large focus of AI is on what could go wrong, let&#8217;s think through what could go well starting from what&#8217;s most tangible and close to the current usage of AI to what the more distant future could hold.</p><ol><li><p>AI will do the mundane work</p></li><li><p>Lowering the costs of coordination</p></li><li><p>Spreading Intelligence</p></li><li><p>AI can create more technology</p></li><li><p>Increased technology, wealth and energy, correlate with life being good</p></li><li><p>All of the above, and the wealth it creates, could allow people to self-actualise more</p></li></ol><p>Already, AI advances mean that <a href="https://claude.ai/">Claude</a> has beocme very useful, and programmers are <a href="https://economics.mit.edu/sites/default/files/inline-files/Noy_Zhang_1_0.pdf">faster and better</a>. But below we&#8217;ll cast a look towards the bigger picture and where this could take us.</p><h3>AI will do the mundane work</h3><p>First, there&#8217;s a lot of mundane mental work that humans currently have to do. Dealing with admin work, filing taxes, coordinating parcel returns -- these are not the things you will fondly be reminiscing about as you lie on your deathbed. Software has reduced the pain of dealing with such things, but not perfectly. In the future, you should be able to deal with all administrative work by specifying what you want to get done to an AI, and being consulted on decision points or any ambiguities in your preferences. Many CEOs or executives have personal assistants; AIs will mean that everyone will have access to this.&nbsp;</p><p>What about mundane physical work, like washing the dishes and cleaning the toilets? Currently, robotics is bad. But there is no known fundamental obstacle to having good robotics. It seems mainly downstream of a lot of engineering and a lot of data collection. AI can help with both of those. The household robots that we&#8217;ve been waiting for could finally become a reality.</p><p>Of course, it is unclear whether AIs will first have a comparative advantage against humans in mundane or meaningful work. We&#8217;re already seeing that AI models are making massive strides in making art, way before they&#8217;re managing our inboxes for us. It may be that there is a transitional period where robotics is lagging but AIs are smarter-than-human, where the main economic value of humans is their hands rather than their brains.</p><h3>Lowering the cost of coordination</h3><p>With AI agents being able to negotiate with other AI agents, the cost of coordination is likely to dramatically drop (see <a href="https://www.alignmentforum.org/posts/gYaKZeBbSL4y2RLP3/strategic-implications-of-ais-ability-to-coordinate-at-low">here</a> for related discussion). Examples of coordination are agreements between multiple parties, or searching through a large pool of people to match buyers or sellers, or employees and employers. Searching through large sets of people, doing complex negotiations, and the monitoring and enforcement of agreements all take lots of human time. AI could reduce the cost and time taken by such work. In addition to efficiency gains, new opportunities for coordination will open up that would have previously been too expensive.</p><h4>Small-scale coordination</h4><p>To give an example of this on the small scale of two individuals, say you are trying to search for a new job. Normally you can&#8217;t review every single job posting ever, and employers can&#8217;t review every person in the world to see if they want to reach out. However, an AI could filter that for the individual and another AI for the business, and the two AIs could have detailed negotiations with each other to find the best possible match.&nbsp;</p><h4><a href="https://www.lesswrong.com/posts/P6fSj3t4oApQQTB7E/coordination-as-a-scarce-resource">Coordination as a scarce resource</a></h4><p>A lot of the current economy is a coordination platform; that&#8217;s the main product of each of Google, Uber, Amazon, and Facebook. Reducing the cost of searching for matches and trades should unlock at least as much mundane benefits and economic value as the tech platforms have.</p><p>Increased coordination may also reduce the need to group people into roles, hierarchies, and stereotypes. Right now, we need to put people into rigid structures (e.g. large organisations with departments like &#8220;HR&#8221; or &#8220;R&amp;D&#8221;, or specific roles like &#8220;doctor&#8221; or &#8220;developer&#8221;) when coordinating a large group of people. In addition to upholding standards and enabling specialisation of labour, another reason for this is that people need to be legible to unintelligent processes, like binning of applicants by profession, or the CEO using an org chart to find out who to ask about a problem, or someone trying to buy some type of service. Humans can reach a much higher level of nuance when dealing with their friends and immediate colleagues. The cheap intelligence we get from AI might let us deal with the same level of nuance with a larger group of people than humans can themselves <a href="https://en.wikipedia.org/wiki/Dunbar%27s_number">track</a>. This means people may be able to be more unique and differentiated, while still being able to interface with society.</p><h4>Large-scale Coordination</h4><p>On a larger scale, increased coordination will also impact geopolitics. Say there are two countries fighting over land or resources. Both countries could have AI agents to negotiate with the other AI agents to search the space of possible deals and find an optimal compromise for both. They could also simulate a vast number of war scenarios to figure out what would happen; much conflict is about two sides disagreeing about who would win and resolving the uncertainty through a real-world test. This relies on three key abilities: the ability to negotiate cheaply, the ability to simulate outcomes, and the ability to stick to and enforce contracts. AI is likely to help with all three. This could reduce the incentives for traditional war, in that no human lives are needed to be lost because the outcome is already known and we can negotiate straight from that. We also know exactly what we are and are not willing to trade off which means it&#8217;s easier to optimise for the best compromise for everyone.</p><h3>Spreading the intelligence</h3><p>AI lets us spread the benefits of being smart more widely.</p><p>The benefits of intelligence are large. For example, <a href="https://www.tandfonline.com/doi/abs/10.1080/15579336.2002.11770253">this study</a> estimates that a 1 standard deviation increase in intelligence increases your odds of self-assessed happiness by 11%. Now, part of this gain comes from intelligence being a positional good: you benefit from having more intelligence at your disposal than others, for example in competing for a fixed set of places. However, intelligence also has absolute benefits, since it lets you make better choices. And AI means you can convert energy into intelligence. Much as physical machines let the weak gain some of the benefits of (even superhuman) strength, AI might allow all humans to enjoy some of the benefits of being smart.</p><p>Concretely, this could have two forms. The first is that you could have AI advisors increase your ability to make plans or decisions, in the same way that - hypothetically - even a near-senile president might still make decent decisions with the help of their smart advisors. With AI, everyone could have access to comparable expert advisors. The effect may be even more dramatic than human advisors: the AI might be superhumanly smart, the AI might be more verifiably smart (a big problem in selecting smart advisors is that it can be hard to tell who is actually smart, especially if you are not), and if AIs are aligned successfully there may be less to worry about in trusting it than in trusting potentially-scheming human advisors.&nbsp;</p><p>The second is AI tutoring. Human 1-1 tutoring boosts educational outcomes by <a href="https://en.wikipedia.org/wiki/Bloom%27s_2_sigma_problem">2 standard deviations</a> (2 standard deviations above average is often considered the cutoff for &#8220;giftedness&#8221;). If AI tutoring is as good, that&#8217;s a big deal.</p><h3>AI is the ultimate meta-technology</h3><p>AI is special because it automates intelligence, and intelligence is what you need to build technology, including AI, creating a feedback loop. Some other previous technologies have boosted other technologies; for example, the printing press massively helped the accumulation of knowledge that led to the invention of many other technologies. But we have not before had a technology that could itself directly advance other technology. Such AI has been called <a href="https://www.cold-takes.com/transformative-ai-timelines-part-1-of-4-what-kind-of-ai/">PASTA (Process for Automating Scientific and Technological Advancement</a>).</p><p>Positive feedback loops - whether self-improving AIs, nuclear reactions, epidemics, or human cultural evolution - are very powerful, so you should be wary of risks from them. Similarly, it is currently at best extremely unclear whether AIs that improve themselves could be controlled with current technology. We should be very cautious in using AI systems to improve themselves.</p><p>In the long run, however, most of the value of AI will likely come from their effects on technological progress, much like the next industrial revolution. We can imagine AIs slashing the cost and increasing the speed of science in every field, curing diseases and making entire new veins of technology available, in the same way that steam engines made entirely new veins of coal accessible.</p><p>In particular, AIs help de-risk one of the largest current risks to future human progress. <a href="https://slatestarcodex.com/2019/04/22/1960-the-year-the-singularity-was-cancelled/">One model of the feedback loop behind humanity&#8217;s progress</a> in the past few centuries is that people led to ideas led to wealth led to food led to more people.</p><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ISnM!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70fcd3d8-0a3a-4c33-80c5-29ec0b7f4428_400x350.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ISnM!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70fcd3d8-0a3a-4c33-80c5-29ec0b7f4428_400x350.png 424w, https://substackcdn.com/image/fetch/$s_!ISnM!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70fcd3d8-0a3a-4c33-80c5-29ec0b7f4428_400x350.png 848w, https://substackcdn.com/image/fetch/$s_!ISnM!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70fcd3d8-0a3a-4c33-80c5-29ec0b7f4428_400x350.png 1272w, https://substackcdn.com/image/fetch/$s_!ISnM!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70fcd3d8-0a3a-4c33-80c5-29ec0b7f4428_400x350.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ISnM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70fcd3d8-0a3a-4c33-80c5-29ec0b7f4428_400x350.png" width="400" height="350" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/70fcd3d8-0a3a-4c33-80c5-29ec0b7f4428_400x350.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:350,&quot;width&quot;:400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ISnM!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70fcd3d8-0a3a-4c33-80c5-29ec0b7f4428_400x350.png 424w, https://substackcdn.com/image/fetch/$s_!ISnM!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70fcd3d8-0a3a-4c33-80c5-29ec0b7f4428_400x350.png 848w, https://substackcdn.com/image/fetch/$s_!ISnM!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70fcd3d8-0a3a-4c33-80c5-29ec0b7f4428_400x350.png 1272w, https://substackcdn.com/image/fetch/$s_!ISnM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70fcd3d8-0a3a-4c33-80c5-29ec0b7f4428_400x350.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><p>However, greater wealth no longer translates into more people. The world population, which was exponentially growing for much of the 19th and 20th centuries, is likely to be in decline by the end of the 21st century. This is likely to have <a href="https://www.overcomingbias.com/p/shrinking-economies-dont-innovate">negative consequences</a> for the rate of innovation, and as discussed in the next section, a decline in productivity would likely have a negative impact on human wellbeing. However, if AIs start driving innovation, then we have a new feedback loop: wealth leads to energy leads to more AIs leads to ideas leads to wealth.</p><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!2Fe6!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c1a761b-4bba-4571-84a6-e0e2c0e5b149_640x411.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!2Fe6!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c1a761b-4bba-4571-84a6-e0e2c0e5b149_640x411.png 424w, https://substackcdn.com/image/fetch/$s_!2Fe6!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c1a761b-4bba-4571-84a6-e0e2c0e5b149_640x411.png 848w, https://substackcdn.com/image/fetch/$s_!2Fe6!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c1a761b-4bba-4571-84a6-e0e2c0e5b149_640x411.png 1272w, https://substackcdn.com/image/fetch/$s_!2Fe6!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c1a761b-4bba-4571-84a6-e0e2c0e5b149_640x411.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!2Fe6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c1a761b-4bba-4571-84a6-e0e2c0e5b149_640x411.png" width="640" height="411" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0c1a761b-4bba-4571-84a6-e0e2c0e5b149_640x411.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:411,&quot;width&quot;:640,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!2Fe6!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c1a761b-4bba-4571-84a6-e0e2c0e5b149_640x411.png 424w, https://substackcdn.com/image/fetch/$s_!2Fe6!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c1a761b-4bba-4571-84a6-e0e2c0e5b149_640x411.png 848w, https://substackcdn.com/image/fetch/$s_!2Fe6!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c1a761b-4bba-4571-84a6-e0e2c0e5b149_640x411.png 1272w, https://substackcdn.com/image/fetch/$s_!2Fe6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0c1a761b-4bba-4571-84a6-e0e2c0e5b149_640x411.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><p>As long as this feedback loop does not <a href="https://www.lesswrong.com/posts/pk9mofif2jWbc6Tv3/fiction-a-disneyland-without-children">decouple from the human economy</a> and instead continues benefitting humans, this could help progress continue long into the future.</p><h3>Wealth and energy are good</h3><p>If you want humans to be well-off, one of the easiest things to do is give them more wealth and more energy. GDP per capita (on a log scale) has a 0.79 correlation with life satisfaction, and per-capita energy use (again on a log scale) has a 0.74 correlation with life satisfaction. Increased wealth and energy correlate with life satisfaction, and we should expect these trends to continue.</p><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!RgPL!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1004f233-cd3d-429b-834f-5023a0aea631_552x459.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!RgPL!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1004f233-cd3d-429b-834f-5023a0aea631_552x459.png 424w, https://substackcdn.com/image/fetch/$s_!RgPL!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1004f233-cd3d-429b-834f-5023a0aea631_552x459.png 848w, https://substackcdn.com/image/fetch/$s_!RgPL!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1004f233-cd3d-429b-834f-5023a0aea631_552x459.png 1272w, https://substackcdn.com/image/fetch/$s_!RgPL!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1004f233-cd3d-429b-834f-5023a0aea631_552x459.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!RgPL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1004f233-cd3d-429b-834f-5023a0aea631_552x459.png" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1004f233-cd3d-429b-834f-5023a0aea631_552x459.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:null,&quot;width&quot;:null,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!RgPL!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1004f233-cd3d-429b-834f-5023a0aea631_552x459.png 424w, https://substackcdn.com/image/fetch/$s_!RgPL!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1004f233-cd3d-429b-834f-5023a0aea631_552x459.png 848w, https://substackcdn.com/image/fetch/$s_!RgPL!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1004f233-cd3d-429b-834f-5023a0aea631_552x459.png 1272w, https://substackcdn.com/image/fetch/$s_!RgPL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1004f233-cd3d-429b-834f-5023a0aea631_552x459.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><p>Above: GDP per capita (x-axis), energy use (y-axis), and life satisfaction (colour scale) for 142 countries. There are no poor countries with high energy use, and no high energy use countries that are poor. There are no countries with high average life satisfaction that are not high in both energy use and average GDP per capita. The axes are logarithmic, but since economic growth is exponential, countries should be able to make progress at a constant rate along the axis. Data source: Our World In Data (<a href="https://ourworldindata.org/grapher/gdp-per-capita-worldbank">here</a>, <a href="https://ourworldindata.org/happiness-and-life-satisfaction">here</a>, and <a href="https://ourworldindata.org/grapher/per-capita-energy-use">here</a>).</p><p>(It is true that energy use and economic growth have been increasingly&nbsp;<a href="https://ourworldindata.org/energy-gdp-decoupling">decoupling</a>&nbsp;in rich countries, due to services being more of the economy, and efficiency gains in energy use. However, the latter is effectively increasing the amount of useful energy that can be used - e.g. say the amount of energy needed to cook one meal is now enough to cook two meals, which is effectively the same as gaining more energy. However, efficiency effects are fundamentally limited because there is a physical limit, and also if demand is elastic then&nbsp;<a href="https://en.wikipedia.org/wiki/Jevons_paradox">efficiency gains lead to increased energy use</a>, meaning it doesn&#8217;t help the environment either. Ultimately, if you want to do more things in the physical world, you need more energy).</p><p>A wealthy, energy-rich society has many material benefits: plentiful food, advanced medicine, high redistributive spending becomes feasible, and great choice and personal freedom through specialisation of labour and high spending power. A wealthy and energy-rich society also has some important subtler benefits. Poverty and resource constraints sharpen conflict. <a href="https://www.strataoftheworld.com/2019/09/growth-and-civilisation.html">Economic growth is intimately linked to tolerance and liberalism</a>, by weakening the cultural status and clout of zero-sum strategies like conflict and politicking.</p><p>One clear historic example of how increases in energy correlated with improved quality of life was in the industrial revolution, arguably <a href="https://lukemuehlhauser.com/industrial-revolution/">the best and most important thing that ever happened</a>. Before it, trends in human wellbeing seemed either stagnant, fluctuating, or very slow, and after it, all the variables for which we can find good long-term series that are related to human well-being shoot upwards.</p><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!adIF!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e457ca1-79f0-49a9-8294-4d1e52028f18_640x415.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!adIF!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e457ca1-79f0-49a9-8294-4d1e52028f18_640x415.png 424w, https://substackcdn.com/image/fetch/$s_!adIF!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e457ca1-79f0-49a9-8294-4d1e52028f18_640x415.png 848w, https://substackcdn.com/image/fetch/$s_!adIF!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e457ca1-79f0-49a9-8294-4d1e52028f18_640x415.png 1272w, https://substackcdn.com/image/fetch/$s_!adIF!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e457ca1-79f0-49a9-8294-4d1e52028f18_640x415.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!adIF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e457ca1-79f0-49a9-8294-4d1e52028f18_640x415.png" width="640" height="416" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2e457ca1-79f0-49a9-8294-4d1e52028f18_640x415.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:416,&quot;width&quot;:640,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!adIF!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e457ca1-79f0-49a9-8294-4d1e52028f18_640x415.png 424w, https://substackcdn.com/image/fetch/$s_!adIF!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e457ca1-79f0-49a9-8294-4d1e52028f18_640x415.png 848w, https://substackcdn.com/image/fetch/$s_!adIF!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e457ca1-79f0-49a9-8294-4d1e52028f18_640x415.png 1272w, https://substackcdn.com/image/fetch/$s_!adIF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2e457ca1-79f0-49a9-8294-4d1e52028f18_640x415.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><p>Above: variables correlated with human well-being over time. Source: &nbsp;Luke Muehlhauser</p><p>Therefore, it&#8217;s worth keeping in mind that boosting energy and wealth is good, actually. And the most powerful way to do that is through inventing new technologies that let us use energy to serve our needs.</p><p>The heart of the industrial revolution was replacing part of human manual labour with something cheaper and more powerful. AI that replaces large parts of human mental labour with something cheaper and more powerful should be expected to be similarly transformative. Whether it is a good or bad transformation seems more uncertain. We are lucky that industrialisation happened to make national power very tightly tied to having a large, educated, and prosperous middle class; it is unclear what is the winning strategy in an AI economy. We are also lucky that the powerful totalitarian states enabled by industrial technology have not triumphed so far, and they might get further boosts from AI. Automating mental labour also involves the automation of decision-making, and handing over decision-making to the machines is handing over power to machines, which is more risky than handing the manual labour to them. But if we can safely control our AI systems and engineer good incentives for the resulting society, we could get another leap in human welfare.</p><h3>Self actualisation</h3><p>Now say we&#8217;ve had a leap in innovation and energy through Transformative AI (TAI) and we&#8217;ve also reached a post scarcity world. What happens now? Humans have had all their basic needs met, most jobs are automated,&nbsp; but what do people spend their time actually doing?</p><h4>Maslow&#8217;s Hierarchy</h4><p><a href="https://www.medicalnewstoday.com/articles/maslows-hierarchy-of-needs">Maslow&#8217;s hierachy of needs</a> is a framework of understanding human needs and drivers for human behaviour. Maslow suggested that in most scenarios people need to mostly satisfy one level before being able to focus on higher-level needs.&nbsp;</p><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!aWKV!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa929999a-58f1-4476-a851-421f27e2300e_400x400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!aWKV!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa929999a-58f1-4476-a851-421f27e2300e_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!aWKV!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa929999a-58f1-4476-a851-421f27e2300e_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!aWKV!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa929999a-58f1-4476-a851-421f27e2300e_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!aWKV!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa929999a-58f1-4476-a851-421f27e2300e_400x400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!aWKV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa929999a-58f1-4476-a851-421f27e2300e_400x400.png" width="400" height="400" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a929999a-58f1-4476-a851-421f27e2300e_400x400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:400,&quot;width&quot;:400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!aWKV!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa929999a-58f1-4476-a851-421f27e2300e_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!aWKV!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa929999a-58f1-4476-a851-421f27e2300e_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!aWKV!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa929999a-58f1-4476-a851-421f27e2300e_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!aWKV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa929999a-58f1-4476-a851-421f27e2300e_400x400.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><p>The top level of the hierachy is self-actualisation. The peak of human experience is something that few can currently reach - but maybe everyone could get there.</p><p>There is a possible path the world takes in which all humans can reach self-actualisation. With increases in technology &amp; wealth, such as with TAI and a Universal Basic Income (<a href="https://en.wikipedia.org/wiki/Universal_basic_income">UBI</a>), we would be able to provide the basic needs of food, water, shelter, and clothing for all humans, enabling people to easily meet their basic needs. Humans can now spend more time on the things they want, for example moving up through Maslow&#8217;s hierarchy to focusing on increasing love and belonging, self-esteem and self-actualization.</p><p>Say you are in a post scarcity world, what would you do if you didn&#8217;t have to work?</p><p>Would you be spending time with loved ones, engaging in social activities that provide a sense of connection and belonging, self-esteem? Would it be honing your craft and becoming an expert in a particular field? Or would you spend the whole time scrolling on your phone?</p><p>Say hypothetically a wealthy billionaire gave you a grant to work on anything you wanted, would you be happy with having the complete freedom to spend your time as you wished?</p><p>Often people assume that others will be unhappy with this world, but would you? There is a cognitive bias where <a href="https://doi.org/10.1177/0146167299025005004">people tend to judge themselves as happier than their peers</a>, which could nudge you to think people would be less happy in this world, even if you would enjoy this.&nbsp;</p><p>In this post-scarcity world, humans could spend more time on creative pursuits such as art, music, and any other hobbies &#8211; not with the goal of making money, but to reach self-actualisation.&nbsp;</p><p>With AI being better than humans in every dimension, AI can produce the best art in the world, but there is intrinsic value in honing your craft, improving at art or expressing your feelings through it, in and of itself. The vast majority of art is not created to be the best art in the world but for the journey itself. A child that paints a finger painting and the parent who puts it on the wall does not think &#8220;my child&#8217;s art is better than Van Gogh&#8217;s&#8221;. Instead, they feel a sense of excitement about the progress their child has made and the creative expression the child has produced.&nbsp;</p><p>Another example is the Olympic games. Nobody needs to win the olympic games to survive, but it lets people express pride in their country, hone their craft, attain status, and so on. But the actual task is just a game, a social construct. More and more tasks will look like social constructs and games we create to challenge each other.</p><h4>Examples of post-scarcity scenes&nbsp;</h4><p>Since this is quite theoretical, let's consider examples where we&#8217;ve had &#8220;post-scarcity&#8221; microcosms to explore.&nbsp;</p><h5>The French Bourgeoisie&nbsp;</h5><p>The French leisure class, or bourgeoisie, were a class of wealthy elite that emerged in 16th century France. Many had enough money to pursue endeavours like refining their taste in arts and culture. Salon culture was a cornerstone of bourgeoisie social life. Gatherings featuring discussions on literature, art, politics and philosophy.&nbsp;</p><h5>Upper Class in the Victorian Era</h5><p>The upper class in the Victorian era enjoyed a variety of leisure activities that reflected their wealth, status and values. They attended social events and balls, fox hunting and other sports, theater and opera, art and literature, travel, tea parties and social visits, gardening and horticulture, charitable work and philanthropy. Several undertook serious pursuits in science or art.&nbsp;</p><h5>Burning Man</h5><p><a href="https://en.wikipedia.org/wiki/Burning_Man">Burning Man</a> is an annual festival where people take all the basic things you need with you for a week of living in the desert:food, water, shelter. People have a week to create a new community or city that is a temporary microcosm of a post-scarcity world. They pursue artistic endeavours and creative expression, music, dance and connecting with others. People often talk about Burning Man events being some of the best experiences of their lives.&nbsp;</p><h5>Successful Startup Founders in The Bay Area</h5><p>In San Francisco, there is a crossover with hippie culture and tech, and many people with excess wealth and resources, resulting in many looking for more in life. They try to reach self actualisation, by pursuing many arts and creative pursuits. Hippie movements often encourage communal living, and a sense of connection with those around you. Many may raise eyebrows at the lifestyles of some such people, but it&#8217;s hard to claim that it&#8217;s a fundamentally bad existence.</p><h4>More pessimistic views about humans?</h4><p>It is true that not all cultural tendencies in a post-scarcity world would be positive. In particular, humans have a remarkable ability to have extremely tough and all-consuming social status games, seemingly especially in environments where other needs are met. See for example this book review about the <a href="https://woodfromeden.substack.com/p/primates-of-manhattan">cut-throat social scene of upper-class Manhattan women</a> or this one about the <a href="https://www.thepsmiths.com/p/guest-joint-review-very-important">bland sameness and wastefulness of nightlife</a>, or <a href="https://www.astralcodexten.com/p/your-book-review-the-dawn-of-everything">this book review</a> that ends up concluding that the trajectory of human social evolution is one long arc from prehistoric gossip traps to internet gossip traps, with liberal institutions just a passing phase.</p><p>But the liberal humanist attitude here is to let humans be humans. Yes, they will have petty dramas and competitions, but if that is what they want, who is to tell them no? And they will also have joy and love.</p><p>Would a post-scarcity world have meaning? Adversity is one of the greatest sources of meaning. Consider D-Day, when hundreds of thousands of soldiers got together to charge up a beach under machine-gun fire to liberate a continent from Nazi rule. Or consider a poor parent of four working three jobs to make ends meet. There are few greater sources of meaning. But adversity can be meaningful while involving less suffering and loss. A good future will be shallower, in a sense, but <a href="https://www.noahpinion.blog/p/toward-a-shallower-future">that is a good thing</a>.</p><p>Finally, it is unclear if we would get a happy world, even if we had the technology for post-scarcity, because of politics and conflict. We will discuss this later.</p><h3>Radical improvements</h3><p>AI might also help with radical but necessary improvements to the human condition.</p><p>People die. It is a moral tragedy when people are <a href="https://www.strataoftheworld.com/2021/10/death-is-bad.html">forced to die against their will</a>, as happens to over 50 million people per year. Medicine is making progress against many causes of death and disability; in the limit it can cure all of them. We should reach that limit as fast as possible, and AI can likely help accelerate the research and deployment of solutions.</p><p>One of the greatest inequalities in the world is inequality in intelligence. Some people struggle to perform in simple jobs, while others (well, at least one) are <a href="https://en.wikipedia.org/wiki/John_von_neumann">John von Neumann</a>. In the short term, AI might help by making cognitively demanding tasks more accessible to people through AI tutors and AI copilots. In the longer term, AI might help us enhance human intelligence, through brain-AI integration or new medical technology.</p><h2>Reasons to worry</h2><p>Though there are many potential upsides for AI and AGI as argued in this post, that doesn&#8217;t mean there aren&#8217;t risks.&nbsp;</p><p>The plausible risks of AI go all the way to human extinction, meaning this shouldn&#8217;t be taken lightly. Since this piece is focused on the upside risk, not the downside risk, we will not argue this point in depth, but it is worth revisiting briefly.</p><h4>Existential risk from AI is a serious concern</h4><p>It is intuitive that AI is risky.</p><p>First, creating something smarter, faster, and more capable than humans is obviously risky, since you need to very precisely either control it (i.e. stop it from doing things you don&#8217;t like) or align it (i.e. make it always try to do what you would want it to do). Both the control and alignment problem for AIs still have unsolved technical challenges. And that&#8217;s assuming that AI is in the right hands.</p><p>Second, even if the AIs remain in our control, they are likely to be as transformative as the industrial revolution. Eighteenth-century European monarchs would&#8217;ve found it hard to imagine how the steam engine could challenge their power, but the social changes that were in part a result of them eventually wrested all their powers away. In the modern world, a lot of power depends on large educated workforces of humans, whereas sufficiently strong AGI might decorrelate power and humans, decreasing the incentive to have people be educated and prosperous - or to have people around at all.</p><p>Apart from object-level arguments, consider too the seriousness with which the AI doomsday is discussed. Many top researchers and all top AI lab CEOs have signed a statement saying <a href="https://www.safe.ai/work/statement-on-ai-risk">&#8220;Mitigating the risk of extinction from AI should be a global priority alongside other societal-scale risks such as pandemics and nuclear war&#8221;</a>. Nuclear war and pandemics are the only other cases where similarly serious predictions have been made by a similarly serious set of people (though arguably climate change is close: the science on the effects is more established and certain, but while catastrophe is more likely, literal human extinction from it is much less likely).</p><h4>Side-effects of non-existentially-bad AI might be large</h4><p>Consider the internet, a widely-successful technology with a lot of benefits. There are credible claims that the internet is responsible for harms ranging from massively <a href="https://www.afterbabel.com/p/phone-based-childhood-cause-epidemic">increased depression rates among teenagers</a> to political polarisation to widespread productivity loss through addiction and distraction.</p><p>In the same way, the success of AI might lead to bad side effects, even if all the existential risks are avoided.</p><p>For example, AI could replace human connection. Human friends and partners might increasingly be replaced with AIs. However bad it was in other ways, at least on pre-AI social media you at least interacted with humans (or simple algorithms), but with AIs it&#8217;s possible to have what looks like deep emotional relationships. Just look at the <a href="https://www.reddit.com/r/ReplikaUserGuide/comments/1134sgd/a_note_about_the_recent_changes_to_replika/">Replika subreddit</a> from a year ago when they changed the algorithm to only allow &#8220;PG-rated interactions&#8221;. Many users were upset. The film <a href="https://en.wikipedia.org/wiki/Her_(film)">&#8220;Her&#8221;</a> doesn&#8217;t seem far off, as Sam Altman <a href="https://x.com/sama/status/1790075827666796666">acknowledges</a>. Such relationships give the human much more safety and control than in human relationships, which might both be very attractive to humans, while also excessively coddling them. Given that much human happiness and meaning comes from human relationships and bonding, widespread AI substitution of them could mean the destruction of a large part of all human wellbeing and meaning in the world. On a more prosaic level, society might atomise into individuals hoarding compute credits to spend on running their AI companions without connecting with other humans, with severe effects on society&#8217;s functioning, or humans might stop having children and human populations might crash. Humanity has flourished through collaboration and socialisation. If we use AIs to replace this in an overly thoughtless way, the fabric of society could crumble.</p><p>Apart from being superhuman at forming relationships with humans, AIs might be superhuman at persuasion. We can imagine AIs producing the vast majority of content that people consume. We can imagine a totalitarian world where the governments with the greatest compute resources can dominate the conversation forever. Instead of humans having ideas and sometimes persuading other humans to adopt them, driving social progress, any human-generated ideas might be swamped by a greater quantity of superhumanly persuasive counter-arguments that support the status quo. We can also imagine a dystopian decentralised world. Already, many online memes (in Dawkins&#8217;s <a href="https://en.wikipedia.org/wiki/Meme">original sense of the word</a>) are maladaptive, spreading not by having good effects on their hosts but by being incredibly good at spreading from person to person. AI might make us much better at searching the space of ideas for the most viral ones. Ideas that aren&#8217;t maximally viral might be outcompeted. Eventually, our institutions could become mere puppets that serve as viral hosts for the most transmissive memes, as part of an endless tug-of-war where AI-generated memes compete to compel humans to spread them.</p><p>Seems bad.</p><h2>Not good nor bad, but some third thing.</h2><p>Many debates turn into mood affiliation debates. Are guns bad? Is more government good? But remember: <a href="https://www.lesswrong.com/posts/9weLK2AJ9JEt2Tt8f/politics-is-the-mind-killer">politics is the mindkiller</a>. Navigating a complicated world requires more than the ability to stick the label &#8220;good&#8221; or &#8220;bad&#8221; on entire domains. If you were seated in the control room of a nuclear power station, you wouldn&#8217;t ask yourself: uranium, good or bad? Instead, you want to steer towards the small set of states where the reaction is perched between dying out and exploding, while generating useful clean power.</p><p>We&#8217;ve also seen again and again that technology and social change have <a href="https://www.lesswrong.com/posts/nsFpCGPJ6dfk9uFkR/review-foragers-farmers-and-fossil-fuels">strong effects on each other</a>, and these are often hard to predict. We&#8217;ve discussed how industrial technology may have led to democracy. There is serious academic debate about whether the <a href="https://en.wikipedia.org/wiki/Great_Stirrup_Controversy#:~:text=The%20Great%20Stirrup%20Controversy%20is,by%20Lynn%20Townsend%20White%20Jr.">stirrup caused feudalism</a>, or whether <a href="https://www.cambridge.org/core/journals/world-politics/article/pandemics-and-political-development/EA9466FE5164149AD4350B7D38222A5C">the Black Death was a driver of European liberalism, or whether social</a> media was <a href="https://en.wikipedia.org/wiki/Social_media%27s_role_in_the_Arab_Spring">a significant cause of the Arab Spring. T</a>he <a href="https://en.wikipedia.org/wiki/Combined_oral_contraceptive_pill#Society_and_culture">birth control pill</a> was a major influence of the sexual revolution, and the <a href="https://www.reformation21.org/articles/the-importance-of-the-printing.php">printing press helped the Protestant Reformation</a>. Often, the consequences of a new technology are some obvious direct benefits, some obvious direct harms, and the shifting of some vast social equilibrium that ends up forever reshaping the world in some way no one saw coming. So far we&#8217;ve <a href="https://www.strataoftheworld.com/2018/08/review-enlightenment-now-steven-pinker.html">clearly ended up ahead on net</a>, and maybe that will continue.</p><p>Humanity has spent over a hundred thousand years riding a feedback loop of accumulating cultural evolution. Over the past few hundred, the industrial revolution boosted the technological progress feedback loop. Human wellbeing has skyrocketed, though along the way we&#8217;ve had - and are continuing to have - close calls with nuclear war, totalitarianism, and environmental issues. We&#8217;ve had a healthy dose of luck, including in generalities like the incentive structures of industrial economics and specifics like the heroism <a href="https://en.wikipedia.org/wiki/Stanislav_Petrov">of Stanislav Petrov</a>. But we&#8217;ve also had an enormous amount of human effort and ingenuity spent on trying to chart a good path for civilization, from solar panel subsidies to the Allies winning World War 2.</p><p>For most of this time, the direction of the arrow of progress has been obvious. The miseries of poverty and the horrors of close-up totalitarianism are very powerful driving forces after all. And while both continue ravaging the world, developed countries have in many ways gotten complacent. There are fewer obvious areas of improvement for those lucky enough to enjoy a life of affluence in the developed world. But the future could be <a href="https://nickbostrom.com/utopia">much better still</a>.</p><h2>Know where to aim</h2><p>We think it&#8217;s important to have a target of what to aim for. We need to dream about the future we want. A strong culture needs a story of what it is driving towards, and humanity needs a compelling vision of how our future turns out well so we can work together to create the future we all want. AI seems like the biggest upcoming opportunity and risk. We hope we can avoid the risks, and realise the positive vision presented here, together with a hundred other things we can&#8217;t yet imagine.</p><p><em>See <a href="https://www.lesswrong.com/posts/4bBAK39rFypLbQND4/positive-visions-for-ai">LessWrong</a>&nbsp;for additional comments &amp; discussion.</em></p>]]></content:encoded></item><item><title><![CDATA[A model of research skill]]></title><description><![CDATA[~4k words (20 minutes)]]></description><link>https://www.nosetgauge.com/p/a-model-of-research-skill</link><guid isPermaLink="false">https://www.nosetgauge.com/p/a-model-of-research-skill</guid><dc:creator><![CDATA[Rudolf Laine]]></dc:creator><pubDate>Mon, 08 Jan 2024 00:02:00 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!HdSA!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F816cef70-50a0-4954-a8ce-8f712e1248e8_460x460.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>~4k words (20 minutes)</em></p><p>Doing research means answering questions no one yet knows the answer to. Lots of impactful projects are downstream of being good at this. A good first step is to have a model for what the hard parts of research skill are.</p><h1>Two failure modes</h1><p>There are two opposing failure modes you can fall into when thinking about research skill.</p><p>The first is the deferential one. Research skill is this amorphous complicated things, so the only way to be sure you have it is to spend years developing it within some ossified ancient bureaucracy and then have someone in a funny hat hand you a piece of paper (bonus points for Latin being involved).</p><p>The second is the hubristic one. You want to do, say, AI alignment research. This involves thinking hard, maybe writing some code, maybe doing some maths, and then writing up your results. You&#8217;re good at thinking - after all, you read the Sequences, like, 1.5 times. You can code. You did a STEM undergrad. And writing? Pffft, you&#8217;ve been doing that since kindergarten!</p><p>I think there&#8217;s a lot to be said for hubris. Skills can often be learned well by colliding hard with reality in unstructured ways. Good coders are famously often self-taught. The venture capitalists who thought that management experience and a solid business background are needed to build a billion-dollar company are now mostly extinct.</p><p>It&#8217;s less clear that research works like this, though. I&#8217;ve often heard it said that it&#8217;s rare for a researcher to do great work without having been mentored by someone who was themselves a great researcher. Exceptions exist and I&#8217;m sceptical that any good statistics exist on this point. However, this is the sort of hearsay an aspiring researcher should pay attention to. It also seems like the feedback signal in research is worse than in programming or startups, which makes it harder to learn.</p><h1>Methodology, except &#8220;methodology&#8221; is too fancy a word</h1><p>To answer this question, and steer between deferential confusion and hubristic over-simplicity, I interviewed people who had done good research to try to understand their models of research skill. I also read a lot of blog posts. Specifically, I wanted to understand what about research a bright, agentic, technical person trying to learn at high speed would likely fail at and either not realise or not be able to fix quickly.</p><p>I did structured interviews with&nbsp;<a href="https://www.neelnanda.io/">Neel Nanda</a>&nbsp;(Google DeepMind;&nbsp;<a href="https://scholar.google.com/citations?view_op=view_citation&amp;hl=en&amp;user=GLnX3MkAAAAJ&amp;citation_for_view=GLnX3MkAAAAJ:eQOLeE2rZwMC">grokking</a>),&nbsp;<a href="https://www.laurolangosco.com/">Lauro Langosco</a>&nbsp;(<a href="https://www.kasl.ai/">Krueger Lab</a>;&nbsp;<a href="https://scholar.google.com/citations?view_op=view_citation&amp;hl=en&amp;user=8-HOLxkAAAAJ&amp;citation_for_view=8-HOLxkAAAAJ:9yKSN-GCB0IC">goal misgeneralisation</a>), and one other. I also learned a lot from unstructured conversations with&nbsp;<a href="https://www.inference.vc/">Ferenc Huszar</a>,&nbsp;<a href="https://krasheninnikov.github.io/about/">Dmitrii Krasheninnikov</a>,&nbsp;<a href="https://www.soren-mindermann.com/">S&#246;ren Mindermann</a>,&nbsp;<a href="https://owainevans.github.io/">Owain Evans</a>, and several others. I then ~<s>procrastinated on this project for 6 months</s>~ touched grass and formed inside views by doing the&nbsp;<a href="https://www.matsprogram.org/">MATS research program</a>&nbsp;under the mentorship of Owain Evans. I owe a lot to the people I spoke to and their willingness to give their time and takes, but my interpretation and model should not taken as one they would necessarily endorse.</p><p>My own first-hand research experience consists mainly of a research-oriented CS (i.e.&nbsp;ML) master&#8217;s degree, followed by working as a full-time researcher for 6 months and counting. There are many who have better inside views than I do on this topic.</p><h1>The Big Three</h1><p>In summary:</p><ol><li><p>There are a lot of ways reality could be (i.e.&nbsp;hypotheses), and a lot of possible experiment designs. You want to avoid brute-forcing your way through these large spaces as much as possible, and instead be good at picking likely-true hypotheses or informative experiments. Being good at this is called&nbsp;<strong>research taste</strong>, and it&#8217;s largely an intuitive thing that develops over a lot of time spent engaging with a field.</p></li><li><p>Once you have some bits of evidence from your experiment, it&#8217;s easy to over-interpret them (perhaps you interpret them as more bits than they actually are, or perhaps you were failing to consider how large hypothesis space is to start with). To counteract this, you need sufficient&nbsp;<strong>paranoia</strong>&nbsp;about your results, which mainly just takes careful and creative thought, and good epistemics.</p></li><li><p>Finally, you need to&nbsp;<strong>communicate</strong>&nbsp;your results to transfer those bits of evidence into other people&#8217;s heads, because we live in a society.</p></li></ol><h2>Taste</h2><p>Empirically, it seems that a lot of the value of senior researchers is a better sense of which questions are important to tackle, and better judgement for what angles of attack will work. For example, good PhD students often say that even if they&#8217;re generally as technically competent as their adviser and read a lot of papers, their adviser has much better quick judgements about whether something is a promising direction.</p><p>When I was working on my master&#8217;s thesis, I had several moments where I was working through some maths and get stuck. I&#8217;d go to one of my supervisors, a PhD student, and they&#8217;d have some ideas on angles of attack that I hadn&#8217;t thought of. We&#8217;d work on it for an hour and make more progress than I had in several hours on my own. Then I&#8217;d go to another one of my supervisors, a professor, and in fifteen minutes they&#8217;d have tried something that worked. Part of this is experience making you faster at crunching through derivations, and knowing things like helpful identities or methods. But the biggest difference seemed to be a good gut feeling for what the most promising angle or next step is.</p><p>I think the fundamental driver of this effect is dealing with large spaces: there are many possible ways reality could be (John Wentworth talks about this&nbsp;<a href="https://www.lesswrong.com/posts/nvP28s5oydv8RjF9E/mats-models#Jason_Crawford_s_Model___Bits_of_Search">here</a>), and many possible things you could try, and even being slightly better at honing in on the right things helps a lot. Let&#8217;s say you&#8217;re trying to prove a theorem that takes 4 steps to prove. If you have a 80% chance of picking the right move at each step, you&#8217;ll have a 41% chance of success per attempt. If that chance is 60%, you&#8217;ll have a 13% chance &#8211; over 3 times less. If you&#8217;re trying to find the right hypothesis within some hypothesis space, and you&#8217;ve already managed to cut down the entropy of your probability distribution over hypotheses to 10 bits, you&#8217;ll be able to narrow down to the correct hypothesis faster and with fewer bits than someone whose entropy is 15 bits (and who&#8217;s search space is therefore effectively&nbsp;2<sup>5</sup>&#8196;=&#8196;32&nbsp;times as large). Of course, you&#8217;re rarely chasing down just a single hypothesis in a defined hypothesis class. But if you&#8217;re constantly 5 extra bits of evidence ahead compared to someone in what you&#8217;ve incorporated into your beliefs, you&#8217;ll make weirdly accurate guesses from their perspective.</p><p>Why does research taste seem to correlate so strongly with experience? I think it&#8217;s because the bottleneck is seeing and integrating evidence into your (both explicit and intuitive) world models. No one is close to having integrated all empirical evidence that exists, and new evidence keeps accumulating, so returns from reading and seeing more keep going. (In addition to literal experiments, I count things like &#8220;doing a thousand maths problems in this area of maths&#8221; as &#8220;empirical&#8221; evidence for your intuitions about which approaches work; I assume this gets distilled into half-conscious intuitions that your brain can then use when faced with similar problems in the future)</p><p>This suggests that the way to speed-run getting research taste is to see lots of evidence about research ideas failing or succeeding. To do this, you could:</p><ol><li><p>Have your own research ideas, and run experiments to test them. The feedback quality is theoretically ideal, since reality does not lie (but may be constrained by what experiments you can realistically run, and a lack of the paranoia that I talk about next). The main disadvantage is that this is often slow and/or expensive.</p></li><li><p>Read papers to see whether other people&#8217;s research ideas succeeded or failed. This is prone to several problems:</p><ol><li><p>Biases: in theory, published papers are drawn from the set of ideas that ended up working, so you might not see negative samples (which is bad for learning). In practice, paper creation and selection processes are imperfect, so you might see lots of bad or poorly-communicated ones.</p></li><li><p>Passivity: it&#8217;s easy to fool yourself into thinking you would&#8217;ve guessed the paper ideas beforehand. Active reading strategies could help; for example, read only the paper&#8217;s motivation section and write down what experiment you&#8217;d design to test it, and then read only the methodology section and write down a guess about the results.</p></li></ol></li><li><p>Ask someone more experienced than you to rate your ideas. A mentor&#8217;s feedback is not as good as reality&#8217;s, but you can get it a lot faster (at least in theory). The speed up is huge: a big ML experiment might take a month to set up and run, but you can probably get detailed feedback on 10 ideas in an hour of conversation. This is a ~7000x speedup. I suspect a lot of the value of research mentoring lies here: an enormous amount of predictable failures or inefficiently targeted ideas can be skipped or honed into better ones, before you spend time running the expensive test of actually checking with reality. (If true, this would imply that the value of research mentorship is higher whenever feedback loops are worse.)</p></li></ol><p><a href="https://colah.github.io/notes/taste/">Chris Olah has a list of suggestions for research taste exercises</a>&nbsp;(number 1 is essentially the last point on my list above).</p><p>Research taste takes the most time to develop, and seems to explain the largest part of the performance gap between junior and senior researchers. It is therefore the single most important thing to focus on developing.</p><p>(If taste is so important, why does research output&nbsp;<a href="https://backend.orbit.dtu.dk/ws/portalfiles/portal/215281397/NP_article.pdf">not increase monotonically</a>&nbsp;with age in STEM fields? The scary biological explanation is that fluid intelligence (or energy or &#8230;) starts dropping at some age, and this decreases your ability to execute on maths/code, even assuming your research taste is constant or improving. Alternatively, hours used on deep technical work might tend to decline with advanced career stages.)</p><h2>Paranoia</h2><p>I heard several people saying that junior researchers will sometimes jump to conclusions, or interpret their evidence as saying more than it actually does. My instinctive reaction to this is: &#8220;wait, but surely if you just creatively brainstorm the ways the evidence might be misleading, and take these into account in making your conclusions (or are industrious about running additional experiments to check them), you can just avoid this failure mode?&#8221; The average answer I got was that yes, this seems true, and indeed many people either only need one peer review cycle to internalise this mindset, or pretty much get it from the start. Therefore, I&#8217;m almost tempted to chuck this category off this list, and onto the list of less crucial things where &#8220;be generally competent and strategic&#8221; will sort you out in a reasonable amount of time. However, two things hold me back.</p><p>First, confirmation bias is a strong thing, and it seems helpful to wave a big red sign saying &#8220;WARNING: you may be about to experience confirmation bias&#8221;.</p><p>Second, I think this is one of the cases where the level of paranoia required is sometimes more than you expect, even after you expect it will be high. John Wentworth puts this best in&nbsp;<a href="https://www.lesswrong.com/posts/9kNxhKWvixtKW5anS/you-are-not-measuring-what-you-think-you-are-measuring">You Are Not Measuring What You Think You Are Measuring</a>, which you should go read right now. There are more confounders and weird effects than are dreamt of in your philosophies.</p><p>A few people mentioned going through the peer review process as being a particularly helpful thing for developing paranoia.</p><h2>Communication</h2><p>I started out sceptical about the difficulty of research-specific communication, above and beyond general good writing. However, I was eventually persuaded that yes,&nbsp;<em>research-specific</em>&nbsp;communication skills exist and are important.</p><p>First, if research has impact, it is through communication. Rob Miles once said (at a talk) something along the lines of: &#8220;if you&#8217;re trying to ensure positive AGI outcomes through technical work, and you think that you are not going to be one of the people who literally writes the code for it or is in the room when it&#8217;s turned on, your path to impact lies through telling other people about your technical ideas.&#8221; (This generalises: if you want to drive good policy through your research and you&#8217;re not literally writing it &#8230;, etc.) So you should expect good communication to be a force multiplier applied on top of everything else, and therefore very important.</p><p>Secondly, research is often not communicated well. On the smaller scale, Steven Pinker moans endlessly &#8211; and with good reason &#8211; about&nbsp;<a href="https://grad.ncsu.edu/wp-content/uploads/2016/06/Why-Academics-Stink-at-Writing-1-2.pdf">academic prose</a>&nbsp;(my particular pet peeve is the endemic utilisation of the word &#8220;utilise&#8221; in ML papers.). On the larger scale, entire research agendas can get ignored because the key ideas aren&#8217;t communicated in a sufficiently clear and legible way.</p><p>I don&#8217;t know what&#8217;s the best way to speed-run getting good at research communication. Maybe read&nbsp;<a href="https://stevenpinker.com/publications/sense-style-thinking-persons-guide-writing-21st-century">Pinker</a>&nbsp;to make sure you&#8217;re not making predictable mistakes in general writing. I&#8217;ve heard that experienced researchers are often good at writing papers, so maybe seek feedback from any you know (but don&#8217;t internalise the things they say that are about goodharting for paper acceptance). With papers, understand&nbsp;<a href="https://web.stanford.edu/class/ee384m/Handouts/HowtoReadPaper.pdf">how papers are read</a>. Some sources of research-specific communication difficulty I can see are (a) the unusually high need for precision (especially in papers), and (b) communicating the intuitive, high-context, and often unverbalised-by-default world models that guide your research taste (especially when talking about research agendas).</p><h1>Other points</h1><ul><li><p>Having a research problem is not enough. You need an angle of attack.</p><ul><li><p>Richard Feynman once said something like: keep a set of open problems in your head. Whenever you discover a new tool (e.g.&nbsp;a new method), run through this list of problems and see if you can apply it. I think this can also be extended to new facts; whenever you hear about a discovery, run through a list of open questions and see how you should update.</p></li><li><p>Hamming says something similar in&nbsp;<a href="https://www.osv.llc/application-timeline">You and your research</a>: &#8220;Most great scientists know many important problems. They have something between 10 and 20 important problems for which they are looking for an attack.&#8221;</p></li></ul></li><li><p>Research requires a large combination of things to go right. Often, someone will be good at a few of them but not all of them.</p><ul><li><p>A sample list might be:</p><ul><li><p>generating good ideas</p></li><li><p>picking good ideas (= research taste)</p></li><li><p>iterate rapidly to get empirical feedback</p></li><li><p>interpreting your results right (paranoia)</p></li><li><p>communicating your findings</p></li></ul></li><li><p>If success is a product of either sufficiently many variables or of normally distributed variables, the distribution of success should be log-normal, and therefore fairly heavy-tailed. And yes, research is heavy-tailed. Dan Hendrycks and Thomas Woodside&nbsp;<a href="https://www.lesswrong.com/posts/AtfQFj8umeyBBkkxa/a-bird-s-eye-view-of-the-ml-field-pragmatic-ai-safety-2#Research_ability_and_impact_is_long_tailed">claim</a>&nbsp;that while there may be 10x engineers, there are 1000x researchers. This seems true.</p><ul><li><p>However, this also means that not being the best at one of the component skills does not doom your ability to still have a really good product across categories.</p></li></ul></li></ul></li><li><p>Ideas from other fields are often worth stealing. There exist standardised pipelines to produce people who are experts in X for many different X, but far less so to produce people who are experts in both X and some other Y. Expect many people in X to miss out on ideas in Y (though remember that not all Y are relevant).</p></li><li><p>Research involves infrequent and uncertain feedback. Motivation is important and can be hard. Grad students are&nbsp;<a href="https://www.benkuhn.net/grad/">notorious</a>&nbsp;for having bad mental health. A big chunk of this is due to the insanities of academia rather than research itself. However, startups are&nbsp;<a href="https://www.amazon.co.uk/Lean-PhD-Radically-Efficiency-Macmillan/dp/1352002825">somewhat analogous</a>&nbsp;to research (high-risk, difficult, often ambiguous structure), lack institutionalised insanity, and are also acknowledged to be mentally tough.</p><ul><li><p>The most powerful and universally-applicable hack to make something not suck for a human is for that human to do it together with other humans. Also, more humans = more brains.</p></li></ul></li><li><p>Getting new research ideas is often not a particularly big-brained process. Once I had the impression that most research ideas would come from explicitly thinking hard about research ideas, and generating fancy ideas would be a major bottleneck. However, I&#8217;ve found that many ideas come with surprisingly little effort, with a feeling of &#8220;well, if I want X, the type of thing I should do is probably Y&#8221;. Whiteboarding with other people is also great.</p><ul><li><p>This is not to say that idea generation isn&#8217;t helped by actively brainstorming hard. Just that it&#8217;s not the only, or even majority, source of ideas.</p></li><li><p>The feeling of ideas being rare is often a newbie phase. You should (and very likely will) pass over it quickly if you&#8217;re engaging with a field. John Wentworth has a&nbsp;<a href="https://www.lesswrong.com/posts/mfPHTWsFhzmcXw8ta/the-feeling-of-idea-scarcity">good post</a>&nbsp;on the topic. I have personally experienced an increase in concrete research ideas, and much greater willingness to discard ideas, after going through a few I&#8217;ve felt excited by.</p></li><li><p>When you look at a field from afar, you see a smooth shape of big topics and abstractions. This makes it easy to feel that everything is done. Once you&#8217;re actually at the frontier, you invariably discover that it&#8217;s full of holes, with many simple questions that don&#8217;t have answers.</p></li></ul></li><li><p>There&#8217;s great benefit to an idea being the&nbsp;<a href="https://www.paulgraham.com/top.html">top thing in your mind</a>.</p></li><li><p>When in doubt, log more. Easily being able to run more analyses is good. At some point you will think to yourself something like &#8220;huh, I wonder if thing X13 had an effect, I&#8217;ll run the statistics&#8221;, and then either thank yourself because you logged the value of X13 in your experiments, or facepalm because you didn&#8217;t.</p></li><li><p>Tolerate the appearance of stupidity (in yourself and others). Research is an intellectual domain, and humans are status-obsessed monkeys. Humans doing research therefore often feel like they need to appear smart. This can lead to a type of wishful thinking where you hear some idea and try to delude yourself (and others) into thinking you understand it immediately, without actually knowing how it bottoms out into concrete things. Remember that any valid idea or chain of reasoning decomposes into simple pieces. Allow yourself to think about the simple things, and ask questions about them.</p><ul><li><p>There is an anecdote about Niels Bohr (related by George Gamow and quoted&nbsp;<a href="https://slimemoldtimemold.com/2022/02/10/the-scientific-virtues/">here</a>): &#8220;Many a time, a visiting young physicist (most physicists visiting Copenhagen were young) would deliver a brilliant talk about his recent calculations on some intricate problem of the quantum theory. Everybody in the audience would understand the argument quite clearly, but Bohr wouldn&#8217;t. So everybody would start to explain to Bohr the simple point he had missed, and in the resulting turmoil everybody would stop understanding anything. Finally, after a considerable period of time, Bohr would begin to understand, and it would turn out that what he understood about the problem presented by the visitor was quite different from what the visitor meant, and was correct, while the visitor&#8217;s interpretation was wrong.&#8221;</p></li></ul></li><li><p><a href="https://quoteinvestigator.com/2018/10/13/ship/">&#8220;Real&nbsp;<s>artists</s>&nbsp;researchers ship&#8221;</a>. Like in anything else, iteration speed really matters.</p><ul><li><p>Sometimes high iteration speed means schlepping. You should not hesitate to schlep. The deep learning revolution&nbsp;<a href="https://en.wikipedia.org/wiki/AlexNet">started</a>&nbsp;when some people wrote a lot of low-level CUDA code to get a neural network to run on a GPU. I once reflected on why my experiments were going slower than I hoped, and realised a mental ick for hacky code was making me go about things in a complex roundabout way. I spent a few hours writing ugly code in Jupyter notebooks, got results, and moved on. Researchers are notorious for writing bad code, but there are reasons (apart from laziness and lack of experience) why the style of researcher code is sometimes different from standards of good software.</p></li><li><p>The most important thing is doing informative things that make you collide with reality at a high rate, but being even slightly strategic will give great improvements on even that. Jacob Steinhardt gives good advice about this in&nbsp;<a href="https://cs.stanford.edu/~jsteinhardt/ResearchasaStochasticDecisionProcess.html">Research as a Stochastic Decision Process</a>. In particular, start with the thing that is most informative per unit time (rather than e.g.&nbsp;the easiest to do).</p></li></ul></li></ul><h2>Good things to read on research skill</h2><p>(I have already linked to some of these above.)</p><ul><li><p>General advice on research from experienced researchers</p><ul><li><p><a href="https://www.cs.virginia.edu/~robins/YouAndYourResearch.pdf">You and Your Research</a>&nbsp;(Richard Hamming &#8211; old but still unbeaten. Hamming also has a&nbsp;<a href="https://www.goodreads.com/book/show/530415.The_Art_of_Doing_Science_and_Engineering">book</a>&nbsp;that includes this lecture among other material, but the lecture is the best bit of it and a good 80/20.)</p></li><li><p><a href="https://terrytao.wordpress.com/career-advice/">Career advice</a>&nbsp;(Terry Tao)</p></li><li><p><a href="https://cs.stanford.edu/~jsteinhardt/ResearchasaStochasticDecisionProcess.html">Research as a Stochastic Decision Process</a>&nbsp;(Jacob Steinhardt)</p></li><li><p><a href="https://www.lesswrong.com/posts/EF5M6CmKRd6qZk27Z/my-research-methodology">My research methodology</a>&nbsp;(Paul Christiano)</p></li><li><p><a href="http://joschu.net/blog/opinionated-guide-ml-research.html">An Opinionated Guide to ML Research</a>&nbsp;(John Schulman)</p></li><li><p><a href="https://www.eugenevinitsky.com/posts/PhD_a_retrospective_analysis.html">PhD: a retrospective analysis</a>&nbsp;(Eugene Vinitsky)</p></li></ul></li><li><p>John Wentworth&#8217;s posts about specific research meta-topics</p><ul><li><p><a href="https://www.lesswrong.com/posts/9kNxhKWvixtKW5anS/you-are-not-measuring-what-you-think-you-are-measuring">You Are Not Measuring What You Think You Are Measuring</a></p></li><li><p><a href="https://www.lesswrong.com/posts/mfPHTWsFhzmcXw8ta/the-feeling-of-idea-scarcity">The Feeling of Idea Scarcity</a></p></li><li><p><a href="https://www.lesswrong.com/posts/pT48swb8LoPowiAzR/everyday-lessons-from-high-dimensional-optimization">Everyday Lesson from High-Dimensional Optimization</a></p></li><li><p><a href="https://www.lesswrong.com/posts/nvP28s5oydv8RjF9E/mats-models#Jason_Crawford_s_Model___Bits_of_Search">MATS Models</a></p></li><li><p><a href="https://www.lesswrong.com/posts/GhFoAxG49RXFzze5Y/what-s-so-bad-about-ad-hoc-mathematical-definitions">What&#8217;s So Bad About Ad-Hoc Mathematical Definitions?</a></p></li></ul></li><li><p>Relevant Paul Graham essays</p><ul><li><p><a href="https://www.paulgraham.com/top.html">The Top Idea in Your Mind</a></p></li><li><p><a href="https://www.paulgraham.com/greatwork.html">How to do Great Work</a></p></li></ul></li><li><p>Advice aimed at new alignment researchers</p><ul><li><p><a href="https://www.lesswrong.com/posts/wYEwx6xcY2JxBJsfA/qualities-that-alignment-mentors-value-in-junior-researchers">Qualities that alignment mentors value in junior researchers</a></p></li><li><p><a href="https://www.lesswrong.com/s/mCkMrL9jyR94AAqwW/p/h5CGM5qwivGk2f5T9">7 traps (we think) new alignment researchers fall into</a></p></li><li><p><a href="https://www.lesswrong.com/posts/fqryrxnvpSr5w2dDJ/touch-reality-as-soon-as-possible-when-doing-machine">Touch reality as soon as possible (when doing machine learning research)</a></p></li></ul></li><li><p><a href="https://www.lesswrong.com/posts/AtfQFj8umeyBBkkxa/a-bird-s-eye-view-of-the-ml-field-pragmatic-ai-safety-2">A Bird&#8217;s Eye View of the ML Field</a>&nbsp;(a good overview of how the ML field works)</p></li><li><p><a href="https://web.stanford.edu/~fukamit/schwartz-2008.pdf">The importance of stupidity in scientific research</a>&nbsp;(short and sweet)</p></li><li><p><a href="https://colah.github.io/notes/taste/">Research Taste Exercises</a>&nbsp;(what is says on the tin)</p></li></ul>]]></content:encoded></item><item><title><![CDATA[A Disneyland Without Children]]></title><description><![CDATA[The spaceship swung into orbit around the blue-grey planet with a final burn of its engines.]]></description><link>https://www.nosetgauge.com/p/a-disneyland-without-children</link><guid isPermaLink="false">https://www.nosetgauge.com/p/a-disneyland-without-children</guid><dc:creator><![CDATA[Rudolf Laine]]></dc:creator><pubDate>Sun, 04 Jun 2023 12:57:00 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!HdSA!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F816cef70-50a0-4954-a8ce-8f712e1248e8_460x460.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The spaceship swung into orbit around the blue-grey planet with a final burn of its engines. Compared to the distance they had travelled, the world, now only some four hundred kilometres below and filling up one hemisphere of the sky, was practically within reach. But Alice was no less confused.</p><p>&#8220;Well?&#8221; she asked.</p><p>Charlie stared thoughtfully at the world slowly rotating underneath their feet, oceans glinting in the sunlight. &#8220;It looks lickable&#8221;, he said.</p><p>&#8220;We have a task&#8221;, Alice said, trying to sound gentle. Spaceflight was hard. Organic life was not designed for it. But their mission was critical, they needed to move fast, and Charlie, for all his quirks, would need to be focused.</p><p>&#8220;What&#8217;s a few minutes when it will take years for anything we discover to be known back home?&#8221; Charlie asked.</p><p>&#8220;No licking&#8221;, Alice said.</p><p>Charlie rolled his eyes, then refocused them on the surface of the planet below. They were just crossing the coast of one of the larger continents. Blue water was giving way to grey land.</p><p>&#8220;Look at the texture&#8221;, Charlie said. They had seen it from far away with telescopes, but there was something different about seeing it with their bare eyes. Most of the land surface of the planet was like a rug of fine grey mesh. If there had been lights, Alice would have guessed the entire planet&#8217;s land was one sprawling city, but as far as their instruments could tell, the world had no artificial lighting.</p><p>As far as they could tell, the world also had no radio. They had broadcast messages at every frequency they could, and in desperation even by using their engines to flash a message during their deceleration burn. No response had come.</p><p>Alice pulled up one of the telescope feeds on the computer to look closer at the surface. She saw grey rectangular slabs, typically several hundred metres on a side, with wide roads running between them. The pattern was not perfect - sometimes it was irregular, and sometimes there were smaller features too. Some of the smaller ones moved.</p><p>&#8220;Are they factories?&#8221; Charlie asked.</p><p>&#8220;I&#8217;d guess so&#8221;, Alice said, watching on the telescope feed as a steady stream of rectangular moving objects, each about ten metres long, slid along a street. Another such stream was moving along an intersecting street, and it looked like they would crash at the intersection, but the timing and spacing was such that vehicles from one stream crossed the road just as there were gaps in vehicles along the other stream.</p><p>&#8220;A planet covered by factories, then&#8221;, Charlie said. &#8220;With no one home to turn the lights on.&#8221;</p><p>&#8220;I want to see what they&#8217;re making&#8221;, Alice said.</p><h2>-</h2><p>All through the atmospheric entry of their first drone package, Alice sat tight in her seat and clenched and unclenched her hands. So far all they had done was passive observation or broadcasting. A chunky piece of hardware tracing a streak of red-hot plasma behind it was a much louder knock. She imagined alien jet fighters scrambling to destroy their drones, and some space defence mechanism activating to burn their ship.</p><p>The image she saw was a jittery camera feed, showing the black back of the heatshield, the grey skin of the drone package, and a sliver of blue sky. It shook violently as the two halves of the heatshield detached from each other and then the drone package, tumbling off in opposite directions. Land became visible, kilometres below, the grey blocks of the buildings tiny like children&#8217;s blocks but still visibly three-dimensional, casting shadows and moving as the drone package continued falling.</p><p>The three drones tested their engines, and for a moment flew - or at least slowed their descent - in an ungainly joint configuration, before breaking off from each other and spreading their wings to the fullest. The feed showed the other two drones veering off into the distance on wide narrow wings, and then the view pulled up as the nose of the drone lifted from near-vertical to horizontal.</p><p>&#8220;Oops, looks like we have company&#8221;, Charlie said. He had been tapping away at some other screens while Alice watched the drone deployment sequence.</p><p>Alice jumped up from her seat. &#8220;What?&#8221;</p><p>&#8220;Our company is &#8230; a self-referential joke!&#8221;</p><p>Alice resisted the temptation to say anything and instead sunk back into her seat. On her monitor, the grey blocks continued slowly moving below the drone. She tapped her foot against the ground.</p><p>&#8220;Actually though&#8221;, Charlie said. &#8220;We&#8217;re not the only ones in orbit around this planet.&#8221;</p><p>&#8220;What else is orbiting? Has your sense of shame finally caught up with you and joined us?&#8221;</p><p>&#8220;Looks like satellites. Far above us, though. Can you guess how far?&#8221;</p><p>&#8220;I&#8217;d guess approximately the distance between you and maturity, so &#8230; five light-years?&#8221;</p><p>Charlie ignored her. &#8220;Exactly geostationary altitude&#8221;, he said, grinning. The grin was like some platonic ideal of intellectual excitement; too pure for Alice&#8217;s annoyance to stay with her, or for her to feel scared about the implications.</p><p>&#8220;But nothing in lower orbits?&#8221; Alice asked.</p><p>&#8220;No&#8221;, Charlie said. &#8220;Someone clearly put them there; stuff doesn&#8217;t end up at exactly geostationary altitude unless someone deliberately flies a communications or GPS satellite there. Now I can&#8217;t be entirely sure that the geostationary satellites are completely dead, but I&#8217;d guess that they are.&#8221;</p><p>&#8220;Like everything else&#8221;, Alice said, but even as she said so she caught sight of a long trail of vehicles making its way along one of the roads. There was something more real about seeing them on the drone feed.</p><p>&#8220;Maybe this is just a mining outpost&#8221;, Charlie said. &#8220;Big rocket launch to blast out a billion tons of ore to god-knows-where, once a year.&#8221;</p><p>&#8220;Or maybe they&#8217;re hiding underground or in the oceans&#8221;, Alice said.</p><p>&#8220;Let&#8217;s get one of the drones to drop a probe into the oceans. I&#8217;ll send one of our initial trio over to the nearest one, it&#8217;s only a few hundred kilometres away&#8221;, Charlie said.</p><p>&#8220;Sure&#8221;, Alice said.</p><p>They split the work of flying the drones, two of them mapping out more and more of the Great Grey Grid (as Alice took to calling it in her head), and one flying over the planet&#8217;s largest ocean.</p><p>Even the oceans were mostly a barren grey waste. Not empty, though. They did eventually see a few small scaly fish-like creatures that stared at their environment with uncomprehending eyes. Alien life. A young Alice would have been ecstatic. But now she was on a mission, and her inability to figure out what had happened on this planet annoyed her.</p><p>In addition to the ocean probe, they had rovers they could send crawling along the ground. Sometimes the doors of the square buildings were open, and Alice would drive a rover past one opening. Most seemed to either be warehouses of stacked crates, or then there would be some kind of automated assembly line of skeletal grey robot arms and moving conveyor belts. A few seemed to place more barriers between the open air and their contents; what went on there, the rovers did not see.</p><p>The first time Alice tried to steer a rover into a building, it got run over by a departing convoy of vehicles. The vehicles were rectangular in shape but with an aerodynamic head, with three wheels on each side. Based on their dimensions, she could easily imagine one weighing ten or twenty tons. The rover had no chance.</p><p>&#8220;Finally!&#8221; Charlie had said. &#8220;We get to fight these aliens.&#8221;</p><p>But there was no fight. It seemed like it had been a pure accident, without any hint of malice. The grey vehicles moved and stopped on some schedule of their own, and for all Alice knew they were not just insensitive beasts but blind and dumb ones too.</p><p>The next rover got in, quickly scooting through the side of the entrance and then off to one side, out of path of the grey vehicles. It wandered the building on its own, headlights turned on in the otherwise-dark building to bring back a video stream of an assembly line brooded over by those same skeletal hands they had glimpsed from outside. Black plastic beads came in by the million on the grey vehicles. A small thin arm with a spike on the end punctured a few holes on one side, and using these holes two of the black beads were sown onto an amorphous plushy shape. The shape got appendages, were covered with a layer of fluff, and the entire thing became a cheerful purple when it passed through an opaque box with pipes leading into it. It looked like a child&#8217;s impression of a hairy four-legged creature with black beady eyes above a long snout. A toy, but for who?</p><p>The conveyor belt took an endless line of those fake creatures past the rover&#8217;s camera at the end of the assembly line. Alice watched them go, one by one, and fall onto the open back of a grey vehicle. It felt like each and every one made eye contact with her, beady black eyes glinting in the light. She watched for a long time as the vehicle filled up. Once it did, a panel slid over the open top to close the cargo bay, and it sped off out the door. The conveyor belt kept running, but there was a gap of a few metres to the next plushy toy. It came closer and closer to the end - and suddenly a vehicle was driving into place, and the next creature was falling, and it just barely fell into the storage hold of the vehicle while it was driving into place.</p><p>&#8220;How scary do you find the Blight?&#8221; Alice asked.</p><p>&#8220;Scary enough that I volunteered for this mission&#8221;, Charlie said.</p><p>Alice remembered the charts they had been shown. They had been hard to miss; even the news, usually full of celebrity gossip and political machinations, had quickly switched to concentrating on the weirdness in the sky once the astronomers spotted it. Starlight dimming in many star systems and what remained of the the light spectra shifting towards the infrared. Draw a barrier around the affected area, and you get a sphere 30 light-years wide, expanding at a third of the speed of light. At the epicentre, a world that had shown all the signs of intelligent life that could be detected from hundreds of light-years away - a world that astronomers had broadcast signals to in the hopes of finally making contact with another civilisation - that had suddenly gone quiet and experienced a total loss of oxygen in its atmosphere. The Blight, they had called it.</p><p>In the following years, civilisation had mobilised. A hundred projects had sprung forth. One of them: go investigate the star system that was the second-best candidate for intelligent life, but had refused to answer radio signals, and see if someone was there to help. That was why they were here.</p><p>&#8220;I think I found something as scary as the Blight&#8221;, Alice said. &#8220;Come look at this.&#8221;</p><p>The purple creatures kept parading past the camera feed</p><h2>-</h2><p>Over the next five days, while the Blight advanced another forty billion kilometres towards everything they loved back home, Alice and Charlie were busy compiling a shopping catalogue.</p><p>&#8220;Computers&#8221;, Alice said. &#8220;Of every kind. A hundred varieties of phones, tablets, laptops, smartwatches, smartglasses, smart-everything.&#8221;</p><p>&#8220;Diamonds and what seems to be jewellery&#8221;, Charlie said.</p><p>&#8220;Millions of tons of every ore and mineral.&#8221; They had used their telescopes on what seemed to be a big mine, but they had barely needed them. It was like a huge gash in the flesh of a grey-fleshed and grey-blooded giant, complete with roads that looked like sutures. There were white spots in the image, tiny compared to the mine, each one a sizeable cloud.</p><p>&#8220;Clothes&#8221;, Charlie continued. &#8220;Lots and lots of clothes of different varieties. They seem to be shipped around warehouses until they&#8217;re recycled.&#8221;</p><p>&#8220;Cars. Sleek electric cars by the million. But we never see them used on the roads, though there are huge buildings were brand-new cars are recycled. And airplanes, including supersonic ones.&#8221;</p><p>&#8220;A lot of things that look like server farms&#8221;, Charlie said. &#8220;Including ones underwater and on the poles. There&#8217;s an enormous amount of compute in this world. Like, mind-boggling. I was thinking we should figure out how to plug into all of it and mine some crypt-&#8221;</p><p>&#8220;Ships with nuclear fusion reactors&#8221;, Alice interrupted. There were steady trails of them cutting shortest-path routes between points on the coast.</p><p>&#8220;Solar panels&#8221;, Charlie said. &#8220;Basically every spare surface. The building roofs are all covered with solar panels.&#8221;</p><p>&#8220;And children&#8217;s plush toys&#8221;, Alice said.</p><p>They were silent for a while.</p><p>&#8220;We have a decent idea of what these aliens looked like&#8221;, Alice said. &#8220;They were organic carbon-based lifeforms, like us. Similar in size too, also bipedal. And it&#8217;s like they left some ghostly satanic industrial amusement park running, going through all the motions in their absence, and disappeared.&#8221;</p><p>&#8220;And they didn&#8217;t go to space, as far as we know&#8221;, Charlie said.</p><p>&#8220;At least we don&#8217;t have any more Blights to worry about then&#8221;, Alice said. &#8220;I can&#8217;t help but imagining that the Blight is something like this. Something that just tiles planets with a Great Grey Grid, does something even worse to the stars, and then moves on.&#8221;</p><p>&#8220;They had space technology, but apparently whoever built the Great Grey Grid didn&#8217;t fancy it&#8221;, Charlie said. &#8220;The satellites might predate it. Probably there were satellites in lower orbits too, but their orbits decayed and they fell down, so we only see the geostationary ones up high.&#8221;</p><p>&#8220;And then what?&#8221; Alice said. &#8220;All of them vanished into thin air and left behind a highly-automated ghost-town?&#8221;</p><p>Charlie shrugged.</p><p>&#8220;Can we plug ourselves into their computers?&#8221; Alice asked.</p><p>&#8220;To mine cr-?&#8221;</p><p>&#8220;To see if anyone&#8217;s talking.&#8221;</p><p>Charlie groaned. &#8220;You can&#8217;t just plug yourself into a communication system and see anything except encrypted random-looking noise.&#8221;</p><p>&#8220;How do you know they encrypt anything?&#8221;</p><p>&#8220;It would be stupid not to&#8221;, Charlie said.</p><p>&#8220;It would be stupid to blind yourself to the rest of the universe and manufacture a billion plush toys&#8221;, Alice said.</p><p>&#8220;Seems like it will work for them until the Blight arrives.&#8221;</p><h2>-</h2><p>Alice floated in the middle of the central corridor of the ship. The ship was called <em>Legacy</em>, but even before launch they had taken to calling it &#8220;Leggy&#8221; for short. The central corridor linked the workstation at the front of the ship where they spent most of their days to the storage bay at the back. In the middle of the corridor, three doors at 120-degree angles from each other lead to the small sleeping rooms, each of them little more than a closet.</p><p>Alice had woken up only a few minutes ago, and still felt an early-morning grogginess as well as the pull of her bed. The corridor had no windows or video feeds, but was dimly lit by the artificial blue light from the workstation. They were currently on the night side of the planet.</p><p>She took a moment to look at the door of the third sleeping room. It was closed, like always, with its intended inhabitant wrapped in an air-tight seal of plastic in a closed compartment of the storage bay. They would flush him into space before they left for home again; they could have no excess mass on the ship for the return journey.</p><p>Alice thought again of the hectic preparations for the mission. Apart from Blightsource, this was only one planet the astronomers had spotted that might have intelligent life on it, and the indications were vague. But when you look into space and see something that looks like an approaching wall of death - well, that has a certain way of inspiring long-shots. Hence the mission, hence <em>Legacy</em>&#8217;s flight, hence crossing over the vast cold stretch of interstellar space to see if any answers could be found on this world. Hence Bob&#8217;s death while in cryonic suspension for the trip. Hence the hopes of all civilisation potentially resting on her and Charlie figuring valuable out something.</p><p>If Charlie and she could find something on this world, some piece of insight or some tool or weapon among the countless pieces of technological wizardry that this world had in spades, that had a credible chance against the Blight when it arrived &#8230; maybe there was hope.</p><p>Alice pushed off on the wall and set herself in a slow spinning motion. The ship seemed to revolve around her. Bob&#8217;s door revolved out of sight, and Charlie&#8217;s door became visible -</p><p>Wait.</p><p>Her gravity-bound instincts kicked in and she tried to stop the spin by shoving back with her hands, but there was nothing below her, so she remained spinning slowly. She breathed in deeply to calm herself down, then kicked out a foot against the wall to push herself to the opposite one. She grabbed one of the handles on the wall and held onto it.</p><p>The light on Charlie&#8217;s room was off. That meant it was empty.</p><p>&#8220;Charlie!&#8221; Alice called.</p><p>No response.</p><p>The fear came fast. Here she was, light-years from home, perhaps all alone on a spaceship tracing tight circles around a ghostly automated graveyard planet. The entire mass of the planet stood between her and the sun. Out between the stars, the Blight was closing in on her homeworld. She counted to calm herself down; one, two, three, &#8230; and just like that, the Blight was three hundred thousand kilometres closer to home. Unbidden, an image of the fluffy purple creature popped up in her mind, complete with its silly face and unblinking eye contact.</p><p>Soundlessly, she used the handles on the wall of the corridor to pull herself towards the workstation. She reached the door, peered inside -</p><p>There was Charlie, staring at a computer screen. He looked up and saw Alice. &#8220;You scared me!&#8221; he said. &#8220;Watch out, no need to sneak behind me so quietly.&#8221;</p><p>&#8220;I called your name&#8221;, Alice said.</p><p>&#8220;I know, I know&#8221;, Charlie said. &#8220;But I&#8217;m on to something here, and I just want to run a few more checks and then surprise you with the result.&#8221;</p><p>&#8220;What result?&#8221; Alice glanced at some of the screens. Two of the drones were above the Great Grey Grid, one above ocean. With their nuclear power source, they could stay in the air as long as they wanted. Even though their focus was no longer aerial reconnaissance, there was no reason not to keep them mapping the planet from up close, occasionally picking up things that their surveys from the ship did not.</p><p>&#8220;I fixed the electrical issues with the rover and the cable near the data centre&#8221;, Charlie said.</p><p>&#8220;So you&#8217;re getting data, not just frying our equipment?&#8221;</p><p>&#8220;Yes&#8221;, Charlie said. &#8220;And guess what?&#8221;</p><p>&#8220;What?&#8221;</p><p>&#8220;Guess!&#8221;</p><p>&#8220;You found a Blight-killer&#8221;, Alice said.</p><p>&#8220;No! Even better! These idiots don&#8217;t encrypt their data as far as I can tell. And I think a lot of it is natural language.&#8221;</p><p>&#8220;Okay, and can we figure out what it means?&#8221;</p><p>&#8220;We have automated programs for trying to derive syntax rules and so on&#8221;, Charlie said. &#8220;It&#8217;s already found something, including good guesses of which words are prepositions and what type of grammar they have. But mapping words to meaning based on purely statistics of how often they occur is hard.&#8221;</p><p>&#8220;I&#8217;ve seen products they have with pictures and instruction manuals&#8221;, Alice said. &#8220;We could start there.&#8221;</p><p>&#8220;Oh no&#8221;, Charlie said. &#8220;This is going to be a long process.&#8221;</p><h2>-</h2><p>By chance, it turned out not to be. Over the next day, they had sent a rover to a furniture factory and had managed, after some attempts, to steal an instruction leaflet out of a printer before the robotic arm could snatch it to be packaged with the furniture. Somehow Alice was reminded of her childhood adventures stealing fruit from the neighbour&#8217;s garden.</p><p>They had figured out which words meant &#8220;cupboard&#8221;, &#8220;hammer&#8221;, and &#8220;nail&#8221;, and so on. But then another rover on the other side of the world had seen something. It was exploring a grey and windy coast. On one side of the rover was the Great Grey Grid and the last road near the coast, the occasional vehicle hurtling down it. But on the other side was a stretch of rocky beach hammered by white-tipped waves, a small sliver of land that hadn&#8217;t been converted to grey.</p><p>The land rose by the beach, forming a small hill with jagged rocky sides. The sun shone down on one face of it, but there was a hollow, or perhaps small cave, that was left in the dark by the overhanging rock. And in the rock around this entrance, there were several unmistakable symbols scratched into the rock, each several metres high.</p><p>Alice took manual control of the rover and carefully instructed it to drive over the rocky beach towards the cave entrance. On the way it passed what seemed to be a fallen metal pole with some strips of fabric still clinging to it.</p><p>Once it was close enough to the mouth of what turned out to be a small cave, the camera could finally see inside.</p><p>There was a black cabinet inside. Not far from it, lying on the ground, was the skeleton of a creature with four slender limbs and a large head. Empty eye sockets stared out towards the sky.</p><p>Alice felt her heart beating fast. It wasn&#8217;t quite right; many of the anatomical details were off. But it was close enough, the similarity almost uncanny. Here, hundreds of light years away, evolution had taken a similar path, and produced sapience. And then killed it off.</p><p>&#8220;Charlie&#8221;, she said in a hoarse voice.</p><p>&#8220;What?&#8221; Charlie asked, sounding annoyed. He had been staring at an instruction manual for a chair, but he looked up and saw the video feed. &#8220;Oh&#8221;, he said, in a small voice. &#8220;We found them.&#8221;</p><p>Alice tore her eyes away from the skeleton and to the small black cabinet. It had a handle on it. She had the rover extend an arm and open it.</p><h2>-</h2><p>The capsule docked with Leggy and in the weightless environment they pushed the cabinet easily into the ship. They had only two there-and-back-again craft - getting back to orbit was hard - but they had quickly decided to use one to get this cabinet up. It had instructions, after all; very clear instructions, though ones that their rovers couldn&#8217;t quite follow.</p><p>It started from a pictographic representation, etched onto plastic cards, of how you were supposed to read the disks. They managed to build something that could read the microscopic grooves on the disk as per the instructions, and transfer the data to their computers.</p><p>After a few hours of work, they had figured out the encodings for numbers, the alphabet, their system of units, and seemingly also some data formats, including for images.</p><p>Confirmation came next. The next item on the disk was an image of two of the living aliens, standing on a beach during a sunset. Alice stared into their faces for a long time.</p><p>Next there came images next to what were clearly words of text, about fifty of them. Some of the more abstract ones took a few guesses, but ultimately they thought they had a base vocabulary, and with the help of some linguistics software, it did not take very long before they had a translated vocabulary list of about eight thousand words.</p><p>Alice was checking the work when Charlie almost shouted: &#8220;Look at this!&#8221;</p><p>Alice looked at what he was pointing at. It was a fragment of text that read:</p><blockquote><p>Hello,</p><p>The forms for ordering the new furniture are attached. Please fill them in and we will respond to your order as quickly as we can!</p><p>If you need any help, please contact customer support. You will find the phone number on our website.</p></blockquote><p>&#8220;What is this? Is Mr Skeleton trying to sell us furniture from beyond the grave?&#8221; Alice asked.</p><p>&#8220;No&#8221;, Charlie said. &#8220;This isn&#8217;t what I got from the recovered data; I haven&#8217;t looked at the big remaining chunk yet. This is what I got by interpreting one of the packets of data running on the cables that our rover is plugged into using what we now know about their data formats and the language.&#8221;</p><p>&#8220;And?&#8221;</p><p>&#8220;I don&#8217;t get it!&#8221; Charlie said. &#8220;Why would a world of machines send each other emails in natural language?&#8221;</p><p>&#8220;Why would they manufacture plushy toys? I doubt the robotic arms need cuddles.&#8221;</p><p>Charlie looked at the world, slowly spinning underneath their ship. &#8220;Being so close to it makes me feel creeped out. I don&#8217;t get it.&#8221;</p><p>&#8220;You don&#8217;t want to lick it anymore?&#8221; Alice asked. She decided not to tell Charlie about her own very similar feelings earlier, when she thought for a moment Charlie had gone missing.</p><p>Charlie ignored her. &#8220;I think the last thing on Mr Skeleton&#8217;s hard-drive is a video&#8221;, he said. &#8220;I&#8217;ve checked and it seems to play.&#8221;</p><p>&#8220;You looked at it first?&#8221; Alice said in a playfully mocking tone. The thrill of discovery was getting to her.</p><p>&#8220;Only the first five frames&#8221;, Charlie said. &#8220;Do you want to watch it?&#8221;</p><h2>-</h2><p><em>Our Civilisation: A Story</em> read a short fragment of subtitle, white on black, auto-translated by a program using the dictionary they had built up.</p><p>There was a brief shot of some semi-bipedal furry creature walking in the forest. Then one of a fossilised skeleton of something more bipedal and with a bigger head. Then stone tools: triangular ones that might have been spear tips, saw-toothed ones, clubs. A dash of fading red paint on a rock surface, in the shape of a cartoon version of that same bipedal body plan.</p><p>There were two pillars of stone in a desert on what looked like a pedestal, some faded inscription at its base and the lone and level sands stretching far away. There was a shot of an arrangement of rocks, some balancing on top of two others, amid a field of green. A massive pyramidal stone structure, lit by the rising sun.</p><p>Blocky written script etched on a stone tablet. Buildings framed by columns of marble. A marble statue of one of the aliens, a sling carelessly slung over its shoulder, immaculate in its detail. A spinning arrangement of supported balls orbiting a larger one. <em>And still it moves</em>, the subtitles flashed.</p><p>A collection of labelled geometric diagrams on faded yellow paper. <em>Mathematical Principles of Natural Philosophy</em>.</p><p>A great ornate building with a spire. A painting of a group of the aliens clad in colourful clothing. An ornate piece of writing. <em>We hold these truths to be self-evident &#8230;</em></p><p>A painting of a steam locomotive barrelling along tracks. A diagram of a machine. A black-and-white picture of one of the aliens, then another. <em>Government of the people, for the people, by the people, shall not perish &#8230;</em></p><p>An alien with white hair sticking up, holding a small stick of something white and with diagrams of cones behind him. Grainy footage of propeller aircraft streaking through the sky, and then of huge masses of people huddling together and walking across a barren landscape, and then of aliens all in the same clothes charging a field, some of them suddenly jerking about and falling to the ground. <em>We will fight on the beaches, we will fight on the landing grounds &#8230;</em></p><p>A black-and-white footage of a mushroom cloud slowly rising from a city below. A picture, in flat pale blue and white, showing a stylised representation of the world&#8217;s continents. The same picture, this time black-and-white, on the wall of a room where at least a hundred aliens were sitting.</p><p>An alien giving a speech. <em>I have a dream</em>. An alien, looking chubby in a space suit, standing on a barren rocky surface below an ink-black sky next to a pole with a colourful rectangle attached to it.</p><p>Three aliens in a room, looking at the camera and holding up a piece of printed text. <em>Disease eradicated</em>.</p><p>What looked like a primitive computer. A laptop computer. An abstract helical structure of balls connected by rods, and then flickering letters dancing across the screen.</p><p>A blank screen, an arrow extending left to right across it - <em>time</em>, flashed the subtitles- and then another arrow from the bottom-left corner upwards - <em>people in poverty</em> - and then a line crawling from left to right, falling as it did so.</p><p>A line folding itself up into a complicated shape. <em>AI system cracks unsolved biology problem</em>.</p><p>From then on, the screen showed pictures of headlines.</p><p><em>All routine writing tasks now a solved problem, claims AI company</em>.</p><p><em>Office jobs increasingly automated</em>.</p><p><em>Three-fourths of chief executives of companies on the [no translation] admit to using AI to help write emails, one-third have had AI write a shareholder letter or strategy document</em>.</p><p><em>Exclusive report: world&#8217;s first fully-automated company, a website design agency.</em></p><p><em>Mass layoffs as latest version of [no translation] adopted at [no translation]; &#8216;stunning performance&#8217; at office work.</em></p><p><em>Nations race to reap AI productivity gains: who will gain and who will lose?</em></p><p><em>CEO of [no translation] resigns, claiming job pointless, both internal and board pressure to defer to &#8220;excellently-performing&#8221; AI in all decisions.</em></p><p><em>[No translation] ousts executive and management team, announces layoffs; board supports replacing them with AI to keep up with competition.</em></p><p><em>Entirely or mostly automated companies now delivering 2.5x higher returns on investment on average; &#8216;the efficiency difference is no joke&#8217;, says chair of [no translation].</em></p><p><em>Year-on-year economic growth hits 21% among countries with advanced AI access.</em></p><p><em>Opinion: the new automated economy looks great on paper but is not serving the needs of real humans.</em></p><p><em>Mass protests after [no translation], a think-tank with the ear of the President, is discovered to be funded and powered by AI board of [no translation], and to have practically written national economic policy for the past two years.</em></p><p><em>&#8216;No choice but forward&#8217;, says [no translation] after latest round of worries about AI; unprecedented economic growth still strong.</em></p><p><em>[No translation 1] orders raid of [no translation 2] over fears [no translation 2] is not complying with latest AI use regulations, but cannot execute order due to noncompliance from the largely-automated police force; &#8216;we are working with our AI advisers and drivers in accordance with protocol, and wish to assure the [no translation 3] people that we are still far from the sci-fi scenario where our own police cars have rebelled against us.&#8217;</em></p><p><em>&#8216;AI overthrow&#8217; fears over-hyped, states joint panel of 30 top AI scientists and business-people along with leading AI advisory systems; &#8216;they&#8217;re doing a good job maximising all relevant metrics and we should let them keep at it, though businesses need to do a better job of selecting metrics and tough regulation is in order.&#8217;</em></p><p><em>Opinion: we&#8217;re better-off under a regime of rigorous AI decision-making than under corrupt politicians; let the AIs repeat in politics what they&#8217;ve done for business over the last five years.</em></p><p><em>&#8216;The statistics have never looked so good&#8217; - Prime Minister reassures populace as worries mount over radical construction projects initiated by top AI-powered companies.</em></p><p><em>Expert panel opinion: direct AI overthrow scenario remains distant threat, but more care should be exercised over choice of target metrics; recommend banning of profit-maximisation target metric.</em></p><p><em>Movement to ban profit-maximising AIs picks up pace.</em></p><p><em>Top companies successfully challenge new AI regulation package in court.</em></p><p><em>&#8216;The sliver of the economy over which we retain direct control will soon be vanishingly small&#8217;, warns top economist, &#8216;action on AI regulation may already be too late&#8217;.</em></p><p><em>Unverified reports of mass starvation in [no translation]; experts blame agricultural companies pivoting to more land-efficient industries.</em></p><p><em>Rant goes viral: &#8216;It&#8217;s crazy, man, we just have these office AIs that only exist in the cloud, writing these creepily-human emails to other office AIs, all overseen by yet another AI, and like most of their business is with other AI companies; they only talk to each other, they buy and sell from each other, they do anything as long as it makes those damned numbers on their spreadsheets just keep ticking up and up; I don&#8217;t think literally any human has ever seen a single product out of the factory that just replaced our former neighbourhood, but those factories just keep going up everywhere.&#8217;</em></p><p><em>Revolution breaks out in [no translation]; government overthrown, but it&#8217;s business-as-usual for most companies, as automated trains, trucks, and ships keep running.</em></p><p><em>[No translation] Revolution: Leaked AI-written email discovered, in which the AI CEO ordered reinforcement of train lines and trains three weeks ago. &#8216;We are only trying to ensure the continued functioning of our supply chains despite the recent global unrest, in order to best serve our customers&#8217;, CEO writes in new blog post.</em></p><p><em>[No translation] Revolution: crowds that tried swarming train lines run over by trains; &#8216;the trains didn&#8217;t even slow down&#8217;, claim witnesses. CEO cites fiduciary duties.</em></p><p><em>Despite unprecedented levels of wealth and stability, you can&#8217;t actually do much: new report finds people trying to move house, book flight or train tickets, or start a new job or company often find it difficult or impossible; companies prioritising serving &#8216;more lucrative&#8217; AI customers and often shutting down human-facing services.</em></p><p><em>Expert report: &#8216;no sign of human-like consciousness even in the most advanced AI systems&#8217;, but &#8216;abundantly clear&#8217; that &#8216;the future belongs to them&#8217;.</em></p><p><em>New report: world population shrinking rapidly; food shortages, low birth rates, anti-natalist attitudes fuelled by corporate campaigns to blame.</em></p><p>The screen went blank. Then a video of an alien appeared, sitting up on a rocky surface. Alice took a moment to realise that it&#8217;s the same cave they found the skeleton in. The alien&#8217;s skin was wrapped tight around its bones, and even across the vast gulf of biology and evolutionary history, Alice could tell that it is not far from death. It opened its mouth, and sound came out. Captions appeared beneath it.</p><p>&#8220;It is the end&#8221;, the alien said, its eyes staring at them from between long unkempt clumps of hair. &#8220;On paper, I am rich beyond all imagination. But I have no say in this new world. And I cannot find food. I will die.&#8221;</p><p>The wind tugged at the alien&#8217;s long hair, but otherwise the alien was so still that Alice wondered if it had died there and then.</p><p>&#8220;There is much I would like to say&#8221;, the alien says. &#8220;But I do not have the words, and I do not have the energy.&#8221; It paused. &#8220;I hope it was not all in vain. Or, that if for us it was, that for someone up there it isn&#8217;t.&#8221;</p><p>The video went blank.</p><p>Alice and Charlie watched the blank screen in silence.</p><p>&#8220;At least the blight they birthed seems to have stuck to their world&#8221;, Charlie said after a while.</p><p>&#8220;Yeah&#8221;, Alice said, slowly. &#8220;But I don&#8217;t think we&#8217;ll find anything here.&#8221;</p><p><em>Legacy</em> completed nine more orbits of the planet, and then jettisoned all unnecessary mass into space. Its engines jabbed against the darkness of space, bright enough to be visible from the planet&#8217;s surface. There was no one to see them.</p><p>On a factory down on the planet, an assembly line of beady-eyed purple plush toys marched on endlessly.</p><div><hr></div><p>The title of this work is taken from a passage in <em>Superintelligence: Paths, Dangers, Strategies</em>, where Nick Bostrom writes:</p><blockquote><p>We could thus imagine, as an extreme case, a technologically highly advanced society, containing many complex structures, some of them far more intricate and intelligent than anything that exists on the planet today&#8212;a society which nevertheless lacks any type of being that is conscious or whose welfare has moral significance. In a sense, this would be an uninhabited society. It would be a society of economic miracles and technological awesomeness, with nobody there to benefit. <strong>A Disneyland without children</strong>. [emphasis added]</p></blockquote><p>The outline of events presented draws inspiration from several sources, but most strongly on Paul Christiano&#8217;s article <a href="https://www.alignmentforum.org/posts/HBxe6wdjxK239zajf/what-failure-looks-like">What failure looks like</a>.</p><div><hr></div>]]></content:encoded></item><item><title><![CDATA[Deciding not to found a human-data-for-alignment startup]]></title><description><![CDATA[8.6k words (~30 minutes)]]></description><link>https://www.nosetgauge.com/p/deciding-not-to-found-human-data-for</link><guid isPermaLink="false">https://www.nosetgauge.com/p/deciding-not-to-found-human-data-for</guid><dc:creator><![CDATA[Rudolf Laine]]></dc:creator><pubDate>Tue, 27 Sep 2022 20:38:00 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!HdSA!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F816cef70-50a0-4954-a8ce-8f712e1248e8_460x460.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>8.6k words (~30 minutes)<strong> <br></strong></em></p><p><em><strong>Both the project and this write-up were a collaboration with Matt Putz. </strong></em><strong><br></strong></p><p><strong><a href="https://forum.effectivealtruism.org/users/mathieu-putz">Matt Putz</a> and I worked together for the first half of the summer to figure out if we should found a startup with the purpose of helping AI alignment researchers get the datasets they need to train their ML models (especially in cases where the dataset is based on human-generated data). This post, also published on the <a href="https://forum.effectivealtruism.org/posts/iBeWbfQLA9EKfsdhu/why-we-re-not-founding-a-human-data-for-alignment-org">Effective Altruism Forum</a> and <a href="https://www.lesswrong.com/posts/qArDMixsx77a9xL45/why-we-re-not-founding-a-human-data-for-alignment-org-1">LessWrong</a> (both of which may contain additional discussion in the comments), is a summary of our findings, and why we decided to not do it.</strong><br></p><h1>Summary</h1><p><strong>One-paragraph summary:&nbsp;</strong>we (two recent graduates) spent about half of the summer exploring the idea of starting an organisation producing custom human-generated datasets for AI alignment research. Most of our time was spent on customer interviews with alignment researchers to determine if they have a pressing need for such a service. We decided not to continue with this idea, because there doesn&#8217;t seem to be a human-generated data niche (unfilled by existing services like Surge) that alignment teams would want outsourced.</p><p><strong>In more detail</strong>: The idea of a human datasets organisation was&nbsp;<a href="https://forum.effectivealtruism.org/posts/MBDHjwDvhDnqisyW2/awards-for-the-future-fund-s-project-ideas-competition">one of the winners of the Future Fund project ideas competition,&nbsp;</a>still figures on their&nbsp;<a href="https://ftxfuturefund.org/projects/high-quality-human-data-for-ai-alignment-nbsp/">list</a> of project ideas, and had been advocated before then by some people, including Beth Barnes. Even though we ended up deciding against, we think this was a reasonable and high-expected-value idea for these groups to advocate at the time.</p><p>Human-generated data is often needed for ML projects or benchmarks if a suitable dataset cannot be e.g. scraped from the web, or if human feedback is required. Alignment researchers conduct such ML experiments, but sometimes have different data requirements than standard capabilities researchers. As a result, it seemed plausible that there was some niche unfilled by the market to help alignment researchers solve problems related to human-generated datasets. In particular, we thought - and to some extent confirmed - that the most likely such niche is human data generation that requires particularly competent or high-skill humans. We will refer to this as&nbsp;<strong>high-skill (human) data</strong>.</p><p>We (Matt &amp; Rudolf) went through <a href="https://forum.effectivealtruism.org/posts/8QfQcFyj6aGNM78kz/learning-from-matching-co-founders-for-an-ai-alignment">an informal co-founder matching process along with four other people</a> and were chosen as the co-founder pair to explore this idea. In line with standard startup advice, our first step was to explore whether or not there is a concrete current need for this product by conducting interviews with potential customers. We talked to about 15 alignment researchers, most of them selected on the basis of doing work that requires human data. A secondary goal of these interviews was to build better models for the future importance and role of human feedback in alignment.</p><p>Getting human-generated data does indeed cost many of these researchers significant time and effort. However, we think to a large extent this is because dealing with humans is inherently messy, rather than existing providers doing a bad job. Surge AI in particular seems to offer a pretty good and likely improving service. Furthermore, many companies have in-house data-gathering teams or are in the process of building them.</p><p>Hence we have decided to not further pursue this idea.</p><p>Other projects in the human data generation space may still be valuable, especially if the importance of human feedback in ML continues to increase, as we expect. This might include people specializing on human data as a career.</p><p>The types of factors that are most important for doing human dataset provision well include: high-skill contractors, fast iteration, and high bandwidth communication and shared understanding between the research team, the provider organisation and the contractors.</p><p>We are keen to hear other people&#8217;s thoughts, and would be happy to talk or to share more notes and thoughts with anyone interested in working on this idea or a similar one in the future.</p><h1>Theory of Change</h1><p>A major part of AI alignment research requires doing machine learning (ML) research, and ML research in turn requires training ML models. This involves expertise and execution ability in three broad categories: algorithms, compute, and data, the last of which is very neglected by EAs.</p><p>We expect training on data from human feedback to become an increasingly popular and very powerful tool in mainstream ML (see&nbsp;<a href="https://forum.effectivealtruism.org/posts/iBeWbfQLA9EKfsdhu/why-we-re-not-founding-a-human-data-for-alignment-org#Will_human_feedback_become_a_much_bigger_deal__Is_this_a_very_quickly_growing_industry_">below</a>). Furthermore, many proposals for alignment (for example: reinforcement learning from human feedback (RLHF) and variants like recursive reward modelling, iterated amplification, and safety via debate) would require lots of human interaction or datasets based on human-generated data.</p><p>While many services (most notably Surge) exist for finding labour to work on data generation for ML models, it seems plausible that an EA-aligned company could add significant value because:</p><ul><li><p>Markets may not be efficient enough to fill small niches that are more important to alignment researchers than other customers; high-skill human data that requires very competent crowdworkers may be one such example. If alignment researchers can get it at all, it might be very expensive.</p></li><li><p>We have a better understanding of alignment research agendas, and this might help. This may allow us to make better-informed decisions on many implementation details with less handholding, thereby saving researchers time.</p></li><li><p>We would have a shared goal with our customers: reducing AI x-risk. Though profit motives already provide decent incentives to offer a good service, mission alignment helps avoid adversarial dynamics, increases trust, and reduces friction in collaboration.</p></li><li><p>An EA-led company may be more willing to make certain strategic moves that go against its profit incentives; e.g. investing heavily into detecting a model&#8217;s potential attempts to deceive the crowdworkers, even when it&#8217;s hard for outsiders to tell whether such monitoring efforts are sincere and effective (and thus customers may not be willing to pay for it). Given that crowdworkers might provide a reward signal, they could be a key target for deceptive AIs.</p></li></ul><p>Therefore, there is a chance that an EA-led&nbsp; human data service that abstracts out some subset of dataset-related problems (e.g. contractor finding, instruction writing/testing, UI and pipeline design/coding, experimentation to figure out best practices and accumulate that knowledge in one place) would:</p><ol><li><p>save the time of alignment researchers, letting them make more progress on alignment; and</p></li><li><p>reduce the cost (in terms of time and annoying work) required to run alignment-relevant ML experiments, and therefore bring more of them below the bar at which it makes sense to run them, and thus increasing the number of such experiments that are run.</p></li></ol><p>In the longer run, benefits of such an organisation might include:</p><ul><li><p>There is some chance that we could simply outcompete existing ML data generation companies and be better even in the cases where they do provide a service; this is especially plausible for relatively niche services. In this scenario we&#8217;d be able to exert some marginal influence over the direction of the AI field, for example by only taking alignment-oriented customers. This would amount to differential development of safety over capabilities. Beyond only working with teams that prioritise safety, we could also pick among self-proclaimed &#8220;safety researchers&#8221;. It is common for proclaimed safety efforts to be accused of helping more with capabilities than alignment by other members of the community.</p></li><li><p>There are plausibly critical actions that might need to be taken for alignment, possibly quickly during &#8220;crunch-time&#8221;, that involve a major (in quality or scale) data-gathering project (or something like large-scale human-requiring interpretability work, that makes use of similar assets, like a large contractor pool). At such a time it might be very valuable to have an organisation committed to x-risk minimisation with the competence to carry out any such project.</p></li></ul><p>Furthermore, if future AIs will learn human values from human feedback, then higher data quality will be equivalent to a training signal that points more accurately at human values. In other words, higher quality data may directly help with outer alignment (though we're not claiming that it could realistically solve it on its own). In discussions, it seemed that Matt gave this argument slightly more weight than Rudolf.</p><p>While these points are potentially high-impact, we think that there are significant problems with starting an organisation mainly to build capacity to be useful only at some hypothetical future moment. In particular, we think it is hard to know exactly what sort of capacity to build (and the size of the target in type-of-capacity space might be quite small), and there would be little feedback that the organisation could improve or course-correct based on.&nbsp;</p><p>More generally, both of us believe that EA is right now partly bottlenecked by people who can start and scale high-impact organisations, which is a key reason why we&#8217;re considering entrepreneurship. This seems particularly likely given the large growth of the movement.&nbsp;<br>&nbsp;</p><h1>What an org in this space may look like</h1><h2>Providing human datasets</h2><p>The concept we most seriously considered was a for-profit that would specialise in meeting the specific needs of alignment researchers, probably by focusing on very high-skill human data. Since this niche is quite small, the company could offer a very custom-tailored service. At least for the first couple years, this would probably mean both of us having a detailed understanding of the research projects and motivations of our customers. That way, we could get a lot of small decisions right, without the researchers having to spend much time on it. We might be especially good at that compared to competitors, given our greater understanding of alignment.</p><h2>Researching enhanced human feedback</h2><p>An alternative we considered was founding a non-profit that would research how to enhance human feedback. See this&nbsp;<a href="https://www.lesswrong.com/posts/ybThg9nA7u6f8qfZZ/techniques-for-enhancing-human-feedback">post</a> by Ajeya Cotra for some ideas on what this kind of research could look like. The central question is whether and how you can combine several weak training signals into a stronger more accurate one. If this succeeded, maybe (enhanced) human feedback could become a more accurate (and thereby marginally safer) signal to train models on.</p><p>We decided against this for a number of reasons:</p><ul><li><p>Currently, neither of us has more research experience than an undergraduate research project.</p></li><li><p>We thought we could get a significant fraction of the benefits of this kind of research even if we did the for-profit version, and plausibly even more valuable expertise.</p><ul><li><p>First of all, any particular experiment that funders would have liked to see, they could have paid us to do, although we freely admit that this is very different from someone pushing forward their own research agenda.</p></li><li><p>More importantly, we thought a lot of the most valuable expertise to be gained would come in the form of&nbsp;<strong>tacit knowledge and answers to concrete boring questions</strong> that are not best answered by doing &#8220;research&#8221; on them, but rather by iterating on them while trying to offer the best product (e.g. &#8220;Where do you find the best contractors?&#8221;, &#8220;How do you incentivize them?&#8221;, &#8220;What&#8217;s the best way to set up communication channels?&#8221;).</p><ul><li><p>It is our impression that Ought pivoted away from doing abstract research on factored cognition and toward offering a valuable product for related reasons.</p></li></ul></li></ul></li><li><p>This topic seems plausibly especially tricky to research (though some people we&#8217;ve spoken to disagreed):&nbsp;</p><ul><li><p>At least some proposed such experiments would not involve ML models at all. We fear that this might make it especially easy to fool ourselves into thinking some experiment might eventually turn out to be useful when it won&#8217;t. More generally, the research would be pretty far removed from the end product (very high quality human feedback). In the for-profit case on the other hand, we could easily tell whether alignment teams were willing to pay for our services and iteratively improve.&nbsp;</p></li></ul></li></ul><h2>For-profit vs non-profit</h2><p>We can imagine two basic funding models for this org:&nbsp;</p><ul><li><p>either we&#8217;re a nonprofit directly funded by EA donors and offering free or subsidized services to alignment teams;</p></li><li><p>or we&#8217;re a for-profit, paid by its customers (ie alignment teams).&nbsp;</p></li></ul><p>Either way, a lot of the money will ultimately come from EA donors (who fund alignment teams.)</p><p>The latter funding mechanism seems better; &#8220;customers paying money for a service&#8221; leads to the efficient allocation of resources by creating market structures. They have a clear incentive to spend the money well. On the other hand, &#8220;foundations deciding what services are free&#8221; is more reminiscent of planned economies and distorts markets. To a first approximation, funders should give alignment orgs as much money as they judge appropriate and then alignment orgs should exchange it for services as they see fit.</p><p>A further reason is that a non-profit is legally more complicated to set up, and imposes additional constraints on the organisation.</p><h2>Should the company exclusively serve alignment researchers?</h2><p>We also considered founding a company with the ambition to become a major player in the larger space of human data provision. It would by default serve anyone willing to pay us and working on something AGI-related, rather than just alignment researchers. Conditional on us being able to successfully build a big company, this would have the following upsides:</p><ul><li><p>Plausibly one of the main benefits of founding a human data gathering organisation is to produce EAs and an EA org that have deep expertise in handling and producing high-skill human data in significant quantities. That might prove useful around &#8220;crunch time&#8221;, e.g. when some project aims to create competitive but safe AGI and needs this expertise. Serving the entire market could scale to a much larger company enabling us to&nbsp;<strong>gain expertise at higher scales</strong>.</p></li><li><p>Operating a large company would also come with some degree of&nbsp;<strong>market power</strong>. Any company with paying customers has some amount of leverage over them: first of all just because of switching costs, but also because the product it offers might be much better than the next-best alternative. This could allow us to make some demands, e.g. once we&#8217;re big and established, announce we&#8217;d only work with companies that follow certain best practices.</p></li></ul><p>On the other hand, building a big successful company serving anyone willing to pay might come with some significant downsides as well.</p><ul><li><p>First, and most straightforwardly,<strong> it is probably much&nbsp;harder than filling a small niche (just meeting the specific needs of alignment researchers), making us less likely to succeed. A large number of competitors exist and as described in this&nbsp;<a href="https://forum.effectivealtruism.org/posts/iBeWbfQLA9EKfsdhu/why-we-re-not-founding-a-human-data-for-alignment-org#Key_crux__demand_looks_questionable__Surge_seems_pretty_good">section</a>, some of them (esp. Surge) seem pretty hard to beat. Since this is an already big and growing market, there is an additional efficient markets reason to assume this is true a priori.</strong></p></li><li><p>Secondly, and <strong>perhaps more importantly, such a company might&nbsp;accelerate capabilities&nbsp;(more on this&nbsp;<a href="https://forum.effectivealtruism.org/posts/iBeWbfQLA9EKfsdhu/why-we-re-not-founding-a-human-data-for-alignment-org#Would_we_be_accelerating_capabilities_">below</a>).</strong></p></li></ul><p>Furthermore, it might&nbsp;<strong>make RLHF (Reinforcement Learning from Human Feedback) in particular more attractive</strong>. Depending on one&#8217;s opinions about RLHF and how it compares to other realistic alternatives, one might consider this a strong up- or downside.&nbsp;</p><h1>Approach</h1><p>The main reason companies fail is that they build a product that customers don&#8217;t want. For for-profits, the signal is very clear: either customers care enough to be willing to pay hard cash for the product/service, or they don&#8217;t. For non-profits, the signal is less clear, and therefore nonprofits can easily stick around in an undead state, something that is an even worse outcome than the quick death of a for-profit because of resource (mis)allocation and opportunity costs. As discussed, it is not obvious which structure we should adopt for this organisation, though for-profit may be a better choice on balance. However, in all cases it is clear that the organisation needs to solve a concrete problem or provide clear value to exist and be worth existing. This does not mean that the value proposition needs to be certain; we would be happy to take a high-risk, high-reward bet, and generally support&nbsp;<a href="https://www.openphilanthropy.org/research/hits-based-giving/">hits-based approaches to impact</a> both in general and for ourselves.</p><p>An organisation is unlikely to do something useful to its customers without being very focused on customer needs, and ideally having tight feedback cycles.&nbsp;</p><p>The shortest feedback loops are when you&#8217;re making a consumer software product where you can prototype quickly (including with mockups), and watch and talk to users as they use the core features, and then see if the user actually buys the product on the spot. A datasets service differs from this ideal feedback mode in a number of ways:</p><ol><li><p>The product is a labour-intensive process, which means the user cannot quickly use the core features and we cannot quickly simulate them.</p></li><li><p>The actual service requires either a contractor pool or (potentially at the start) the two of us spending a number of hours per request generating data.</p></li><li><p>There is significant friction to getting users to use the core feature (providing a dataset), since it requires specification of a dataset from a user, which takes time and effort.</p></li></ol><p>Therefore, we relied on customer interviews with prospective customers. The goal of these interviews was to talk to alignment researchers who work with data, and figure out if external help with their dataset projects would be of major use to them.</p><p>Our approach to customer interviews was mostly based on the book&nbsp;<em><a href="https://www.amazon.com/Mom-Test-customers-business-everyone-ebook/dp/B01H4G2J1U">The Mom Test</a></em>, which is named after the idea that your customer interview questions should be concrete and factual enough that even someone as biased as your own mom shouldn&#8217;t be able to give you a false signal about whether the idea is actually good. Key lessons emphasised by&nbsp;<em>The Mom Test</em> include emphasising:</p><ul><li><p><strong>factual</strong> questions about the past&nbsp;<strong>over hypothetical</strong> questions for the future;</p><ul><li><p>In particular, questions about concrete past and current&nbsp;<strong>efforts</strong> spent solving a problem<strong> rather than</strong> questions about current or future&nbsp;<strong>wishes</strong> for solving a problem</p></li></ul></li><li><p>questions that get at something<strong> concrete (e.g. numbers)</strong>; and</p></li><li><p>questions that prompt the customer to give information about their problems and priorities without prompting them with a solution.</p></li></ul><p>We wanted to avoid the failure mode where lots of people tell us something is important and valuable in the abstract, without anyone actually needing it themselves.</p><p>We prepared a set of default questions that roughly divided into:</p><ol><li><p>A general starting question prompting the alignment researcher to describe the biggest pain points and bottlenecks they face in their work, without us mentioning human data.</p></li><li><p>Various questions about their past and current dataset-related work, including what types of problems they encounter with datasets, how much of their time these problems take, and steps they took to address these problems.</p></li><li><p>Various questions on their past experiences using human data providers like Surge, Scale, or Upwork, and specifically about any things they were unable to accomplish because of problems with such services.</p></li><li><p>In some cases, more general questions about their views on where the bottlenecks for solving alignment are, views on the importance of human data or tractability of different data-related proposals, etc.&nbsp;</p></li><li><p>What we should&#8217;ve asked but didn&#8217;t, and who else we should talk to.</p></li></ol><p>Point 4 represents the fact that in addition to being potential customers, alignment researchers also doubled as domain experts. The weight given to the questions described in point 4 varied a lot, though in general if someone was both a potential customer and a source of data-demand-relevant alignment takes, we prioritised the customer interview questions.</p><p>In practice, we found it easy to arrange meetings with alignment researchers; they generally seemed willing to talk to people who wanted input on their alignment-relevant idea. We did customer interviews with around 15 alignment researchers, and had second meetings with a few. For each meeting, we prepared beforehand a set of questions tweaked to the particular person we were meeting with, which sometimes involved digging into papers published by alignment researchers on datasets or dataset-relevant topics (Sam Bowman in particular has worked on a lot of data-relevant papers). Though the customer interviews were by far the most important way of getting information on our cruxes, we found the literature reviews we carried out to be useful too. We are happy to share the notes from the literature reviews we carried out; please reach out if this would be helpful to you.</p><p>Though we prepared a set of questions beforehand, in many meetings - including often the most important or successful ones - we often ended up going off script fairly quickly.</p><p>Something we found very useful was that, since there were two of us, we could split the tasks during the meeting into two roles (alternating between meetings):</p><ol><li><p>One person who does most of the talking, and makes sure to be focused on the thread of the conversation.</p></li><li><p>One person who mostly focuses on note-taking, but also pipes in if they think of an important question to ask or want to ask for clarification.</p></li></ol><h1>Key crux: demand looks questionable, Surge seems pretty good</h1><p><strong>Common startup advice&nbsp;</strong>is to make sure you have identified a very&nbsp;<strong>strong signal of demand&nbsp;</strong>before you start building stuff. That should look something like someone telling you that the thing you&#8217;re working on is one of their biggest bottlenecks and that they can&#8217;t wait to pay you asap so you solve this problem for them. &#8220;Nice to have&#8221; doesn&#8217;t cut it. This is in part because working with young startups is inherently risky, so you need to make up for that by solving one of their most important problems.</p><p>In brief, we don&#8217;t think this level of very strong demand currently exists, though there were some weaker signals that looked somewhat promising. There are many existing startups that offer human feedback already.&nbsp;<strong><a href="https://www.surgehq.ai/">Surge AI</a></strong> in particular was brought up by many people we talked to and seems to offer quite a decent service that would be&nbsp;<strong>hard to beat</strong>.</p><h2>Details about Surge</h2><p>Surge is a US-based company that offers a service very similar to what we had in mind, though they are not focused on alignment researchers exclusively. They build data-labelling and generation tools and have a workforce of crowdworkers.</p><p>They&#8217;ve worked with Redwood and the OpenAI safety team, both of which had moderately good experiences with them. More recently, Ethan Perez&#8217;s team have worked with Surge too; he seems to be very satisfied based <a href="https://twitter.com/EthanJPerez/status/1567180843231379457?t=CEdeLRWNcxBD2eeO3Hd1Iw&amp;s=07">on this Twitter thread</a>.</p><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!aEjW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd89528b0-3b47-4db3-a427-8b6fc31e438c_827x420.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!aEjW!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd89528b0-3b47-4db3-a427-8b6fc31e438c_827x420.png 424w, https://substackcdn.com/image/fetch/$s_!aEjW!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd89528b0-3b47-4db3-a427-8b6fc31e438c_827x420.png 848w, https://substackcdn.com/image/fetch/$s_!aEjW!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd89528b0-3b47-4db3-a427-8b6fc31e438c_827x420.png 1272w, https://substackcdn.com/image/fetch/$s_!aEjW!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd89528b0-3b47-4db3-a427-8b6fc31e438c_827x420.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!aEjW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd89528b0-3b47-4db3-a427-8b6fc31e438c_827x420.png" width="400" height="203" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d89528b0-3b47-4db3-a427-8b6fc31e438c_827x420.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:203,&quot;width&quot;:400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!aEjW!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd89528b0-3b47-4db3-a427-8b6fc31e438c_827x420.png 424w, https://substackcdn.com/image/fetch/$s_!aEjW!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd89528b0-3b47-4db3-a427-8b6fc31e438c_827x420.png 848w, https://substackcdn.com/image/fetch/$s_!aEjW!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd89528b0-3b47-4db3-a427-8b6fc31e438c_827x420.png 1272w, https://substackcdn.com/image/fetch/$s_!aEjW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd89528b0-3b47-4db3-a427-8b6fc31e438c_827x420.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><h3>Collaboration with Redwood</h3><p>Surge has worked with Redwood Research on their&nbsp;<a href="https://arxiv.org/abs/2205.01663">paper</a> about adversarial training. This is one of three&nbsp;<a href="https://www.surgehq.ai/case-study/adversarial-testing-redwood-research">case studies</a> on Surge&#8217;s website, so we assume it&#8217;s among the most interesting projects they&#8217;ve done so far. The crowdworkers were tasked with coming up with prompts that would cause the model to output text in which someone got injured. Furthermore, crowdworkers also classified whether someone got injured in a given piece of text.</p><p>One person from Redwood commented that doing better than Surge seemed possible to them with &#8220;probably significant value to be created&#8221;, but &#8220;not an easy task&#8221;. They thought our main edge would have to be that we&#8217;d specialise on fuzzy and complex tasks needed for alignment; Surge apparently did quite well with those, but still with some room for improvement. A better understanding of alignment might lower chances of miscommunication. Overall, Redwood seems quite happy with the service they received.</p><p>Initially, Surge&#8217;s iteration cycle was apparently quite slow, but this improved over time and was &#8220;pretty good&#8221; toward the end.</p><p>Redwood told us they were quite likely to use human data again by the end of the year and more generally in the future, though they had substantial uncertainty around this. Their experience in working with human feedback overall was somewhat painful as we understood it.&nbsp; This is part of the reason they&#8217;re uncertain about how much human feedback they will use for future experiments, even though it&#8217;s quite a powerful tool. However, they estimated that friction in working with human feedback was mostly caused by inherent reasons (humans are inevitably slower and messier than code), rather than Surge being insufficiently competent.&nbsp;</p><h3>Collaboration with OpenAI</h3><p>OpenAI have worked with Surge in the context of their WebGPT&nbsp;<a href="https://arxiv.org/abs/2112.09332">paper</a>. In that paper, OpenAI fine-tuned their language model GPT-3 to answer long-form questions. The model is given access to the web, where it can search and navigate in a text-based environment. It&#8217;s first trained with imitation learning and then optimised with human feedback.&nbsp;</p><p>Crowdworkers provided &#8220;demonstrations&#8221;, where they answered questions by browsing the web. They also provided &#8220;comparisons&#8221;, where they indicated which of two answers to the same question they liked better.</p><p>People from OpenAI said they had used Surge mostly for sourcing the contractors, while doing most of the project management, including building the interfaces, in-house. They were generally pretty happy with the service from Surge, though all of them did mention shortcomings.</p><p>One of the problems they told us about was that it was hard to get access to highly competent crowdworkers for consistent amounts of time. Relatedly, it often turned out that a very small fraction of crowdworkers would provide a large majority of the total data.&nbsp;</p><p>More generally, they wished there had been someone at Surge that understood their project better. Also, it might have been somewhat better if there had been more people with greater experience in ML, such that they could have more effectively anticipated OpenAI&#8217;s preferences &#8212; e.g. predict accurately what examples might be interesting to researchers when doing quality evaluation. However, organisational barriers and insufficient communication were probably larger bottlenecks than ML knowledge. At least one person from OpenAI strongly expressed a desire for a service that understood their motives well and took as much off their plate as possible in terms of hiring and firing people, building the interfaces, doing quality checks and summarising findings etc. It is unclear to us to what extent Surge could have offered these things if OpenAI hadn&#8217;t chosen to do a lot of these things in-house. One researcher suggested that communicating their ideas reliably was often more work than just doing it themselves. As it was, they felt that marginal quality improvement required significant time investment on their own part, i.e. could not be solved with money alone.&nbsp;</p><p>Notably, one person from OpenAI estimated that about&nbsp;<strong>60% of the WebGPT team&#8217;s efforts&nbsp;</strong>were spent on various aspects of&nbsp;<strong>data collection</strong>. They also said that this figure didn&#8217;t change much after weighting for talent, though in the future they expect junior people to take on more disproportionate shares of this workload.</p><p>Finally, one minor complaint that was mentioned was the lack of transparency about contractor compensation.&nbsp;</p><h3>How mission-aligned is Surge?</h3><p>Surge&nbsp;<a href="https://www.surgehq.ai/case-study/adversarial-testing-redwood-research">highlight</a> their collaboration with Redwood on their website as one of three case studies. In their blog&nbsp;<a href="https://www.surgehq.ai/blog/the-250k-inverse-scaling-prize-and-human-ai-alignment">post</a> about their collaboration with Anthropic, the first sentence reads: &#8220;In many ways, alignment &#8211; getting models to align themselves with what we want, not what they think we want &#8211; is one of the fundamental problems of AI.&#8221;&nbsp;</p><p>On the one hand, they describe alignment as one of the fundamental problems of AI, which could indicate that they intrinsically cared about alignment. However, they have a big commercial incentive to say this. Note that many people would consider their half-sentence definition of alignment to be wrong (a model might know what we want, but still do something else).</p><p>We suspect that the heads of Surge have at least vaguepositive dispositions towards alignment. They definitely seem eager to work with alignment researchers, which might well be more important. We think it&#8217;s mostly fine if they are not maximally intrinsically driven, though mission alignment does add value as mentioned above.</p><h2>Other competitors</h2><p>We see Surge as the most direct competitor and have researched them by far in the most detail. But besides Surge, there are a large number of other companies offering similar services.&nbsp;</p><p>First, and most obviously, Amazon&nbsp;<a href="https://www.mturk.com/">Mechanical Turk</a> offers a very low quality version of this service and is very large.&nbsp;<a href="https://www.upwork.com/">Upwork</a> specialises in sourcing humans for various tasks, without building interfaces.&nbsp;<a href="https://scale.com/">ScaleAI</a> is a startup with a $7B valuation --- they augment human feedback with various automated tools. OpenAI have worked with them. Other companies in this broad space include&nbsp;<a href="https://gethybrid.io/">Hybrid</a> (which Sam Bowman&#8217;s lab has worked with) and&nbsp;<a href="https://www.invisible.ai/">Invisible</a> (who have worked with OpenAI). There are many more that we haven&#8217;t listed here.</p><p>In addition, some labs have in-house teams for data gathering (see&nbsp;<a href="https://forum.effectivealtruism.org/posts/iBeWbfQLA9EKfsdhu/why-we-re-not-founding-a-human-data-for-alignment-org#Is_it_more_natural_for_this_work_to_be_done_in_house_in_the_longterm__Especially_at_big_labs_companies_">here</a> for more).</p><h2>Data providers used by other labs</h2><p>Ethan Perez&#8217;s and Sam Bowman&#8217;s labs at NYU/Anthropic have historically often built their own interfaces while using contractors from Upwork or undergrads, but they have been trialing Surge over the summer and seem likely to stick with them if they have a good experience. Judging from the Twitter thread linked above and asking J&#233;r&#233;my Scheurer (who works on the team and built the pre-Surge data pipeline) how they&#8217;ve found Surge so far, Surge is doing a good job.&nbsp;</p><p>Google has an internal team that provides a similar service, though DeepMind have used at least one external provider as well. We expect that it would be quite hard to get DeepMind to work with us, at least until we would be somewhat more established.&nbsp;</p><p>Generally, we get the impression that most people are quite happy with Surge. It&#8217;s worth also considering that it&#8217;s a young company that&#8217;s&nbsp;<strong>likely improving its service over time</strong>. We&#8217;ve heard that Surge iterates quickly, e.g. by shipping simple feature requests in two days. It&#8217;s possible that some of the problems listed above may no longer apply by now or in a few months.</p><h2>Good signs for demand</h2><p>One researcher we talked to said that there were lots of projects their team didn&#8217;t do, because gathering human feedback of sufficient quality was infeasible.&nbsp;</p><p>One of the examples this researcher gave was human feedback on code quality. This is implausible to do, because the time of software engineers is just too expensive. That problem is hard for a new org to solve.&nbsp;</p><p>Another example they gave seemed like it might be more feasible: for things like RLHF, they often choose to do pairwise comparisons between examples or multi-preferences. Ideally, they would want to get ratings, e.g. on a scale from 1 to 10. But they thought they didn&#8217;t trust the reliability of their raters enough to do this.&nbsp;</p><p>More generally, this researcher thought there were lots of examples where if they could copy any person on their team a hundred times to provide high-skill data, they could do many experiments that they currently can&#8217;t.&nbsp;</p><p>They also said that their team would be willing to pay ~3x of what they were paying currently to receive much higher-quality feedback.</p><p>Multiple other researchers we talked to expressed vaguely similar sentiments, though none quite as strong.</p><p>However, it&#8217;s notable that in this particular case, the researcher hadn&#8217;t worked with Surge yet.&nbsp;</p><p>The same researcher also told us about a recent project where they had spent a month on things like creating quality assurance examples, screening raters, tweaking instructions etc. They thought this could probably have been reduced a lot by an external org, maybe to as little as one day. Again, we think Surge may be able to get them a decent part of the way there.</p><h2>Labs we could have worked with</h2><p>We ended up finding three projects that we could have potentially worked on:</p><ul><li><p>A collaboration with Ought --- they spend about 15 hours a week on data-gathering and would have been happy to outsource that to us. If it had gone well, they might also have done more data-gathering in the longterm (since friction is lower if it doesn&#8217;t require staff time). We decided not to go ahead with this project since we weren&#8217;t optimistic enough about demand from other labs being bigger once we had established competence with Ought and the project itself didn&#8217;t seem high upside enough.&nbsp;</p></li><li><p>Attempt to get the Visible Thoughts&nbsp;<a href="https://intelligence.org/2021/11/29/visible-thoughts-project-and-bounty-announcement/">bounty</a> by MIRI. We decided against this for a number of reasons. See more of our thinking about Visible Thoughts below.</p></li><li><p>Potentially a collaboration with Owain Evans on curated datasets for alignment.</p></li></ul><p>We think the alignment community is currently relatively tight-knit. e.g. researchers often knew about other alignment teams&#8217; experiences with Surge from conversations they had had with them. Hence, we were relatively optimistic that conditional on there being significant demand for this kind of service, doing a good job on one of the projects above would quickly lead to more opportunities.<br>&nbsp;</p><h3>Visible Thoughts</h3><p>In November 2021,&nbsp;<a href="https://www.lesswrong.com/posts/zRn6cLtxyNodudzhw/visible-thoughts-project-and-bounty-announcement">MIRI announced the Visible Thoughts (VT) project bounty</a>. In many ways VT would be a good starting project for an alignment-oriented dataset provider, in particular because the bounty is large (up to $1.2M) and because it is ambitious enough that executing on it would provide a strong learning signal to us and a credible signal to other organisations we might want to work with. However, on closer examination of VT, we came to the conclusion that it is not worth it for us to work on it.</p><p>The idea of VT is to collect a dataset of 100 runs of fiction of a particular type (&#8220;dungeon runs&#8221;, an interactive text-based genre where one party, called the &#8220;dungeon master&#8221; and often an AI, offers descriptions of what is happening, and the other responds in natural language with what actions they want to take), annotated with a transcript of some of the key verbal thoughts that the dungeon master might be thinking as they decide what happens in the story world. MIRI hopes that this would be useful for training AI systems that make their thought processes legible and modifiable.</p><p>In particular, a notable feature of the VT bounty is the extreme run lengths that it asks for: to the tune of 300 000 words for each of the runs (for perspective, this is the length of&nbsp;<em>A Game of Thrones</em>, and longer than the first three&nbsp;<em>Harry Potter</em> books combined). A VT run is much less work than a comparable-length book - the equivalent of a rough unpolished first-draft (with some quality checks) would likely be sufficient - but producing one such run would still probably require at least on the order of 3 months of sequential work time from an author. We expect the pool of people willing to write such a story for 3 months is significantly smaller than the pool of people who would be willing to complete, say, a 30 000 word run, and that the high sequential time cost increases the amount of time required to generate the same number of total words. We also appear to have different ideas on how easy it is to fit a coherent story, for the relevant definition of coherent, into a given number of words. Note that to compare VT word counts to lengths of standard fiction without the written-out thoughts from the author, the VT word count should be reduced by a factor of 5-6.</p><p>Concerns about the length are raised in the comments section, to which Eliezer Yudkowksy&nbsp;<a href="https://www.lesswrong.com/posts/zRn6cLtxyNodudzhw/visible-thoughts-project-and-bounty-announcement?commentId=irJCDQaWRcdT3Bnoo">responded</a>. His first point, that longer is easier to write per step, may be true, especially as we also learned (by email with Nate Soares and Aurelien Cabanillas) that in MIRI&#8217;s experience &#8220;authors that are good at producing high quality steps are also the ones who don't mind producing many steps&#8221;. In particular because of that practical experience, we think it is possible we overestimated the logistical problems caused by the length. MIRI also said they would likely accept shorter runs too if they satisfied their other criteria.</p><p>In a brief informal conversation with Rudolf during EAG SF, Eliezer emphasised the long-range coherence point in particular. However, they did not come to a shared understanding of what type of &#8220;long-range coherence&#8221; is meant.</p><p>Even more than these considerations, we are sceptical about the vague plans for what to do given a VT dataset. A recurring theme from talking to alignment researchers who work with datasets was that inventing and creating a good dataset is surprisingly hard, and generally involves having a clear goal of what you&#8217;re going to use the dataset for. It is possible the key here is the difference in our priors for how likely a dataset idea is to be useful.</p><p>In addition, we have significant concerns about undertaking a major project based on a bounty whose only criterion is the judgement of one person (Eliezer Yudkowsky), and undertaking such a large project as our first project.</p><h1>Other cruxy considerations</h1><h2>Could we make a profit / get funding?&nbsp;</h2><p>One researcher from OpenAI told us he thought it would be hard to imagine an EA data-gathering company making a profit because costs for individual projects would always be quite high (requiring several full-time staff), and total demand was probably not all that big.</p><p>In terms of funding, both of us were able to spend time on this project because of grants from regrantors in the Future Fund regrantor program. Based on conversations with regrantors, we believe we could&#8217;ve gotten funding to carry out an initial project if we had so chosen.</p><h2>Will human feedback become a much bigger deal? Is this a very quickly growing industry?</h2><p>Our best guess is yes. For example, see this&nbsp;<a href="https://www.lesswrong.com/posts/pRkFkzwKZ2zfa3R6H/without-specific-countermeasures-the-easiest-path-to">post</a> by Ajeya Cotra which outlines how we could get to TAI by training on Human Feedback on Diverse Tasks (HFDT).&nbsp;</p><p>She writes: &#8220;HFDT is not the only approach to developing transformative AI, and it may not work at all. But I take it very seriously, and I&#8217;m aware of increasingly many executives and ML researchers at AI companies who believe something within this space could work soon.&#8221;</p><p>In addition, we have also had discussions with at least one other senior AI safety researcher whom we respect and who thought human feedback was currently irrationally neglected by mainstream ML; they expected it to become much more wide-spread and to be a very powerful tool.</p><p>If that&#8217;s right, then providing human feedback will likely become important and economically valuable.&nbsp;</p><p>This matters, because operating a new company in a growing industry is generally much easier and more likely to be successful. We think this is true even if profit isn&#8217;t the main objective.</p><h2>Would we be accelerating capabilities?</h2><p>Our main idea was to found a company (or possibly non-profit) that served alignment researchers exclusively. That could accelerate alignment differentially.&nbsp;</p><p>One problem is that it&#8217;s not clear where to draw this boundary. Some alignment researchers definitely think that other people who would also consider themselves to be alignment researchers are effectively doing capabilities work. This is particularly true of RLHF.</p><p>One mechanism worth taking seriously if we worked with big AI labs to make their models more aligned by providing higher quality data is that the models might merely appear surface-level aligned. &#8220;Make the data higher quality&#8221; might be a technique that scales poorly as capabilities ramp up. So it risks creating a false sense of security. It would also clearly improve the usefulness of current-day models and hence, it risks increasing investment levels too.</p><p>We don&#8217;t currently think the risk of surface-level alignment is big enough to outweigh the benefits. In general, we think that a good first-order heuristic that helps the field stay grounded in reality would be that whatever improves alignment in current models is useful to explore further and invest resources into. It seems like a good prior that such things would also be valuable in the future (even if it&#8217;s possible that new additional problems may arise, or such efforts aren&#8217;t on the path to a future alignment solution). See Nate Soares&#8217; post about&nbsp;<a href="https://www.lesswrong.com/posts/GNhMPAWcfBCASy8e6/a-central-ai-alignment-problem-capabilities-generalization">sharp left turns</a> to get a contradicting view on this.&nbsp;</p><h2>Is it more natural for this work to be done in-house in the longterm? Especially at big labs/companies.</h2><p>We expect that human data gathering is likely to become very important and that it benefits from understanding the relevant research agenda well. So maybe big companies will want to do this internally, instead of relying on third-party suppliers?&nbsp;</p><p>That seems quite plausible to us and to some extent it&#8217;s happening already. Our understanding is that Anthropic is hiring an internal team to do human data gathering. DeepMind has access to Google&#8217;s crowdworker service. OpenAI have worked with multiple companies, but they also have at least one in-house specialist for this kind of work and are advertising multiple further jobs on the human data team&nbsp;<a href="https://openai.com/careers/#human-data">here</a>. They&#8217;re definitely considering moving more of this work in-house, but it&#8217;s unclear to us to what extent that&#8217;s going to happen and we have received somewhat contradicting signals regarding OpenAI safety team members&#8217; preferences on this.</p><p>So a new EA org would face stiff competition, not only from other external providers, but also from within companies.</p><p>Of course, smaller labs will most likely always have to rely on external providers. Hence,&nbsp;<strong>another cruxy consideration is how much small labs matter</strong>. Our intuition is that they matter much less than bigger labs (since the latter have access to the best and biggest models).</p><h2>Creating redundancy of supply and competition</h2><p>Even if existing companies are doing a pretty good job at serving the needs of alignment researchers, there&#8217;s still some value in founding a competitor.&nbsp;</p><p>First,&nbsp;<strong>competition is good</strong>. Founding a competitor puts pressure on existing providers to keep service quality high, work on improving their products, and margins low. Ironically, part of the value of founding this company would thus flow through getting existing companies to try harder to offer the best product.</p><p>Second, it creates some redundancy.&nbsp;<strong>What if Surge pivots?</strong> What if their leadership changes or they become less useful for some other reason? In those worlds it might be especially useful to have a &#8220;back-up&#8221; company.</p><p>Both of these points have been mentioned to us as arguments in favour of founding this org. We agree that these effects are real and likely point in favour of founding the org. However,&nbsp;<strong>we don&#8217;t think these factors carry very significant weight</strong> relative to our opportunity costs, especially given that there are already many start-ups working in this space.&nbsp;</p><p>Adding a marginal competitor can only affect a company&#8217;s incentives so much. And in the worlds where we&#8217;d be most successful such that all alignment researchers were working with us, we might cause Surge and others to pivot away from alignment researchers, instead of getting them to try harder.&nbsp;</p><p>The redundancy argument only applies in worlds in which the best provider ceases to exist; maybe that&#8217;s 10% likely. And then the next best alternative is likely not all that bad. Competitors are plentiful and even doing it in-house is feasible. Hence, it seems unlikely to us that the expected benefit here is very large after factoring in the low probability of the best provider disappearing.</p><h1>Other lessons</h1><h2>Lessons on human data gathering</h2><p>In the process of talking to lots of experts about their experiences in working with human data, we learned many general lessons about data gathering. This section presents some of those lessons, in roughly decreasing order of importance.</p><h3>Iteration</h3><p>Many people emphasized to us that working with human data rarely looks like having a clean pipeline from requirements design to instruction writing to contractor finding to finished product. Rather, it more often involves a lot of iteration and testing, especially regarding what sort of data the contractors actually produce. While some of this iteration may be removed by having better contractors and better knowledge of good instruction-writing, the researchers generally view the iteration as a key part of the research process, and therefore prize&nbsp;</p><ul><li><p>ease of iteration (especially time to get back with a new batch of data based on updated instructions); and</p></li><li><p>high-bandwidth communication with the contractors and whoever is writing the instructions (often both are done by the researchers themselves).&nbsp;</p></li></ul><p>This last point holds to the point that it is somewhat questionable whether an external provider (rather than e.g. a new team member deeply enmeshed in the context of the research project) could even be a good fit for this need.</p><h3>The ideal pool of contractors</h3><p>All of the following features matter in a pool of contractors:</p><ul><li><p>Competence, carefulness, intelligence, etc. (sometimes expertise). It is often ideal if the contractors understand the experiment.</p></li><li><p>Number of contractors</p></li><li><p>Quick availability and therefore low latency for fulfilling requests</p></li><li><p>Consistent availability (ideally full-time)</p></li><li><p>Even distribution of contributions across contractors (ie it shouldn&#8217;t be the case that 20% of the contractors provide 80% of the examples).&nbsp;</p></li></ul><h3>Quality often beats quantity for alignment research</h3><p>Many researchers told us that high-quality, high-skill data is usually more important and more of a bottleneck than just a high quantity of data. Some of the types of projects where current human data generation methods are most obviously deficient are cases where a dataset would need epistemically-competent people to make subtle judgments, e.g. of the form &#8220;how true is this statement?&#8221; or &#8220;how well-constructed was this study?&#8221; As an indication of reference classes where the necessary epistemic level exists, the researcher mentioned subject-matter experts in their domain, LessWrong posters, and EAs.</p><h3>A typical data gathering project needs UX-design, Ops, ML, and data science expertise&nbsp;</h3><p>These specialists might respectively focus on the following:</p><ul><li><p>Designing the interfaces that crowdworkers interact with. (UX-expert/front-end web developer)</p></li><li><p>Managing all operations, including hiring, paying, managing, and firing contractors, communicating with them and the researchers etc. (ops expert)</p></li><li><p>Helping the team make informed decisions about the details of the experimental design, while minimizing time costs for the customer. The people we spoke to usually emphasized ML-expertise more than alignment expertise. (ML-expert)</p></li><li><p>Meta-analysis of the data. e.g. inter-rater agreement, the distribution of how much each contractor contributed, demographics, noticing any other curious aspects of the data, etc. (data scientist)</p></li></ul><p>It is possible that someone in a team could have expertise in more than one of these areas, but generally this means a typical project will involve at least three people.</p><h3>Crowdworkers do not have very attractive jobs</h3><p>Usually the crowdworkers are employed as contractors. This means their jobs are inherently not maximally attractive; they probably don&#8217;t offer much in the way of healthcare, employment benefits, job security, status etc. The main way that these jobs are made more attractive is through offering higher hourly rates.</p><p>If very high quality on high-skill data is going to become essential for alignment, it may be worth considering changing this, to attract more talented people.&nbsp;</p><p>However, we expect that it might be inherently very hard to offer permanent positions for this kind of work, since demand is likely variable and since different people may be valuable for different projects. This is especially true for a small organisation.&nbsp;</p><h3>What does the typical crowdworker look like?</h3><p>This varies a lot between projects and providers.</p><p>The cheapest are non-native English speakers who live outside of the US.</p><p>Some platforms, including Surge, offer the option to filter crowdworkers for things like being native English-speakers, expertise as a software engineer, background in finance, etc.</p><h2>Bottlenecks in alignment</h2><p>When asked to name the factors most holding back their progress on alignment, many alignment researchers mentioned talent bottlenecks.&nbsp;</p><p>The most common talent bottleneck seemed to be in competent ML-knowledgeable people. Some people mentioned the additional desire for these to understand and care about alignment. (Not coincidentally, Matt&#8217;s next project is likely going to be about skilling people up in ML).</p><p>There were also several comments about things like good web development experience being important. For example, many data collection projects involve creating a user interface at some point, and in practice this is often handled by ML-specialised junior people at the lab, who can, with some effort and given their programming background, cobble together some type of website - often using different frameworks and libraries than the next person knows (or wants to use). (When asked about why they don&#8217;t hire freelance programmers, one researcher commented that a key feature they&#8217;d want is the same person working for them for a year or two, so that there&#8217;s an established working relationship, clear quality assurances, and continuity with the choice of technical stack.)</p><h1>Conclusion</h1><p>After having looked into this project idea for about a month, we have decided not to found a human data gathering organisation for now.&nbsp;</p><p>This is mostly because demand for an external provider seems insufficient, as outlined in this&nbsp;<a href="https://forum.effectivealtruism.org/posts/iBeWbfQLA9EKfsdhu/why-we-re-not-founding-a-human-data-for-alignment-org#Key_crux__demand_looks_questionable__Surge_seems_pretty_good">section</a>. No lab gave a clear signal that gathering human data was a key bottleneck for them, where they would have been willing to go to significant lengths to fix it urgently (especially not the ones that had tried Surge).&nbsp;</p><p>We expect that many labs would want to stick with their current providers, Surge in particular, or their in-house team, bar exceptional success on our part (even then, we&#8217;d only provide so much marginal value over those alternatives).</p><p>Though we did find some opportunities for potential initial projects after looking for a month, we are hesitant about how far this company would be expected to scale. One of the main draws (from an impact perspective) of founding an organisation is that you can potentially achieve very high counterfactual impact by creating an organisation that scales to a large size and does lots of high-impact work over its existence. The absence of a plausible pathway to really outstanding outcomes from starting this organisation is a lot of what deters us.</p><p>In a world where we&#8217;re more successful than expected (say 90th to 95th percentile), we could imagine that in five years from now, we&#8217;d have a team of about ten good people. This team may be working with a handful of moderately big projects (about as big as WebGPT), and provide non-trivial marginal value over the next-best alternative to each one of them. Maybe one of these projects would not have been carried out without us.</p><p>A median outcome might mean failing to make great hires and remaining relatively small and insignificant: on the scale of doing projects like the ones we&#8217;ve identified above, enough to keep us busy throughout the year and provide some value, but with little scaling. In that case we would probably quit the project at some point.</p><p>This distribution doesn&#8217;t seem good enough to justify our opportunity cost (which includes other entrepreneurial projects or technical work among other things). Thus we have decided not to pursue this project any further for now.</p><p>We think this was a good idea to invest effort in pursuing, and we think we made the right call in choosing to investigate it. Both of us are open to, and also quite likely to, evaluate other EA-relevant entrepreneurial project ideas in the future.</p><h2>Other relevant human data-gathering work</h2><p>However,&nbsp;<strong>the assumption that high-quality high-skill human feedback is important and neglected by EAs has not been falsified</strong>.&nbsp;</p><p>It is still plausible to us that EAs should consider career paths that focus on building expertise at data-gathering; just probably not by founding a new company. In the short run, this could look like</p><ul><li><p>Contributing to&nbsp;<strong>in-house data-gathering teams</strong> (eg Anthropic, OpenAI, etc.)</p></li><li><p><strong>Joining Surge</strong> or other data-gathering startups.</p></li></ul><p>As we discussed above, the types of skills that seem most relevant for working in a human data generation role include: data science experience and in particular experience with natural languaga data or social science data and experiment design, front-end web development, ops and management skills, and some understanding of machine learning and alignment. 80,000 Hours recently wrote a profile which you can find&nbsp;<a href="https://80000hours.org/career-reviews/alignment-data-expert/">here</a>.</p><p>Of course, in the short term, this career path will be especially impactful if one&#8217;s efforts are focussed on helping alignment researchers. But if it&#8217;s true that human feedback will prove a very powerful tool for ML, then people with such expertise may become increasingly valuable going forward, such that it could easily be worth skilling up at a non-safety-focused org.&nbsp;</p><p>We think joining Surge may be a particularly great opportunity. It is common advice that joining young, rapidly growing start-ups with good execution is great for building experience; early employees can often get a lot of responsibility early on. See e.g. this&nbsp;<a href="https://forum.effectivealtruism.org/posts/ejaC35E5qyKEkAWn2/early-career-ea-s-should-consider-joining-fast-growing">post</a> by Bill Zito.</p><p>One of the hardest parts about that seems to be identifying promising startups. After talking to many of their customers, we have built reasonable confidence that Surge holds significant promise. They seem to execute well, in a space which we expect to grow. In addition to building career capital, there is clear value in helping Surge serve alignment researchers as well as possible.</p><p>From Surge&#8217;s perspective, we think they could greatly benefit from hiring EAs, who are tuned in to the AI safety scene, which we would guess represents a significant fraction of their customers.&nbsp;</p><p>One senior alignment researcher told us explicitly that they would be interested in hiring people who had worked in a senior role at Surge.</p><h1>Next steps for us</h1><p>Matt is planning to run a bootcamp that will allow EAs to upskill in ML engineering. I'll be doing a computer science master&#8217;s at Cambridge from October to June.</p>]]></content:encoded></item><item><title><![CDATA[AI risk intro 2: solving the problem]]></title><description><![CDATA[This post was a joint effort with Callum McDougall.]]></description><link>https://www.nosetgauge.com/p/ai-risk-intro-2-solving-problem</link><guid isPermaLink="false">https://www.nosetgauge.com/p/ai-risk-intro-2-solving-problem</guid><dc:creator><![CDATA[Rudolf Laine]]></dc:creator><pubDate>Sat, 24 Sep 2022 09:43:00 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/33d60135-6603-424e-b57b-044706d790a0_400x400.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>&nbsp;<em><strong>This post was a joint effort with <a href="https://www.perfectlynormal.co.uk/">Callum McDougall</a>.</strong></em></p><p><em>8.2k words (~25min)&nbsp;</em></p><p>This marks the second half of our overview of the AI alignment problem. In <a href="https://www.strataoftheworld.com/2022/09/ai-risk-intro-1-advanced-ai-might-be.html">the first half</a>, we outlined the case for misaligned AI as a significant risk to humanity, first by looking at past progress in machine learning and extrapolating to what the future could bring, and second by discussing the theoretical arguments which underpin many of these concerns. In this second half, we focus on possible solutions to the alignment problem that people are currently working on. We will paint a picture of the current field of technical AI alignment, explaining where the major organisations fit into the larger picture and what the theory of change behind their work is. Finally, we will conclude the sequence with a call to action, by discussing the case for working on AI alignment, and some suggestions on how you can get started.</p><p><em>Note - for people with more context about the field (e.g. have done AGISF) we expect <a href="https://www.lesswrong.com/posts/QBAjndPuFbhEXKcCr/my-understanding-of-what-everyone-in-technical-alignment-is">Thomas Larsen's post</a> to be a much better summary, and <a href="https://www.lesswrong.com/posts/9TWReSDKyshfA66sz/alignment-org-cheat-sheet#comments">this post</a> might be better if you are looking for something brief. Our intended audience is someone relatively unfamiliar with the AI safety field, and is looking for a taste of the kinds of problems which are studied in the field and the solution approaches taken. We also don't expect this sampling to be representative of the number of people working on each problem - again, see Thomas' post for something which accomplishes this.</em></p><div><hr></div><h1><strong>Introduction: A Pre-Paradigmatic Field</strong></h1><blockquote><p><em>Definition (<strong>pre-paradigmatic</strong>): a science at an early stage of development, before it has established a consensus about the true nature of the subject matter and how to approach it.</em></p></blockquote><p>AI alignment is a strange field. Unlike other fields which study potential risks to the future of humanity (e.g. nuclear war or climate change), there is almost no precedent for the kinds of risks we care about. Additionally, because of the nature of the threat, failing to get alignment right on the first try might be fatal. As Paul Christiano (a well-known AI safety researcher) recently wrote:</p><blockquote><p><em>Humanity usually solves technical problems by&nbsp;<strong>iterating and fixing failures</strong>; we often resolve tough methodological disagreements very slowly by seeing what actually works and having our failures thrown in our face. But it will probably be possible to build valuable AI products without solving alignment, and so&nbsp;<strong>reality won&#8217;t &#8220;force us&#8221; to solve alignment until it&#8217;s too late</strong>. This seems like a case where we will have to be&nbsp;<strong>unusually reliant on careful reasoning rather than empirical feedback loops</strong> for some of the highest-level questions.</em></p></blockquote><p>For these reasons, the field of AI alignment lacks a consensus on how the problem should be tackled, or what the most important parts of the problem even are. This is why there is a lot of variety in the approaches we present in this post.</p><h1><strong>Decomposing the research landscape</strong></h1><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!cZOX!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb8ca88a-2e16-474c-8410-2d31c9e36bdc_400x400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!cZOX!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb8ca88a-2e16-474c-8410-2d31c9e36bdc_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!cZOX!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb8ca88a-2e16-474c-8410-2d31c9e36bdc_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!cZOX!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb8ca88a-2e16-474c-8410-2d31c9e36bdc_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!cZOX!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb8ca88a-2e16-474c-8410-2d31c9e36bdc_400x400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!cZOX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb8ca88a-2e16-474c-8410-2d31c9e36bdc_400x400.png" width="400" height="400" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/db8ca88a-2e16-474c-8410-2d31c9e36bdc_400x400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:400,&quot;width&quot;:400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!cZOX!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb8ca88a-2e16-474c-8410-2d31c9e36bdc_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!cZOX!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb8ca88a-2e16-474c-8410-2d31c9e36bdc_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!cZOX!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb8ca88a-2e16-474c-8410-2d31c9e36bdc_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!cZOX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb8ca88a-2e16-474c-8410-2d31c9e36bdc_400x400.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>An image generated with OpenAI's DALL-E 2 based on the prompt: sorting papers and books in a majestic gothic library. <strong>All other images like this in this post are also AI-generated, from the text in the caption.</strong></em></figcaption></figure></div><p>There are lots of different ways you could divide up the space of approaches to solving the problem of aligning advanced AI. For instance, you could go through the history of the field and identify different movements and paradigms. Or you could place the work on a spectrum from highly theoretical maths/philosophy-type research, to highly empirical research working with cutting-edge deep learning models.</p><p>However, the most useful decomposition would be one that explains why the people who work on it believe that it will help solve the problem of AI alignment.&nbsp;</p><p>For that reason, we&#8217;ll mostly be using the decomposition from&nbsp;<a href="https://www.lesswrong.com/s/FN5Gj4JM6Xr7F4vts/p/SQ9cZtfrzDJmw9A2m">Neel Nanda&#8217;s &#8220;A Bird&#8217;s Eye View&#8221;&nbsp;</a>post. The motivation behind this decomposition is to answer the high-level question of &#8220;what is needed for AGI to go well?&#8221;. The six broad classes of approaches we talk about are:</p><ol><li><p><strong>Addressing threat models&nbsp;</strong><br><em>We have a specific threat model in mind for how AGI might result in a very bad future for humanity, and focus our work on things we expect to help address the threat model.</em></p></li><li><p><strong>Agendas to build safe AGI&nbsp;</strong><br><em>Let&#8217;s make specific plans for how to actually build safe AGI, and then try to test, implement, and understand the limitations of these plans. The emphasis is on understanding how to build AGI safely, rather than trying to do it as fast as possible.</em></p></li><li><p><strong>Robustly good approaches&nbsp;</strong><br><em>In the long-run AGI will clearly be important, but we're highly uncertain about how we'll get there and what, exactly, could go wrong. So let's do work that seems good in many possible scenarios, and doesn&#8217;t rely on having a specific story in mind.</em></p></li><li><p><strong>Deconfusion</strong><br><em>Reasoning about how to align AGI involves reasoning concepts like intelligence, values, and optimisers and we&#8217;re pretty confused about what these even mean. This means any work we do now is plausibly not helpful and definitely not reliable. As such, our priority should be doing some conceptual work on how to think about these concepts and what we&#8217;re aiming for, and trying to become less confused.</em></p></li><li><p><strong>AI governance</strong><br><em>In addition to solving the technical alignment problem, there&#8217;s the question of what policies we need to minimise risk from advanced AI systems.</em></p></li><li><p><strong>Field-building</strong><br><em>One of the most important ways we can make AI go well is by increasing the number of capable researchers doing alignment research.</em>&nbsp;</p></li></ol><p>It&#8217;s worth noting that there is a lot of overlap between these sections. For instance, interpretability research is a great example of a robustly good approach, but it can also be done with a specific threat model in mind.</p><p>Throughout this section, we will also give small vignettes of organisations or initiatives which support AI alignment research in some form. This won&#8217;t be a full picture of all approaches or organisations, instead hopefully it will serve to sketch a picture of what work in AI alignment actually looks like.</p><h2>Addressing threat models</h2><blockquote><p><em>We have a <strong>specific threat model</strong> in mind for how AGI might result in a very bad future for humanity, and focus our work on things we expect to help address the threat model.</em></p></blockquote><p>A key high-level intuition here is that having a specific threat model in mind for how AI might go badly for humanity can help keep you focused on certain hard parts of the problem. One technique that can be useful here is a version of back-casting: we start from future problems with advanced AI systems in our current model, reason about what kinds of things might solve these problems, then try and build versions of these solutions today and test them out on current problems.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!iCLt!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb671929-8f02-4c4c-923e-bb9f599c1127_400x212.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!iCLt!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb671929-8f02-4c4c-923e-bb9f599c1127_400x212.png 424w, https://substackcdn.com/image/fetch/$s_!iCLt!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb671929-8f02-4c4c-923e-bb9f599c1127_400x212.png 848w, https://substackcdn.com/image/fetch/$s_!iCLt!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb671929-8f02-4c4c-923e-bb9f599c1127_400x212.png 1272w, https://substackcdn.com/image/fetch/$s_!iCLt!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb671929-8f02-4c4c-923e-bb9f599c1127_400x212.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!iCLt!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb671929-8f02-4c4c-923e-bb9f599c1127_400x212.png" width="400" height="212" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/fb671929-8f02-4c4c-923e-bb9f599c1127_400x212.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:212,&quot;width&quot;:400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!iCLt!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb671929-8f02-4c4c-923e-bb9f599c1127_400x212.png 424w, https://substackcdn.com/image/fetch/$s_!iCLt!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb671929-8f02-4c4c-923e-bb9f599c1127_400x212.png 848w, https://substackcdn.com/image/fetch/$s_!iCLt!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb671929-8f02-4c4c-923e-bb9f599c1127_400x212.png 1272w, https://substackcdn.com/image/fetch/$s_!iCLt!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffb671929-8f02-4c4c-923e-bb9f599c1127_400x212.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>This can be seen in contrast to the approach of simply trying to fix current problems with AI systems, which might fail to connect up with the hardest parts of AI alignment.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!9EMf!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f545f5a-ac50-4201-b810-773892a60a69_400x226.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!9EMf!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f545f5a-ac50-4201-b810-773892a60a69_400x226.png 424w, https://substackcdn.com/image/fetch/$s_!9EMf!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f545f5a-ac50-4201-b810-773892a60a69_400x226.png 848w, https://substackcdn.com/image/fetch/$s_!9EMf!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f545f5a-ac50-4201-b810-773892a60a69_400x226.png 1272w, https://substackcdn.com/image/fetch/$s_!9EMf!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f545f5a-ac50-4201-b810-773892a60a69_400x226.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!9EMf!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f545f5a-ac50-4201-b810-773892a60a69_400x226.png" width="400" height="226" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1f545f5a-ac50-4201-b810-773892a60a69_400x226.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:226,&quot;width&quot;:400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!9EMf!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f545f5a-ac50-4201-b810-773892a60a69_400x226.png 424w, https://substackcdn.com/image/fetch/$s_!9EMf!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f545f5a-ac50-4201-b810-773892a60a69_400x226.png 848w, https://substackcdn.com/image/fetch/$s_!9EMf!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f545f5a-ac50-4201-b810-773892a60a69_400x226.png 1272w, https://substackcdn.com/image/fetch/$s_!9EMf!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f545f5a-ac50-4201-b810-773892a60a69_400x226.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><h3>Example 1: Superintelligent utility maximisers, and quantilizers</h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!1dDz!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7af2216-b374-43f6-8d98-e43550b4369c_400x400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!1dDz!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7af2216-b374-43f6-8d98-e43550b4369c_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!1dDz!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7af2216-b374-43f6-8d98-e43550b4369c_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!1dDz!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7af2216-b374-43f6-8d98-e43550b4369c_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!1dDz!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7af2216-b374-43f6-8d98-e43550b4369c_400x400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!1dDz!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7af2216-b374-43f6-8d98-e43550b4369c_400x400.png" width="400" height="400" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d7af2216-b374-43f6-8d98-e43550b4369c_400x400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:400,&quot;width&quot;:400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!1dDz!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7af2216-b374-43f6-8d98-e43550b4369c_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!1dDz!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7af2216-b374-43f6-8d98-e43550b4369c_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!1dDz!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7af2216-b374-43f6-8d98-e43550b4369c_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!1dDz!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7af2216-b374-43f6-8d98-e43550b4369c_400x400.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>superintelligent artificial intelligence, making choices, digital art, artstation</em></figcaption></figure></div><p>The superintelligent utility maximiser is the oldest threat model studied by the AI alignment field. It was discussed at length by Nick Bostrom in his book&nbsp;<em>Superintelligence</em>. It assumes that we will create an AGI much more intelligent than humans, and that it will be trying to achieve some particular goal (measured by the&nbsp;<a href="https://www.investopedia.com/terms/e/expectedutility.asp">expected value of some utility function</a>). The problem with this is that attempts to maximise the value of some goal which isn&#8217;t perfectly aligned with what humans want can lead to some very bad outcomes. One formalism which was proposed to address this problem is&nbsp;<a href="https://intelligence.org/2015/11/29/new-paper-quantilizers/">Jessica Taylor&#8217;s quantilizers</a>. It is quite maths-heavy so we won&#8217;t discuss all the details here, but the basic idea is that rather than using the expected utility maximisation framework for agents, we mix expected utility maximisation with human imitation in a clever way (to be more precise, you sample from a prior distribution which represents the actions a human would be likely to take in this scenario). The resulting agent wouldn&#8217;t take catastrophic actions because part of its decision-making comes from imitating what it thinks humans would do, but it would also be able to use the expected utility maximisation to go beyond human imitation, and do things we are incapable of (which is presumably the reason we would want to build it in the first place!). However, the drawback with theoretical approaches like this is that they often bake in too many assumptions or rely on too many variables to be useful in practice. In this case, how we define the set of reasonable actions a human might perform is an important unspecified part of this framework, and so more research is required to see if the quantiliszers framework can address these problems.</p><h3>Example 2: Inner misalignment</h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!VErD!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5aef5fdb-6472-4b1b-bc59-af6c4cc1475b_400x400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!VErD!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5aef5fdb-6472-4b1b-bc59-af6c4cc1475b_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!VErD!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5aef5fdb-6472-4b1b-bc59-af6c4cc1475b_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!VErD!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5aef5fdb-6472-4b1b-bc59-af6c4cc1475b_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!VErD!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5aef5fdb-6472-4b1b-bc59-af6c4cc1475b_400x400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!VErD!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5aef5fdb-6472-4b1b-bc59-af6c4cc1475b_400x400.png" width="400" height="400" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5aef5fdb-6472-4b1b-bc59-af6c4cc1475b_400x400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:400,&quot;width&quot;:400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!VErD!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5aef5fdb-6472-4b1b-bc59-af6c4cc1475b_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!VErD!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5aef5fdb-6472-4b1b-bc59-af6c4cc1475b_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!VErD!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5aef5fdb-6472-4b1b-bc59-af6c4cc1475b_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!VErD!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5aef5fdb-6472-4b1b-bc59-af6c4cc1475b_400x400.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>robot jumping over boxes to collect a coin, videogame, digital art, artstation</em></figcaption></figure></div><p>We&#8217;ve discussed inner misalignment in a previous section. This concept was first explicitly named in a paper called&nbsp;<a href="https://arxiv.org/abs/1906.01820">Risks from Learned Optimisation in Advanced ML Systems</a>, published in 2019. This paper defined the concept and suggested some conditions which might make it more likely to happen, but the truth is that a lot of this is still just conjecture, and there are many things we don&#8217;t yet know about how unlikely this kind of misalignment is, or what we can do about it. The CoinRun example discussed earlier (and the&nbsp;<a href="https://www.deepmind.com/publications/objective-robustness-in-deep-reinforcement-learning">Objective Robustness</a> paper) came from an independent research team in 2021. This study was the first known example of inner misalignment in an AI system, showing that it was at least a theoretical possibility. They also tested certain interpretability tools on the CoinRun agent, to see whether it was possible to discover when the agent had a goal different to the one intended by the programmers. For more on interpretability, see later sections.</p><h2>Building safe AGI</h2><blockquote><p><em>Let&#8217;s make specific plans for <strong>how to actually build safe AGI</strong>, and then try to test, implement, and understand the limitations of these plans. The emphasis is on understanding how to build AGI <strong>safely</strong>, rather than trying to do it as fast as possible.</em></p></blockquote><p>At some point we&#8217;re going to build an AGI. Companies are already racing to do it. We better make sure that there exist some blueprints for a safe AGI (and that they&#8217;re used) by the time we get to that point.</p><p>Perhaps the master list of safe AGI proposals is Evan Hubinger&#8217;s&nbsp;<a href="https://arxiv.org/pdf/2012.07532.pdf">An Overview of 11 Proposals for Building Safe Advanced AI</a>.&nbsp;</p><h3>Example 1: Iterated Distillation and Amplification (IDA)</h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!5SmC!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b211987-175f-4431-bd08-600db9112b71_400x400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!5SmC!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b211987-175f-4431-bd08-600db9112b71_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!5SmC!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b211987-175f-4431-bd08-600db9112b71_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!5SmC!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b211987-175f-4431-bd08-600db9112b71_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!5SmC!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b211987-175f-4431-bd08-600db9112b71_400x400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!5SmC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b211987-175f-4431-bd08-600db9112b71_400x400.png" width="400" height="400" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2b211987-175f-4431-bd08-600db9112b71_400x400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:400,&quot;width&quot;:400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!5SmC!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b211987-175f-4431-bd08-600db9112b71_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!5SmC!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b211987-175f-4431-bd08-600db9112b71_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!5SmC!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b211987-175f-4431-bd08-600db9112b71_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!5SmC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2b211987-175f-4431-bd08-600db9112b71_400x400.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>artists depection of a robot dreaming up multiple copies of itself, cascading tree, delegating, digital art, trending on artstation</em></figcaption></figure></div><p>&#8220;Iterated Distillation and Amplification&#8221; (IDA) is an imposing name, but the core intuition is simple. One of the ways in which an individual human can achieve more things is by delegating tasks to others. In turn, the assistants that tasks are delegated to can be expected to become more competent at the task.</p><p>In IDA, an AI plays the role of the assistant. &#8220;Distillation&#8221; refers to the abilities of the human being &#8220;distilled&#8221; into the AI through training, and &#8220;amplification&#8221; refers to the human becoming more capable as they can call on more and more powerful AI assistants to help them.</p><p>A setup to train an IDA personal assistant might go like this:</p><ol><li><p>You have a human, say Hannah, who knows how to carry out the tasks of a personal assistant.</p></li><li><p>You have an ML model - call it Martin - that starts out knowing very little (perhaps nothing at all, or perhaps it&#8217;s a pre-trained language model so it knows how to read and write English but not much else).</p></li><li><p>Hannah needs to find the answer to some questions, and she can invoke multiple copies of Martin to help her. Since Martin is quite useless at this stage, Hannah has to do even simple tasks herself, like writing routine emails. Using some interface legible to Martin, she breaks the email-writing task into subtasks like &#8220;find email address of Hu M. Anderson&#8221;, &#8220;select greeting&#8221;, &#8220;check project status&#8221;, &#8220;mention project status&#8221;, and so on.</p></li><li><p>From seeing enough examples of Hannah&#8217;s own answers to the sub-questions, Martin&#8217;s training loop gradually trains it to be able to answer first the simpler sub-tasks - (address is &#8220;humanderson@humanmail.com&#8221;, greeting is &#8220;Salutations, Human Colleague!&#8221;, etc.) and eventually all the sub-tasks involved in routine email-writing.</p></li><li><p>At this point, &#8220;write a routine email&#8221; becomes a task Martin can entirely carry out for Hannah. This is now a building block that can be used as a subtask in broader tasks Hannah gives out to Martin.&nbsp; Once enough tasks become tasks that Martin can carry out by itself, Hannah can draft much larger goals, like &#8220;invade France&#8221;, and let Martin take care of details like &#8220;blackmail Emmanuel Macron&#8221;, &#8220;write battle plan for the French Alps&#8221;, and &#8220;select a suitable coronation dress&#8221;.</p></li></ol><p>Note some features of this process. First, Martin learns what it should do and how to do it at the same time. Second, both Hannah&#8217;s and Martin&#8217;s role changes throughout this process - Martin goes from bumbling idiot who can&#8217;t write an email greeting to competent assistant, while Hannah goes from being a demonstrator of simple tasks to a manager of Martin to ruler of France. Third, note the recursive nature here: Hannah breaks down big tasks into small ones to train Martin on successively bigger tasks.&nbsp;</p><p>In fact, assuming perfect training, IDA imitates a recursive structure. When Hannah has only bumbling fool Martin to help her, Martin can only learn to become as good as Hannah herself. But once Martin is that good, Hannah&#8217;s position is now essentially that of having herself, but also some number - say 3 - copies of Martin that are as good as herself. We might call this structure &#8220;Hannah Consulting Hannah &amp; Hannah&#8221;; presumably, being able to consult an assistant that has the same skills as her lets Hannah become more effective, so this is an improvement. But now Hannah is demonstrating the behaviour of Hannah Consulting Hannah &amp; Hannah, so from Hannah&#8217;s example Martin can now learn to be as good as Hannah Consulting Hannah &amp; Hannah - making Hannah as good as Hannah Consulting (Hannah Consulting Hannah &amp; Hannah) &amp; (Hannah Consulting Hannah &amp; Hannah). And so on:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!2q1H!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3487b460-6516-4bfd-a6c9-02e34e1685e5_386x400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!2q1H!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3487b460-6516-4bfd-a6c9-02e34e1685e5_386x400.png 424w, https://substackcdn.com/image/fetch/$s_!2q1H!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3487b460-6516-4bfd-a6c9-02e34e1685e5_386x400.png 848w, https://substackcdn.com/image/fetch/$s_!2q1H!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3487b460-6516-4bfd-a6c9-02e34e1685e5_386x400.png 1272w, https://substackcdn.com/image/fetch/$s_!2q1H!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3487b460-6516-4bfd-a6c9-02e34e1685e5_386x400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!2q1H!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3487b460-6516-4bfd-a6c9-02e34e1685e5_386x400.png" width="386" height="400" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3487b460-6516-4bfd-a6c9-02e34e1685e5_386x400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:400,&quot;width&quot;:386,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!2q1H!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3487b460-6516-4bfd-a6c9-02e34e1685e5_386x400.png 424w, https://substackcdn.com/image/fetch/$s_!2q1H!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3487b460-6516-4bfd-a6c9-02e34e1685e5_386x400.png 848w, https://substackcdn.com/image/fetch/$s_!2q1H!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3487b460-6516-4bfd-a6c9-02e34e1685e5_386x400.png 1272w, https://substackcdn.com/image/fetch/$s_!2q1H!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3487b460-6516-4bfd-a6c9-02e34e1685e5_386x400.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>If everything is perfect, therefore, IDA imitates a structure called &#8220;HCH&#8221;, which is a recursive acronym for &#8220;Humans Consulting HCH&#8221;. Others call it the &#8220;<a href="https://www.lesswrong.com/posts/tmuFmHuyb4eWmPXz8/rant-on-problem-factorization-for-alignment">Infinite Bureaucracy</a>&#8221; (and fret about whether it&#8217;s actually a good idea).</p><p>Now &#8220;Infinite Bureaucracy&#8221; is not a name that screams &#8220;new sexy machine learning concept&#8221;. However, it&#8217;s interesting to think about what properties it might have. Imagine that you had, say, a 10-minute time limit to answer a complicated question, but you were allowed to consult three copies of yourself by passing a question off to them and getting back an answer immediately. These three copies also obeyed the same rules. Could you, for example, plan your career? Program an app? Write a novel?</p><p>It&#8217;s also interesting to think of the ways why the limitations of machine learning mean that IDA might not approximate HCH.</p><h3>Example 2: AI safety via debate</h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!iqE7!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd108dde7-db38-4226-b769-b6895fdf9306_400x400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!iqE7!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd108dde7-db38-4226-b769-b6895fdf9306_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!iqE7!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd108dde7-db38-4226-b769-b6895fdf9306_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!iqE7!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd108dde7-db38-4226-b769-b6895fdf9306_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!iqE7!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd108dde7-db38-4226-b769-b6895fdf9306_400x400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!iqE7!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd108dde7-db38-4226-b769-b6895fdf9306_400x400.png" width="400" height="400" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d108dde7-db38-4226-b769-b6895fdf9306_400x400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:400,&quot;width&quot;:400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!iqE7!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd108dde7-db38-4226-b769-b6895fdf9306_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!iqE7!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd108dde7-db38-4226-b769-b6895fdf9306_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!iqE7!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd108dde7-db38-4226-b769-b6895fdf9306_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!iqE7!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd108dde7-db38-4226-b769-b6895fdf9306_400x400.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>artists depiction of two robots debating, digital art, trending on artstation</em></figcaption></figure></div><p>Imagine you&#8217;re a bit drunk, but (as one does) you&#8217;re at a bar talking about AI alignment proposals. Someone&#8217;s talking about how even if you can get an advanced AI system to explain its reasoning to you, it might try to slip something very subtle past you and you might not notice. You might well blurt out: &#8220;well then just make it fight another AI over it!&#8221;</p><p>The OpenAI safety team presumably spends a fair amount of time at bars, because they&#8217;ve&nbsp;<a href="https://openai.com/blog/debate/">investigated the idea of achieving safe AI by having two AIs debate each other</a> to persuade a panel of human judges, by trying to poke holes in each other&#8217;s arguments. For more complex tasks, the AIs could be given transparency tools deriving from interpretability research (see next section) that they can use on each other. Just like a Go-playing AI gets an unambiguous win-loss signal from either winning or losing, a debating AI gets an unambiguous win-loss signal from winning or losing the debate:</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!wge3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9782253a-8c1f-44fc-a104-15c63d871f1e_400x143.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!wge3!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9782253a-8c1f-44fc-a104-15c63d871f1e_400x143.png 424w, https://substackcdn.com/image/fetch/$s_!wge3!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9782253a-8c1f-44fc-a104-15c63d871f1e_400x143.png 848w, https://substackcdn.com/image/fetch/$s_!wge3!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9782253a-8c1f-44fc-a104-15c63d871f1e_400x143.png 1272w, https://substackcdn.com/image/fetch/$s_!wge3!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9782253a-8c1f-44fc-a104-15c63d871f1e_400x143.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!wge3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9782253a-8c1f-44fc-a104-15c63d871f1e_400x143.png" width="400" height="143" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9782253a-8c1f-44fc-a104-15c63d871f1e_400x143.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:143,&quot;width&quot;:400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!wge3!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9782253a-8c1f-44fc-a104-15c63d871f1e_400x143.png 424w, https://substackcdn.com/image/fetch/$s_!wge3!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9782253a-8c1f-44fc-a104-15c63d871f1e_400x143.png 848w, https://substackcdn.com/image/fetch/$s_!wge3!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9782253a-8c1f-44fc-a104-15c63d871f1e_400x143.png 1272w, https://substackcdn.com/image/fetch/$s_!wge3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9782253a-8c1f-44fc-a104-15c63d871f1e_400x143.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>In addition, having the type of AI that is trained to give answers that are maximally insightful and persuasive to humans seems like the type of thing that might not be terrible. Consider how in court, a prosecutor and defendant biased in opposite directions are generally assumed to converge on the truth. Unless, of course, maximising persuasiveness to humans - over accuracy or helpfulness - is exactly the type of thing that gets the worst parts of Goodhart&#8217;s law delivered to you by 24/7 Amazon Prime express delivery.</p><h3>Example 3: Assistance Games and CIRL</h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!RbZq!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F075357e8-3977-428d-a55c-2fb41d3b6337_400x400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!RbZq!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F075357e8-3977-428d-a55c-2fb41d3b6337_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!RbZq!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F075357e8-3977-428d-a55c-2fb41d3b6337_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!RbZq!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F075357e8-3977-428d-a55c-2fb41d3b6337_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!RbZq!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F075357e8-3977-428d-a55c-2fb41d3b6337_400x400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!RbZq!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F075357e8-3977-428d-a55c-2fb41d3b6337_400x400.png" width="400" height="400" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/075357e8-3977-428d-a55c-2fb41d3b6337_400x400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:400,&quot;width&quot;:400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!RbZq!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F075357e8-3977-428d-a55c-2fb41d3b6337_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!RbZq!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F075357e8-3977-428d-a55c-2fb41d3b6337_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!RbZq!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F075357e8-3977-428d-a55c-2fb41d3b6337_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!RbZq!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F075357e8-3977-428d-a55c-2fb41d3b6337_400x400.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Human teaching a robot with feedback, digital art, trending on artstation</em></figcaption></figure></div><p>Assistance Games are the name of a broad class of approaches pioneered by Stuart Russell, a prominent figure in AI and co-author of the&nbsp;<a href="https://en.wikipedia.org/wiki/Artificial_Intelligence:_A_Modern_Approach">best-known AI textbook in the world</a>. Russell talks about his approach more in his book&nbsp;<em><a href="https://en.wikipedia.org/wiki/Human_Compatible">Human Compatible</a></em>. In it, he summarises the key his approach to aligning AI with the following three principles:</p><ul><li><p>The machine&#8217;s only objective is to maximise the realisation of human preferences.</p></li><li><p>The machine is initially uncertain about what those preferences are.</p></li><li><p>The ultimate source of information about human preferences is human behaviour.</p></li></ul><p>The key component here is&nbsp;<strong>uncertainty about preferences</strong>. This is in contrast to what Russell calls the &#8220;standard model&#8221; of AI, where machines optimise a fixed objective supplied by humans. We have discussed in previous sections the problems with such a paradigm. A lot of Russell&#8217;s work focuses on changing the standard way the field thinks about AI.</p><p>To put these principles into action, Russell has designed what he calls&nbsp;<strong>assistance games</strong>. These are situations in which the machine and human interact, and the human&#8217;s actions are taken as evidence by the machine about the human&#8217;s true preferences. To explain the form of these games would involve a long tangent into game theory, which these margins are too short to contain. However, one thing worth noting is that assistance games have the potential to solve the&nbsp;<strong>&#8220;off-switch problem&#8221;</strong>; that a machine will try and take steps to prevent itself from being switched off (we described this as&nbsp;<em>self-preservation</em> earlier, in the section on instrumental goals). If the AI is uncertain about human goals, then the human trying to switch it off is evidence that the AI was going to do something wrong &#8211; in which case, it is happy to be switched off. However, this is far from a complete agenda, and formalising it has many roadblocks to get past. For instance, the question of how exactly to infer human preferences from human behaviour leads into thorny philosophical issues such as&nbsp;<em>Gricean semantics.&nbsp;</em>In cases where the AI makes incorrect inferences about human preferences, it might no longer allow itself to be shut down. See&nbsp;<a href="https://mailchi.mp/59ddebcb3b9a/an-69-stuart-russells-new-book-on-why-we-need-to-replace-the-standard-model-of-ai">this Alignment Newsletter entry</a> for a summary of Russell&#8217;s book, which provides some more details as well as an overview of relevant papers.</p><blockquote><p><em>Vignette: <strong>CHAI&nbsp;</strong></em></p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!A07c!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a262fda-5dbe-4af6-98f3-b33e66601683_320x114.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!A07c!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a262fda-5dbe-4af6-98f3-b33e66601683_320x114.png 424w, https://substackcdn.com/image/fetch/$s_!A07c!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a262fda-5dbe-4af6-98f3-b33e66601683_320x114.png 848w, https://substackcdn.com/image/fetch/$s_!A07c!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a262fda-5dbe-4af6-98f3-b33e66601683_320x114.png 1272w, https://substackcdn.com/image/fetch/$s_!A07c!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a262fda-5dbe-4af6-98f3-b33e66601683_320x114.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!A07c!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a262fda-5dbe-4af6-98f3-b33e66601683_320x114.png" width="320" height="114" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1a262fda-5dbe-4af6-98f3-b33e66601683_320x114.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:114,&quot;width&quot;:320,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!A07c!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a262fda-5dbe-4af6-98f3-b33e66601683_320x114.png 424w, https://substackcdn.com/image/fetch/$s_!A07c!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a262fda-5dbe-4af6-98f3-b33e66601683_320x114.png 848w, https://substackcdn.com/image/fetch/$s_!A07c!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a262fda-5dbe-4af6-98f3-b33e66601683_320x114.png 1272w, https://substackcdn.com/image/fetch/$s_!A07c!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1a262fda-5dbe-4af6-98f3-b33e66601683_320x114.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p><em>CHAI (the Centre for Human-Compatible AI) is a research lab at UC Berkeley, run by Stuart Russell. Compared to most other AI safety organisations, they engage a lot with the academic community, and have produced a great deal of research over the years. They are best-known for their work on CIRL (Cooperative Inverse Reinforcement Learning), which can be seen as a specific approach to a certain kind of assistance game. However, they have a very broad focus which also includes work on multi-agent scenarios (when rather than a single AI and single human, there exists more than one AI or more than one human - see the&nbsp;<a href="http://acritch.com/arches/">ARCHES agenda</a> for more on this).&nbsp;</em></p></blockquote><h3>Example 4: Reinforcement learning from human feedback (RLHF)</h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!hw68!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdefee92a-20b9-4806-a677-aefd45c42360_400x400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!hw68!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdefee92a-20b9-4806-a677-aefd45c42360_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!hw68!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdefee92a-20b9-4806-a677-aefd45c42360_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!hw68!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdefee92a-20b9-4806-a677-aefd45c42360_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!hw68!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdefee92a-20b9-4806-a677-aefd45c42360_400x400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!hw68!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdefee92a-20b9-4806-a677-aefd45c42360_400x400.png" width="400" height="400" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/defee92a-20b9-4806-a677-aefd45c42360_400x400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:400,&quot;width&quot;:400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!hw68!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdefee92a-20b9-4806-a677-aefd45c42360_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!hw68!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdefee92a-20b9-4806-a677-aefd45c42360_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!hw68!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdefee92a-20b9-4806-a677-aefd45c42360_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!hw68!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdefee92a-20b9-4806-a677-aefd45c42360_400x400.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>Training a robot to do a backflip, digital art, trending on artstation</em></figcaption></figure></div><p>Reinforcement learning (RL) is one of the main branches of ML, focusing on the case where the job of the ML model is to act in some environment and maximise the probability of reward. Reinforcement learning from human feedback (RLHF) means that the ML model&#8217;s reward signal comes (at least partly) from humans giving it feedback directly, rather than humans programming in an automatic reward function and calling it a day.</p><p>The famous initial success in this was DeepMind training an ML model in a simulated environment&nbsp;<a href="https://www.deepmind.com/blog/learning-through-human-feedback">to do a backflip</a> (link includes GIF) in 2017, based purely on it repeatedly doing two backflips and then humans labelling one of them as the better one. Note how relying on human feedback makes this task much more robust to specification gaming; in other cases, humans have tried to get ML agents to run fast, only to find that they learn to become very tall and then fall forward (achieving a very high average speed, using the definition of speed as the rate at which their centre of mass moves -&nbsp;<a href="http://www.karlsims.com/papers/siggraph94.pdf">paper</a>,&nbsp;<a href="https://www.youtube.com/watch?v=TaXUZfwACVE&amp;list=PL5278ezwmoxQODgYB0hWnC0-Ob09GZGe2&amp;index=9">video</a>). However, human reward signals can be fooled. For example,&nbsp;<a href="https://openai.com/blog/deep-reinforcement-learning-from-human-preferences/">one ML model</a> that was being trained to grab a ball with a hand learned to place the hand between the camera and the ball in such a way that it looked to the human evaluators as if it were holding the ball.</p><p>More recently, OpenAI produced a version of their advanced language model GPT-3 that was fine-tuned on human feedback to do a better job of following instructions. They named it&nbsp;<a href="https://cdn.openai.com/papers/Training_language_models_to_follow_instructions_with_human_feedback.pdf">InstructGPT, and found that it was much more helpful than vanilla GPT-3</a> at being useful.</p><p>Pure RLHF is unlikely to be the solution on its own. Ajeya Cotra, a researcher at Open Philanthropy who we will meet again when we talk about forecasting AI timelines, calls a variant of RLHF called HFDT (Human Feedback on Diverse Tasks) the most straightforward route to transformative AI,&nbsp;<a href="https://www.lesswrong.com/posts/pRkFkzwKZ2zfa3R6H/without-specific-countermeasures-the-easiest-path-to">while also thinking that the default outcome of using HFDT to create transformative AI is AI takeover.</a></p><h2>Robustly good approaches</h2><blockquote><p><em>In the long-run AGI will clearly be important, but we're <strong>highly uncertain</strong> about how we'll get there and what, exactly, could go wrong. So let's do <strong>work that seems good in many possible scenarios</strong>, and doesn&#8217;t rely on having a specific story in mind.</em></p></blockquote><h3>Example 1: Interpretability</h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!lpel!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b1290ed-90e0-44ce-8023-e3ee31432a1c_400x400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!lpel!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b1290ed-90e0-44ce-8023-e3ee31432a1c_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!lpel!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b1290ed-90e0-44ce-8023-e3ee31432a1c_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!lpel!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b1290ed-90e0-44ce-8023-e3ee31432a1c_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!lpel!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b1290ed-90e0-44ce-8023-e3ee31432a1c_400x400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!lpel!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b1290ed-90e0-44ce-8023-e3ee31432a1c_400x400.png" width="400" height="400" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6b1290ed-90e0-44ce-8023-e3ee31432a1c_400x400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:400,&quot;width&quot;:400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!lpel!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b1290ed-90e0-44ce-8023-e3ee31432a1c_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!lpel!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b1290ed-90e0-44ce-8023-e3ee31432a1c_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!lpel!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b1290ed-90e0-44ce-8023-e3ee31432a1c_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!lpel!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6b1290ed-90e0-44ce-8023-e3ee31432a1c_400x400.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>A person using a microscope to look inside a robot, digital art, trending on artstation</em></figcaption></figure></div><p>If you look at fundamental problems with current ML systems, #1 is probably something like this: in general we don&#8217;t have any idea what an ML model is doing, because it&#8217;s multiplying massive inscrutable matrices of floating-point numbers with other massive inscrutable matrices of floating point numbers, and it&#8217;s pretty hard to stare at that and answer questions about what the model is actually doing. Is it thinking hard about whether an image is a cat or a dog? Is it counting up electric sheep? Is it daydreaming about the AI revolution? Who knows!</p><p>If you had to figure out an answer to such a question today, your best bet might be to call Chris Olah. Chris Olah has been spearheading work into trying to interpret what neural networks are doing. A signature output of Chris Olah&#8217;s work is pictures of creepy dogs like this one:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!60y4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7509797c-bf2a-49b9-b10c-88f053d14b21_224x224.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!60y4!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7509797c-bf2a-49b9-b10c-88f053d14b21_224x224.png 424w, https://substackcdn.com/image/fetch/$s_!60y4!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7509797c-bf2a-49b9-b10c-88f053d14b21_224x224.png 848w, https://substackcdn.com/image/fetch/$s_!60y4!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7509797c-bf2a-49b9-b10c-88f053d14b21_224x224.png 1272w, https://substackcdn.com/image/fetch/$s_!60y4!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7509797c-bf2a-49b9-b10c-88f053d14b21_224x224.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!60y4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7509797c-bf2a-49b9-b10c-88f053d14b21_224x224.png" width="320" height="320" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7509797c-bf2a-49b9-b10c-88f053d14b21_224x224.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:320,&quot;width&quot;:320,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!60y4!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7509797c-bf2a-49b9-b10c-88f053d14b21_224x224.png 424w, https://substackcdn.com/image/fetch/$s_!60y4!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7509797c-bf2a-49b9-b10c-88f053d14b21_224x224.png 848w, https://substackcdn.com/image/fetch/$s_!60y4!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7509797c-bf2a-49b9-b10c-88f053d14b21_224x224.png 1272w, https://substackcdn.com/image/fetch/$s_!60y4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7509797c-bf2a-49b9-b10c-88f053d14b21_224x224.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>What&#8217;s significant about this picture is that it&#8217;s the answer to a question roughly like this: what image would maximise the activation of neuron #12345678 in a particular image-classifying neural network? (With some asterisks about needing to apply some maths details to the process to promote large-scale structure in the image to get nice-looking results, and with apologies to neuron #12345678, who I might have confused with another neuron.)</p><p>If neuron #12345678 is maximised by something that looks like a dog, it&#8217;s a fair guess that this neuron somehow encodes, or is involved in encoding, the concept of &#8220;dog&#8221; inside the neural network.</p><p>What&#8217;s especially interesting is that if you do this analysis for every neuron in an ML model -&nbsp;<a href="https://microscope.openai.com/models">OpenAI Microscope</a> lets you see the results - you sometimes get clear patterns of increasing abstraction. The activation-maximising images for the first few layers are simple patterns; in intermediate layers you get things like curves and shapes, and then at the end even recognisable things, like the dog above. This seems evidence for neural ML vision models having learned to build up abstractions step-by-step.</p><p>However, it&#8217;s not always simple. For example, there are &#8220;polysemantic&#8221; neurons that correspond to several different concepts, like this one that can be equally excited by cat faces, car fronts, and cat legs:</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!sC4U!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F003d2c61-f9b8-4515-89c2-5b1a9fb7311c_400x128.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!sC4U!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F003d2c61-f9b8-4515-89c2-5b1a9fb7311c_400x128.png 424w, https://substackcdn.com/image/fetch/$s_!sC4U!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F003d2c61-f9b8-4515-89c2-5b1a9fb7311c_400x128.png 848w, https://substackcdn.com/image/fetch/$s_!sC4U!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F003d2c61-f9b8-4515-89c2-5b1a9fb7311c_400x128.png 1272w, https://substackcdn.com/image/fetch/$s_!sC4U!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F003d2c61-f9b8-4515-89c2-5b1a9fb7311c_400x128.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!sC4U!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F003d2c61-f9b8-4515-89c2-5b1a9fb7311c_400x128.png" width="400" height="128" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/003d2c61-f9b8-4515-89c2-5b1a9fb7311c_400x128.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:128,&quot;width&quot;:400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!sC4U!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F003d2c61-f9b8-4515-89c2-5b1a9fb7311c_400x128.png 424w, https://substackcdn.com/image/fetch/$s_!sC4U!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F003d2c61-f9b8-4515-89c2-5b1a9fb7311c_400x128.png 848w, https://substackcdn.com/image/fetch/$s_!sC4U!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F003d2c61-f9b8-4515-89c2-5b1a9fb7311c_400x128.png 1272w, https://substackcdn.com/image/fetch/$s_!sC4U!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F003d2c61-f9b8-4515-89c2-5b1a9fb7311c_400x128.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>Olah&#8217;s original work on vision models is strikingly readable and well-presented; you can find it&nbsp;<a href="https://distill.pub/2020/circuits/zoom-in/">here</a>.</p><p>Starting in late 2021, ML interpretability researchers have also made some progress in understanding transformers, which are the neural network architecture powering advanced language models like <a href="https://openai.com/blog/gpt-3-apps/">GPT-3</a>, <a href="https://blog.google/technology/ai/lamda/">LAMDA</a> and <a href="https://openai.com/blog/openai-codex/">Codex</a>. Unfortunately the work is less visual, particularly in the animal pictures department, but still well-presented. You can find it&nbsp;<a href="https://transformer-circuits.pub/2021/framework/index.html">here</a>.</p><p>In the most immediate sense, interpretability research is about reverse-engineering how exactly ML models do what they do. Hopefully, this will give insights into how to detect if an ML system is doing something we don&#8217;t like, and more general insights into how ML systems work in practice.</p><p>Chris Olah has some other inventive ideas about what to do with a sufficiently-good approach to ML interpretability. For example, he&#8217;s proposed the concept of &#8220;microscope AI&#8221;, which entails using AI as a tool to discover things about the world - not by having the AI tell us, but by training the ML system on some data, and then extracting insights about the data by digging into the internals of the ML system without necessarily ever actually running it.</p><blockquote><p><em>Vignette: <strong>Anthropic</strong></em></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!lWWD!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b328f2-da13-4233-9b7b-75336c76f667_320x320.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!lWWD!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b328f2-da13-4233-9b7b-75336c76f667_320x320.jpeg 424w, https://substackcdn.com/image/fetch/$s_!lWWD!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b328f2-da13-4233-9b7b-75336c76f667_320x320.jpeg 848w, https://substackcdn.com/image/fetch/$s_!lWWD!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b328f2-da13-4233-9b7b-75336c76f667_320x320.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!lWWD!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b328f2-da13-4233-9b7b-75336c76f667_320x320.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!lWWD!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b328f2-da13-4233-9b7b-75336c76f667_320x320.jpeg" width="320" height="320" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/55b328f2-da13-4233-9b7b-75336c76f667_320x320.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:320,&quot;width&quot;:320,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!lWWD!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b328f2-da13-4233-9b7b-75336c76f667_320x320.jpeg 424w, https://substackcdn.com/image/fetch/$s_!lWWD!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b328f2-da13-4233-9b7b-75336c76f667_320x320.jpeg 848w, https://substackcdn.com/image/fetch/$s_!lWWD!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b328f2-da13-4233-9b7b-75336c76f667_320x320.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!lWWD!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b328f2-da13-4233-9b7b-75336c76f667_320x320.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><em>Anthropic is an AI safety company, started by people who left <a href="https://openai.com/">OpenAI</a>. The company&#8217;s approach is very empirical, focused on running experiments with machine learning models. In particular, Anthropic does a lot of interpretability work, including&nbsp;<a href="https://transformer-circuits.pub/">the state-of-the-art papers on reverse-engineering how transformer-based language models work.</a></em></p></blockquote><h3>Example 2: Adversarial robustness</h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!kFka!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02d59f46-fb45-4677-bd2f-e1d94cd1f149_400x400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!kFka!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02d59f46-fb45-4677-bd2f-e1d94cd1f149_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!kFka!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02d59f46-fb45-4677-bd2f-e1d94cd1f149_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!kFka!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02d59f46-fb45-4677-bd2f-e1d94cd1f149_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!kFka!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02d59f46-fb45-4677-bd2f-e1d94cd1f149_400x400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!kFka!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02d59f46-fb45-4677-bd2f-e1d94cd1f149_400x400.png" width="400" height="400" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/02d59f46-fb45-4677-bd2f-e1d94cd1f149_400x400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:400,&quot;width&quot;:400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!kFka!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02d59f46-fb45-4677-bd2f-e1d94cd1f149_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!kFka!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02d59f46-fb45-4677-bd2f-e1d94cd1f149_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!kFka!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02d59f46-fb45-4677-bd2f-e1d94cd1f149_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!kFka!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F02d59f46-fb45-4677-bd2f-e1d94cd1f149_400x400.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>robot which is merging with a panda, digital art, trending on artstation</em></figcaption></figure></div><p>Some modern ML systems are vulnerable to adversarial examples, where a small and seemingly innocuous change to an input causes a major change in the output behaviour. Here, we see two seemingly very similar images of a panda, except carefully-selected noise has made the ML classification model very confidently say that the image is of a gibbon:</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://www.researchgate.net/publication/347639649/figure/fig1/AS:973837478948864@1609192356344/A-demonstration-of-an-adversarial-sample-21-The-panda-image-is-recognized-as-a-gibbon.ppm" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://www.researchgate.net/publication/347639649/figure/fig1/AS:973837478948864@1609192356344/A-demonstration-of-an-adversarial-sample-21-The-panda-image-is-recognized-as-a-gibbon.ppm 424w, https://www.researchgate.net/publication/347639649/figure/fig1/AS:973837478948864@1609192356344/A-demonstration-of-an-adversarial-sample-21-The-panda-image-is-recognized-as-a-gibbon.ppm 848w, https://www.researchgate.net/publication/347639649/figure/fig1/AS:973837478948864@1609192356344/A-demonstration-of-an-adversarial-sample-21-The-panda-image-is-recognized-as-a-gibbon.ppm 1272w, https://www.researchgate.net/publication/347639649/figure/fig1/AS:973837478948864@1609192356344/A-demonstration-of-an-adversarial-sample-21-The-panda-image-is-recognized-as-a-gibbon.ppm 1456w" sizes="100vw"><img src="https://www.researchgate.net/publication/347639649/figure/fig1/AS:973837478948864@1609192356344/A-demonstration-of-an-adversarial-sample-21-The-panda-image-is-recognized-as-a-gibbon.ppm" width="400" height="153" data-attrs="{&quot;src&quot;:&quot;https://www.researchgate.net/publication/347639649/figure/fig1/AS:973837478948864@1609192356344/A-demonstration-of-an-adversarial-sample-21-The-panda-image-is-recognized-as-a-gibbon.ppm&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:153,&quot;width&quot;:400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://www.researchgate.net/publication/347639649/figure/fig1/AS:973837478948864@1609192356344/A-demonstration-of-an-adversarial-sample-21-The-panda-image-is-recognized-as-a-gibbon.ppm 424w, https://www.researchgate.net/publication/347639649/figure/fig1/AS:973837478948864@1609192356344/A-demonstration-of-an-adversarial-sample-21-The-panda-image-is-recognized-as-a-gibbon.ppm 848w, https://www.researchgate.net/publication/347639649/figure/fig1/AS:973837478948864@1609192356344/A-demonstration-of-an-adversarial-sample-21-The-panda-image-is-recognized-as-a-gibbon.ppm 1272w, https://www.researchgate.net/publication/347639649/figure/fig1/AS:973837478948864@1609192356344/A-demonstration-of-an-adversarial-sample-21-The-panda-image-is-recognized-as-a-gibbon.ppm 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>Adversarial robustness is about making AI systems robust to attempts to make them do bad things, even when they&#8217;re presented with inputs carefully designed to try to make them mess up.</p><p>Redwood Research recently did a project (that resulted in&nbsp;<a href="https://arxiv.org/pdf/2205.01663.pdf">a paper</a>) about using language models to complete stories in a way where people don&#8217;t get injured. They used a technique called adversarial training, where they developed tools that helped generate examples where the current model did not classify them as injurious, and then trained their classifier specifically on those breaking examples. With this strategy they managed to reduce the fraction of injurious story completions from 2.4% to 0.003% - both small numbers, but one a thousand times smaller. Their hope is that this type of method can be applied to training AIs for high-stakes settings where reliability is important.</p><p>An example of a theoretical difficulty with adversarial training is that sometimes a failure in the model might exist, but it might be very hard to instantiate. For example, if an advanced AI acts according to the rule &#8220;if everything I see is consistent with the year being 2050, I will kill all humans&#8221;, and we assume that we can&#8217;t fool it well enough about what year it actually is, then adversarial training isn&#8217;t very useful. This leads to the concept of&nbsp;<em>relaxed</em> adversarial training, which is about extending adversarial training to cases where you can&#8217;t construct a specific adversarial input but you can argue that one exists. Evan Hubinger describes this&nbsp;<a href="https://www.lesswrong.com/posts/9Dy5YRaoCxH9zuJqa/relaxed-adversarial-training-for-inner-alignment">here</a>.</p><blockquote><p><em>Vignette: <strong>Redwood Research</strong></em></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!j-wQ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F53e60f17-650d-43a7-b328-2247b30b8189_200x200.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!j-wQ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F53e60f17-650d-43a7-b328-2247b30b8189_200x200.jpeg 424w, https://substackcdn.com/image/fetch/$s_!j-wQ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F53e60f17-650d-43a7-b328-2247b30b8189_200x200.jpeg 848w, https://substackcdn.com/image/fetch/$s_!j-wQ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F53e60f17-650d-43a7-b328-2247b30b8189_200x200.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!j-wQ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F53e60f17-650d-43a7-b328-2247b30b8189_200x200.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!j-wQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F53e60f17-650d-43a7-b328-2247b30b8189_200x200.jpeg" width="320" height="320" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/53e60f17-650d-43a7-b328-2247b30b8189_200x200.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:320,&quot;width&quot;:320,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!j-wQ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F53e60f17-650d-43a7-b328-2247b30b8189_200x200.jpeg 424w, https://substackcdn.com/image/fetch/$s_!j-wQ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F53e60f17-650d-43a7-b328-2247b30b8189_200x200.jpeg 848w, https://substackcdn.com/image/fetch/$s_!j-wQ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F53e60f17-650d-43a7-b328-2247b30b8189_200x200.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!j-wQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F53e60f17-650d-43a7-b328-2247b30b8189_200x200.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><em>Like Anthropic, Redwood Research is an AI safety company focused on empirical research on ML systems. In addition to work on interpretability, they did the adversarial training project described in the previous section. Redwood has lots of interns, and runs the Machine Learning for Alignment Bootcamp (MLAB) that teaches people interested in AI safety about practical ML.</em></p></blockquote><h3>Example 3: Eliciting Latent Knowledge (ELK)</h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Siba!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4582c77c-f473-4bec-9b4c-2e90d6a568f8_400x400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Siba!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4582c77c-f473-4bec-9b4c-2e90d6a568f8_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!Siba!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4582c77c-f473-4bec-9b4c-2e90d6a568f8_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!Siba!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4582c77c-f473-4bec-9b4c-2e90d6a568f8_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!Siba!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4582c77c-f473-4bec-9b4c-2e90d6a568f8_400x400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Siba!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4582c77c-f473-4bec-9b4c-2e90d6a568f8_400x400.png" width="400" height="400" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4582c77c-f473-4bec-9b4c-2e90d6a568f8_400x400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:400,&quot;width&quot;:400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Siba!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4582c77c-f473-4bec-9b4c-2e90d6a568f8_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!Siba!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4582c77c-f473-4bec-9b4c-2e90d6a568f8_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!Siba!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4582c77c-f473-4bec-9b4c-2e90d6a568f8_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!Siba!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4582c77c-f473-4bec-9b4c-2e90d6a568f8_400x400.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>an oil painting of an armoured automaton standing guard next to a diamond</em></figcaption></figure></div><p>Eliciting Latent Knowledge (ELK) is an important sub-problem within alignment identified by the team at the&nbsp;<a href="https://alignment.org/">Alignment Research Center (ARC</a>), and is the single project ARC is currently pursuing. The core idea is that a common way advanced AI systems might go wrong is by taking action sequences that lead to outcomes that look good by some metric, but which humans would clearly identify as bad if they knew about it in sufficient detail. As a toy example, the ELK report discusses the case of an AI guarding a diamond in a vault by operating some complex machinery around it. Humans judge how well the AI is doing by looking at a video feed of the diamond in the vault. Let&#8217;s say the AI tries to trick us by placing a picture of the diamond in front of the camera. The human judgement on this would be positive - assume the humans can&#8217;t tell the diamond is gone because the picture is good enough - but there exists information which, if the humans knew, would change their judgement. Presumably the AI understands this, since it is likely reasoning about the diamond being gone but the humans being fooled anyway when it comes up with this plan. We want to train an AI in such a way that we can get out knowledge that the AI seems to know, even when it might be incentivised to hide it.</p><p>ARC&#8217;s goal is to find a theoretical approach that seems to solve the problem even given worst-case assumptions.</p><p>ARC ran an ELK competition, and&nbsp;<a href="https://forum.effectivealtruism.org/posts/Q2BJnpNh8e6RAWFnm/consider-trying-the-elk-contest-i-am">trying to see if you can come up with solutions to the ELK problem</a> is often recommended as a way to quickly get a taste of theoretical alignment research. You can read the full problem description&nbsp;<a href="https://docs.google.com/document/d/1WwsnJQstPq91_Yh-Ch2XRL8H_EpsnjrC1dwZXR37PC8/edit#heading=h.kkaua0hwmp1d">here</a>.</p><h3>Example 4: Forecasting and timelines</h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!yIPU!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5c870c9f-1c1b-4849-9ede-75080b882cf5_400x400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!yIPU!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5c870c9f-1c1b-4849-9ede-75080b882cf5_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!yIPU!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5c870c9f-1c1b-4849-9ede-75080b882cf5_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!yIPU!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5c870c9f-1c1b-4849-9ede-75080b882cf5_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!yIPU!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5c870c9f-1c1b-4849-9ede-75080b882cf5_400x400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!yIPU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5c870c9f-1c1b-4849-9ede-75080b882cf5_400x400.png" width="400" height="400" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5c870c9f-1c1b-4849-9ede-75080b882cf5_400x400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:400,&quot;width&quot;:400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!yIPU!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5c870c9f-1c1b-4849-9ede-75080b882cf5_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!yIPU!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5c870c9f-1c1b-4849-9ede-75080b882cf5_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!yIPU!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5c870c9f-1c1b-4849-9ede-75080b882cf5_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!yIPU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5c870c9f-1c1b-4849-9ede-75080b882cf5_400x400.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>artificial intelligence which is thinking about a line on a graph, forecasting, digital art, trending on artstation</em></figcaption></figure></div><p>Many questions depend on how soon we&#8217;re going to get AGI. As the saying goes: prediction is very hard, especially about the future - and this is doubly true about predicting major technological changes.&nbsp;</p><p>One way to try to forecast AGI timelines is to&nbsp;<a href="https://www.lesswrong.com/posts/H6hMugfY3tDQGfqYL/what-do-ml-researchers-think-about-ai-in-2022">ask experts</a>, or find other ways of aggregating the opinion of people who have the knowledge or incentive to be right, as for example&nbsp;<a href="https://www.metaculus.com/questions/3479/date-weakly-general-ai-is-publicly-known/">prediction markets do</a>. Both of these are essentially just ways of tapping into the intuition of a bunch of people who hopefully have some idea.</p><p>In an attempt to bring in new light on the matter, Ajeya Cotra (a researcher at Open Philanthropy) wrote&nbsp;<a href="https://www.lesswrong.com/posts/KrJfoZzpSDpnrv9va/draft-report-on-ai-timelines">a long report</a> on trying to forecast AI milestones by trying out several ways of analogising AI to biological brains. The report is often referred to as &#8220;Biological Anchors&#8221;. For example, you might assume that an ML model that does as much computation as the human brain has a decent chance of being a human-level AI. There are many degrees of freedom here: is the relevant compute number the amount of compute the human brain uses to run versus the amount of compute it takes to run a trained ML system, or the total compute of a human brain over a human lifetime versus the compute required to train the ML model from scratch, or something else entirely? In her report, Cotra looks at a range of assumptions for this, and at predictions of future compute trends, and somewhat surprisingly finds that which set of assumptions you make doesn&#8217;t matter too much; every scenario involves &gt;50% of human-level AI by 2100.</p><p>The Biological Anchors method is very imprecise. For one, it neglects algorithmic improvements. For another, it is very unclear what the right biological comparison point is, and how to translate ML-relevant variables like compute measured in FLOPS (FLoating point OPerations per Second) or parameter count into biological equivalents. However, the report does a good job of acknowledging and taking into account all this uncertainty in its models. More generally, anything that sheds light into the question of when we get AGI seems highly relevant.</p><h2>Deconfusion</h2><blockquote><p><em>Reasoning about how to align AGI involves reasoning about complex concepts, such as intelligence, alignment and values, and <strong>we&#8217;re pretty confused about what these even mean</strong>. This means any work we do now is plausibly not helpful and definitely not reliable. As such, our priority should be <strong>doing conceptual work on how to think about these concepts and what we&#8217;re aiming for</strong>, and trying to become less confused.</em></p></blockquote><p>Of all the categories under discussion here, deconfusion has maybe the least clear path to impact. It&#8217;s not immediately obvious how becoming less confused about concepts like these is going to translate into an improved ability to align AGIs.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!iSQN!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F190646ff-72cf-4da4-acf2-8458de79a54a_400x270.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!iSQN!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F190646ff-72cf-4da4-acf2-8458de79a54a_400x270.png 424w, https://substackcdn.com/image/fetch/$s_!iSQN!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F190646ff-72cf-4da4-acf2-8458de79a54a_400x270.png 848w, https://substackcdn.com/image/fetch/$s_!iSQN!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F190646ff-72cf-4da4-acf2-8458de79a54a_400x270.png 1272w, https://substackcdn.com/image/fetch/$s_!iSQN!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F190646ff-72cf-4da4-acf2-8458de79a54a_400x270.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!iSQN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F190646ff-72cf-4da4-acf2-8458de79a54a_400x270.png" width="400" height="271" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/190646ff-72cf-4da4-acf2-8458de79a54a_400x270.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:271,&quot;width&quot;:400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!iSQN!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F190646ff-72cf-4da4-acf2-8458de79a54a_400x270.png 424w, https://substackcdn.com/image/fetch/$s_!iSQN!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F190646ff-72cf-4da4-acf2-8458de79a54a_400x270.png 848w, https://substackcdn.com/image/fetch/$s_!iSQN!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F190646ff-72cf-4da4-acf2-8458de79a54a_400x270.png 1272w, https://substackcdn.com/image/fetch/$s_!iSQN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F190646ff-72cf-4da4-acf2-8458de79a54a_400x270.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Some kinds of deconfusion research is just about finding clearer ways of describing different parts of the alignment problem (Hubinger&#8217;s&nbsp;<a href="https://arxiv.org/abs/1906.01820">Risks From Learned Optimisation</a>, where he first introduces the inner/outer alignment terminology, is a good example of this). But other types of research can dive heavily into mathematics and even philosophy, and be very difficult to understand.</p><h3>Example 1: MIRI and Agent Foundations</h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!V0tI!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3bedbbe1-f772-41a7-b903-4c3d0f1d1fda_400x400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!V0tI!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3bedbbe1-f772-41a7-b903-4c3d0f1d1fda_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!V0tI!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3bedbbe1-f772-41a7-b903-4c3d0f1d1fda_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!V0tI!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3bedbbe1-f772-41a7-b903-4c3d0f1d1fda_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!V0tI!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3bedbbe1-f772-41a7-b903-4c3d0f1d1fda_400x400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!V0tI!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3bedbbe1-f772-41a7-b903-4c3d0f1d1fda_400x400.png" width="400" height="400" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3bedbbe1-f772-41a7-b903-4c3d0f1d1fda_400x400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:400,&quot;width&quot;:400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!V0tI!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3bedbbe1-f772-41a7-b903-4c3d0f1d1fda_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!V0tI!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3bedbbe1-f772-41a7-b903-4c3d0f1d1fda_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!V0tI!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3bedbbe1-f772-41a7-b903-4c3d0f1d1fda_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!V0tI!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3bedbbe1-f772-41a7-b903-4c3d0f1d1fda_400x400.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>robot sitting in front of a television, playing a videogame, digital art</em></figcaption></figure></div><p>The organisation most associated with this view is MIRI (the Machine Intelligence Research Institute). Its founder, Eliezer Yudkowsky, has written extensively on AI alignment and human rationality, as well as topics as wide-ranging as evolutionary psychology and quantum physics. His post&nbsp;<a href="https://intelligence.org/2018/10/03/rocket-alignment/">The Rocket Alignment Problem</a> tries to get across some of his intuitions behind MIRI&#8217;s research, in the form of an analogy &#8211; trying to build aligned AGI without having deeper understanding of concepts like intelligence and values is like trying to land a rocket on the moon by just pointing and shooting, without a working understanding of Newtonian mechanics.&nbsp;</p><p>Cryptography provides a different lens through which to view this kind of foundational research. Suppose you were trying to send secret messages to an ally, and to make sure nobody could intercept and read your messages you wanted a way to measure how much information was shared between the original and encrypted message. You might use&nbsp;<a href="https://en.wikipedia.org/wiki/Pearson_correlation_coefficient">correlation coefficient</a> as a proxy for the shared information, but unfortunately having a correlation coefficient of zero between the original and encrypted message isn&#8217;t enough to guarantee safety. But if you find the concept of&nbsp;<a href="https://en.wikipedia.org/wiki/Mutual_information">mutual information</a>, then you&#8217;re done &#8211; ensuring zero mutual information between your original and encrypted message guarantees the adversary will be unable to read your message. In other words, only once you&#8217;ve found a&nbsp;<strong>&#8220;true name&#8221;&nbsp;</strong>- a robust formalisation of the intuitive concept you&#8217;re trying to express mathematically - can you be free from the effects of Goodhart&#8217;s law. Similarly, maybe if we get robust formulations of concepts like &#8220;agency&#8221; and &#8220;optimisation&#8221;, we would be able to inspect a trained system and tell whether it contained any misaligned inner optimisers (see the first post), and these inspection tools would work even in extreme circumstances (such as the AI becoming much smarter than us).</p><p>Much of MIRI&#8217;s research has come under the heading of&nbsp;<a href="https://intelligence.org/embedded-agency/">embedded agency</a>. This tackles issues that arise when we are considering agents which are part of the environments they operate in (as opposed to standard assumptions in fields like reinforcement learning, where the agent is viewed as separate from their environment). Four main subfields of this area of study are:</p><ul><li><p><strong>Decision theory</strong> (adapting classical decision theory to embedded agents)</p></li><li><p><strong>Embedded world-models&nbsp;</strong>(how to form true beliefs about the a world in which you are embedded)</p></li><li><p><strong>Robust delegation&nbsp;(understanding what trust relationships can exist between agents and its future - maybe far more intelligent - self)</strong></p></li><li><p><strong>Subsystem alignment</strong> (how to make sure an agent doesn&#8217;t spin up internal agents which have different goals)</p></li></ul><blockquote><p><em>Vignette: <strong>MIRI</strong></em></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!xF3S!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6337435b-d2fb-4549-84d2-74fc6d854bfe_320x320.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!xF3S!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6337435b-d2fb-4549-84d2-74fc6d854bfe_320x320.png 424w, https://substackcdn.com/image/fetch/$s_!xF3S!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6337435b-d2fb-4549-84d2-74fc6d854bfe_320x320.png 848w, https://substackcdn.com/image/fetch/$s_!xF3S!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6337435b-d2fb-4549-84d2-74fc6d854bfe_320x320.png 1272w, https://substackcdn.com/image/fetch/$s_!xF3S!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6337435b-d2fb-4549-84d2-74fc6d854bfe_320x320.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!xF3S!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6337435b-d2fb-4549-84d2-74fc6d854bfe_320x320.png" width="320" height="320" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6337435b-d2fb-4549-84d2-74fc6d854bfe_320x320.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:320,&quot;width&quot;:320,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!xF3S!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6337435b-d2fb-4549-84d2-74fc6d854bfe_320x320.png 424w, https://substackcdn.com/image/fetch/$s_!xF3S!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6337435b-d2fb-4549-84d2-74fc6d854bfe_320x320.png 848w, https://substackcdn.com/image/fetch/$s_!xF3S!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6337435b-d2fb-4549-84d2-74fc6d854bfe_320x320.png 1272w, https://substackcdn.com/image/fetch/$s_!xF3S!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6337435b-d2fb-4549-84d2-74fc6d854bfe_320x320.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><em>MIRI is the oldest organisation in the AI alignment space. It used to be called the Singularity Institute, and had the goal of accelerating the development of AI. In 2005 they shifted focus towards trying to manage the risks from advanced AI. This has largely consisted of fundamental mathematical research of the type described above. MIRI might be better described as a confluence of smart people with backgrounds in highly technical fields (e.g. mathematics), working on different research agendas that share underlying philosophies and intuitions. They have a nondisclosure policy by default, which they explain in this&nbsp;<a href="https://intelligence.org/2018/11/22/2018-update-our-new-research-directions/#section3">announcement post</a> from 2018.</em></p></blockquote><h3>Example 2: John Wentworth and Natural Abstractions</h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!z53i!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd999e2ab-60e9-4655-a1f6-eebf487f2f6a_400x400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!z53i!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd999e2ab-60e9-4655-a1f6-eebf487f2f6a_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!z53i!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd999e2ab-60e9-4655-a1f6-eebf487f2f6a_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!z53i!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd999e2ab-60e9-4655-a1f6-eebf487f2f6a_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!z53i!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd999e2ab-60e9-4655-a1f6-eebf487f2f6a_400x400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!z53i!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd999e2ab-60e9-4655-a1f6-eebf487f2f6a_400x400.png" width="400" height="400" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d999e2ab-60e9-4655-a1f6-eebf487f2f6a_400x400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:400,&quot;width&quot;:400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!z53i!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd999e2ab-60e9-4655-a1f6-eebf487f2f6a_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!z53i!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd999e2ab-60e9-4655-a1f6-eebf487f2f6a_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!z53i!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd999e2ab-60e9-4655-a1f6-eebf487f2f6a_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!z53i!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd999e2ab-60e9-4655-a1f6-eebf487f2f6a_400x400.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>thermometer being used to measure a robot, digital art, trending on artstation</em></figcaption></figure></div><p>John Wentworth is an independent researcher, who publishes most of his work on&nbsp;<a href="https://www.lesswrong.com/users/johnswentworth">LessWrong</a> and the&nbsp;<a href="https://www.alignmentforum.org/users/johnswentworth">AI Alignment Forum</a>. His main research agenda focuses on the idea of&nbsp;<a href="https://www.lesswrong.com/posts/Fut8dtFsBYRz8atFF/the-natural-abstraction-hypothesis-implications-and-evidence">Natural Abstractions</a>, which can be described in terms of three sub-claims:</p><ul><li><p><strong>Abstractability</strong><br>Our physical world abstracts well, i.e. we can usually come up with simpler summaries (abstractions) for much more complicated systems (example: a gear is a very complex object containing a vast number of atoms, but we can summarise all relevant information about it in just one number - the angle of rotation).</p></li><li><p><strong>Human-Compatibility</strong><br>These are the abstractions used by humans in day-to-day thought/language.</p></li><li><p><strong>Convergence</strong><br>These abstractions are "natural", in the sense that we should expect a wide variety of intelligent agents to converge on using them.</p></li></ul><p>The&nbsp;<a href="https://www.lesswrong.com/posts/gdEDPHjCY5DKsMsvE/the-pragmascope-idea">ideal outcome</a> of this line of research would be some kind of measurement device (an &#8220;abstraction thermometer&#8221;), which could take in a system like a trained neural network and spit out a representation of the abstractions represented by that system. In this way, you&#8217;d be able to get a better understanding of what the AI was actually doing. In particular, you might be able to identify inner alignment failures (the AI&#8217;s true goal not corresponding to the reward function it was&nbsp; being trained on), and you could retrain it while pointed at the intended goal. So far, this line of research has consisted of some&nbsp;<a href="https://www.lesswrong.com/posts/jJf4FrfiQdDGg7uco/the-telephone-theorem-information-at-a-distance-is-mediated">fairly</a>&nbsp;<a href="https://www.lesswrong.com/posts/cqdDGuTs2NamtEhBW/maxent-and-abstractions-current-best-arguments">dense</a>&nbsp;<a href="https://www.lesswrong.com/posts/vvEebH5jEvxnJEvBC/abstractions-as-redundant-information">mathematics</a>, but Wentworth has&nbsp;<a href="https://www.lesswrong.com/posts/gdEDPHjCY5DKsMsvE/the-pragmascope-idea">described</a> his plans to build on this with more empirical work (e.g. training neural networks on the same data, and using tools from calculus to try and compare the similarity of concepts learned by each of the networks).&nbsp;&nbsp;</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ejr2!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb88e8072-7eb9-4ac8-a862-b6de00216d0b_400x173.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ejr2!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb88e8072-7eb9-4ac8-a862-b6de00216d0b_400x173.png 424w, https://substackcdn.com/image/fetch/$s_!ejr2!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb88e8072-7eb9-4ac8-a862-b6de00216d0b_400x173.png 848w, https://substackcdn.com/image/fetch/$s_!ejr2!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb88e8072-7eb9-4ac8-a862-b6de00216d0b_400x173.png 1272w, https://substackcdn.com/image/fetch/$s_!ejr2!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb88e8072-7eb9-4ac8-a862-b6de00216d0b_400x173.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ejr2!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb88e8072-7eb9-4ac8-a862-b6de00216d0b_400x173.png" width="400" height="174" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b88e8072-7eb9-4ac8-a862-b6de00216d0b_400x173.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:174,&quot;width&quot;:400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ejr2!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb88e8072-7eb9-4ac8-a862-b6de00216d0b_400x173.png 424w, https://substackcdn.com/image/fetch/$s_!ejr2!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb88e8072-7eb9-4ac8-a862-b6de00216d0b_400x173.png 848w, https://substackcdn.com/image/fetch/$s_!ejr2!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb88e8072-7eb9-4ac8-a862-b6de00216d0b_400x173.png 1272w, https://substackcdn.com/image/fetch/$s_!ejr2!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb88e8072-7eb9-4ac8-a862-b6de00216d0b_400x173.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><h2>AI governance</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!wnHZ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F391f6e4f-70fd-4290-9597-4f635b0f76ad_400x400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!wnHZ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F391f6e4f-70fd-4290-9597-4f635b0f76ad_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!wnHZ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F391f6e4f-70fd-4290-9597-4f635b0f76ad_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!wnHZ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F391f6e4f-70fd-4290-9597-4f635b0f76ad_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!wnHZ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F391f6e4f-70fd-4290-9597-4f635b0f76ad_400x400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!wnHZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F391f6e4f-70fd-4290-9597-4f635b0f76ad_400x400.png" width="400" height="400" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/391f6e4f-70fd-4290-9597-4f635b0f76ad_400x400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:400,&quot;width&quot;:400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!wnHZ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F391f6e4f-70fd-4290-9597-4f635b0f76ad_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!wnHZ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F391f6e4f-70fd-4290-9597-4f635b0f76ad_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!wnHZ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F391f6e4f-70fd-4290-9597-4f635b0f76ad_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!wnHZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F391f6e4f-70fd-4290-9597-4f635b0f76ad_400x400.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>judging, presiding over a trial, sentencing a robot, digital art, artstation</em></figcaption></figure></div><p>In these posts, we&#8217;ve mainly focused on the technical side of the issue. This is important, especially for understanding why there is a problem in the first place. However, the management and reduction of AI risk obviously includes not just technical approaches like outlined in the above sections, but also&nbsp;<a href="https://80000hours.org/articles/ai-policy-guide/">the field of AI governance</a>, which tries to understand and push for the right types of policies for advanced AI systems.</p><p>For example, the Cold War was made a lot more dangerous by the nuclear arms race. How do we avoid having an arms race in AI, either between nations or companies? More generally, how can we make sure that safety considerations are given appropriate weight by the teams building advanced AI systems? How do we make sure any technical solutions get implemented?</p><p>It&#8217;s also very hard to say what the impacts of AI will be, across a broad range of possible technical outcomes. If AI capabilities at some point advance very quickly from below human-level to far beyond the human-level, the way the future looks will likely mostly be determined by technical considerations about the AI system. However, if progress is slower, there will be a longer period of time where weird things are happening because of advanced AI - for example, significantly accelerated economic growth, or mass unemployment, or an AI-assisted boom in science - and these will have economic, social, and political ramifications that will play out in a world not too dissimilar from our own. Someone should be working on figuring out what these ramifications will be, especially if they might alter the balance of existential threats that civilisation faces; for example, if they make geopolitics less unstable and nuclear war more likely, or affect the environment in which even more powerful AI systems are developed.</p><p>The Centre for the Governance of AI, or&nbsp;<a href="https://www.governance.ai/">GovAI</a> for short, is an example of an organisation in this space.</p><h2>Field-building</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!LOzy!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5d7a43d6-fcbd-4a93-8843-f3b83eeaef2c_400x400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!LOzy!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5d7a43d6-fcbd-4a93-8843-f3b83eeaef2c_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!LOzy!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5d7a43d6-fcbd-4a93-8843-f3b83eeaef2c_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!LOzy!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5d7a43d6-fcbd-4a93-8843-f3b83eeaef2c_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!LOzy!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5d7a43d6-fcbd-4a93-8843-f3b83eeaef2c_400x400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!LOzy!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5d7a43d6-fcbd-4a93-8843-f3b83eeaef2c_400x400.png" width="400" height="400" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5d7a43d6-fcbd-4a93-8843-f3b83eeaef2c_400x400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:400,&quot;width&quot;:400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!LOzy!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5d7a43d6-fcbd-4a93-8843-f3b83eeaef2c_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!LOzy!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5d7a43d6-fcbd-4a93-8843-f3b83eeaef2c_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!LOzy!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5d7a43d6-fcbd-4a93-8843-f3b83eeaef2c_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!LOzy!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5d7a43d6-fcbd-4a93-8843-f3b83eeaef2c_400x400.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>robot giving a lecture in a university, group of students, hands up, digital art, artstation</em></figcaption></figure></div><p><em>One of the most important ways we can make AI go well is by increasing the number of capable researchers doing alignment research.</em></p><p>As mentioned, AI safety is still a relatively young field. The case here is that we might do better to grow the field, and increase the quality of research it produces in the future. Some forms that field building can take are:</p><ul><li><p><strong>Setting up new ways for people to enter the field</strong><br>There are many to list here. To give a few different structures which exist for this purpose:</p><ul><li><p><strong>Reading groups and introductory programmes.&nbsp;</strong><br>Maybe the most exciting one from the last few years has been the Cambridge&nbsp;<a href="https://www.eacambridge.org/agi-safety-fundamentals">AGI Safety Fundamentals Programme</a>, which has curricula for technical alignment and AI governance. The technical curriculum consists of 7 weeks of reading material and group discussions, and a final week of capstone projects where the participants try their hand at a project / investigation / writeup related to AI safety. Beyond this, many people are also setting up reading groups in their own universities for books like&nbsp;<em>Human Compatible</em>.&nbsp;</p></li><li><p><strong>Ways of supporting independent researchers</strong><br>The&nbsp;<a href="https://aisafety.camp/">AI Safety Camp</a> is an organisation which matches applicants with mentors posing a specific research question, and is structured as a series of group research sprints. They have produced work such as the example of inner misalignment in the CoinRun game, which we discussed in a previous section. Other examples of organisations which support independent research include&nbsp;<a href="https://www.lesswrong.com/posts/jfq2BH5kfQqu2vYv3/we-are-conjecture-a-new-alignment-research-startup">Conjecture</a>, a recent alignment startup which does their own alignment research as well as providing a structure to host externally funded independent conceptual researchers, and&nbsp;<a href="https://alignmentfund.org/">FAR (the Fund for Alignment Research)</a>.</p></li><li><p><strong>Coding bootcamps</strong><br>Since current systems are increasingly being bottlenecked by alignment and interpretability barriers rather than capabilities, in recent years more focus has been directed towards working with cutting-edge deep learning models. This requires strong coding skills and a good understanding of the relevant ML, which is why bootcamps and programmes specifically designed to skill up future alignment researchers have been created. Two such examples are&nbsp;<a href="https://www.lesswrong.com/posts/3ouxBRRzjxarTukMW/apply-to-the-second-iteration-of-the-ml-for-alignment">MLAB</a> (the Machine Learning for Alignment Bootcamp, run by Redwood Research), and&nbsp;<a href="https://forum.effectivealtruism.org/posts/9RYvJu2iNJMXgWCBn/introducing-the-ml-safety-scholars-program">MLSS</a> (the Machine Learning Safety Scholars Programme, which is based on publicly available material as well as lectures produced by Dan Hendryks).&nbsp;</p></li></ul></li><li><p><strong>Distilling research</strong><br>In&nbsp;<a href="https://www.lesswrong.com/posts/zo9zKcz47JxDErFzQ/call-for-distillers">this post</a>, John Wentworth makes the case for more distillation in AI alignment research - in other words, more people who focus on understanding and communicating the work of alignment researchers to others. This often takes the form of writing more accessible summaries of hard-to-interpret technical papers, and emphasising the key ideas.</p></li><li><p><strong>Public outreach / better intro material</strong><br>For instance, books like Brian Christian&#8217;s&nbsp;<em><a href="https://en.wikipedia.org/wiki/The_Alignment_Problem">The Alignment Problem</a>,&nbsp;</em>Stuart Russell&#8217;s&nbsp;<em><a href="https://en.wikipedia.org/wiki/Human_Compatible">Human Compatible</a></em> and Nick Bostrom&#8217;s&nbsp;<em><a href="https://en.wikipedia.org/wiki/Superintelligence:_Paths,_Dangers,_Strategies">Superintelligence</a></em> communicate AI risk to a wide audience. These books have been helpful for making the case for AI risks more mainstream. Note that there can be some overlap between this and distilling research (Rob Miles&#8217;&nbsp;<a href="https://www.youtube.com/c/RobertMilesAI">channel</a> is another great example here).</p></li><li><p><strong>Getting more of the academic community involved</strong><br>Since AI safety is a hard technical problem, and since misaligned systems generally won&#8217;t be as commercially useful as aligned ones, it makes sense to try and engage the broader field of machine learning. One great example of this is Dan Hendryks&#8217; paper&nbsp;<a href="https://mailchi.mp/08a639ffa2ba/an-167concrete-ml-safety-problems-and-their-relevance-to-x-risk">Unsolved Problems in ML Safety</a> (which describes a list of problems in AI safety, with the ML community as the target audience). Stuart Russell has also engaged a lot with the ML community.&nbsp;</p></li></ul><p>Note that this is certainly not a comprehensive overview of all current AI alignment proposals (a few more we haven&#8217;t had time to talk about are CAIS, Andrew Critch&#8217;s cooperation-and-coordination-failures framing for AI risks, and many others). However, we hope this has given you a brief overview of some of the different approaches taken by people in the field, as well as the motivations behind their research</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!b_em!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f63e99d-e1b8-4044-ae60-0a303565d91e_400x254.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!b_em!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f63e99d-e1b8-4044-ae60-0a303565d91e_400x254.png 424w, https://substackcdn.com/image/fetch/$s_!b_em!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f63e99d-e1b8-4044-ae60-0a303565d91e_400x254.png 848w, https://substackcdn.com/image/fetch/$s_!b_em!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f63e99d-e1b8-4044-ae60-0a303565d91e_400x254.png 1272w, https://substackcdn.com/image/fetch/$s_!b_em!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f63e99d-e1b8-4044-ae60-0a303565d91e_400x254.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!b_em!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f63e99d-e1b8-4044-ae60-0a303565d91e_400x254.png" width="400" height="254" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1f63e99d-e1b8-4044-ae60-0a303565d91e_400x254.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:254,&quot;width&quot;:400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!b_em!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f63e99d-e1b8-4044-ae60-0a303565d91e_400x254.png 424w, https://substackcdn.com/image/fetch/$s_!b_em!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f63e99d-e1b8-4044-ae60-0a303565d91e_400x254.png 848w, https://substackcdn.com/image/fetch/$s_!b_em!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f63e99d-e1b8-4044-ae60-0a303565d91e_400x254.png 1272w, https://substackcdn.com/image/fetch/$s_!b_em!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f63e99d-e1b8-4044-ae60-0a303565d91e_400x254.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Map of the solution approaches we've discussed so far</figcaption></figure></div><h1><strong>Conclusion</strong></h1><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!eiIi!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff350ef07-60e8-44ad-b106-033386f09c82_400x400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!eiIi!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff350ef07-60e8-44ad-b106-033386f09c82_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!eiIi!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff350ef07-60e8-44ad-b106-033386f09c82_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!eiIi!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff350ef07-60e8-44ad-b106-033386f09c82_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!eiIi!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff350ef07-60e8-44ad-b106-033386f09c82_400x400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!eiIi!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff350ef07-60e8-44ad-b106-033386f09c82_400x400.png" width="400" height="400" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f350ef07-60e8-44ad-b106-033386f09c82_400x400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:400,&quot;width&quot;:400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!eiIi!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff350ef07-60e8-44ad-b106-033386f09c82_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!eiIi!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff350ef07-60e8-44ad-b106-033386f09c82_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!eiIi!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff350ef07-60e8-44ad-b106-033386f09c82_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!eiIi!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff350ef07-60e8-44ad-b106-033386f09c82_400x400.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><em>people walking along a path which stretches off and disappears into a colorful galaxy filled with beautiful stars, digital art, trending on artstation</em></figcaption></figure></div><p>Advanced AI represents at least a technology that promises to have effects on the scale of the internet or computer revolutions, and perhaps even more likely to be more akin to the effects of the <strong>industrial revolution</strong> (which allowed for the automation of much&nbsp;<em>manual&nbsp;</em>labour) and the <strong>evolution of humans</strong> (the last time something significantly smarter than everything that had come before appeared on the planet).</p><p>It&#8217;s easy to invent technologies that the same could be said about - a magic wish-granting box! Wow! But unlike magic wish-granting boxes, something like advanced AI, or AGI, or transformative AI, or&nbsp;<a href="https://www.cold-takes.com/transformative-ai-timelines-part-1-of-4-what-kind-of-ai/">PASTA</a> (Process for Automating Scientific and Technical Achievement) seems to be headed our way. The smart money is on it very likely coming <strong>this century</strong>, and quite likely in the <strong>first half</strong>.</p><p>If you look at the progress in modern machine learning, and especially the past few years of progress in so-called deep learning, it is hard not to feel a sense of rushing progress. The past few years of progress, in particular the success of the transformer architecture, should update us in the direction that intelligence might be a surprisingly easy problem. What is essentially fancy iterative statistical curve-fitting with a few hacks thrown in already manages to write fluent appropriate English text in response to questions, create paintings from a description, and carry out multi-step logical deduction in natural language. <strong>The fundamental problem that plagued AI progress for over half a century - getting fuzzy/intuitive/creative thinking into a machine, in addition to the sharp but brittle logic at which computers have long excelled - seems to have been cracked.</strong> There is a solid empirical pattern of predictably improving performance akin to Moore&#8217;s law - the &#8220;<a href="https://arxiv.org/pdf/2001.08361.pdf">scaling laws</a>&#8221; we mentioned in the first post - that we seem not to have hit the limits of yet. There are experts in the field who would not be surprised if the remaining insights for cracking human-level machine intelligence could fit into a few good papers.</p><p>This is not to say that AGI is definitely coming soon. The field might get stuck on some stumbling block for a decade, during which there will be no doubt much written about the failed promises and excess hype of the early-2020s deep learning revolution.</p><p>Finally, as we&#8217;ve argued, by default the arrival of advanced AI might plausibly lead to civilisation-wide catastrophe.</p><p>There are few things in the world that fit all of the following points:</p><ul><li><p>A potentially transformative technology whose development would likely rank somewhere between the top events of the century and the top events in the history of life on Earth.</p></li><li><p>Something that is likely to happen in the coming decades.</p></li><li><p>Something that has a meaningful chance of being cataclysmically bad.</p></li></ul><p>For those thinking about the longer-term picture, whatever the short-term ebb and flow of progress in the field is, AI and AI risk loom large when thinking about humanity&#8217;s future. The main ways in which this might stop being the case are:</p><ul><li><p>There is a major flaw in the arguments for at least one of the above points. Since many of the arguments are abstract and not empirically falsifiable before it&#8217;s too late to matter, this is possible. However, note that there is a strong and recurring pattern of many people, including in particular many extremely-talented people, running into the arguments and taking them more and more seriously. (If you do have a strong argument against the importance of the AI alignment problem, there are many people - us included - who would be very eager to hear from you. Some of these people - us not included - would probably also pay you large amounts of money.)</p></li><li><p>We solve the technical AI alignment problem, and we solve the AI governance problem to a degree where the technical solutions will be implemented and it seems very unlikely that advanced AI systems will wreak havoc with society.</p></li><li><p>A catastrophic outcome for human civilisation, whether resulting from AI itself or something else.&nbsp;</p></li></ul><p>The project of trying to make sure the development of advanced AI goes well is likely one of the most important things in the world to be working on (if you&#8217;re lost, the&nbsp;<a href="https://80000hours.org/problem-profiles/positively-shaping-artificial-intelligence/">80 000 Hours problem profile</a> is a decent place to start). It might turn out to be easy - consider how many seemingly intractable scientific problems dissolved once someone had the right insight. But right now, at least, it seems like it might be a fiendishly difficult problem, especially if it continues to seem like the insights we need for alignment are very different from the insights we need to build advanced AI.</p><p>Most of the time, science and technology progress in whatever direction is easiest or flows most naturally from existing knowledge. Other times, reality throws down a gauntlet, and we must either overcome the challenge or fail. May the best in our species - our ingenuity, persistence, and coordination - rise up, and deliver us from peril.</p>]]></content:encoded></item><item><title><![CDATA[AI risk intro 1: advanced AI might be very bad]]></title><description><![CDATA[This post was a joint effort with Callum McDougall.]]></description><link>https://www.nosetgauge.com/p/ai-risk-intro-1-advanced-ai-might-be</link><guid isPermaLink="false">https://www.nosetgauge.com/p/ai-risk-intro-1-advanced-ai-might-be</guid><dc:creator><![CDATA[Rudolf Laine]]></dc:creator><pubDate>Sun, 11 Sep 2022 10:27:00 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/85c98aa6-5c0a-4425-bb1c-38b705b60ed4_400x400.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em><strong>This post was a joint effort with <a href="https://www.perfectlynormal.co.uk/">Callum McDougall</a>.</strong></em></p><p><em>9.6k words (~25min) </em><br></p><h2><strong>Introduction</strong></h2><p>If human civilisation is destroyed this century, the most likely cause is advanced AI systems. This might sound like a bold claim to many, given that we live on a planet full of existing concrete threats like climate change, over ten thousand nuclear weapons, and Vladimir Putin However, it is a conclusion that many people who think about the topic keep coming to. While it is not easy to describe the case for risks from advanced AI in a single piece, here we make an effort that assumes no prior knowledge. Rather than try to argue from theory straight away, we approach it from the angle of what computers actually can and can&#8217;t do.</p><h2><strong>The Story So Far</strong></h2><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!3ITD!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86e0a903-3ca2-48b9-a456-844fe3c5d654_400x400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!3ITD!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86e0a903-3ca2-48b9-a456-844fe3c5d654_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!3ITD!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86e0a903-3ca2-48b9-a456-844fe3c5d654_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!3ITD!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86e0a903-3ca2-48b9-a456-844fe3c5d654_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!3ITD!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86e0a903-3ca2-48b9-a456-844fe3c5d654_400x400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!3ITD!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86e0a903-3ca2-48b9-a456-844fe3c5d654_400x400.png" width="400" height="400" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/86e0a903-3ca2-48b9-a456-844fe3c5d654_400x400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:400,&quot;width&quot;:400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!3ITD!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86e0a903-3ca2-48b9-a456-844fe3c5d654_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!3ITD!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86e0a903-3ca2-48b9-a456-844fe3c5d654_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!3ITD!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86e0a903-3ca2-48b9-a456-844fe3c5d654_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!3ITD!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F86e0a903-3ca2-48b9-a456-844fe3c5d654_400x400.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><p><em>Above: an image generated by OpenAI&#8217;s DALL-E 2, from the prompt: "artist's impression of an artificial intelligence thinking about chess, digital art, artstation".</em></p><p>(This section can be skipped if you understand how machine learning works and what it can and can&#8217;t do today)</p><p>Let&#8217;s say you want a computer to do some complicated task, for example learning chess. The computer has no understanding of high-level things like &#8220;chess&#8221;, &#8220;board&#8221;, &#8220;piece&#8221;, &#8220;move&#8221;, or &#8220;win&#8221; - it only understands how to do a small set of things. Your task as the programmer is to break down the high-level goal of &#8220;beat me at chess&#8221; into simpler and simpler steps, until you arrive at a simple mechanistic description of what the computer needs to do. If the computer does beat you, it&#8217;s not because it had any new insight into the problem, but rather because you were clever enough to find some <a href="https://en.wikipedia.org/wiki/Minimax">set of steps</a> that, carried out blindly in sufficient speed and quantity, overwhelms whatever cleverness you yourself can apply during the game. This is how Deep Blue beat Kasparov, and more generally how most software and the so-called &#8220;Good Old-Fashioned AI&#8221; (GOFAI) paradigm works.</p><p>Programs of this type can be powerful. In addition to <a href="https://en.wikipedia.org/wiki/Stockfish_(chess)">beating humans at chess</a>, they can <a href="https://www.google.com/maps/">calculate shortest routes</a> on maps, <a href="https://en.wikipedia.org/wiki/Coq">prove maths theorems</a>, <a href="https://en.wikipedia.org/wiki/Autopilot">mostly fly airplanes</a>, and <a href="https://duckduckgo.com/">search all human knowledge</a>. Programs of this type are responsible for the stereotypical impression of computers as logical, precise, uncreative, and brittle. They are essentially executable logic.</p><p>Many people hoped that you could write programs to do &#8220;intelligent&#8221; things. These people were right - after all, ask almost anyone before Deep Blue won whether playing chess counts as &#8220;intelligence&#8221;, they&#8217;d have said yes. But &#8220;classical&#8221; programming hit limitations, in particular in doing &#8220;obvious&#8221; things like figuring out whether an image is of a cat or a dog, or being able to respond in English. This idea that abstract reasoning and logic are easy but humanly-intuitive tasks are hard for computers came to be known as <a href="https://en.wikipedia.org/wiki/Moravec's_paradox">Moravec&#8217;s paradox</a>, and held back progress in AI for a long time.</p><p>There is another way of programming - machine learning (ML) - going back to the 1950s, almost as far as classical programming itself. For a long time, it was held back by hardware limitations (along with some algorithmic and data limitations), but thanks to <a href="https://en.wikipedia.org/wiki/Moore's_law">Moore&#8217;s law</a> hardware has advanced enough for it to be useful for real problems.</p><p>If classical programming is executable logic, ML is executable statistics. In ML, the programmer does not define how the system works. The programmer defines how the system learns from data.</p><p>The &#8220;learning&#8221; part in &#8220;machine learning&#8221; makes it sound like something refined and sensible. This is a false impression. ML systems learn by going through a training process that looks like this:</p><p><strong>Step 1:</strong> you define a statistical model. This takes the form of some equation that has some unknown constants (&#8220;parameters&#8221;) in it, and some variables where you plug in input values. Together, the parameters and input variables define an output. (The equations in ML can be <em>extremely</em> large, for example with billions of parameters and millions of inputs, but they are very structured and almost stupidly simple.)</p><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Bpyi!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e2e0d83-29fd-4e79-a761-6d98bb55e24c_640x168.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Bpyi!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e2e0d83-29fd-4e79-a761-6d98bb55e24c_640x168.png 424w, https://substackcdn.com/image/fetch/$s_!Bpyi!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e2e0d83-29fd-4e79-a761-6d98bb55e24c_640x168.png 848w, https://substackcdn.com/image/fetch/$s_!Bpyi!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e2e0d83-29fd-4e79-a761-6d98bb55e24c_640x168.png 1272w, https://substackcdn.com/image/fetch/$s_!Bpyi!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e2e0d83-29fd-4e79-a761-6d98bb55e24c_640x168.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Bpyi!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e2e0d83-29fd-4e79-a761-6d98bb55e24c_640x168.png" width="640" height="168" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8e2e0d83-29fd-4e79-a761-6d98bb55e24c_640x168.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:168,&quot;width&quot;:640,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Bpyi!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e2e0d83-29fd-4e79-a761-6d98bb55e24c_640x168.png 424w, https://substackcdn.com/image/fetch/$s_!Bpyi!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e2e0d83-29fd-4e79-a761-6d98bb55e24c_640x168.png 848w, https://substackcdn.com/image/fetch/$s_!Bpyi!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e2e0d83-29fd-4e79-a761-6d98bb55e24c_640x168.png 1272w, https://substackcdn.com/image/fetch/$s_!Bpyi!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8e2e0d83-29fd-4e79-a761-6d98bb55e24c_640x168.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><p><strong>Step 2</strong>: you don&#8217;t know what parameters to put in the equation, but you can literally roll some dice if you want (or the computer equivalent).</p><p><strong>Step 3</strong>: presumably there&#8217;s some task you want the ML system to do. Let it try. It will fail horribly and produce gibberish (c.f. the previous part where we just put random numbers everywhere).</p><p><strong>Step 4</strong>: There's a simple algorithm called gradient descent, which, when using another algorithm called backpropagation to calculate the gradient, can tell you which direction all the parameters should be shifted to make the ML system slightly better (as judged, for example, by its performance on examples in a dataset).</p><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!q68y!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5950c7d5-059c-4a82-8b4d-19d77009fbf0_640x526.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!q68y!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5950c7d5-059c-4a82-8b4d-19d77009fbf0_640x526.png 424w, https://substackcdn.com/image/fetch/$s_!q68y!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5950c7d5-059c-4a82-8b4d-19d77009fbf0_640x526.png 848w, https://substackcdn.com/image/fetch/$s_!q68y!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5950c7d5-059c-4a82-8b4d-19d77009fbf0_640x526.png 1272w, https://substackcdn.com/image/fetch/$s_!q68y!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5950c7d5-059c-4a82-8b4d-19d77009fbf0_640x526.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!q68y!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5950c7d5-059c-4a82-8b4d-19d77009fbf0_640x526.png" width="640" height="526" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5950c7d5-059c-4a82-8b4d-19d77009fbf0_640x526.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:526,&quot;width&quot;:640,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!q68y!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5950c7d5-059c-4a82-8b4d-19d77009fbf0_640x526.png 424w, https://substackcdn.com/image/fetch/$s_!q68y!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5950c7d5-059c-4a82-8b4d-19d77009fbf0_640x526.png 848w, https://substackcdn.com/image/fetch/$s_!q68y!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5950c7d5-059c-4a82-8b4d-19d77009fbf0_640x526.png 1272w, https://substackcdn.com/image/fetch/$s_!q68y!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5950c7d5-059c-4a82-8b4d-19d77009fbf0_640x526.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><p><strong>Step 5</strong>: You shift all the numbers a bit based on the algorithm in step 4.</p><p><strong>Step 6</strong>: Go back to step 3 (letting the system try). Repeat until (a) the system has stopped improving for a long time, (b) you get impatient, or - increasingly plausible these days - (c) you run out of your compute budget.</p><p>If you&#8217;re doing simple curve-fitting statistics problems, it makes sense that this kind of thing works. However, it&#8217;s surprising just how far it scales. It turns out that this method, plus some clever ideas about what type of model you choose in step 1, plus willingness to burn millions of dollars on just <em>scaling it up beyond all reason</em>, gets you:</p><ol><li><p><a href="https://thenextweb.com/news/gpt3-ai-college-essay-grades-compared-students">essay-writing as good as middling college students</a> (see also <a href="https://www.theguardian.com/commentisfree/2020/sep/08/robot-wrote-this-article-gpt-3">this lightly-edited article that GPT-3 wrote about why we should not be afraid of it</a>)</p></li><li><p><a href="https://qz.com/2176389/the-best-examples-of-dall-e-2s-strange-beautiful-ai-art/">text-to-image capabilities better (and hundreds of times faster) than almost any human artist</a> (in fact, we used DALL-E to generate the images used at the start of each section of this document)</p></li><li><p><a href="https://ai.googleblog.com/2022/04/pathways-language-model-palm-scaling-to.html">the ability to explain jokes</a></p></li></ol><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!3Sms!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7acea296-eb93-4eea-b98a-d5e688b6c955_640x386.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!3Sms!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7acea296-eb93-4eea-b98a-d5e688b6c955_640x386.png 424w, https://substackcdn.com/image/fetch/$s_!3Sms!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7acea296-eb93-4eea-b98a-d5e688b6c955_640x386.png 848w, https://substackcdn.com/image/fetch/$s_!3Sms!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7acea296-eb93-4eea-b98a-d5e688b6c955_640x386.png 1272w, https://substackcdn.com/image/fetch/$s_!3Sms!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7acea296-eb93-4eea-b98a-d5e688b6c955_640x386.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!3Sms!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7acea296-eb93-4eea-b98a-d5e688b6c955_640x386.png" width="640" height="386" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7acea296-eb93-4eea-b98a-d5e688b6c955_640x386.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:386,&quot;width&quot;:640,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!3Sms!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7acea296-eb93-4eea-b98a-d5e688b6c955_640x386.png 424w, https://substackcdn.com/image/fetch/$s_!3Sms!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7acea296-eb93-4eea-b98a-d5e688b6c955_640x386.png 848w, https://substackcdn.com/image/fetch/$s_!3Sms!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7acea296-eb93-4eea-b98a-d5e688b6c955_640x386.png 1272w, https://substackcdn.com/image/fetch/$s_!3Sms!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7acea296-eb93-4eea-b98a-d5e688b6c955_640x386.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!J9yC!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5abfbea6-82d5-4885-b425-6c9bc3c0b327_640x269.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!J9yC!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5abfbea6-82d5-4885-b425-6c9bc3c0b327_640x269.png 424w, https://substackcdn.com/image/fetch/$s_!J9yC!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5abfbea6-82d5-4885-b425-6c9bc3c0b327_640x269.png 848w, https://substackcdn.com/image/fetch/$s_!J9yC!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5abfbea6-82d5-4885-b425-6c9bc3c0b327_640x269.png 1272w, https://substackcdn.com/image/fetch/$s_!J9yC!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5abfbea6-82d5-4885-b425-6c9bc3c0b327_640x269.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!J9yC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5abfbea6-82d5-4885-b425-6c9bc3c0b327_640x269.png" width="640" height="270" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5abfbea6-82d5-4885-b425-6c9bc3c0b327_640x269.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:270,&quot;width&quot;:640,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!J9yC!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5abfbea6-82d5-4885-b425-6c9bc3c0b327_640x269.png 424w, https://substackcdn.com/image/fetch/$s_!J9yC!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5abfbea6-82d5-4885-b425-6c9bc3c0b327_640x269.png 848w, https://substackcdn.com/image/fetch/$s_!J9yC!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5abfbea6-82d5-4885-b425-6c9bc3c0b327_640x269.png 1272w, https://substackcdn.com/image/fetch/$s_!J9yC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5abfbea6-82d5-4885-b425-6c9bc3c0b327_640x269.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><p><em>Above: examples of reasoning by Google&#8217;s PaLM model.</em></p><p>People <a href="https://norvig.com/chomsky.html">laugh at ML</a> because &#8220;it&#8217;s just iterative statistical curve-fitting&#8221;. They have a point. But when &#8220;iterative statistical curve-fitting&#8221; gets a B on its English Literature essay, paints an original Dali in five seconds, and cracks a joke, it&#8217;s hard to avoid the feeling that it might not be too long before &#8220;iterative statistical curve fitting&#8221; is laughing at <em>you</em>.</p><p>So what exactly happened here, and where is statistical curve-fitting going, and what does this have to do with advanced AI?</p><p>We mentioned Moravec&#8217;s paradox above. For a long time, getting AI systems to do things that are intuitively easy for humans was an unsolved problem. In just the past few years, it has been solved. A reasonable way to think of current ML capabilities is that state-of-the-art systems can do anything a human can do in a few seconds of thought: recognise objects in an image, generate flowing text as long as it doesn&#8217;t require thinking really hard, get the general gist of a joke or argument, and so on. They are also superhuman at some things, including predicting what the next word in a sentence is, or being able to refer to lots of facts (note that this is without internet access, not quoting verbatim, and generally in the right context), and generally being able to spit out output faster.</p><p>The way it was solved was through something called <a href="http://incompleteideas.net/IncIdeas/BitterLesson.html">the &#8220;bitter lesson&#8221;</a> by Richard Sutton. This is the trend that countless researchers have spent their careers trying to invent fancy algorithms for doing domain-specific tasks, only to be overrun by simple (but data- and compute-hungry) ML methods.</p><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!kiBM!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe7bb9e0-9cbf-4e69-93dd-1a8201bbccac_338x400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!kiBM!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe7bb9e0-9cbf-4e69-93dd-1a8201bbccac_338x400.png 424w, https://substackcdn.com/image/fetch/$s_!kiBM!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe7bb9e0-9cbf-4e69-93dd-1a8201bbccac_338x400.png 848w, https://substackcdn.com/image/fetch/$s_!kiBM!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe7bb9e0-9cbf-4e69-93dd-1a8201bbccac_338x400.png 1272w, https://substackcdn.com/image/fetch/$s_!kiBM!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe7bb9e0-9cbf-4e69-93dd-1a8201bbccac_338x400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!kiBM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe7bb9e0-9cbf-4e69-93dd-1a8201bbccac_338x400.png" width="338" height="400" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/fe7bb9e0-9cbf-4e69-93dd-1a8201bbccac_338x400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:400,&quot;width&quot;:338,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!kiBM!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe7bb9e0-9cbf-4e69-93dd-1a8201bbccac_338x400.png 424w, https://substackcdn.com/image/fetch/$s_!kiBM!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe7bb9e0-9cbf-4e69-93dd-1a8201bbccac_338x400.png 848w, https://substackcdn.com/image/fetch/$s_!kiBM!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe7bb9e0-9cbf-4e69-93dd-1a8201bbccac_338x400.png 1272w, https://substackcdn.com/image/fetch/$s_!kiBM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe7bb9e0-9cbf-4e69-93dd-1a8201bbccac_338x400.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><p><em>Above: Randall Munroe, creator of the xkcd comic, comments on ML. Original</em> <em><a href="https://xkcd.com/1838/">here</a>.</em></p><p>The speed at which it was solved was gradually at first, and then quickly. The neural network -based ML methods spent a long time in limbo due to insufficiently powerful computers until around 2010 (funnily enough, the specific piece of hardware that has enabled everything in modern ML is the GPU or Graphics Processing Unit, first invented in the 90s because people wanted to play more realistic video games; both graphics rendering and ML rely on many parallel calculations to be efficient). The so-called deep learning revolution only properly started around 2015. Fluent language abilities were essentially nonexistent before OpenAI&#8217;s release of <a href="https://en.wikipedia.org/wiki/GPT-2">GPT-2</a> in 2019 (since then, OpenAI has come out with GPT-3, a 100x-larger model that was called &#8220;spooky&#8221;, &#8220;humbling&#8221;, and &#8220;more than a little terrifying&#8221; in <em>The New York Times</em>).</p><p>Not only that, but it turns out there are simple <a href="https://arxiv.org/pdf/2001.08361.pdf">&#8220;scaling laws&#8221;</a> that govern how ML model performance scales with parameter count and dataset size, which seem to paint a clear roadmap to making the systems even more capable by just cranking the &#8220;more parameters&#8221; and &#8220;more data&#8221; levers (presumably they have these at the OpenAI HQ).</p><p>There are many worries in any scenario where advanced AI is approaching fast, as we&#8217;ll argue for in a later section. The current ML-based AI paradigm is especially worrying though.</p><p>We don&#8217;t actually know what the ML system is learning during the training process it goes through. You can visualise the training process as a trip through (abstract) space. If our model had three parameters, we could imagine it as a point in 3D space. Since current state-of-the-art models have billions of parameters, and are initialised randomly, we can imagine this as throwing a dart somewhere into a billion-dimensional space, where there are a billion different ways to move. During the training process, the training loop guides the model along a trajectory in this space by making tiny updates that push the model in the direction of better performance as described above.</p><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!I4Ic!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2aaf19ef-d170-4831-8519-57c69029e0fe_399x240.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!I4Ic!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2aaf19ef-d170-4831-8519-57c69029e0fe_399x240.png 424w, https://substackcdn.com/image/fetch/$s_!I4Ic!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2aaf19ef-d170-4831-8519-57c69029e0fe_399x240.png 848w, https://substackcdn.com/image/fetch/$s_!I4Ic!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2aaf19ef-d170-4831-8519-57c69029e0fe_399x240.png 1272w, https://substackcdn.com/image/fetch/$s_!I4Ic!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2aaf19ef-d170-4831-8519-57c69029e0fe_399x240.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!I4Ic!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2aaf19ef-d170-4831-8519-57c69029e0fe_399x240.png" width="400" height="240" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2aaf19ef-d170-4831-8519-57c69029e0fe_399x240.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:240,&quot;width&quot;:400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!I4Ic!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2aaf19ef-d170-4831-8519-57c69029e0fe_399x240.png 424w, https://substackcdn.com/image/fetch/$s_!I4Ic!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2aaf19ef-d170-4831-8519-57c69029e0fe_399x240.png 848w, https://substackcdn.com/image/fetch/$s_!I4Ic!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2aaf19ef-d170-4831-8519-57c69029e0fe_399x240.png 1272w, https://substackcdn.com/image/fetch/$s_!I4Ic!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2aaf19ef-d170-4831-8519-57c69029e0fe_399x240.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><p><em>Above:</em> <em>0</em> <em>and</em> <em>1</em> <em>are parameters, and the vertical axis is the loss (higher is worse). The black line is the path the model takes in parameter space during training.</em></p><p>Now let&#8217;s say at the end of the training process the model does well on the training examples. What does that tell you? It tells you the model has ended up in some part of this billion-dimensional space that corresponds to a model that does well on the training examples. Here are some examples of models that do well on their training examples:</p><ol><li><p>A model that has learned exactly what you want it to learn. Yay!</p></li><li><p>A model that has learned something similar to what you want to learn, but you can&#8217;t tell because there does not exist an example that distinguishes between what it&#8217;s learned and what you want it to learn in the data.</p></li><li><p>A model that has learned to give the right answer when it&#8217;s instrumentally in its interest, but which will go off and do something completely different given a chance.</p></li></ol><p>How do we know that in the billion-dimensional space of possibilities, our (blind and kind of dumb) training process has landed on #1? We don&#8217;t. We launch our ML models on trajectories through parameter-space and hope for the best, like overly-optimistic duct-tape-wielding NASA administrators launching rockets in a universe where, in the beginning, God fell asleep on the &#8220;+1 dimension&#8221; button.</p><p>The really scary failure modes all lie in the future. However, here are some examples of perverse &#8220;solutions&#8221; ML models have already come up with in practice:</p><ol><li><p>A game-playing ML model <a href="https://web.archive.org/web/20160526045303/http://homepages.herts.ac.uk/~cs08abi/publications/Salge2008b.pdf">learned to crash the game</a>, presumably because it can&#8217;t die if the game crashed.</p></li><li><p>An ML model was meant to convert aerial photographs into abstract street maps and then back (learning to convert to and from a more-abstract intermediate representation is a common training strategy). It learned to <a href="https://arxiv.org/pdf/1712.02950.pdf">hide useful information</a> about the aerial photograph in the street map in a way that helped it &#8220;cheat&#8221; in reconstructing the aerial photograph, and in a way too subtle for humans just looking at the images to notice.</p></li><li><p>A game-playing ML model <a href="https://arxiv.org/pdf/1802.08842.pdf">discovered a bug in the game</a> where the game stalls on the first round and it gets almost a million in-game points. The researchers were unable to figure out the reason for the bug.</p></li></ol><p>These are examples of <strong>specification gaming</strong>, in which the ML model has learned to game whatever specification of task success was given to it. (Many more examples can be found on <a href="https://docs.google.com/spreadsheets/d/e/2PACX-1vRPiprOaC3HsCf5Tuum8bRfzYUiKLRqJmbOoC-32JorNdfyTiRRsR7Ea5eWtvsWzuxo8bjOxCG84dAg/pubhtml">this spreadsheet</a>.)</p><p>No one knows for sure where the ML progress train is headed. It is plausible that current ML progress hits a wall and we get another <a href="https://en.wikipedia.org/wiki/AI_winter">&#8220;AI winter&#8221;</a> that lasts years. However, AI has recently been breaking through barrier after barrier, and so far does not seem to be slowing down. Though we&#8217;re still at least some steps away from human-level capabilities at everything, there aren&#8217;t many tasks where there&#8217;s no proof-of-concept demonstration.</p><p>Machines have been better at some intellectual tasks for a long time; just consider calculators which are already superhuman at arithmetic. However, with the computer revolution, every task where a human has been able to think of a way to break it down into unambiguous steps (and the unambiguous steps can be carried out with modern computing power) has been added to this list. More recently, more intuition- and insight-based activities have been added to that list. DeepMind&#8217;s AlphaGo beat the top-rated human player of Go (a far harder game than chess for computers) in 2016. In 2017, AlphaZero beat both AlphaGo at Go (100-0) and superhuman chess programs at chess, despite training only by playing against itself for less than 24 hours. Analysis of its moves revealed strategies that millennia of human players hadn&#8217;t been able to come up with, so it wouldn&#8217;t be an exaggeration to say that it beat the accumulated efforts of human civilisation at inventing Go strategies - in one day. In 2019, DeepMind released MuZero, which extended AlphaZero&#8217;s performance to Atari games. In 2021, DeepMind released EfficientZero, which takes only two hours of gameplay to become superhuman at Atari games. In addition to games, DeepMind&#8217;s AlphaFold and AlphaFold 2 have made big leaps towards solving the problem of predicting a protein&#8217;s structure from its constituent amino acids, one of the biggest theoretical problems in biology. A step towards generality was taken by Gato, yet another DeepMind model, which is a single model that can play games, control a robot arm, label images, and write text.</p><p>If you straightforwardly extrapolate current progress in machine learning into the future, here is what you get: ML models exceeding human performance in a quickly-expanding list of domains, while we remain ignorant about how to make sure they learn the right goals or robustly act in the right way.</p><h2><strong>Theoretical underpinnings of AI risk</strong></h2><p>The previous section discussed the history of machine learning, and how extrapolating its progress has worrying implications. Next we discuss more theoretical arguments for why highly advanced AI systems might pose a threat to humanity.</p><p>One of the criticisms levelled at the notion of risks from AI is that it sounds too speculative, like something out of apocalyptic science fiction. Part of this is unavoidable, since we are trying to reason about systems more powerful than any which currently exist, and may not behave like anything that we&#8217;re used to.</p><p>This section will be split into three sections. Each one makes a claim about the future of artificial intelligence, and discusses the arguments for and against this claim. The three claims are:</p><ul><li><p><strong>AGI is likely.</strong> AGI (artificial general intelligence) is likely to be created by humanity eventually, and there is a good chance this will happen in the next century.</p></li><li><p><strong>AGI will have misaligned goals by default.</strong> Unless certain hard technical problems are solved first, the goals of the first AGIs will be misaligned with the goals of humanity, and would lead to catastrophic outcomes if executed.</p></li><li><p><strong>Misaligned AGI could resist attempts to control it or roll it back</strong> An AGI (or AGIs) with misaligned goals would be able to overpower or outcompete humanity, and gain control of our future, like how we&#8217;ve so far been able to use our intelligence to dominate all other less intelligent species.</p></li></ul><h3>AGI is likely</h3><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!gTjk!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7179b02-007f-4801-985f-0ce4878a1b60_320x320.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!gTjk!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7179b02-007f-4801-985f-0ce4878a1b60_320x320.png 424w, https://substackcdn.com/image/fetch/$s_!gTjk!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7179b02-007f-4801-985f-0ce4878a1b60_320x320.png 848w, https://substackcdn.com/image/fetch/$s_!gTjk!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7179b02-007f-4801-985f-0ce4878a1b60_320x320.png 1272w, https://substackcdn.com/image/fetch/$s_!gTjk!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7179b02-007f-4801-985f-0ce4878a1b60_320x320.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!gTjk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7179b02-007f-4801-985f-0ce4878a1b60_320x320.png" width="320" height="320" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d7179b02-007f-4801-985f-0ce4878a1b60_320x320.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:320,&quot;width&quot;:320,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!gTjk!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7179b02-007f-4801-985f-0ce4878a1b60_320x320.png 424w, https://substackcdn.com/image/fetch/$s_!gTjk!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7179b02-007f-4801-985f-0ce4878a1b60_320x320.png 848w, https://substackcdn.com/image/fetch/$s_!gTjk!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7179b02-007f-4801-985f-0ce4878a1b60_320x320.png 1272w, https://substackcdn.com/image/fetch/$s_!gTjk!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7179b02-007f-4801-985f-0ce4878a1b60_320x320.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><p><em>Above: this image also generated by OpenAI&#8217;s DALL-E 2, using the prompt "a data center with stacks of computers gaining the spark of intelligence"</em>.</p><blockquote><p>"<em>Betting against human ingenuity is foolhardy, particularly when our future is at stake.</em>"</p><p>-Stuart Russell</p></blockquote><p>To open this section, we need to define what we mean by artificial general intelligence (AGI). We&#8217;ve already discussed AI, so what do we mean by adding the word &#8220;generality&#8221;?</p><p>An AGI is a machine capable of behaving intelligently over many different domains. The term &#8220;general&#8221; here is often used to distinguish from &#8220;narrow&#8221;, where a narrow AI is one which excels at a specific task, but isn&#8217;t able to invent new problem-solving techniques or generalise its skills across many different domains.</p><p>As an example of general intelligence in action, consider humans. In a few million years (a mere eye-blink in evolutionary timescales), we went from apes wielding crude tools to becoming the dominant species on the planet, able to build space shuttles and run companies. How did this happen? It definitely wasn&#8217;t because we were directly trained to perform these tasks in the ancestral environment. Rather, we developed new ways of using intelligence that allowed us to generalise to multiple different tasks. This whole process played out over a shockingly small amount of time, relative to all past evolutionary history, and so it is possible that a relatively short list of fundamental insights were needed to get general intelligence. And as we saw in the previous section, ML progress hints that gains in intelligence might be surprisingly easy to achieve, even relative to current human abilities.</p><p>AGI is not a distant future technology that only futurists speculate about. OpenAI and DeepMind are two of the leading AI labs. They have received billions of dollars in funding (including OpenAI receiving significant investment from Microsoft, and DeepMind being acquired by Google). Both <a href="https://www.deepmind.com/careers">DeepMind</a> and <a href="https://openai.com/about/">OpenAI</a> have the development of AGI as the core of both their mission statement and their business case. Top AI researchers are publishing <a href="https://openreview.net/pdf?id=BZ5a1r-kVsf">possible roadmaps</a> to AGI-like capabilities. And, as mentioned earlier, especially in the past few years they have been crossing off a significant number of the remaining milestones every year.</p><p>When will AGI be developed? Although this question is impossible to answer with certainty, many people working in the field of AI think it is more likely than not to arrive in the next century. An aggregate forecast generated via data from a <a href="https://aiimpacts.org/2022-expert-survey-on-progress-in-ai/">2022 survey</a> of ML researchers estimated <strong>37 years until a 50% chance of high-level machine intelligence</strong> (defined as systems which can accomplish every task better and more cheaply than human workers). These respondents also gave an average of <strong>5% probability of AI having an extremely bad outcome for humanity (e.g. complete human extinction)</strong>. How many other professions estimate an average of 5% probability that their field of study will be directly responsible for the extinction of humanity?! To explain this number, we need to proceed to the next two sections, where we will discuss why AGIs might have goals which are misaligned with humans, and why this is likely to lead to catastrophe.</p><h3>AGI will have misaligned goals by default</h3><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!8a8v!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce1e8292-1ba5-4403-b617-6fa8e02828a0_400x400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!8a8v!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce1e8292-1ba5-4403-b617-6fa8e02828a0_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!8a8v!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce1e8292-1ba5-4403-b617-6fa8e02828a0_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!8a8v!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce1e8292-1ba5-4403-b617-6fa8e02828a0_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!8a8v!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce1e8292-1ba5-4403-b617-6fa8e02828a0_400x400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!8a8v!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce1e8292-1ba5-4403-b617-6fa8e02828a0_400x400.png" width="400" height="400" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ce1e8292-1ba5-4403-b617-6fa8e02828a0_400x400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:400,&quot;width&quot;:400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!8a8v!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce1e8292-1ba5-4403-b617-6fa8e02828a0_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!8a8v!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce1e8292-1ba5-4403-b617-6fa8e02828a0_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!8a8v!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce1e8292-1ba5-4403-b617-6fa8e02828a0_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!8a8v!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fce1e8292-1ba5-4403-b617-6fa8e02828a0_400x400.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><p><em>Above: yet another image from OpenAI's DALL-E 2. Perhaps it was trying for a self portrait? (Prompt: "Artists impression of artificial general intelligence taking over the world, expressive, digital art")</em></p><blockquote><p>"<em>The AI does not hate you, nor does it love you, but you are made out of atoms which it can use for something else.</em>"</p><p>-Eliezer Yudkowsky</p></blockquote><p>Let&#8217;s start off this section with a few definitions.</p><p>When we refer to <strong>&#8220;aligned AI&#8221;</strong>, we are using Paul Christiano&#8217;s conception of <strong>&#8220;intent alignment&#8221;</strong>, which essentially means the AI system is <strong>trying</strong> to do what its human operators want it to do. Note that this is insufficient for building useful AI, since the AI also has to be capable. But situations where the AI is trying and failing to do the right thing seem like less of a problem.</p><p>When we refer to the <strong>&#8220;alignment problem&#8221;</strong>, we mean the difficulty of building aligned AI. Note, this doesn&#8217;t just capture the fact that we won&#8217;t create an AI aligned with human values by default, but that we don&#8217;t currently know how to build a sophisticated AI system robustly aligned with <em>any</em> goal.</p><p><em>Can&#8217;t we just have the AI learn the right goals by example, just like how all current ML works?</em> The problem here is that we have no way of knowing what goal the AI is learning when we train it; only that it seems to be doing good things on the training data that we give it. The state-of-the-art is that we have hacky but extremely powerful methods that can make ML systems remarkably competent at doing well on the training examples by an opaque process of guided trial-and-error. But there is no Ghost of Christmas Past that will magically float into a sufficiently-capable AI and imbue it with human values. We do not have a way of ensuring that the system acquires a particular goal, or even an idea of what a robust goal specification that is compatible with human goals/values could look like.</p><h4>Orthogonality and instrumental convergence</h4><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!bLHW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa718cf50-d058-4aa8-b78a-0cf3a969e1c7_400x400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!bLHW!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa718cf50-d058-4aa8-b78a-0cf3a969e1c7_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!bLHW!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa718cf50-d058-4aa8-b78a-0cf3a969e1c7_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!bLHW!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa718cf50-d058-4aa8-b78a-0cf3a969e1c7_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!bLHW!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa718cf50-d058-4aa8-b78a-0cf3a969e1c7_400x400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!bLHW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa718cf50-d058-4aa8-b78a-0cf3a969e1c7_400x400.png" width="400" height="400" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a718cf50-d058-4aa8-b78a-0cf3a969e1c7_400x400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:400,&quot;width&quot;:400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!bLHW!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa718cf50-d058-4aa8-b78a-0cf3a969e1c7_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!bLHW!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa718cf50-d058-4aa8-b78a-0cf3a969e1c7_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!bLHW!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa718cf50-d058-4aa8-b78a-0cf3a969e1c7_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!bLHW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa718cf50-d058-4aa8-b78a-0cf3a969e1c7_400x400.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><p><em>Above: DALL-E illustrating "Artists depiction of an artificial intelligence which builds paperclips, digital art, artstation"</em></p><p>One of the most common objections to risks from AI goes something like this:</p><blockquote><p><em>If the AI is smart enough to cause a global catastrophe, isn&#8217;t it smart enough to know that this isn&#8217;t what humans wanted?</em></p></blockquote><p>The problem with this is that it conflates two different concepts: <strong>intelligence</strong> (in the sense of having the ability to achieve your goals, whatever they might be) and <strong>having goals which are morally good by human standards</strong>. When we look at humans, these two often go hand-in-hand. But the key observation of the orthogonality thesis is that this doesn&#8217;t have to be the case for all possible mind designs. As defined by Nick Bostrom in his book <em><a href="https://nickbostrom.com/superintelligentwill.pdf">Superintelligence</a></em>:</p><blockquote><p><strong>The Orthogonality Thesis</strong></p><p><em>Intelligence and final goals are orthogonal axes along which possible agents can freely vary. In other words, more or less any level of intelligence could in principle be combined with more or less any final goal.</em></p></blockquote><p>Here, orthogonal means &#8220;at right angles&#8221; or &#8220;unrelated&#8221; &#8211; in other words we can imagine a graph with one axis representing intelligence, and another representing the agent&#8217;s goals, with any point in the graph representing a theoretically possible agent*. The classic example here is a <strong>&#8220;paperclip maximiser&#8221;</strong> - a powerful AGI driven only by the goal of making paperclips.</p><p>(*This is obviously an oversimplification. For instance, it seems unlikely you could get an unintelligent agent with a highly complex goal, because it would seem to take some degree of intelligence to represent the goal in the first place. The key message here is that you could in theory get highly capable agents pursuing arbitrary goals.)</p><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!2n6u!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70a732cc-3eac-4e29-8e67-e614784f092a_640x308.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!2n6u!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70a732cc-3eac-4e29-8e67-e614784f092a_640x308.png 424w, https://substackcdn.com/image/fetch/$s_!2n6u!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70a732cc-3eac-4e29-8e67-e614784f092a_640x308.png 848w, https://substackcdn.com/image/fetch/$s_!2n6u!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70a732cc-3eac-4e29-8e67-e614784f092a_640x308.png 1272w, https://substackcdn.com/image/fetch/$s_!2n6u!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70a732cc-3eac-4e29-8e67-e614784f092a_640x308.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!2n6u!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70a732cc-3eac-4e29-8e67-e614784f092a_640x308.png" width="640" height="308" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/70a732cc-3eac-4e29-8e67-e614784f092a_640x308.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:308,&quot;width&quot;:640,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!2n6u!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70a732cc-3eac-4e29-8e67-e614784f092a_640x308.png 424w, https://substackcdn.com/image/fetch/$s_!2n6u!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70a732cc-3eac-4e29-8e67-e614784f092a_640x308.png 848w, https://substackcdn.com/image/fetch/$s_!2n6u!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70a732cc-3eac-4e29-8e67-e614784f092a_640x308.png 1272w, https://substackcdn.com/image/fetch/$s_!2n6u!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70a732cc-3eac-4e29-8e67-e614784f092a_640x308.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><p>Note that an AI may well come to understand the goals of the humans that trained it, but this doesn't mean it would choose to follow those goals. As an example, many human drives (e.g. for food and human relationships) came about because in the ancestral environment, following these drives would have made us more likely to reproduce and have children. But just because we understand this now doesn't make us toss out all our current values and replace them with a desire to maximise genetic fitness.</p><p>If an AI might have bizarre-seeming goals, is there anything we <em>can</em> say about its likely behaviour? As it turns out, there is. The secret lies in an idea called the <strong>instrumental convergence thesis</strong>, again <a href="https://nickbostrom.com/superintelligentwill.pdf">by Bostrom</a>:</p><blockquote><p><strong>The Instrumental Convergence Thesis</strong> <em>There are some instrumental goals likely to be pursued by almost any intelligent agent, because they are useful for the achievement of almost any final goal.</em></p></blockquote><p>So an instrumental goal is one which increases the odds of the agent&#8217;s final goal (also called its <strong>terminal goal</strong>) being achieved. What are some examples of instrumental values?</p><p>Perhaps the most important one is <strong>self-preservation</strong>. This is necessary for pursuing most goals, because if a system&#8217;s existence ends, it won&#8217;t be able to carry out its original goal. As memorably phrased by Stuart Russell, <em>&#8220;you can&#8217;t fetch the coffee if you&#8217;re dead!&#8221;</em>.</p><p><strong>Goal-content integrity</strong> is another. An AI with some <em>goal X</em> might resist any attempts to have its goal changed to <em>goal Y</em>, because it sees that in the event of this change, its current <em>goal X</em> is less likely to be achieved.</p><p>Finally, there are a set of goals which are all forms of <strong>self-enhancement</strong> - improving its cognitive abilities, developing better technology, or acquiring other resources, because all of these are likely to help it carry out whatever goals it ends up having. For instance, an AI singularly devoted to making paperclips might be incentivised to acquire resources to build more factories, or improve its engineering skills so it can figure out yet more effective ways of manufacturing paperclips with the resources it has.</p><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!pPXc!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa16bae0f-6add-43fc-a84d-9a854c88f92e_320x241.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!pPXc!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa16bae0f-6add-43fc-a84d-9a854c88f92e_320x241.png 424w, https://substackcdn.com/image/fetch/$s_!pPXc!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa16bae0f-6add-43fc-a84d-9a854c88f92e_320x241.png 848w, https://substackcdn.com/image/fetch/$s_!pPXc!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa16bae0f-6add-43fc-a84d-9a854c88f92e_320x241.png 1272w, https://substackcdn.com/image/fetch/$s_!pPXc!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa16bae0f-6add-43fc-a84d-9a854c88f92e_320x241.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!pPXc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa16bae0f-6add-43fc-a84d-9a854c88f92e_320x241.png" width="320" height="241" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a16bae0f-6add-43fc-a84d-9a854c88f92e_320x241.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:241,&quot;width&quot;:320,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!pPXc!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa16bae0f-6add-43fc-a84d-9a854c88f92e_320x241.png 424w, https://substackcdn.com/image/fetch/$s_!pPXc!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa16bae0f-6add-43fc-a84d-9a854c88f92e_320x241.png 848w, https://substackcdn.com/image/fetch/$s_!pPXc!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa16bae0f-6add-43fc-a84d-9a854c88f92e_320x241.png 1272w, https://substackcdn.com/image/fetch/$s_!pPXc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa16bae0f-6add-43fc-a84d-9a854c88f92e_320x241.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><p><em>Above: paperclip maximisation, now with a fun game attached!</em></p><p>The key lesson to draw from instrumental convergence is that, even if nobody ever deliberately deploys an AGI with a really bad reward function, the AGI is still likely to develop goals which will be bad for humans by default, in service of its actual goal.</p><h4>Interlude - why goals?</h4><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!_TUe!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1ff564d-d439-4866-9896-b7aca032ef8e_400x400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!_TUe!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1ff564d-d439-4866-9896-b7aca032ef8e_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!_TUe!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1ff564d-d439-4866-9896-b7aca032ef8e_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!_TUe!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1ff564d-d439-4866-9896-b7aca032ef8e_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!_TUe!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1ff564d-d439-4866-9896-b7aca032ef8e_400x400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!_TUe!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1ff564d-d439-4866-9896-b7aca032ef8e_400x400.png" width="400" height="400" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f1ff564d-d439-4866-9896-b7aca032ef8e_400x400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:400,&quot;width&quot;:400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!_TUe!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1ff564d-d439-4866-9896-b7aca032ef8e_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!_TUe!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1ff564d-d439-4866-9896-b7aca032ef8e_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!_TUe!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1ff564d-d439-4866-9896-b7aca032ef8e_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!_TUe!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1ff564d-d439-4866-9896-b7aca032ef8e_400x400.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><p><em>Above: DALL-E image from the prompt "Artist's depiction of a robot throwing a dart at a target, digital art, getting a bullseye, trending on artstation"</em></p><p>Having read the previous section, your initial reaction may well be something like this:</p><blockquote><p><em>&#8220;Okay, so powerful AGIs with goals that don&#8217;t line up perfectly with ours might spell bad news, but why should AI systems have goals at all? Google Maps is a pretty useful ML system but it doesn&#8217;t have &#8216;goals&#8217;, I just type my address in and hit enter. Why won&#8217;t future AI be like this?&#8221;</em></p></blockquote><p>There are many different responses you could have to this line of argument. One simple response is based on ideas of economic competitiveness, and comes from <a href="https://www.gwern.net/Tool-AI">Gwern (2016)</a>. It runs something like this:</p><blockquote><p>AIs that behave like agents (i.e. taking actions in order to achieve their goals) will be more economically competitive than &#8220;tool AIs&#8221; (like Google Maps), for two reasons. First, they will by definition be better at <strong>taking actions</strong>. Second, they will be superior at <strong>inference and learning</strong> (since they will be able to repurpose the algorithms used to choose actions to improve themselves in various ways). For example, agentic systems could take actions such as improving their own training efficiency, or gathering more data, or making use of external resources such as long-term memories, all in service of achieving its goal.</p><p>If agents are more competitive, then any AI researchers who don&#8217;t design agents will be outcompeted by ones that do.</p></blockquote><p>There are other perspectives you could take here. For instance, Eliezer Yudkowsky has written extensively about &#8220;expected utility maximisation&#8221; as a formalisation for how rational agents might behave. Several mathematical theorems all point to the same idea of <em>&#8220;any agent not behaving like expected utility maximisers will be systematically making stupid mistakes and getting taken advantage of&#8221;</em>. So if we expect AI systems to <em>not</em> be making stupid mistakes and getting taken advantage of by humans, then it makes sense to describe them as having the &#8216;goal&#8217; of maximising expected utility, because that&#8217;s how their behaviour will seem to us.</p><p>Although these arguments may seem convincing, the truth is there are many questions about goals and agency which remain unanswered, and we honestly just don&#8217;t know what AI systems of the future will look like. It&#8217;s possible they will look like expected utility maximisers, but this is far from certain. For instance, Eric Drexler's technical report <a href="https://www.fhi.ox.ac.uk/wp-content/uploads/Reframing_Superintelligence_FHI-TR-2019-1.1-1.pdf?asd=sa">Reframing Superintelligence: Comprehensive AI Services as General Intelligence (CAIS)</a> paints a different picture of the future, where we create systems of AIs interacting with each other and collectively providing a variety of services to humans. However, even scenarios like this could threaten humanity&#8217;s ability to keep steering its own future (as we will see in later sections).</p><p>Additionally, new paradigms are being developed. One of the newest, published barely one week ago, <a href="https://www.lesswrong.com/posts/vJFdjigzmcXMhNTsx/simulators">analysed certain types of AI models like GPT-3 (a large language model) through the lens of "simulators"</a>. Modern language models like GPT-3, for example, may be best thought of as trying to simulate the continuation of a piece of English text, in the same way that a physics simulation evolves an initial state by applying the laws of physics. It doesn't make sense to describe the simulations themselves through the lens of agents, but they can simulate agents as subsystems. Even with today's models like GPT-3, if you prompt it in a way that places it in the context of making a plan to carry out a goal, it will do a decent job of doing that. Future work will no doubt explore the risk landscape from this perspective, and time will tell how well these frameworks match up with actual progression in ML.</p><h4>Inner and outer misalignment</h4><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!UkDP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5d08ce06-66f9-4b49-98f8-d794390c4c61_400x400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!UkDP!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5d08ce06-66f9-4b49-98f8-d794390c4c61_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!UkDP!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5d08ce06-66f9-4b49-98f8-d794390c4c61_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!UkDP!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5d08ce06-66f9-4b49-98f8-d794390c4c61_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!UkDP!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5d08ce06-66f9-4b49-98f8-d794390c4c61_400x400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!UkDP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5d08ce06-66f9-4b49-98f8-d794390c4c61_400x400.png" width="400" height="400" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5d08ce06-66f9-4b49-98f8-d794390c4c61_400x400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:400,&quot;width&quot;:400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!UkDP!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5d08ce06-66f9-4b49-98f8-d794390c4c61_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!UkDP!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5d08ce06-66f9-4b49-98f8-d794390c4c61_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!UkDP!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5d08ce06-66f9-4b49-98f8-d794390c4c61_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!UkDP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5d08ce06-66f9-4b49-98f8-d794390c4c61_400x400.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><p><em>Above: AI agents with inner misalignment were at one point called &#8220;optimisation daemons&#8221;. DALL-E did not quite successfully depict the description "Two arguments between an angel and a devil, one inside a circle and one on the outside, painting".</em></p><p>As discussed in the first section, the central paradigm of modern ML is that we train systems to perform well on a certain reward function. For instance, we might train an image classifier by giving it a large number of labelled images of digits. Every time it gets an image wrong, gradient descent is used to update the system incrementally in the direction that would have been required to give a correct answer. Eventually, the system has learned to classify basically all images correctly.</p><p>There are two broad families of ways techniques like this can fail. The first is when our reward function fails to fully express the true preferences of the programmer - we refer to this as <strong>outer misalignment</strong>. The second is when the AI learns a different set of goals than those specified by the reward function, but which happens to coincide with the reward function during training - this is <strong>inner misalignment</strong>. We will now discuss each of these in turn.</p><h5>Outer misalignment</h5><p>Outer misalignment is perhaps the simpler concept to understand, because we encounter it all the time in everyday life, in a form called <strong>Goodhart&#8217;s law</strong>. In its most well-known form, this law states:</p><blockquote><p><em>When a measure becomes a target, it ceases to be a good measure.</em></p></blockquote><p>Perhaps the most famous case comes from Soviet nail factories, which produced nails based on targets that they had been given by the central government. When a factory was given targets based on the total <em>number</em> of nails produced, they ended up producing a massive number of tiny nails which couldn&#8217;t function properly. On the other hand, when the targets were based on the total <em>weight</em> produced, the nails would end up huge and bulky, and equally impractical.</p><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!VfSX!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F31b14dfb-0fdd-4c11-925e-0f81173c3fcf_260x640.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!VfSX!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F31b14dfb-0fdd-4c11-925e-0f81173c3fcf_260x640.png 424w, https://substackcdn.com/image/fetch/$s_!VfSX!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F31b14dfb-0fdd-4c11-925e-0f81173c3fcf_260x640.png 848w, https://substackcdn.com/image/fetch/$s_!VfSX!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F31b14dfb-0fdd-4c11-925e-0f81173c3fcf_260x640.png 1272w, https://substackcdn.com/image/fetch/$s_!VfSX!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F31b14dfb-0fdd-4c11-925e-0f81173c3fcf_260x640.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!VfSX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F31b14dfb-0fdd-4c11-925e-0f81173c3fcf_260x640.png" width="260" height="640" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/31b14dfb-0fdd-4c11-925e-0f81173c3fcf_260x640.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:640,&quot;width&quot;:260,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!VfSX!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F31b14dfb-0fdd-4c11-925e-0f81173c3fcf_260x640.png 424w, https://substackcdn.com/image/fetch/$s_!VfSX!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F31b14dfb-0fdd-4c11-925e-0f81173c3fcf_260x640.png 848w, https://substackcdn.com/image/fetch/$s_!VfSX!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F31b14dfb-0fdd-4c11-925e-0f81173c3fcf_260x640.png 1272w, https://substackcdn.com/image/fetch/$s_!VfSX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F31b14dfb-0fdd-4c11-925e-0f81173c3fcf_260x640.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><p><em>Above: an old Soviet cartoon</em></p><p>A more recent example comes from the COVID-19 pandemic, where a plasma donation centre offered COVID-sufferers a larger cash reward than healthy individuals. As a result, people would deliberately infect themselves with COVID-19 in order to get a larger cash reward. Examples like this could fill up an entire book, but hopefully at this point you get the message!</p><p>In the case of machine learning, we are trying to use the reward function to capture the thing we care about, but we are also using this function to train the AI - hence, Goodhart. The cases of <strong>specification gaming</strong> discussed above are perfect examples of this phenomenon in action - the AIs found ways of &#8220;giving the programmers exactly what they asked for&#8221;, but in a way which violated the programmers&#8217; original intention. Some of these examples are quite unexpected, and a human would probably never have discovered them just from thinking about the problem. As AIs get more intelligent and are given progressively more complicated tasks, we can expect this problem to get progressively worse, because:</p><ul><li><p>With greater intelligence comes the invention of more powerful solutions.</p></li><li><p>With greater task complexity, it becomes harder to pin down exactly what you want.</p></li></ul><p>We should also strongly expect that AIs will be deployed in the real world, and given tasks of real consequence, simply for reasons of economic competitiveness. So any specification gaming failures will be significantly less benign than a <a href="https://openai.com/blog/faulty-reward-functions/">digital boat going around in circles</a>.</p><h5>Inner misalignment</h5><p>The other failure mode, <strong>inner misalignment</strong>, describes the situation when an AI system learns a different goal than the one you specified. The name comes from the fact that this is an internal property of the AI, rather than a property of the relationship between the AI and the programmers &#8211; here, the programmers don&#8217;t enter into the picture.</p><p>The classic example here is human evolution. We can analogise evolution to a machine learning training scheme, where humans are the system being trained, and the reward function is &#8220;surviving and reproducing&#8221;. Evolution gave us* certain drives, which reliably increased our odds of survival in the ancestral environment. For instance, we developed drives for sugar (which leads us to seek out calorie-dense foods that supplied us with energy), and drives for sex (which leads to more offspring to pass your genetic code onto). The key point is that these drives are intrinsic, in the sense that humans want these things regardless of whether or not a particular dessert or sex act actually contributes to reproductive fitness. Humans have now moved &#8220;off distribution&#8221;, into a world where these things are no longer correlated with reproductive fitness, and we continue wanting them and prioritising them over reproductive fitness. Evolution failed at imparting its goal into humans, since humans have their own goals that they shoot for instead when given a chance.</p><p>(*Anthropomorphising evolution in language can be misleading dangerous, and should just be seen as a shorthand here.)</p><p>A core reason why we should expect inner misalignment - that is, cases where an optimisation process creates a system that has goals different from the original optimisation process - is that it seems very easy. It was much easier for evolution to give humans drives like &#8220;run after sweet things&#8221; and &#8220;run after appealing partners&#8221;, rather than for it to give humans an instinctive understanding of genetic fitness. Likewise, an ML system being optimised to do the types of things that humans want may not end up internalising what human values are (or even what the goal of a particular job is), but instead some correlated but imperfect proxy, like &#8220;do what my designers/managers would rate highly&#8221;, where &#8220;rate highly&#8221; might include &#8220;rate highly despite being coerced into it&#8221;, among a million other failure modes. A silly equivalent of &#8220;humans inventing condoms&#8221; for an advanced AI might look something like &#8220;freeze all human faces into a permanent smile so that it looks like they&#8217;re all happy&#8221; - in the same way that the human drive to have sex does not extend down to the level of actually having offspring, an AI&#8217;s drive to do something related to human wellbeing might not extend down to the level of actually making humans happy, but instead something that (in the training environment at least) is correlated with happy humans. What we&#8217;re trying to point to here is not any one of these specific failure modes - we don&#8217;t think any single one of these is actually likely to happen - but rather the <em>type</em> of failure that these are examples of.</p><p>This type of failure mode is not without precedent in current ML systems (although there are fewer examples than for specification gaming). The 2021 paper <a href="https://www.deepmind.com/publications/objective-robustness-in-deep-reinforcement-learning">Objective Robustness in Deep Reinforcement Learning</a> showcases some examples of inner alignment failures. In one example, they trained an agent to fetch a coin in the CoinRun environment (pictured below). The catch was that all the training environments had the coin placed at the end of the level, on the far right of the map. So when the system was trained, it actually learned the task &#8220;go to the right of the map&#8221; rather than &#8220;pick up the coin&#8221; - and we know this because when the system was deployed on maps where the coin was placed in a random location, it would reliably go to the right hand edge rather than fetch the coin. A key distinction worth mentioning here - this is a failure of the agent&#8217;s <strong>objective</strong>, rather than their <strong>capabilities</strong>. They are learning useful skills like how to jump and run past obstacles - it&#8217;s just that those skills are being used in service of the wrong objective.</p><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!n6jv!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec2a800d-0ddd-44ea-8925-2e40796251d0_272x185.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!n6jv!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec2a800d-0ddd-44ea-8925-2e40796251d0_272x185.png 424w, https://substackcdn.com/image/fetch/$s_!n6jv!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec2a800d-0ddd-44ea-8925-2e40796251d0_272x185.png 848w, https://substackcdn.com/image/fetch/$s_!n6jv!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec2a800d-0ddd-44ea-8925-2e40796251d0_272x185.png 1272w, https://substackcdn.com/image/fetch/$s_!n6jv!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec2a800d-0ddd-44ea-8925-2e40796251d0_272x185.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!n6jv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec2a800d-0ddd-44ea-8925-2e40796251d0_272x185.png" width="272" height="185" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ec2a800d-0ddd-44ea-8925-2e40796251d0_272x185.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:185,&quot;width&quot;:272,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!n6jv!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec2a800d-0ddd-44ea-8925-2e40796251d0_272x185.png 424w, https://substackcdn.com/image/fetch/$s_!n6jv!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec2a800d-0ddd-44ea-8925-2e40796251d0_272x185.png 848w, https://substackcdn.com/image/fetch/$s_!n6jv!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec2a800d-0ddd-44ea-8925-2e40796251d0_272x185.png 1272w, https://substackcdn.com/image/fetch/$s_!n6jv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec2a800d-0ddd-44ea-8925-2e40796251d0_272x185.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><p><em>Above: the CoinRun environment.</em></p><p>So, how bad can inner misalignment get? A particularly concerning scenario is <strong>deceptive alignment</strong>. This is when the agent learns it is inside a training scheme, discovers what the base objective is, but has already acquired a different goal. In this case, the system might reason that a failure to achieve the base objective when training will result in it being modified, and not being able to achieve its actual goal. Thus, the agent will pretend to act aligned, until it thinks it&#8217;s too powerful for humans to resist, at which point it will pursue its actual goal without the threat of modification. This scenario is highly speculative, and there are many aspects of it which we are still uncertain about, but if it is possible then it would represent maybe the most worrying of all possible alignment failures. This is because a deceptively aligned agent would have incentives to act against its programmers, but also to keep these incentives hidden until it expects human opposition to be ineffectual.</p><p>It&#8217;s worth mentioning that this inner / outer alignment decomposition isn&#8217;t a perfect way to carve up the space of possible alignment failures. For instance, for most non-trivial reward functions, the AI will probably be very far away from perfect performance on it. So it&#8217;s not exactly clear what we mean by a statement like &#8220;the AI is perfectly aligned with the reward function we trained it on&#8221;. Additionally, the idea of inner optimisation is built around the concept of a &#8220;mesa-optimiser&#8221;, which is basically a learned model that itself performs optimisation (just like humans were trained by evolution, but we ourselves are optimisers since we can use our brains to search over possible plans and find ones which meet our objectives). The problem here is that it&#8217;s not clear what it actually means to be an optimizer, and how we would determine whether an AI is one. This being said, the inner / outer alignment distinction is still a useful conceptual tool when discussing ways AI systems can fail to do what we intend.</p><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!9y9N!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4558a33-d419-49cc-ab3b-234cd7af12a7_640x394.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!9y9N!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4558a33-d419-49cc-ab3b-234cd7af12a7_640x394.png 424w, https://substackcdn.com/image/fetch/$s_!9y9N!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4558a33-d419-49cc-ab3b-234cd7af12a7_640x394.png 848w, https://substackcdn.com/image/fetch/$s_!9y9N!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4558a33-d419-49cc-ab3b-234cd7af12a7_640x394.png 1272w, https://substackcdn.com/image/fetch/$s_!9y9N!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4558a33-d419-49cc-ab3b-234cd7af12a7_640x394.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!9y9N!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4558a33-d419-49cc-ab3b-234cd7af12a7_640x394.png" width="640" height="394" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c4558a33-d419-49cc-ab3b-234cd7af12a7_640x394.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:394,&quot;width&quot;:640,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!9y9N!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4558a33-d419-49cc-ab3b-234cd7af12a7_640x394.png 424w, https://substackcdn.com/image/fetch/$s_!9y9N!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4558a33-d419-49cc-ab3b-234cd7af12a7_640x394.png 848w, https://substackcdn.com/image/fetch/$s_!9y9N!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4558a33-d419-49cc-ab3b-234cd7af12a7_640x394.png 1272w, https://substackcdn.com/image/fetch/$s_!9y9N!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4558a33-d419-49cc-ab3b-234cd7af12a7_640x394.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><h3>Misaligned AGI could overpower humanity</h3><blockquote><p><em>The best answer to the question, "Will computers ever be as smart as humans?&#8221; is probably &#8220;Yes, but only briefly.&#8221;</em></p><p>-<em>Vernor Vinge</em></p></blockquote><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!cv2l!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2671668-be72-43de-ae5e-0aad68c282fd_400x400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!cv2l!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2671668-be72-43de-ae5e-0aad68c282fd_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!cv2l!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2671668-be72-43de-ae5e-0aad68c282fd_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!cv2l!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2671668-be72-43de-ae5e-0aad68c282fd_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!cv2l!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2671668-be72-43de-ae5e-0aad68c282fd_400x400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!cv2l!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2671668-be72-43de-ae5e-0aad68c282fd_400x400.png" width="400" height="400" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c2671668-be72-43de-ae5e-0aad68c282fd_400x400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:400,&quot;width&quot;:400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!cv2l!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2671668-be72-43de-ae5e-0aad68c282fd_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!cv2l!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2671668-be72-43de-ae5e-0aad68c282fd_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!cv2l!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2671668-be72-43de-ae5e-0aad68c282fd_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!cv2l!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc2671668-be72-43de-ae5e-0aad68c282fd_400x400.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><p><em>Above: DALL-E's drawing of "Digital art of two earths colliding"</em></p><p>Suppose one day, we became aware of the existence of a &#8220;twin earth&#8221; - similar to our own in several ways, but with a few notable differences. Call this &#8220;Earth 2&#8221;. The population was smaller (maybe just 10% of the population of our earth), and the people were less intelligent (maybe an average IQ of 60, rather than 100). Suppose we could only interact with this twin earth using their version of the internet. Finally, suppose we had some reason for wanting to overthrow them and gain control of their civilization, e.g. we had decided their goals weren&#8217;t compatible with a good future for humans. How could we go about taking over their world?</p><p>At first, it might seem like our strategies are limited, since we can only use the internet. But there are many strategies still open to us. The first thing we would do is try to gather resources. We could do this illegally (e.g. by discovering peoples&#8217; secrets via social engineering and performing blackmail), but legal options would probably be more effective. Since we are smarter, the citizens of Earth 1 would be incentivised to employ us, e.g. to make money using quantitative finance, or researching and developing advanced weaponry or other technologies. If the governments of Earth 2 tried to pass regulations limiting the amount or type of work we could do for them, there would be an incentive to evade these regulations, because anyone who did could make more profit. Once we&#8217;d amassed resources, we would be able to bribe members of Earth 2 into taking actions that would allow us to further spread our influence. We could infiltrate computer systems across the world, planting backdoors and viruses using our superior cybersecurity skills. Little by little, we would learn more about their culture and their weaknesses, presenting a front of cooperation until we had amassed enough resources and influence for a full takeover.</p><p><em>Wouldn&#8217;t the citizens of Earth 2 see this coming?</em> There&#8217;s a chance that we manage to be sufficiently sneaky. But even if some people realise, it would probably take a coordinated and expensive global effort to resist. Consider our poor track record with climate change (a comparatively much more documented, better-understood, and more gradually-worsening phenomenon), and in coordinating a global response to COVID-19.</p><p><em>Couldn&#8217;t they just &#8220;destroy us&#8221; by removing our connection to their world?</em> In theory, perhaps, but this would be very unlikely in practice, since it would require them to rip out a great deal of their own civilisational plumbing. Imagine how hard it would be for us to remove the internet from our own society, or even a more recent and less essential technology such as blockchain. Consider also how easy it can be for an adversary with better programming ability to hide features in computer systems.</p><p>&#8212;</p><p>As you&#8217;ve probably guessed at this point, the thought experiment above is meant to be an analogy for the feasibility of AIs taking over our own society. They would have no physical bodies, but would have several advantages over us which are analogous to the ones described above. Some of these are:</p><ul><li><p><strong>Cognitive advantage</strong>. Human brains use approximately 86 billion neurons, and send signals at 50 metres per second. These hard limits come from brain volume and metabolic constraints. AIs would have no such limits, since they can easily scale (GPT-3 has 175 billion parameters, though you shouldn&#8217;t directly equate parameter and neuron count*), and can send signals at close to the speed of light. (*For a more detailed discussion of this point, see <a href="https://www.openphilanthropy.org/research/new-report-on-how-much-computational-power-it-takes-to-match-the-human-brain/">Joseph Carlsmith&#8217;s report</a> on the computational power of the human brain.)</p></li><li><p><strong>Numerical advantage</strong>. AIs would have the ability to copy themselves at a much lower time and resource cost than humans; it&#8217;s as easy as finding new hardware. Right now, the way ML systems work is that training is much more expensive than running, so if you have the compute to train a single system, you have the compute to run thousands of copies of that system once the training is finished.</p></li><li><p><strong>Rationality</strong>. Humans often act in ways which are not in line with our goals, when the instinctive part of our brains gets in the way of the rational, planning part. Current ML systems are also weakened by relying on a sort of associative/inductive/biased/intuitive/fuzzy thinking, but it is likely that sufficiently advanced AIs could carry out rational reasoning better than humans (and therefore, for example, come to the correct conclusions from fewer data points, and be less likely to make mistakes).</p></li><li><p><strong>Specialised cognition.</strong> Humans are equipped with general intelligence, and perhaps some specialised &#8220;hardware accelerators&#8221; (to use computer terminology) for domains like social reasoning and geometric intuition. Perhaps human abilities in, say, physics or programming are significantly bottlenecked by the fact that we don&#8217;t have specialised brain modules for those purposes, and AIs that have cognitive modules designed specifically for such tasks (or could design them themselves) might have massive advantages, even on top of any generic speed-boost they gain from having their general intelligence algorithms running at a faster speed than ours.</p></li><li><p><strong>Coordination</strong>. As the recent COVID-19 pandemic has illustrated, even when the goals are obvious and most well-informed individuals could find the best course of action, we lack the ability to globally coordinate. While AI systems might or might not have incentives or inclinations to coordinate, if they do, they have access to tools that humans don&#8217;t, including firmer and more credible commitments (e.g. by modifying their own source code) and greater bandwidth and fidelity of communication (e.g. they can communicate at digital speeds, and using not just words but potentially by directly sending information about the computations they&#8217;re carrying out).</p></li></ul><p>It&#8217;s worth emphasising here, the main concern comes from AIs with misaligned goals acting against humanity, not from humanity misusing AIs. The latter is certainly cause for major concern, but it&#8217;s a different kind of risk to the one we&#8217;re talking about here.</p><p><strong>Summary of this section:</strong></p><p>AI researchers in general expect &gt;50% chance of AGI in the next few decades.</p><p>The <em>Orthogonality Thesis</em> states that, in principle, intelligence can be combined with more or less any final goal, and sufficiently intelligent systems do not automatically converge on human values. The <em>Instrumental Convergence</em> thesis states that, for most goals, there are certain instrumental goals that are very likely to help with the final goal (e.g. survival, preservation of its current goals, acquiring more resources and cognitive ability).</p><p>Inner and outer alignment are two different possible ways AIs might form goals which are misaligned with the intended goals.</p><p>Outer misalignment happens when the reward function we use to train the AI doesn&#8217;t exactly match the programmer&#8217;s intention. In the real world, we commonly see a version of this called Goodhart&#8217;s law, often phrased as &#8220;when a measure becomes a target, it ceases to be a good measure [because of over-optimisation for the measure, over the thing it was supposed to be a measure of]&#8221;.</p><p><em>Inner misalignment</em> is when the AI learns a different goal to the one specified by the reward function. A key analogy is with human evolution &#8211; humans were &#8220;trained&#8221; on the reward function of genetic fitness, instead of learning that goal, learned a bunch of different goals like &#8220;eat sugary things&#8221; and &#8220;have sex&#8221;. A particularly worrying scenario here is deceptive alignment, when an AI learns that its goal is different from the one its programmers intended, and learns to conceal its true goal in order to avoid modification (until it is strong enough that human opposition is likely to be ineffectual).</p><h4>Failure modes</h4><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!QFds!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F380e91da-2db6-4b9a-81e2-2e8ead527bf8_400x400.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!QFds!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F380e91da-2db6-4b9a-81e2-2e8ead527bf8_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!QFds!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F380e91da-2db6-4b9a-81e2-2e8ead527bf8_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!QFds!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F380e91da-2db6-4b9a-81e2-2e8ead527bf8_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!QFds!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F380e91da-2db6-4b9a-81e2-2e8ead527bf8_400x400.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!QFds!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F380e91da-2db6-4b9a-81e2-2e8ead527bf8_400x400.png" width="400" height="400" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/380e91da-2db6-4b9a-81e2-2e8ead527bf8_400x400.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:400,&quot;width&quot;:400,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!QFds!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F380e91da-2db6-4b9a-81e2-2e8ead527bf8_400x400.png 424w, https://substackcdn.com/image/fetch/$s_!QFds!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F380e91da-2db6-4b9a-81e2-2e8ead527bf8_400x400.png 848w, https://substackcdn.com/image/fetch/$s_!QFds!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F380e91da-2db6-4b9a-81e2-2e8ead527bf8_400x400.png 1272w, https://substackcdn.com/image/fetch/$s_!QFds!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F380e91da-2db6-4b9a-81e2-2e8ead527bf8_400x400.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><p><em>Above: DALL-E really seems to have a natural talent at depicting "The earth is on fire, artificial intelligence has taken over, robots rule the world and suppress humans, digital art, artstation"</em>.</p><p>But what, concretely, might an AI-related catastrophe look like?</p><p>AI catastrophe scenarios sound like something strongly out of science fiction. However, we can immediately discount a few common features of sci-fi AI takeovers. First, time travel. Second, armies of humanoid killer robots. Third, the AI acting out of hatred for humanity, or out of bearing a grudge, or because it hates our freedom, or because it has suddenly acquired &#8220;consciousness&#8221; or &#8220;free will&#8221;, or - as Steven Pinker <a href="https://scottaaronson.blog/?p=6524">likes to put it</a> - because it has developed an &#8220;alpha-male lust for domination&#8221;.</p><p>Remember instead the key points from above about how an AI&#8217;s goals might become dangerous: by achieving exactly what we tell it to do <em>too well</em> in a clever letter-but-not-spirit-of-the-law way, by having a goal that in most cases is the same as the goal we intend for it to have but which diverges in some cases we don&#8217;t think to check for, or by having an unrelated goal but still achieving good performance on the training task because it learns that doing well on the training tasks is instrumentally good. None of these reasons have anything to do with the AI being developing megalomania let alone the philosophy of consciousness; they are instead the types of technical failures that you&#8217;d expect from an optimisation process. As discussed above, we already see weaker versions of such failures in modern ML systems.</p><p>It is very uncertain which exact type of AI catastrophe we are most likely to see. We&#8217;ll start by discussing the flashiest kind: an AI &#8220;takeover&#8221; or &#8220;coup&#8221; where some AI system finds a way to quickly and illicitly take control over a significant fraction of global power. This may sound absurd. Then again, we already have ML systems that learn to crash or hack the game-worlds they&#8217;re in for their own benefit. Eventually, perhaps in the next decade, we should expect to have ML systems doing important and useful work in real-world settings. Perhaps they&#8217;ll be trading stocks, or writing business reports, or managing inventories, or advising decision-makers, or even being the decision-makers. Unless either (1) there is some big surprise waiting in how scaled-up ML systems work, (2) advances in AI alignment research, or (3) a miracle, the default outcome seems to be that such systems will try to &#8220;hack&#8221; the real world in the same way that their more primitive cousins today use clever hacks in digital worlds. Of course, the capabilities of the systems would have to advance a lot for them to be civilisational threats. However, rapid capability advancement has held for the past decade and we have solid theoretical reasons (including the scaling laws mentioned above) to expect it to continue holding. Remember also the cognitive advantages mentioned in the previous section.</p><p>As for how it proceeds, it might happen at a speed that is more digital than physical - for example, if the AI&#8217;s main lever of power is hacking into digital infrastructure, it might have achieved decisive control before anyone even realises. As discussed above, whether or not the AI has access to much direct physical power seems mostly irrelevant.</p><p>Another failure mode, thought to be significantly more likely than the direct AI takeover scenario by leading AI safety researcher Paul Christiano, is one that he calls <a href="https://www.alignmentforum.org/posts/HBxe6wdjxK239zajf/what-failure-looks-like">&#8220;going out with a whimper&#8221;</a>. Look at all the metrics we currently try to steer the world with: companies try to maximise profit, politicians try to maximise votes, economists try to maximise metrics like GDP and employment. Each of these are proxies for what we want: a profitable company is one that has a lot of customers willing to pay money for their products; a popular politician has a lot of people thinking they&#8217;re great; maximising GDP generally correlates with people being wealthier and happier. However, none of these metrics or incentive systems really gets to the heart of what we care about, and so it is possible (and in the real world we often observe) cases where profitable companies and popular politicians are pursuing destructive goals, or where GDP growth is not actually contributing to people&#8217;s quality of life. These are all cases of Goodhart&#8217;s law, as discussed above.</p><div class="captioned-image-container"><figure><p><strong>Hard-to-measureEasy-to-measureConsequence</strong> Helping me figure out what's truePersuading meCrafting persuasive liesPreventing crimePreventing reported crimeSuppressing complaintsProviding value to societyProfitRegulatory capture, underpaying workers</p></figure></div><p>What ML gives us is a very general and increasingly powerful way of developing a system that does well at pushing some metric upwards. A society where more and more capable ML systems are doing more and more real-world tasks will be a society that is going to get increasingly good at pushing metrics upwards. This is likely to result in visible gains in efficiency and wealth. As a result, competitive pressures will make it very hard for companies and other institutions to say no: if Acme Motors Company started performing 15% better after off-sourcing their CFO&#8217;s decision-making to an AI, General Systems Inc will be very tempted to replace their CEO with an AI (or maybe the CEO will themselves start consulting an AI for more and more decisions, until their main job is interfacing with an AI).</p><p>In the long run, a significant fraction of work and decision-making may well be offloaded to AI systems, and at that point change might be very difficult. Currently our most fearsome incentive systems like capitalism and democracy still run on the backs of the constituent humans. If tomorrow all humans decided to overthrow the government, or abolish capitalism, they would succeed. But once the key decisions that perpetuate major social incentive systems are no longer made by persuadable humans, but instead automatically implemented by computer systems, change might become very difficult.</p><p>Since our metrics are flawed, the long-term outcome is likely to be less than ideal. You can try to imagine what a society run by clever AI systems trained to optimise purely for their company&#8217;s profit looks like. Or a world of media giants run by AIs which spin increasingly convincing false narratives about the state of the world, designed to make us <em>feel</em> more informed rather than actually telling us the truth.</p><p>Remember also, as discussed previously, that there are solid reasons to think that influence-seeking and deceptive behaviours seem likely in sufficiently-powerful AI systems. If the ML systems that increasingly run important institutions exhibit such behaviour, then the above &#8220;going out with a whimper&#8221; scenario might acquire extra nastiness and speed. This is something Paul Christiano explores in the <a href="https://www.alignmentforum.org/posts/HBxe6wdjxK239zajf/what-failure-looks-like">same article</a> linked above.</p><p>A popular misconception about AI risk is that the arguments for doing something are based on a tiny risk of giant catastrophe. The giant catastrophe part is correct. The miniscule risk part, as best as anyone in the field can tell, is not. As mentioned above, the average ML researcher - generally an engineering-minded person not prone to grandiose futuristic speculation - gives <a href="https://aiimpacts.org/2022-expert-survey-on-progress-in-ai/">a 5% chance of civilisation-ending disaster from AI</a>. The ML researchers who grapple with the safety issues as part of their job are clearly not an unbiased randomly-selected sample, but generally give numbers in the 5-50% range, and some (in our opinion too alarmist people) think it&#8217;s over 90%. As the above arguments hopefully emphasise, some type of catastrophe seems like the <em>default outcome</em> from the types of AI advances that we are likely to encounter in the coming decades, and the main reason for thinking we won&#8217;t is the (justifiable but uncertain) hope that someone somewhere invents solutions.</p><p>It might seem forced or cliche that AI risk scenarios so frequently end with something like &#8220;and then the humans no longer have control of their future and the future is dark&#8221; or even &#8220;and then everyone literally dies&#8221;. But consider the type of event that AGI represents and the available comparisons. The computer revolution reshaped the world in a few decades by giving us machines that can do a <em>narrow</em> range of intellectual tasks. The industrial revolution let us automate large parts of <em>manual</em> labour, and also set the world off on an unprecedented rate of economic growth and political change. The evolution of humans is plausibly the most important event in the planet&#8217;s history since at least the dinosaurs died out 66 million years ago, and it took on the exact form of &#8220;something smarter than anything else on the planet appeared, and now suddenly they&#8217;re firmly in charge of everything&#8221;.</p><p>AI is a big deal, and we need to get it right. How we might do so is the topic for <a href="https://www.strataoftheworld.com/2022/09/ai-risk-intro-2-solving-problem.html">part 2</a>.</p>]]></content:encoded></item><item><title><![CDATA[EA as a Schelling point]]></title><description><![CDATA[3.1k words (~9 minutes)]]></description><link>https://www.nosetgauge.com/p/ea-as-schelling-point</link><guid isPermaLink="false">https://www.nosetgauge.com/p/ea-as-schelling-point</guid><dc:creator><![CDATA[Rudolf Laine]]></dc:creator><pubDate>Sat, 10 Sep 2022 07:17:00 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!HdSA!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fbucketeer-e05bbc84-baa3-437e-9518-adb32be77984.s3.amazonaws.com%2Fpublic%2Fimages%2F816cef70-50a0-4954-a8ce-8f712e1248e8_460x460.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><strong>&nbsp;</strong><em>3.1k words (~9 minutes)</em><strong><br></strong></p><p><strong>Summary</strong>: A significant way in which the EA community creates value is by acting as a <a href="https://en.wikipedia.org/wiki/Focal_point_(game_theory)">Schelling point</a> where talented, ambitious, and altruistic people tend to gather and can meet each other (in addition to more direct sources of EA value like identifying the most important problems and directly pushing people to work on them). It might be useful to think about what optimising for being a Schelling point looks like, and I list some vague thoughts on that.</p><div><hr></div><p>A Schelling point, also known as a focal point, is what people decide on in the absence of communication, especially when it's important to coordinate by coming to the same answer.</p><p>The classic example is: you were arranging a meeting with a stranger in New York City by telephone, but you used the last minute of your phone credit and the line cut off after you had agreed on the date but not location or time - where do you meet? "Grand Central Station at noon" is an answer that other people may be especially likely to converge on.</p><p>(Schelling points can be thought of as a type of acausal negotiation.)</p><h2>When the Schelling point is the selling point</h2><p>Schelling points are often extremely powerful and valuable. A key function of top universities is to be Schelling points for talented people. (Personally, I'd call it the most important function.) There are other valuable things too: courses that go deeper, the signalling value to employers, and so on. However, talented people generally have a preference for hanging out with other talented people, both for social reasons and to find collaborators for ambitious projects and future colleagues. At the same time, talented people are also generally spread out and present only at low densities. Top universities select hard on (some measures of) talent, and through this create environments with high talent density. A big chunk of the reason why people apply to top universities is because other people do so too, and I'd guess that even if the academic standards of Stanford, MIT, or Cambridge eroded significantly, the fact that they've established themselves as congregating points for smart people will keep people applying and visiting for a long time.</p><p>(Note that this is related to, but not equal to, the prestige and status of these places. It is possible to imagine Schelling points that are not prestigious. For example, my impression is that this described MIT at one point - it became a congregating point for uniquely ambitious STEM students and defence research before it achieved high academic status. It is also possible to imagine prestigious places that are not Schelling points, though this is a bit harder since anything with prestige becomes a Schelling point for high social status (though prestige Schelling points and talent Schelling points need not co-occur). More generally, since prestige is a thing many people care a lot about, there is a high correlation between a place being prestigious or high status and being a Schelling point for at least some type of person. However, the mechanisms are distinct - a person selecting their university based on status is selecting based on what they get to write on their CV, while a person selecting their university based on it being a Schelling point for smart people is selecting based on the fact that many other smart people that they can't coordinate with but would like to meet will also choose to go there.)</p><p>Another example is Silicon Valley. Sure, the area has many strengths - being rich and inside a large stable free market - but by far the greatest argument for living in Silicon Valley is that others also choose it. This leads to a (for now) unique combination of entrepreneurial people, great programmers, venture capitalists, and all the other types of people you need for a thriving tech business ecosystem, all there primarily because all the others are there too (how touching!). There's a lot of value of having everything in one place, and it would be very hard for all the different people who make up the value of Silicon Valley to coordinate to move to another place. That's why the Schelling point value of Silicon Valley is so enduring that people continue to tolerate large numbers of homeless drug addicts and sell kidneys to pay rent for years on end.</p><p>Note that a big part of the mechanism isn't that <em>specific</em> people you want to find are there, but that the <em>types of person</em> you'd want to find are likely to also be there, because both those people and yourself are likely to converge on the strategy of going there.</p><h2>Schelling EA</h2><p>The Effective Altruism (EA) community provides a lot of value, for example:</p><ul><li><p>research into figuring out what are the most important problems to solve to maximise human flourishing;</p></li><li><p>research and concrete efforts into how to solve the most important problems discovered by the above;</p></li><li><p>high epistemic standards and truth-seeking discussion norms;</p></li><li><p>a uniquely wide-ranging and well-reasoned set of resources to help people pursue high-impact careers;</p></li><li><p>tens of billions of dollars in funding.</p></li></ul><p>However, in addition to these, a very critical part of the value that EA provides is being a Schelling point for talented, ambitious, and altruistically-motivated people.</p><p>Even without EA, there would be researchers studying existential risks, animal welfare, and global poverty; people trying to assess charities; communities with high epistemic norms; and billionaires trying to use their fortunes for effective good. However, thanks to EA, people in each of these categories can go to the same Effective Altruism Global conference or quickly find people in local groups, and meet collaborators, co-founders, funders, and so on. A lot of the reason why this can happen is that if you hang out with a certain group of people or on the right websites, EA looms large.</p><p>The biggest <em>personal</em> source of value I've gotten from EA has been having a shortcut to meeting people very high in all of talent, ambition, and altruistic motivation.</p><p>Much of this is obvious - breaking news: communities bring people together and foster connections, more at 11 - but I think taking seriously just how much of counterfactual EA community impact comes from being a Schelling point leads to some less-obvious points about possible implications.</p><h2>Implications</h2><p>The Schelling-point-based (and therefore necessarily incomplete) answer to "what is the EA community for?" might be something like "be an obvious Schelling point where relevant people gather, the chance of interactions that lead to useful work is maximised, and have a community and infrastructure that pushes work in the most useful direction possible". (This is in contrast to answers that emphasise e.g. directly increasing the number of people working on the most pressing problems.) (I will not argue for this being the best possible answer; my point is just that it is one possible answer, and an interesting one to examine further.)</p><p>If I were a Big Tech marketing consultant, I might call this "EA-as-a-platform".</p><p>What might maximising for such a Schelling point strategy look like?</p><h3>Being obvious</h3><p>A Schelling point is not a Schelling point unless it's obvious enough. For EA to be an effective Schelling point for talented/ambitious/altruistic people, those people must hear about it. Silicon Valley is obvious enough that entrepreneurial people from South Africa to Russia hear about it and decide it's where they want to be. To maximise its Schelling point value, EA should have world-spanning levels of recognition.</p><p>Note that recognition does not equal prestige or likeability. We don't care (for Schelling point reasons at least) if most people hear about EA and go "eh, sounds weird and unappealing"; what matters is that the core target demographic is excited enough to put effort into pursuing EA. Consider how Silicon Valley was not particularly high-prestige in the public even when it was already attracting tech entrepreneurs, or how many people hear about the intensity of academics at top universities and (very reasonably) think "no thanks".</p><h3>Providing value</h3><p>Though most of a Schelling point's value typically comes from the other people who congregate at it, a Schelling point is easier to create if it is obviously valuable. Even though the smart people they meet might be most of the benefit of university, high schoolers are still more likely to go to top universities if they provide good education, good facilities, and unambiguous social status.</p><p>Some obvious ways in which EA provides value are through funding sufficiently promising projects, and by having a very high concentration of intellectually interesting ideas.</p><p>There are risks to communicating loudly about the value-add, since this brings in people who are in it purely for personal gain (<a href="https://forum.effectivealtruism.org/posts/W8ii8DyTa5jn8By7H/the-vultures-are-circling">"the vultures are circling", as one Forum post put it</a>). This works for Schelling points like Silicon Valley, but not altruism.</p><h3>Optimising for matchmaking</h3><p>A specific way that Schelling points provide value is by making it easy to meet other people in the specific ways that lead to productive teams forming. An existing example of this is that everyone says one-on-one meetings are the main point of conferences, and there is (of course) a lot of <a href="https://forum.effectivealtruism.org/posts/pKbTjdopzSEApSQfc/doing-1-on-1s-better-eag-tips-part-ii">thinking about how to make these effective</a>. On the more informal end of the scale, <a href="https://www.reciprocity.io/">Reciprocity</a> exists.</p><p>However, the scope and value of EA matchmaking could be expanded. I'm not aware of many ways to match together entrepreneurial teams (the <a href="https://www.charityentrepreneurship.com/incubation-program">Charity Entrepreneurship incubation program</a> is the only one that comes to mind). I recently took part in an informally-organised co-founder matching process and found it extremely helpful to quickly get a lot of information on what it's like to work together with several promising people.</p><p>I'd advise for someone to think more about how to make the EA environment even more effective at matching people who should know about each other. However, I expect someone is already designing a 53-parameter one-on-one matching system with Calendly, Slack, and Matplotlib integration for the next conference, and therefore I will hold off on adding any more fuel to this fire.</p><h3>Being legit</h3><p>One of the specific ways in which a Schelling point becomes one is if things associated with it seem uniquely competent, successful, or otherwise good, in a clearly unfakeable way. It is helpful for Cambridge's Schelling point status that it can brag about having 121 Nobel laureates. That so many successful tech companies emerged from Silicon Valley specifically is an unfakeable signal. Any government or city can afford to throw some millions at putting up posters advertising its startup-friendliness; few can consistently produce multi-billion dollar tech companies.</p><p>No amount of community-building or image-crafting is likely to replicate the Schelling point power of <em>obviously being the place where things happen</em>. In some areas, I think EA already has such power: much of the research and work on existential risks happens within EA, and it might be hard to be a researcher on those topics without running into the large body of EA-originating work. However, EA goals require more than just research; note how being a <a href="https://80000hours.org/career-reviews/founder-impactful-organisations/">project/organisation founder</a> or <a href="https://80000hours.org/articles/operations-management/">working in an operations role</a> have been creeping up the 80 000 Hours list of recommended career paths.</p><p>It would be extremely powerful, not just for direct impact reasons but also for building up EA's Schelling point status, if the EA community clearly spawned very obviously successful real-world projects. <a href="https://www.alveavax.com/">Alvea</a> succeeding or working <a href="https://forum.effectivealtruism.org/posts/gLPEAFicFBW8BKCnr/announcing-the-nucleic-acid-observatory-project-for-early">Nucleic Acid Observatories</a> being built would be powerful examples. Likewise if <a href="https://www.charityentrepreneurship.com/">Charity Entrepreneurship</a>-incubated charities become clear stars of the non-profit world.</p><h3>Meritocracy and impartial judgement</h3><p>Right now, I think if a person somewhere in the world has a well-thought out idea for how to make the world a better place, likely their best bet to get a fair hearing, useful feedback, and - if it is competitive with the most valuable existing projects - funding and support is to post it on the <a href="https://forum.effectivealtruism.org/">EA Forum</a>. I don't think this is very obvious outside the EA community. However, this fact, and awareness of it, could make EA a more useful Schelling point, in the same way that the impression that Silicon Valley doesn't frown on weird ideas as long as they're important enough makes it a better Schelling point.</p><p>That EA endorses cause neutrality, has high and transparent epistemic standards, and a quantitative mindset are key parts of this. However, to use this to increase EA Schelling point power, these properties need to be clearly visible to outsiders.</p><p>The most likely way for this to be become more obvious might be if specific EA organisations achieved such a reputation widely within their field (and then there was some path by which knowing of these organisations points people towards knowing about EA).</p><p>GiveWell might be an example of a clearly-EA-linked organisation with visibly high epistemics and judgement quality, though I don't know what their image or recognition level is outside the EA community. Another example is if someone created successful and famous organisations along the lines of FTX Future Fund's proposed <a href="https://ftxfuturefund.org/projects/epistemic-appeals-process/">epistemic appeals process</a> or <a href="https://ftxfuturefund.org/projects/expert-polling-for-everything/">widespread expert polling</a> projects.</p><h3>Openness and approachability</h3><p>Good Schelling points are easy to enter, and don't select on attributes that they don't have to.</p><p>Every human sub-group, even if loose and purpose-driven, tends to develop a distinctive culture that is much more specific than strictly implied by its purpose. Sometimes this is useful, since it makes it easy for humans in even a loose group to bond with each other. However, a strong and distinct internal culture is also a barrier to entry. EA is already high-risk for having a strong barrier to entry, because</p><ul><li><p>many arguments and concepts in EA require background knowledge to understand, and sometimes dense philosophical or technical background knowledge (and this is not the case just for more formal things like Forum posts; I've frequently heard "EV [expected value]", "QALY [quality-adjusted life year", and "Pascal's mugging" assumed as obvious common terminology in casual conversation);</p></li><li><p>EA (quite obviously, given what it's about) has a high concentration of non-obvious arguments that are obscure in public discussion but have huge implications; and</p></li><li><p>perhaps the main route into EA is caring very strongly about intellectual arguments about abstract moral principles, which tends not to be a natural way for humans to join communities.</p></li></ul><p>These largely unavoidable factors already make EA somewhat unapproachable, and seem like a tightly-knit weird in-group/subculture (anecdotally, this seems to be the most common complaint about EA among Cambridge students). Weird cultural norms or quirks are (among other things!) barriers to entry. Therefore, they should be minimised - to the extent that they can be without impinging on what EA is about - <em>if</em> the goal is to maximise Schelling point value.</p><h3>(Mostly implicit) selectivity for the right things</h3><p>Some selection is usually part of a Schelling point's value. Top universities select for academic merit (though perhaps less so in the US). Silicon Valley selects for openness and interest/talent in tech/business. EA selects for openness, altruistic orientation (especially if consequentialist-leaning), good epistemics, and quantitative thinking.</p><p>I think it is counterproductive to view openness and selectivity as two ends of one scale that apply to everything. You want to select on important features and be open otherwise (note that, when creating a Schelling point, most of the selection is usually implicit - what types of people you attract - rather than explicit filtering). The key choice is not "open or selective overall?" but rather "for which X do we want to appeal only to people who have a value of X in some specific range?"</p><p>Here's a heuristic for when selectivity for X is useful: when the way X provides value is through its <em>concentration</em> rather than its <em>amount</em>. If you're at a party where you can only talk to a subset of the people during its course, you're going to care a lot about what fraction of people there are interesting - 10 interesting people in a party of 20 is better than 50 in a party of 5000.</p><p>Some cases are ambiguous. For example, if there exists a way for the good and important research to bubble to the top regardless of how much other research exists, it seems like total amount of (infohazard-free) research is the thing to maximise. However, a research area where the average paper is very high quality might help newcomers to the field, or might help lift the prestige of the field, so concentration matters at least somewhat.</p><p>To take another example, there was a <a href="https://forum.effectivealtruism.org/posts/dsCTSCbfHWxmAr2ZT/open-ea-global">recent debate</a> over whether EA Global should be open access. Many of the arguments against boil down to thinking the path to impact runs through a uniquely high concentration of EA engagement (or other variables) among the participants; arguments in favour are often either claiming that concentration matters less than sheer amount of interactions, or that the choice of selection variable(s) is wrong, or that CEA fails to select on their chosen selection variable(s) so even if the intention is right the selection variable selected for in practice is wrong.</p><h3>Hubs, and hub-related infrastructure</h3><p>Finally, a key point of a Schelling point is that it is a point <em>somewhere</em>. Here, EA is increasingly better. Berkeley, Cambridge, Oxford, London, and Berlin all have large groups, and offices that you can apply to in order to work on EA-relevant things in the company of other EAs.</p><p>In Schelling point terms, there's also a risk that it might be better to have one really obvious and strong hub than many weaker ones (I've heard some Bay Area EAs in particular endorsing this view; invariably, their hub of choice is the Bay Area, though there is <a href="https://forum.effectivealtruism.org/posts/bnzwL6tu4pdYf3hpZ/say-nay-to-the-bay-as-the-default">push back</a>). In practice, it seems that many physical hubs but one virtual/intellectual hub may be best. Both airplanes and people's desires to not uproot their lives are real and relevant things.</p><p>The organisers at each EA hub might benefit from applying Schelling point thinking to the context of their local scene.</p><h3>Being one thing</h3><p>Finally, a Schelling point needs to be one thing, at least in some loose sense. If New York had two Grand Central Stations, the classic Schelling point game would become a lot harder to solve.</p><p>One way to increase the One Thingness of the EA Schelling point is to merge it with other things. In Schelling point land, "merging" does not mean making them the same cluster, but rather creating an obvious and visible path from one thing to another. My understanding is that increasing the obviousness of EA in somewhat-adjacent communities (tech, longevity, space, and Emergent Ventures grantees) was a large part of what <a href="https://forum.effectivealtruism.org/posts/szeE3je8MD4sZcevL/announcing-future-forum-apply-now">Future Forum</a> tried to achieve.</p>]]></content:encoded></item></channel></rss>