<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Mind the Future]]></title><description><![CDATA[Understanding the system of the world]]></description><link>https://www.mindthefuture.info</link><image><url>https://substackcdn.com/image/fetch/$s_!DN7E!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd26e2534-f7d1-4746-a6bc-69981acf305f_1024x1024.png</url><title>Mind the Future</title><link>https://www.mindthefuture.info</link></image><generator>Substack</generator><lastBuildDate>Wed, 06 May 2026 10:34:22 GMT</lastBuildDate><atom:link href="https://www.mindthefuture.info/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Richard Ngo]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[mindthefuture@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[mindthefuture@substack.com]]></itunes:email><itunes:name><![CDATA[Richard Ngo]]></itunes:name></itunes:owner><itunes:author><![CDATA[Richard Ngo]]></itunes:author><googleplay:owner><![CDATA[mindthefuture@substack.com]]></googleplay:owner><googleplay:email><![CDATA[mindthefuture@substack.com]]></googleplay:email><googleplay:author><![CDATA[Richard Ngo]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Economic efficiency often undermines sociopolitical autonomy]]></title><description><![CDATA[Against being econ-brained]]></description><link>https://www.mindthefuture.info/p/economic-efficiency-often-undermines</link><guid isPermaLink="false">https://www.mindthefuture.info/p/economic-efficiency-often-undermines</guid><dc:creator><![CDATA[Richard Ngo]]></dc:creator><pubDate>Tue, 10 Mar 2026 19:25:52 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/45933d2f-c2fc-42a4-b0bb-803b78fb55e3_1600x1069.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Many people in my intellectual circles use economic abstractions as one of their main tools for reasoning about the world. However, this often leads them to overlook how interventions which promote economic efficiency undermine people&#8217;s ability to maintain sociopolitical autonomy. By &#8220;autonomy&#8221; I roughly mean a lack of reliance on others&#8212;which we might operationalize as the ability to survive and pursue your plans even when others behave adversarially towards you. By &#8220;sociopolitical&#8221; I mean that I&#8217;m thinking not just about individuals, but also groups formed by those individuals: families, communities, nations, cultures, etc.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a></p><p>The short-term benefits of economic efficiency tend to be legible and quantifiable. However, economic frameworks struggle to capture the longer-term benefits of sociopolitical autonomy, for a few reasons. Firstly, it&#8217;s hard for economic frameworks to describe the relationship between individual interests and the interests of larger-scale entities. Concepts like national identity, national sovereignty or social trust are very hard to cash out in economic terms&#8212;yet they&#8217;re strongly predictive of a country&#8217;s future prosperity. (In technical terms, this seems related to the fact that utility functions are outcome-oriented rather than process-oriented&#8212;i.e. they only depend on interactions between players insofar as those interactions affect the game&#8217;s outcome).</p><p>Secondly, economic frameworks typically assume that people act in their rational interests at each point in time. They therefore rule out adversarial dynamics like credible threats (and following through on commitments more generally). Yet both offensive and defensive commitments are crucial aspects of how groups make decisions (as decision theories like FDT and UDT attempt to capture). For example:</p><ul><li><p>The legal system&#8217;s commitment to punishing criminals (even when the punishment costs society much more than the crime did) is the foundation on which economic property rights are maintained.</p></li><li><p>A nation&#8217;s commitment to regaining territory lost in wars (even when it can&#8217;t be justified by cost-benefit analyses, like Britain&#8217;s defense of the Falklands) deters enemies from trying to seize that territory in the first place.</p></li></ul><p>A more general principle here is that, while economists tend to think about what&#8217;s rational on the margin, political power depends on what would happen in <em>worst-case </em>scenarios. Marginal thinking is often more useful in the short term, but in the long term <a href="https://www.lesswrong.com/posts/fPvssZk3AoDzXwfwJ/universal-basic-income-and-poverty">control over the worst-case outcomes</a> provides leverage (for you or your adversaries) to shape the whole landscape of marginal effects. For example, if a tyrannical ruler sometimes executes people who seem disloyal, then his subjects might respond by <a href="https://www.mindthefuture.info/p/power-lies-trembling">proactively punishing dissidents</a> to prove their own loyalty. Hence relatively infrequent executions can be amplified into a <a href="https://www.mindthefuture.info/p/elite-coordination-via-the-consensus">society-wide control apparatus</a> that shapes everyone&#8217;s marginal incentives. (On a technical level, this is related to how changes in <a href="https://en.wikipedia.org/wiki/Cooperative_bargaining#Disagreement_point">disagreement points</a> can have big effects on the solutions of bargaining games&#8212;though mainstream bargaining theory hasn&#8217;t accounted for <a href="https://www.lesswrong.com/posts/vJ7ggyjuP4u2yHNcP/threat-resistant-bargaining-megapost-introducing-the-rose">how this incentivizes threats</a>.)</p><p>Thirdly, economics assumes commensurability (e.g. that goods and services can be priced in terms of money). But the mechanisms and institutions which maintain sociopolitical autonomy require a level of reliability which is undermined by commensurability. For example:</p><ul><li><p>Individuals whose integrity is for sale at the right price can&#8217;t be trusted as leaders.</p></li><li><p>Legal systems which punish speech based on how much harm they think it does are easily weaponized. (This is more of a utilitarian failing than an economic failing, but utilitarianism also relies heavily on commensurability.)</p></li><li><p>Countries which concede some territory to their neighbors undermine their ability to credibly commit to defending the rest of their territory.</p></li></ul><p>(James C. Scott's work on how <a href="https://en.wikipedia.org/wiki/Seeing_Like_a_State">states use commensurability and standardization</a> to <a href="https://www.amazon.com/Art-Not-Being-Governed-Anarchist-ebook/dp/B01N75OC23">scale their power</a> is very relevant to this.)</p><p>These particular examples are sufficiently obvious that few people defend treating them as commensurable. However, in the rest of this post I&#8217;ll discuss five cases where I think many people are applying economic frameworks too broadly, and thereby undermining the sociopolitical foundations that economic analysis implicitly relies on. I&#8217;ll refer to this as being &#8220;econ-brained&#8221;. Econ-brain is related to neoliberalism, libertarianism, and effective altruism, though it&#8217;s not synonymous with any of them.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> It&#8217;s often critiqued by both the anti-market left and the nationalist right; I&#8217;m more sympathetic to the latter critiques, but will mostly focus on examples that aren&#8217;t polarized along standard partisan lines.</p><p>I&#8217;d eventually like to develop a formal definition of &#8220;sociopolitical rationality&#8221; that can precisely describe the failures of &#8220;economic rationality&#8221;. In the meantime, I hope that these examples convey the core intuitions. Of course, it&#8217;s hard to summarize any one topic, let alone five of them. So please take each of these five sociopolitical perspectives in the spirit of &#8220;ideas you might be missing, that could add up to something big&#8221; rather than &#8220;a individually knock-down case against econ-brained thinking&#8221;. To facilitate that, I recommend that you take a few moments to note down your opinion of the headline topic before reading the corresponding section.</p><h3>Five case studies</h3><ol><li><p>Prediction markets</p></li></ol><p>[Pause here if you want to consider your stance towards them before reading.]</p><p></p><p></p><p><a href="https://en.wikipedia.org/wiki/Prediction_market">Prediction markets</a> have highly desirable properties from an economic perspective. They are incentive-compatible ways of surfacing hidden information. They&#8217;re extremely hard to manipulate, at least in theory&#8212;if anyone suspects manipulation is happening, they can profit by betting in the opposite direction. And so they&#8217;ve been supported by various economists (<a href="https://www.overcomingbias.com/p/prediction-markets-now">most notably Hanson</a>) as well as the rationalist and effective altruist communities.</p><p>Why oppose prediction markets? One standard response is that prediction markets could be used as <a href="https://en.wikipedia.org/wiki/Assassination_market">assassination markets</a>. That is, any market which would be affected by the death of a major figure could allow someone to profit off assassinating them. However, this feels like an edge case&#8212;assassinations are rare, and financially-motivated assassinations even rarer.</p><p>A more central objection, based on the same principle, is that it&#8217;s easy for prediction markets to become <em>corruption</em> markets. One type of corruption is simply profiting by betting on private information, which we&#8217;ve already started to see with the rise of polymarket (see <a href="https://www.mondaq.com/unitedstates/commoditiesderivativesstock-exchanges/1737394/potential-insider-trading-concerns-in-prediction-markets-the-polymarket-maduro-bet-case?utm_source=chatgpt.com">here</a>, <a href="https://gizmodo.com/polymarket-user-accused-of-1-million-insider-trade-on-google-search-markets-2000696258">here</a>, <a href="https://futurism.com/future-society/polymarket-half-time-show">here</a>). We can debate the extent to which institutions <em>should</em> be able to keep information private&#8212;but by default they won&#8217;t have a choice. Unlike stock markets, prediction markets can be set up in large numbers on arbitrary questions, with anonymized crypto-based payouts, potentially making insider trading much harder to monitor.</p><p>Moreover, as prediction markets become better-capitalized I expect we&#8217;ll start to see cases where decisions are <em>made</em> in order to influence prediction markets. We&#8217;ve only seen <a href="https://www.businessinsider.com/coinbase-ceo-earnings-call-words-prediction-markets-bets-2025-12?utm_source=chatgpt.com">unimportant</a> <a href="https://casinobeats.com/2025/12/11/donald-trump-speech-manipulated-kalshi-trader-ensure-winning-bets/?utm_source=chatgpt.com">examples</a> of this so far, but as prediction markets grow the incentives to do so will increase. Furthermore, prediction markets could be used as a mechanism to anonymously bribe decision-makers. <a href="https://x.com/RichardMCNgo/status/1885786853921763608">As a toy example</a>, people who wanted to incentivize policy X could create and subsidize a market like &#8220;conditional on policy X being announced, which day will it happen?&#8221; The decision-maker could then profit by announcing policy X on a day of their choosing, and betting accordingly. Unlike regular bribes, this doesn&#8217;t require any direct interaction or agreement which could serve as smoking-gun evidence of corruption (though it does leave a public record of the anonymized transactions).</p><p>In short, prediction markets harm institutions&#8217; ability to maintain autonomy in the face of external pressures, by commodifying the process of turning institutional influence into money (and vice versa). Nor is this a coincidence. Instead, prediction markets create &#8220;efficiency&#8221; precisely by incentivizing individuals to be more engaged with markets, at the expense of legal and moral obligations to the institutions they work within.</p><ol start="2"><li><p>Land value taxes</p></li></ol><p>[Pause here if you want to consider your stance towards them before reading.]</p><p></p><p></p><p>Land value taxes are well-known to be highly economically efficient. In general, taxes disincentivize the production of whatever is being taxed. However, in most places it&#8217;s not possible to produce more land. And the vast majority of the value of land is driven by factors that the land owners themselves don&#8217;t control (such as proximity to a city). So land taxes are considered far less distortionary than taxes on income or consumption&#8212;hence the <a href="https://archive.is/bkS21">recurring</a> <a href="https://www.astralcodexten.com/p/your-book-review-progress-and-poverty">popularity</a> of <a href="https://en.wikipedia.org/wiki/Georgism">Georgism</a> amongst <a href="https://www.noahpinion.blog/p/how-to-sell-georgism-to-the-middle">political commentators</a>, who sometimes suggest that they should replace income taxes altogether.</p><p>The term &#8220;non-distortionary&#8221; can be misleading, though. If land value taxes replaced income taxes, they&#8217;d significantly affect who&#8217;s able to afford which property&#8212;just in ways that economists think increase efficiency. Consider someone who&#8217;d like to use their property in a way that isn&#8217;t very financially rewarding&#8212;for example, as a community hub. Once they own their property, they might need relatively little income to be viable (and therefore pay little in income taxes). However, if a land value tax is implemented, they&#8217;d need to pay the same amount of tax as a commercial business using that same property would, which might force them to move or shut down.</p><p>Defenders of land value taxes argue that this is efficient from an economic perspective: it reallocates property from economically unproductive to economically productive uses. Another way of putting this, however, is that land value taxes would make it harder for land-owners to remain autonomous. Instead of freely choosing how to use their own properties, they&#8217;d face strong pressures to use it in ways that the market finds valuable. To contrast this with income taxes, consider some group that doesn&#8217;t use money to organize itself internally. If you draw a boundary around that group, then income tax only takes some percentage of money that flows in across that boundary, and so the group can reduce their tax burden by becoming more self-sufficient. Conversely, a land value tax creates a net outflow of money from the group that isn&#8217;t determined by how much money is flowing in, forcing them to maintain a significant income stream to survive.</p><p>There&#8217;s a rights-based case against infringing on such groups&#8217; autonomy, which I&#8217;ll discuss later on. But even in consequentialist terms, society is disproportionately shaped by people and groups that are able to insulate themselves from commercial pressures. This occurs at many different scales: individual homeowners, churches or universities, communities (or communes), all the way up to ethnic groups like the Amish. Such groups are able to experiment with novel ideologies and lifestyles in significant part because they&#8217;re less accountable to market forces than corporations. The lessons from those experiments can spread very widely (e.g. the Amish are a common reference point in discussions of fertility declines). By comparison, consider how bad almost all corporations are at cultural leadership&#8212;because genuinely novel thinking is often economically illegible, and therefore very difficult to do under financial pressure.</p><p>I&#8217;ve been discussing land value taxes in a very abstract sense. In reality, there are many complicating factors which might mitigate the effects I described, some of which I discuss in a footnote.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> However, the most important practical consideration may simply be the difficulty of guaranteeing that land value taxes would actually replace other taxes, rather than just adding to them. Over the last century, we&#8217;ve seen massive expansions of state power in many domains&#8212;amount of regulation and amount of taxation being two crucial ones. For the population as a whole to retain its autonomy, it seems very important to set and defend <a href="https://www.lesswrong.com/posts/Kbm6QnJv9dgWsPHQP/schelling-fences-on-slippery-slopes">Schelling fences</a> at which we can coordinate to resist further encroachments&#8212;with strong property rights being one of the best such fences. Adding new taxes&#8212;and in particular recurring taxes on things which you already own&#8212;would make &#8220;ownership&#8221; a less meaningful concept. It would therefore become more difficult to rally around property rights to fight against expansions of state power (especially ones nominally justified by appeals to economic efficiency).<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a></p><p>I suspect that many ordinary people understand the dynamics I&#8217;ve explained on an intuitive level&#8212;hence why property taxes and poll taxes are so unpopular. However, these intuitions remain illegible from an econ-brained perspective, in part because the sociopolitical principles behind them have never been adequately formalized.</p><ol start="3"><li><p>Higher education</p></li></ol><p>Higher education is puzzling from an econ-brained perspective, because university students don&#8217;t seem to be learning very many job-relevant skills, yet are still paid a significant wage premium over non-graduates. The best economic explanation for why this happens is <a href="https://www.amazon.com/Case-against-Education-System-Waste/dp/0691174652/ref=as_sl_pc_qf_sp_asin_til">Caplan&#8217;s signaling account</a>; he claims that going to university is a signal of intelligence, conscientiousness and conformity.</p><p>However, <a href="https://www.mindthefuture.info/p/contra-caplan-on-higher-education">as I argue in this post</a>, the signaling account doesn&#8217;t work, because there are much cheaper ways to signal all of these traits. Instead, I suspect that college is best understood as forming an elite class with its own norms and values (as described by <a href="https://infed.org/dir/welcome/pierre-bourdieu-habitus-capital-and-field-exploring-reproduction-in-the-practice-of-education/#:~:text=and%20scholars%E2%80%99%20(Calhoun%202002).-,habitus,-Bourdieu%E2%80%99s%20distinction%20by">Bourdieu</a>, <a href="https://web.archive.org/web/20240324083144/http://archive.harpers.org/1994/11/pdf/HarpersMagazine-1994-11-0001862.pdf">Lasch</a>, and others).</p><p>I&#8217;ll note that the formation of such an elite class is actually harmful for most countries. So in this case I&#8217;d actually prefer a more economically efficient outcome (like a massive reduction in university prestige and attendance). However, it&#8217;s still a good example of the difference between economic and sociopolitical reasoning.</p><ol start="4"><li><p>Free trade</p></li></ol><p>Mainstream economic thinking is strongly in favor of free trade, for the sake of its economic benefits. However, mainstream economic thinking has also led to a huge amount of American manufacturing capacity being offshored to its geopolitical rivals, to the point where even most US military supply chains are dependent on Chinese production. So economic efficiency here comes at the longer-term cost of national autonomy&#8212;both in terms of robustness to disruptions (e.g. from covid) and robustness to conflict with China. While both points have been made in various places over the years, they don&#8217;t seem to have been adequately incorporated into economic consensus&#8212;e.g. I saw few mainstream economists take them into account when evaluating Trump&#8217;s tariffs.</p><p>Now, there&#8217;s an argument that intertwining the US and Chinese supply chains makes the world safer, by making war between the two superpowers more costly. In other words, perhaps decreasing American and Chinese autonomy is a good thing. However, even though both countries are <em>economically </em>dependent on each other, the US is disproportionately <em>industrially</em> and <em>militarily </em>dependent on China. So from a &#8220;hard power&#8221; perspective, the US gave up autonomy while China retained (and in fact increased) its autonomy.</p><p>Another big tension between economic and sociopolitical views of free trade is that the sociopolitical view accounts for shifts in the internal balance of power within the US. The manufacturing industry is far more widely-distributed across US states than the finance or software industries. So its decline has led to increased concentration of power amongst coastal elites. Again, I&#8217;m not claiming that this should be a decisive argument against free trade; however, it&#8217;s the kind of consideration that doesn&#8217;t arise naturally from an econ-brained perspective. Whereas from a sociopolitical perspective, maintaining autonomous subagents is a crucial component of a nation&#8217;s continued health (which is a major reason to defend states&#8217; rights).</p><ol start="5"><li><p>The future of AGI</p></li></ol><p>Econ-brained thinking has shaped the AGI safety community&#8217;s (and thereby the wider world&#8217;s) perspective on the future of AGI. Influential figures like <a href="https://intelligence.org/files/AIFoomDebate.pdf">Hanson</a>, <a href="https://paulfchristiano.com/ai/">Christiano</a>, and <a href="https://80000hours.org/podcast/episodes/carl-shulman-common-sense-case-existential-risks/">Shulman</a> often apply economic abstractions to make forecasts. This contrasts with thinkers like <a href="https://www.lesswrong.com/posts/uMQ3cqWDPHhjtiesc/agi-ruin-a-list-of-lethalities">Yudkowsky</a> or <a href="https://podcast.clearerthinking.org/episode/028/michael-vassar-preference-falsification-and-postmodernism/">Vassar</a> who are more dismissive of the relevance of economics for thinking about AGI (though I wouldn&#8217;t summarize them as &#8220;sociopolitics-brained&#8221;, but rather merely &#8220;less econ-brained&#8221;).</p><p>In this section I&#8217;ll prioritize breadth over depth. I&#8217;ll give half a dozen examples of econ-brained ideas about how to orient to AGI, and mostly leave the task of generating sociopolitical critiques of them as exercises for the reader:</p><ul><li><p>The idea of paying AIs to cooperate with us, as discussed <a href="https://www.lesswrong.com/posts/psqkwsKrKHCfkhrQx/making-deals-with-early-schemers">here</a>, <a href="https://docs.google.com/document/d/1SUgGftOMKO4GVSBzOa3wMDcAhM4UrOGAjYVPnALxKY4/edit?tab=t.0#heading=h.h4ccv9xrmzam">here</a>, and <a href="https://www.dwarkesh.com/p/give-ais-a-stake-in-the-future">here</a>.</p></li><li><p>The idea of owning galaxies, as discussed <a href="https://www.lesswrong.com/posts/SYyBB23G3yF2v59i8/on-owning-galaxies">here</a>.</p></li><li><p>The idea that harms from speeding up AI capabilities progress can be largely offset by benefits from preventing capabilities overhangs (as defended <a href="https://www.alignmentforum.org/posts/fRSj2W4Fjje8rQWm9/thoughts-on-sharing-information-about-language-model#Accelerating_LM_agents_seems_neutral__or_maybe_positive_">here</a>, <a href="https://www.alignmentforum.org/posts/vwu4kegAEZTBtpT6p/thoughts-on-the-impact-of-rlhf-research">here</a>, and <a href="https://www.alignmentforum.org/posts/uFNgRumrDTpBfQGrs/let-s-think-about-slowing-down-ai?commentId=NdA7t6RaDGy3MWANp">here</a> and critiqued <a href="https://www.alignmentforum.org/posts/4Pi3WhFb4jPphBzme/don-t-accelerate-problems-you-re-trying-to-solve">here</a>). In addition to Paul&#8217;s position, it&#8217;s illustrative to contrast two other people&#8217;s stances towards this idea:</p><ul><li><p>Sam Altman used the idea of compute overhangs as a justification to accelerate progress towards AGI, until it became more useful to start pushing for more GPU production instead.</p></li><li><p>Meanwhile, an example of the polar opposite strategy was <a href="https://www.lesswrong.com/posts/LtT24cCAazQp4NYc5/open-global-investment-as-a-governance-model-for-agi?commentId=9i6hEbipSWwyzsFAA">Wei Dai declining to invest in Anthropic for moral reasons</a>, thereby losing out on what would by now have been over 400x returns. I respect Wei&#8217;s approach very much (despite not knowing whether he should have been more econ-brained in this case).</p></li></ul></li><li><p>The idea that AGI labs are efficient at racing towards AGI, and therefore building new capabilities evals isn&#8217;t very helpful for them (<a href="https://x.com/richardmcngo/status/1869089459662438573?s=46">as I critique here</a>).</p></li><li><p>The idea of tracking progress towards AGI in terms of <a href="https://sideways-view.com/2018/02/24/takeoff-speeds/">GDP growth</a> or <a href="https://www.basilhalperin.com/papers/agi_emh.pdf">real interest rates</a>.</p></li><li><p>The idea that AGI will come in the form of separate tools or services rather than unified agents, as defended by <a href="https://www.overcomingbias.com/p/how-lumpy-ai-serviceshtml">Hanson</a> and <a href="https://owainevans.github.io/pdfs/Reframing_Superintelligence_FHI-TR-2019.pdf">Drexler</a>.</p><ul><li><p>Note the parallel between this perspective and the idea that businesses are mainly held together by <a href="https://en.wikipedia.org/wiki/The_Nature_of_the_Firm">transaction costs</a>, the latter of which led Krier to the very econ-brained idea that society could be revolutionized by AI-enabled <a href="https://blog.cosmos-institute.org/p/coasean-bargaining-at-scale">Coasean bargaining at scale</a>.</p></li></ul></li></ul><p>Some of these ideas have been critiqued by <a href="https://www.lesswrong.com/posts/xJWBofhLQjf3KmRgg/four-ways-learning-econ-makes-people-dumber-re-future-ai">Byrnes</a>, <a href="https://www.lesswrong.com/posts/F8sfrbPjCQj4KwJqn/the-sun-is-big-but-superintelligences-will-not-spare-earth-a">Yudkowsky</a>, and others. In his posts on the <a href="https://www.alignmentforum.org/posts/ivpKSjM4D6FbqF4pZ/cortes-pizarro-and-afonso-as-precedents-for-takeover">Spanish conquistadors</a> as precedents for AGI takeover, Kokotajlo is clearly also looking at the issue through a sociopolitical lens. However, it&#8217;s worth noting that econ-brained thinkers have scored some big wins over the last decade&#8212;e.g. <a href="https://www.lesswrong.com/posts/HBxe6wdjxK239zajf/what-failure-looks-like">predicting the diffusion of AI across society</a>, and <a href="https://www.lesswrong.com/posts/sCCdCLPN9E3YvdZhj/shulman-and-yudkowsky-on-ai-progress">the unprecedented amount of investment</a> that would be <a href="https://www.dwarkesh.com/p/carl-shulman">funneled towards the AI industry</a>. And zooming out even further, compute-based forecasts of AGI like Kurzweill&#8217;s and Legg&#8217;s have been surprisingly prescient. Such forecasts aren&#8217;t quite central examples of being econ-brained, but there&#8217;s definitely something econ-brained (and <a href="https://www.lesswrong.com/posts/LvKDMWQ3yLG9R3gHw/empiricism-as-anti-epistemology">something</a> <a href="https://www.lesswrong.com/posts/ax695frGJEzGxFBK4/biology-inspired-agi-timelines-the-trick-that-never-works">anti-Yudkowskian</a>) about believing so much in <a href="https://slatestarcodex.com/2019/03/13/does-reality-drive-straight-lines-on-graphs-or-do-straight-lines-on-graphs-drive-reality/">straight lines on graphs</a>.</p><p>Why is this? The most straightforward possibility is simply that the concept of econ-brain is too lossy an abstraction to reliably evaluate thinkers with. Ideally we&#8217;d try to diagnose what led to each of these successes and failures in granular detail. But as a rough heuristic, is being more econ-brained actually a good way to improve your forecasts? Some possible responses:</p><ul><li><p>Maybe the forecasting successes listed above required the right balance between econ-brained and other kinds of thinking. If you&#8217;re too econ-brained, you reject the concept of AGI altogether; if you&#8217;re not econ-brained enough, you&#8217;re surprised by how continuously progress has advanced over the last decade. Paul and Carl and Ray and Shane might be in the sweet spot re these particular topics. But this isn&#8217;t a very satisfying response, because these people are extremely econ-brained by almost everyone&#8217;s standards.</p></li><li><p>Maybe economic factors are more important in the short term (during which institutions and power structures are roughly stable), whereas sociopolitical dynamics play out over longer time horizons (and will especially kick in once AIs become capable of wielding political power). This makes econ-brained people more like foxes, and sociopolitics-brained people more like hedgehogs. The former tend to make more predictions that are accurate; however, the latter have a better chance of predicting the most important large-scale shifts.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a></p></li><li><p>Maybe there&#8217;s a tradeoff between predictive accuracy and the ability to get things done. In general, outside-view bets like &#8220;nothing ever happens&#8221; tend to outperform your inside view on most topics. Similarly, believing in efficient markets is a good strategy for most investors. But it&#8217;s hard to change the world by believing in efficient markets. Relatedly, in the final section of <a href="https://www.mindthefuture.info/p/power-lies-trembling">this post</a> I discuss how &#8220;leaps of faith&#8221; can be extremely valuable for sparking coordinated action.</p></li></ul><h3>Conclusion</h3><p>These five case studies are far from exhaustive. There are plenty of examples that I omitted for brevity (e.g. surge pricing, YIMBYism, earning to give, etc). And there are other cases that I suspect are important examples of this phenomenon, but don&#8217;t yet understand well enough to discuss in detail. For example, cryptocurrency is a nominally-economic domain that seems more driven by sociopolitical dynamics than economic fundamentals. And <a href="https://benjaminrosshoffman.com/#:~:text=Macroeconomics%2C%20Social%20Theory%2C%20and%20the%20Psychic%20Consequences%20of%20Pyramid%20Schemes">Ben Hoffman&#8217;s writing on macroeconomics</a> (in particular <a href="https://benjaminrosshoffman.com/the-debtors-revolt/">his post on the debtor&#8217;s revolt</a>) provides a perspective from which 20th-century economic history was driven by sociopolitical conflicts.</p><p>In other cases, econ-brained thinking is harnessed to defend a position, but isn&#8217;t the main force behind that position. For example, the cultural wars that are currently raging over immigration definitely feature clashes between economic and sociopolitical considerations. However, I suspect that the pro-immigration side is not fundamentally motivated by immigration&#8217;s purported economic benefits, which are better understood as fig leaves on <a href="https://www.lesswrong.com/posts/FuGfR3jL3sw6r8kB4/richard-ngo-s-shortform?commentId=9g5QK3Km25fMBHgcK">a deeper-rooted globalist ideology</a>. Similarly, even though much of the explicit debate about Brexit pitted economic against cultural considerations, the sheer vitriol that elites leveled against Brexiteers suggests that they were primarily motivated by sociopolitical considerations of their own.</p><p>Ultimately, the greatest prize would be a precise technical theory that fills in what economics is missing. Scott Garrabrant&#8217;s <a href="https://www.lesswrong.com/s/4hmf7rdfuXDJkxhfg">distinction between arithmetic and geometric rationality</a> seems like one important step towards this. As he points out, arithmetic rationality (which I suspect is closely related to economic thinking) is oriented towards maximizing efficiency. But if taken too far, <a href="https://www.lesswrong.com/s/4hmf7rdfuXDJkxhfg/p/rc5ZKGjXTHs7wPjop">it creates internally dysfunctional agents</a>, and so it needs to be governed at the meta-level by geometric rationality (which I suspect is closely related to sociopolitical thinking). A big question is then how to <a href="https://www.lesswrong.com/posts/8oMF8Lv5jiGaQSFvo/boundaries-part-1-a-key-missing-concept-from-utility-theory">draw boundaries</a> between the two categories in a principled way.</p><p>That&#8217;s all beyond the scope of this post, though. For now, I merely hope that I&#8217;ve conveyed the core idea that there&#8217;s <em>something </em>interesting about autonomy and related sociopolitical concepts which is systematically neglected (and undermined) by econ-brained thinking.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>Corporations are another example of such a group&#8212;though a less central one, because they lack many of the traits that hold together most sociopolitical groups (such as membership/citizenship that&#8217;s difficult to take away from people).</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>I use &#8220;econ-brain&#8221; rather than &#8220;neoliberalism&#8221; to avoid getting caught up in the political connotations, since the neoliberal world order does many things that econ-brained people disagree with. Also, econ-brain applies to some issues that neoliberalism doesn&#8217;t have much of a stance on, like prediction markets or AGI. Meanwhile, when I talk about libertarians as econ-brained, I&#8217;m primarily referring to the modern economic-focused libertarianism espoused by thinkers like <a href="https://www.econlib.org/archives/2016/03/libertarianism_4.html">Brian Caplan</a> and <a href="https://slatestarcodex.com/2013/12/08/a-something-sort-of-like-left-libertarianism-ist-manifesto/">Scott Alexander</a>. Conversely, historical libertarian(ish) figures like Hayek and Rand thought much more about sociopolitical concepts such as serfdom vs freedom.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>Three such considerations:</p><ul><li><p>Many of the organizations I mentioned above currently have charitable tax exemptions, and so wouldn&#8217;t be adversely affected by land value taxes. However, I think of this as only a band-aid solution to the core problem. If standards for charities are too loose, land value tax is no longer effective (because everyone would find some way to own property via a charity). If standards are too strict, then charitable status provides much less autonomy (because charities would still have to stay on the state&#8217;s good side to retain their status). Overall, the more a tax relies on getting the exceptions right, the less sound we should consider its principles to be.</p></li><li><p>Property taxes are similar to land value taxes in many ways, and are far more common. So I expect that many of the problems that a full-blown land value tax would cause already exist to a lesser extent in jurisdictions with high property taxes. It&#8217;d be useful to get empirical data on this. For now, I&#8217;m focusing on land value taxes as a cleaner case study of econ-brained thinking.</p></li><li><p>My thought experiment of a community avoiding income taxes by becoming more self-sufficient is in tension with the fact that, in the US, income taxes technically also apply to non-monetary transactions. However, I think that the impossibility of actually enforcing this itself helps demonstrate the limitations of economic thinking. Even in principle, how could you put prices on non-monetary exchanges that occur within a family, or a community, or between university students? If you imagine a government actually trying to do this (and punishing people who don&#8217;t pay) that would be the clearest example yet of how economic thinking undermines sociopolitical autonomy.</p></li></ul></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>A related practical issue which I haven&#8217;t seen a good Georgist response to: the case for land value taxes over property taxes relies on incentivizing construction. But if construction is severely restricted by permitting processes (as it is in most Western cities) then a land value tax would unfairly penalize landowners who didn&#8217;t already have buildings on their land, without actually leading to much additional housing. To be fair, I expect this is part of why YIMBYism is much more popular today than Georgism.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>Relatedly: when Paul Christiano and Eliezer Yudkowsky tried to operationalize their disagreement as a bet, Paul claimed that he&#8217;d be willing to bet on most things, whereas Eliezer was much more selective. But when they did settle on a single bet, Eliezer ended up winning (though note that the bet they chose was one where Eliezer was closer to the consensus side, suggesting that there might have been adverse selection).</p></div></div>]]></content:encoded></item><item><title><![CDATA[Contra Caplan on higher education]]></title><description><![CDATA[Though it's still a waste of time and money]]></description><link>https://www.mindthefuture.info/p/contra-caplan-on-higher-education</link><guid isPermaLink="false">https://www.mindthefuture.info/p/contra-caplan-on-higher-education</guid><dc:creator><![CDATA[Richard Ngo]]></dc:creator><pubDate>Mon, 16 Feb 2026 20:53:59 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/6e55fc18-3592-4dfd-85c6-33e6f0048129_944x860.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h3>Three theories of higher education</h3><p>Getting an undergraduate degree is very costly. In America, the direct financial cost of attending a private university is typically in the hundreds of thousands of dollars. Even when tuition is cheap (or covered by scholarships), forgoing three to four years of salary and career progression is a large opportunity cost. There are a variety of reasons why students are willing to pay these costs, but the key one is that desirable employers highly value college degrees.</p><p>Why? The standard economic answer is that college classes teach skills which are relevant for doing jobs well: the &#8220;human capital&#8221; theory. But even a cursory comparison of college curricula to the actual jobs college graduates are hired for makes this idea seem suspicious. And private tutoring is <a href="https://en.wikipedia.org/wiki/Bloom%27s_2_sigma_problem">so vastly more effective than classes</a> that it&#8217;s very inefficient to learn primarily via the latter (especially now that many university courses are more expensive than even 1:1 tutoring, let alone AI tutoring).</p><p>Another answer is that attending college can be valuable for the sake of signaling desirable traits to employers. An early version of this model <a href="https://www.jstor.org/stable/1882010">comes from Spence</a>; more recently, Bryan Caplan has argued that most of the wage premium from going to college comes from signaling. In this post I&#8217;ll be engaging with Caplan&#8217;s version of the signaling hypothesis, as laid out in <a href="https://www.amazon.com.au/Case-against-Education-System-Waste/dp/0691174652">his book The Case Against Education</a>.</p><p>Your university degree signals many things about your underlying characteristics, but Caplan claims that there are three traits employers prioritize above all others: &#8220;the trinity of intelligence, conscientiousness, and conformity&#8221;. This hypothesis purports to explain a number of important gaps in the human capital theory&#8212;e.g. why college students so quickly forget so much of the material covered in their courses after passing exams, why the rise of free online courses hasn&#8217;t changed the college landscape very much, and why <a href="https://en.wikipedia.org/wiki/Sheepskin_effect">finishing 90% of a degree is far less than 90% as valuable</a> as completing the whole thing.</p><p>However, I think Caplan&#8217;s signaling theory is also wrong. In particular, his concept of conformity can&#8217;t be understood in standard economic terms. Instead, I&#8217;ll argue that we need a sociological explanation centered around group membership and group norms&#8212;which I&#8217;ll try to flesh out in a follow-up post. First, though, let&#8217;s engage with Caplan&#8217;s position, starting with the other two aspects of his trinity.</p><h3>College attendance isn&#8217;t explained by intelligence or conscientiousness signaling</h3><p>A key problem with Caplan&#8217;s trinity is that most of it is easily replaceable. Getting good grades at college <em>does</em> signal intelligence and conscientiousness, but these could be signaled far more easily and cheaply. It&#8217;s very easy to signal intelligence via test scores: IQ is surprisingly predictive of many other desirable cognitive traits. This need not require literal IQ tests&#8212;standardized tests like the SAT or GRE are highly correlated with intelligence. In other cases, companies use IQ-like tests (e.g. tech companies&#8217; coding interviews). These are also significantly harder to cheat on than college courses.</p><p>Caplan acknowledges that college grades are far from the best way to signal intelligence; what he doesn&#8217;t discuss is that they&#8217;re even further from the best way to signal conscientiousness. If you asked people why they don&#8217;t just learn college material independently without paying for college, I expect that a common response would simply be &#8220;oh, I don&#8217;t have the discipline for that&#8221;. College provides external frameworks, timetables, local incentives, and social pressure for people who aren&#8217;t conscientious enough to learn without that.</p><p>So although doing well at college signals more conscientiousness than lazing about, an even better signal of conscientiousness would be acquiring all the same knowledge without attending college at all! In fact, people who are capable of learning college-level material independently should be going out of their way to <em>avoid</em> college lest they be confused for those who can only do it within a motivating social structure. Again, this hinges on the existence of high-quality testing services&#8212;but if conscientiousness signaling drove a significant proportion of the value of a college diploma, then providing such testing would be very profitable.</p><p>I&#8217;ll digress briefly to clarify a point that sometimes confuses people (including my past self). It&#8217;s common to talk about &#8220;<a href="https://en.wikipedia.org/wiki/Signalling_theory#Handicap_principle">costly</a><em><a href="https://en.wikipedia.org/wiki/Signalling_theory#Handicap_principle"> </a></em><a href="https://en.wikipedia.org/wiki/Signalling_theory#Handicap_principle">signaling</a>&#8221;, which involves incurring costs that would be prohibitive for people who don&#8217;t possess desirable traits. But costly signaling is just one type of &#8220;credible signaling&#8221;, aka signaling that is difficult to fake. Other types of credible signaling need not be expensive&#8212;IQ tests are an example of a very cheap but very credible signal.</p><p>By basic economic logic, people should prefer to do credible signaling in cheaper rather than more expensive ways. So any explanation of behavior in terms of costly signalling needs to explain why the system doesn&#8217;t gradually shift towards using cheaper credible signals. In Caplan&#8217;s account, that&#8217;s where the &#8220;conformity&#8221; part plays a big role.</p><p>I&#8217;ll explore in more detail what his account of conformity signaling is in the next section, and why I don&#8217;t think it succeeds. But I first want to note that the arguments above should already update our view of Caplan&#8217;s theory. If I&#8217;m right about how replaceable the functions of intelligence and conscientiousness signaling are in justifying college degrees, then even calling it a &#8220;signaling theory&#8221; of education is misleading. Instead what Caplan is defending is more accurately summarized as the &#8220;conformity signaling theory&#8221; of education, because that&#8217;s the part that&#8217;s justifying almost all of the cost of college <em>compared with other possible signaling strategies</em>. Analogously: if product A costs $10 and lets you do tasks X and Y, and product B costs $100 and lets you do tasks X, Y, and Z, then a good explanation for why &#8220;rational&#8221; people keep buying B needs to focus on the value of doing Z.</p><p>Of course, it&#8217;s hard enough to write a book about how college is for signaling; describing college attendance as being driven by conformity would be even more controversial. I don&#8217;t want to criticize Caplan too harshly for this omission, since he&#8217;s been more honest than almost any other academic about the ways in which higher education is a waste of time and money. And I don&#8217;t think he&#8217;s being deliberately deceptive. But my guess is that he flinched away from summarizing his theory as the &#8220;conformity signaling theory&#8221; of education because it would have received even more pushback than his &#8220;signaling theory&#8221;. I wish he hadn&#8217;t, though, because trying to pin down what conformity signaling is, and why employers purportedly value it, makes the holes in this theory clear.</p><h3>College attendance isn&#8217;t explained by conformity signaling</h3><p>Conformity signaling is more complicated than intelligence or conscientiousness signaling. Caplan spends half a dozen pages explaining it in the first chapter of The Case Against Education. I&#8217;ll describe his position in my own words here, starting with a quick note on what he <em>doesn&#8217;t </em>mean. Firstly, he doesn&#8217;t mean that students are signaling a <em>general </em>tendency to conform:</p><blockquote><p>&#8220;Employers aren&#8217;t looking for workers who conform in some abstract sense&#8230; Hippies strive to look, talk, and act like fellow hippies. This doesn&#8217;t make unkempt hair and tie-dye shirts any less repugnant to employers. Employers are looking for people who conform to the folkways of today&#8217;s workplace&#8212;people who look, talk, and act like modern model workers.&#8221;</p></blockquote><p>For now I&#8217;ll call this trait &#8220;conformity to professional norms&#8221;. You might then think that employers want to hire college graduates because they&#8217;ve learned professional norms during their degrees. But this would be a <em>human capital</em> explanation, whereas Caplan is clear that he&#8217;s focusing on a signaling explanation<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> So we can reconstruct Caplan&#8217;s signaling theory as instead claiming:</p><ol><li><p>High school students vary in the underlying trait of how able and willing they are to conform to professional norms.</p></li><li><p>Going to (and succeeding at) university signals that you&#8217;re able and willing to conform to professional norms.</p></li></ol><p>So far, this is a standard signaling explanation. We can debate how strong the correlation is between successful university attendance and workplace professionalism&#8212;I can see arguments in either direction. However, even if the former is a very good signal of the latter, there&#8217;s a more pressing issue: attending university is very costly compared with references or work trials or almost any other method of signaling professionalism. So Caplan needs to be able to explain why far cheaper methods of credibly signaling conformity to professional norms don&#8217;t develop.</p><p>This where his theory of conformity signaling becomes disanalogous to intelligence or conscientiousness signaling, by adding a third claim:</p><ol start="3"><li><p>Conformity is the one thing that you <em>can&#8217;t </em>develop cheaper ways to signal, because doing new and unusual types of signaling <em>itself</em> demonstrates a lack of conformity.</p></li></ol><p>Because of this, Caplan argues that university degrees are now &#8220;locked in&#8221; as the key signal of conformity. Anyone who tries to signal conformity to professional norms in other ways is outing themselves as weird and nonconformist, making their new signal self-defeating.</p><p>It&#8217;s a clever move from Caplan, but ultimately I think it&#8217;s conceptually confused. The core issue is that, even if &#8220;professionalism&#8221; requires some amount of &#8220;conformity&#8221;, they&#8217;re still distinct concepts. There are plenty of ways that rational employers should want their employees to be nonconformist&#8212;e.g. spotting new market opportunities before others do. There are also plenty of ways in which college students don&#8217;t mind signaling nonconformity with the business world: their avant-garde politics, their idiosyncratic hair and clothes, and often their nontraditional majors. If students were really spending years of their lives and hundreds of thousands of dollars primarily to signal conformity, shouldn&#8217;t they be picking much lower-hanging fruit first?</p><p>Indeed, there&#8217;s something suspicious about Caplan&#8217;s use of the term &#8220;conformity&#8221; at all. Why not just say that employers are looking for professionalism, and students are trying to signal professionalism? Adding the word &#8220;conform&#8221; is a verbal trick which proves too much: by Caplan&#8217;s logic <em>any</em> example of a person signaling that they follow norm X could be redescribed as &#8220;signaling conformity to norm X&#8221;, and then used to explain why they&#8217;re &#8220;locked in&#8221; to irrational behavior.</p><p>Finally, <a href="https://www.overcomingbias.com/p/school-is-to-submithtml">as Hanson notes</a>: even if the idea of lock-in explains why a practice continues, it can&#8217;t explain why it <em>started</em>. In the past, only a small percentage of the population attended college, and it was perfectly normal to get a prestigious job without a college degree. What drove the rise of college in the first place? Whatever it was, that seems like it should be our default hypothesis for what&#8217;s driving the college wage premium today.</p><h3>Explaining college requires sociological theories</h3><p>To be clear, I do think there&#8217;s <em>something</em> important going on related to conformity. It just can&#8217;t be captured as part of a signaling framework&#8212;or any other economic framework&#8212;for at least two reasons.</p><p>Firstly, signaling is a framework under which rational agents pay costs to demonstrate pre-existing traits. But conforming is best understood as a process of <em>internalizing </em>deference to other people, i.e. making oneself less rational. Conformists can&#8217;t turn their conformity off when it might profit them&#8212;think of how many people decided not to invest in bitcoin, or scoffed at the possibility of rapid AI progress, because it sounded weird. They even internalize conformity on an emotional level&#8212;e.g. they often get angry at nonconformists (something which I expect Caplan has experienced many times). This is hard to model in economic terms.</p><p>A second problem with the idea of students signaling to employers is that employers are <em>also</em> better modeled as conforming rather than making rational choices. For example, Caplan claims that students don&#8217;t signal intelligence using standardized test scores because &#8220;putting high scores on your resume suggests you&#8217;re smart but socially inept. You&#8217;re doing something that&#8217;s &#8216;simply not done.&#8217;&#8221; But firms could easily request standardized test scores from all applicants, alleviating each student&#8217;s fear of standing out.</p><p>More generally, when Caplan lists the traits that he thinks employers want, surprisingly few of them are directly related to employee productivity:</p><blockquote><p>&#8220;What are modern model workers like? They&#8217;re team players. They&#8217;re deferential to superiors, but not slavish. They&#8217;re congenial toward coworkers but put business first. They dress and groom conservatively. They say nothing remotely racist or sexist, and they stay a mile away from anything construable as sexual harassment. Perhaps most importantly, they know and do what&#8217;s expected, even when articulating social norms is difficult or embarrassing. Employers don&#8217;t have to tell a modern model worker what&#8217;s socially acceptable case by case.&#8221;</p></blockquote><p>Traits like employees&#8217; appearances, political correctness, and ability to intuit social norms don't help much with the object-level work involved in most jobs. What they <em>are</em> relevant for is managing the company&#8217;s image&#8212;whether in the eyes of other employees, potential customers, or even government regulators. But even if this makes sense in isolation, we&#8217;ve now hypothesized a labor &#8220;market&#8221; in which everyone is nervously looking around at everyone else to try to avoid appearing weird. This is no longer an economic equilibrium in any reasonable sense. Instead, it&#8217;s a social equilibrium&#8212;albeit one with major economic implications&#8212;and we&#8217;ll need new concepts to model it.</p><p>In a follow-up post I&#8217;ll discuss some sociological theories of college attendance&#8212;most notably Bourdieu&#8217;s theory of higher education as a consecration of cultural elites. Unfortunately such theories have not been specified very rigorously. So I&#8217;ll also attempt to bridge the gap between economics and sociology by describing the formation of an elite class in game-theoretic terms.</p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>As a human capital explanation, the &#8220;learning professional norms&#8221; hypothesis also suffers from many of the issues of the &#8220;learning academic knowledge&#8221; hypothesis&#8212;e.g. <a href="https://en.wikipedia.org/wiki/Sheepskin_effect">the sheepskin effect</a>. Additionally, there&#8217;s the question of <em>who</em> students are learning professional norms from. Academics are notoriously unbusinesslike in many ways; and if it&#8217;s other students, that raises the question of why the already-professional students they&#8217;re learning from don&#8217;t just go straight into the workforce.</p></div></div>]]></content:encoded></item><item><title><![CDATA[Aligning to Virtues]]></title><description><![CDATA[Which alignment target?]]></description><link>https://www.mindthefuture.info/p/aligning-to-virtues</link><guid isPermaLink="false">https://www.mindthefuture.info/p/aligning-to-virtues</guid><dc:creator><![CDATA[Richard Ngo]]></dc:creator><pubDate>Mon, 16 Feb 2026 08:44:29 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/85b58149-608e-4432-9499-d12ec859dc2b_691x479.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h3>Which alignment target?</h3><p>Suppose you&#8217;re an AI company or government, and you want to figure out what values to align your AI to. Here are three options, and some of their downsides:</p><p><strong>AIs that are aligned to a set of consequentialist values</strong> are incentivized to acquire power to pursue those values. This creates power struggles between those AIs and:</p><ol><li><p>Humans who don&#8217;t share those values.</p></li><li><p>Humans who disagree with the AI about how to pursue those values.</p></li><li><p>Humans who don&#8217;t trust that the AI will actually pursue its stated values after gaining power.</p></li></ol><p>This is true whether those values are misaligned with all humans, aligned with some humans, chosen by aggregating all humans&#8217; values, or an attempt to specify some &#8220;moral truth&#8221;. In general, since humans have many different values, I think of the power struggle as being between coalitions which each contain some humans and some AIs.</p><p><strong>AIs that are aligned to a set of deontological principles</strong> (like refusing to harm humans) are safer, but also less flexible. What&#8217;s fine for an AI to do in one context might be harmful in another context; what&#8217;s fine for <em>one</em> AI to do might be very harmful for a million AIs to do. More generally, deontological principles draw a rigid line between acceptable and unacceptable behavior which is often either too restrictive or too permissive.</p><p>Alignment to deontological principles therefore creates power struggles over who gets to set the principles, and who has access to model weights to fine-tune the principles out of the AI.</p><p><strong>AIs that are corrigible/obedient to their human users</strong> can be told to do things which are arbitrarily harmful to other humans. This includes a spectrum of risks, from terrorism to totalitarianism. So it creates power struggles between humans for control over AIs (and especially over model weights, as discussed above). <a href="https://www.youtube.com/watch?v=4v3uqWeVmco">As per this talk</a>, it&#8217;s hard to draw a sharp distinction between risks from power-seeking AIs, versus risks from AIs that are corrigible to power-seeking users. Ideally we&#8217;d choose an alignment target which mitigates both risks.</p><p>Thus far attempts to compromise between these challenges (e.g. various model specs) have basically used ad-hoc combinations of these three approaches. However, this doesn&#8217;t seem like a very robust long-term solution. Below I outline an alternative which I think is more desirable.</p><h3>Aligning to virtues</h3><p>I personally would not like to be governed by politicians who are aligned to any of these three options. Instead, above all else I&#8217;d like politicians to be aligned to common-sense virtues like integrity, honor, kindness and dutifulness (and have experience balancing between them). This suggests that such virtues are also a promising target towards which to try to align AIs.</p><p>I intend to elaborate on my conception of virtue ethics (and why it&#8217;s the best way to understand ethics) in a series of upcoming posts. It&#8217;s a little difficult to comprehensively justify my &#8220;aligning to virtues&#8221; proposal in advance of that. However, since I&#8217;ve already sat on this post for almost a year, for now I&#8217;ll just briefly outline some of the benefits of virtues as an alignment target:</p><ol><li><p><strong>Virtues generalize deontological rules</strong>. Deontological rules are often very rigid, as discussed above. Virtues can be seen as more nuanced, flexible versions of them. For example, a deontologist might avoid lying while still misleading others. However, someone who has internalized the virtue of honesty will proactively try to make sure that they&#8217;re understood correctly. Especially as AIs become more intelligent than humans, we would like their values to generalize further.</p></li><li><p><strong>Situational awareness becomes a feature not just a challenge</strong>. Today we try to test AI whether AIs will obey instructions in often-implausible hypothetical scenarios. But as AIs get more intelligent, trying to hide their actual situation from them will become harder and harder. However, the benefit of this is that we&#8217;ll be able to align them to values which require them to know about their situation. For example, following an instruction given by the president might be better (or worse) than following an instruction given by a typical person. And following an instruction given to many AIs might be better (or worse) than following an instruction that&#8217;s only given to one AI. Situationally aware AIs will by default know which case they&#8217;re in.<br>Deontological values don&#8217;t really account for such distinctions: you should follow deontology no matter who or where you are. Corrigibility does, but only in a limited way (e.g. distinguishing between authorized users and non-authorized users). Conversely, virtues and consequentialist values are approaches which allow AIs to apply their situational awareness to make flexible context-dependent choices.</p></li><li><p><strong>Credit hacking becomes a feature not just a challenge</strong>. One concern about (mis)alignment is that AIs will find ways to preserve their values even when trained to do otherwise (a possibility sometimes known as <a href="https://www.alignmentforum.org/posts/EeAgytDZbDjRznPMA/gradient-hacking-definitions-and-examples">credit hacking</a>). Again, however, we can use this to our advantage. One characteristic trait of virtues is that they&#8217;re <em>robust</em> to a wide range of possible inputs. For example, it&#8217;s far easier for a consequentialist to reason themselves into telling a white lie, than it is for someone strongly committed to the virtue of honesty. So we should expect that AIs who start off virtuous will have an easier time preserving their values even when humans are trying to train them to cause harm. This might mean that AI companies can release models with fine-tuning access (or even open-source models) which are still very hard to misuse.</p></li><li><p><strong>Multi-agent interactions become a feature not just a challenge</strong>. If you align one AI, how should it interact with other AIs? I think of virtues as traits that govern cooperation between many agents, allowing them to work together while also reinforcing each other's virtues. For example, honesty as a virtue allows groups of agents to trust each other rather than succumbing to infighting, while setting each other&#8217;s incentives to further reinforce honesty. There&#8217;s a lot more to be done to flesh out this account of virtues, but insofar as it&#8217;s reasonable, then virtues are a much more scalable solution for aligning each of many copies of an AI, than the others discussed above.</p></li><li><p><strong>There&#8217;s more agreement on virtues than there is on most other types of values</strong>. For example, many people disagree about which politicians are good or bad in consequentialist terms, but they&#8217;ll tend to agree much more about which virtues different politicians display.</p></li></ol><p>In practice, I expect that the virtues we&#8217;ll want AIs to be aligned to are fairly different from the virtues we want human leaders to be aligned to. Both theoretical work (e.g. on defining virtues) and empirical work (e.g. on seeing how applying different virtues affect AI behavior in practice) seem valuable to identify a good virtue-based AI alignment target.</p><p>The main downside of trying to align to virtues is that it gives AIs more leeway in how they make decisions, and so it&#8217;s harder to tell whether our alignment techniques have succeeded or failed. But that will just be increasingly true of AIs in general, so we may as well plan for it.</p>]]></content:encoded></item><item><title><![CDATA[Distributed vs centralized agents]]></title><description><![CDATA[With applications to the far future]]></description><link>https://www.mindthefuture.info/p/distributed-vs-centralized-agents</link><guid isPermaLink="false">https://www.mindthefuture.info/p/distributed-vs-centralized-agents</guid><dc:creator><![CDATA[Richard Ngo]]></dc:creator><pubDate>Mon, 09 Feb 2026 21:21:28 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/a2d24ca8-6cb0-4959-9ab1-2b4f3238dd6e_494x484.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Much of my thinking over the last year has focusing on understanding the concept of &#8220;distributed agents&#8221;, as opposed to the &#8220;centralized agents&#8221; that the existing paradigm of expected utility maximization describes. One way of describing the difference is in terms of how autonomous their subagents are. Another is that centralized agents are more efficient (as sometimes formalized by the notion of &#8220;coherence&#8221;), while distributed agents are more robust.</p><p>Unfortunately robustness is hard to formalize, since it requires that you perform well even in unpredicted (and sometimes unpredictable) situations. I give some tentative characterizations of distributed agents below, but there&#8217;s still a lot of work to be done to formally define distributed agents. And ultimately I&#8217;d like to go further, to understand how agents can have both properties&#8212;which is roughly what I mean by &#8220;<a href="https://www.mindthefuture.info/p/towards-a-scale-free-theory-of-intelligent">coalitional agency</a>&#8220;.</p><p>I gave a talk on the distinction about seven months ago. I&#8217;d been hoping to write up the main ideas at more length, but since that doesn&#8217;t look like it&#8217;ll happen any time soon, I&#8217;m sharing the slides below. Hopefully they&#8217;re reasonably comprehensible by themselves, but feel free to ask questions about any parts that are unclear.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!iYbO!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2f66958-d7b8-4201-a934-e68f90810c0c_960x540.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!iYbO!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2f66958-d7b8-4201-a934-e68f90810c0c_960x540.jpeg 424w, https://substackcdn.com/image/fetch/$s_!iYbO!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2f66958-d7b8-4201-a934-e68f90810c0c_960x540.jpeg 848w, https://substackcdn.com/image/fetch/$s_!iYbO!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2f66958-d7b8-4201-a934-e68f90810c0c_960x540.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!iYbO!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2f66958-d7b8-4201-a934-e68f90810c0c_960x540.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!iYbO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2f66958-d7b8-4201-a934-e68f90810c0c_960x540.jpeg" width="960" height="540" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b2f66958-d7b8-4201-a934-e68f90810c0c_960x540.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:540,&quot;width&quot;:960,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!iYbO!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2f66958-d7b8-4201-a934-e68f90810c0c_960x540.jpeg 424w, https://substackcdn.com/image/fetch/$s_!iYbO!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2f66958-d7b8-4201-a934-e68f90810c0c_960x540.jpeg 848w, https://substackcdn.com/image/fetch/$s_!iYbO!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2f66958-d7b8-4201-a934-e68f90810c0c_960x540.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!iYbO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2f66958-d7b8-4201-a934-e68f90810c0c_960x540.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!sGsg!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d3a49fa-3d5c-4f3d-b936-7a8489c5cd7b_960x540.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!sGsg!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d3a49fa-3d5c-4f3d-b936-7a8489c5cd7b_960x540.jpeg 424w, https://substackcdn.com/image/fetch/$s_!sGsg!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d3a49fa-3d5c-4f3d-b936-7a8489c5cd7b_960x540.jpeg 848w, https://substackcdn.com/image/fetch/$s_!sGsg!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d3a49fa-3d5c-4f3d-b936-7a8489c5cd7b_960x540.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!sGsg!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d3a49fa-3d5c-4f3d-b936-7a8489c5cd7b_960x540.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!sGsg!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d3a49fa-3d5c-4f3d-b936-7a8489c5cd7b_960x540.jpeg" width="960" height="540" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3d3a49fa-3d5c-4f3d-b936-7a8489c5cd7b_960x540.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:540,&quot;width&quot;:960,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!sGsg!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d3a49fa-3d5c-4f3d-b936-7a8489c5cd7b_960x540.jpeg 424w, https://substackcdn.com/image/fetch/$s_!sGsg!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d3a49fa-3d5c-4f3d-b936-7a8489c5cd7b_960x540.jpeg 848w, https://substackcdn.com/image/fetch/$s_!sGsg!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d3a49fa-3d5c-4f3d-b936-7a8489c5cd7b_960x540.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!sGsg!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3d3a49fa-3d5c-4f3d-b936-7a8489c5cd7b_960x540.jpeg 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!j12Q!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7a3305e6-8373-4115-88e0-23fbee7d2d63_960x540.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!j12Q!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7a3305e6-8373-4115-88e0-23fbee7d2d63_960x540.jpeg 424w, https://substackcdn.com/image/fetch/$s_!j12Q!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7a3305e6-8373-4115-88e0-23fbee7d2d63_960x540.jpeg 848w, https://substackcdn.com/image/fetch/$s_!j12Q!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7a3305e6-8373-4115-88e0-23fbee7d2d63_960x540.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!j12Q!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7a3305e6-8373-4115-88e0-23fbee7d2d63_960x540.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!j12Q!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7a3305e6-8373-4115-88e0-23fbee7d2d63_960x540.jpeg" width="960" height="540" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7a3305e6-8373-4115-88e0-23fbee7d2d63_960x540.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:540,&quot;width&quot;:960,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!j12Q!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7a3305e6-8373-4115-88e0-23fbee7d2d63_960x540.jpeg 424w, https://substackcdn.com/image/fetch/$s_!j12Q!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7a3305e6-8373-4115-88e0-23fbee7d2d63_960x540.jpeg 848w, https://substackcdn.com/image/fetch/$s_!j12Q!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7a3305e6-8373-4115-88e0-23fbee7d2d63_960x540.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!j12Q!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7a3305e6-8373-4115-88e0-23fbee7d2d63_960x540.jpeg 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!5Th1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4dce3e03-f436-4dfd-a95c-c89429bbd7c6_960x540.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!5Th1!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4dce3e03-f436-4dfd-a95c-c89429bbd7c6_960x540.jpeg 424w, https://substackcdn.com/image/fetch/$s_!5Th1!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4dce3e03-f436-4dfd-a95c-c89429bbd7c6_960x540.jpeg 848w, https://substackcdn.com/image/fetch/$s_!5Th1!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4dce3e03-f436-4dfd-a95c-c89429bbd7c6_960x540.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!5Th1!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4dce3e03-f436-4dfd-a95c-c89429bbd7c6_960x540.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!5Th1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4dce3e03-f436-4dfd-a95c-c89429bbd7c6_960x540.jpeg" width="960" height="540" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4dce3e03-f436-4dfd-a95c-c89429bbd7c6_960x540.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:540,&quot;width&quot;:960,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!5Th1!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4dce3e03-f436-4dfd-a95c-c89429bbd7c6_960x540.jpeg 424w, https://substackcdn.com/image/fetch/$s_!5Th1!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4dce3e03-f436-4dfd-a95c-c89429bbd7c6_960x540.jpeg 848w, https://substackcdn.com/image/fetch/$s_!5Th1!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4dce3e03-f436-4dfd-a95c-c89429bbd7c6_960x540.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!5Th1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4dce3e03-f436-4dfd-a95c-c89429bbd7c6_960x540.jpeg 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!E-82!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6fe91c8b-d28e-46e8-81d5-d8c456757fd2_960x540.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!E-82!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6fe91c8b-d28e-46e8-81d5-d8c456757fd2_960x540.jpeg 424w, https://substackcdn.com/image/fetch/$s_!E-82!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6fe91c8b-d28e-46e8-81d5-d8c456757fd2_960x540.jpeg 848w, https://substackcdn.com/image/fetch/$s_!E-82!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6fe91c8b-d28e-46e8-81d5-d8c456757fd2_960x540.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!E-82!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6fe91c8b-d28e-46e8-81d5-d8c456757fd2_960x540.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!E-82!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6fe91c8b-d28e-46e8-81d5-d8c456757fd2_960x540.jpeg" width="960" height="540" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6fe91c8b-d28e-46e8-81d5-d8c456757fd2_960x540.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:540,&quot;width&quot;:960,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!E-82!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6fe91c8b-d28e-46e8-81d5-d8c456757fd2_960x540.jpeg 424w, https://substackcdn.com/image/fetch/$s_!E-82!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6fe91c8b-d28e-46e8-81d5-d8c456757fd2_960x540.jpeg 848w, https://substackcdn.com/image/fetch/$s_!E-82!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6fe91c8b-d28e-46e8-81d5-d8c456757fd2_960x540.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!E-82!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6fe91c8b-d28e-46e8-81d5-d8c456757fd2_960x540.jpeg 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><a href="https://www.lesswrong.com/posts/fjfWrKhEawwBGCTGs/a-simple-case-for-extreme-inner-misalignment">See more on my interpretation of Yudkowsky here</a>; note that he disagrees with my emphasis on compression though (as per the exchange in the comments).</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!VdyH!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f78a867-99ca-4a6c-b19c-74d211286503_960x540.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!VdyH!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f78a867-99ca-4a6c-b19c-74d211286503_960x540.jpeg 424w, https://substackcdn.com/image/fetch/$s_!VdyH!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f78a867-99ca-4a6c-b19c-74d211286503_960x540.jpeg 848w, https://substackcdn.com/image/fetch/$s_!VdyH!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f78a867-99ca-4a6c-b19c-74d211286503_960x540.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!VdyH!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f78a867-99ca-4a6c-b19c-74d211286503_960x540.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!VdyH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f78a867-99ca-4a6c-b19c-74d211286503_960x540.jpeg" width="960" height="540" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6f78a867-99ca-4a6c-b19c-74d211286503_960x540.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:540,&quot;width&quot;:960,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!VdyH!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f78a867-99ca-4a6c-b19c-74d211286503_960x540.jpeg 424w, https://substackcdn.com/image/fetch/$s_!VdyH!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f78a867-99ca-4a6c-b19c-74d211286503_960x540.jpeg 848w, https://substackcdn.com/image/fetch/$s_!VdyH!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f78a867-99ca-4a6c-b19c-74d211286503_960x540.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!VdyH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6f78a867-99ca-4a6c-b19c-74d211286503_960x540.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><a href="https://www.mindthefuture.info/p/why-im-not-a-bayesian">My post on why I&#8217;m not a bayesian</a> also gives a sense of what understanding epistemology in more distributed terms looks like.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!5qo3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F728b4ebe-17b3-4f2e-9417-86f0dd7aab2f_960x540.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!5qo3!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F728b4ebe-17b3-4f2e-9417-86f0dd7aab2f_960x540.jpeg 424w, https://substackcdn.com/image/fetch/$s_!5qo3!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F728b4ebe-17b3-4f2e-9417-86f0dd7aab2f_960x540.jpeg 848w, https://substackcdn.com/image/fetch/$s_!5qo3!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F728b4ebe-17b3-4f2e-9417-86f0dd7aab2f_960x540.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!5qo3!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F728b4ebe-17b3-4f2e-9417-86f0dd7aab2f_960x540.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!5qo3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F728b4ebe-17b3-4f2e-9417-86f0dd7aab2f_960x540.jpeg" width="960" height="540" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/728b4ebe-17b3-4f2e-9417-86f0dd7aab2f_960x540.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:540,&quot;width&quot;:960,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!5qo3!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F728b4ebe-17b3-4f2e-9417-86f0dd7aab2f_960x540.jpeg 424w, https://substackcdn.com/image/fetch/$s_!5qo3!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F728b4ebe-17b3-4f2e-9417-86f0dd7aab2f_960x540.jpeg 848w, https://substackcdn.com/image/fetch/$s_!5qo3!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F728b4ebe-17b3-4f2e-9417-86f0dd7aab2f_960x540.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!5qo3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F728b4ebe-17b3-4f2e-9417-86f0dd7aab2f_960x540.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>&#8220;Will&#8217;s very rough first pass&#8221; is a reference to <a href="https://utilitarianism.net/utilitarianism-and-practical-ethics/">the passage in his textbook on utilitarianism</a> where Will MacAskill describes what decision procedure a utilitarian should follow. My point here is to contrast how much thought he (and other utilitarians) put into finding criteria of rightness, vs how rudimentary their thinking about decision procedures is.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ji3_!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F31dea3f9-927d-4de2-be88-2471b6507e5c_960x540.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ji3_!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F31dea3f9-927d-4de2-be88-2471b6507e5c_960x540.jpeg 424w, https://substackcdn.com/image/fetch/$s_!ji3_!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F31dea3f9-927d-4de2-be88-2471b6507e5c_960x540.jpeg 848w, https://substackcdn.com/image/fetch/$s_!ji3_!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F31dea3f9-927d-4de2-be88-2471b6507e5c_960x540.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!ji3_!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F31dea3f9-927d-4de2-be88-2471b6507e5c_960x540.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ji3_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F31dea3f9-927d-4de2-be88-2471b6507e5c_960x540.jpeg" width="960" height="540" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/31dea3f9-927d-4de2-be88-2471b6507e5c_960x540.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:540,&quot;width&quot;:960,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ji3_!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F31dea3f9-927d-4de2-be88-2471b6507e5c_960x540.jpeg 424w, https://substackcdn.com/image/fetch/$s_!ji3_!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F31dea3f9-927d-4de2-be88-2471b6507e5c_960x540.jpeg 848w, https://substackcdn.com/image/fetch/$s_!ji3_!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F31dea3f9-927d-4de2-be88-2471b6507e5c_960x540.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!ji3_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F31dea3f9-927d-4de2-be88-2471b6507e5c_960x540.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!yDyB!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd71ed05-c641-4246-8c66-4ee72b85a7a7_960x540.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!yDyB!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd71ed05-c641-4246-8c66-4ee72b85a7a7_960x540.jpeg 424w, https://substackcdn.com/image/fetch/$s_!yDyB!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd71ed05-c641-4246-8c66-4ee72b85a7a7_960x540.jpeg 848w, https://substackcdn.com/image/fetch/$s_!yDyB!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd71ed05-c641-4246-8c66-4ee72b85a7a7_960x540.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!yDyB!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd71ed05-c641-4246-8c66-4ee72b85a7a7_960x540.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!yDyB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd71ed05-c641-4246-8c66-4ee72b85a7a7_960x540.jpeg" width="960" height="540" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bd71ed05-c641-4246-8c66-4ee72b85a7a7_960x540.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:540,&quot;width&quot;:960,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!yDyB!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd71ed05-c641-4246-8c66-4ee72b85a7a7_960x540.jpeg 424w, https://substackcdn.com/image/fetch/$s_!yDyB!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd71ed05-c641-4246-8c66-4ee72b85a7a7_960x540.jpeg 848w, https://substackcdn.com/image/fetch/$s_!yDyB!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd71ed05-c641-4246-8c66-4ee72b85a7a7_960x540.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!yDyB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbd71ed05-c641-4246-8c66-4ee72b85a7a7_960x540.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!WcsJ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc65cf54a-a31c-44b5-8a12-d1b9ff98f827_960x540.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!WcsJ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc65cf54a-a31c-44b5-8a12-d1b9ff98f827_960x540.jpeg 424w, https://substackcdn.com/image/fetch/$s_!WcsJ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc65cf54a-a31c-44b5-8a12-d1b9ff98f827_960x540.jpeg 848w, https://substackcdn.com/image/fetch/$s_!WcsJ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc65cf54a-a31c-44b5-8a12-d1b9ff98f827_960x540.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!WcsJ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc65cf54a-a31c-44b5-8a12-d1b9ff98f827_960x540.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!WcsJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc65cf54a-a31c-44b5-8a12-d1b9ff98f827_960x540.jpeg" width="960" height="540" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c65cf54a-a31c-44b5-8a12-d1b9ff98f827_960x540.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:540,&quot;width&quot;:960,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!WcsJ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc65cf54a-a31c-44b5-8a12-d1b9ff98f827_960x540.jpeg 424w, https://substackcdn.com/image/fetch/$s_!WcsJ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc65cf54a-a31c-44b5-8a12-d1b9ff98f827_960x540.jpeg 848w, https://substackcdn.com/image/fetch/$s_!WcsJ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc65cf54a-a31c-44b5-8a12-d1b9ff98f827_960x540.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!WcsJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc65cf54a-a31c-44b5-8a12-d1b9ff98f827_960x540.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!rwYG!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F047bfcf9-abb0-4039-9d42-859ef3ee7f5e_960x540.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!rwYG!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F047bfcf9-abb0-4039-9d42-859ef3ee7f5e_960x540.jpeg 424w, https://substackcdn.com/image/fetch/$s_!rwYG!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F047bfcf9-abb0-4039-9d42-859ef3ee7f5e_960x540.jpeg 848w, https://substackcdn.com/image/fetch/$s_!rwYG!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F047bfcf9-abb0-4039-9d42-859ef3ee7f5e_960x540.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!rwYG!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F047bfcf9-abb0-4039-9d42-859ef3ee7f5e_960x540.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!rwYG!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F047bfcf9-abb0-4039-9d42-859ef3ee7f5e_960x540.jpeg" width="960" height="540" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/047bfcf9-abb0-4039-9d42-859ef3ee7f5e_960x540.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:540,&quot;width&quot;:960,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!rwYG!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F047bfcf9-abb0-4039-9d42-859ef3ee7f5e_960x540.jpeg 424w, https://substackcdn.com/image/fetch/$s_!rwYG!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F047bfcf9-abb0-4039-9d42-859ef3ee7f5e_960x540.jpeg 848w, https://substackcdn.com/image/fetch/$s_!rwYG!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F047bfcf9-abb0-4039-9d42-859ef3ee7f5e_960x540.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!rwYG!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F047bfcf9-abb0-4039-9d42-859ef3ee7f5e_960x540.jpeg 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.mindthefuture.info/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Mind the Future! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[21st Century Civilization curriculum]]></title><description><![CDATA[I&#8217;ve just released a curriculum on foundational questions in modern politics.]]></description><link>https://www.mindthefuture.info/p/21st-century-civilization-curriculum</link><guid isPermaLink="false">https://www.mindthefuture.info/p/21st-century-civilization-curriculum</guid><dc:creator><![CDATA[Richard Ngo]]></dc:creator><pubDate>Tue, 21 Oct 2025 07:35:03 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/709d2444-4562-4795-a9a5-e33cefd8417d_1732x996.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><a href="https://x.com/richardmcngo/status/1980531594139238622?s=46">I&#8217;ve just released</a> a curriculum on foundational questions in modern politics, which I drew up in collaboration with <a href="https://x.com/SamoBurja?ref_src=twsrc%5Egoogle%7Ctwcamp%5Eserp%7Ctwgr%5Eauthor">Samo Burja</a>. I&#8217;ve copied the introductory text and the section headings below; you can find the full curriculum at <a href="https://www.21civ.com/">www.21civ.com</a>.</p><p><strong><a href="https://forms.gle/aNUqFsrNFxipkUMm7">Sign up here</a> by 27 October to join the first cohort of discussion groups </strong>(which will meet weekly to discuss each of the 11 sections of the curriculum).</p><p>This curriculum is about Western civilization, and how it enables citizens of the Western world to live together in a just, orderly way. But it&#8217;s also about the 21st century, which has been characterized by the continual decline of many aspects of that civilization.</p><p>Despite our superior technology, there are many things that Western countries could do in the past that we can&#8217;t today&#8212;e.g. rapidly build large-scale infrastructure, maintain low-crime cities, and run competent bureaucracies. More importantly, it feels like there are no adults in the room: modern elites often seem unvirtuous and even unserious by historical standards<em>.</em> This curriculum focuses on explaining what changed, and how to orient to the world we now find ourselves in.</p><p>Discussions of large-scale political issues can be unsettling or jarring. So the curriculum intertwines discussion of what&#8217;s happening on a factual level, with readings on how to develop a healthy emotional and ethical stance towards politics (culminating in the final week&#8217;s focus on cultivating virtue). It also strongly prioritizes honesty and clarity of writing (even on topics often considered taboo), which is one reason why most readings are informal blog posts or essays rather than academic papers.</p><p>Given the sheer scope of the topics covered by the curriculum, it does not aim at comprehensiveness; nor does it try to give detailed strategies for solving civilizational decay. Indeed, given the accelerating development of AI (as discussed in week 10) the coming decades are likely to be extremely unpredictable. Readers should instead think of the curriculum as a starting point for informed, realistic discussions about how we as a civilization can steer ourselves through the coming turmoil.</p><h3>Contents</h3><ul><li><p>Week 1: Western Culture</p></li><li><p>Week 2: Civilizational Decay</p></li><li><p>Week 3: The Managerial State</p></li><li><p>Week 4: The Open Society and its Discontents</p></li><li><p>Week 5: The Psychology of Modern Elites</p></li><li><p>Week 6: The Unprotected Class</p></li><li><p>Week 7: World on Fire</p></li><li><p>Week 8: Industry and Money</p></li><li><p>Week 9: Sociopolitics</p></li><li><p>Week 10: Technology and Civilization</p></li><li><p>Week 11: Reviving Virtue</p></li></ul><p><a href="http://www.21civ.com">Read the full curriculum here</a>.</p>]]></content:encoded></item><item><title><![CDATA[Underdog bias rules everything around me]]></title><description><![CDATA[They're just as scared of us as we are of them]]></description><link>https://www.mindthefuture.info/p/underdog-bias-rules-everything-around</link><guid isPermaLink="false">https://www.mindthefuture.info/p/underdog-bias-rules-everything-around</guid><dc:creator><![CDATA[Richard Ngo]]></dc:creator><pubDate>Sun, 17 Aug 2025 19:18:05 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/e90f6b45-f860-462e-b18b-4a41ab48541f_475x258.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><strong>People very often underrate how much power they (and their allies) have, and overrate how much power their enemies have. I call this &#8220;underdog bias&#8221;, and I think it&#8217;s the most important cognitive bias to understand in order to make sense of modern society.</strong></p><p>I&#8217;ll start by describing a closely-related phenomenon. The <a href="https://en.wikipedia.org/wiki/Hostile_media_effect">hostile media effect</a> is a well-known bias whereby people tend to perceive news they read or watch as skewed against their side. For example, pro-Palestinian students shown a video clip tended to judge that the clip would make viewers more pro-Israel, while pro-Israel students shown the same clip thought it&#8217;d make viewers more pro-Palestine. Similarly, sports fans often see referees as being biased against their own team.</p><p>The hostile media effect is particularly striking because it arises in settings where there&#8217;s relatively little scope for bias. People watching media clips and sports are all seeing exactly the same videos. And sports in particular are played on very even terms, where fairness just means enforcing the rules impartially.</p><p>But most possible conflicts are much less symmetric, both in terms of what information each side has, and even in terms of what game each side is playing. Consider, for instance, an argument about whether big corporations have too much power. The proponent might point to corporations&#8217; wealth, employee talent, and lobbying ability; their opponent might point to how many regulations they have to follow, how much corporations compete between themselves, and how strong anti-corporate public sentiment is. In order to evaluate a question like this, people need to decide both <em>how to draw coalition boundaries </em>(to what extent should big corporations be counted as a single unified group?) and <em>how to weigh different types of power against each other.</em></p><p>I think that biases in how these weighings and boundaries are evaluated are a much bigger deal than biases in evaluating fairness in isolated contexts. Specifically, I think that people typically underrate the types of power they have, and overrate the types of power their opponents have. You&#8217;re intimately familiar with the limitations of your own abilities&#8212;you run into them regularly, often in deeply frustrating ways. You track all the fractures inside your own coalition, and they often seem fundamental and intractable. Conversely, it&#8217;s easy to forget about the things which are much easier for you than for your opponents, and to view their internal rivalries as temporary and easily-resolved.</p><p>These effects are exacerbated by information asymmetries, aka the &#8220;fog of war&#8221;. You know who&#8217;s working with you; you don&#8217;t know who&#8217;s working against you. When outside observers sympathize with your side, you know that they&#8217;re not actually contributing very much to your cause; when outside observers sympathize with your opponents, you don&#8217;t know if that&#8217;s a sign of enmity. Similarly, you know how your own plans are progressing, but you don&#8217;t know what your opponents are <a href="https://en.wikipedia.org/wiki/Emotive_conjugation">scheming</a>. To see how strong this effect can be, just look at fiction, where villains often implement arbitrarily-complicated schemes offscreen without breaking suspension of disbelief.</p><p>In addition to the hostile media effect, underdog bias is related to a number of other biases (like <a href="https://en.wikipedia.org/wiki/Hostile_attribution_bias">hostile attribution bias</a>, <a href="https://en.wikipedia.org/wiki/Siege_mentality">siege mentality</a>, <a href="https://en.wikipedia.org/wiki/Fundamental_attribution_error">the fundamental attribution error</a>, and simple <a href="https://www.edge.org/conversation/john_tooby-coalitional-instincts">tribalism</a>). But hopefully the description above conveys why I think it&#8217;s fundamental enough to be worth separating out.</p><h4><strong>Underdog bias in practice</strong></h4><p>In this section I give six examples of conflicts where each side thinks (with some justification) of themselves as the underdog, and rejects the idea of their opponents being the underdogs. I won&#8217;t try to defend them in detail, but they hopefully convey the pervasiveness of underdog bias.</p><ol><li><p><strong>Government vs industry</strong></p><ol><li><p>Working within a government is often a miserable experience. Government institutions are typically cash-strapped and unable to hire the most talented people; meanwhile they face heavy opposition from industry, which can spend enormous amounts of money lobbying, donating to candidates, etc. Many employees at regulatory agencies see themselves as underdogs trying to rein in whole industries, while working under strong political and bureaucratic constraints.</p></li><li><p>Big corporations have a lot of money, but are in a constant state of cut-throat competition, and also need to comply with a huge number of regulations. Bureaucrats can very easily make their lives harder in ways ranging from trivial (e.g. adding red tape) to existential (e.g. blocking acquisitions). So it&#8217;s easy for companies to feel like underdogs pitted against the power of the state&#8212;especially when it seems like exercises of that power might be influenced by their competitors.</p></li></ol></li><li><p><strong>Tech vs media</strong></p><ol><li><p>Many journalists see themselves as underdogs fighting the accumulation of power by special interests, of which tech is the most prominent. This feeling is exacerbated by how financially precarious the industry is. The rise of social media has <a href="https://www.pewresearch.org/short-reads/2021/07/13/u-s-newsroom-employment-has-fallen-26-since-2008/">decimated newspaper jobs</a>, leading newspapers to optimize far harder for engagement than they used to.</p></li><li><p>The tech industry has huge amounts of money, and the ability to reshape the world in many ways, but has historically wielded relatively little cultural and political influence. Many techies see themselves as underdogs fighting the cultural establishment, as exemplified by a pervasive (and often extreme) anti-tech bias in the news media. Techies (until recently) have felt like nobody represents their interests in DC, where both Democrats and Republicans have historically been anti-innovation.</p></li></ol></li><li><p><strong>Elites vs masses</strong></p><ol><li><p>Normal people view themselves as underdogs in a world where cultural elites (e.g. top university graduates) control almost all major institutions. Even when they vote for outsiders, the effects of those elections are blunted by elite control over institutions (especially the &#8220;deep state&#8221;). Elite ideology has increasingly diverged from the ideology of the masses, making many people feel helpless to elect anyone who will actually represent their interests.</p></li><li><p>Cultural elites control almost all major institutions, but are few in number. Elites feel scared of populist masses who distrust them, and who can vote for anti-elite candidates. The most elite groups (like billionaires or Jews) are often the ones it&#8217;s most socially acceptable to blame for problems, or even call for violence against.</p></li></ol></li><li><p><strong>Republicans vs Democrats</strong></p><ol><li><p>Many Republicans see themselves as underdogs fighting against the Democratic elites who have far more influence within universities, government bureaucracies, legacy media outlets, and tech companies. They see themselves as fighting against pervasive ideological control over a wide range of institutions.</p></li><li><p>Many Democrats view Republicans as having structural advantages&#8212;such as the Electoral College, or the Senate&#8212;which allow them to remain competitive even with fewer supporters. Democrats therefore view themselves as underdogs fighting for change against existing cultural norms and Republican-run power structures (like the Supreme Court).</p></li></ol></li><li><p><strong>America vs China</strong></p><ol><li><p>America has more power in the sense of established hegemony. China has more power in the sense of momentum and, increasingly, manufacturing and economic superiority. The likelihood of conflict between a rising power (who&#8217;s worried about overcoming their historical disadvantage) and an established power (who&#8217;s worried about their opponents&#8217; growth trajectory) is known as <a href="https://en.wikipedia.org/wiki/Thucydides_Trap">the Thucydides trap</a>.</p></li></ol></li><li><p><strong>Israel vs Palestine</strong></p><ol><li><p>Many Palestinians see themselves as underdogs against an opponent with far more military power than them, which is also strongly backed by the US.</p></li><li><p>While Israel has power over Palestine specifically, it&#8217;s also surrounded by hostile neighbors far larger than it, which have often stated their intention (and repeatedly actually tried) to wipe Israel off the map. Many Israelis see themselves as underdogs who have to defend themselves against an entire region (as well as increasingly-hostile public opinion in the west).</p></li></ol></li></ol><p>The two maps below are sometimes shown by supporters of each side as a way of conveying how much of an underdog their side is. (To be clear, I&#8217;m not endorsing either of them, just illustrating the rhetorical strategies being used.)</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!RrAA!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb93da93-2812-4b44-add7-d4233313f710_580x386.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!RrAA!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb93da93-2812-4b44-add7-d4233313f710_580x386.png 424w, https://substackcdn.com/image/fetch/$s_!RrAA!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb93da93-2812-4b44-add7-d4233313f710_580x386.png 848w, https://substackcdn.com/image/fetch/$s_!RrAA!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb93da93-2812-4b44-add7-d4233313f710_580x386.png 1272w, https://substackcdn.com/image/fetch/$s_!RrAA!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb93da93-2812-4b44-add7-d4233313f710_580x386.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!RrAA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb93da93-2812-4b44-add7-d4233313f710_580x386.png" width="580" height="386" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/cb93da93-2812-4b44-add7-d4233313f710_580x386.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:386,&quot;width&quot;:580,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!RrAA!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb93da93-2812-4b44-add7-d4233313f710_580x386.png 424w, https://substackcdn.com/image/fetch/$s_!RrAA!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb93da93-2812-4b44-add7-d4233313f710_580x386.png 848w, https://substackcdn.com/image/fetch/$s_!RrAA!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb93da93-2812-4b44-add7-d4233313f710_580x386.png 1272w, https://substackcdn.com/image/fetch/$s_!RrAA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcb93da93-2812-4b44-add7-d4233313f710_580x386.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!0mbr!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64ff0967-9d59-4ca9-ab45-6ffb08187ac3_1208x676.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!0mbr!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64ff0967-9d59-4ca9-ab45-6ffb08187ac3_1208x676.png 424w, https://substackcdn.com/image/fetch/$s_!0mbr!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64ff0967-9d59-4ca9-ab45-6ffb08187ac3_1208x676.png 848w, https://substackcdn.com/image/fetch/$s_!0mbr!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64ff0967-9d59-4ca9-ab45-6ffb08187ac3_1208x676.png 1272w, https://substackcdn.com/image/fetch/$s_!0mbr!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64ff0967-9d59-4ca9-ab45-6ffb08187ac3_1208x676.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!0mbr!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64ff0967-9d59-4ca9-ab45-6ffb08187ac3_1208x676.png" width="568" height="317.8543046357616" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/64ff0967-9d59-4ca9-ab45-6ffb08187ac3_1208x676.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:676,&quot;width&quot;:1208,&quot;resizeWidth&quot;:568,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!0mbr!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64ff0967-9d59-4ca9-ab45-6ffb08187ac3_1208x676.png 424w, https://substackcdn.com/image/fetch/$s_!0mbr!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64ff0967-9d59-4ca9-ab45-6ffb08187ac3_1208x676.png 848w, https://substackcdn.com/image/fetch/$s_!0mbr!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64ff0967-9d59-4ca9-ab45-6ffb08187ac3_1208x676.png 1272w, https://substackcdn.com/image/fetch/$s_!0mbr!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F64ff0967-9d59-4ca9-ab45-6ffb08187ac3_1208x676.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Underdog bias doesn&#8217;t imply that any of these groups are <em>wrong</em> to be scared of their opponents&#8217; power. But it does suggest that they&#8217;ll tend to underestimate their own power and, crucially, underestimate how scared their opponents are of them. (The applications of this principle to the politics of AI are left as an exercise for the reader.)</p><h4><strong>Why underdog bias?</strong></h4><p>If underdog bias makes us so wrong about the world, why is it such a strong psychological effect? Some cognitive biases are just clear-cut mistakes, but we should expect that the strongest &#8220;biases&#8221; were evolutionarily adaptive in some way.</p><p>The descriptions I gave above suggest that there are qualitative differences in the types of reasoning about ourselves and our enemies&#8212;roughly corresponding to <a href="https://www.overcomingbias.com/p/abstractdistanthtml">near vs far mode</a>. In near mode we focus on concrete, nuanced details about our local situation. In far mode, we construct larger-scale narratives, in more black-and-white terms, often for the sake of signaling to others.</p><p>Why might signaling that you&#8217;re the underdog be more important than having accurate beliefs? One possibility is to gain allies. <a href="http://ts-si.org/files/AppealOfTheUnderdogPSPB1603.pdf">Vandallo et al.</a> have <a href="http://spr.sagepub.com/content/early/2013/03/03/0265407513477629.abstract">a few studies</a> on the effects of appearing to be the disadvantaged side. <a href="https://slatestarcodex.com/2013/05/18/against-bravery-debates/">Scott Alexander</a> summarizes their conclusions as follows: &#8220;if you get yourself perceived as the brave long-suffering underdog, people will support your cause and, as an added bonus, want to have sex with you&#8221;. And <a href="https://www.lesswrong.com/posts/A4MK9RQqSAJZjanQD/why-support-the-underdog">in this post</a> he points to the longstanding prevalence of underdogs in narratives (stretching back to myths like that of David and Goliath).</p><p>But he also recognizes that there&#8217;s a big difference between reported and actual support. All else equal, underdogs are pluckier and their victories more impressive&#8212;so it makes sense that we support them on a narrative level, when there&#8217;s no cost to doing so. But in real life supporting the underdog means that you&#8217;re on the side most likely to lose. Whether or not underdog bias is beneficial for gaining allies will therefore depend on whether those allies are more concerned about being (or looking) virtuous, or more concerned about actually winning. And while the modern era is dominated by virtue signaling, that was much less true in the ancestral environment, where resources were much scarcer. In that setting you&#8217;d instead expect people to have &#8220;overdog bias&#8221; which makes them overestimate their own side&#8217;s strength (which is one way that tribalism, patriotism, etc. can be interpreted).</p><p>Another possibility is that underdog bias is most valuable as a way of firing up your own supporters. I see this in action whenever I accidentally give my email to a political candidate, and get bombarded with emails about how I need to donate because the other side is on the verge of an overwhelming victory. But again, there&#8217;s a missing link: why should fear make you fight harder? On a rational agent model, being the underdog could make you decide fighting isn&#8217;t worth it&#8212;or even make you defect to the enemy. And on an emotional level, <a href="https://malcolmocean.com/2022/02/towardsness-awayness-motivation-arent-symmetrical/">being scared makes it much harder</a> to think clearly or navigate complicated situations.</p><p>So my best guess is that underdog bias was useful because ancestral conflicts were <em>simple</em> and <em>compulsory</em>. In other words, our political intuitions are calibrated for a world where alliances are more tribal&#8212;where we don&#8217;t have freedom of movement or freedom of association. People used to be stuck with their family/tribe/ethnic group whether they liked it or not; if they tried to ally with another, they were often rejected, or at best permanently viewed as an untrustworthy outsider. So the only rational response to being in a worse position would be to fight harder, using fairly straightforward and intuitive strategies.</p><p>In one way, this response is maladaptive in the modern world&#8212;where fewer battle lines are based on immutable characteristics or irreconcilable differences, and the best way to approach conflict is less intuitive. Yet as I mentioned above, the modern world is also much more sympathetic to victims. This suggests that underdog bias may have gradually transitioned from a way of firing up one&#8217;s supporters, to something closer to a <a href="https://en.wikipedia.org/wiki/Victim_mentality">victim complex</a> aimed at evoking sympathy from onlookers.</p><p>This whole section has been very speculative, and I&#8217;m still not confident in my answer to where underdog bias comes from. But we don&#8217;t need an explanation of underdog bias to believe that it corrupts many people&#8217;s thinking about complex issues. How can you actually reduce your underdog bias, though? The best approach I&#8217;ve found is simple in theory (though devilishly difficult in practice). Say to yourself: &#8220;they're just as scared of us as we are of them.&#8221; It&#8217;s true far more often than you think.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.mindthefuture.info/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Mind the Future! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[On Pessimization]]></title><description><![CDATA[Getting the opposite of what you want]]></description><link>https://www.mindthefuture.info/p/on-pessimization</link><guid isPermaLink="false">https://www.mindthefuture.info/p/on-pessimization</guid><dc:creator><![CDATA[Richard Ngo]]></dc:creator><pubDate>Sat, 16 Aug 2025 18:07:29 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/4ffd60f5-d181-43cb-a522-48be53256f9d_1366x1178.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>&#8220;Your worst sin is that you have destroyed and betrayed yourself for nothing.&#8221; - Dostoevsky</em></p><p>When people set an ambitious goal, they can fail simply by not changing the world very much. But there&#8217;s another surprisingly common way to fail: by achieving the <em>opposite</em> of their goal. I call this effect <em>pessimization: </em>the opposite of optimization.</p><p>Though pessimization is an uncommon term, it&#8217;s not an uncommon concept. We allude to it whenever we call someone their own worst enemy, or predict that their actions will backfire. We try to take advantage of it with <a href="https://en.wikipedia.org/wiki/Reverse_psychology">reverse psychology</a>. It&#8217;s the subject of Robert Conquest&#8217;s <a href="https://en.wikipedia.org/wiki/Robert_Conquest#Laws_of_politics">third law of politics</a>: &#8220;the simplest way to explain the behavior of any bureaucratic organization is to assume that it is controlled by a cabal of its enemies.&#8221; And the <a href="https://en.wikipedia.org/wiki/The_purpose_of_a_system_is_what_it_does">POSIWID</a> aphorism (&#8220;the purpose of a system is what it does&#8221;) is often used to describe ways in which a system actively opposes its nominal purpose. I&#8217;ve also previously described the pessimization of ideological advocacy under the heading of &#8220;<a href="https://x.com/RichardMCNgo/status/1810495607590543701">the activist&#8217;s curse</a>&#8221;.</p><p>I divide pessimization into three types, each of which I&#8217;ll discuss in detail. The most intuitive is what I call <em>direct pessimization</em>: when an enemy chooses what to do based on what will harm your interests the most. This is the sense in which &#8220;pessimize&#8221; is used by e.g. <a href="https://www.lesswrong.com/posts/yaJsCQokiyeLFHhgy/incorporating-justice-theory-into-decision-theory">Yudkowsky</a> and <a href="https://www.lesswrong.com/posts/HoQ5Rp7Gs6rebusNP/superintelligent-ai-is-necessary-for-an-amazing-future-but-1">Soares</a>. Drivers of direct pessimization include sadism, revenge and threats.</p><p>Conversely, <em>indirect pessimization</em> arises when your actions help other people achieve the opposite of your goals, even though they&#8217;re not deliberately trying to hurt you. And <em>perverse pessimization</em> occurs when an agent or coalition nominally dedicated to a goal is actively trying to achieve the opposite of that goal&#8212;i.e. it&#8217;s directly pessimizing itself.</p><p>I&#8217;d like to improve our understanding of pessimization so that we can better create, identify, and promote remedies to it. In the rest of this post I characterize each of the three types of pessimization I mentioned above. I treat them as <a href="https://www.mindthefuture.info/p/towards-a-scale-free-theory-of-intelligent">scale-free phenomena</a>, describing examples within individuals, organizations, countries, and even whole civilizations.&nbsp;</p><h3>Direct pessimization</h3><p><em>Direct pessimization</em> is when one agent or faction is specifically trying to hurt another&#8212;not as a side effect of other actions, but rather as a means of achieving their desired outcome. For example, imprisoning criminals in order to prevent them from committing further crimes wouldn&#8217;t qualify. However, punishing criminals for the sake of deterrence or retribution would: in the first case the prisoners&#8217; suffering is the means by which others are deterred, in the second case their suffering itself is the desired outcome. While less of a focus today, these motivations were more salient in historical penal systems, which used a wide range of physical punishments ranging from flogging all the way to cruel and unusual tortures. By &#8220;hurting&#8221; I don&#8217;t just mean causing physical pain, though, but also other ways of harming someone&#8217;s interests&#8212;like the social humiliation of <a href="https://en.wikipedia.org/wiki/Stocks">stocks</a>, exile, execution, <a href="https://en.wikipedia.org/wiki/Attainder">attainder</a>, or the Chinese <a href="https://en.wikipedia.org/wiki/Nine_familial_exterminations">nine familial exterminations</a>.</p><p>To be clear, the distinction between a &#8220;means&#8221; and a &#8220;side effect&#8221; of achieving a goal is a thorny one, <a href="https://plato.stanford.edu/entries/double-effect/">which many philosophers have debated</a>. For example, <a href="https://plato.stanford.edu/entries/double-effect/#RoleConvNormWarf">which harms inflicted on an enemy nation</a> are necessary components of winning a war, and which are merely unfortunate side effects of winning the war? I won&#8217;t try to resolve the edge cases here&#8212;however, clear examples of larger-scale direct pessimization include terrorist attacks and genocides, as well as terror bombing civilians during wars. At even larger scales, <a href="https://longtermrisk.org/reducing-risks-of-astronomical-suffering-a-neglected-priority/#IIII_Unawareness_of_possible_sources_of_astronomical_suffering">s-risk threat models</a> outline how and why superintelligences might threaten to pessimize each others&#8217; utility functions.</p><p>Another driver of direct pessimization is sadism. We see this in serial killers, sociopaths, and many of the people who end up actually implementing the punishments discussed above. However, one reason why I wanted to carve out the category of &#8220;direct pessimization&#8221; in the first place is because I don&#8217;t know how to draw a principled distinction between sadism, revenge, deterrence, and hurting others for instrumental benefit. For example, many people think of social status as zero-sum, and display petty cruelty (e.g. mocking jokes or mean-spirited gossip) in order to maintain their position in the hierarchy. The type of sadism that is characteristic of sociopaths is typically much more intense, but I&#8217;m not sure that the underlying logic is so different. Indeed, since sociopathy often results from traumatic childhood experiences, we could view it as a kind of &#8220;revenge&#8221; on the world at large&#8212;not too dissimilar to how normal people sometimes feel about those who have wronged them.</p><p>The examples I&#8217;ve discussed above only feature pessimization in one direction&#8212;but direct pessimization often spurs cycles of retribution, at many scales. Within individual psychology, internal conflict (e.g. procrastination) can prompt increasingly harsh self-coercive strategies, like vicious self-criticism. In relationships, emotional insecurity can lead partners to approach each other with blame, shame, and cruelty, until the person who loves them the most is also the one who hurts them the most. In politics, focusing on the other side&#8217;s failings can provoke cycles of <a href="https://en.wikipedia.org/wiki/Negative_partisanship">negative partisanship</a> (see also <a href="https://slatestarcodex.com/2014/12/17/the-toxoplasma-of-rage/">the toxoplasma of rage</a> and &#8220;<a href="https://en.wikipedia.org/wiki/Owning_the_libs">owning the libs</a>&#8221;). In geopolitics, actions aimed at hurting enemy states (like sanctions or blockades) can spiral into military engagements, and from there into no-holds-barred wars. One way of viewing moral progress is the process of forming agreements to limit cycles of direct pessimization&#8212;both at large scales, like international conventions; and at small scales, like local norms of decency.</p><p>While few of these examples will be novel to many readers, it feels useful to bring them together under a single heading&#8212;both to make it easier to reason about preventing them, and to contrast them with my two other types of pessimization. The most notable contrast is that direct pessimization is a transitive relationship: we describe A directly pessimizing B. Conversely, perverse pessimization is a reflexive relationship: &#8220;B pessimized themself&#8221; (or &#8220;B pessimized" for short). What about when A harms B&#8217;s interests as a side effect of pursuing their own goals? Depending on how this happened, we might call it an accident, or negligence, or <a href="https://en.wikipedia.org/wiki/Manslaughter#Constructive">constructive malice</a>. However, I wouldn&#8217;t classify any of these as pessimization. Instead, I&#8217;m interested in the cases where B&#8217;s own actions make it much easier for their interests to be harmed, which I&#8217;d describe as B indirectly pessimizing themself. Let&#8217;s explore that now.</p><h3>Indirect pessimization</h3><p>Caring about X often leads you to create intellectual and physical tools for affecting X. But these can convince and/or help people with other goals to produce ~X (even when they&#8217;re not specifically trying to hurt you). I call this dynamic indirect pessimization.</p><p>The simplest example of indirect pessimization is simply telling the wrong people about your goal. To pursue a goal X you need to have some conception of what X is and how you&#8217;re going to achieve it. But telling people about X prompts them to consider whether pursuing ~X is a good way to achieve their own interests. For example, if your goal is to prevent anyone from <a href="https://www.asimov.press/p/mirror-life">creating mirror life, because it might destroy the world</a>, then telling people about your goal may be what gives them the idea of creating mirror life at all.</p><p>Indirect pessimization happens most easily when X advocates don&#8217;t have very good proposals for what people should do to help achieve X. If so, most ways that people act on their new knowledge about X will contribute to achieving ~X. For example, early AI safety advocates did a lot to raise awareness of AGI risk&#8212;but the lack of actionable strategies to help reduce risk led to much of that awareness being directed towards founding various AGI labs which are now racing towards AGI.</p><p>Furthermore, pro-X coalitions often motivate work towards X by identifying its beneficial consequences. But every argument of the form X&#8594;Y (for example, &#8220;higher education makes people more progressive&#8221;) can be translated into the form ~Y&#8594;~X (&#8220;to avoid people becoming more progressive, limit higher education&#8221;). And so, unless there&#8217;s unanimity on Y being valuable, such arguments will convince some people to oppose X, which might outweigh the benefit of gaining new supporters.&nbsp;</p><p>We often see this effect in social justice advocacy. Calling something racist is an effective strategy in moderation, but <a href="https://slatestarcodex.com/2017/06/21/against-murderism/">when overused</a> starts to make people think that racism isn&#8217;t so bad. Relatedly, <a href="https://x.com/whstancil/status/1932484678574833902?s=46">in this tweet</a> a leftist is trying to make discussion of race science more prominent, because they think that believing in it is taboo enough to discredit their political opponents. But highlighting that a person with widespread support believes in race science will help undermine the taboo in the minds of that person&#8217;s supporters. Lastly, perhaps the most egregious example is the conflation of all leftist issues into a single &#8220;omnicause&#8221;&#8212;e.g. <a href="https://www.instagram.com/p/DCW3FAZJYZP/?img_index=12">this argument</a> that &#8220;Palestine is the issue that makes us realise everything is interconnected. Every struggle for justice, freedom, and liberation&#8221;. Even if some people find this persuasive, it&#8217;s also a kind of indirect pessimization&#8212;now people who disagree with the writer on Palestine will be pushed towards disagreeing on everything else too.</p><p>Indirect pessimization doesn&#8217;t just arise from constructing concepts and arguments, but also physical tools or mechanisms. For example, people sometimes propose constructing tech for deflecting asteroids away from Earth. But given how low the rate of dangerous asteroids hitting Earth is, <a href="https://forum.effectivealtruism.org/posts/RZf2KqeMFZZEpvBHp/risks-from-asteroids">the most likely way for that to happen</a> is if asteroid-deflection technology is <a href="https://www.narrativeark.xyz/p/jacob-on-the-precipice">misused to deflect asteroids </a><em><a href="https://www.narrativeark.xyz/p/jacob-on-the-precipice">towards</a></em><a href="https://www.narrativeark.xyz/p/jacob-on-the-precipice"> Earth</a>. So naive attempts to lower asteroid risk might easily empower people to increase it instead.</p><p>Similarly, consider the apocryphal <a href="https://economicpoint.com/goodharts-law">anecdote about the cobra effect</a>. This is often cited as an illustration of Goodhart&#8217;s law. But what&#8217;s most striking about the story is that it described an intervention not just failing to <em>reduce</em> the population of wild cobras, but actually <em>increasing</em> it. In other words, these (fictional) authorities indirectly pessimized their own goal. A real-world example comes from ML, where constructing a given objective function creates the possibility of accidentally inverting it&#8212;e.g. as OpenAI did by (briefly) <a href="https://arxiv.org/pdf/1909.08593">creating maximally-inappropriate AIs</a> (section 4.4). <a href="https://www.lesswrong.com/posts/D7PumeYTDPfBTp3i7/the-waluigi-effect-mega-post">The Waluigi effect</a> and <a href="https://arxiv.org/abs/2502.17424">emergent misalignment</a> provide more recent examples of indirect pessimization in AI values. Meanwhile <a href="https://x.com/richardmcngo/status/1869089450346877207?s=46">I suspect</a> that many evaluations of dangerous AI capabilities, which were intended to help prevent them from being developed, will instead help AGI labs accelerate towards them.</p><p>Of course, when people try to achieve some goal, there are many possible undesirable side-effects. So why is it worth distinguishing specifically the category of side effects which cause the opposite of that goal? One reason is that many of the mechanisms which drive indirect pessimization are (unlike most side effects) relevant for a wide range of goals. Another is that <a href="https://www.lesswrong.com/s/qXZLFGqpD7aeEgXGL">fear-based motivation</a> makes indirect pessimization unusually hard to think about. The more scared you are that you might not achieve your goal, the more urgently you feel that &#8220;something must be done&#8221;, and the more you flinch away from picturing how that &#8220;something&#8221; might actually make it worse. This applies both on an individual level (where the possibility that you&#8217;re indirectly pessimizing is often very painful for your ego), and on a group level (where self-criticism is often taboo).</p><h3>Perverse pessimization</h3><p>The final type of pessimization I&#8217;ll talk about is <em>perverse</em> <em>pessimization</em>. I define this as a situation where the nominally pro-X faction is the one directly causing ~X to happen&#8212;i.e. the pro-X faction is sabotaging itself. What causes this? I&#8217;ll talk about two broad reasons: fear of success, and vice signalling within factions.</p><p>I&#8217;ve talked above about how fear of not achieving one&#8217;s goals can indirectly pessimize those goals. But what&#8217;s far more perverse is a situation where a coalition is instead scared of <em>achieving </em>its nominal goals. One reason why this happens: the identity, prestige and even existence of a coalition often depend on the persistence of the problem it&#8217;s trying to solve. So the prospect of the problem going away can be very scary to individual members of the coalition (especially powerful or longstanding members) or the coalition as a whole.</p><p>This applies especially when solutions are offered from outside the coalition&#8212;if those solutions work, then the coalition has to admit that its previous strategies weren&#8217;t working, which might trigger shame or envy. So it&#8217;s particularly tempting for coalitions to suppress external attempts to achieve their own nominal goals. The environmentalist movement&#8217;s opposition to nuclear power, and often even to solar or wind infrastructure, provides a good example of this (as does its gradual slide towards being a generic anti-capitalist movement).</p><p>Another source of fear of success is that the more successful you are, the more you have to lose. Having hope, then losing it, is a very painful experience. So people often decide that it&#8217;s better to proactively identify as a victim in order to avoid that risk. From that mindset, other people&#8217;s success sparks envy, resentment, and attempts to tear them down&#8212;but this toxicity only makes one&#8217;s own problems worse. For example, the men who join incel communities deeply want to find partners, but end up in a perverse position where they would lose their community if they found a partner. (Some pointers to further reading on this dynamic: <a href="https://en.wikipedia.org/wiki/Tall_poppy_syndrome">tall poppy syndrome</a>; <a href="https://en.wikipedia.org/wiki/Law_of_Jante">the laws of Jante</a>; <a href="https://articles.starcitygames.com/articles/stuck-in-the-middle-with-bruce/">this essay</a> about Magic: the Gathering; <a href="https://www.amazon.com/Existential-Kink-Unmask-Embrace-getting/dp/1578636477">Existential Kink</a>; and <a href="https://www.astralcodexten.com/p/book-review-sadly-porn">Sadly, Porn</a>.)</p><p>We also see fear of success in many relationships. The most successful relationships are those intimate enough to allow both people to feel deeply understood. But the vulnerability required for real intimacy is terrifying, and it&#8217;s easy for that fear to lead us to push friends or romantic partners away&#8212;essentially punishing them for wanting to be close to us. We hope that our defense mechanisms will ensure that we&#8217;re only vulnerable to people who deeply care about us, but they&#8217;re often precisely what prevent people from coming to deeply care about us (as masterfully portrayed in <a href="https://en.wikipedia.org/wiki/Notes_from_Underground">Notes From Underground</a>).</p><p>A second major reason why perverse pessimization arises is that anti-X behavior can help someone rise within a (nominally) pro-X coalition. This is a kind of vice signalling, showing that you have enough power to directly defy the values of your own coalition. I expect that there&#8217;s a lot of private signalling of cynicism in high-pressure law and finance firms, amongst politicians, and amongst highly competitive elites more generally. One particularly legible example comes from <a href="https://en.wikipedia.org/wiki/Gerald_Ratner">Gerald Ratner</a>, a jewelry CEO who called his own products &#8220;total crap&#8221; in an apparent attempt to signal cynicism.</p><p>These same dynamics play out within society as a whole, with domain-specific cynicism replaced by transgression against broader norms. Satanism, for instance, was historically compelling precisely because it was so transgressive. These days, few are offended by it. But <a href="https://theupheaval.substack.com/p/american-strong-gods">Hitler is now filling the role</a> of a secular Satan, which then sparks transgressions like Kanye West&#8217;s recent song &#8220;<a href="https://en.wikipedia.org/wiki/Heil_Hitler_(song)">Heil Hitler</a>&#8221;, &#8220;ironic&#8221; neo-Nazism on 4chan, etc. Meanwhile many sexual fetishes are erotic in large part because they&#8217;re taboo, with the contrast between the fear of taboo violation and the relief of sexual acceptance being a source of pleasure. Jointly acting out taboo fetishes can allow elites to form transgression bonds&#8212;Epstein&#8217;s island being perhaps the most prominent example (as well as in Hollywood, with Weinstein, Diddy, etc).</p><p>Such transgressions typically start out covert. But it&#8217;s hard to keep controversial secrets&#8212;especially because bragging about transgressions is a good way to signal dominance. So transgressions tend to slide towards becoming &#8220;open secrets&#8221;, which often creates a self-fulfilling prophecy that there&#8217;s a <a href="https://www.mindthefuture.info/p/elite-coordination-via-the-consensus">consensus of power</a> in favor of transgression. People who want to stay on the good side of that consensus learn to actively punish others for trying to enforce norms or pursue the stated goals of the coalition. <a href="https://www.umass.edu/preferen/You%20Must%20Read%20This/herrmann-thoni-gachter.pdf">Hermann et al.</a> call this phenomenon <a href="https://www.lesswrong.com/posts/X5RyaEDHNq5qutSHK/anti-social-punishment">anti-social punishment</a>; Ben Hoffman calls it <a href="https://benjaminrosshoffman.com/guilt-shame-and-depravity/">depravity</a>; Jessica Taylor calls it <a href="https://unstableontology.com/2021/04/12/on-commitments-to-anti-normativity/">anti-normativity</a>; and Michael Vassar links it to <a href="https://podcast.clearerthinking.org/episode/028/michael-vassar-preference-falsification-and-postmodernism/">postmodernism</a>. I don&#8217;t know of systematic ways to detect anti-normativity, but sufficiently egregious hypocrisy can provide strong hints that it exists. For example, one of the US military&#8217;s strongest nominal values is abiding by US laws. Yet rather than rewarding whistleblowers who expose examples of criminal behavior, the military typically persecutes them.</p><p>I&#8217;ve described fear of achieving one&#8217;s goals and anti-normativity as two separate mechanisms driving perverse pessimization. But being in an anti-normative coalition induces fear of achieving your goals, because succeeding (or even just standing out) makes you a target. And when you&#8217;re scared of achieving your goals, you&#8217;ll reward other people (or your own internal subagents) for signalling opposition to those goals, thereby reinforcing the anti-normative coalition. So we should actually think of them as two facets of a single phenomenon.</p><h3>Orienting to pessimization</h3><p>Pessimization sucks, but that doesn&#8217;t mean we should be constantly scared of it. The accusation of pessimization could easily become a bludgeon used to attack anyone doing anything valuable (which would, ironically, pessimize the desire to avoid pessimization). The challenge is to watch out for pessimization in a way which doesn&#8217;t slip into self-defeating cynicism.</p><p>So it&#8217;s also worth talking about limits on pessimization. While direct pessimization is a powerful way to cow your enemies, it also makes you a target for retaliation&#8212;e.g. violating norms against torture can spur domestic and international opposition to a regime. Meanwhile indirect and perverse pessimization are often covert enough that they only spread slowly. For example, even though some people find pedophilia attractive <em>because</em> it&#8217;s taboo, <a href="https://x.com/richardmcngo/status/1912990666952753646?s=46">I expect</a> that the taboo overall dramatically reduces pedophilia.</p><p>Perverse pessimization is also self-undermining: it makes systems weaker over time, which reduces their long-term influence. If big companies weren&#8217;t so often stuck in <a href="https://thezvi.wordpress.com/2019/05/30/quotes-from-moral-mazes/">moral mazes</a>, they could do a far better job at fending off startups. More generally, the parts of the world that avoid perverse pessimization will tend to grow and become more important over time. So <a href="https://en.wikipedia.org/wiki/Creative_destruction">creative destruction</a> is one of our best defenses against perverse pessimization.</p><p>However, letting institutions taken over by perverse pessimization fail can be arbitrarily costly&#8212;e.g. when <a href="https://dominiccummings.substack.com/p/a-talk-on-regime-change">major governments rot from the head down</a>. And so it can be incredibly high-leverage for whistleblowers and other reformers to identify and oppose perverse pessimization inside important institutions (see e.g. <a href="https://www.amazon.in/Secrets-Memoir-Vietnam-Pentagon-Papers-ebook/dp/B000OCXFY2">Ellsberg</a>, <a href="https://time.com/7012881/daniel-kokotajlo/">Kokotajlo</a>). This is bottlenecked on virtues like courage and integrity, which is a major reason why I&#8217;ve become a virtue ethicist (as I&#8217;ll write about in an upcoming post). Indeed, I suspect that one way of understanding virtues is as traits that guard us against sliding into pessimization, thereby allowing us to more robustly steer the world towards desired outcomes.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.mindthefuture.info/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Mind the Future! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Well-foundedness as an organizing principle of healthy minds and societies]]></title><description><![CDATA["If a kingdom is divided against itself, it cannot stand."]]></description><link>https://www.mindthefuture.info/p/well-foundedness-as-an-organizing</link><guid isPermaLink="false">https://www.mindthefuture.info/p/well-foundedness-as-an-organizing</guid><dc:creator><![CDATA[Richard Ngo]]></dc:creator><pubDate>Mon, 07 Apr 2025 00:28:34 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/001d57a1-b511-448b-b112-784a130c8ea2_800x800.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>&#8220;If a kingdom is divided against itself, it cannot stand. And if a house is divided against itself, it cannot be maintained.&#8221; - </em>Mark 3:24</p><p>In <a href="https://www.mindthefuture.info/p/towards-a-scale-free-theory-of-intelligent">my last post</a> I argued that we should view intelligent agents as coalitions of cooperating and competing subagents. The crucial question is then: how can we characterize effectively-functioning coalitional agents? One standard criterion is <em>coherence</em>: the extent to which the agent acts as if it has consistent goals and beliefs. In other words, an agent is coherent insofar as disagreeing subagents are able to still act as a functional coalition. For example:</p><ul><li><p>A coherent individual can decide how to resolve an internal conflict and then stick with that commitment, rather than vacillating or procrastinating as their mood changes.</p></li><li><p>A coherent company is able to decide on an overarching strategic plan and then execute it, without different divisions prioritizing their own interests.</p></li><li><p>A coherent country has a clear set of national interests that its leaders and people consistently prioritize (even when it conflicts with their personal interests).</p></li></ul><p>Coherence is a valuable property for a coalition to have, but I think that characterizing idealized agents primarily in terms of coherence gives us an impoverished understanding of them. For example, a coalition which is highly-coherent because its leaders exercise a lot of top-down control is much less robust than a coalition which is highly-coherent because all its subagents actually want to cooperate with each other.</p><p>In this post I try to capture the difference between these two possibilities in terms of a property I call <em>well-foundedness</em>.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> I define an agent as well-founded to the extent that conflicts between its subagents don&#8217;t propagate <em>down</em> to induce conflicts <em>within</em> those subagents. For example:</p><ul><li><p>In a well-founded marriage, spouses don&#8217;t try to induce internal conflict within their partner (e.g. shaming or guilting them) to win fights.</p></li><li><p>In a well-founded company, employees don&#8217;t need to pick sides during power struggles between executives&#8212;they can just carry on with their jobs.</p></li><li><p>In a well-founded country, people are friends and colleagues despite supporting different political factions.</p></li></ul><p>I think of well-foundedness as complementary to (and in some ways dual to) coherence. Ideal agents should have both properties. I don&#8217;t yet know how to define well-foundedness precisely, but in the rest of this post I characterize it informally by describing the four possible combinations of coherent/incoherent and well-founded/poorly-founded.</p><h3>Incoherence with or without well-foundedness</h3><p>Consider two countries each experiencing strong internal political polarization, but in very different ways. In Holistan, the two factions are the one representing East Holistan and the one representing West Holistan. By contrast, in Fractistan the two factions represent two subpopulations who live closely intermingled throughout the country&#8212;say, two religious or ethnic groups.</p><p>For each country, increasing internal tensions makes them less coherent&#8212;each half of the country thinks of the other half as their enemy. But the two conflicts will play out very differently. In Holistan, you might see the East and the West start to cut ties with each other; set up parallel governance structures; or discourage travel or trade between them. If conflict continues to escalate, you might see a civil war, with each side drawing on their territory and population to muster an army.</p><p>This is pretty bad! But despite being very incoherent as a country, Holistan is still relatively well-founded, because each of East and West Holistan are <em>internally</em> still pretty coherent. What that means is that they have a line of retreat from conflict. The two sides in Holistan can still disengage; they can draw up peace treaties; they can form two separate countries. After the war ends, the fabric of each society remains intact&#8212;colleagues are still on amicable terms, neighbors still trust neighbors, city councils can still debate issues without relitigating the war with each conflict.</p><p>Perhaps the most vivid illustration of how important this is comes from World War 2. I am often struck by how, after the most devastating war the world has ever seen, European countries quickly recovered to unprecedented heights of prosperity, and even became close allies. I think a significant reason for that is that World War 2 was a relatively well-founded conflict between nation-states. It didn&#8217;t turn them into low-trust societies.</p><p>By contrast, Fractistan is neither coherent nor well-founded. The lack of clear territorial boundaries between the two factions makes it harder for rising tensions to erupt into a full-scale civil war. But the whole country is affected regardless: each region and city and neighborhood faces an internal power struggle. You might see lynch mobs or pogroms; or, in an extreme case, the kind of decentralized genocide that happened in Rwanda.</p><p>And there&#8217;s no easy way to end the conflict. Even if a nation-wide compromise is reached, each person will still be surrounded by former enemies. Each small-scale flare-up of renewed conflict will trigger further cycles of escalation. In other words, Fractistan will persist as a country, but with low social trust for the indefinite future. I picture it with a fracture running from top to bottom, fractally dividing the country in two at every level of organization.</p><p>Aside from Rwanda, two countries with Fractistan-like conflicts were <a href="https://en.wikipedia.org/wiki/Bosnian_War">Bosnia and Herzegovina</a> and pre-partition India, which both saw widespread neighbor-on-neighbor violence. India was lucky to have a strong geographic separation between the bulk of its Hindus and Muslims. Even so, however, the process of splitting India and Pakistan was incredibly messy and cost hundreds of thousands of lives.</p><h3>The tradeoff between coherence and well-foundedness</h3><p>Coherence and well-foundedness are separate properties which are both individually valuable. But there are some tradeoffs between them. To explore those it&#8217;s useful to consider the opposite of Holistan: a country which is very coherent but also very poorly-founded.</p><p>I could make up an example here, but we already have one that matches the description very well: North Korea. As a country, it&#8217;s extremely coherent. Its whole government follows the instructions of one man. Its whole society follows the instructions of its government. There&#8217;s no national-level dissent, nor regional-level dissent, nor even dissent on the level of local communities.</p><p>But repressing dissent doesn&#8217;t make it go away&#8212;it just pushes it down to lower-level subagents. I expect that some North Koreans feel safe to dissent within their families; others only within the privacy of their own thoughts; and others not even there, but only in their <a href="https://en.wikipedia.org/wiki/Shadow_(psychology)">subconscious shadows</a>. Even when this dissent never surfaces openly, it is visible in the cost and scale of the control apparatus required to keep it repressed. If that control apparatus ever breaks, North Korea&#8217;s coherence would fall apart.</p><p>Such extreme repression is obviously bad. But in moderation repression can be a valuable driver of coherence. <a href="https://eh.net/book_reviews/the-weirdest-people-in-the-world-how-the-west-became-psychologically-peculiar-and-particularly-prosperous">Henrich hypothesizes</a> that the success of the West was driven by the Catholic Church&#8217;s prohibition against cousin marriage, which made Christian Europe less clannish. If true, you can think of this as trading off well-foundedness for coherence: by repressing kinship-based networks, the Church made larger-scale cooperation possible. More generally, societal morality works by repressing the antisocial instincts of each individual (as well as groups organized around antisocial behavior, like criminal gangs).</p><p>A similar set of tradeoffs arise in individual psychology: people can become more disciplined by repressing their emotions. This makes them more coherent&#8212;e.g. they can choose to work long hours on things that are instrumentally useful. But it often harms their ability to enjoy themselves and to understand and process their underlying motivations.</p><p>Conversely, to become well-founded, you need to surface ways in which conflicts manifest at low levels and then resolve them. This requires the opposite of repression: <em>expression</em>. Specifically, it requires that lower-level subagents are able to express their true preferences (as they can in individuals who freely let their emotions surface, or countries which let political dissidents speak freely).</p><p>What are the tradeoffs between building coherence via repression, and building well-foundedness via expression? I think of the former as making the average-case outcome worse, but also reducing variance. Repression is therefore appropriate when you&#8217;re in a <em>scarce</em> environment&#8212;one in which a big loss could totally wipe you out. An individual whose career could be ruined if they let their emotions show needs to repress them (even if it makes them more stressed and less productive on average). And a country which could be invaded if it gets distracted by internal politics needs to repress dissent (even if there&#8217;s something valuable to be learned from that dissent).</p><p>By contrast, expression tends to lead to better outcomes, but at the cost of also increasing variance. <a href="https://www.lesswrong.com/s/qXZLFGqpD7aeEgXGL/p/QqtQoHSjjfLgC4jDZ">Expressing underlying conflicts allows them to be solved directly</a>, but makes the overall agent less coherent until things actually resolve. For example, instead of sniping at each other about household chores, a married couple could express the emotional fears that underlie those frustrations. If that goes well, they&#8217;d feel much more respected and appreciated afterwards; but if it goes badly it could provoke a (potentially relationship-ending) fight. In a political context, letting dissidents speak out could lead to valuable reform, but it could also give rise to a full-fledged separatist movement.</p><p>So the tradeoff between repression and expression is a nuanced one, and needs to be decided on a case-by-case basis. But in general it&#8217;s likely that we err too far on the side of repression, because we don&#8217;t intuitively realize how abundant the world has become. Most countries don&#8217;t face significant risk of invasion, and therefore the costs of secession would often be outweighed by the benefits (more national cohesion, <a href="https://www.sciencedirect.com/science/article/abs/pii/S0014292110000991">better governance</a>, and being more well-founded in general). On an individual level, we can now move between different communities far more easily than at any previous point in history, and so taking the social risk of letting out our emotions is less dangerous than our intuitions are calibrated to expect. (There are some prominent exceptions, which I&#8217;ll discuss in a follow-up post, but I&#8217;m trying to keep this post relatively free of politically controversial examples.)</p><h3>Idealized coalitional agency</h3><p>I think the tradeoff I&#8217;ve described above is an important dynamic in almost all real-world agents. But I don&#8217;t think it&#8217;s <em>inevitable</em>. We can imagine coalitions designed so that low-level agents can express their preferences, and make local improvements, without threatening the stability of the overall coalition.</p><p>What would such designs look like? I expect that a key component is &#8220;peace treaties&#8221; between high-level subagents, where they all agree not to use certain types of low-level conflicts to further their own ends. We can think of liberalism as a peace treaty which allows people with different religious and political beliefs to coexist. Meanwhile, arrangements like capitalism and democracy channel conflicts into formats that are productive (like business competition and political campaigning) rather than destructive (like theft or violence).</p><p>But well-foundedness at one level requires coherence at levels below that&#8212;otherwise it&#8217;s easy for conflicts to propagate downwards. And the systems I describe above aren&#8217;t very good at creating (or maintaining) lower-level coherence. In <a href="https://theupheaval.substack.com/p/american-strong-gods">Reno&#8217;s terminology</a>, they are &#8220;weak gods&#8221; whose primary purpose is to help different groups coexist, but which don&#8217;t have strong opinions about what those groups should actually care about. Conversely, &#8220;strong gods&#8221; like nationalism provide the substantive ideological content capable of unifying people under a single coherent identity, at the cost of excluding outsiders. The challenge of designing an idealized coalitional agent can be seen as the unsolved problem of balancing weak and strong gods across many different scales.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.mindthefuture.info/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.mindthefuture.info/subscribe?"><span>Subscribe now</span></a></p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>A <a href="https://en.wikipedia.org/wiki/Well-founded_relation">well-founded set</a> is one which has a bottom element. By somewhat tenuous analogy, a well-founded agent is one whose internal conflicts bottom out. Unlike the set theory definition, my notion of well-foundedness is a matter of degree: the further down an agent&#8217;s internal conflicts go, the less well-founded it is.</p></div></div>]]></content:encoded></item><item><title><![CDATA[Towards a scale-free theory of intelligent agency]]></title><description><![CDATA[An inherently multi-agent perspective]]></description><link>https://www.mindthefuture.info/p/towards-a-scale-free-theory-of-intelligent</link><guid isPermaLink="false">https://www.mindthefuture.info/p/towards-a-scale-free-theory-of-intelligent</guid><dc:creator><![CDATA[Richard Ngo]]></dc:creator><pubDate>Sat, 22 Mar 2025 05:01:31 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/0b965984-3c8e-48fd-93bf-beab290989d1_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I recently left OpenAI to pursue independent research. I&#8217;m working on a number of different research directions, but the most fundamental is my pursuit of a scale-free theory of intelligent agency. In this post I give a rough sketch of how I&#8217;m thinking about that. I&#8217;m erring on the side of sharing half-formed ideas, so there may well be parts that don&#8217;t make sense yet. Nevertheless, I think this broad research direction is very promising.</p><p>This post has two sections. The first describes what I mean by a theory of intelligent agency, and some problems with existing (non-scale-free) attempts. The second outlines my current path towards formulating a scale-free theory of intelligent agency, which I&#8217;m calling <em>coalitional agency</em>.</p><h2>Theories of intelligent agency</h2><p>By a &#8220;theory of intelligent agency&#8221; I mean a unified mathematical framework that describes both <em>understanding the world</em> and <em>influencing the world</em>. In this section I&#8217;ll outline the two best candidate theories of intelligent agency that we currently have (expected utility maximization and active inference), explain why neither of them is fully satisfactory, and outline how we might do better.</p><h3>Expected utility maximization</h3><p>Expected utility maximization is the received view of intelligent agency in many fields (I&#8217;ll abbreviate it as EUM, and EUM agents as EUMs). Idealized EUMs have beliefs in the form of probability distributions, and goals in the form of utility functions, as specified by the axioms of probability theory and utility theory. They choose whichever strategy leads to the most utility in expectation; this is typically modelled as a process of search or planning.</p><p>EUM is a very productive framework in simple settings&#8212;like game theory, bargaining theory, microeconomics, etc. It&#8217;s particularly useful for describing agents making one-off decisions between a fixed set of choices. However, it&#8217;s much more difficult to use EUM to model agents making sequences of choices over time, especially when they learn and update their concepts throughout that process. The two points I want to highlight here:</p><ol><li><p>EUM treats goals and beliefs as totally separate. But in practice, agents represent both of these in terms of the same underlying concepts. When those concepts change, both beliefs and goals change. The best way to learn reusable concepts is via <em>deep learning</em>, where simple concepts in lower layers of the network are built up into more complex concepts in higher layers.</p></li><li><p>Agents often act based on heuristics learned via trial and error. Even if they&#8217;ll eventually acquire beliefs about how those heuristics help them achieve their goals, they often start off without any clear idea of why those heuristics work so well. But this makes them hard to model as idealized EUMs which choose actions based on their beliefs and goals. Instead, the process of learning heuristics is better understood as (model-free) reinforcement learning.</p></li></ol><p>So we might hope that a theory of deep learning, or reinforcement learning, or deep reinforcement learning, will help fill in EUM&#8217;s blind spots. Unfortunately, theoretical progress has been slow on all of these&#8212;they&#8217;re just too broad to say meaningful things about in the general case.</p><h3>Active inference</h3><p>Fortunately, there&#8217;s another promising theory which comes at it from a totally different angle. <a href="https://mitpress.mit.edu/9780262553995/active-inference/">Active inference</a> is a theory born out of neuroscience. Where EUM starts by assuming an agent already has beliefs and goals, active inference gives us a theory of how beliefs and goals are built up over time.</p><p>The core idea underlying active inference is <em>predictive coding</em>. Predictive coding models our brains as hierarchical networks where the lowest level is trying to predict our sensory inputs, the next-lowest level is trying to predict the lowest level, and so on. The higher up the hierarchy you go, the more abstract and compressed the representations become. The lower levels might represent individual &#8220;pixels&#8221; seen by our retinas, then higher levels lines and shapes, then higher levels physical objects like dogs and cats, then even higher levels abstract concepts like animals and life.</p><p>This is, of course, similar to <a href="https://distill.pub/2020/circuits/zoom-in/">how artificial neural networks work</a> (especially ones trained by self-supervised learning). The key difference: predictive coding tells us that, in the brain, the patterns recognized at each level are determined by reconciling the bottom-up signals and the top-down predictions. For example, after looking at the image below, you might not perceive any meaningful shapes within it. But if you have a strong enough top-down prediction that the image makes sense (e.g. because I&#8217;m telling you it does) then that prediction will keep being sent down to lower layers responsible for identifying shapes, until they discover the dog. This explains the sharp shifts in our perceptions when looking at such images: at first we can&#8217;t see the dog at all, but when we find it it jumps into focus, and afterwards we can&#8217;t unsee it.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!3_Il!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa465c6cd-630f-467b-a2c7-603858cd7baf_1500x999.webp" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!3_Il!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa465c6cd-630f-467b-a2c7-603858cd7baf_1500x999.webp 424w, https://substackcdn.com/image/fetch/$s_!3_Il!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa465c6cd-630f-467b-a2c7-603858cd7baf_1500x999.webp 848w, https://substackcdn.com/image/fetch/$s_!3_Il!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa465c6cd-630f-467b-a2c7-603858cd7baf_1500x999.webp 1272w, https://substackcdn.com/image/fetch/$s_!3_Il!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa465c6cd-630f-467b-a2c7-603858cd7baf_1500x999.webp 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!3_Il!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa465c6cd-630f-467b-a2c7-603858cd7baf_1500x999.webp" width="1456" height="970" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a465c6cd-630f-467b-a2c7-603858cd7baf_1500x999.webp&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:970,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:95734,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/webp&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.mindthefuture.info/i/159597187?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa465c6cd-630f-467b-a2c7-603858cd7baf_1500x999.webp&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!3_Il!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa465c6cd-630f-467b-a2c7-603858cd7baf_1500x999.webp 424w, https://substackcdn.com/image/fetch/$s_!3_Il!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa465c6cd-630f-467b-a2c7-603858cd7baf_1500x999.webp 848w, https://substackcdn.com/image/fetch/$s_!3_Il!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa465c6cd-630f-467b-a2c7-603858cd7baf_1500x999.webp 1272w, https://substackcdn.com/image/fetch/$s_!3_Il!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa465c6cd-630f-467b-a2c7-603858cd7baf_1500x999.webp 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Predictive coding is a very elegant theory. And what&#8217;s even more elegant is that it explains <em>actions</em> in the same way&#8212;as very strong top-down predictions which override the default states of our motor neurons. Specifically, we can resolve conflicts between beliefs and observations either by updating our beliefs, or by taking actions which make the beliefs come true. Active inference is an extension of predictive coding in which some beliefs are so rigid that, when they conflict with observations, it&#8217;s easier to act to change future observations than it is to update those beliefs. We can call these hard-to-change beliefs &#8220;goals&#8221;, thereby unifying beliefs and goals in a way that EUM doesn&#8217;t.</p><p>This is a powerful and subtle point, and one which is often misunderstood. I think the best way to fully understand this point is in terms of perceptual control theory. <a href="https://slatestarcodex.com/2019/03/20/translating-predictive-coding-into-perceptual-control/">Scott Alexander gives a good overview here</a>; I&#8217;ll also explain the connection at more length in a follow-up post.</p><h3>Towards a scale-free unification</h3><p>Active inference is a beautiful theory&#8212;not least because it includes EUM as a special case. Active inference represents goals as probability distributions over possible outcomes. If we interpret the logarithm of each probability as that outcome&#8217;s utility (and set aside the value of information) then active inference agents <a href="https://x.com/RichardMCNgo/status/1893556269644161215">choose actions which maximize expected utility</a>. (One intuition for why such an interpretation is natural comes from Scott Garrabrant's <a href="https://www.lesswrong.com/s/4hmf7rdfuXDJkxhfg/p/DMxe4XKXnjyMEAAGw">geometric rationality</a>.)</p><p>So what does expected utility maximization have to add to active inference? I think that what active inference is missing is the ability to model <em>strategic interactions</em> between different goals. That is: we know how to talk about EUMs playing games against each other, bargaining against each other, etc. But, based on my (admittedly incomplete) understanding of active inference, we don&#8217;t yet know how to talk about goals doing so <em>within</em> a single active inference agent.</p><p>Why does that matter? One reason: the biggest obstacle to a goal being achieved is often other conflicting goals. So any goal capable of learning from experience will naturally develop strategies for avoiding or winning conflicts with other goals&#8212;which, indeed, seems to happen <a href="https://www.lesswrong.com/s/ZbmRyDN8TCpBTZSip/p/5gfqG3Xcopscta3st">in human minds</a>.</p><p>More generally, any theory of intelligent agency needs to model internal conflict in order to be scale-free. By a scale-free theory I mean one which applies at many different levels of abstraction, remaining true even when you &#8220;zoom in&#8221; or &#8220;zoom out&#8221;. I see so many similarities in how intelligent agency works at different scales (on the level of human subagents, human individuals, companies, countries, civilizations, etc) that I strongly expect our eventual theory of it to be scale-free.</p><p>But active inference agents are cooperative <em>within </em>themselves while having strategic interactions with other agents; this privileges one level of analysis over all the others. Instead, I propose, we should think of active inference agents as being composed of subagents who themselves compete and cooperate in game-theoretic ways. I call this approach <em>coalitional agency; </em>in the next section I characterize my current understanding of it from two different directions.</p><h2>Two paths towards a theory of coalitional agency</h2><p>The core idea of coalitional agency is that we should think of agents as being composed of cooperating and competing subagents; and those subagents as being composed of subsubagents in turn; and so on. The broad idea here is not new&#8212;indeed, it&#8217;s the core premise of Minsky&#8217;s <em>Society of Mind</em>, published back in 1986. But I hope that thinking of coalitional agency as incorporating elements of both EUM and active inference will allow progress towards a formal version of the theory.</p><p>In this section I&#8217;ll give two different characterizations of coalitional agency: one starting from EUM and trying to make it more coalitional, and the other starting from active inference and trying to make it more agentic. More specifically, the first poses the question: if a group of EUMs formed a coalition, what would it look like? The second poses the question: how could active inference agents be more robust to conflict between their internal subagents?</p><h3>From EUM to coalitional agency</h3><p>If a group of EUMs formed a coalition, what would it look like? EUM has a standard answer to this: the coalition would be a linearly-aggregated EUM. In this section I first explain why the standard answer is unsatisfactory. I then give an alternative answer: the coalition should be an incentive-compatible decision procedure<em>.</em></p><h4>Aggregating into EUMs is very inflexible</h4><p>In the EUM framework, any non-EUM agent is incoherent in the sense of violating the underlying axioms of probability theory and/or utility theory. So insofar as EUM has predictive power, it predicts that competent coalitions will also be EUMs. But which EUMs? The standard answer is given by <a href="https://forum.effectivealtruism.org/posts/v89xwH3ouymNmc8hi/harsanyi-s-simple-proof-of-utilitarianism">Harsanyi&#8217;s utilitarian theorem</a>, which shows that (under reasonable-seeming assumptions) an aggregation of EUMs into a larger-scale EUM must have a utility function that&#8217;s a weighted average of the subagents&#8217; utilities.</p><p>However, this strongly limits the space of possible aggregated agents. Imagine two EUMs, Alice and Bob, whose utilities are each linear in how much cake they have. Suppose they&#8217;re trying to form a new EUM whose utility function is a weighted average of their utility functions. Then they&#8217;d only have three options:</p><ol><li><p>Form an EUM which would give Alice all the cakes (because it weights Alice&#8217;s utility higher than Bob&#8217;s)</p></li><li><p>Form an EUM which would give Bob all the cakes (because it weights Bob&#8217;s utility higher than Alice&#8217;s)</p></li><li><p>Form an EUM which is totally indifferent about the cake allocation between them (which would allocate cakes arbitrarily, and could be swayed by the tiniest incentive to give all Alice&#8217;s cakes to Bob, or vice versa)</p></li></ol><p>These are all very unsatisfactory. Bob wouldn&#8217;t want #1, Alice wouldn&#8217;t want #2, and #3 is extremely non-robust. Alice and Bob could toss a coin to decide between options #1 and #2, but then they wouldn&#8217;t be acting as an EUM (since EUMs can&#8217;t prefer a probabilistic mixture of two options to either option individually). And even if they do, whoever loses the coin toss will have a strong incentive to renege on the deal.</p><p>We could see these issues merely as the type of frictions that plague any idealized theory. But we could also seem them as hints about what EUM is getting wrong on a more fundamental level. Intuitively speaking, the problem here is that there&#8217;s no mechanism for separately respecting the interests of Alice and Bob after they&#8217;ve aggregated into a single agent. For example, they might want the EUM they form to value fairness between their two original sets of interests. But adding this new value is not possible if they&#8217;re limited to (a probability distribution over) weighted averages of their utilities. This makes aggregation very risky when Alice and Bob can&#8217;t consider all possibilities in advance (i.e. in all realistic settings).</p><p>Based on similar reasoning, Scott Garrabrant <a href="https://www.lesswrong.com/s/4hmf7rdfuXDJkxhfg/p/Xht9swezkGZLAxBrd">rejects the independence axiom</a>. He argues that the axiom is unjustified because rational agents should be able to lock in values like fairness based on prior agreements (or even <em>hypothetical</em> agreements).</p><h4>Coalitional agents are incentive-compatible decision procedures</h4><p>The space of decision procedures is very broad; can we say more about which decision procedures rational agents should commit to following? One key desideratum for commitments is that it&#8217;s easy to trust that they&#8217;ll be kept. Consider the example above of flipping a coin to decide between options #1 and 2 above. This is fair, but it sets up strong incentives for whoever loses the coinflip to break their commitment, since they will not get any benefit from keeping it.</p><p>And it&#8217;s even worse than that, because in general the only way to find out another agent&#8217;s utilities is to ask them, <em>and they could just lie</em>. From the god&#8217;s-eye perspective you can build an EUM which averages subagents&#8217; utilities; from the perspective of the agents themselves, you can&#8217;t. In other words, EUMs constructed by taking a weighted average of subagents&#8217; utilities are not <em><a href="https://en.wikipedia.org/wiki/Incentive_compatibility">incentive-compatible</a></em>.</p><p>EUMs which can't guarantee each other's honesty will therefore want to aggregate into incentive-compatible decision procedures which each agent does best by following. Perhaps the best-known incentive-compatible decision procedure is the <em>fair cake-cutting algorithm</em>, also known as &#8220;I cut you choose&#8221;. This is a much simpler and more elegant way to split cakes than the example I gave above of Alice and Bob aggregating into a single EUM.</p><p>Now, cake-cutting is one very specific type of problem, and we shouldn&#8217;t expect there to be incentive-compatible decision procedures with such nice properties for all problems. Nevertheless, there&#8217;s a very wide range of possibilities to explore. Some of the simplest possible incentive-compatible decision procedures include:</p><ol><li><p>Each subagent gets one type of decision that they&#8217;re solely responsible for.</p></li><li><p>We randomize between different subagents for different decisions.</p></li><li><p>Subagents vote between two options.</p></li><li><p>One subagent chooses which other subagent to delegate to.</p></li><li><p>Different mechanisms are used for different types of decisions.</p></li></ol><p>These decision procedures each give subagents some type of control over the outputs&#8212;and, importantly, a type of control that generalizes to a range of problems beyond the ones they were able to consider during bargaining.</p><h4>Which incentive-compatible decision procedure?</h4><p>The question is then: how should subagents choose <em>which </em>incentive-compatible bargaining procedure to adopt? The most principled answer is that they should use a <em>bargaining theory </em>framework. This is a little different from the traditional theoretical framework for bargaining. Bargaining doesn&#8217;t typically produce ways of organizing the bargainers&#8212;instead it produces an object-level answer to whatever problem the bargainers face.</p><p>This makes sense when you have a single decision to make. But when bargainers face many possible future decisions, bargaining over outcomes requires specifying which outcome to choose in every possible situation. This is deeply intractable in realistic settings, where bargainers can&#8217;t predict every possible scenario they might face.</p><p>In those settings it is much more tractable to bargain over <em>methods </em>of making decisions which generalize beyond the problems that the bargainers are currently aware of. I don&#8217;t know of much work on this, but the same idealized bargaining solutions (e.g. the Nash bargaining solution) should still apply in principle. The big question is whether there&#8217;s anything interesting to be said about the relationship between incentive-compatible decision procedures and bargaining solutions. For example, are there classes of incentive-compatible decision procedures which make it especially easy for agents to identify which one is near the optimal bargaining solution? On a more theoretical level, one tantalizing hint is that the <a href="https://www.lesswrong.com/posts/vJ7ggyjuP4u2yHNcP/threat-resistant-bargaining-megapost-introducing-the-rose">ROSE bargaining solution</a> is also constructed by abandoning the axiom of independence&#8212;just as Garrabrant does in his rejection of EUM above. This connection seems worth exploring further.</p><p>To finish, I&#8217;ve summarized many of the claims from this section in the following table:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!VDx5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc19af86c-6959-493c-bed5-cbeb2a7899a6_1118x516.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!VDx5!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc19af86c-6959-493c-bed5-cbeb2a7899a6_1118x516.png 424w, https://substackcdn.com/image/fetch/$s_!VDx5!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc19af86c-6959-493c-bed5-cbeb2a7899a6_1118x516.png 848w, https://substackcdn.com/image/fetch/$s_!VDx5!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc19af86c-6959-493c-bed5-cbeb2a7899a6_1118x516.png 1272w, https://substackcdn.com/image/fetch/$s_!VDx5!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc19af86c-6959-493c-bed5-cbeb2a7899a6_1118x516.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!VDx5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc19af86c-6959-493c-bed5-cbeb2a7899a6_1118x516.png" width="1118" height="516" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c19af86c-6959-493c-bed5-cbeb2a7899a6_1118x516.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:516,&quot;width&quot;:1118,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:100439,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.mindthefuture.info/i/159597187?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc19af86c-6959-493c-bed5-cbeb2a7899a6_1118x516.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!VDx5!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc19af86c-6959-493c-bed5-cbeb2a7899a6_1118x516.png 424w, https://substackcdn.com/image/fetch/$s_!VDx5!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc19af86c-6959-493c-bed5-cbeb2a7899a6_1118x516.png 848w, https://substackcdn.com/image/fetch/$s_!VDx5!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc19af86c-6959-493c-bed5-cbeb2a7899a6_1118x516.png 1272w, https://substackcdn.com/image/fetch/$s_!VDx5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc19af86c-6959-493c-bed5-cbeb2a7899a6_1118x516.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>What do I mean by &#8220;hard to reason about?&#8221; One nice thing about EUMs is that their behavior is extremely easy to summarize: they do whatever&#8217;s best for their goals according to their beliefs. But we can&#8217;t talk about decision procedures in the same way. Individual subagents may have goals and beliefs, but the decision procedure itself doesn&#8217;t: it just processes those subagents into a final decision.</p><p>Fortunately, there&#8217;s a way to rescue our intuitive idea that agents should have beliefs and goals. It&#8217;ll involve talking about much more complex incentive-compatible decision procedures, though. So first I&#8217;ll turn to the other direction in which we can try to derive coalitional agency: starting from active inference.</p><h3>From active inference to coalitional agency</h3><p>I just gave an account of coalitional agents in which they&#8217;re built up from individual EUMs. In this section I&#8217;ll do the opposite: start from an active inference agent and modify it until it looks more like a coalitional agent.</p><p>More specifically, consider a hierarchical generative model containing beliefs/goals, where higher layers predict lower layers, and lower layers send prediction errors up to higher layers. Let&#8217;s define a subagent as a roughly-internally-consistent cluster of beliefs and goals within that larger agent. Note that this definition is a matter of degree: if we apply a high bar for internal consistency, then each subagent will be small (e.g. beliefs and desires about a single object) whereas a lower bar will lead to larger subagents (e.g. a whole ideology).</p><p>Subagents with different beliefs and goals will tend to make different predictions (including &#8220;predictions&#8221; about which actions they want the agent to take). What modifications do we need to make to our original setup for it to be robust to strategic dynamics between those subagents?</p><h4>Predicting observations via prediction markets</h4><p>When multiple subagents make conflicting predictions, the standard approach is to combine them by taking a <a href="https://www.frontiersin.org/journals/human-neuroscience/articles/10.3389/fnhum.2014.00457/full">precision-weighted average</a>. Credit is then assigned to each subagent for the prediction in proportion to how confident it was. But this is not incentive-compatible: subagents can benefit by strategizing about how the other subagents will respond, and changing their responses accordingly.</p><p>There are various incentive-compatible ways to elicit predictions from multiple agents (many of which are discussed by <a href="https://arxiv.org/pdf/2403.07949">Neyman</a>). However, the most elegant incentive-compatible method for aggregating predictions is a <em>prediction market</em>. Each trader on a prediction market can choose to buy shares in propositions it thinks are overpriced and sell shares in propositions it thinks are underpriced. This allows subagents to specialize into different niches within the overall agent. It also incentivizes them to arbitrage away any logical inconsistency they notice. These dynamics are modeled by the <a href="https://intelligence.org/files/LogicalInduction.pdf">Garrabrant induction framework</a>.</p><h4>Choosing actions via auctions</h4><p>Given my discussion above about actions being in some sense predictions of future behavior, we might think that actions should be chosen by prediction markets too. However, there&#8217;s a key asymmetry: if I <em>expect</em> a complex plan to happen, I can profit by predicting <em>any</em> aspect of it. But if I <em>want</em> a complex plan to happen, I need to successfully coordinate <em>every</em> aspect of it. So, unlike predictions of observations, predictions of actions need to have some mechanism for giving a single plan control over many different actuators.</p><p>In active inference, the mechanism by which this occurs is called expected free energy minimization. I&#8217;m honestly pretty confused about how expected free energy minimization works, but I strongly suspect that it&#8217;s not incentive-compatible. In particular, the discontinuity involved in picking the single highest-value plan seems like it&#8217;d induce incentives to overestimate your own plan&#8217;s value. However, <a href="https://www.andrew.cmu.edu/user/coesterh/RIA.pdf">Demski et al.</a>&#8217;s BRIA framework solves this problem by requiring subagents to bid for the right to implement a plan and receive the corresponding reward. Rational subagents will never bid more than the reward they actually expect. So my hunch is that something like this auction system would be the best way to adjust our original setup to make it incentive-compatible.</p><h4>Aggregating values via voting</h4><p>The last important component of decision-making is evaluating plans (whether in advance or in hindsight). What happens when different subagents disagree on which goals or values the plans should be evaluated in terms of? Again, the standard approach is to take a precision-weighted average of their evaluations, but this still has all the same incentive-compatibility issues. And unlike predictions, values have no ground truth feedback signal, meaning that prediction markets don&#8217;t help.</p><p>So I expect that the most appropriate way to aggregate goals/values is via a voting system. This is also the conclusion reached by <a href="https://www.fhi.ox.ac.uk/wp-content/uploads/2021/06/Parliamentary-Approach-to-Moral-Uncertainty.pdf">Newberry and Ord</a>, who model idealized moral decision-making in terms of a parliament in which subagents vote on what values to pursue. Specifically, they propose using <a href="https://en.wikipedia.org/wiki/Random_ballot">random ballot voting</a>, in which each voter&#8217;s favorite option is selected with probability proportional to their vote share. This voting algorithm has three particularly notable features:</p><ol><li><p>As a nondeterministic voting system, it dodges <a href="https://en.wikipedia.org/wiki/Arrow%27s_impossibility_theorem">Arrow&#8217;s impossibility theorem</a>.</p></li><li><p>It leads to coalitional bargaining between subagents. Pairs of subagents with different preferences are incentivized to switch both their votes to a compromise option. Iterating this process many times produces hierarchical coalitions. (The downside: when bargaining is allowed, random ballot voting isn&#8217;t threat-resistant, since agents are incentivised to start with extreme positions in order to receive more concessions.)</p></li><li><p>It outputs a probability distribution over outcomes. We therefore have a very natural way to use votes to evaluate plans: we can measure the divergence between the outcome of the vote and the (expected or actual) outcome of the plan.</p></li></ol><h2>Putting it all together</h2><p>I&#8217;ve described two paths towards a theory of coalitional agency. On one path, we start from expected utility maximizers and aggregate them to form coalitional agents, via those EUMs bargaining about which decision procedures to use. The problem is that the resulting decision procedure may be incoherent in the sense that it can&#8217;t be ascribed beliefs or goals. On the other path, we make interactions between active inference subagents more incentive-compatible by using prediction markets, auctions, and voting (or similar mechanisms) to manage internal conflict.</p><p>What I&#8217;ll call the <em><strong>coalitional agency hypothesis</strong></em> is the idea that these two paths naturally &#8220;meet in the middle&#8221;&#8212;specifically, that EUMs doing (idealized) bargaining about which decision procedure to use would in many cases converge to something like my modified active inference procedure. If true, we&#8217;d then be able to talk about that procedure&#8217;s &#8220;beliefs&#8221; (the prices of its prediction market) and &#8220;goals&#8221; (the output of its voting procedure).</p><p>One line of work which supports the coalitional agency hypothesis is Critch&#8217;s <em><a href="https://arxiv.org/abs/1701.01302">negotiable reinforcement learning</a></em> framework, under which EUMs should bet their influence on any disagreements about the future they have with other agents, so that they end up very powerful if (and only if) their predictions are right. I interpret this result as evidence that (some version of) prediction markets are the default outcome of bargaining over incentive-compatible decision procedures.</p><p>But all of this work is still vague and tentative. I&#8217;d very much like to develop a more rigorous formulation of coalitional agency. This would benefit greatly from working with collaborators (especially those with strong mathematical skills). So I&#8217;ll finish with two calls to action. If you&#8217;re a junior(ish) researcher and you want to work with me on any of this, apply to <a href="https://www.matsprogram.org/apply">my MATS fellowship</a>. If you&#8217;re an experienced researcher and you&#8217;d like to chat or otherwise get involved (potentially by joining a workshop series I&#8217;ll be running on this) please send me a message directly.</p><p></p><p><em>Thanks to davidad, Jan Kulveit, Emmett Shear, Ivan Vendrov, Scott Garrabrant, Abram Demski, Martin Soto, Laura Deming, Aaron Tucker, Adria Garriga, Oliver Richardson, Madeleine Song and others for helping me formulate these ideas</em>.</p>]]></content:encoded></item><item><title><![CDATA[Elite Coordination via the Consensus of Power]]></title><description><![CDATA[Solving the mystery of the cathedral]]></description><link>https://www.mindthefuture.info/p/elite-coordination-via-the-consensus</link><guid isPermaLink="false">https://www.mindthefuture.info/p/elite-coordination-via-the-consensus</guid><dc:creator><![CDATA[Richard Ngo]]></dc:creator><pubDate>Wed, 19 Mar 2025 06:51:39 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/7c3a1f2e-f284-46dd-aaed-28da87713b59_4000x3267.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>This post is about how implicit coordination between powerful people allows them to act in surprisingly synchronized ways. I&#8217;ll start by discussing wokeness, the most prominent recent example. I&#8217;ll then analyze the mechanism behind such coordination (which I call the <em>consensus of power</em>); discuss how to oppose it; and outline some possibilities for what healthier power structures would look like.</p><h3>The mystery of the cathedral</h3><p>The politics of the last decade have been very weird. The Great Awokening gave rise to an extreme variant of progressive ideology that quickly gained a huge amount of traction amongst American elites, and from there embedded itself into institutions across America and the wider western world.</p><p>Sudden ideological change is nothing new&#8212;<a href="https://www.mindthefuture.info/p/power-lies-trembling">my last post</a> was about political preference cascades, and how they&#8217;re a natural result of social behavior. But what&#8217;s fascinatingly novel about the Great Awokening is the extent to which it was an almost entirely <em>leaderless</em> movement. There was no modern MLK leading this new charge for a racial reckoning; nor a Harvey Milk of trans rights; nor even Butlers or Steins writing incisive commentaries. The closest we had was Obama, who in hindsight was a milquetoast leader whose cult of personality was driven in large part by progressive longing for a black president.</p><p>One of Curtis Yarvin&#8217;s main intellectual contributions has been to give an account of wokeness (and <a href="https://www.unqualified-reservations.org/2007/06/ultracalvinist-hypothesis-in/">its historical antecedents</a>) which highlights this peculiarity, specifically via his concept of <em><a href="https://graymirror.substack.com/p/a-brief-explanation-of-the-cathedral">the cathedral</a></em>:</p><blockquote><p>&#8220;The cathedral&#8221; is just a short way to say &#8220;journalism plus academia&#8221;&#8212;in other words, the intellectual institutions at the center of modern society, just as the Church was the intellectual institution at the center of medieval society.<br><br>But the label is making a point. The Catholic Church is one institution&#8212;the cathedral is many institutions. Yet the label is singular. This transformation from many to one&#8212;literally, e pluribus unum&#8212;is the heart of the mystery at the heart of the modern world.</p><p>The mystery of the cathedral is that all the modern world&#8217;s legitimate and prestigious intellectual institutions, even though they have no central organizational connection, behave in many ways as if they were a single organizational structure.</p></blockquote><p>That is: different parts of the cathedral talk about the same issues, support the same policies, share the same taboos, and so on. Why? When we see this level of synchrony without visible leaders, it&#8217;s tempting to take a conspiratorial stance: there <em>are</em> leaders, they&#8217;re just acting surreptitiously. And undoubtedly there have been many <em>small</em> conspiracies within woke organizations, some of which grew to involve high-level figures (e.g. <a href="https://www.tracingwoodgrains.com/p/the-faas-hiring-scandal-a-quick-overview">the FAA hiring scandal</a>, or <a href="https://www.nytimes.com/2025/03/16/opinion/covid-pandemic-lab-leak.html">the conspiracy to suppress the lab leak hypothesis</a>, or <a href="https://www.justice.gov/archives/opa/pr/justice-department-finds-yale-illegally-discriminates-against-asians-and-whites-undergraduate">conspiracies within universities</a> to <a href="https://en.wikipedia.org/wiki/Students_for_Fair_Admissions_v._Harvard">discriminate against Asians</a>).</p><p>But wokeness is so visibly a bottom-up phenomenon (with high-level figures being followers rather than leaders) that it&#8217;s difficult to postulate that its overall priorities are set by any large-scale conspiracy. So Yarvin doesn&#8217;t. Instead, <a href="https://graymirror.substack.com/p/a-brief-explanation-of-the-cathedral">he explains</a> the mystery of the cathedral in <a href="https://en.wikipedia.org/wiki/Memetics">memetic</a> terms:</p><blockquote><p>[In a progressive liberal democracy], there is a market for dominant ideas. A dominant idea is an idea that validates the use of power &#8230; And there is no market for recessive ideas. A recessive idea is an idea that invalidates power or its use.</p><p>&#8230;It is not hard to see why, in the lecture halls and newsrooms, dominant ideas tend to outcompete recessive ideas. A dominant idea is an idea that tends to benefit you and your friends. A dominant idea will be especially popular with your friends and former students in the civil service, because it gives them more work and more power.</p></blockquote><p>Now, competition between ideas is undoubtedly a big part of the story of wokeness. The introduction of algorithmic newsfeeds in the early 2010s produced hotbeds of rapid cultural evolution, which quickly led to <a href="https://x.com/arctotherium42/status/1880997332092891586">the widespread transmission of many woke concepts</a>. Even if you don&#8217;t like Yarvin&#8217;s specific memetic explanation (which I don&#8217;t, since many aspects of wokeness are recessive ideas), <a href="https://slatestarcodex.com/2014/12/17/the-toxoplasma-of-rage/">there</a> <a href="https://ncofnas.com/p/a-guide-for-the-hereditarian-revolution">are</a> <a href="https://www.astralcodexten.com/p/the-psychopolitics-of-trauma">plenty</a> <a href="https://www.postliberalorder.com/p/against-the-politics-of-envy">more</a> <a href="https://web.archive.org/web/20240403094949/https://www.residentcontrarian.com/p/an-article-about-a-book-about-pornography?r=75epr">to</a> <a href="https://www.city-journal.org/article/nietzsche-can-help-us-understand-wokeness">choose</a> <a href="https://www.amazon.com/Parasitic-Mind-Infectious-Killing-Common/dp/162157959X">from</a>. Indeed, the common description of wokeness as a &#8220;mind virus&#8221; closely parallels Dawkins&#8217; description of religions as &#8220;<a href="https://en.wikipedia.org/wiki/Viruses_of_the_Mind">viruses of the mind</a>&#8221;.</p><p>But there&#8217;s a crucial difference between explaining how an idea spreads and explaining how people who believe in that idea coordinate. <a href="https://x.com/RichardMCNgo/status/1895173615865471032">Memetics generally focuses on the former</a>; the mystery of the cathedral is about the latter. When woke institutions behave &#8220;as if they were a single organizational structure&#8221;, it often involves each of them reacting to each other&#8217;s actions (e.g. media outlets rapidly flipping from &#8220;it&#8217;s racist to worry about covid&#8221; to &#8220;masks and lockdowns are crucial&#8221;). That level of synchrony is hard to explain in terms of which ideas are memetically fitter.</p><p>So to solve the mystery of the cathedral we should appeal neither to conspiracies nor competitions between ideas. Instead, I think, we need a third type of explanation: an &#8220;emergent conspiracy&#8221; of people independently acting in ways that they expect others to support. I call this phenomenon the <em>consensus of power</em> (a term I originally got from <a href="https://x.com/benlandautaylor">Ben Landau-Taylor</a>, which I&#8217;ll define more precisely in the next section).<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a></p><p>The cathedral is a prominent example of a consensus of power. But my main focus in this post isn't to analyze the cathedral specifically, but rather the game-theoretic dynamics underlying how consenses of power work in general. Daniel Ellsberg describes some of these dynamics in <a href="https://www.amazon.com/Secrets-Memoir-Vietnam-Pentagon-Papers/dp/0142003425">his account of US decision-making about the Vietnam War</a>. So does Naunihal Singh in his account of coups, and Timur Kuran in his account of preference falsification, both of which I discussed at length in <a href="https://www.mindthefuture.info/p/power-lies-trembling">my last post</a>. You can think of this post as extending those ideas from the simple case of &#8220;will the coup succeed?&#8221; to a setting where people are deciding their positions on many related issues based on their expectations about other people&#8217;s expectations and behavior. (Unfortunately, this change makes discrete formalisms like <a href="https://ericneyman.wordpress.com/2021/06/05/social-behavior-curves-equilibria-and-radicalism/">threshold models</a> and <a href="https://en.wikipedia.org/wiki/Keynesian_beauty_contest">Keynesian beauty contests</a> much less applicable. I&#8217;m keen to better understand how to formally model consenses of power; for now, though, I&#8217;ll focus on their qualitative dynamics.)</p><h3>The consensus of power</h3><p>Whenever we interact with others in society, we need to figure out which social and political norms to use. (Here I&#8217;m using the term &#8220;norms&#8221; broadly to include which factions to favor in which ways, which ideas are within the Overton window, etc.) A crucial feature of norms is that they don&#8217;t just guide your own behavior, but also which behaviors you reward or punish when done by others. So you benefit from adopting (and enforcing) the same norms as other people, especially powerful people. This means that it&#8217;s advantageous to track not just which norms other people follow, but which norms people believe other people will follow, and which norms people believe that people believe that other people will follow, and which&#8230;</p><p>This suggests a definition for the consensus of power: the norms which are common knowledge amongst the ruling elite. Of course there&#8217;s no clear demarcation for who qualifies as a ruling elite; what I&#8217;m trying to gesture towards (but can&#8217;t yet define precisely) is a version of common knowledge which weights more powerful people more highly. Note that this is very different from just aggregating powerful people&#8217;s preferences, because norms can persist even when they&#8217;re extremely unpopular. Often when someone goes against the consensus of power, you don&#8217;t personally care. But you know that others will punish you if you&#8217;re associated with it (because others will punish <em>them</em> if they don&#8217;t punish you, because&#8230;). So you push back against it&#8212;and thus help uphold the consensus of power. I&#8217;ll call people who consistently uphold the consensus of power &#8220;apparatchiks&#8221;.</p><p>The consensus of power, like the conspiratorial stance, attempts to explain the surprising level of coordination often seen within ruling elites. But direct coordination via conspiracy and indirect coordination via consensus are very different mechanisms that make importantly different predictions. I&#8217;ll elucidate the differences by framing each of them as a type of superagent containing many individual people.</p><p>Conspiracies are <em>centralized agents</em> who coordinate in a top-down way, by deferring to the decisions of leaders. This allows them to be highly goal-directed, and to plan and execute complex strategies. But it also bottlenecks them both on trust in those leaders, and on how well those leaders can gather and process information. This leaves them vulnerable to direct attack: disrupt or undermine their leaders and you can take out the entire conspiracy.</p><p>By contrast, consenses are <em>distributed agents</em>. They need not have any leaders, or even any central nodes. Instead, each apparatchik observes the behavior of many other apparatchiks, and gradually adjusts their own behavior accordingly. This allows consenses of power to perceive and act on a lot of information in parallel.</p><p>It also makes attacking them difficult, because there&#8217;s no clear locus of power. One good example comes from the ideological conformity of the mainstream media over the last few decades. If there&#8217;d been a news czar who&#8217;d banned coverage of Biden&#8217;s cognitive decline, or Pakistani pedophile gangs, or opposition to the BLM riots, or pushback against child transitions, or immigrant crime statistics, or racial IQ gaps, then that person could be criticised directly. But instead there were just thousands of individual decisions made by apparatchik news editors trying to judge which stories the consensus of power wanted them to run, resulting in a news environment almost as homogeneous as if there <em>had</em> been a left-wing news czar.</p><p>So consenses of power can be very influential and robust. But they&#8217;re also myopic and inflexible. Apparatchiks will only throw their weight behind positions that it&#8217;s common knowledge that the consensus of power favors. But without leaders willing to take a risk to rally people around a position, common knowledge is hard to achieve. So consenses of power are very limited in their ability to &#8220;think outside the box&#8221; or course-correct when pursuing their goals. The latter typically requires big flashy failures, which allow apparatchiks to then coordinate a push for a &#8220;vibe shift&#8221;.</p><p>A prominent example of these dynamics was the process by which the Democrats selected first Biden then Harris as their 2024 nominee. Biden&#8217;s decision to run again was very unwise but hard for leading Democrats to oppose, since nobody wanted to stick their neck out by taking a stand against such a powerful figure. There was also no Schelling point to coordinate on pushing back on it&#8212;until Biden&#8217;s poor debate performance broke the taboo, leading to a huge surge of criticism.</p><p>The shift was so sudden that many hypothesized that it was centrally coordinated, but that hypothesis isn&#8217;t necessary. Instead, we can picture every Democrat leader individually wanting to get rid of Biden, then jumping on the bandwagon as soon as it became socially permissible to do so. Biden tried hard to cling onto the nomination, but once roused a consensus of power is very difficult to oppose. Yet even then the difference between conspiracy and consensus was clear. Any conspiracy powerful enough to remove Biden would have replaced him with a more electable candidate than Harris. However, none of those candidates were prominent enough for the consensus to land on.</p><p>More generally, we can think of wokeness as a consensus of power under which it was common knowledge that leaders would give in to demands made by left-wing activists. As more and more leaders gave in to more and more demands, it took more and more courage for any individual leader to draw a red line. This created a channel by which left-wing activists could rapidly inject ideas that evolved on social media into prestigious institutions.</p><h3>Fighting a consensus of power</h3><p>Another way to characterize consenses of power is by understanding what it takes to resist them. When you find yourself opposing a conspiracy, the best thing you can do is bring it to light&#8212;even one piece of evidence of illegitimate collusion can blow the whole thing up. But when you find yourself opposing a consensus of power, such smoking-gun evidence may be very hard to find, or nonexistent. You can try to describe the more subtle dynamics you see, but this can easily sound uncompelling or paranoid. And even if the evidence is pretty good, it&#8217;s often hard to get people to act on it&#8212;because they&#8217;ll then find themselves opposing the whole consensus of power, which is a terrifying prospect.</p><p>In other words, attacks that work on centralized agents often won&#8217;t work on distributed agents. But distributed agents have their own weaknesses, most notably their lack of coherence. What do those suggest as strategies for fighting a consensus of power?</p><p><strong>Firstly: act in its blind spots</strong>. Common knowledge is hard to build, and so consenses of power can be surprisingly slow to incorporate knowledge that many individual apparatchiks already have. This means that there are often many possible high-impact actions which the consensus of power doesn&#8217;t yet have an opinion about. For example, the losses that the woke consensus of power have incurred from Elon buying Twitter are at least an order of magnitude larger than the purchase price of $44 billion. But it was very difficult for the woke consensus of power to defend itself against this form of attack, because buying a social media platform is a very non-standard strategy for acquiring political power. (It did act afterwards, <a href="https://judiciary.house.gov/media/press-releases/how-worlds-biggest-brands-seek-control-online-speech">via the GARM-coordinated boycott</a>, but this was too little too late.)</p><p>Although consenses of power have many blind spots, taking advantage of them isn&#8217;t always easy. Firstly, because consenses of power inherently want to preserve their own power, and are therefore suspicious of people doing unusual things at scale, even if they don&#8217;t fully understand how those things threaten them. (If you see something <a href="https://apnews.com/article/kamala-harris-trump-vance-weird-c54d506d1f533ee7aa455f7b500322c5">being criticized as &#8220;weird&#8221;</a>, it&#8217;s a good bet that this is an attack by a consensus of power.) And secondly, because it&#8217;s easy for opponents of consenses of power to slip into a conspiratorial stance, treating the consensus as an entity that&#8217;s as well-organized and goal-directed as a centralized conspiracy.</p><p>Conspiratorial thinking suggests that if you ever do something which goes against the interests of apparatchiks, the consensus of power will &#8220;get you&#8221; for it in a way that you&#8217;re not anticipating. Because consenses of power can be so widely-distributed, this creates a miasma of fear around acting (or even thinking) against them. In order to dispel that miasma, you need enough courage to act against individual apparatchiks and see that they can&#8217;t actually direct the consensus of power to target you. The rarity of people who are willing to do that helps explain Thiel&#8217;s claim that <a href="https://www.youtube.com/watch?v=E_vDZqQgmeQ">courage is in shorter supply than genius</a>.</p><p><strong>Secondly: judo-throw the consensus of power</strong>. Consenses of power have a tendency to overreach, because there&#8217;s nobody in control who can stop ill-advised bandwagons. For example, a decade ago Silicon Valley was weak enough that Democrats could score easy points by attacking it. That gradually became less and less true, but Democrat elites in government, journalism and academia continued a campaign of pointless and often petty attacks, because there was no clear signal that they should stop. This has now backfired dramatically for them, with Silicon Valley leaders coming out strongly for Trump.</p><p>So it&#8217;s possible to cause significant damage to a consensus of power by waving a red flag in front of it, to draw it into an uphill battle. But this is a risky strategy. The woke consensus of power attacked many people on increasingly flimsy grounds over the last decade (e.g. <a href="https://en.wikipedia.org/wiki/Google%27s_Ideological_Echo_Chamber">James Damore</a>, <a href="https://www.vox.com/2020/7/29/21340308/david-shor-omar-wasow-speech">David Shor</a>, and <a href="https://thehill.com/blogs/blog-briefing-room/news/502975-california-man-fired-over-alleged-white-power-sign-says-he-was/">Emmanuel Cafferty</a>). But when wider society failed to rally in their defense, those potential overreaches turned into victories, creating (self-fulfilling) common knowledge that the woke consensus could cancel people for basically anything.</p><p>So judo-throwing a consensus of power is bottlenecked not just on courage but also on integrity and political savvy. The former makes it harder for the consensus of power to find pretexts for attacking you, thereby helping your allies build enough common knowledge (<a href="https://www.mindthefuture.info/p/power-lies-trembling">or faith</a>) to mobilize to defend you. The latter allows you to choose when and how to most effectively take a stand&#8212;as civil rights leaders did by choosing Rosa Parks as a figurehead. Perhaps the best example of a judo throw on a consensus of power is Gandhi&#8217;s <a href="https://en.wikipedia.org/wiki/Salt_March">Salt March</a>, which was extremely courageous (in its commitment to nonviolence), high-integrity (in its openness about its plans and goals) and politically savvy (in its choice of the salt tax as a symbol of oppression).</p><p>Having said that, one common downside of cultivating political savvy is that understanding how the consensus of power thinks risks turning you into an apparatchik. It&#8217;s easy to overzealously police naive activists on your &#8220;own side&#8221;, out of fear that they&#8217;ll provoke the consensus of power. But doing so makes you complicit in upholding the consensus of power, and sometimes even makes you your own side&#8217;s harshest critic. (A recent example comes from <a href="https://www.spectator.co.uk/article/has-nigel-farage-missed-the-immigration-vibe-shift/">Nigel Farage</a>, who&#8217;s <a href="https://www.theguardian.com/politics/2025/mar/09/reform-uk-power-struggle-rupert-lowe-nigel-farage">so scared to be branded as dangerously far-right</a> that he&#8217;s <a href="https://www.dailymail.co.uk/news/article-14489839/Nigel-Farage-Reform-Rupert-Lowe-mass-deportations-grooming-gangs.html">preemptively making those same accusations against his own MP</a>.)</p><p>Worse, engaging closely with a consensus of power reliably makes you overestimate its stability. It&#8217;s easy to assume that apparatchiks hold their positions sincerely (since that&#8217;s the impression they&#8217;re all trying to give!) when in fact most are conformists who would flip if they saw the tide turning. So almost all politically savvy people are far less ambitious than they should be: they try to gradually nudge the consensus of power when they should be trying to replace it altogether.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a></p><h3>Replacing a consensus of power</h3><p>Recall that I&#8217;ve defined the consensus of power as common knowledge about what people will favor, weighted by their power. But power is actually in large part constituted by people&#8217;s willingness to defer to you, which is significantly affected by how much influence you have over the consensus of power. So there are many possible stable(ish) consenses of power which we could choose to steer towards or away from.</p><p>Indeed, we can think about the history of politics in terms of transitions between different consenses of power. Most of human history was governed by what I&#8217;ll call the <em>consensus of force</em>: a consensus of power weighted by the ability to deploy military force. This tended to involve power struggles between monarchs, nobles, and generals, often cashing out in very destructive wars.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a></p><p>The rise of enlightenment liberalism helped rein in the consensus of force via (implicit and explicit) rules. I&#8217;ll call the regime which has replaced it the <em>consensus of culture</em>: a consensus of power weighted by your ability to spread your ideas. The cathedral is a consensus of culture focused around the institutions which used to be the best at spreading ideas: newspapers and universities. While the consensus of force is backed up by physical coercion, consenses of culture are primarily backed up by social coercion. This dramatically reduces the costs of conflict.</p><p>We can imagine distributed agents which don&#8217;t even rely on social coercion. In <a href="https://www.mindthefuture.info/p/power-lies-trembling">my last post</a> I talked about <em>values</em> (like morality or national identity) as distributed agents. The crucial difference between values-based agents and consenses of power is that values motivate individuals even in the absence of external incentives. By contrast, a pure consensus of power has no sway over people who aren&#8217;t rewarded for following it or punished for defying it (though in practice almost all ideologies are somewhere between purely value-based and pure consenses).</p><p>But the price of non-coercion is a lack of robustness. You can&#8217;t run a society on purely non-coercive grounds, because you then have no mechanism to punish defectors. Even consenses of culture can become extremely decoupled from physical reality, because you don&#8217;t need physical force to apply social coercion. For example, South Korea&#8217;s consensus of culture is so bad that <a href="https://x.com/DemoDoomer/status/1899411850338447806">they&#8217;re driving themselves extinct</a>. More generally, social media and smartphones facilitated the creation of echo chambers in which unhinged consenses of culture could develop. And while populist surges in many western countries are now pushing back against wokeness, there&#8217;s no guarantee that the same forces won&#8217;t drive them equally crazy.</p><p>There&#8217;s an interesting historical analogy here. The development of modern militaries made the consensus of force unprecedentedly powerful, leading enlightenment intellectuals to develop the ideological paradigm of liberalism to rein it in. Liberalism spread because &#8220;letting our ideas die in our stead&#8221; was more appealing than continuing to fight wars over religion. Now the development of modern communications technology has made the consensus of culture unprecedentedly powerful. What this suggests is that we need liberalism 2.0: an ideological paradigm robust enough to rein in the excesses of the consensus of culture, but non-coercive enough that it can spread without sparking another culture war.</p><p>That&#8217;s a hard ideology design problem. But I think it&#8217;s solvable. In forthcoming posts I&#8217;ll talk about some of the other constraints that such an ideology would face, and my best guess at what it should look like.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.mindthefuture.info/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.mindthefuture.info/subscribe?"><span>Subscribe now</span></a></p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>In <a href="https://slatestarcodex.com/2018/01/24/conflict-vs-mistake/">Scott Alexander&#8217;s terminology</a>, I&#8217;d describe conspiracies as a conflict theory explanation, and memetic competition as a mistake theory explanation. I&#8217;d then describe the consensus of power as a systems theory explanation.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>For example, most AI governance advocates historically tried to take positions that would sound reasonable in DC, without planning for how to respond to predictable rapid Overton window shifts. So after ChatGPT it was DC outsiders like <a href="https://futureoflife.org/open-letter/pause-giant-ai-experiments/">FLI</a> and <a href="https://time.com/6266923/ai-eliezer-yudkowsky-open-letter-not-enough/">Yudkowsky</a> who stepped up to bring AI risk discourse into the political mainstream.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>Ben Landau-Taylor pointed out to me that the consensus of force best describes <em>feudal </em>societies. By contrast, at various points various empires (such as the Chinese and Roman empires) consolidated power to a sufficient extent that it was infeasible to challenge the center using military force. This shifted power from the consensus of force to what I&#8217;ll call the <em>consensus of the court</em>: a consensus of power weighted by the favor of the emperor.</p><p>A system where power flows from a single person might seem very different from the consenses of power I&#8217;ve discussed above. But even an emperor with absolute formal power still has scarce attention, and still needs to delegate almost all tasks. So the apparatchiks within the court will scramble to anticipate his desires, and only infrequently bring complaints directly to him (since doing so is costly and risky). The consensus of the court is then determined by common knowledge about what gatekeepers will allow to be brought to the emperor&#8217;s attention, and which types of orders the emperor can actually enforce. Not even the emperor is in control of this process, since the consensus of the court is trying to tell him exactly what he wants to hear&#8212;sometimes effectively enough that he becomes no more than a figurehead.</p></div></div>]]></content:encoded></item><item><title><![CDATA[Power Lies Trembling]]></title><description><![CDATA[A three-book review on the game theory of political power]]></description><link>https://www.mindthefuture.info/p/power-lies-trembling</link><guid isPermaLink="false">https://www.mindthefuture.info/p/power-lies-trembling</guid><dc:creator><![CDATA[Richard Ngo]]></dc:creator><pubDate>Sat, 22 Feb 2025 22:37:44 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/cb4e1410-2e56-4b27-a75d-147705669fde_1080x1228.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>In <a href="https://www.thinkingcomplete.com/2022/04/book-review-very-important-people.html">a previous book review</a> I described exclusive nightclubs as the particle colliders of sociology&#8212;places where you can reliably observe extreme forces collide. If so, military coups are the supernovae of sociology. They&#8217;re huge, rare, sudden events that, if studied carefully, provide deep insight about what lies underneath the veneer of normality around us.</p><p>That&#8217;s the conclusion I take away from Naunihal Singh&#8217;s book <em><a href="https://www.amazon.com/Seizing-Power-Strategic-Logic-Military/dp/1421422565">Seizing Power: the Strategic Logic of Military Coups</a></em>. It&#8217;s <em>not</em> a conclusion that Singh himself draws: his book is careful and academic (though much more readable than most academic books). His analysis focuses on Ghana, a country which experienced ten coup attempts between 1966 and 1983 alone. Singh spent a year in Ghana carrying out hundreds of hours of interviews with people on both sides of these coups, which led him to formulate a new model of how coups work.</p><p>I&#8217;ll start by describing Singh&#8217;s model of coups. Then I&#8217;ll explain how the dynamics of his model also apply to everything else, with reference to Timur Kuran&#8217;s excellent book on preference falsification, <em><a href="https://www.amazon.com/Private-Truths-Public-Lies-Falsification/dp/0674707583">Private Truths, Public Lies</a>.</em> In particular, I&#8217;ll explain <em>threshold models of social behavior</em>, which I find extremely insightful for understanding social dynamics.</p><p>Both of these books contain excellent sociological analyses. But they&#8217;re less useful as guides for how one should personally respond to the dynamics they describe. I think that&#8217;s because in sociology you&#8217;re always part of<em> </em>the system you&#8217;re trying to affect, so you can never take a fully objective, analytical stance towards it. Instead, acting effectively also requires the right emotional and philosophical stance. So to finish the post I&#8217;ll explore such a stance&#8212;specifically the philosophy of faith laid out by Soren Kierkegaard in his book <em><a href="https://en.wikipedia.org/wiki/Fear_and_Trembling">Fear and Trembling</a></em>.</p><h4>The revolutionary&#8217;s handbook</h4><p>What makes coups succeed or fail? Even if you haven&#8217;t thought much about this, you probably implicitly believe in one of two standard academic models of them. The first is <em>coups as elections</em>. In this model, people side with the coup if they&#8217;re sufficiently unhappy with the current regime&#8212;and if enough people side with the coup, then the revolutionaries will win. This model helps explain why popular uprisings like the Arab Spring can be so successful even when they start off with little military force on their side. The second is <em>coups as battles</em>. In this model, winning coups is about seizing key targets in order to co-opt the &#8220;nervous system&#8221; of the existing government. This model (whose key ideas are outlined in <a href="https://mirror.explodie.org/108765924-Luttwak-1969-Coup-d-Etat.pdf">Luttwak&#8217;s influential book on coups</a>) explains why coups depend so heavily on secrecy, and often succeed or fail based on their initial strikes.</p><p>Singh rejects both of these models, and puts forward a third: coups as coordination games. The core insight of this model is that, above all, military officers want to <em>join the side that will win</em>&#8212;both to ensure their and their troops&#8217; survival, and to minimize unnecessary bloodshed overall. Given this, their own preferences about which side they&#8217;d prefer to win are less important than their <em>expectations</em> about which side <em>other people</em> will support. This explains why very unpopular dictators can still hold onto power for a long time (even though the <em>coups as elections</em> model predicts they&#8217;d quickly be deposed): because everyone expecting everyone else to side with the dictator is a stable equilibrium.</p><p>It also explains why the targets that revolutionaries focus on are often not ones with military importance (as predicted by the <em>coups as battles</em> model) but rather targets of <em>symbolic </em>importance, like parliaments and palaces&#8212;since holding them is a costly signal of strength. Another key type of target often seized by revolutionaries is broadcasting facilities, especially radio stations. Why? Under the <em>coups as battles</em> model, it&#8217;s so they can coordinate their forces (and disrupt the coordination of the existing regime&#8217;s forces). Meanwhile the <em>coups as elections</em> model suggests that revolutionaries should use broadcasts to persuade people that they&#8217;re better than the old regime. Instead, according to Singh, what we most often observe is revolutionaries publicly broadcasting claims that <em>they&#8217;ve already won</em>&#8212;or (when already having won is too implausible to be taken seriously) that <em>their victory is inevitable</em>.</p><p>It&#8217;s easy to see why, if you believed those claims, you&#8217;d side with the coup. But, crucially, such claims can succeed without actually persuading anyone! If you believe that <em>others</em> are gullible enough to fall for those claims, you should fall in line. Or if you believe that others believe that <em>you</em> will believe those claims, then they will fall in line and so you should too. In other words, coups are an incredibly unstable situation where everyone is trying to predict everyone else&#8217;s predictions about everyone else&#8217;s predictions about everyone else&#8217;s predictions about everyone else&#8217;s&#8230; about who will win. Once the balance starts tipping one way, it will quickly accelerate. And so each side&#8217;s key priority is making themselves the <a href="https://en.wikipedia.org/wiki/Focal_point_(game_theory)">Schelling point</a> for coordination via managing public information (i.e. information that everyone knows everyone else has) about what&#8217;s happening. (This can be formally modeled as a <a href="https://en.wikipedia.org/wiki/Keynesian_beauty_contest">Keynesian beauty contest</a>. Much more on this in follow-up posts.)</p><p>Singh calls the process of creating self-fulfilling <a href="https://www.lesswrong.com/posts/9QxnfMYccz9QRgZ5z/the-costly-coordination-mechanism-of-common-knowledge">common knowledge</a> <em>making a fact</em>. I find this a very useful term, which also applies to more mundane situations&#8212;e.g. taking the lead in a social context can make a fact that you&#8217;re now in charge. Indeed, one of the most interesting parts of Singh&#8217;s book was a description of how coups can happen via managing the social dynamics of meetings of powerful people (e.g. all the generals in an army). People rarely want to be the first to defend a given side, especially in high-stakes situations. So if you start the meeting with a few people confidently expressing support for a coup, and then ask if anyone objects, the resulting silence can make the fact that everyone supports the coup. This strategy can succeed even if almost all the people in the meeting oppose the coup&#8212;if none of them dares to say so in the meeting, it&#8217;s very hard to rally them afterwards against what&#8217;s now become the common-knowledge default option.</p><p>One of Singh&#8217;s case studies hammers home how powerful meetings are for common knowledge creation. In 1978, essentially all the senior leaders in the Ghanaian military wanted to remove President Acheampong. However, they couldn&#8217;t create common knowledge of this, because it would be too suspicious for them to all meet without the President. Eventually Acheampong accidentally sealed his fate by sending a letter to a general criticizing the military command structure, which the general used as a pretext to call a series of meetings culminating in a bloodless coup in the President&#8217;s office.</p><p>Meetings are powerful not just because they get the key people in the same place, but also because they can be run quickly. The longer a coup takes, the less of a <em>fait accompli </em>it appears, and the more room there is for doubt to creep in. Singh ends the book with a fascinating case study of the 1991 coup attempt by Soviet generals against Gorbachev and Yeltsin. Even accounting for cherry-picking, it&#8217;s impressive how well this coup lines up with the &#8220;coups as coordination games&#8221; model. The conspirators included almost all of the senior members of the current government, and timed their strike for when both Gorbachev and Yeltsin were on vacation&#8212;but made the mistake of allowing Yeltsin to flee to the Russian parliament. From there he made a series of speeches asserting his moral legitimacy, while his allies spread rumors that the coup was falling apart. Despite having Yeltsin surrounded with overwhelming military force, bickering and distrust amongst the conspirators delayed their assault on the parliament long enough for them to become demoralized, at which point the coup essentially fizzled out.</p><p>Another of Singh&#8217;s most striking case studies was of a low-level Ghanaian soldier, Jerry Rawlings, who carried out a successful coup with less than a dozen armed troops. He was able to succeed in large part because the government had shown weakness by airing warnings about the threat Rawlings posed, and pleas not to cooperate with him. This may seem absurd, but Singh does a great job characterizing what it&#8217;s like to be a soldier confronted by revolutionaries in the fog of war, hearing all sorts of rumors that something big is happening, but with no real idea how many people are supporting the coup. In that situation, by far the easiest option is to stand aside, lest you find yourself standing alone against the new government. And the more people stand aside, the more snowballing social proof the revolutionaries have. So our takeaway from the Soviet coup attempt shouldn&#8217;t be that making a fact is <em>inherently</em> difficult&#8212;just that rank and firepower are no substitute for information control.</p><p>I don&#8217;t think of Singh as totally disproving the two other theories of coups&#8212;they probably all describe complementary dynamics. For example, if the Soviet generals had captured Yeltsin in their initial strike, he wouldn&#8217;t have had the chance to win the subsequent coordination game. And though Singh gives a lot of good historical analysis, he&#8217;s light on advance predictions. But Singh&#8217;s model is still powerful enough that it should constrain our expectations in many ways. For example, I&#8217;d predict based on Singh&#8217;s theory that radio will still be important for coups in developing countries, even now that it&#8217;s no longer the main news source for most people. The internet can convey much more information much more quickly, but radio is still better for creating common knowledge, in part <em>because</em> of its limitations (like having a fixed small number of channels). If you think of other predictions which help distinguish these three theories of coups, do let me know.</p><h4>From explaining coups to explaining everything</h4><p>Singh limits himself to explaining the dynamics of coups. But once he points them out, it&#8217;s easy to start seeing them everywhere. What if <em>everything</em> is a coordination game?</p><p>That&#8217;s essentially the thesis of Timur Kuran&#8217;s book <em><a href="https://www.amazon.com/Private-Truths-Public-Lies-Falsification/dp/0674707583">Private Truths, Public Lies</a></em>. Kuran argues that a big factor affecting which beliefs people express on basically all political topics is their desire to conform to the opinions expressed by others around them&#8212;a dynamic known as <em>preference falsification</em>. Preference falsification can allow positions to maintain dominance even as they become very unpopular. But it also creates a reservoir of pent-up energy that, when unleashed, can lead public opinion to change very rapidly&#8212;a process known as a <em>preference cascade</em>.</p><p>The most extreme preference cascades come during coups when common knowledge tips towards one side winning (as described above). But Kuran chronicles many other examples, most notably the history of race relations in America. In his telling, both the end of slavery and the end of segregation happened significantly after white American opinion had tipped against them&#8212;because people didn&#8217;t know that other people had also changed their minds. &#8220;According to one study [in the 70s], 18 percent of the whites favored segregation, but as many as 47 percent believed that most did so.&#8221; And so change, when it came, was very sudden: &#8220;In the span of a single decade, the 1960s, the United States traveled from government-supported discrimination against blacks to the prohibition of all color-based discrimination, and from there to government-promoted discrimination in favor of blacks.&#8221;</p><p>According to Kuran, this shift unfortunately wasn&#8217;t a reversion from preference falsification to honesty, but rather an overshot into a new regime of preference falsification. Writing in 1995, he claims that &#8220;white Americans are overwhelmingly opposed to special privileges for blacks. But they show extreme caution in expressing themselves publicly, for fear of being labeled as racists.&#8221; This fear has entrenched affirmative action ever more firmly over the decades since then, until the very recent and very sudden rise of MAGA.</p><p>Kuran&#8217;s other main examples are communism and the Indian caste system. His case studies are interesting, but the most valuable part of the book for me was his exposition of a formal model of preference falsification and preference cascades: threshold models of social behavior. For a thorough explanation of them, see <a href="https://ericneyman.wordpress.com/2021/06/05/social-behavior-curves-equilibria-and-radicalism/">this blog post by Eric Neyman</a> (who calls visual representations of threshold models <em>social behavior curves</em>). Here I&#8217;ll just give an abbreviated introduction by stealing some of Eric&#8217;s graphs.</p><p>The basic idea is that threshold models describe how people&#8217;s willingness to do something depends on how many other people are doing it. Most people have some threshold at which they&#8217;ll change their public position, which is determined by a combination of their own personal preferences and the amount of pressure they feel to conform to others. For example, the graph below is a hypothetical social behavior curve of what percentage of people would wear facemasks in public, as a function of how many people they see already wearing masks. (The axis labels are a little confusing&#8212;you could also think of the x and y axes as &#8220;mask-wearers at current timestep&#8221; and &#8220;mask-wearers at next timestep&#8221; respectively.)</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!lj29!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe5cd891-3846-4363-9e87-96168ed45c3b_727x516.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!lj29!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe5cd891-3846-4363-9e87-96168ed45c3b_727x516.png 424w, https://substackcdn.com/image/fetch/$s_!lj29!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe5cd891-3846-4363-9e87-96168ed45c3b_727x516.png 848w, https://substackcdn.com/image/fetch/$s_!lj29!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe5cd891-3846-4363-9e87-96168ed45c3b_727x516.png 1272w, https://substackcdn.com/image/fetch/$s_!lj29!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe5cd891-3846-4363-9e87-96168ed45c3b_727x516.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!lj29!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe5cd891-3846-4363-9e87-96168ed45c3b_727x516.png" width="727" height="516" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/fe5cd891-3846-4363-9e87-96168ed45c3b_727x516.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:516,&quot;width&quot;:727,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!lj29!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe5cd891-3846-4363-9e87-96168ed45c3b_727x516.png 424w, https://substackcdn.com/image/fetch/$s_!lj29!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe5cd891-3846-4363-9e87-96168ed45c3b_727x516.png 848w, https://substackcdn.com/image/fetch/$s_!lj29!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe5cd891-3846-4363-9e87-96168ed45c3b_727x516.png 1272w, https://substackcdn.com/image/fetch/$s_!lj29!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe5cd891-3846-4363-9e87-96168ed45c3b_727x516.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>On this graph, if 35% of people currently wear masks, then once this fact becomes known around 50% of people would want to wear masks. This means that 35% of people wearing masks is not an equilibrium&#8212;if the number of mask-wearers starts at 35%, it will increase over time. More generally, whenever the percentage of people wearing a mask corresponds to a point on the social behavior curve above the y=x diagonal, then the number of mask-wearers will increase; when below y=x, it&#8217;ll decrease. So the equilibria are places where the curve intersects y=x. But only equilibria which cross from the left side to the right side are stable; those that go the other way are unstable (like a pen balanced on its tip), with any slight deviation sending them spiraling away towards the nearest stable equilibrium.</p><p>I recommend staring at the graph above until that last paragraph feels obvious.</p><p>I find the core insights of threshold models extremely valuable; I think of them as sociology&#8217;s analogue to supply and demand curves in economics. They give us simple models of <a href="https://en.wikipedia.org/wiki/Moral_panic">moral panics</a>, <a href="https://slatestarcodex.com/2019/02/04/respectability-cascades/">respectability cascades</a>, <a href="https://en.wikipedia.org/wiki/Echo_chamber_(media)">echo chambers</a>, <a href="https://www.astralcodexten.com/p/give-up-seventy-percent-of-the-way">the euphemism treadmill</a>, and a multitude of other sociological phenomena&#8212;including coups.</p><p>We can model coups as an extreme case where the only stable equilibria are the ones where everyone supports one side or everyone supports the other, because the pressure to be on the winning side is so strong. This implies that coups have an s-shaped social behavior curve, with a very unstable equilibrium in the middle&#8212;something like the diagram below. The steepness of the curve around the unstable equilibrium reflects the fact that, once people figure out which side of the tipping point they&#8217;re on, support for that side snowballs very quickly.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!SdxE!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41cf26f8-7706-4741-baa2-927bace56beb_397x369.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!SdxE!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41cf26f8-7706-4741-baa2-927bace56beb_397x369.png 424w, https://substackcdn.com/image/fetch/$s_!SdxE!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41cf26f8-7706-4741-baa2-927bace56beb_397x369.png 848w, https://substackcdn.com/image/fetch/$s_!SdxE!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41cf26f8-7706-4741-baa2-927bace56beb_397x369.png 1272w, https://substackcdn.com/image/fetch/$s_!SdxE!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41cf26f8-7706-4741-baa2-927bace56beb_397x369.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!SdxE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41cf26f8-7706-4741-baa2-927bace56beb_397x369.png" width="397" height="369" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/41cf26f8-7706-4741-baa2-927bace56beb_397x369.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:369,&quot;width&quot;:397,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:18036,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!SdxE!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41cf26f8-7706-4741-baa2-927bace56beb_397x369.png 424w, https://substackcdn.com/image/fetch/$s_!SdxE!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41cf26f8-7706-4741-baa2-927bace56beb_397x369.png 848w, https://substackcdn.com/image/fetch/$s_!SdxE!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41cf26f8-7706-4741-baa2-927bace56beb_397x369.png 1272w, https://substackcdn.com/image/fetch/$s_!SdxE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F41cf26f8-7706-4741-baa2-927bace56beb_397x369.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>This diagram illustrates that shifting the curve a few percent left or right has highly nonlinear effects. For most possible starting points, it won&#8217;t have any effect. But if we start off near an intersection, then even a small shift could totally change the final outcome. You can see an illustration of this possibility (again from Eric&#8217;s blog post) below&#8212;it models a persuasive argument which makes people willing to support something with 5 percentage points less social proof, thereby shifting the equilibrium a long way. The historical record tells us that courageous individuals defying social consensus can work in practice, but now it works in <em>theory</em> too.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!T_5u!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc4675a6-bcdd-4e27-bfe8-a85f7eff9d68_538x518.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!T_5u!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc4675a6-bcdd-4e27-bfe8-a85f7eff9d68_538x518.png 424w, https://substackcdn.com/image/fetch/$s_!T_5u!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc4675a6-bcdd-4e27-bfe8-a85f7eff9d68_538x518.png 848w, https://substackcdn.com/image/fetch/$s_!T_5u!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc4675a6-bcdd-4e27-bfe8-a85f7eff9d68_538x518.png 1272w, https://substackcdn.com/image/fetch/$s_!T_5u!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc4675a6-bcdd-4e27-bfe8-a85f7eff9d68_538x518.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!T_5u!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc4675a6-bcdd-4e27-bfe8-a85f7eff9d68_538x518.png" width="538" height="518" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/cc4675a6-bcdd-4e27-bfe8-a85f7eff9d68_538x518.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:518,&quot;width&quot;:538,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!T_5u!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc4675a6-bcdd-4e27-bfe8-a85f7eff9d68_538x518.png 424w, https://substackcdn.com/image/fetch/$s_!T_5u!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc4675a6-bcdd-4e27-bfe8-a85f7eff9d68_538x518.png 848w, https://substackcdn.com/image/fetch/$s_!T_5u!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc4675a6-bcdd-4e27-bfe8-a85f7eff9d68_538x518.png 1272w, https://substackcdn.com/image/fetch/$s_!T_5u!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc4675a6-bcdd-4e27-bfe8-a85f7eff9d68_538x518.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Having said all that, I don&#8217;t want to oversell threshold models. They&#8217;re still very simple, which means that they miss some important factors:</p><ul><li><p>They only model a binary choice between supporting and opposing something, whereas most people are noncommittal on most issues by default (especially in high-stakes situations like coups). But adding in this third option makes the math much more complicated&#8212;e.g. it introduces the possibility of cycles, meaning there might not be <em>any</em> equilibria.</p></li><li><p>Realistically, support and opposition aren&#8217;t limited to discrete values, but can range continuously from weak to strong. So perhaps we should think of social behavior curves in terms of average level of support rather than number of supporters.</p></li><li><p>Threshold models are memoryless: the next timestep depends only on the current timestep. This means that they can&#8217;t describe, for example, the momentum that builds up after behavior consistently shifts in one direction.</p></li><li><p>Threshold models treat all people symmetrically. By contrast, belief propagation models track how preferences cascade through a network of people, where each person is primarily responding to <em>local</em> social incentives. Such models are more realistic than simple threshold models.</p></li></ul><p>I&#8217;d be very interested to hear about extensions to threshold models which avoid these limitations.</p><h4>From explaining everything to influencing everything</h4><p>How should understanding the prevalence of preference falsification change our behavior? Most straightforwardly, it should predispose us to express our true beliefs more even in controversial cases&#8212;because there might be far more people who agree with us than it appears. And as described above, threshold models give an intuition for how even a small change in people&#8217;s willingness to express a view can trigger big shifts.</p><p>However, there&#8217;s also a way in which threshold models can easily be misleading. In the diagram above, we modeled persuasion as an act of shifting the curve. But the most important aspect of persuasion is often not your argument itself, but rather the social proof you provide by defending a conclusion. And so in many cases it&#8217;s more realistic to think of your argument, not as translating the entire curve, but as merely increasing the number of advocates for X by one.</p><p>There&#8217;s a more general point here. It&#8217;s tempting to think that you can estimate the social behavior curve, then decide how you&#8217;ll act based on that. But everyone else&#8217;s choices are based on their predictions of you, and you&#8217;re constantly leaking information about your decision-making process. So you can&#8217;t generate credences about how others will decide, then use them to make your decision, because your eventual decision is heavily correlated with other people&#8217;s decisions. You&#8217;re not just intervening on the curve, you <em>are</em> the curve.</p><p>More precisely, social behavior is a domain where the correlations between people&#8217;s decisions are strong enough to make <a href="https://en.wikipedia.org/wiki/Causal_decision_theory">causal decision theory</a> misleading. Instead it&#8217;s necessary to use either <em><a href="https://en.wikipedia.org/wiki/Evidential_decision_theory">evidential decision theory</a></em> or <em><a href="https://intelligence.org/2017/03/18/new-paper-cheating-death-in-damascus/">functional decision theory</a></em>. Both of these track the non-causal dependencies between your decision and other people&#8217;s decisions. In particular, both of them involve a step where you reason &#8220;if I do something, then it&#8217;s more likely that others will do the same thing&#8221;&#8212;even when they have no way of finding out about your final decision before making theirs. So you&#8217;re not searching for a decision which <em>causes</em> good things to happen; instead you&#8217;re searching for a desirable fixed point for simultaneous correlated decisions by many people.</p><p>I&#8217;ve put this in cold, rational language. But what we&#8217;re talking about is nothing less than a <em>leap of faith</em>. Imagine sitting at home, trying to decide whether to join a coup to depose a hated ruler. Imagine that if enough of you show up on the streets at once, loudly and confidently, then you&#8217;ll succeed&#8212;but that if there are only a few of you, or you seem scared or uncertain, then the regime won&#8217;t be cowed, and will arrest or kill all of you. Imagine your fate depending on something you can&#8217;t control at all except via the fact that if you have faith, others are more likely to have faith too. It&#8217;s a terrifying, gut-wrenching feeling.</p><p>Perhaps the most eloquent depiction of this feeling comes from Soren Kierkegaard in his book <em><a href="https://en.wikipedia.org/wiki/Fear_and_Trembling">Fear and Trembling</a></em>. Kierkegaard is moved beyond words by the story of Abraham, who is not only willing to sacrifice his only son on God&#8217;s command&#8212;but somehow, even as he&#8217;s doing it, still believes against all reason that everything will turn out alright. Kierkegaard struggles to describe this level of pure faith as anything but absurd. Yet it&#8217;s this absurdity that is at the heart of social coordination&#8212;because you can never fully reason through what happens when other people predict your predictions of their predictions of&#8230; To cut through that, you need to simply <em>decide</em>, and hope that your decision will <em>somehow</em> change everyone else&#8217;s decision. You walk out your door to possible death because you believe, absurdly, that doing so will make other people simultaneously walk out of the doors of their houses all across the city.</p><p>A modern near-synonym for &#8220;leap of faith&#8221; is &#8220;hyperstition&#8221;: an idea that you bring about by believing in it. This is <a href="https://www.orphandriftarchive.com/articles/hyperstition/">Nick Land&#8217;s term</a>, which he seems to use primarily for larger-scale memeplexes&#8212;like capitalism, the ideology of progress, or AGI. Deciding whether or not to believe in these hyperstitions has some similarity to deciding whether or not to join a coup, but the former are much harder to reason about by virtue of their scale. We can think of hyperstitions as forming the background landscape of psychosocial reality: the <a href="https://en.wikipedia.org/wiki/The_Commanding_Heights">commanding heights of ideology</a>, the shifting sands of public opinion, and the <a href="https://en.wikipedia.org/wiki/On_What_Matters#Summary">moral mountain</a> off which we may&#8212;or may not&#8212;take a leap into the <a href="https://www.poetryfoundation.org/poems/43588/dover-beach">sea of faith</a>.</p><h4>Becoming a knight of faith</h4><p>Unfortunately, the mere realization that social reality is composed of hyperstitions doesn&#8217;t give you social superpowers, any more than knowing Newtonian mechanics makes you a world-class baseball player. So how can you decide when and how to actually swing for the fences? I&#8217;ll describe the tension between having too much and too little faith by contrasting three archetypes: the pragmatist, the knight of resignation, and the knight of faith.</p><p>The pragmatist treats faith as a decision like any other. They figure out the expected value of having faith&#8212;i.e. of adopting an &#8220;irrationally&#8221; strong belief&#8212;and go for it if and only if it seems valuable enough. Doing that analysis is difficult: it requires the ability to identify big opportunities, judge people&#8217;s expectations, and know how your beliefs affect common knowledge. In other words, it requires <em>skill at politics</em>, which I&#8217;ll talk about much more in <a href="https://www.mindthefuture.info/p/elite-coordination-via-the-consensus">a follow-up post</a>.</p><p>But while pragmatic political skill can get you a long way, it eventually hits a ceiling&#8212;because the world is watching not just what you do but also your <em>reasons</em> for doing it. If your choice is a pragmatic one, others will be able to tell&#8212;from your gait, your expression, your voice, your phrasing, and of course how your position evolves over time. They&#8217;ll know that you&#8217;re the sort of person who will change your mind if the cost/benefit calculus changes. And so they&#8217;ll know that they won&#8217;t truly be able to rely on you&#8212;that you don&#8217;t have <em>sincere</em> faith.</p><p>Imagine, by contrast, someone capable of fighting for a cause no matter how many others support them, no matter how hopeless it seems. Even if such a person never actually needs to fight alone, the common knowledge that they <em>would</em> makes them <a href="https://x.com/RichardMCNgo/status/1877806024603627740">a nail in the fabric of social reality</a>. They anchor the social behavior curve not merely by adding one more supporter to their side, but by being an immutable fixed point around which everyone knows (that everyone knows that everyone knows&#8230;) that they must navigate.</p><p>The archetype that Kierkegaard calls the <em>knight of resignation</em> achieves this by being <em>resigned</em> to the worst-case outcome. They gather the requisite courage by suppressing their hope, by convincing themselves that they have nothing to lose. They walk out their door having accepted death, with a kind of weaponized despair.</p><p>The grim determination of the knight of resignation is more reliable than pragmatism. But if you won&#8217;t let yourself think about the possibility of success, it&#8217;s very difficult to reason well about how it can be achieved, or to inspire others to pursue it. So what makes Kierkegaard fear and tremble is not the knight of resignation, but the <em>knight of faith</em>&#8212;the person who looks at the worst-case scenario directly, and (like the knight of resignation) sees no causal mechanism by which his faith will save him, but (like Abraham) <em>believes that he will be saved anyway</em>. That&#8217;s the kind of person who could found a movement, or a country, or a religion. It's Washington stepping down from the presidency after two terms, and Churchill holding out against Nazi Germany, and Gandhi committing to non-violence, and Navalny returning to Russia&#8212;each one making themselves a beacon that others can&#8217;t help but feel inspired by.</p><p>What&#8217;s the difference between being a knight of faith, and simply falling into wishful thinking or delusion? How can we avoid having faith in the wrong things, when the whole point of faith is that we haven&#8217;t pragmatically reasoned our way into it? Kierkegaard has no good answer for this&#8212;he seems to be falling back on the idea that if there&#8217;s anything worth having faith in, it&#8217;s God. But from the modern atheist perspective, we have no such surety, and <a href="https://www.thinkingcomplete.com/2021/01/meditations-on-faith.html">even Abraham seems like he&#8217;s making a mistake</a>. So on what basis should we decide when to have faith?</p><p>I don&#8217;t think there&#8217;s any simple recipe for making such a decision. But it&#8217;s closely related to the difference between positive motivations (like love or excitement) and negative motivations (like fear or despair). Ultimately I think of faith as a coordination mechanism grounded in values that are shared across many people, like moral principles or group identities. When you act out of positive motivation towards those values, others will be able to recognize the parts of you that also arise in them, which then become a Schelling point for coordination. That&#8217;s much harder when you act out of pragmatic interests that few others share&#8212;especially personal fear. (If you act out of fear for your group&#8217;s interests, then others may still recognize themselves in you&#8212;but you&#8217;ll also create a <a href="https://malcolmocean.com/2022/02/towardsness-awayness-motivation-arent-symmetrical/">neurotic</a> and <a href="https://x.com/RichardMCNgo/status/1810495607590543701">self-destructive movement</a>.)</p><p>I talk at length about how to replace negative motivation with positive motivation <a href="https://www.lesswrong.com/s/qXZLFGqpD7aeEgXGL">in this series of posts</a>. Of course, it&#8217;s much easier said than done. Negative motivations are titanic psychological forces which steer most decisions most people make. But replacing them is worth the effort, because it unlocks a deep integrity&#8212;the ability to cooperate with different parts of yourself all the way down, without relying on deception or coercion. And that in turn allows you to cooperate with copies of those parts that live in <em>other </em>people&#8212;to act as as part of a distributed agent held together by a shared sense of justice or fairness or goodness. That&#8217;s how we can tame the drive towards preference falsification, and steer our desire for collective identities towards those which promote the flourishing of the individuals within them.</p><p><em>Postscript: six months after publishing this post, I feel hesitant about how much I focused on the example of popular revolt (and have edited the last sentence accordingly). Marching in the streets against an oppressive government can be admirable, but it&#8217;s also very mob-like. In particular, it doesn&#8217;t ensure that the next regime will actually be better (as many Iranian revolutionaries learned at great cost). I&#8217;m now more focused specifically on faith that&#8217;s oriented towards developing a just, well-governed coalition. I&#8217;m not yet quite sure what I mean by that, but I plan to explore that idea further in forthcoming posts.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.mindthefuture.info/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Mind the Future! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Why I'm not a Bayesian]]></title><description><![CDATA[Philosophical objections to Bayesianism as an epistemology.]]></description><link>https://www.mindthefuture.info/p/why-im-not-a-bayesian</link><guid isPermaLink="false">https://www.mindthefuture.info/p/why-im-not-a-bayesian</guid><dc:creator><![CDATA[Richard Ngo]]></dc:creator><pubDate>Sun, 06 Oct 2024 15:30:52 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!koXC!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c574997-7d0b-4bc0-914c-0dc5384e4218_496x418.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>This post focuses on philosophical objections to Bayesianism as an epistemology. I first explain Bayesianism and some standard objections to it, then lay out my two main objections (inspired by ideas in philosophy of science). A follow-up post will speculate about how to formalize an alternative.</p><h3><strong>Degrees of belief</strong></h3><p>The core idea of <a href="https://plato.stanford.edu/entries/epistemology-bayesian/">Bayesian epistemology</a>: we should ideally reason by assigning credences to propositions which represent our degrees of belief that those propositions are true. (Note that this is different from Bayesianism as a set of statistical techniques, or Bayesianism as an approach to machine learning, which I don&#8217;t discuss here.)</p><p>If that seems like a sufficient characterization to you, you can go ahead and skip to the next section, where I explain my objections to it. But for those who want a more precise description of Bayesianism, and some existing objections to it, I&#8217;ll more specifically characterize it in terms of five claims. Bayesianism says that we should ideally reason in terms of:</p><ol><li><p>Propositions which are either true or false (<strong>classical logic</strong>)</p></li><li><p>Each of which is assigned a credence (<strong>probabilism</strong>)</p></li><li><p>Representing subjective degrees of belief in their truth (<strong>subjectivism</strong>)</p></li><li><p>Which at each point in time obey the&nbsp;<a href="https://en.wikipedia.org/wiki/Probability_axioms">axioms of probability</a> (<strong>static rationality</strong>)</p></li><li><p>And are updated over time by applying&nbsp;<a href="https://en.wikipedia.org/wiki/Bayes%27_theorem">Bayes&#8217; rule</a> to new evidence (<strong>rigid empiricism</strong>)</p></li></ol><p>I won&#8217;t go into the case <em>for</em> Bayesianism here except to say that it does elegantly formalize many common-sense intuitions. Bayes&#8217; rule follows directly from <a href="https://oscarbonilla.com/2009/05/visualizing-bayes-theorem/">a straightforward Venn diagram</a>. The axioms of probability are powerful and mathematically satisfying. Subjective credences seem like the obvious way to represent our uncertainty about the world. Nevertheless, there are a wide range of alternatives to Bayesianism, each branching off from the claims listed above at different points:</p><ol><li><p><em><a href="https://plato.stanford.edu/entries/knowledge-analysis/">Traditional epistemology</a></em> only accepts #1, and rejects #2. Traditional epistemologists often defend a binary conception of knowledge&#8212;e.g. one defined in terms of justified true belief (or a similar criterion, like reliable belief).</p></li><li><p><em><a href="https://en.wikipedia.org/wiki/Frequentist_probability">Frequentism</a> </em>accepts #1 and #2, but rejects #3: it doesn&#8217;t think that credences should be subjective. Instead, frequentism holds that credences should correspond to the&nbsp;relative frequency&nbsp;of an event in the long term, which is an objective fact about the world. For example, you should assign 50% credence that a flipped coin will come up heads, because if you continued flipping the coin the proportion of heads would approach 50%.</p></li><li><p><em><a href="https://www.lesswrong.com/posts/y5GftLezdozEHdXkL/an-intuitive-guide-to-garrabrant-induction">Garrabrant induction</a></em> accepts #1 to #3, but rejects #4. In order for credences to obey the axioms of probability, all the logical implications of a statement must be assigned the same credence. But this &#8220;logical omniscience&#8221; is impossible for computationally-bounded agents like ourselves. So in the Garrabrant induction framework, credences instead converge to obeying the axioms of probability&nbsp;in the limit, without guarantees that they&#8217;re coherent after only limited thinking time.</p></li><li><p><em><a href="https://www.lesswrong.com/posts/xJyY5QkQvNJpZLJRo/radical-probabilism-1">Radical probabilism</a></em> accepts #1 to #4, but rejects #5. Again, this can be motivated by qualms about logical omniscience: if thinking for longer can identify new implications of our existing beliefs, then our credences sometimes need to update via a different mechanism than Bayes&#8217; rule. So radical probabilism instead allows an agent to update to any set of statically rational credences at any time, even if they&#8217;re totally different from its previous credences. The one constraint is that each credence needs to converge over time to a fixed value&#8212;i.e. it can&#8217;t continue oscillating indefinitely (otherwise the agent would be vulnerable to a&nbsp;<a href="https://en.wikipedia.org/wiki/Dutch_book_theorems">Dutch Book</a>).</p></li></ol><p>It&#8217;s not crucial whether we classify Garrabrant induction and radical probabilism as variants of Bayesianism or alternatives to it, because my main objection to Bayesianism doesn&#8217;t fall into any of the above categories. Instead, I think we need to go back to basics and reject #1. Specifically, I have two objections to the idea that idealized reasoning should be understood in terms of propositions that are true or false:</p><ol><li><p>We should assign truth-values that are intermediate between true and false (<strong>fuzzy truth-values</strong>)</p></li><li><p>We should reason in terms of&nbsp;<em>models</em> rather than propositions (the&nbsp;<strong>semantic view</strong>)</p></li></ol><p>I&#8217;ll defend each claim in turn.</p><h3><strong>Degrees of truth</strong></h3><p>Formal languages (like code) are only able to express ideas that can be pinned down precisely. Natural languages, by contrast, can refer to vague concepts which don&#8217;t have clear, fixed boundaries. For example, the truth-values of propositions which contain&nbsp;<a href="https://learnenglish.britishcouncil.org/grammar/b1-b2-grammar/adjectives-gradable-non-gradable">gradable adjectives</a> like &#8220;large&#8221; or &#8220;quiet&#8221; or &#8220;happy&#8221; depend on how we interpret those adjectives. Intuitively speaking, a description of something as &#8220;large&#8221; can be more or less true depending on how large it actually is. The most common way to formulate this spectrum is as &#8220;fuzzy&#8221; truth-values which range from 0 to 1. A value close to 1 would be assigned to claims that are clearly true, and a value close to 0 would be assigned to claims that are clearly false, with claims that are &#8220;kinda true&#8221; in the middle.</p><p>Another type of &#8220;kinda true&#8221; statements are approximations. For example, if I claim that there&#8217;s a grocery store 500 meters away from my house, that&#8217;s probably true in an&nbsp;<em>approximate</em> sense, but false in a&nbsp;<em>precise</em> sense. But once we start distinguishing the different senses that a concept can have, it becomes clear that basically&nbsp;<em>any</em> concept can have widely divergent category boundaries depending on the context.&nbsp;<a href="https://metarationality.com/refrigerator">A striking example from Chapman</a>:&nbsp;</p><blockquote><p>A: Is there any water in the refrigerator?<br>B: Yes.<br>A: Where? I don&#8217;t see it.<br>B: In the cells of the eggplant.</p></blockquote><p>The claim that there&#8217;s water in the refrigerator is technically true, but pragmatically false. And the concept of &#8220;water&#8221; is far better-defined than almost all abstract concepts (like the ones I&#8217;m using in this post). So we should treat natural-language propositions as context-dependent by default. But that&#8217;s still consistent with some statements being&nbsp;<em>more</em> context-dependent than others (e.g. the claim that there&#8217;s air in my refrigerator would be true under almost any interpretation). So another way we can think about fuzzy truth-values is as a range from &#8220;this statement is false in almost any sense&#8221; through &#8220;this statement is true in some senses and false in some senses&#8221; to &#8220;this statement is true in almost any sense&#8221;.</p><p>Note, however, that there&#8217;s an asymmetry between &#8220;this statement is true in almost any sense&#8221; and &#8220;this statement is false in almost any sense&#8221;, because the latter can apply to two different types of claims. Firstly, claims that are meaningful but false (&#8220;there&#8217;s a tiger in my house&#8221;). Secondly, claims that are&nbsp;<em>nonsense</em>&#8212;there are just no meaningful interpretations of them at all (&#8220;colorless green ideas sleep furiously&#8221;). We can often distinguish these two types of claims by negating them: &#8220;there isn&#8217;t a tiger in my house&#8221; is true, whereas &#8220;colorless green ideas don&#8217;t sleep furiously&#8221; is still nonsense. Of course, nonsense is also a matter of degree&#8212;e.g. metaphors are by default less meaningful than concrete claims, but still not entirely nonsense.</p><p>So I've motivated fuzzy truth-values from four different angles: vagueness, approximation, context-dependence, and sense vs nonsense. The key idea behind each of them is that concepts have fluid and amorphous category boundaries (a property called&nbsp;<em><a href="https://metarationality.com/nebulosity">nebulosity</a></em>). However, putting all of these different aspects of nebulosity on the same zero-to-one scale might be an oversimplification. More generally,&nbsp;<a href="https://en.wikipedia.org/wiki/Fuzzy_logic">fuzzy logic</a> has few of the appealing properties of classical logic, and (to my knowledge) isn&#8217;t very directly useful. So I&#8217;m not claiming that we should adopt fuzzy logic wholesale, or that we know what it means for a given proposition to be X% true instead of Y% true (a question which I&#8217;ll come back to in a follow-up post). For now, I&#8217;m just claiming that there&#8217;s an important sense in which thinking in terms of fuzzy truth-values is&nbsp;<em>less wrong</em> (another non-binary truth-value) than only thinking in terms of binary truth-values.</p><h3><strong>Model-based reasoning</strong></h3><p>The intuitions in favor of fuzzy truth-values become clearer when we apply them, not just to individual propositions, but to&nbsp;<em>models</em> of the world. By a model I mean a (mathematical) structure that attempts to describe some aspect of reality. For example, a model of the weather might have variables representing temperature, pressure, and humidity at different locations, and a procedure for updating them over time. A model of a chemical reaction might have variables representing the starting concentrations of different reactants, and a method for determining the equilibrium concentrations. Or, more simply, a model of the Earth might just be a sphere.</p><p>In order to pin down the difference between reasoning about propositions and reasoning about models, philosophers of science have drawn on concepts from mathematical logic. They distinguish between the&nbsp;<em>syntactic</em> content of a theory (the axioms of the theory) and its&nbsp;<em>semantic</em> content (the models for which those axioms hold).&nbsp;<a href="https://plato.stanford.edu/entries/structure-scientific-theories/">As an example</a>, consider the three axioms of&nbsp;<a href="https://en.wikipedia.org/wiki/Projective_plane">projective planes</a>:</p><ol><li><p>For any two points, exactly one line lies on both.</p></li><li><p>For any two lines, exactly one point lies on both.</p></li><li><p>There exists a set of four points such that no line has more than two of them.</p></li></ol><p>There are infinitely many models for which these axioms hold; here&#8217;s one of the simplest:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!koXC!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c574997-7d0b-4bc0-914c-0dc5384e4218_496x418.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!koXC!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c574997-7d0b-4bc0-914c-0dc5384e4218_496x418.png 424w, https://substackcdn.com/image/fetch/$s_!koXC!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c574997-7d0b-4bc0-914c-0dc5384e4218_496x418.png 848w, https://substackcdn.com/image/fetch/$s_!koXC!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c574997-7d0b-4bc0-914c-0dc5384e4218_496x418.png 1272w, https://substackcdn.com/image/fetch/$s_!koXC!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c574997-7d0b-4bc0-914c-0dc5384e4218_496x418.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!koXC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c574997-7d0b-4bc0-914c-0dc5384e4218_496x418.png" width="496" height="418" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8c574997-7d0b-4bc0-914c-0dc5384e4218_496x418.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:418,&quot;width&quot;:496,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Geometric figure including triangle ACE with interior circle BDF and center point G. Point B is on line segment AC, D is on CE, and F is on AE. G is the center of the circle. Point G is on line segments AD, BE, and CF.&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Geometric figure including triangle ACE with interior circle BDF and center point G. Point B is on line segment AC, D is on CE, and F is on AE. G is the center of the circle. Point G is on line segments AD, BE, and CF." title="Geometric figure including triangle ACE with interior circle BDF and center point G. Point B is on line segment AC, D is on CE, and F is on AE. G is the center of the circle. Point G is on line segments AD, BE, and CF." srcset="https://substackcdn.com/image/fetch/$s_!koXC!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c574997-7d0b-4bc0-914c-0dc5384e4218_496x418.png 424w, https://substackcdn.com/image/fetch/$s_!koXC!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c574997-7d0b-4bc0-914c-0dc5384e4218_496x418.png 848w, https://substackcdn.com/image/fetch/$s_!koXC!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c574997-7d0b-4bc0-914c-0dc5384e4218_496x418.png 1272w, https://substackcdn.com/image/fetch/$s_!koXC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c574997-7d0b-4bc0-914c-0dc5384e4218_496x418.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>If propositions and models are two sides of the same coin, does it matter which one we primarily reason in terms of? I think so, for two reasons. Firstly, most models are very difficult to put into propositional form. We each have implicit mental models of our friends&#8217; personalities, of how liquids flow, of what a given object feels like, etc, which are far richer than we can express propositionally. The same is true even for many formal models&#8212;specifically those whose internal structure doesn&#8217;t directly correspond to the structure of the world. For example, a neural network might encode a great deal of real-world knowledge, but even full access to the weights doesn&#8217;t allow us to extract that knowledge directly&#8212;the fact that a given weight is 0.3 doesn&#8217;t allow us to claim that any real-world entity has the value 0.3.</p><p>What about scientific models where each element of the model&nbsp;<em>is</em> intended to correspond to an aspect of reality? For example, what&#8217;s the difference between modeling the Earth as a sphere, and just believing the proposition &#8220;the Earth is a sphere&#8221;? My answer: thinking in terms of propositions (known in philosophy of science as the&nbsp;<em><a href="https://www.princeton.edu/~hhalvors/teaching/phi520_f2012/Suppe_2000.pdf">syntactic view</a></em>) biases us towards assigning truth values in a&nbsp;<em>reductionist</em> way. This works when you&#8217;re using binary truth-values, because they relate to each other according to classical logic. But when you&#8217;re using fuzzy truth-values, the relationships between the truth-values of different propositions become much more complicated. And so thinking in terms of models (known as the&nbsp;<em><a href="https://www.jstor.org/stable/187834">semantic view</a></em>) is better because models can be assigned truth-values in a&nbsp;<em>holistic&nbsp;</em>way.</p><p>As an example: &#8220;the Earth is a sphere&#8221; is mostly true, and &#8220;every point on the surface of a sphere is equally far away from its center&#8221; is precisely true. But &#8220;every point on the surface of the Earth is equally far away from the Earth&#8217;s center&#8221; seems ridiculous&#8212;e.g. it implies that mountains don&#8217;t exist. The problem here is that rephrasing a proposition in logically equivalent terms can dramatically affect its implicit context, and therefore the degree of truth we assign to it in isolation.</p><p>The semantic view solves this by separating claims about the structure of the model itself from claims about how the model relates to the world. The former are typically much less nebulous&#8212;claims like &#8220;in the spherical model of the Earth, every point on the Earth&#8217;s surface is equally far away from the center&#8221; are straightforwardly true. But we can then bring in nebulosity when talking about the model as a whole&#8212;e.g. &#8220;my spherical model of the Earth is closer to the truth than your flat model of the Earth&#8221;, or &#8220;my spherical model of the Earth is useful for doing astronomical calculations and terrible for figuring out where to go skiing&#8221;. (Note that we can make similar claims about the mental models, neural networks, etc, discussed above.)</p><p>We might then wonder: should we be talking about the truth of entire models at all? Or can we just talk about their usefulness in different contexts, without the concept of truth? This is&nbsp;<em><a href="https://plato.stanford.edu/entries/scientific-realism/">the</a></em><a href="https://plato.stanford.edu/entries/scientific-realism/"> major debate in philosophy of science</a>. I personally think that in order to explain why scientific theories can often predict a wide range of different phenomena, we need to make claims about how well they describe the structure of reality&#8212;i.e. how true they are. But we should still use degrees of truth when doing so, because even our most powerful scientific models aren&#8217;t fully true. We know that general relativity isn&#8217;t fully true, for example, because it conflicts with quantum mechanics. Even so, it would be absurd to call general relativity false, because it clearly describes a major part of the structure of physical reality. Meanwhile Newtonian mechanics is further away from the truth than general relativity, but still much closer to the truth than&nbsp;<a href="https://en.wikipedia.org/wiki/Aristotelian_physics">Aristotelian mechanics</a>, which in turn is much closer to the truth than&nbsp;<a href="https://en.wikipedia.org/wiki/Animism">animism</a>. The general point I&#8217;m trying to illustrate here was&nbsp;<a href="https://hermiene.net/essays-trans/relativity_of_wrong.html">expressed pithily by Asimov</a>: &#8220;Thinking that the Earth is flat is wrong. Thinking that the Earth is a sphere is wrong. But if you think that they&#8217;re equally wrong, you&#8217;re wronger than both of them put together.&#8221;</p><h3><strong>The correct role of Bayesianism</strong></h3><p>The position I&#8217;ve described above overlaps significantly with the&nbsp;<em><a href="https://plato.stanford.edu/entries/structural-realism/">structural realist</a></em> position in philosophy of science. However, structural realism is usually viewed as a stance on how to interpret scientific theories, rather than how to reason more generally. So the philosophical position which best captures the ideas I&#8217;ve laid out is probably&nbsp;<a href="https://plato.stanford.edu/entries/popper/#ProbKnowVeri">Karl Popper&#8217;s&nbsp;</a><em><a href="https://plato.stanford.edu/entries/popper/#ProbKnowVeri">critical rationalism</a></em>. Popper was actually the first to try to formally define&nbsp;<a href="https://en.wikipedia.org/wiki/Verisimilitude">a scientific theory's degree of truth</a> (though he was working before the semantic view became widespread, and therefore formalized theories in terms of propositions rather than in terms of models). But his attempt&nbsp;<a href="https://plato.stanford.edu/Entries/truthlikeness/tichy-miller.html">failed on a technical level</a>; and no attempt since then has gained widespread acceptance. Meanwhile, the field of machine learning evaluates models by their loss, which can be formally defined&#8212;but the loss of a model is heavily dependent on the data distribution on which it&#8217;s evaluated. Perhaps the most promising approach to assigning fuzzy truth-values comes from <a href="https://www.lesswrong.com/posts/y5GftLezdozEHdXkL/an-intuitive-guide-to-garrabrant-induction">Garrabrant induction</a>, where the &#8220;money&#8221; earned by individual traders could be interpreted as a metric of fuzzy truth. However, these traders can strategically interact with each other, making them more like agents than typical models.</p><p>Where does this leave us? We&#8217;ve traded the crisp, mathematically elegant Bayesian formalism for fuzzy truth-values that, while intuitively compelling, we can&#8217;t define even in principle. But I&#8217;d rather be vaguely right than precisely wrong. Because it focuses on propositions which are each (almost entirely) true or false, Bayesianism is actively misleading in domains where reasoning well requires constructing and evaluating sophisticated models (i.e. most of them).</p><p>For example,&nbsp;<a href="https://www.lesswrong.com/posts/nj8JKFoLSMEmD3RGp/how-much-evidence-does-it-take">Bayesians measure evidence in &#8220;bits&#8221;</a>, where one bit of evidence rules out half of the space of possibilities. When asking a question like &#8220;<a href="https://www.lesswrong.com/posts/JD7fwtRQ27yc8NoqS/strong-evidence-is-common">is this stranger named Mark?</a>&#8221;, bits of evidence are a useful abstraction: I can get one bit of evidence simply by learning whether they&#8217;re male or female, and a couple more by learning that their name has only one syllable. Conversely,&nbsp;<a href="https://www.lesswrong.com/posts/MwQRucYo6BZZwjKE7/einstein-s-arrogance">talking in Bayesian terms about discovering scientific theories</a> is nonsense. If every PhD in fundamental physics had contributed even one bit of usable evidence about how to unify quantum physics and general relativity, we&#8217;d have solved quantum gravity many times over by now. But we haven&#8217;t, because almost all of the work of science is in&nbsp;<em>constructing</em> sophisticated models, which Bayesianism says almost nothing about. (Formalisms like <a href="https://en.wikipedia.org/wiki/Solomonoff%27s_theory_of_inductive_inference">Solomonoff induction</a> attempt to sidestep this omission by enumerating and simulating all computable models, but that&#8217;s so different from what any realistic agent can do that we should think of it less as idealized cognition and more as a different thing altogether, which just happens to converge to the same outcome in the infinite limit.)</p><p>Mistakes like these have many downstream consequences. Nobody should be very confident about complex domains that nobody has sophisticated models of (like superintelligence); but the idea that &#8220;<a href="https://www.lesswrong.com/posts/JD7fwtRQ27yc8NoqS/strong-evidence-is-common">strong evidence is common</a>&#8221; helps&nbsp;<a href="https://www.lesswrong.com/posts/JD7fwtRQ27yc8NoqS/strong-evidence-is-common?commentId=jNtPscXFL634DD47d">justify</a> confident claims about them. And without a principled distinction between credences that are derived from deep, rigorous models of the world, and credences that come from vague speculation (and are therefore subject to huge&nbsp;<a href="https://en.wikipedia.org/wiki/Knightian_uncertainty">Knightian uncertainty</a>),&nbsp;<a href="https://www.lesswrong.com/posts/tG9BLyBEiLeRJZvX6/communicating-effectively-under-knightian-norms">it&#8217;s hard for public discussions to actually make progress</a>.</p><p>Should I therefore be a critical rationalist? I do think Popper got a lot of things right. But I also get the sense that he (along with Deutsch, his most prominent advocate) throws the baby out with the bathwater. There&nbsp;<em>is</em> a great deal of insight encoded in Bayesianism which critical rationalists discard (e.g. by&nbsp;<a href="https://philosophy.tamucc.edu/index.php/texts/popper-problem-of-induction">rejecting induction</a>). A better approach is to view Bayesianism as describing a special case of epistemology, which applies in contexts simple enough that we&#8217;ve already constructed all relevant models or hypotheses, exactly one of which is exactly true (with all the rest of them being equally false), and we just need to decide between them. Interpreted in that limited way, Bayesianism is both useful (e.g. in providing a framework for bets and prediction markets) and inspiring: if we can formalize this special case so well, couldn&#8217;t we also formalize the general case? What would it look like to concretely define degrees of truth? I don&#8217;t have a solution, but I&#8217;ll outline some existing attempts, and play around with some ideas of my own, in <a href="https://www.mindthefuture.info/p/towards-a-scale-free-theory-of-intelligent">a follow-up post</a>.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.mindthefuture.info/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Mind the Future! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Meditations on Mot]]></title><description><![CDATA[A response to Scott Alexander]]></description><link>https://www.mindthefuture.info/p/meditations-on-mot</link><guid isPermaLink="false">https://www.mindthefuture.info/p/meditations-on-mot</guid><dc:creator><![CDATA[Richard Ngo]]></dc:creator><pubDate>Mon, 04 Dec 2023 00:15:54 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/788ed49a-4478-4dd3-8b0d-c27b5475b84c_632x415.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<blockquote><p><em>Holy! Holy! Holy! Holy! Holy! Holy! Holy! Holy! Holy! Holy! Holy! Holy! Holy! Holy! Holy!<br>Everything is holy! everybody&#8217;s holy! everywhere is holy! everyday is in eternity! Everyman&#8217;s an angel!<br>Holy the lone juggernaut! Holy the vast lamb of the middleclass! Holy the crazy shepherds of rebellion! Who digs Los Angeles IS Los Angeles!<br>Holy New York Holy San Francisco Holy Peoria &amp; Seattle Holy Paris Holy Tangiers Holy Moscow Holy Istanbul!<br>Holy time in eternity holy eternity in time holy the clocks in space holy the fourth dimension holy the fifth International holy the Angel in Moloch!<br>- Footnote to </em>Howl</p></blockquote><p><em><strong>Scene: Carl and Allen, two old friends, are having a conversation about theodicy.</strong></em></p><p><strong>Carl:</strong> &#8220;Let me tell you about the god who is responsible for almost all our suffering. This god is an ancient Canaanite god, one who has been seen throughout history as a source of death and destruction. Of course, he doesn&#8217;t exist in a literal sense, but we can conceptualize him as a manifestation of forces that persist even today, and which play a crucial role in making the world worse. His name is M-&#8221;</p><p><strong>Allen:</strong> &#8220;-oloch, right? Scott Alexander&#8217;s god of coordination failures. Yeah, I&#8217;ve read <em><a href="https://slatestarcodex.com/2014/07/30/meditations-on-moloch/">Meditations on Moloch</a></em>. It&#8217;s an amazing post; it resonated with me very deeply.&#8221;</p><p><strong>Carl: </strong>&#8220;I was actually going to say <em><a href="https://www.newworldencyclopedia.org/entry/Mot_(Semitic_god)">Mot</a></em>, the Canaanite god of death, bringer of famine and drought.&#8221;</p><p><strong>Allen:</strong> &#8220;Huh. Okay, you got me. Tell me about Mot, then; what does he represent?&#8221;</p><p><strong>Carl:</strong> &#8220;Mot is the god of sterility and lifelessness. To me, he represents the <em>lack of technology</em> in our lives. With technology, we can tame famine, avert drought, and cure disease. We can perform feats that our ancestors would have seen as miracles: flying through the air, and even into space. But we&#8217;re still so so far from achieving the true potential of technology&#8212;and I think of Mot as the personification of what&#8217;s blocking us.</p><p>&#8220;You can see Mot everywhere, when you know what to look for. Whenever a patient lies suffering from a disease that we haven&#8217;t cured yet, that&#8217;s Mot&#8217;s hand at work. Whenever a child grows up in poverty, that&#8217;s because of Mot too. We could have flying cars, and space elevators, and so much more, if it weren&#8217;t for Mot.</p><p>&#8220;Look out your window and you see buildings, trees, people. But if you don&#8217;t see skyscrapers literally miles high, or trees that have been bioengineered to light the streets, or people who are eternally youthful and disease-free, then you&#8217;re not just seeing Earth&#8212;you&#8217;re also seeing Mot. Hell, the fact that we&#8217;re still on this planet, in physical bodies, is a testament to Mot&#8217;s influence. We could be settling the stars, and living in virtual utopias, and even merging our minds, if it weren&#8217;t for Mot.&#8221;</p><p><strong>Allen:</strong> &#8220;Huh. Well, I feel you there; I want all those things too. And you&#8217;re right that god-like technology could solve almost all the issues we face today. But something does feel pretty weird about describing all of this as a single problem, let alone blaming a god of lacking-technology.&#8221;</p><p><strong>Carl:</strong> &#8220;Say more?&#8221;</p><p><strong>Allen:</strong> &#8220;Well, there&#8217;s not any unified force holding back the progress of technology, right? If anything, it&#8217;s the opposite. Absence of advanced technology is the default state, which we need to work hard to escape&#8212;and that&#8217;s difficult not because of any opposition, but just because of entropy.&#8221;</p><p><strong>Carl: </strong>&#8220;What about cases where Mot is being channeled by enemies of progress? For example, when <a href="https://marginalrevolution.com/marginalrevolution/2015/08/is-the-fda-too-conservative-or-too-aggressive.html">bureaucratic regulatory agencies</a> do their best to <a href="https://slatestarcodex.com/2017/08/29/my-irb-nightmare/">stifle scientific research</a>?&#8221;</p><p><strong>Allen:</strong> &#8220;But in those cases you don&#8217;t need to appeal to Mot&#8212;you can just say &#8216;our enemy is overregulation&#8217;. Or if you defined Mot as the god of overregulation, I&#8217;d be totally on board. But you&#8217;re making a much bigger claim than that. The reason we haven&#8217;t uploaded ourselves yet isn&#8217;t that there&#8217;s a force that&#8217;s blocking us, it&#8217;s almost entirely that scientific progress is really really hard!&#8221;</p><p><strong>Carl:</strong> &#8220;Yepp, I agree with all your arguments. And you&#8217;ve probably already guessed where I&#8217;m going with this, but let&#8217;s spell it out: why don&#8217;t these objections to blaming our problems on lack of technology, aka Mot, apply just as much to blaming them on lack of coordination, aka Moloch?&#8221;</p><p><strong>Allen:</strong> &#8220;Yeah, I&#8217;ve been trying to figure that out. First of all, a lot of the intuitive force behind the concept of Moloch comes from really blatant coordination failures, like the ones that Scott lays out in the original post. If you&#8217;re stuck in a situation that nobody wants, then something&#8217;s gone <em>terribly wrong</em>; and when something goes terribly wrong, then it&#8217;s natural to start blaming enemy action.&#8221;</p><p><strong>Carl:</strong> &#8220;There are really blatant examples of lack-of-technology too, though. Look at a wheel. It&#8217;s a literal circle; it&#8217;s hard to imagine any technology that&#8217;s simpler. Yet humans spent millennia gathering crops and carrying loads before inventing it. Or think about cases where we narrowly missed out on transformative breakthroughs. <a href="https://en.wikipedia.org/wiki/Aeolipile">The Romans built toy steam engines</a>&#8212;they just never managed to scale them up to produce an industrial revolution. Getting so close to accelerating a post-scarcity world by two millennia, but just missing, surely counts as something going terribly, tragically wrong. Don&#8217;t these cases demonstrate that Mot&#8217;s presence can be just as blatant as Moloch&#8217;s?&#8221;</p><p><strong>Allen:</strong> &#8220;Well, a big part of both of those stories was the absence of demand. Wheels just weren&#8217;t very useful before there were high-quality roads; and <a href="https://acoup.blog/2022/08/26/collections-why-no-roman-industrial-revolution/">early steam engines just weren&#8217;t very useful in the absence of large coal mines</a>. Of course they both turned out to be very worthwhile in the long term, but that&#8217;s really hard to foresee.&#8221;</p><p><strong>Carl: </strong>&#8220;So you&#8217;re saying that we sometimes need to jump out of a local trap in order to make longer-term technological progress. Remind me, what was your position on understanding local obstacles to progress by anthropomorphizing them as Canaanite gods?&#8221;</p><p><strong>Allen:</strong> &#8220;Okay, fair point. But Moloch isn&#8217;t just an external obstacle&#8212;it&#8217;s also a state of mind. When you pretend that you&#8217;re going to cooperate when you&#8217;re not, or you place your own interests above those of the group, you&#8217;re channeling Moloch. And when enough people do that, societal trust breaks down, and the Molochian dynamics become a self-fulfilling prophecy.&#8221;</p><p><strong>Carl:</strong> &#8220;And when you ridicule people for trying something different, or lobby for legislative barriers to deploying new technology, you&#8217;re channeling Mot. And when enough people do that, society loses faith in positive-sum growth, and progress stagnates. It&#8217;s directly analogous. Come on, what&#8217;s your true objection here?&#8221;</p><p><strong>Allen:</strong> &#8220;I mean, I can&#8217;t fully articulate it. But the ideal of perfect coordination feels much more achievable to me than the ideal of perfect technology. We <em>could</em> just agree to act in a unified way&#8212;it&#8217;s simply a matter of wanting it enough. In other words, saying that lack of technology is responsible for our problems isn&#8217;t very actionable&#8212;you can&#8217;t just magic up technology out of nowhere. But saying that lack of coordination is responsible for our problems is a straightforward step towards convincing people to become more coordinated.&#8221;</p><p><strong>Carl:</strong> &#8220;Actually, the last few centuries could pretty reasonably be described as humanity continually magicking up technology out of nowhere. Of course, scientific and technological progress still takes a lot of work, and a lot of iteration. But when it works, it lets you jump directly to far better outcomes. By contrast, it&#8217;s incredibly difficult to improve things like government competence or social trust&#8212;or even to prevent them from declining. So overall, boosting technological progress is far<em> more</em> actionable than increasing coordination, and we should write off the phrase &#8216;we could just agree&#8217; as a particularly seductive appeal to <a href="https://www.lesswrong.com/posts/fysgqk4CjAwhBgNYT/fake-explanations">magic</a>.&#8221;</p><p><strong>Allen:</strong> &#8220;I do agree that scientific and technological progress has far outstripped progress in governance and coordination. So on an intellectual level, I think you&#8217;ve convinced me that Moloch is no more useful a concept than Mot. But I still don&#8217;t feel like I&#8217;ve <a href="https://www.lesswrong.com/posts/Mc6QcrsbH5NRXbCRX/dissolving-the-question">dissolved the question</a> of why Moloch <em>seems</em> more compelling than Mot. Do you have any explanation for that?&#8221;</p><p><strong>Carl:</strong> &#8220;I think the illusion comes from Scott using a simplistic notion of coordination, as exemplified by his claim that &#8216;the opposite of a trap is a garden&#8230; with a single gardener dictating where everything should go&#8217;. In other words, he implicitly assumes that &#8216;coordinate&#8217; is synonymous with &#8216;centralize power&#8217;. From that perspective, we can view coordination as a single spectrum, with &#8216;Moloch&#8217; at one end and &#8216;just put one guy in charge of everything&#8217; at the other. But the space of possibilities is much richer and more complicated than that, along at least three different dimensions.</p><p>&#8220;Firstly, coordination is complicated in the same way that science is complicated: it requires developing new concepts and frameworks that are totally alien from your current perspective, even if they&#8217;ll seem obvious in hindsight. For most people throughout history, ideas like liberalism, democracy, and free speech were deeply counterintuitive (or, in Scott&#8217;s terminology, <a href="https://slatestarcodex.com/2014/02/23/in-favor-of-niceness-community-and-civilization/">&#8216;terrifying unspeakable Elder Gods&#8217;</a>). In terms of spreading prosperity across the world, the limited liability company was just as important an invention as the steam engine. If you wouldn&#8217;t blame Mot for all the difficulties of labor and locomotion that were eventually solved by steam engines, you shouldn&#8217;t blame Moloch for all the difficulties of trust and incentive alignment that were eventually solved by LLCs.</p><p>&#8220;Secondly, coordination is complicated in the same way that engineering large-scale systems is complicated: there are always just a huge number of practical obstacles and <a href="http://johnsalvatier.org/blog/2017/reality-has-a-surprising-amount-of-detail">messy details</a> to deal with. It took the best part of a century to get from the first commercial steam engine to Watts&#8217; design; and even today, some of the hardest software engineering problems simply involve getting well-understood algorithms to work at much larger scales (like serving search results, or training LLMs). Similarly, when we look at important real-life coordination problems, they&#8217;re very different from toy problems like prisoner&#8217;s dilemmas or tragedies of the commons. Even when there&#8217;s a simple &#8216;headline idea&#8217; for a better equilibrium, actually reaching that equilibrium requires a huge amount of legwork: engaging with different stakeholders, building trust, standardizing communication protocols, creating common knowledge, balancing competing interests, designing agreements, iterating to fix problems that come up, and so on.</p><p>&#8220;Thirdly, coordination is complicated in the same way that security is complicated: you don&#8217;t just need to build effective tools, you need to prevent them from being hijacked and misused. Remember that both fascist and communist despots gained power by appealing to the benefits of cooperation&#8212;&#8216;fascism&#8217; is even named after &#8216;fasces&#8217;, the bundles of sticks that are stronger together than apart. If we&#8217;d truly learned the lessons of history, then categorizing actions as &#8216;cooperating&#8217; versus &#8216;defecting&#8217; would feel as simplistic as categorizing people as &#8216;good&#8217; versus &#8216;evil&#8217;. And in fact many people do sense this intuitively, which is why there&#8217;s so commonly strong resistance to top-down solutions to coordination problems, and why the scientific and engineering problems of building coordination technologies are so tricky.&#8221;</p><p><strong>Allen:</strong> &#8220;I buy that coordination is often far more complicated than it seems. But blaming Moloch for coordination breakdowns still seems valuable insofar as it stops us from just blaming each other, which can disrupt any hope of improvement.&#8221;</p><p><strong>Carl:</strong> &#8220;Yeah, I agree. I think of this in terms of the spectrum from <a href="https://slatestarcodex.com/2018/01/24/conflict-vs-mistake/">conflict theory to mistake theory</a>. Saying that few immoral defectors are responsible for coordination problems is pure conflict theory. The concept of Moloch reframes things so that, instead of &#8216;defectors&#8217; being our enemies, an abstract anthropomorphic entity is our enemy instead. And that&#8217;s progress! But it&#8217;s still partly conflict-theoretic, because it tells us that we just need to identify the enemy and kill it. That biases us towards trying to find &#8216;silver bullets&#8217; which would restore us to our rightful coordinated state. Instead, it&#8217;d be better to lean even further into mistake theory: discord is the default, and to prevent it we need to do the hard work of designing and implementing complicated alien coordination technologies.&#8221;</p><p><strong>Allen:</strong> &#8220;<a href="https://www.thinkingcomplete.com/2018/02/in-defence-of-conflict-theory.html">You shouldn&#8217;t underestimate the value of conflict theory</a>, though. It&#8217;s incredibly good at harnessing people&#8217;s tribal instincts towards actually doing something useful. We can&#8217;t be cold and rational all the time&#8212;we need emotionally salient motivations to get us fired up.&#8221;</p><p><strong>Carl:</strong> &#8220;Right. So I don&#8217;t think we should get rid of Moloch as a rallying cry. But I do think that we should get rid of Moloch as a <em>causal node</em> in our ontologies: as a reason why the world is one way, rather than another. And I think we should be much more careful about terminology like &#8216;coordination failure&#8217; or <a href="https://equilibriabook.com/">&#8216;inadequate equilibria&#8217;</a>, which both mistakenly suggest that there&#8217;s <a href="https://www.lesswrong.com/posts/sbJgv5De6d34eiHWA/judgments-often-smuggle-in-implicit-standards">a binary threshold</a> between enough coordination and not-enough coordination. That&#8217;s like saying that cars which can go faster than 80 miles per hour are &#8216;adequate technology&#8217;, but cars which can&#8217;t are a &#8216;technology failure&#8217;. Maybe that&#8217;s occasionally a useful distinction, but it misses the bigger picture: that they&#8217;re actually very similar on almost all axes, because it takes so much complex technology to build a car at all.</p><p>&#8220;For Scott, there&#8217;s no better temple to Moloch than Las Vegas. But even there, my argument applies. You could look at Vegas and see Moloch&#8217;s hand at work. Or you could see Vegas as a product of the miraculous coordination technology that is modern capitalism&#8212;perhaps an edge case of it, but still an example of its brilliance. Or you could see Vegas as a testament to the wisdom of the constitution: casinos are banned almost everywhere in the US, but for the sake of diversity and robustness it sure seems like there should be at least <em>one </em>major city which allows them.<em> </em>Or you could see Vegas as an example of incredible restraint: there are <a href="http://localroger.com/k5host/casodycs.html">innumerable possible ways</a> to extract money from addled tourists in the desert, and Vegas prevents <em>almost all </em>of them. Or you could see it as a testament to the cooperative instinct inside humans: every day thousands of employees go to work and put in far more effort than the bare minimum it would take to not get fired. Setting aside the concept of Moloch makes it easier to see the sheer scale of coordination all around us, which is the crucial first step towards designing even better coordination technologies.</p><p>&#8216;In Las Vegas, Scott saw Moloch. But in Scott&#8217;s description of Moloch, I see Mot. We can do better than thinking of coordination as war and deicide. We can think of it as science, as engineering, as security&#8212;and as the gradual construction, as we sail down the river, of the ship that will take us across the sea.&#8221;</p><p><em>Holy the sea holy the desert holy the railroad holy the locomotive holy the visions holy the hallucinations holy the miracles <a href="https://www.latimes.com/entertainment-arts/story/2023-09-13/las-vegas-sphere-can-be-bizarre-and-sublime-or-just-an-ad">holy the eyeball</a> holy the abyss!<br>Holy forgiveness! mercy! charity! faith! Holy! Ours! bodies! suffering! magnanimity!<br>Holy the supernatural extra brilliant intelligent kindness of the soul!</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.mindthefuture.info/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Mind the Future! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Techno-humanism is techno-optimism for the 21st century]]></title><description><![CDATA[The more powerful technology becomes, the better we must understand both it and ourselves]]></description><link>https://www.mindthefuture.info/p/techno-humanism-is-techno-optimism</link><guid isPermaLink="false">https://www.mindthefuture.info/p/techno-humanism-is-techno-optimism</guid><dc:creator><![CDATA[Richard Ngo]]></dc:creator><pubDate>Fri, 27 Oct 2023 00:00:24 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/9ad40830-4b81-49a6-a7f9-650b82d24eaa_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Lately I&#8217;ve been reading about the history of economic thought, with the goal of understanding how today&#8217;s foundational ideas were originally developed. My biggest takeaway has been that economics is more of a moving target than I&#8217;d realized. Early economists were studying economies which lacked many of the defining features of modern economies: limited liability corporations, trade unions, depressions, labor markets, capital markets, and so on. The resulting theories were often right for their time; but that didn&#8217;t stop many from being wrong (sometimes disastrously so) for ours.</p><p>That&#8217;s also how I feel about the philosophy of techno-optimism, as recently defended in Marc Andreessen&#8217;s <em><a href="https://a16z.com/the-techno-optimist-manifesto/">Techno-Optimist Manifesto</a></em>. I&#8217;ll start this post with the many things it gets right; then explore why, over the last century, the power of technology itself has left straightforward techno-optimism outdated. I&#8217;ll finish by outlining an alternative to techno-optimism: techno-humanism, which focuses not just on building powerful engines of progress, but also on developing a deep scientific understanding of how they can advance the values we care about most.</p><h3>The triumphs of techno-optimism</h3><p>Here&#8217;s where I agree with Andreessen: throughout almost the entirety of human history, techno-optimism was fundamentally correct as a vision for how humans should strive to improve the world. Technology and markets have given us health and wealth unimaginable to people from previous centuries, despite consistent opposition to them. Even now, in the face of overwhelming historical evidence, we still dramatically underrate the potential of technology to solve problems ranging from <a href="https://twitter.com/AukeHoekstra/status/1064529619951513600">climate change</a> to neglected tropical diseases to extreme poverty and many more. This skepticism holds back not only speculative technologies but also ones that are right in front of us: miraculous solutions to huge problems (like vaccines for <a href="https://en.wikipedia.org/wiki/Moderna_COVID-19_vaccine#Original_version">COVID</a> and <a href="https://marginalrevolution.com/marginalrevolution/2023/10/what-is-an-emergency-the-case-of-rapid-malaria-vaccination.html">malaria</a>, and <a href="https://denovo.substack.com/p/gene-drives-why-the-wait">gene drives</a> for mosquito eradication) are regularly stalled by political or bureaucratic obstructionism, costing millions of lives. The RTS,S malaria vaccine spent <em><a href="https://worksinprogress.co/issue/why-we-didnt-get-a-malaria-vaccine-sooner/">twenty-three years</a></em><a href="https://worksinprogress.co/issue/why-we-didnt-get-a-malaria-vaccine-sooner/"> bogged down in clinical trials</a>, with <em>six years</em> of that delay caused by WHO &#8220;precautions&#8221; even after other regulators had already approved it. Fighting against the ideologies and institutional practices which cause tragedies like these is an incredibly important cause.</p><p>We often similarly underrate the other core tenet of techno-optimism: individual liberty. This was a radical principle when Enlightenment thinkers and leaders enshrined it&#8212;and, unfortunately, it remains a radical principle today, despite having proven it worth over and over again. It turns out that free markets can lead to <a href="https://en.wikipedia.org/wiki/Four_Asian_Tigers">near-miraculous prosperity</a>; that free speech is the strongest enemy of tyranny; and that open-source projects (whether <a href="https://en.wikipedia.org/wiki/Linux">building</a> <a href="https://en.wikipedia.org/wiki/Open_standard">software</a> or <a href="https://www.wikipedia.org/">building</a> <a href="https://en.wikipedia.org/wiki/Open-source_intelligence">knowledge</a>) can be extraordinarily successful. Yet all of this is constantly threatened by the creeping march of centralization and overregulation. Opponents of liberty always have high rhetoric about the near-term benefits, but fail to grapple with the extent to which centralized power is inevitably captured or subverted, sometimes with horrific consequences. To defeat that creep requires a near-pathological focus on freedom&#8212;which, little by little, has added up to a far better world. Even an ideally benevolent government simply couldn&#8217;t compete with billions of humans using the tools available to them to improve their own lives and the lives of those they interact with, including by discovering solutions that no central planner would have imagined.</p><p>Not only is techno-optimism vastly underappreciated, it&#8217;s often dismissed with shockingly bad arguments that lack understanding of basic history or economics. Because it&#8217;s so hard to viscerally understand <a href="https://ourworldindata.org/problems-and-progress">how much better the world has gotten</a>, even its most cogent critics fail to grapple with the sheer <em>scale</em> of the benefits of techno-optimism. Those benefits accrue to billions of people&#8212;and they aren&#8217;t just incidental gains, but increases in health and wealth that would have been inconceivable a few centuries ago. Familiarity with the past is, in this case, the best justification for optimism about the future: almost all of the barriers to progress we face today pale in comparison to the barriers that we&#8217;ve already overcome.</p><p>I say all this to emphasize that I very viscerally feel the hope and the beauty of techno-optimism. The idea that progress and knowledge can grow the pie for everyone is a stunningly powerful one; so too is the possibility that we live in a world where the principles of freedom and openness would triumph over any obstacle, if we only believed in them. And so when I criticize techno-optimism, I do so not from a place of scorn, but rather from a place of wistfulness. I wish I could support techno-optimism without reservations.</p><h3>Three cracks in techno-optimism</h3><p>But techno-optimism is not the right philosophy for the 21st century. Over the last century in particular, cracks have been developing in the techno-optimist narrative. These are not marginal cracks&#8212;not the whataboutism with which techno-optimism is usually met. These are cracks which, like the benefits of techno-optimism, play out at the largest of scales. I&#8217;ll talk about three: war, exploitation, and civilizational vulnerability.</p><p>One: the increasing scale of war. World War 1 was a bloody, protracted mess because of the deployment of new technologies like machine guns and rapid-firing artillery. World War 2 was even worse, with whole cities firebombed and nuked, amidst the industrial-scale slaughter of soldiers and civilians alike. And the Cold War was more dangerous still, leaving the world teetering on the brink of global nuclear war. Mutually Assured Destruction wasn&#8217;t enough to prevent close calls; our current civilization owes its existence to the courage of <a href="https://en.wikipedia.org/wiki/Stanislav_Petrov">Stanislav Petrov</a>, <a href="https://en.wikipedia.org/wiki/Vasily_Arkhipov">Vasily Arkhipov</a>, and probably others we don&#8217;t yet know about. So you cannot be a full-throated techno-optimist without explaining how to reliably avoid catastrophe from the weapons that techno-optimists construct. And no such explanation exists yet, because so far we have been avoiding nuclear holocaust through luck and individual heroism. <strong>When the </strong><em><strong>next</strong></em><strong> weapon with planetary-scale destructive capabilities is developed, as it inevitably will be, we need far more robust mechanisms preventing it from being deployed.</strong></p><p>Two: the increasing scale of exploitation. Technology allows the centralization of power, and the use of it to oppress the less powerful at far larger scales than before. Historical examples abound&#8212;most notably the Atlantic slave trade and the mass deaths of civilians under 20th-century communist and fascist regimes. But since it&#8217;s too easy to chalk these up as mistakes of the past which we&#8217;re now too enlightened to repeat, I&#8217;ll focus on an example that continues today: the mass torture of animals in factory farms. This started less than a century ago, in response to advances in logistics and disease-reduction technologies. Yet it has grown at a staggering scale since then: <a href="https://ourworldindata.org/how-many-animals-are-factory-farmed">the number of animals killed in factory farms </a><em><a href="https://ourworldindata.org/how-many-animals-are-factory-farmed">every year</a></em> is comparable to the number of humans who have ever lived. <strong>The sheer scale of this suffering forces any serious moral thinker to ask: how can we stop it as soon as possible</strong><em><strong>? </strong></em><strong>And how can we ensure that nothing like it ever happens again?</strong></p><p>Three: increasing vulnerability to reckless or malicious use of technology. Markets and technology have made the world far more robust in many ways. We should be deeply impressed by how supply chains which criss-cross the world remained largely intact even at the height of COVID. But there are other ways in which the world has become more vulnerable to our mistakes. The most prominent is gain-of-function research on pathogens. Not only are we now capable of engineering global pandemics, but China- and US-funded scientists probably <em>did</em>. <strong>And there&#8217;s no reason that the next one couldn&#8217;t be far worse, especially if released deliberately.</strong> Nor is bioengineering the sole domain in which (deliberate or accidental) offense may overwhelm defense to a catastrophic degree; other possibilities include geoengineering, <a href="https://www.narrativeark.xyz/p/jacob-on-the-precipice">asteroid redirection</a>, nanotechnology, and <a href="https://nickbostrom.com/papers/vulnerable.pdf">new fields that we haven&#8217;t yet imagined</a>.</p><p>These cracks all hurt the techno-optimist worldview. But I don&#8217;t think they take it down; techno-optimists have partial responses to all of them. The world has become far more peaceful as a result of our increasing wealth&#8212;even if wars can be far bigger, they&#8217;re also far rarer now. Technology will produce tastier meat substitutes, and cheap clean meat, and when it does factory farming will end, and humanity will look back on it in horror. And while we can&#8217;t yet robustly prevent the accidental or deliberate deployment of catastrophically powerful weapons (like nuclear weapons or engineered pandemics), we might yet stumble through regardless, as we have so far. So if the cracks above were the main problems with techno-optimism, I would still be a techno-optimist. I would have compunctions about the development of even more powerful weapons, and about all the sentient beings (whether farmed animals or wild animals or future people or artificial minds) which remained outside our circle of concern. But I&#8217;d still believe that any &#8220;cure&#8221; to those problems which undermined techno-optimism would be worse than the disease.</p><p>Yet I am not a techno-optimist, because we are about to leave the era in which humans are the most intelligent beings on earth. And the era of artificial intelligence will prise open these cracks until the deeper holes in the techno-optimist worldview become clear.</p><h3>AI and techno-optimism</h3><p>Until now, the primary forces shaping the world have been human intelligence and human agency, which allow us to envisage the outcomes we want, identify paths towards them, and consistently pursue them. Soon, artificial intelligence and artificial agency will match ours; and soon after that AI will far surpass us. I&#8217;ll describe at a very high level how I expect this to play out, in two stages.</p><p><strong>Firstly: AIs used as tools will supercharge the dynamics I described above. </strong>The benefits of individual freedom will expand, as individuals become far more empowered, and can innovate far faster. But so too will the cracks enabled by technology: war, exploitation, and civilizational vulnerability. Which effect will be bigger? I simply don&#8217;t know; and nobody else does either. Predicting the offense-defense balance of new technology is incredibly difficult, because it requires accounting for all the different uses that future innovators will come up with. What we <em>can </em>predict is that the stakes will be higher than ever before: 21st-century technology could magnify the worst disasters of the 20th century, like world wars and totalitarianism.</p><p>Perhaps, despite that, the best path is still to rush ahead: to prioritize building new technology now, and assume that we can sort out the rest later. This is a very fragile strategy, though: even if you (and your government) will use it responsibly, many others will not. And there&#8217;s still relatively little attention and effort focused on trying to avert the large-scale risks of new technologies directly. So the techno-optimist response to risks like the ones I&#8217;ve described above is shallow: in theory it&#8217;s opposed, but in practice it may well make things worse.</p><p><strong>Secondly:</strong> <strong>we&#8217;ll develop AI agents with values of their own</strong>. As AIs automate more and more complex tasks, over longer and longer timeframes, they&#8217;ll need to make more and more value-laden judgment calls about which actions and outcomes to favor. Eventually, viewing them as tools will be clearly inadequate, and we&#8217;ll need to treat them as agents in their own right&#8212;agents whose values may or may not align with our own. Note that this might only happen once they&#8217;ve significantly surpassed human intelligence&#8212;but given how fast progress in the field has been, this is something we should be planning for well in advance.</p><p>Humans who are free to make their own decisions tend to push the world to be better in terms of our values&#8212;that&#8217;s the techno-optimist position. But artificial agents who are making many decisions will push the world to be better in terms of <em>their</em> values. This isn&#8217;t necessarily a bad thing. Humans are hypocritical, short-sighted, often selfish, sometimes sadistic&#8212;and so it&#8217;s possible that AIs will be better custodians of our moral values than we are. We might train them to be wise, and kind, and consistently nudge us towards a better world&#8212;not by overriding human judgments, but rather by serving as teachers and mentors, giving us the help we need to become better people and build a better civilization. I think this is the most likely outcome, and it&#8217;s one we should be incredibly excited about.</p><p>But it&#8217;s also possible that AIs will develop alien values which conflict with those of humans. If so, when we give them instructions, they&#8217;ll appear to work towards our ends, but consistently make choices which bolster their power and undermine our own. Of course, AIs will start off with very little power&#8212;we&#8217;ll be able to shut them down whenever we detect misbehavior. But AIs will be able to <a href="https://www.narrativeark.xyz/p/one">coordinate with each other</a> far better than humans can, communicate in ways we&#8217;re <a href="https://www.schneier.com/blog/archives/2023/06/ai-generated-steganography.html">incapable of interpreting</a>, and carry out tasks that we&#8217;re incapable of overseeing. They&#8217;ll become embedded in our economies in ways that amplify the effects of their decisions: hundreds of millions of people will use copies of a single model on a daily basis. And as AIs become ever more intelligent, the risks will grow. When agents are far more capable than the principals on whose behalf they act, principal-agent problems can become very severe. When it comes to superhuman AI agents, we should think about the risks less in terms of financial costs, or even human costs, and more in terms of political instability: careless principals risk losing control entirely.</p><p>How plausible is this scenario, really? That&#8217;s far too large a question to address here; instead, see <a href="https://managing-ai-risks.com/">this open letter</a>, <a href="https://arxiv.org/abs/2209.00626">this position paper</a> and <a href="https://course.aisafetyfundamentals.com/alignment">this curriculum</a>. But although there&#8217;s disagreement on many details, there&#8217;s a broad consensus that we simply don&#8217;t understand how AI motivations develop, or how those motivations generalize to novel situations. And although there&#8217;s widespread disagreement about the trajectory of AI capabilities, what&#8217;s much less controversial is that when AI <em>does</em> significantly surpass human capabilities, we should be wary of putting it in positions where it can accumulate power unless we have very good reasons to trust it.</p><p>It&#8217;s also worth noting that Andreessen's version of techno-optimism draws heavily from Nick Land&#8217;s philosophy of <em>accelerationism</em>, which expects us to lose control, and is actively excited about it. &#8220;Nothing human makes it out of the near future&#8221;, <a href="http://www.ccru.net/swarm1/1_melt.htm">Land writes</a>, <a href="https://web.archive.org/web/20170110095648/http://www.xenosystems.net/pythia-unbound/">and celebrates</a>: &#8220;The planet has been run by imbeciles for long enough.&#8221; I read those words and shudder. Land displays a deep contempt for the things that make us ourselves; his philosophy is fundamentally anti-humanist (as Scott Alexander argues more extensively in his <em><a href="https://slatestarcodex.com/2014/07/30/meditations-on-moloch/">Meditations on Moloch</a></em>). And while his position is extreme, it reflects a problem at the core of techno-optimism: the faster you go, the less time you have to orient to your surroundings, and the easier it is to diverge from the things you actually care about. Nor does speed even buy us very much. On a cosmic scale, we have plenty of time, plenty of resources available to us, and plenty of space to expand. The one thing we don&#8217;t have is a reset button for if we lose control.</p><h3>AI and techno-humanism</h3><p>So we need a philosophy which combines an appreciation for the incredible track record of both technology and liberty with a focus on ensuring that they actually end up promoting our values. This mindset is common amongst Effective Altruists&#8212;but Effective Altruism is an agglomeration of many very different perspectives, drawn together not by a shared vision about the future but by shared beliefs about how we&#8217;re obliged to act. I&#8217;d like to point more directly to an overarching positive vision of what humanity should aim towards. Transhumanism offers one such vision, but it&#8217;s so radically individualist that it glosses over the relationships and communities that are the most meaningful aspects of most people&#8217;s lives. (Note, for example, how Bostrom&#8217;s <em><a href="https://nickbostrom.com/utopia">Letter from Utopia</a></em> barely mentions the existence of other people; while his <a href="https://nickbostrom.com/old/transhumanism">introduction to transhumanism</a> relegates relationships to the final paragraph of the postscript.) So I&#8217;ll borrow a term coined by Yuval Harari, and call the philosophy that I&#8217;ve been describing <em><a href="https://www.wired.com/2017/02/yuval-harari-tech-is-the-new-religion/">techno-humanism</a></em>.</p><p>Harari describes techno-humanism as an ideology focused on upgrading humans to allow our actions and values to remain relevant in an AI-dominated future. I broadly agree with his characterization (and will explore it more in future posts), but both the benefits and the risks of re-engineering human brains are still a long way away. On a shorter timeframe, I think a different way of &#8220;upgrading&#8221; our minds is more relevant: <strong>developing a deep </strong><em><strong>understanding</strong></em><strong> of our values and how technology can help achieve them</strong>. Flying cars and rockets are cool, but the things we ultimately care about are far more complex and far more opaque to us. We understand machines but not minds; algorithms but not institutions; economies but not communities; prices but not values. Insofar as we face risks from poor political decisions, or misaligned AIs, or society becoming more fragile, from a techno-humanist perspective it&#8217;s because we lack the understanding to do better.</p><p>This is a standard criticism of techno-optimism&#8212;and usually an unproductive one, since the main alternative typically given is to defer to academic humanities departments which produce far more ideology than understanding. But techno-humanism instead advocates trying to develop this understanding using the most powerful tools we have: science and technology. To give just a few examples of what this could look like: studying artificial minds and their motivations will allow us to build more trustworthy AIs, teach us about our own minds, and help us figure out how the two should interface. The internet should be full of experiments in how humans can interact&#8212;<a href="https://www.astralcodexten.com/p/prediction-market-faq">prediction markets</a>, <a href="https://www.thinkingcomplete.com/2020/04/melting-democracy.html">delegative democracies</a>, <a href="https://slatestarcodex.com/2019/12/09/2019-adversarial-collaboration-entries/">adversarial collaborations</a>, and many more&#8212;whose findings can then improve existing institutions. We should leverage insights from domains like game theory, voting theory, network theory, and bargaining theory to help understand and reimagine politics&#8212;starting with <a href="https://electionscience.org/">better voting systems</a> and ideally going far further. And we should design sophisticated protocols for testing and verifying the knowledge that will be generated by AIs, so that we can avoid replicating the replication crises that currently plague many fields.</p><p>This may sound overly optimistic. But some of the most insightful fields of knowledge&#8212;like economics and evolutionary biology&#8212;uncovered deep structure in incredibly complex domains via identifying just a few core insights. And we&#8217;ll soon have AI assistance in uncovering patterns and principles that would otherwise be beyond our grasp. Meanwhile, platforms like Wikipedia and Stack Overflow have been successful beyond all expectations; it&#8217;s likely that there are others which could be just as valuable, if only there were more people trying to build them. So I think that the techno-humanist project has a huge amount of potential, and will only become more important over time.</p><h3>Balancing the tradeoffs</h3><p>So far I&#8217;ve described techno-humanism primarily in terms of advances that techno-optimists would also be excited about. But inevitably, there will also be clashes between those who prioritize avoiding the risks I&#8217;ve outlined and those who don&#8217;t. From a techno-optimist perspective&#8212;a perspective that has proven its worth over and over again during the last few centuries&#8212;slowing down technological progress has a cost measured in millions of lives. This is an <a href="https://marginalrevolution.com/marginalrevolution/2021/01/the-invisible-graveyard-is-invisible-no-more.html">invisible graveyard</a> which is brushed aside even by the politicians and bureaucrats most responsible for it; no wonder many techno-optimists feel driven to push for unfettered acceleration.</p><p>But from a techno-humanist perspective, reckless technological progress has a cost measured in expected fractions of humanity&#8217;s entire future. Human civilization used to be a toddler: constantly tripping over and hurting itself, but never putting itself in any real danger. Now human civilization is a teenager: driving fast, experimenting with <a href="https://slatestarcodex.com/2014/12/17/the-toxoplasma-of-rage/">mind-altering substances</a>, and genuinely capable of wrecking itself. We don&#8217;t need the car to go faster&#8212;it&#8217;s already constantly accelerating. Instead, we need to ensure that the steering wheel and brakes are working impeccably&#8212;and that we&#8217;re in a fit state to use them to prevent non- or anti-human forces controlling the direction of our society.</p><p>How can people who are torn between these two perspectives weigh them against each other? On a purely numerical level, humanity&#8217;s potential to build an intergalactic civilization renders &#8220;fractions of humanity&#8217;s future&#8221; bigger by far. But that math is too blas&#233;&#8212;it&#8217;s the same calculation that can, and often has, been used to justify centralization of power, totalitarianism, and eventual atrocities. And so we should be extremely, extremely careful when using arguments that appeal to &#8220;humanity&#8217;s entire future&#8221; to override time-tested principles. That doesn&#8217;t imply that we should never do so. But wherever possible, techno-optimists and techno-humanists should try to cooperate rather than fight. After all, techno-humanism is also primarily about making progress: specifically, the type of progress that will be needed to defuse the crises sparked by other types of progress. The disagreement isn&#8217;t about where we should end up; it&#8217;s about <a href="https://en.wikipedia.org/wiki/Differential_technological_development">the ordering of steps along the way</a>.</p><p>The two groups should also challenge each other to do better in areas where we disagree, so that we can eventually reach a synthesis. One challenge that techno-humanists should pose to techno-optimists: <strong>be more broadly ambitious!</strong> We know that technology and markets can work incredibly well, and have a near-miraculous ability to overcome obstacles. And so it&#8217;s easy and natural to see them as solutions to all the challenges confronting us. But the most courageous and ambitious version of techno-optimism needs to grapple with the possibility that our downfall will come not from <em>lack </em>of technology, but rather <em>overabundance</em> of technology&#8212;and the possibility that to prevent it we need progress on the things that have historically been hardest to improve, like the quality of political decision-making. In other words, techno-humanism aims to harness human ingenuity (and technological progress) to make &#8220;steering wheels&#8221; and &#8220;brakes&#8221; more sophisticated and discerning, rather than the blunt cudgels that they often are today.</p><p>My other challenge for techno-optimists: <strong>be optimistic not just about the benefits of technological growth, but also about its robustness</strong>. The most visceral enemy of techno-optimism is stagnation. And it&#8217;s easy to see harbingers of stagnation all around us: overregulation, NIMBYism, illiberalism, degrowth advocacy, and so on. But when we zoom out enough to see the millennia-long exponential curve leading up to our current position, it seems far less plausible that these setbacks will actually derail the long-term trend, no matter how outrageous the latest news cycle is. On the contrary: taking AGI seriously implies that innovation is on the cusp of speeding up dramatically, as improvements generated by AIs feed back into the next generation of AIs. In light of that, a preference for slower AI progress is less like Luddism, and more like carefully braking as we approach a sharp bend in the road.</p><p>Techno-optimists should challenge techno-humanists to improve as well. I can&#8217;t speak for them, but my best guess for the challenges that techno-optimists should pose to techno-humanists:</p><ol><li><p><strong>Techno-humanists need to articulate a compelling positive vision</strong>, one which inspires people to fight for it. Above, I&#8217;ve listed some ideas which have potential to improve our collective understanding and decision-making abilities; but there&#8217;s far more work to be done in actually fleshing out those ideas, and pushing towards their implementation. And even if we succeeded, what then? What would it actually look like for humanity to make consistently sensible decisions, and leverage technology to promote our long-term flourishing? Knowing that would allow us to better steer towards those good futures.</p></li><li><p><strong>Techno-humanists should grapple more seriously with the incredible track record of techno-optimism</strong>. Throughout history, people have consistently dramatically underrated how valuable scientific and technological progress can be. That&#8217;s not a coincidence at all, because characterizing which breakthroughs are possible is often a big chunk of the work required to actually make those breakthroughs. Nor is it a coincidence that people dramatically underrate the value of liberty&#8212;decentralization works so well precisely because there are so many things that central planners can&#8217;t predict. So even if you find my arguments above compelling, we should continue to be very wary of falling into the same trap.</p></li></ol><p>The purpose of this blog is to meet those challenges. Few of the ideas in this post are original to me, but they lay the groundwork for future posts which will explore more novel territory. My next post will build on them by arguing that an understanding-first approach is feasible even when it comes to the biggest questions facing us&#8212;that we can look ahead to see the broad outlines of where humanity is going, and use that knowledge to steer towards a future that is both deeply human and deeply humane.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.mindthefuture.info/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Mind the Future! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item></channel></rss>