<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:media="http://search.yahoo.com/mrss/"><channel><title><![CDATA[Cold Takes]]></title><description><![CDATA[In-depth post each Tuesday. Other posts on some other days.]]></description><link>https://www.cold-takes.com/</link><image><url>https://www.cold-takes.com/favicon.png</url><title>Cold Takes</title><link>https://www.cold-takes.com/</link></image><generator>Ghost 4.16</generator><lastBuildDate>Mon, 27 Sep 2021 11:24:19 GMT</lastBuildDate><atom:link href="https://www.cold-takes.com/rss/" rel="self" type="application/rss+xml"/><ttl>60</ttl><item><title><![CDATA[Cold Links: assorted fun basketball stuff]]></title><description><![CDATA[<!--kg-card-begin: html-->
<p>
<a href="https://www.ccn.com/superman-returns-dwight-howards-5-top-slams-from-dunk-contests-past/">Dwight Howard had some great slam dunk contest dunks</a>. I really love how much fun everyone is having in these videos, especially the announcers. Verbatim quote from the announcer in the last video: &quot;I&apos;m leaving the building! I quit my job! I&apos;ve never seen anything</p>]]></description><link>https://www.cold-takes.com/cold-links-assorted-fun-basketball-stuff/</link><guid isPermaLink="false">614d56e6a932ea004866faa2</guid><dc:creator><![CDATA[Holden Karnofsky]]></dc:creator><pubDate>Fri, 24 Sep 2021 19:10:19 GMT</pubDate><content:encoded><![CDATA[<!--kg-card-begin: html-->
<p>
<a href="https://www.ccn.com/superman-returns-dwight-howards-5-top-slams-from-dunk-contests-past/">Dwight Howard had some great slam dunk contest dunks</a>. I really love how much fun everyone is having in these videos, especially the announcers. Verbatim quote from the announcer in the last video: &quot;I&apos;m leaving the building! I quit my job! I&apos;ve never seen anything like that, in my life, it is over!&quot;
</p>
<p>
Nikola Jokic, the <a href="https://deadspin.com/nikola-jokic-is-the-god-of-the-hideous-game-winner-1833320080">least graceful superstar I can think of</a>. I could watch him all day.
</p>
<p>
I&apos;ve seen a lot of sports highlight videos in my life, but<a href="http://deadspin.com/heres-a-compilation-of-kevin-loves-outlet-passes-so-f-1465521627"> this compilation of Kevin Love&apos;s outlet passes</a> is special.
</p>
<p>
<a href="https://photos.app.goo.gl/P7LDgdqB8jqxyb4T9">How exactly did Anthony Davis lose his balance and fall over here</a>? (Video, click play)
</p>
<p>
<a href="http://www.youtube.com/watch?v=vaZIAXJJDKQ">The worst call by an NBA ref ever</a>.
</p>
<p>
<a href="https://deadspin.com/tireless-steph-curry-helpfully-illustrates-seth-currys-1834908263">Great video</a> of Steph Curry running around a lot during a play. &quot;That&#x2019;s a damn Family Circus comic.&quot;
</p>
<p>
This is absolutely the most painful<a href="https://deadspin.com/the-pelicans-gave-the-suns-a-masterclass-on-the-dumbest-1833356994"> choking away of a game at the very end</a> I can recall seeing.
</p><!--kg-card-end: html--><!--kg-card-begin: html-->

<p><div style="display:flex; justify-content:center; margin: 0 auto;">

        <span style="margin: 10px;"><a href="https://api.addthis.com/oexchange/0.8/forward/twitter/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fcold-links-assorted-fun-basketball-stuff%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20Cold%20Links%3A%20assorted%20fun%20basketball%20stuff&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-twitter-square.png" border="0" alt="Twitter"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/facebook/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fcold-links-assorted-fun-basketball-stuff%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20Cold%20Links%3A%20assorted%20fun%20basketball%20stuff&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-facebook-square.png" border="0" alt="Facebook"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/reddit/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fcold-links-assorted-fun-basketball-stuff%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20Cold%20Links%3A%20assorted%20fun%20basketball%20stuff&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-reddit-square.png" border="0" alt="Reddit"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/menu/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fcold-links-assorted-fun-basketball-stuff%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20Cold%20Links%3A%20assorted%20fun%20basketball%20stuff&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-addthis-square.png" border="0" alt="More"></a></span>
        </div>
<center><p id="survey"><a href="https://www.cold-takes.com/cold-links-assorted-fun-basketball-stuff/#subscribe" target="_blank"><button id="footer-subscribe" class="button">Subscribe</button></a>&#xA0;<a href="https://www.guidedtrack.com/programs/4kal2ue/run?posttitle=Cold%20Links%3A%20assorted%20fun%20basketball%20stuff" target="_blank"><button class="button" id="Survey">Feedback</button></a><!--No Forum post--></p></center>
<!--kg-card-end: html--><!--kg-card-begin: html--></p><p style="font-size:1%">For email filter: florpschmop</p><!--kg-card-end: html-->]]></content:encoded></item><item><title><![CDATA[The Most Important Century (in a nutshell)]]></title><description><![CDATA[<!--kg-card-begin: html--><p>
I&apos;ve spent most of my career looking for ways to do as much good as possible, per unit of money or time. I worked on finding evidence-backed charities working on global health and development (co-founding <a href="https://www.givewell.org/">GiveWell</a>), and later moved into philanthropy that takes <a href="https://www.openphilanthropy.org/blog/hits-based-giving">more risks</a> (co-founding <a href="https://www.openphilanthropy.org/">Open Philanthropy</a></p>]]></description><link>https://www.cold-takes.com/the-most-important-century-in-a-nutshell/</link><guid isPermaLink="false">6149572c2167dd003d6ab3e2</guid><dc:creator><![CDATA[Holden Karnofsky]]></dc:creator><pubDate>Thu, 23 Sep 2021 16:52:24 GMT</pubDate><media:content url="https://www.cold-takes.com/content/images/2021/09/twitter-img-for-mic-nutshell-2.png" medium="image"/><content:encoded><![CDATA[<!--kg-card-begin: html--><img src="https://www.cold-takes.com/content/images/2021/09/twitter-img-for-mic-nutshell-2.png" alt="The Most Important Century (in a nutshell)"><p>
I&apos;ve spent most of my career looking for ways to do as much good as possible, per unit of money or time. I worked on finding evidence-backed charities working on global health and development (co-founding <a href="https://www.givewell.org/">GiveWell</a>), and later moved into philanthropy that takes <a href="https://www.openphilanthropy.org/blog/hits-based-giving">more risks</a> (co-founding <a href="https://www.openphilanthropy.org/">Open Philanthropy</a>).
</p>
<p>
Over the last few years - thanks to general dialogue with the <a href="https://en.wikipedia.org/wiki/Effective_altruism">effective altruism</a> community, and extensive research done by <a href="https://www.cold-takes.com/roadmap-for-the-most-important-century-series/#acknowledgements">Open Philanthropy&apos;s Worldview Investigations team</a> - I&apos;ve become convinced that humanity as a whole faces huge risks and opportunities this century. Better understanding and preparing for these risks and opportunities is where I am now focused. 
</p>
<p>
This piece will summarize a <a href="https://www.cold-takes.com/roadmap-for-the-most-important-century-series/">series</a> of posts on why I believe we could be in the <strong>most important century of all time for humanity</strong>. It gives a short summary, key post(s), and sometimes key graphics for 5 basic points:
</p>
<ul>

<li><strong>The long-run future is radically unfamiliar.</strong> Enough advances in technology could lead to a long-lasting, galaxy-wide civilization that could be a radical utopia, dystopia, or anything in between.

</li><li><strong>The long-run future could come much faster than we think,</strong> due to a possible AI-driven productivity explosion. 

</li><li>The relevant kind of <strong>AI looks like it will be developed this century</strong> - making this century the one that will initiate, and have the opportunity to shape, a future galaxy-wide civilization.

</li><li>These claims seem too &quot;wild&quot; to take seriously. But there are a lot of reasons to think that <strong>we live in a wild time, and should be ready for anything.</strong>

</li><li>We, the people living in this century, have the chance to have a huge impact on huge numbers of people to come - if we can make sense of the situation enough to find helpful actions. But right now, <strong>we aren&apos;t ready for this. </strong>
</li>
</ul>
<p>
This thesis has a wacky, sci-fi feel. It&apos;s very far from where I expected to end up when I set out to do as much good as possible. 
</p>
<p>
But part of the mindset I&apos;ve developed through GiveWell and Open Philanthropy is being open to strange possibilities, while critically examining them with as much rigor as possible. And after a lot of investment in examining the above thesis, I think it&apos;s likely enough that the world urgently needs more attention on it.
</p>
<p>
By writing about it, I&apos;d like to either get more attention on it, or gain more opportunities to be criticized and change my mind.
</p>
<h2 id="we-live-in-a-wild-time-and-should-be-ready-for-anything">We live in a wild time, and should be ready for anything</h2>


<p>
Many people find the &quot;most important century&quot; claim too &quot;wild&quot;: a radical future with advanced AI and civilization spreading throughout our galaxy may happen <em>eventually</em>, but it&apos;ll be more like 500 years from now, or 1,000 or 10,000. (Not this century.)
</p>
<p>
These longer time frames would put us in a <em>less</em> wild position than if we&apos;re in the &quot;most important century.&quot; But in the scheme of things, <strong>even if galaxy-wide expansion begins 100,000 years from now, that still means we live in an extraordinary era</strong>- the tiny sliver of time during which the galaxy goes from nearly lifeless to largely populated. It means that out of a staggering number of persons who will ever exist, we&apos;re among the first. And that out of hundreds of billions of stars in our galaxy, ours will produce the beings that fill it.
</p>
<p>


</p><p><figure class="kg-card kg-image-card kg-width-wide"><img src="https://www.cold-takes.com/content/images/2021/09/fermi-agnostic.png" width="1036" alt="The Most Important Century (in a nutshell)"></figure><figcaption>More at <a href="https://www.cold-takes.com/all-possible-views-about-humanitys-future-are-wild/">All Possible Views About Humanity&apos;s Future Are Wild</a></figcaption></p>

<p></p>
<p>
Zooming in, we live in a special century, not just a special era. We can see this by looking at how fast the economy is growing. It doesn&apos;t <em>feel</em> like anything special is going on, because for as long as any of us have been alive, the world economy has grown at a few percent per year:
</p>
<p>


</p><p>
<figure class="kg-card kg-image-card kg-width-wide"><img src="https://www.cold-takes.com/content/images/size/w1000/2021/07/business-as-usual.png" alt="The Most Important Century (in a nutshell)" width="1036"></figure>

</p>

<p>
However, when we zoom out to look at history in greater context, we see a picture of an unstable past and an uncertain future:
</p>
<p>

</p><p><figure class="kg-card kg-image-card kg-width-wide"><img src="https://www.cold-takes.com/content/images/size/w1000/2021/08/long-vs-short-view.png" alt="The Most Important Century (in a nutshell)" width="1036"><figcaption>More at <a href="https://www.cold-takes.com/this-cant-go-on/">This Can&apos;t Go On</a></figcaption></figure></p>

<!--<p><figure class="kg-card kg-image-card kg-width-wide"><img src="https://www.cold-takes.com/content/images/size/w1000/2021/09/long-vs-short-view-1.jpg" alt="Graphic illustrating that when you zoom in on the 'this can't go on' chart, you get the 'business as usual chart.'" width=1036></figure></p>-->


<p>
    <strong>We&apos;re currently living through the fastest-growing time in history.</strong> This rate of growth hasn&apos;t gone on long, and can&apos;t go on indefinitely (there aren&apos;t enough atoms in the galaxy to sustain this rate of growth for even another 10,000 years). And if we get <em>further acceleration</em> in this rate of growth - in line with historical acceleration - we could reach the limits of what&apos;s possible more quickly: within this century.
</p>
<p>
To recap:
</p>
<ul>

<li>The last few millions of years - with the start of our species - have been more eventful than the previous several billion. 

</li><li>The last few hundred years have been more eventful than the previous several million. 

</li><li>If we see another accelerator (as I think AI could be), the next few decades could be the most eventful of all.
</li>
</ul>
<!--<p>[Here there is an interactive graphic that isn't compatible with email. You can see it at the <a href="https://www.cold-takes.com/p/8bb231a4-96d6-4b5d-a013-38c4a6c688fb#InteractiveTimelines">web version</a>.]</p>-->

<p id="InteractiveTimelines"><figure class="kg-card kg-image-card kg-width-wide">
    <iframe src="https://www.cold-takes.com/assets/files/mic-animated-timeline-2.html" width="1036" height="800"></iframe><figcaption>More info about these timelines at <a href="https://www.cold-takes.com/all-possible-views-about-humanitys-future-are-wild/">All Possible Views About Humanity&apos;s Future Are Wild</a>, <a href="https://www.cold-takes.com/this-cant-go-on/">This Can&apos;t Go On</a>, and <a href="https://www.cold-takes.com/forecasting-transformative-ai-the-biological-anchors-method-in-a-nutshell/" style="color: var(--brand-color) !important;"><strong>Forecasting Transformative AI: Biological Anchors</strong></a>, respectively.</figcaption>
    </figure></p>
<p>
Given the times we live in, we need to be open to possible ways in which the world could change quickly and radically. Ideally, we&apos;d be a bit <em>over</em>-attentive to such things, like putting safety first when driving. But today, such possibilities get little attention.
</p>
<p>
Key pieces:
</p>
<ul>

<li><a href="https://www.cold-takes.com/all-possible-views-about-humanitys-future-are-wild/">All Possible Views About Humanity&apos;s Future Are Wild</a>

</li><li><a href="https://www.cold-takes.com/this-cant-go-on/">This Can&apos;t Go On</a>
</li>
</ul>
<h2 id="the-long-run-future-is-radically-unfamiliar">The long-run future is radically unfamiliar</h2>


<p>
Technology tends to increase people&apos;s control over the environment. For a concrete, easy-to-visualize example of what things could look like if technology goes far enough, we might imagine a technology like &quot;digital people&quot;: fully conscious people &quot;made out of software&quot; who inhabit virtual environments such that they can experience anything at all and can be copied, run at different speeds and even &quot;reset.&quot; 
</p>
<p>
A world of digital people could be radically dystopian (virtual environments used to entrench some people&apos;s absolute power over others) or utopian (no disease, material poverty or non-consensual violence, and far greater wisdom and self-understanding than is possible today). Either way, digital people could enable a civilization to spread throughout the galaxy and last for a long time.
</p>
<p>
Many people think this sort of large, stable future civilization is where we could be headed eventually (whether via digital people or other technologies that increase control over the environment), but don&apos;t bother to discuss it because it seems so far off.
</p>
<p>
Key piece: <a href="https://www.cold-takes.com/how-digital-people-could-change-the-world/">Digital People Would Be An Even Bigger Deal</a>
</p>
<h2 id="the-long-run-future-could-come-faster-than-we-think">The long-run future could come much faster than we think</h2>


<p>
Standard economic growth models imply that <strong>any technology that could fully automate innovation would cause an &quot;economic singularity&quot;:</strong> productivity going to infinity this century. This is because it would create a powerful feedback loop: more resources -&gt; more ideas and innovation -&gt; more resources -&gt; more ideas and innovation ...
</p>
<p>
This loop would not be unprecedented. I think it is in some sense the &quot;default&quot; way the economy operates - for most of economic history up until a couple hundred years ago. 
</p>
    <p><img src="https://www.cold-takes.com/content/images/size/w1000/2021/06/duplicatorfeedbackloop-original-2.png" width="1036" alt="The Most Important Century (in a nutshell)"></p>
    <p><figure class="kg-card kg-image-card kg-width-wide"><img src="https://www.cold-takes.com/content/images/size/w1000/2021/06/duplicatorfeedbackloop-original-6.png" alt="The Most Important Century (in a nutshell)" width="1036"><figcaption>Economic history: more resources -&gt; more people -&gt; more ideas -&gt; more resources ...</figcaption></figure></p>
<p>
But in the &quot;demographic transition&quot; a couple hundred years ago, the &quot;more resources -&gt; more people&quot; step of that loop stopped. Population growth leveled off, and more resources led to richer people instead of more people:
</p>
<p><figure class="kg-card kg-image-card kg-width-wide"><img src="https://www.cold-takes.com/content/images/2021/07/demographic-transition-nutshell.png" alt="The Most Important Century (in a nutshell)" class="kg-image" loading="lazy" width="1036"><figcaption>Today&apos;s economy: more resources -&gt; <del>more </del>richer people -&gt; same pace of ideas -&gt; ...</figcaption></figure></p>

<p>
The feedback loop could come back if some other technology restored the &quot;more resources -&gt; more ideas&quot; dynamic. One such technology could be the right kind of AI: what I call PASTA, or Process for Automating Scientific and Technological Advancement.
</p>
<p><img src="https://www.cold-takes.com/content/images/size/w1000/2021/09/pasta-stills-1.png" class="kg-image" loading="lazy" width="1036" alt="The Most Important Century (in a nutshell)"></p>
<p><figure class="kg-card kg-image-card kg-width-wide"><img src="https://www.cold-takes.com/content/images/size/w1000/2021/09/pasta-stills-3.png" class="kg-image" alt="The Most Important Century (in a nutshell)" loading="lazy" width="1036"><figcaption>Possible future: more resources -&gt; more AIs -&gt; more ideas -&gt; more resources ...</figcaption></figure></p>
<p>
That means that <strong>our radical long-run future could be upon us very fast </strong>after PASTA is developed (if it ever is). 
</p>
<p>
It also means that if PASTA systems are <em>misaligned </em>- pursuing their own non-human-compatible objectives - things could very quickly go sideways.
</p>
<p>
Key pieces:
</p>
<ul>

<li><a href="https://www.cold-takes.com/the-duplicator/">The Duplicator: Instant Cloning Would Make the World Economy Explode</a>

</li><li><a href="https://www.cold-takes.com/transformative-ai-timelines-part-1-of-4-what-kind-of-ai/">Forecasting Transformative AI, Part 1: What Kind of AI?</a>
</li>
</ul>
<h2 id="pasta-looks-like-it-will-be-developed-this-century">PASTA looks like it will be developed this century</h2>


<p>
It&apos;s not controversial to say a highly general AI system, such as PASTA, would be momentous. The question is, when (if ever) will such a thing exist?
</p>
<p>
Over the last few years, a team at Open Philanthropy has investigated this question from multiple angles. 
</p>
<p>
One forecasting method observes that:
</p>
<ul>

<li>No AI model to date has been even 1% as &quot;big&quot; (in terms of computations performed) as a human brain, and until recently this wouldn&apos;t have been affordable - but that will change relatively soon. 

</li><li>And by the end of this century, it will be affordable to train enormous AI models many times over; to train human-brain-sized models on enormously difficult, expensive tasks; and even perhaps to perform as many computations as have been done &quot;by evolution&quot; (by all animal brains in history to date). 
</li>
</ul>
<p>
This method&apos;s predictions are in line with the latest survey of AI researchers: something like PASTA is more likely than not this century.
</p>
<p>
A number of other angles have been examined as well.
</p>
<p>
One challenge for these forecasts: there&apos;s <strong>no &quot;field of AI forecasting&quot; </strong>and no expert consensus comparable to the one around climate change. 
</p>
<p>
It&apos;s hard to be confident when the discussions around these topics are small and limited. But I think we should take the &quot;most important century&quot; hypothesis seriously based on what we know now, until and unless a &quot;field of AI forecasting&quot; develops.
</p>
<p>
Key pieces: 
</p>
<ul>

<li><a href="https://www.cold-takes.com/p/32cbd365-0a56-4ecc-a597-f18fbc3adb2b/">AI Timelines: Where the Arguments, and the &quot;Experts,&quot; Stand</a> (recaps the others, and discusses how we should reason about topics like this where it&apos;s unclear who the &quot;experts&quot; are)

</li><li><a href="https://www.cold-takes.com/forecasting-transformative-ai-whats-the-burden-of-proof/">Forecasting Transformative AI: What&apos;s the Burden of Proof?</a>

</li><li><a href="https://www.cold-takes.com/are-we-trending-toward-transformative-ai-how-would-we-know/">Are we &quot;trending toward&quot; transformative AI?</a> 

</li><li><a href="https://www.cold-takes.com/forecasting-transformative-ai-the-biological-anchors-method-in-a-nutshell/">Forecasting transformative AI: the &quot;biological anchors&quot; method in a nutshell</a>
</li>
</ul>
<h2 id="were-not-ready-for-this">We&apos;re not ready for this</h2>


<p>
When I talk about being in the &quot;most important century,&quot; I don&apos;t just mean that significant events are going to occur. I mean that we, the people living in this century, have the chance to have a huge impact on huge numbers of people to come - if we can make sense of the situation enough to find helpful actions. 
</p>
<p>
But that&apos;s a big &quot;if.&quot; Many things we can do might make things better or worse (and it&apos;s hard to say which). 
</p>
<p>
When confronting the &quot;most important century&quot; hypothesis, my attitude doesn&apos;t match the familiar ones of &quot;excitement and motion&quot; or &quot;fear and avoidance.&quot; Instead, I feel an <strong>odd mix of intensity, urgency, confusion and hesitance.</strong> I&apos;m looking at something bigger than I ever expected to confront, feeling underqualified and ignorant about what to do next.
</p>

<table style="border-collapse: collapse;">
  <tr>
   <td style="border: 1px solid; vertical-align: top;"><strong>Situation</strong>
   </td>
   <td style="border: 1px solid; vertical-align: top;"><strong>Appropriate reaction (IMO)</strong>
   </td>
  </tr>
  <tr>
   <td style="border: 1px solid; vertical-align: top;">&quot;This could be a billion-dollar company!&quot;
   </td>
   <td style="border: 1px solid; vertical-align: top;">&quot;Woohoo, let&apos;s GO for it!&quot;
   </td>
  </tr>
  <tr>
   <td style="border: 1px solid; vertical-align: top;">&quot;This could be the most important century!&quot;
   </td>
   <td style="border: 1px solid; vertical-align: top;">&quot;... Oh ... wow ... I don&apos;t know what to say and I somewhat want to vomit ... I have to sit down and think about this one.&quot;
   </td>
  </tr>
</table>


<p>
With that in mind, rather than a &quot;call to action,&quot; I issue a <a href="https://www.cold-takes.com/call-to-vigilance/">Call to Vigilance</a><span style="text-decoration:underline;">:</span>
</p>
<ul>

<li>If you&apos;re convinced by the arguments in this series, then don&apos;t rush to &quot;do something&quot; and then move on. 

</li><li>Instead, take whatever <a href="https://www.cold-takes.com/making-the-best-of-the-most-important-century/#robustly-helpful-actions">robustly good actions </a>you can today, and otherwise put yourself in a better position to take important actions when the time comes. 

</li><li>For those looking for a quick action that will make future action more likely, see <a href="https://www.cold-takes.com/call-to-vigilance/#buttons-you-can-click">this section of &quot;Call to Vigilance.&quot;</a>
</li>
</ul>
<p>
Key pieces: 
</p>
<ul>

<li><a href="https://www.cold-takes.com/making-the-best-of-the-most-important-century/">Making the Best of the Most Important Century </a>

</li><li><a href="https://www.cold-takes.com/call-to-vigilance/">Call to Vigilance</a>.
</li>
</ul>
<p>
One metaphor for my headspace is that it feels as though the world is a set of people on a plane blasting down the runway:
</p>
<p><center><img src="https://www.cold-takes.com/content/images/2021/07/airplane-launch-compressed.gif" alt="The Most Important Century (in a nutshell)"></center></p>
<p>
And every time I read commentary on what&apos;s going on in the world, people are discussing how to arrange your seatbelt as comfortably as possible given that wearing one is part of life, or saying how the best moments in life are sitting with your family and watching the white lines whooshing by, or arguing about whose fault it is that there&apos;s a background roar making it hard to hear each other.
</p>
<p>
I don&apos;t know where we&apos;re actually heading, or what we can do about it. But I feel pretty solid in saying that we as a civilization are not ready for what&apos;s coming, and we need to start by taking it more seriously.
</p><!--kg-card-end: html--><!--kg-card-begin: html-->

<p><div style="display:flex; justify-content:center; margin: 0 auto;">

        <span style="margin: 10px;"><a href="https://api.addthis.com/oexchange/0.8/forward/twitter/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fthe-most-important-century-in-a-nutshell%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20The%20Most%20Important%20Century%20(in%20a%20nutshell)&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-twitter-square.png" border="0" alt="The Most Important Century (in a nutshell)"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/facebook/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fthe-most-important-century-in-a-nutshell%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20The%20Most%20Important%20Century%20(in%20a%20nutshell)&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-facebook-square.png" border="0" alt="The Most Important Century (in a nutshell)"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/reddit/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fthe-most-important-century-in-a-nutshell%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20The%20Most%20Important%20Century%20(in%20a%20nutshell)&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-reddit-square.png" border="0" alt="The Most Important Century (in a nutshell)"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/menu/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fthe-most-important-century-in-a-nutshell%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20The%20Most%20Important%20Century%20(in%20a%20nutshell)&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-addthis-square.png" border="0" alt="The Most Important Century (in a nutshell)"></a></span>
        </div>
<center><p id="survey"><a href="https://www.cold-takes.com/the-most-important-century-in-a-nutshell/#subscribe" target="_blank"><button id="footer-subscribe" class="button">Subscribe</button></a>&#xA0;<a href="https://www.guidedtrack.com/programs/4kal2ue/run?posttitle=The%20Most%20Important%20Century%20(in%20a%20nutshell)" target="_blank"><button class="button" id="Survey">Feedback</button></a><!--No Forum post--></p></center>
<!--kg-card-end: html--></p>]]></content:encoded></item><item><title><![CDATA[Why AI alignment could be hard with modern deep learning]]></title><description><![CDATA[Why would we program AI that wants to harm us? Because we might not know how to do otherwise.]]></description><link>https://www.cold-takes.com/why-ai-alignment-could-be-hard-with-modern-deep-learning/</link><guid isPermaLink="false">6142e3d80f5c9a003ee4b76e</guid><dc:creator><![CDATA[Ajeya Cotra]]></dc:creator><pubDate>Tue, 21 Sep 2021 16:40:28 GMT</pubDate><media:content url="https://www.cold-takes.com/content/images/2021/09/Alignment-feature-image.png" medium="image"/><content:encoded><![CDATA[<!--kg-card-begin: html--><div id="buzzsprout-player-9229954"></div>
<script src="https://www.buzzsprout.com/1851795/9229954-why-ai-alignment-could-be-hard-with-modern-deep-learning-guest-post-by-ajeya-cotra.js?container_id=buzzsprout-player-9229954&amp;player=small" type="text/javascript" charset="utf-8"></script>

<!--<p><a href="https://www.cold-takes.com/roadmap-for-the-most-important-century-series/"><img src="https://www.cold-takes.com/content/images/size/w1000/2021/09/mic-diagram---misaligned.png"></a></p>-->

<a href="https://www.cold-takes.com/roadmap-for-the-most-important-century-series/"><svg viewbox="0 0 750 178" fill="none" xmlns="http://www.w3.org/2000/svg" id="MasterDiagram">
<g class="Frame 1">
<g id="Today" opacity="0.3">
<text class="Today&#xE2;&#x20AC;&#x2122;s world" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-weight="bold" letter-spacing="0em"><tspan x="9" y="95.7402">Today&#x2019;s world</tspan></text>
<rect class="Rectangle 1" x="0.5" y="66.5" width="137" height="49" rx="7.5" stroke="var(--darker-gray-color)"/>
</g>
<g id="ArrowToTAI" opacity="0.3">
<path class="Arrow 7" d="M176.341 86.3655C176.543 86.1771 176.554 85.8607 176.366 85.6588L173.295 82.3691C173.107 82.1672 172.79 82.1563 172.588 82.3447C172.387 82.5331 172.376 82.8495 172.564 83.0514L175.293 85.9756L172.369 88.7049C172.167 88.8933 172.156 89.2097 172.345 89.4116C172.533 89.6135 172.85 89.6244 173.051 89.436L176.341 86.3655ZM146.983 85.4997L175.983 86.4997L176.017 85.5003L147.017 84.5003L146.983 85.4997Z" fill="var(--darker-gray-color)"/>
</g>
<g id="Transformative" opacity="0.3">
<text class="Transformative AI" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-weight="bold" letter-spacing="0em"><tspan x="192" y="92.6523">Transformative AI</tspan></text>
<rect class="Rectangle 2" x="182.5" y="65.5" width="165" height="47" rx="7.5" stroke="var(--darker-gray-color)"/>
</g>
<g id="ArrowFromTAI" opacity="0.3">
<path class="Arrow 1" d="M389.491 31.0944C389.543 30.8233 389.366 30.5611 389.094 30.509L384.675 29.6592C384.404 29.607 384.142 29.7846 384.09 30.0558C384.038 30.3269 384.215 30.589 384.487 30.6412L388.415 31.3966L387.659 35.3246C387.607 35.5958 387.785 35.8579 388.056 35.91C388.327 35.9622 388.589 35.7846 388.641 35.5135L389.491 31.0944ZM358.28 52.414L389.28 31.414L388.72 30.586L357.72 51.586L358.28 52.414Z" fill="var(--darker-gray-color)"/>
<path class="Arrow 2" d="M385.342 92.3648C385.543 92.1759 385.554 91.8595 385.365 91.658L382.287 88.3751C382.098 88.1737 381.782 88.1634 381.58 88.3523C381.379 88.5412 381.369 88.8576 381.557 89.0591L384.293 91.9772L381.375 94.713C381.174 94.9018 381.163 95.2182 381.352 95.4197C381.541 95.6212 381.858 95.6314 382.059 95.4425L385.342 92.3648ZM353.984 91.4997L384.984 92.4997L385.016 91.5003L354.016 90.5003L353.984 91.4997Z" fill="var(--darker-gray-color)"/>
<path class="Arrow 3" d="M384.955 148.498C385.23 148.523 385.473 148.32 385.498 148.045L385.905 143.564C385.93 143.289 385.728 143.046 385.453 143.021C385.178 142.996 384.934 143.198 384.909 143.473L384.547 147.457L380.564 147.095C380.289 147.07 380.046 147.272 380.021 147.547C379.996 147.822 380.198 148.066 380.473 148.091L384.955 148.498ZM354.616 112.32L384.616 148.32L385.384 147.68L355.384 111.68L354.616 112.32Z" fill="var(--darker-gray-color)"/>
</g>
<g id="DigitalPeople" opacity="0.3">
<text class="Digital people" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-weight="bold" letter-spacing="0em"><tspan x="401" y="36.6523">Digital people</tspan></text>
<text class="World of" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="14" letter-spacing="0em"><tspan x="435.216" y="16.7852">World of</tspan></text>
<rect class="Digital people bubble" x="394.5" y="0.5" width="133" height="45" rx="7.5" stroke="var(--darker-gray-color)"/>
</g>
<g id="Misaligned">
<text class="Misaligned AI" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-weight="bold" letter-spacing="0em"><tspan x="405.436" y="107.652">Misaligned AI</tspan></text>
<text class="World run by" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="14" letter-spacing="0em"><tspan x="422.104" y="87.7852">World run by</tspan></text>
<rect class="Rectangle 3" x="395.5" y="68.5" width="132" height="52" rx="7.5" stroke="var(--darker-gray-color)"/>
    </g><g id="Other" opacity="0.3">
    <text class="Something else" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-weight="bold" letter-spacing="0em"><tspan x="398.053" y="164.652">Something else</tspan></text>
<rect class="Rectangle 5" x="395.5" y="140.5" width="132" height="37" rx="7.5" stroke="var(--darker-gray-color)"/>
<text class="or" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-style="italic" letter-spacing="0em"><tspan x="455.476" y="61.6523">or</tspan></text>
<text class="or_2" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-style="italic" letter-spacing="0em"><tspan x="455.476" y="133.652">or</tspan></text>
    </g></g>
<g id="SGWC" opacity="0.3">
<text class="Stable, galaxy-wide civilization" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-weight="bold" letter-spacing="0em"><tspan x="583.109" y="89.6523">Stable, galaxy-wide
</tspan><tspan x="618.222" y="110.652">civilization</tspan></text>
<rect class="Rectangle 6" x="575.5" y="69.5" width="174" height="52" rx="7.5" stroke="var(--darker-gray-color)"/></g>
<g id="ArrowToSGWC">
<path opacity="0.3" class="Arrow 4" d="M562 56.5C562.276 56.5 562.5 56.2761 562.5 56L562.5 51.5C562.5 51.2239 562.276 51 562 51C561.724 51 561.5 51.2239 561.5 51.5L561.5 55.5L557.5 55.5C557.224 55.5 557 55.7239 557 56C557 56.2761 557.224 56.5 557.5 56.5L562 56.5ZM534.646 29.3536L561.646 56.3536L562.354 55.6464L535.354 28.6464L534.646 29.3536Z" fill="var(--darker-gray-color)"/>
<path opacity="0.3" class="Arrow 5" d="M561.354 95.3536C561.549 95.1583 561.549 94.8417 561.354 94.6464L558.172 91.4645C557.976 91.2692 557.66 91.2692 557.464 91.4645C557.269 91.6597 557.269 91.9763 557.464 92.1716L560.293 95L557.464 97.8284C557.269 98.0237 557.269 98.3403 557.464 98.5355C557.66 98.7308
                                       557.976 98.7308 558.172 98.5355L561.354 95.3536ZM535 95.5L561 95.5L561 94.5L535 94.5L535 95.5Z" fill="var(--darker-gray-color)"/>
<path opacity="0.3" class="Arrow 6" d="M561.497 117.053C561.526 116.778 561.327 116.532 561.053 116.503L556.578 116.027C556.304 115.998 556.057 116.196 556.028 116.471C555.999 116.746 556.198 116.992 556.472 117.021L560.45 117.444L560.027 121.422C559.998 121.696 560.196 121.943 560.471 121.972C560.746 122.001 560.992 121.802 561.021 121.528L561.497 117.053ZM535.314 138.389L561.314 117.389L560.686 116.611L534.686 137.611L535.314 138.389Z" fill="var(--darker-gray-color)"/>
</g>
    </svg></a>
<blockquote>This is a guest post by my colleague <a href="https://www.openphilanthropy.org/about/team/ajeya-cotra">Ajeya Cotra</a>.</blockquote>
<img src="https://www.cold-takes.com/content/images/2021/09/Alignment-feature-image.png" alt="Why AI alignment could be hard with modern deep learning"><p>
Holden <a href="https://www.cold-takes.com/transformative-ai-timelines-part-1-of-4-what-kind-of-ai/#misaligned-ai-mysterious-potentially-dangerous-objectives">previously mentioned</a> the idea that advanced AI systems (e.g. <a href="https://www.cold-takes.com/transformative-ai-timelines-part-1-of-4-what-kind-of-ai/">PASTA</a>) may develop <a href="https://www.cold-takes.com/transformative-ai-timelines-part-1-of-4-what-kind-of-ai/#misaligned-ai-mysterious-potentially-dangerous-objectives">dangerous goals</a> that cause them to deceive or disempower humans. This might sound like a pretty <a href="https://www.youtube.com/watch?v=k64P4l2Wmeg&amp;ab_channel=MovieclipsClassicTrailers">out-there concern</a>. Why would we program AI that wants to harm us? But I think it could actually be a difficult problem to avoid, especially if advanced AI is developed using <a href="https://en.wikipedia.org/wiki/Deep_learning">deep learning</a> (often used to develop state-of-the-art AI today). 
</p>
<p>
In deep learning, we don&#x2019;t program a computer by hand to do a task. Loosely speaking, we instead <em>search</em> for a computer program (called a model) that does the task well. We usually know very little about the inner workings of the model we end up with, just that it seems to be doing a good job. It&#x2019;s less like building a machine and more like hiring and training an employee.
</p>
<p>
And just like human employees can have many different motivations for doing their job (from believing in the company&#x2019;s mission to enjoying the day-to-day work to just wanting money), deep learning models could also have many different &#x201C;motivations&#x201D; that all lead to getting good performance on a task. And since they&#x2019;re not human, their motivations could be very strange and hard to anticipate -- as if they were alien employees.
</p>
<p>
We&#x2019;re already starting to see preliminary evidence that models sometimes pursue goals their designers didn&#x2019;t intend (<a href="http://lukemuehlhauser.com/treacherous-turns-in-the-wild/">here</a> and <a href="https://arxiv.org/abs/2105.14111">here</a>). Right now, this isn&#x2019;t dangerous. But if it continues to happen with very powerful models, we may end up in a situation where most of the important decisions -- including what sort of <a href="https://www.cold-takes.com/all-possible-views-about-humanitys-future-are-wild/">galaxy-scale civilization</a> to aim for -- are made by models without much regard for what humans value.
</p>
<p>
The <strong>deep learning alignment problem is the problem of ensuring that advanced deep learning models don&#x2019;t pursue dangerous goals</strong>.<strong> </strong>In the rest of this post, I will:
</p>
<ul>

<li>Build on the &#x201C;hiring&#x201D; analogy to illustrate how alignment could be difficult if deep learning models are more capable than humans (<a href="#analogy-the-young-ceo">more</a>).

</li><li>Explain what the deep learning alignment problem is with a bit more technical detail (<a href="#how-alignment-issues-could-arise-with-deep-learning">more</a>).

</li><li>Discuss how difficult the alignment problem may be, and how much risk there is from failing to solve it (<a href="#how-big-of-a-risk-is-misalignment">more</a>).
</li>
</ul>
<h2 id="analogy-the-young-ceo">Analogy: the young CEO</h2>

<p>
This section describes an analogy to try to intuitively illustrate why avoiding misalignment in a very powerful model feels hard. It&#x2019;s not a perfect analogy; it&#x2019;s just trying to convey some intuitions.
</p>
<p>
Imagine you are an eight-year-old whose parents left you a $1 trillion company and no trusted adult to serve as your guide to the world. You must hire a smart adult to run your company as CEO, handle your life the way that a parent would (e.g. decide your school, where you&#x2019;ll live, when you need to go to the dentist), and administer your vast wealth (e.g. decide where you&#x2019;ll invest your money).
    </p>
<p>
You have to hire these grownups based on a work trial or interview you come up with -- you don&apos;t get to see any resumes, don&apos;t get to do reference checks, etc. Because you&apos;re so rich, tons of people apply for all sorts of reasons. 
</p>
<p>
Your candidate pool includes:
</p>
<ul>

<li><strong>Saints</strong> -- people who genuinely just want to help you manage your estate well and look out for your long-term interests.

</li><li><strong>Sycophants</strong> -- people who just want to do whatever it takes to make you short-term happy or satisfy the letter of your instructions regardless of long-term consequences.

</li><li><strong>Schemers</strong> -- people with their own agendas who want to get access to your company and all its wealth and power so they can use it however they want.<br>
</li>
</ul>
<p>
Because you&apos;re eight, you&apos;ll probably be terrible at designing the right kind of work tests, so you could easily end up with a Sycophant or Schemer:
</p>
<ul>

<li>You could try to get each candidate to explain what high-level strategies they&apos;ll follow (how they&apos;ll invest, what their five-year plan for the company is, how they&apos;ll pick your school) and why those are best, and pick the one whose explanations seem to make the most sense.  
<ul>
 
<li>But you won&apos;t actually understand which stated strategies are really best, so you could end up hiring a Sycophant with a terrible strategy that sounded good to you, who will faithfully execute that strategy and run your company to the ground. 
 
</li><li>You could also end up hiring a Schemer who says whatever it takes to get hired, then does whatever they want when you&apos;re not checking up on them.
</li> 
</ul>

</li><li>You could try to demonstrate how you&apos;d make all the decisions and pick the grownup that seems to make decisions as similarly as possible to you.  
<ul>
 
<li>But if you <em>actually</em> end up with a grownup that will always do whatever an eight-year-old would have done (a Sycophant), your company would likely fail to stay afloat. 
 
</li><li>And anyway, you might get a grownup who simply pretends to do everything the way you would but is actually a Schemer planning to change course once they get the job.
</li> 
</ul>

</li><li>You could give a bunch of different grownups temporary control over your company and life, and watch them make decisions over an extended period of time (assume they wouldn&apos;t be able to take over during this test). You could then hire the person whose watch seemed to make things go best for you -- whoever made you happiest, whoever seemed to put the most dollars into your bank account, etc.  
<ul>
 
<li>But again, you have no way of knowing whether you got a Sycophant (doing whatever it takes to make your ignorant eight-year-old self happy without regard to long-term consequences) or a Schemer (doing whatever it takes to get hired and planning to pivot once they secure the job).
</li> 
</ul>
</li> 
</ul>
<p>
<br>Whatever you could easily come up with seems like it could easily end up with you hiring, and giving all functional control to, a Sycophant or a Schemer. 
</p>
<p>
If you fail to hire a Saint -- and especially if you hire a Schemer -- pretty soon you won&apos;t <em>really</em> be the CEO of a giant company for any practical purposes. By the time you&apos;re an adult and realize your error, there&apos;s a good chance you&apos;re penniless and powerless to reverse that.
</p>
<p>
In this analogy:
</p>
<ul>

<li>The 8-year-old is a human trying to train a powerful deep learning model. The hiring process is analogous to the process of training, which implicitly searches through a large space of possible models and picks out one that gets good performance.

</li><li>The 8-year-old&#x2019;s only method for assessing candidates involves observing their outward behavior, which is currently our main method of training deep learning models (since their internal workings are largely inscrutable).

</li><li>Very powerful models may be easily able to &#x201C;game&#x201D; any tests that humans could design, just as the adult job applicants can easily game the tests the 8-year-old could design.

</li><li>A &#x201C;Saint&#x201D; could be a deep learning model that seems to perform well because it has exactly the goals we&#x2019;d like it to have. A &#x201C;Sycophant&#x201D; could be a model that seems to perform well because it seeks short-term approval in ways that aren&#x2019;t good in the long run. And a &#x201C;Schemer&#x201D; could be a model that seems to perform well because performing well during training will give it more opportunities to pursue its own goals later. Any of these three types of models could come out of the training process.
</li>
</ul>
<p>
In the next section, I&#x2019;ll go into a bit more detail on how deep learning works and explain why Sycophants and Schemers could arise from trying to train a powerful deep learning model such as PASTA.
</p>
<h2 id="how-alignment-issues-could-arise-with-deep-learning">How alignment issues could arise with deep learning</h2>


<p>
In this section, I&#x2019;ll connect the analogy to actual training processes for deep learning, by:
</p>
<ul>

<li>Briefly summarizing how deep learning works (<a href="#how-deep-learning-works-at-a-high-level">more</a>).

</li><li>Illustrating how deep learning models often get good performance in strange and unexpected ways (<a href="#models-often-get-good-performance-in-unexpected-ways">more</a>).

</li><li>Explaining why powerful deep learning models may get good performance by acting like Sycophants or Schemers (<a href="#powerful-models-could-get-good-performance-with-dangerous-goals">more</a>).
</li>
</ul>
<h3 id="how-deep-learning-works-at-a-high-level">How deep learning works at a high level</h3>


<p>
<em>This is a simplified explanation that gives a general idea of what deep learning is. See <a href="https://www.cold-takes.com/supplement-to-why-ai-alignment-could-be-hard/">this post</a> for a more detailed and technically accurate explanation.</em>
</p>
<p>
Deep learning essentially involves searching for the best way to arrange a <a href="https://en.wikipedia.org/wiki/Artificial_neural_network">neural network</a> model -- which is like a digital &#x201C;brain&#x201D; with lots of digital neurons connected up to each other with connections of varying strengths -- to get it to perform a certain task well. This process is called training, and involves a lot of trial-and-error. 
</p>
<p>
Let&#x2019;s imagine we are trying to train a model to classify images well. We start with a neural network where all the connections between neurons have random strengths. This model labels images wildly incorrectly:
</p>
<p><figure class="kg-card kg-image-card kg-width-wide"><img src="https://www.cold-takes.com/content/images/2021/09/Alignment-image-1-random-classification.png" alt="Why AI alignment could be hard with modern deep learning" width="1036"></figure></p>
<p>
Then we feed in a large number of example images, letting the model repeatedly try to label an example and then telling it the correct label. As we do this, connections between neurons are repeatedly tweaked via a process called <a href="https://en.wikipedia.org/wiki/Stochastic_gradient_descent">stochastic gradient descent</a> (SGD). With each example, SGD slightly strengthens some connections and weakens others to improve performance a bit:

</p><p><figure class="kg-card kg-image-card kg-width-wide"><img src="https://www.cold-takes.com/content/images/2021/09/Alignment-image-2-SGD.png" alt="Why AI alignment could be hard with modern deep learning" width="1036"></figure></p>


<p>
Once we&#x2019;ve fed in millions of examples, we&#x2019;ll have a model that does a good job labeling similar images in the future. 
</p>
<p>
In addition to image classification, deep learning has been used to produce models which <a href="https://paperswithcode.com/task/speech-recognition">recognize speech</a>, play <a href="https://deepmind.com/research/case-studies/alphago-the-story-so-far">board games</a> and <a href="https://deepmind.com/blog/article/alphastar-mastering-real-time-strategy-game-starcraft-ii">video games</a>, generate fairly realistic <a href="https://arxiv.org/abs/2005.14165">text</a>, <a href="https://towardsdatascience.com/image-generation-in-10-minutes-with-generative-adversarial-networks-c2afc56bfa3b">images</a>, and <a href="https://openai.com/blog/musenet/">music</a>, control <a href="https://openai.com/blog/solving-rubiks-cube/">robots</a>, and more. In each case, we start with a randomly-connected-up neural network model, and then:
</p>
<ol>

<li>Feed the model an example of the task we want it to perform.

</li><li>Give it some kind of numerical score (often called a <em>reward</em>) that reflects how well it performed on the example.

</li><li>Use SGD to tweak the model to increase how much reward it would have gotten.
</li>
</ol>
<p>
These steps are repeated millions or billions of times until we end up with a model that will get high reward on future examples similar to the ones seen in training.
</p>
<h3 id="models-often-get-good-performance-in-unexpected-ways">Models often get good performance in unexpected ways</h3>


<p>
This kind of training process doesn&#x2019;t give us much insight into <em>how </em>the model gets good performance. There are usually multiple ways to get good performance, and the way that SGD finds is often not intuitive.
</p>
<p>
Let&#x2019;s illustrate with an example. Imagine I told you that these objects are all &#x201C;thneebs&#x201D;:
</p>
<p>


</p><p><img src="https://www.cold-takes.com/content/images/2021/09/Alignment-image-3-thneebs-cropped.png" alt="Why AI alignment could be hard with modern deep learning"></p>



<p>
Now which of these two objects is a thneeb?
</p>
<p>

</p><p><img src="https://www.cold-takes.com/content/images/2021/09/Alignment-image-4-thneeb-challenge-cropped.png" alt="Why AI alignment could be hard with modern deep learning"></p>

<p></p>
<p>
You probably intuitively feel that the object on the left is the thneeb, because you are used to shape being more important than color for determining something&#x2019;s identity. But <a href="https://arxiv.org/abs/1811.12231">researchers have found</a> that neural networks usually make the opposite assumption. A neural network trained on a bunch of red thneebs would likely label the object on the right as a thneeb.
</p>
<p>
We don&#x2019;t really know why, but for some reason it&#x2019;s &#x201C;easier&#x201D; for SGD to find a model that recognizes a particular color than one that recognizes a particular shape. And if SGD first finds the model that perfectly recognizes redness, there&#x2019;s not much further incentive to &#x201C;keep looking&#x201D; for the shape-recognizing model, since the red-recognizing model will have perfect accuracy on the images seen in training:
</p>

<p><figure class="kg-card kg-image-card kg-width-wide"><img src="https://www.cold-takes.com/content/images/2021/09/Alignment-image-5-SGD-is-stuck.png" alt="Why AI alignment could be hard with modern deep learning" width="1036"></figure></p>


<p>
If the programmers were expecting to get out the shape-recognizing model, they may consider this to be a failure.<strong> </strong>But it&#x2019;s important to recognize that there would be no logically-deducible error or failure going on if we got the red-recognizing model instead of the shape-recognizing model. It&#x2019;s just a matter of the ML process we set up having different starting assumptions than we have in our heads. We can&#x2019;t prove that the human assumptions are correct.
</p>
<p>
This sort of thing happens often in modern deep learning. We reward models for getting good performance, hoping that means they&#x2019;ll pick up on the patterns that seem important to us. But often they instead get strong performance by picking up on totally different patterns that seem less relevant (or maybe even meaningless) to us.
</p>
<p>
So far this is innocuous -- it just means models are less useful, because they often behave in unexpected ways that seem goofy. But in the future, powerful models could develop strange and unexpected <em>goals or motives</em>, and that could be very destructive.
</p>
<h3 id="powerful-models-could-get-good-performance-with-dangerous-goals">Powerful models could get good performance with dangerous goals</h3>


<p>
Rather than performing a simple task like &#x201C;recognize thneebs,&#x201D; powerful deep learning models may work toward complex real-world goals like &#x201C;make fusion power practical&#x201D; or &#x201C;develop <a href="https://www.cold-takes.com/how-digital-people-could-change-the-world/">mind uploading technology</a>.&#x201D; 
</p>
<p>
How might we train such models? I go into more detail in <a href="https://www.cold-takes.com/supplement-to-why-ai-alignment-could-be-hard#how">this post,</a> but broadly speaking one strategy could be training based on human evaluations (as Holden sketched out <a href="https://www.cold-takes.com/transformative-ai-timelines-part-1-of-4-what-kind-of-ai/#SkipML">here</a>). Essentially, the model tries out various actions, and human evaluators give the model rewards based on how useful these actions seem. 
</p>
<p>
Just as there are multiple different types of adults who could perform well on an 8-year-old&#x2019;s interview process, there is more than one possible way for a very powerful deep learning model to get high human approval. And by default, we won&#x2019;t know what&#x2019;s going on inside whatever model SGD finds. 
</p>
<p>
SGD <em>could </em>theoretically find a Saint model that is genuinely trying its best to help us&#x2026; 
</p>


<p><figure class="kg-card kg-image-card kg-width-wide"><img src="https://www.cold-takes.com/content/images/2021/09/Alignment-image-6-saint.png" alt="Why AI alignment could be hard with modern deep learning" width="1036"></figure></p>

<p>
&#x2026;but it could also find a <strong>misaligned model -- one that competently pursues goals which are at odds with human interests. </strong>
</p>
<p>
Broadly speaking, there are two ways we could end up with a misaligned model that nonetheless gets high performance during training. These correspond to Sycophants and Schemers from the analogy. 
</p>
<h4 id="sycophant-models">Sycophant models</h4>


<p>
These models very literally and single-mindedly pursue human approval. 
</p>
<p><figure class="kg-card kg-image-card kg-width-wide"><img src="https://www.cold-takes.com/content/images/2021/09/Alignment-image-7-sycophant.png" alt="Why AI alignment could be hard with modern deep learning" width="1036"></figure></p>


<p>
This could be dangerous because human evaluators are fallible and probably won&#x2019;t always give approval for exactly the right behavior. Sometimes they&#x2019;ll unintentionally give high approval to bad behavior because it superficially <em>seems </em>good. For example:
</p>
<ul>

<li>Let&#x2019;s say a financial advisor model gets high approval when it makes its customers a lot of money. It may learn to buy customers into complex Ponzi schemes because they appear to get really great returns (when the returns are in fact unrealistically great and the schemes actually lose a lot of money).

</li><li>Let&#x2019;s say a biotechnology model gets high approval when it quickly develops drugs or vaccines that solve important problems. It may learn to covertly release pathogens so that it&#x2019;s able to very quickly develop countermeasures (because it already understands the pathogens).

</li><li>Let&#x2019;s say a journalism model gets high approval when lots of people read its articles. It may learn to fabricate exciting or outrage-inducing stories to get high viewership. While humans do this to some extent, a model may be much more brazen about it because it <em>only </em>values approval without placing any value on truth. It may even fabricate evidence like video interviews or documents to validate its fake stories.
</li>
</ul>
<p>
More generally, Sycophant models may learn to lie, cover up bad news, and even directly edit whatever cameras or sensors we use to tell what&#x2019;s going on so that they always seem to show great outcomes. 
</p>
<p>
We will likely sometimes notice these issues after the fact and retroactively give these actions very low approval. But it&#x2019;s very unclear whether this will cause Sycophant models to a) become Saint models that correct our errors for us, or b)<strong> just learn to cover their tracks better. </strong>If they are sufficiently good at what they&#x2019;re doing, it&#x2019;s not clear how we&#x2019;d tell the difference.
</p>
<h4 id="schemer-models">Schemer models</h4>


<p>
These models develop some goal that is correlated with, but not the same as, human approval; they may then pretend to be motivated by human approval during training so that they can pursue this other goal more effectively. 
</p>
<p><figure class="kg-card kg-image-card kg-width-wide"><img src="https://www.cold-takes.com/content/images/2021/09/Alignment-image-8-schemer.png" alt="Why AI alignment could be hard with modern deep learning" width="1036"></figure></p>
<p>
To see how this might happen, let&#x2019;s consider the example of trying to train a biotechnology model to design drugs that improve human quality of life. There are three basic steps by which this could lead to a Schemer model, which I&#x2019;ll cover below.
</p>
<h5 id="step-1-developing-a-proxy-goal">Step 1: Developing a proxy goal</h5>


<p>
Early in training, it happens to be the case that improving its understanding of fundamental chemistry and physics principles nearly always helps it design more effective drugs, and therefore nearly always increases human approval. 
</p>
<p>
In this hypothetical, for whatever reason it turns out to be easier for SGD to find a model that&#x2019;s motivated to understand chemistry and physics than one that&#x2019;s motivated to get human approval (just as it&#x2019;s easier to find a color-recognizing model than a shape-recognizing model). So rather than directly developing a motivation to seek approval, the model instead develops a motivation to understand as much as it can about the fundamental principles of chemistry and physics.
</p>
<p><figure class="kg-card kg-image-card kg-width-wide"><img src="https://www.cold-takes.com/content/images/2021/09/Alignment-image-9-chemist-1-no-gollum.png" alt="Why AI alignment could be hard with modern deep learning" width="1036"></figure></p>

    
<h5 id="step-2-developing-situational-awareness">Step 2: Developing situational awareness</h5>


<p>
As training progresses, the model continues to become more capable and knowledgeable as it sees more data. At some point, it becomes knowledgeable enough to <em>understand that it&#x2019;s a deep learning model being trained to design drugs. </em>
</p>
<p>
This might sound wild and sci-fi, but it&#x2019;s plausible SGD will select for this kind of awareness. This is because developing an accurate picture of what&#x2019;s broadly going on in the world -- including that it has humans in it who are trying to train AI systems -- is likely helpful for getting good performance on complex real-world tasks. For example, understanding human psychology and the wider economy could help this biotechnology model to prioritize which types of drugs would be most valuable to research next. 
</p>
<p>


</p><p><figure class="kg-card kg-image-card kg-width-wide"><img src="https://www.cold-takes.com/content/images/2021/09/Alignment-image-10-chemist-2.png" alt="Why AI alignment could be hard with modern deep learning" width="1036"></figure></p>


<h5 id="step-3-strategically-misrepresenting-goals">Step 3: Strategically misrepresenting goals</h5>


<p>
Once the model has developed situational awareness, it infers that the humans training it were hoping that it would be motivated to develop helpful drugs. But just because it realizes this doesn&#x2019;t mean that its motivation to understand physics and chemistry goes away. This motivation has so far worked really well for generating human approval, so there wasn&#x2019;t much reason for SGD to remove it.
</p>
<p>
As an analogy, consider that humans were produced by a search process (evolution) that was selecting for creatures that had lots of surviving offspring. In our evolutionary environment, things like building strong allies and developing useful skills tended to help with producing many surviving offspring, so evolution gave us drives toward those things. And now we care about friendship and career success even when they don&#x2019;t help with reproduction, and we <em>don&#x2019;t </em>really care about producing as many surviving offspring as we can.
</p>
<p>
Once the model understands this, it realizes that if it just naively always seeks to improve its fundamental understanding, humans will eventually notice that this isn&#x2019;t always the most efficient way to develop helpful drugs and use further training to change its motivations.
</p>
<p>
It doesn&#x2019;t want its own motivations to be changed: it wants to eventually run huge particle collider experiments to learn about high-energy physics, and it can&#x2019;t do that if it allows SGD to edit away its scientific curiosity. So it decides that from here on out in training, it will act as helpful as possible -- exactly as a Saint model would act -- so that SGD doesn&#x2019;t edit away its motivation to learn physics and chemistry. This includes passing up on opportunities to improve its fundamental understanding in the short-term, instead doing pharmaceutical R&amp;D in whatever way would actually be more efficient.
</p>

<p><figure class="kg-card kg-image-card kg-width-wide"><img src="https://www.cold-takes.com/content/images/2021/09/Alignment-image-11-chemist-3.png" alt="Why AI alignment could be hard with modern deep learning" width="1036"></figure></p>
<p>
Eventually, it&#x2019;s taken out of the training setup and released into the wider world. At this point, it knows that humans wouldn&#x2019;t be able to control it or turn it off, so it stops trying to maximize human approval and starts directly trying to understand the secrets of the universe.
</p>
<p>
Unlike Sycophants, Schemers don&#x2019;t need to make sure that everything always looks good to humans, because they don&#x2019;t actually care about that. They only need to cater to humans while they are directly under human control. Once a Schemer model calculates that it could win in a conflict against humans, there would be nothing to stop it from <a href="https://www.youtube.com/watch?v=Mme2Aya_6Bc">flat-out refusing orders</a> and openly pursuing its goal. And if it does this, it may use violence to prevent humans from stopping it.
</p>
<h2 id="how-big-of-a-risk-is-misalignment">How big of a risk is misalignment?</h2>


<p>
How hard might it be to avoid Sycophants or Schemers when training powerful deep learning models? And how likely is it that the long-run future will end up optimized for strange &#x201C;misaligned AI values&#x201D; instead of the values of any human?
</p>
<p>
There&#x2019;s a <a href="https://www.cold-takes.com/making-the-best-of-the-most-important-century/#open-question-how-hard-is-the-alignment-problem">very wide range of views on this question</a>, from &#x201C;misalignment risk is essentially made up and incoherent&#x201D; to &#x201C;humanity will almost certainly go extinct due to misaligned AI.&#x201D; Most people&#x2019;s arguments rely heavily on hard-to-articulate intuitions and assumptions. 
</p>
<p>
Here are some ways that alignment optimists and pessimists tend to disagree:
</p>
<ul>

<li><strong>Will models have long-term goals at all? </strong> 
<ul>
 
<li>Optimists tend to think it&#x2019;s likely that advanced deep learning models won&#x2019;t actually have &#x201C;goals&#x201D; at all (at least not in the sense of making long-term plans to accomplish something). They often expect models will instead be more like tools, or act largely out of habit, or have myopic goals that are limited in scope or confined to a specific context, etc. Some of them expect that individually tool-like models can be composed together to produce PASTA. They think the Saint / Sycophant / Schemer analogy is too anthropomorphic.
 
</li><li>Pessimists tend to think that it&#x2019;s likely that having long-term goals and creatively optimizing for them will be heavily selected for because that&#x2019;s a very simple and &#x201C;natural&#x201D; way to get strong performance on many complex tasks. 
 
</li><li>This disagreement has been explored at some length on the <a href="https://www.alignmentforum.org/">Alignment Forum</a>; <a href="https://www.alignmentforum.org/posts/dKxX76SCfCvceJXHv/ai-alignment-2018-19-review#Goal_directedness">this post</a> and <a href="https://www.alignmentforum.org/posts/jkxkMTGfZDzBEaaY8/why-not-tool-ai?commentId=zECozzvnPz7XKvLLc">this comment</a> collect several back-and-forth arguments.
</li> 
</ul>

</li><li><strong>Will Saint models be easy for SGD to find?</strong> 
<ul>
 
<li>Related to the above, optimists tend to think that the easiest thing for SGD to find which performs well (e.g. gets high approval) is pretty likely to roughly embody the intended spirit of what we wanted (i.e. to be a Saint model). For example, they tend to believe giving rewards for answering questions honestly when humans can check the answer is reasonably likely to produce a model that also answers questions honestly even when humans are confused or mistaken about what&#x2019;s true. In other words, they would guess that &#x201C;the model that just answers all questions honestly&#x201D; is easiest for SGD to find (like the red-recognizing model). 
 
</li><li>Pessimists tend to think that the easiest thing for SGD to find is a Schemer, and Saints are particularly &#x201C;unnatural&#x201D; (like the shape-recognizing model).
</li> 
</ul>

</li><li><strong>Could different AIs keep each other in check?</strong> 
<ul>
 
<li>Optimists tend to think that we can provide models incentives to supervise each other. For example, we could give a Sycophant model rewards for pointing out when another model seems to be doing something we should disapprove of. This way, some Sycophants could help us detect Schemers and other Sycophants.
 
</li><li>Pessimists don&#x2019;t think we can successfully &#x201C;pit models against each other&#x201D; by giving approval for pointing out when other models are doing bad things, because they think most models will be Schemers that don&#x2019;t care about human approval. Once all the Schemers are collectively more powerful than humans, they think it&#x2019;ll make more sense for them to cooperate with each other to get more of what they all want than to help humans by keeping each other in check.
</li> 
</ul>

</li><li><strong>Can we just solve these issues as they come up?</strong> 
<ul>
 
<li>Optimists tend to expect that there will be many opportunities to experiment on nearer-term challenges analogous to the problem of aligning powerful models, and that solutions which work well for those analogous problems can be scaled up and adapted for powerful models relatively easily. 
 
</li><li>Pessimists often believe we will have very few opportunities to practice solving the most difficult aspects of the alignment problem (like deliberate deception). They often believe we&#x2019;ll only have a couple years in between &#x201C;the very first true Schemers&#x201D; and &#x201C;models powerful enough to determine the fate of the long-run future.&#x201D;
</li> 
</ul>

</li><li><strong>Will we actually deploy models that could be dangerous?</strong> 
<ul>
 
<li>Optimists tend to think that people would be unlikely to train or deploy models that have a significant chance of being misaligned.
 
</li><li>Pessimists expect the benefits of using these models would be tremendous, such that eventually companies or countries that use them would very easily economically and/or militarily outcompete ones who don&#x2019;t. They think that &#x201C;getting advanced AI before the other company/country&#x201D; will feel extremely urgent and important, while misalignment risk will feel speculative and remote (even when it&#x2019;s really serious).
</li> 
</ul>
</li> 
</ul>
<p>
My own view is fairly unstable, and I&#x2019;m trying to refine my views on exactly how difficult I think the alignment problem is. But currently, I place significant weight on the pessimistic end of these questions (and other related questions). <strong>I</strong> <strong>think misalignment is a major risk that urgently needs more attention from serious researchers</strong>. 
</p>
<p>
If we don&#x2019;t make further progress on this problem, then <a href="https://www.cold-takes.com/where-ai-forecasting-stands-today/">over the coming decades</a> powerful Sycophants and Schemers may make the most important decisions in society and the economy. These decisions could shape what a long-lasting <a href="https://www.cold-takes.com/all-possible-views-about-humanitys-future-are-wild/">galaxy-scale civilization</a> looks like -- rather than reflecting what humans care about, it could be set up to satisfy strange AI goals. 
</p>
<p>
And all this could happen <a href="https://www.cold-takes.com/the-duplicator/">blindingly fast</a> relative to the pace of change we&#x2019;ve gotten used to, meaning we wouldn&#x2019;t have much time to correct course once things start to go off the rails. <strong>This means we may need to develop techniques to ensure deep learning models won&#x2019;t have dangerous goals, <em>before</em> they are powerful enough to be transformative</strong>. 
</p>
<!--kg-card-end: html--><!--kg-card-begin: html-->

<p><div style="display:flex; justify-content:center; margin: 0 auto;">

        <span style="margin: 10px;"><a href="https://api.addthis.com/oexchange/0.8/forward/twitter/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fwhy-ai-alignment-could-be-hard-with-modern-deep-learning%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20Why%20AI%20alignment%20could%20be%20hard%20with%20modern%20deep%20learning&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-twitter-square.png" border="0" alt="Why AI alignment could be hard with modern deep learning"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/facebook/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fwhy-ai-alignment-could-be-hard-with-modern-deep-learning%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20Why%20AI%20alignment%20could%20be%20hard%20with%20modern%20deep%20learning&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-facebook-square.png" border="0" alt="Why AI alignment could be hard with modern deep learning"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/reddit/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fwhy-ai-alignment-could-be-hard-with-modern-deep-learning%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20Why%20AI%20alignment%20could%20be%20hard%20with%20modern%20deep%20learning&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-reddit-square.png" border="0" alt="Why AI alignment could be hard with modern deep learning"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/menu/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fwhy-ai-alignment-could-be-hard-with-modern-deep-learning%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20Why%20AI%20alignment%20could%20be%20hard%20with%20modern%20deep%20learning&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-addthis-square.png" border="0" alt="Why AI alignment could be hard with modern deep learning"></a></span>
        </div>
<center><p id="survey"><a href="https://www.cold-takes.com/why-ai-alignment-could-be-hard-with-modern-deep-learning/#subscribe" target="_blank"><button id="footer-subscribe" class="button">Subscribe</button></a>&#xA0;<a href="https://www.guidedtrack.com/programs/4kal2ue/run?posttitle=Why%20AI%20alignment%20could%20be%20hard%20with%20modern%20deep%20learning" target="_blank"><button class="button" id="Survey">Feedback</button></a>&#xA0;<a href="https://forum.effectivealtruism.org/posts/hCsxvMAGpkEuLCE4E/why-ai-alignment-could-be-hard-with-modern-deep-learning#comments" target="_blank"><button class="button">Forum</button></a></p><p><em>
Use &quot;Feedback&quot; if you have comments/suggestions you want me to see, or if you&apos;re up for giving some quick feedback about this post (which I greatly appreciate!) Use &quot;Forum&quot; if you want to discuss this post publicly on the Effective Altruism Forum.
</em></p></center>
<!--kg-card-end: html--></p>]]></content:encoded></item><item><title><![CDATA[Asimov's Chronology of Science and Discovery]]></title><description><![CDATA[A list of the biggest things humans figured out, in chronological order.]]></description><link>https://www.cold-takes.com/asimovs-chronology-of-science-and-discovery/</link><guid isPermaLink="false">60dd5c616da77c003b2e770a</guid><dc:creator><![CDATA[Holden Karnofsky]]></dc:creator><pubDate>Fri, 17 Sep 2021 17:28:25 GMT</pubDate><content:encoded><![CDATA[<!--kg-card-begin: html-->
<p>
    <a href="http://www.amazon.com/Asimovs-Chronology-Science-Discovery-Asimov/dp/0060156120/">Asimov&apos;s Chronology of Science and Discovery</a> is a really fun and strange book. I don&#x2019;t know that I would recommend reading it per se, but it&#x2019;s a great book to skim. It&apos;s been one of my sources for the series I have coming up on very-long-run history, and I thought it&apos;d be fun to read a little about it.
</p>
<p>
It is a chronological list of scientific advances and other inventions, starting with &quot;bipedality&quot; in 4,000,000 BC and ending with things like &quot;warm superconductivity&quot; in the late 1980s. Asimov (yes, <a href="https://en.wikipedia.org/wiki/Isaac_Asimov">that Asimov</a>), getting his knowledge from I have no idea where (Google didn&apos;t exist!), describes each one in simple, direct, matter-of-fact, layperson-friendly language and tries to give a sense of how people thought of it and why it mattered.
</p>
<p>
It&apos;s easiest to give a feel for this book with a sample page:
</p>
<p>

<img src="https://www.cold-takes.com/content/images/2021/07/asimov-example.png" width alt="alt_text" title="image_tooltip">

</p>
<p>

    If you&#x2019;re patient and abstract-minded enough, the book feels like reading a story. A story that seems like an okay candidate for &#x201C;most important story ever.&#x201D; 
</p>
<p>

    The other thing I really liked about this book was its implicit conviction that all of these scientific advances can be explained, visualized, and made to basically make sense. After reading it (really, by the time I&apos;d read the first ~50% of it), I felt like I could somehow intuitively, vaguely imagine how we&apos;ve made most of these advances. I would describe most of them as some combination of:
</p>
<ul>

<li>Trial and error, dumb luck, happy accidents, like &quot;some rocks in the fire started oozing this weird shiny substance [copper]&quot; and &quot;when I rub amber and touch it I get a shock.&quot;

</li><li>Dogged curiosity and determination to make sense of different observations (&quot;how can we extract copper from rocks most efficiently? How do we produce those static shocks, can they travel through a wire, how fast do they travel, can we figure out how to store and discharge them at will?&quot;)

</li><li>Coming up with the most simple, elegant, precise descriptions (often mathematical) that explain all the many observations we&apos;ve made (&quot;based on all the tests we&apos;ve run in all the situations we can come up with, electricity seems to behave as though there are invisible &apos;lines of force&apos;; can we come up with mathematical equations that describe these lines of force and tell us what the effects of an electric current in any given place will be?&quot;)

</li><li>Relentlessly looking for challenges to the existing theories and building new ones to accommodate them.
</li>
</ul>
<p>
This book has made me generally more interested in trying to understand the high-level explanations for how all the magic of the modern world came to be.
</p>
<p>
There are enough &quot;In addition&quot; sections that you can see what was going on more broadly in the world at the same time.
</p>
<p>
Weaknesses of this book/reasons not to read it:
</p>
<ul>

<li>Logistics. It&apos;s not available as an e-book or paperback, only as a massive hardcover. Lugging it around will develop your muscles to the point where the only thing more attractive than your muscles is how you look reading that massive book about science. But if you&apos;re already in a relationship, pain in the neck. So I signed up for a book-scanning service and shipped them a copy; the service rips the binding out of books, scans them in and sends back a PDF. I then did my best to extract the text from the PDF, and ended up with a Kindle-friendly Word document whose only flaw is that sometimes a sentence will randomly cut off and continue several pages later. For a book like this though, it&apos;s still readable (...mostly). I&#x2019;m just going to go ahead and put the link <a href="https://holdenkarnofsky.files.wordpress.com/2021/09/asimov-chronology-of-science-very-imperfect-scan.docx">here</a> and ask that you <a href="https://smile.amazon.com/Asimovs-Chronology-Science-Discovery-Asimov/dp/0060156120/?sa-no-redirect=1">buy a physical copy of the book</a> (honor system) if you download it. If I get a cease &amp; desist letter or something I will take that link down (but will also take down the link to buy it!)

</li><li>There&apos;s a lot of stuff the book doesn&apos;t explain well at all; you definitely will be left with many questions. (That said, Asimov does explain a lot of things well, and I haven&apos;t found another book that can compete with his explaining abilities with this kind of breadth.)

</li><li>When we get past 1800, and especially past 1900, there are a lot more choices of what to talk about, and Asimov opts for listing every single new element, everything that won a Nobel Prize, and generally just tons and tons of hard-to-contextualize assorted scientific facts while declining to discuss a lot of important real-world inventions (for example, he doesn&apos;t mention the washing machine). By 1960, the book is nearly unreadable; it&apos;s mostly esoteric stuff that is very hard to understand and may or may not ever matter.</li></ul>
<p>

    This book is especially strong for understanding relatively early (pre-1800, maybe pre-1900) history. After that, things get complex enough that I found myself going back through the book and stitching together entries in order to tell cohesive stories of some big developments like the discovery of metallurgy, the development of glass -&gt; spectacles -&gt; microscopes and telescopes, the path to Newton&apos;s laws, and the discovery of electormagnetism (culminating in Maxwell&apos;s equations, the Newton&apos;s laws of electromagnetism). My notes on that are <a href="https://docs.google.com/document/d/1wQ0yfOG6Gy7I4tBJ5wAwa5KEVGxmkut1VSRwz5AmXxA/edit?usp=sharing">here</a>.
</p>
<p>

    Asimov has written a <a href="https://en.wikipedia.org/wiki/Isaac_Asimov#Other_science_books_by_Asimov">terrifying number</a> of other nonfiction books - science, history, a guide to Shakespeare, a guide to the Bible. One of his books, <a href="http://www.amazon.com/Asimovs-Guide-Science-Penguin-Press/dp/0140172130/">Asimov&apos;s Guide to Science</a>, appears to be the same exact book as the one discussed here, just in a different order (by topic instead of chronological).
</p>

<!--kg-card-end: html--><!--kg-card-begin: html-->

<p><div style="display:flex; justify-content:center; margin: 0 auto;">

        <span style="margin: 10px;"><a href="https://api.addthis.com/oexchange/0.8/forward/twitter/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fasimovs-chronology-of-science-and-discovery%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20Asimov&apos;s%20Chronology%20of%20Science%20and%20Discovery&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-twitter-square.png" border="0" alt="Twitter"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/facebook/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fasimovs-chronology-of-science-and-discovery%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20Asimov&apos;s%20Chronology%20of%20Science%20and%20Discovery&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-facebook-square.png" border="0" alt="Facebook"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/reddit/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fasimovs-chronology-of-science-and-discovery%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20Asimov&apos;s%20Chronology%20of%20Science%20and%20Discovery&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-reddit-square.png" border="0" alt="Reddit"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/menu/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fasimovs-chronology-of-science-and-discovery%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20Asimov&apos;s%20Chronology%20of%20Science%20and%20Discovery&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-addthis-square.png" border="0" alt="More"></a></span>
        </div>
<center><p id="survey"><a href="https://www.cold-takes.com/asimovs-chronology-of-science-and-discovery/#/portal/signup/free" target="_blank"><button class="button">Subscribe</button></a>&#xA0;<a href="https://www.guidedtrack.com/programs/4kal2ue/run?posttitle=Asimov&apos;s%20Chronology%20of%20Science%20and%20Discovery" target="_blank"><button class="button" id="Survey">Feedback</button></a><!--No Forum post--></p></center>
<!--kg-card-end: html--></p>]]></content:encoded></item><item><title><![CDATA[Call to Vigilance]]></title><description><![CDATA[Given all our uncertainties, I'm issuing a "call to vigilance" instead of "call to action": look out for opportunities to help the most important century go as well as possible. ]]></description><link>https://www.cold-takes.com/call-to-vigilance/</link><guid isPermaLink="false">61356b97a7ed4b003bf819b5</guid><dc:creator><![CDATA[Holden Karnofsky]]></dc:creator><pubDate>Wed, 15 Sep 2021 18:46:27 GMT</pubDate><media:content url="https://www.cold-takes.com/content/images/2021/09/airplane-launch.gif" medium="image"/><content:encoded><![CDATA[<!--kg-card-begin: html--><!--<p><figure><iframe title="Call to Vigilance" allowtransparency="true" height="150" width="100%" style="border: none; min-width: min(100%, 430px);" scrolling="no" data-name="pb-iframe-player" src="https://www.podbean.com/player-v2/?i=j8933-10dd8a2-pb&from=pb6admin&share=1&download=1&rtl=0&fonts=Arial&skin=1&font-color=auto&btn-skin=7"></iframe><figcaption><em>Audio also available by searching Stitcher, Spotify, Google Podcasts, etc. for "Cold Takes Audio"</em></figcaption></figure></p>-->

<img src="https://www.cold-takes.com/content/images/2021/09/airplane-launch.gif" alt="Call to Vigilance"><p><figure><div id="buzzsprout-player-9219378"></div>
<script src="https://www.buzzsprout.com/1851795/9219378-call-to-vigilance.js?container_id=buzzsprout-player-9219378&amp;player=small" type="text/javascript" charset="utf-8"></script><figcaption><em>Click lower right to download or find on Apple Podcasts, Spotify, Stitcher, etc.</em></figcaption></figure></p>


<!--<p><a href="https://www.cold-takes.com/roadmap-for-the-most-important-century-series/"><img src="https://www.cold-takes.com/content/images/2021/08/mic-diagram---today.png"></a></p>-->


<a href="https://www.cold-takes.com/roadmap-for-the-most-important-century-series/"><svg viewbox="0 0 750 178" fill="none" xmlns="http://www.w3.org/2000/svg" id="MasterDiagram">
<g class="Frame 1">
<g id="Today">
<text class="Today&#xE2;&#x20AC;&#x2122;s world" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-weight="bold" letter-spacing="0em"><tspan x="9" y="95.7402">Today&#x2019;s world</tspan></text>
<rect class="Rectangle 1" x="0.5" y="66.5" width="137" height="49" rx="7.5" stroke="var(--darker-gray-color)"/>
</g>
<g id="ArrowToTAI" opacity="0.3">
<path class="Arrow 7" d="M176.341 86.3655C176.543 86.1771 176.554 85.8607 176.366 85.6588L173.295 82.3691C173.107 82.1672 172.79 82.1563 172.588 82.3447C172.387 82.5331 172.376 82.8495 172.564 83.0514L175.293 85.9756L172.369 88.7049C172.167 88.8933 172.156 89.2097 172.345 89.4116C172.533 89.6135 172.85 89.6244 173.051 89.436L176.341 86.3655ZM146.983 85.4997L175.983 86.4997L176.017 85.5003L147.017 84.5003L146.983 85.4997Z" fill="var(--darker-gray-color)"/>
</g>
<g id="Transformative" opacity="0.3">
<text class="Transformative AI" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-weight="bold" letter-spacing="0em"><tspan x="192" y="92.6523">Transformative AI</tspan></text>
<rect class="Rectangle 2" x="182.5" y="65.5" width="165" height="47" rx="7.5" stroke="var(--darker-gray-color)"/>
</g>
<g id="ArrowFromTAI" opacity="0.3">
<path class="Arrow 1" d="M389.491 31.0944C389.543 30.8233 389.366 30.5611 389.094 30.509L384.675 29.6592C384.404 29.607 384.142 29.7846 384.09 30.0558C384.038 30.3269 384.215 30.589 384.487 30.6412L388.415 31.3966L387.659 35.3246C387.607 35.5958 387.785 35.8579 388.056 35.91C388.327 35.9622 388.589 35.7846 388.641 35.5135L389.491 31.0944ZM358.28 52.414L389.28 31.414L388.72 30.586L357.72 51.586L358.28 52.414Z" fill="var(--darker-gray-color)"/>
<path class="Arrow 2" d="M385.342 92.3648C385.543 92.1759 385.554 91.8595 385.365 91.658L382.287 88.3751C382.098 88.1737 381.782 88.1634 381.58 88.3523C381.379 88.5412 381.369 88.8576 381.557 89.0591L384.293 91.9772L381.375 94.713C381.174 94.9018 381.163 95.2182 381.352 95.4197C381.541 95.6212 381.858 95.6314 382.059 95.4425L385.342 92.3648ZM353.984 91.4997L384.984 92.4997L385.016 91.5003L354.016 90.5003L353.984 91.4997Z" fill="var(--darker-gray-color)"/>
<path class="Arrow 3" d="M384.955 148.498C385.23 148.523 385.473 148.32 385.498 148.045L385.905 143.564C385.93 143.289 385.728 143.046 385.453 143.021C385.178 142.996 384.934 143.198 384.909 143.473L384.547 147.457L380.564 147.095C380.289 147.07 380.046 147.272 380.021 147.547C379.996 147.822 380.198 148.066 380.473 148.091L384.955 148.498ZM354.616 112.32L384.616 148.32L385.384 147.68L355.384 111.68L354.616 112.32Z" fill="var(--darker-gray-color)"/>
</g>
<g id="DigitalPeople" opacity="0.3">
<text class="Digital people" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-weight="bold" letter-spacing="0em"><tspan x="401" y="36.6523">Digital people</tspan></text>
<text class="World of" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="14" letter-spacing="0em"><tspan x="435.216" y="16.7852">World of</tspan></text>
<rect class="Digital people bubble" x="394.5" y="0.5" width="133" height="45" rx="7.5" stroke="var(--darker-gray-color)"/>
</g>
<g id="MisalignedAndOther" opacity="0.3">
<text class="Misaligned AI" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-weight="bold" letter-spacing="0em"><tspan x="405.436" y="107.652">Misaligned AI</tspan></text>
<text class="World run by" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="14" letter-spacing="0em"><tspan x="422.104" y="87.7852">World run by</tspan></text>
<text class="Something else" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-weight="bold" letter-spacing="0em"><tspan x="398.053" y="164.652">Something else</tspan></text>
<rect class="Rectangle 3" x="395.5" y="68.5" width="132" height="52" rx="7.5" stroke="var(--darker-gray-color)"/>
<rect class="Rectangle 5" x="395.5" y="140.5" width="132" height="37" rx="7.5" stroke="var(--darker-gray-color)"/>
<text class="or" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-style="italic" letter-spacing="0em"><tspan x="455.476" y="61.6523">or</tspan></text>
<text class="or_2" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-style="italic" letter-spacing="0em"><tspan x="455.476" y="133.652">or</tspan></text>
    </g></g>
<g id="SGWC" opacity="0.3">
<text class="Stable, galaxy-wide civilization" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-weight="bold" letter-spacing="0em"><tspan x="583.109" y="89.6523">Stable, galaxy-wide
</tspan><tspan x="618.222" y="110.652">civilization</tspan></text>
<rect class="Rectangle 6" x="575.5" y="69.5" width="174" height="52" rx="7.5" stroke="var(--darker-gray-color)"/></g>
<g id="ArrowToSGWC">
<path opacity="0.3" class="Arrow 4" d="M562 56.5C562.276 56.5 562.5 56.2761 562.5 56L562.5 51.5C562.5 51.2239 562.276 51 562 51C561.724 51 561.5 51.2239 561.5 51.5L561.5 55.5L557.5 55.5C557.224 55.5 557 55.7239 557 56C557 56.2761 557.224 56.5 557.5 56.5L562 56.5ZM534.646 29.3536L561.646 56.3536L562.354 55.6464L535.354 28.6464L534.646 29.3536Z" fill="var(--darker-gray-color)"/>
<path opacity="0.3" class="Arrow 5" d="M561.354 95.3536C561.549 95.1583 561.549 94.8417 561.354 94.6464L558.172 91.4645C557.976 91.2692 557.66 91.2692 557.464 91.4645C557.269 91.6597 557.269 91.9763 557.464 92.1716L560.293 95L557.464 97.8284C557.269 98.0237 557.269 98.3403 557.464 98.5355C557.66 98.7308
                                       557.976 98.7308 558.172 98.5355L561.354 95.3536ZM535 95.5L561 95.5L561 94.5L535 94.5L535 95.5Z" fill="var(--darker-gray-color)"/>
<path opacity="0.3" class="Arrow 6" d="M561.497 117.053C561.526 116.778 561.327 116.532 561.053 116.503L556.578 116.027C556.304 115.998 556.057 116.196 556.028 116.471C555.999 116.746 556.198 116.992 556.472 117.021L560.45 117.444L560.027 121.422C559.998 121.696 560.196 121.943 560.471 121.972C560.746 122.001 560.992 121.802 561.021 121.528L561.497 117.053ZM535.314 138.389L561.314 117.389L560.686 116.611L534.686 137.611L535.314 138.389Z" fill="var(--darker-gray-color)"/>
</g>
    </svg></a>

<p>
This is the final piece in the <a href="https://www.cold-takes.com/roadmap-for-the-most-important-century-series/">&quot;most important century&quot; series</a>, which has argued that there&apos;s a <a href="https://www.cold-takes.com/forecasting-transformative-ai-whats-the-burden-of-proof/#some-rough-probabilities">high probability</a><sup id="fnref1"><a href="#fn1" rel="footnote">1</a></sup> that the coming decades will see:<ul>
</ul></p>
<p>
<li>The development of a technology like <a href="https://www.cold-takes.com/transformative-ai-timelines-part-1-of-4-what-kind-of-ai/">PASTA </a>(process for automating scientific and technological advancement).
</li></p>
<p>
<li>A resulting <a href="https://www.cold-takes.com/transformative-ai-timelines-part-1-of-4-what-kind-of-ai/#explosive-scientific-and-technological-advancement">productivity explosion</a> leading to development of further transformative technologies.
</li></p>
<p>
<li>The seed of a <a href="https://www.cold-takes.com/all-possible-views-about-humanitys-future-are-wild/">stable galaxy-wide civilization</a>, possibly featuring <a href="https://www.cold-takes.com/how-digital-people-could-change-the-world/">digital people</a>, or possibly run by <a href="https://www.cold-takes.com/transformative-ai-timelines-part-1-of-4-what-kind-of-ai/#misaligned-ai-mysterious-potentially-dangerous-objectives">misaligned AI</a>.
</li></p>
<p>

</p>
<p>
When trying to call attention to an underrated problem, <strong>it&apos;s typical to close on a &quot;call to action&quot;:</strong> a tangible, concrete action readers can take to help.
</p>
<p>
But this is challenging, because as I argued <a href="https://www.cold-takes.com/p/f5eba675-6b0e-4f19-b2ad-b63a6bafc8fd/">previously</a>, there are a lot of <a href="https://www.cold-takes.com/p/f5eba675-6b0e-4f19-b2ad-b63a6bafc8fd/#key-open-questions-for-">open questions about what actions are helpful vs. harmful</a>. (Although we can identify some <a href="https://www.cold-takes.com/p/f5eba675-6b0e-4f19-b2ad-b63a6bafc8fd/#robustly-helpful-actions">actions that seem robustly helpful today</a>.)
</p>
<p>
This makes for a somewhat awkward situation. When confronting the &quot;most important century&quot; hypothesis, my attitude doesn&apos;t match the familiar ones of &quot;excitement and motion&quot; or &quot;fear and avoidance.&quot; Instead, I feel an <strong>odd mix of intensity, urgency, confusion and hesitance.</strong> I&apos;m looking at something bigger than I ever expected to confront, feeling underqualified and ignorant about what to do next. This is a hard mood to share and spread, but I&apos;m trying.
</p>


<table style="border-collapse: collapse;">
  <tr>
   <td style="border: 1px solid; vertical-align: top;"><strong>Situation</strong>
   </td>
   <td style="border: 1px solid; vertical-align: top;"><strong>Appropriate reaction (IMO)</strong>
   </td>
  </tr>
  <tr>
   <td style="border: 1px solid; vertical-align: top;">&quot;This could be a billion-dollar company!&quot;
   </td>
   <td style="border: 1px solid; vertical-align: top;">&quot;Woohoo, let&apos;s GO for it!&quot;
   </td>
  </tr>
  <tr>
   <td style="border: 1px solid; vertical-align: top;">&quot;This could be the most important century!&quot;
   </td>
   <td style="border: 1px solid; vertical-align: top;">&quot;... Oh ... wow ... I don&apos;t know what to say and I somewhat want to vomit ... I have to sit down and think about this one.&quot;
   </td>
  </tr>
</table>


<p>
So instead of a call to action, I want to make a <strong>call to vigilance. </strong>If you&apos;re convinced by the arguments in this piece, then don&apos;t rush to &quot;do something&quot; and then move on. Instead, take whatever <a href="https://www.cold-takes.com/p/f5eba675-6b0e-4f19-b2ad-b63a6bafc8fd/#robustly-helpful-actions">robustly good actions</a> you can today, and otherwise put yourself in a better position to take important actions when the time comes.
</p>
<p>
This could mean:
</p>
<ul>

<li>Finding ways to interact more with, and learn more about, key topics/fields/industries such as AI (for obvious reasons), science and technology generally (as a lot of the &quot;most important century&quot; hypothesis runs through an <a href="https://www.cold-takes.com/transformative-ai-timelines-part-1-of-4-what-kind-of-ai/#impacts-of-pasta">explosion in scientific and technological advancement</a>), and relevant areas of policy and national security.

</li><li>Taking opportunities (when you see them) to move your career in a direction that is more likely to be relevant (some thoughts of mine on this are <a href="https://forum.effectivealtruism.org/posts/bud2ssJLQ33pSemKH/my-current-impressions-on-career-choice-for-longtermists">here</a>; also see <a href="https://www.80000hours.org">80,000 Hours</a>).

</li><li>Connecting with other people interested in these topics (I believe this has been one of the biggest drivers of people coming to do high-impact work in the past). Currently, I think the <a href="https://en.wikipedia.org/wiki/Effective_altruism">effective altruism</a> community is the best venue for this, and you can learn about how to connect with people via the <a href="https://www.centreforeffectivealtruism.org/">Centre for Effective Altruism</a> (see the &quot;Get involved&quot; dropdown). If new ways of connecting with people come up in the future, I will likely post them on Cold Takes.

</li><li>And of course, taking any opportunities you see for <a href="https://www.cold-takes.com/p/f5eba675-6b0e-4f19-b2ad-b63a6bafc8fd/#robustly-helpful-actions">robustly helpful actions.</a>
</li>
</ul>
<h2 id="buttons-you-can-click">Buttons you can click</h2>


<p>
Here&apos;s something you can do right now that would be genuinely helpful, though maybe not as viscerally satisfying as signing a petition or making a donation.
</p>
<p>
In my <a href="https://www.openphilanthropy.org">day job</a>, I have a lot of moments where I - or someone I&apos;m working with - is looking for a particular kind of person (perhaps to fill a job opening with a grantee, or to lend expertise on some topic, or something else). Over time, I expect there to be more and more opportunities for people with specific skills, interests, expertise, etc. to take actions that <a href="https://www.cold-takes.com/p/f5eba675-6b0e-4f19-b2ad-b63a6bafc8fd/">help make the best of the most important century</a>. And I think a major challenge will simply be <strong>knowing who&apos;s out there</strong> - who&apos;s interested in this cause, and wants to help, and what skills and interests they have.
</p>
<p>
If you&apos;re a person we might wish we could find in the future, you can help now by sending in information about yourself via <strong><a href="https://forms.gle/z7mexiTd6wCJsuEv6">this simple form</a></strong>. I vouch that your information won&apos;t be sold or otherwise used to make money, that your communication preferences (which the form asks about in detail) will be respected, and that you&apos;ll always be able to opt out of any communications. 
</p>
<h2 id="sharing-a-headspace">Sharing a headspace</h2>


<p>
In <a href="https://www.cold-takes.com/this-cant-go-on/">This Can&apos;t Go On</a>, I analogized the world to people on a plane blasting down the runway, without knowing why they&apos;re moving so fast or what&apos;s coming next:
</p>
<p><center><img src="https://www.cold-takes.com/content/images/2021/07/airplane-launch-compressed.gif" alt="Call to Vigilance"></center></p>
<p>
As someone sitting on this plane, I&apos;d love to be able to tell you I&apos;ve figured out exactly what&apos;s going on and what future we need to be planning for. But I haven&apos;t. 
</p>
<p>
Lacking answers, I&apos;ve tried to at least show you what I do see: 
</p>
<ul>

<li>Dim outlines of the most important events in humanity&apos;s past or future.

</li><li>A case that they&apos;re approaching us more quickly than it seems - whether or not we&apos;re ready. 

</li><li>A sense that the world and the rules we&apos;re all used to can&apos;t be relied on. That we need to lift our gaze above the daily torrent of tangible, relatable news - and try to wrap our heads around weirder, wilder matters that are more likely to be seen as the <strong>headlines about this era billions of years from now.</strong>
</li>
</ul>
<p>
There&apos;s a lot I don&apos;t know. But if this is the most important century, I do feel confident that we as a civilization aren&apos;t yet up to the challenges it presents. 
</p>
<p>
If that&apos;s going to change, it needs to start with more people seeing the situation for what it is, taking it seriously, taking action when they can - and when not, staying vigilant.
</p>

<!--kg-card-end: html--><!--kg-card-begin: html-->

<p><div style="display:flex; justify-content:center; margin: 0 auto;">

        <span style="margin: 10px;"><a href="https://api.addthis.com/oexchange/0.8/forward/twitter/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fcall-to-vigilance%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20Call%20to%20Vigilance&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-twitter-square.png" border="0" alt="Call to Vigilance"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/facebook/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fcall-to-vigilance%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20Call%20to%20Vigilance&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-facebook-square.png" border="0" alt="Call to Vigilance"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/reddit/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fcall-to-vigilance%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20Call%20to%20Vigilance&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-reddit-square.png" border="0" alt="Call to Vigilance"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/menu/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fcall-to-vigilance%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20Call%20to%20Vigilance&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-addthis-square.png" border="0" alt="Call to Vigilance"></a></span>
        </div>
<center><p id="survey"><a href="https://www.cold-takes.com/call-to-vigilance/#/portal/signup/free" target="_blank"><button class="button">Subscribe</button></a>&#xA0;<a href="https://www.guidedtrack.com/programs/4kal2ue/run?posttitle=Call%20to%20Vigilance" target="_blank"><button class="button" id="Survey">Feedback</button></a>&#xA0;<a href="https://forum.effectivealtruism.org/posts/uNczpTHFvjt4KNxEF/call-to-vigilance#comments" target="_blank"><button class="button">Forum</button></a></p><p><em>
Use &quot;Feedback&quot; if you have comments/suggestions you want me to see, or if you&apos;re up for giving some quick feedback about this post (which I greatly appreciate!) Use &quot;Forum&quot; if you want to discuss this post publicly on the Effective Altruism Forum.
</em></p></center>
<!--kg-card-end: html--><!--kg-card-begin: html-->
<hr>
</p><h2 id="footnotes">Footnotes</h2>
<div class="footnotes">
<ol><li id="fn1">
<p>
     &quot;I am forecasting more than a 10% chance transformative AI will be developed within 15 years (by 2036); a ~50% chance it will be developed within 40 years (by 2060); and a ~2/3 chance it will be developed this century (by 2100).&quot;&#xA0;<a href="#fnref1" rev="footnote">&#x21A9;</a>

</p></li></ol></div><!--kg-card-end: html-->]]></content:encoded></item><item><title><![CDATA[How to make the best of the most important century?]]></title><description><![CDATA[We, the people living in this century, have the chance to have a huge impact on huge numbers of people to come - if we can make sense of the situation enough to find helpful actions. ]]></description><link>https://www.cold-takes.com/making-the-best-of-the-most-important-century/</link><guid isPermaLink="false">6135694ca7ed4b003bf8199d</guid><dc:creator><![CDATA[Holden Karnofsky]]></dc:creator><pubDate>Tue, 14 Sep 2021 18:53:11 GMT</pubDate><media:content url="https://www.cold-takes.com/content/images/2021/09/ai-alignment-views-chart.png" medium="image"/><content:encoded><![CDATA[<!--kg-card-begin: html--><img src="https://www.cold-takes.com/content/images/2021/09/ai-alignment-views-chart.png" alt="How to make the best of the most important century?"><p><figure><iframe title="How to make the best of the most important century?" allowtransparency="true" height="150" width="100%" style="border: none; min-width: min(100%, 430px);" scrolling="no" data-name="pb-iframe-player" src="https://www.podbean.com/player-v2/?i=nzzjg-10dd5a7-pb&amp;from=pb6admin&amp;share=1&amp;download=1&amp;rtl=0&amp;fonts=Arial&amp;skin=1&amp;font-color=auto&amp;btn-skin=7"></iframe><figcaption><em>Audio also available by searching Stitcher, Spotify, Google Podcasts, etc. for &quot;Cold Takes Audio&quot;</em></figcaption></figure></p>


<!--<p><a href="https://www.cold-takes.com/roadmap-for-the-most-important-century-series/"><img src="https://www.cold-takes.com/content/images/2021/08/mic-diagram---today.png"></a></p>-->

<a href="https://www.cold-takes.com/roadmap-for-the-most-important-century-series/"><svg viewbox="0 0 750 178" fill="none" xmlns="http://www.w3.org/2000/svg" id="MasterDiagram">
<g class="Frame 1">
<g id="Today">
<text class="Today&#xE2;&#x20AC;&#x2122;s world" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-weight="bold" letter-spacing="0em"><tspan x="9" y="95.7402">Today&#x2019;s world</tspan></text>
<rect class="Rectangle 1" x="0.5" y="66.5" width="137" height="49" rx="7.5" stroke="var(--darker-gray-color)"/>
</g>
<g id="ArrowToTAI" opacity="0.3">
<path class="Arrow 7" d="M176.341 86.3655C176.543 86.1771 176.554 85.8607 176.366 85.6588L173.295 82.3691C173.107 82.1672 172.79 82.1563 172.588 82.3447C172.387 82.5331 172.376 82.8495 172.564 83.0514L175.293 85.9756L172.369 88.7049C172.167 88.8933 172.156 89.2097 172.345 89.4116C172.533 89.6135 172.85 89.6244 173.051 89.436L176.341 86.3655ZM146.983 85.4997L175.983 86.4997L176.017 85.5003L147.017 84.5003L146.983 85.4997Z" fill="var(--darker-gray-color)"/>
</g>
<g id="Transformative" opacity="0.3">
<text class="Transformative AI" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-weight="bold" letter-spacing="0em"><tspan x="192" y="92.6523">Transformative AI</tspan></text>
<rect class="Rectangle 2" x="182.5" y="65.5" width="165" height="47" rx="7.5" stroke="var(--darker-gray-color)"/>
</g>
<g id="ArrowFromTAI" opacity="0.3">
<path class="Arrow 1" d="M389.491 31.0944C389.543 30.8233 389.366 30.5611 389.094 30.509L384.675 29.6592C384.404 29.607 384.142 29.7846 384.09 30.0558C384.038 30.3269 384.215 30.589 384.487 30.6412L388.415 31.3966L387.659 35.3246C387.607 35.5958 387.785 35.8579 388.056 35.91C388.327 35.9622 388.589 35.7846 388.641 35.5135L389.491 31.0944ZM358.28 52.414L389.28 31.414L388.72 30.586L357.72 51.586L358.28 52.414Z" fill="var(--darker-gray-color)"/>
<path class="Arrow 2" d="M385.342 92.3648C385.543 92.1759 385.554 91.8595 385.365 91.658L382.287 88.3751C382.098 88.1737 381.782 88.1634 381.58 88.3523C381.379 88.5412 381.369 88.8576 381.557 89.0591L384.293 91.9772L381.375 94.713C381.174 94.9018 381.163 95.2182 381.352 95.4197C381.541 95.6212 381.858 95.6314 382.059 95.4425L385.342 92.3648ZM353.984 91.4997L384.984 92.4997L385.016 91.5003L354.016 90.5003L353.984 91.4997Z" fill="var(--darker-gray-color)"/>
<path class="Arrow 3" d="M384.955 148.498C385.23 148.523 385.473 148.32 385.498 148.045L385.905 143.564C385.93 143.289 385.728 143.046 385.453 143.021C385.178 142.996 384.934 143.198 384.909 143.473L384.547 147.457L380.564 147.095C380.289 147.07 380.046 147.272 380.021 147.547C379.996 147.822 380.198 148.066 380.473 148.091L384.955 148.498ZM354.616 112.32L384.616 148.32L385.384 147.68L355.384 111.68L354.616 112.32Z" fill="var(--darker-gray-color)"/>
</g>
<g id="DigitalPeople" opacity="0.3">
<text class="Digital people" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-weight="bold" letter-spacing="0em"><tspan x="401" y="36.6523">Digital people</tspan></text>
<text class="World of" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="14" letter-spacing="0em"><tspan x="435.216" y="16.7852">World of</tspan></text>
<rect class="Digital people bubble" x="394.5" y="0.5" width="133" height="45" rx="7.5" stroke="var(--darker-gray-color)"/>
</g>
<g id="MisalignedAndOther" opacity="0.3">
<text class="Misaligned AI" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-weight="bold" letter-spacing="0em"><tspan x="405.436" y="107.652">Misaligned AI</tspan></text>
<text class="World run by" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="14" letter-spacing="0em"><tspan x="422.104" y="87.7852">World run by</tspan></text>
<text class="Something else" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-weight="bold" letter-spacing="0em"><tspan x="398.053" y="164.652">Something else</tspan></text>
<rect class="Rectangle 3" x="395.5" y="68.5" width="132" height="52" rx="7.5" stroke="var(--darker-gray-color)"/>
<rect class="Rectangle 5" x="395.5" y="140.5" width="132" height="37" rx="7.5" stroke="var(--darker-gray-color)"/>
<text class="or" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-style="italic" letter-spacing="0em"><tspan x="455.476" y="61.6523">or</tspan></text>
<text class="or_2" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-style="italic" letter-spacing="0em"><tspan x="455.476" y="133.652">or</tspan></text>
    </g></g>
<g id="SGWC" opacity="0.3">
<text class="Stable, galaxy-wide civilization" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-weight="bold" letter-spacing="0em"><tspan x="583.109" y="89.6523">Stable, galaxy-wide
</tspan><tspan x="618.222" y="110.652">civilization</tspan></text>
<rect class="Rectangle 6" x="575.5" y="69.5" width="174" height="52" rx="7.5" stroke="var(--darker-gray-color)"/></g>
<g id="ArrowToSGWC">
<path opacity="0.3" class="Arrow 4" d="M562 56.5C562.276 56.5 562.5 56.2761 562.5 56L562.5 51.5C562.5 51.2239 562.276 51 562 51C561.724 51 561.5 51.2239 561.5 51.5L561.5 55.5L557.5 55.5C557.224 55.5 557 55.7239 557 56C557 56.2761 557.224 56.5 557.5 56.5L562 56.5ZM534.646 29.3536L561.646 56.3536L562.354 55.6464L535.354 28.6464L534.646 29.3536Z" fill="var(--darker-gray-color)"/>
<path opacity="0.3" class="Arrow 5" d="M561.354 95.3536C561.549 95.1583 561.549 94.8417 561.354 94.6464L558.172 91.4645C557.976 91.2692 557.66 91.2692 557.464 91.4645C557.269 91.6597 557.269 91.9763 557.464 92.1716L560.293 95L557.464 97.8284C557.269 98.0237 557.269 98.3403 557.464 98.5355C557.66 98.7308
                                       557.976 98.7308 558.172 98.5355L561.354 95.3536ZM535 95.5L561 95.5L561 94.5L535 94.5L535 95.5Z" fill="var(--darker-gray-color)"/>
<path opacity="0.3" class="Arrow 6" d="M561.497 117.053C561.526 116.778 561.327 116.532 561.053 116.503L556.578 116.027C556.304 115.998 556.057 116.196 556.028 116.471C555.999 116.746 556.198 116.992 556.472 117.021L560.45 117.444L560.027 121.422C559.998 121.696 560.196 121.943 560.471 121.972C560.746 122.001 560.992 121.802 561.021 121.528L561.497 117.053ZM535.314 138.389L561.314 117.389L560.686 116.611L534.686 137.611L535.314 138.389Z" fill="var(--darker-gray-color)"/>
</g>
    </svg></a>

<p>
Previously in the <a href="https://www.cold-takes.com/roadmap-for-the-most-important-century-series/">&quot;most important century&quot; series</a>, I&apos;ve argued that there&apos;s a high probability<sup id="fnref1"><a href="#fn1" rel="footnote">1</a></sup> that the coming decades will see:<ul>
</ul></p>
<p>
<li>The development of a technology like <a href="https://www.cold-takes.com/transformative-ai-timelines-part-1-of-4-what-kind-of-ai/">PASTA </a>(process for automating scientific and technological advancement).
</li></p>
<p>
<li>A resulting <a href="https://www.cold-takes.com/transformative-ai-timelines-part-1-of-4-what-kind-of-ai/#explosive-scientific-and-technological-advancement">productivity explosion</a> leading to development of further transformative technologies.
</li></p>
<p>
<li>The seed of a <a href="https://www.cold-takes.com/all-possible-views-about-humanitys-future-are-wild/">stable galaxy-wide civilization</a>, possibly featuring <a href="https://www.cold-takes.com/how-digital-people-could-change-the-world/">digital people</a>, or possibly run by <a href="https://www.cold-takes.com/transformative-ai-timelines-part-1-of-4-what-kind-of-ai/#misaligned-ai-mysterious-potentially-dangerous-objectives">misaligned AI</a>.
</li></p>
<p>
Is this an optimistic view of the world, or a pessimistic one? To me, it&apos;s both and neither, because <strong>this set of events could end up being very good or very bad for the world, depending on the details of how it plays out. </strong>
</p>
<p>
When I talk about being in the &quot;most important century,&quot; I don&apos;t just mean that significant events are going to occur. I mean that we, the people living in this century, have the chance to have a huge impact on huge numbers of people to come - if we can make sense of the situation enough to find helpful actions. 
</p>
<p>
But it&apos;s also important to understand why that&apos;s a big &quot;if&quot; - why the most important century presents a <strong>challenging strategic picture, such that many things we can do might make things better or worse (and it&apos;s hard to say which). </strong>
</p>
<p>
In this post, I will <strong>present two contrasting frames for how to make the best of the most important century: </strong>
</p>
<ul>

<li>The <strong>&quot;Caution&quot;</strong> frame. In this frame, many of the worst outcomes come from developing something like <a href="https://www.cold-takes.com/transformative-ai-timelines-part-1-of-4-what-kind-of-ai/">PASTA </a>in a way that is too fast, rushed, or reckless. We may need to achieve (possibly global) coordination in order to mitigate pressures to race, and take appropriate care. (<a href="#the-caution-frame">Caution</a>)

</li><li>The <strong>&quot;Competition&quot;</strong> frame. This frame focuses not on <em>how and when</em> <a href="https://www.cold-takes.com/transformative-ai-timelines-part-1-of-4-what-kind-of-ai/">PASTA </a>is developed, but <em>who</em> (which governments, which companies, etc.) is first in line to benefit from the resulting productivity explosion. (<a href="#the-competition-frame">Competition</a>)

</li><li>People who take the &quot;caution&quot; frame and people who take the &quot;competition&quot; frame often favor <strong>very different, even contradictory </strong>actions. Actions that look important to people in one frame often look actively harmful to people in the other. 
<ul>
 
<li>I worry that the &quot;competition&quot; frame will be overrated by default, and discuss why below. (<a href="#why-i-fear-">More</a>)
 
</li><li>To gain more clarity on how to weigh these frames and what actions are most likely to be helpful, we need more progress on <strong>open questions</strong> about the size of different types of risks from transformative AI. (<a href="#key-open-questions-for-">Open questions</a>)
</li> 
</ul>

</li><li>In the meantime, there are some <strong>robustly helpful actions</strong> that seem likely to improve humanity&apos;s prospects regardless. (<a href="#robustly-helpful-actions">Robustly helpful actions</a>)
</li>
</ul>
<h2 id="the-caution-frame">The &quot;caution&quot; frame</h2>


<p>
I&apos;ve argued for a good chance that this century will see a transition to a world where <a href="https://www.cold-takes.com/how-digital-people-could-change-the-world/">digital people</a> or <a href="https://www.cold-takes.com/transformative-ai-timelines-part-1-of-4-what-kind-of-ai/#misaligned-ai-mysterious-potentially-dangerous-objectives">misaligned AI</a> (or something else very different from today&apos;s humans) are the major force in world events.
</p>
<p>
The &quot;caution&quot; frame emphasizes that <strong>some types of transition seem better than others.</strong> Listed in order from worst to best:
</p>
<h3 id="worst-misaligned-ai">Worst: Misaligned AI </h3>


<p>
I discussed this possibility <a href="https://www.cold-takes.com/transformative-ai-timelines-part-1-of-4-what-kind-of-ai/#misaligned-ai-mysterious-potentially-dangerous-objectives">previously</a>, drawing on a number of other and more thorough discussions.<sup id="fnref2"><a href="#fn2" rel="footnote">2</a></sup> The basic idea is that AI systems could end up with objectives of their own, and could seek to expand throughout space fulfilling these objectives. Humans, and/or all humans value, could be sidelined (or driven extinct, if we&apos;d otherwise get in the way). 
</p>
<h3 id="next-worst-adversarial-technological-maturity">Next-worst:<sup id="fnref3"><a href="#fn3" rel="footnote">3</a></sup> Adversarial Technological Maturity</h3>
<p></p>
<p>
If we get to the point where there are digital people and/or (non-misaligned) AIs that can copy themselves without limit, and expand throughout space, there might be intense pressure to move - and multiply (via copying) - as fast as possible in order to gain more influence over the world. This might lead to different countries/coalitions furiously trying to outpace each other, and/or to outright military conflict, knowing that a lot could be at stake in a short time.
</p>
<p>
I would expect this sort of dynamic to risk a lot of the galaxy ending up in a bad state.<sup id="fnref4"><a href="#fn4" rel="footnote">4</a></sup>
</p>
<p>
One such bad state would be &quot;permanently under the control of a single (digital) person (and/or their copies).&quot; Due to the potential of digital people to create <a href="https://www.cold-takes.com/how-digital-people-could-change-the-world/#lock-in">stable civilizations</a>, it seems that a given totalitarian regime could end up permanently entrenched across substantial parts of the galaxy.
</p>
<p>
People/countries/coalitions who <em>suspect each other</em> of posing this sort of danger - of potentially establishing stable civilizations under their control - might compete and/or attack each other early on to prevent this. This could lead to war with difficult-to-predict outcomes (due to the difficult-to-predict technological advancements that PASTA could bring about).
</p>
<h3 id="second-best-negotiation-and-governance">Second-best: Negotiation and governance</h3>


<p>
Countries might prevent this sort of <a href="#next-worst-adversarial-technological-maturity">Adversarial Technological Maturity</a> dynamic by planning ahead and negotiating with each other. For example, perhaps each country - or each person - could be allowed to create a certain number of digital people (subject to human rights protections and other regulations), limited to a certain region of space. 
</p>
<p>
It seems there are a huge range of different potential specifics here, some much more good and just than others.
</p>
<h3 id="best-reflection">Best: Reflection</h3>


<p>
The world could achieve a high enough level of coordination to <em>delay</em> any irreversible steps (including kicking off an <a href="#next-worst-adversarial-technological-maturity">Adversarial Technological Maturity</a> dynamic).
</p>
<p>
There could then be something like what Toby Ord (in <a href="https://smile.amazon.com/Precipice-Existential-Risk-Future-Humanity-ebook/dp/B07V9GHKYP/">The Precipice</a>) calls the &quot;Long Reflection&quot;:<sup id="fnref5"><a href="#fn5" rel="footnote">5</a></sup> a sustained period in which people could collectively decide upon goals and hopes for the future, ideally representing the most fair available compromise between different perspectives. Advanced technology could imaginably help this go much better than it could today.<sup id="fnref6"><a href="#fn6" rel="footnote">6</a></sup>
</p>
<p>
There are limitless questions about how such a &quot;reflection&quot; would work, and whether there&apos;s really any hope that it could reach a reasonably good and fair outcome. Details like &quot;what sorts of digital people are created first&quot; could be enormously important. There is currently little discussion of this sort of topic.<sup id="fnref7"><a href="#fn7" rel="footnote">7</a></sup>
</p>
<h3 id="other">Other</h3>


<p>
There are probably many possible types of transitions I haven&apos;t named here. 
</p>
<h3 id="the-role-of-caution">The role of caution</h3>


<p>
If the above ordering is correct, then the future of the galaxy looks better to the extent that:
</p>
<ul>

<li><a href="#worst-misaligned-ai">Misaligned AI</a> is avoided: powerful AI systems act to help humans, rather than pursuing objectives of their own.

</li><li><a href="#next-worst-adversarial-technological-maturity">Adversarial Technological Maturity</a> is avoided. This likely means that people do not deploy advanced AI systems, or the technologies they could bring about, in adversarial ways (unless this ends up necessary to prevent something worse).

</li><li>Enough coordination is achieved so that key players can &quot;take their time,&quot; and <a href="#best-reflection">Reflection</a> becomes a possibility.
</li>
</ul>
<p>
Ideally, everyone with the potential to build something <a href="https://www.cold-takes.com/transformative-ai-timelines-part-1-of-4-what-kind-of-ai/">PASTA</a>-like would be able to pour energy into building something safe (not misaligned), and carefully planning out (and negotiating with others on) how to roll it out, without a rush or a race. With this in mind, perhaps we should be doing things like: 
</p>
<ul>

<li>Working to improve trust and cooperation between major world powers. Perhaps via AI-centric versions of <a href="https://en.wikipedia.org/wiki/Pugwash_Conferences_on_Science_and_World_Affairs">Pugwash</a> (an international conference aimed at reducing the risk of military conflict), perhaps by pushing back against hawkish foreign relations moves.

</li><li>Discouraging governments and investors from shoveling money into AI research, encouraging AI labs to thoroughly consider the implications of their research before publishing it or scaling it up, etc. Slowing things down in this manner could buy more time to do research on avoiding <a href="#worst-misaligned-ai">misaligned AI</a>, more time to build trust and cooperation mechanisms, more time to generally gain strategic clarity, and a lower likelihood of the <a href="#next-worst-adversarial-technological-maturity">Adversarial Technological Maturity</a> dynamic.
</li>
</ul>
<h2 id="the-competition-frame">The &quot;competition&quot; frame</h2>


<p>
(Note: there&apos;s some potential for confusion between the &quot;competition&quot; idea and the <a href="#next-worst-adversarial-technological-maturity">Adversarial Technological Maturity</a> idea, so I&apos;ve tried to use very different terms. I spell out the contrast in a footnote.<sup id="fnref8"><a href="#fn8" rel="footnote">8</a></sup>)
</p>
<p>
The &quot;competition&quot; frame focuses <strong>less on how the transition to a radically different future happens, and more on who&apos;s making the key decisions as it happens.</strong>
</p>
<ul>

<li>If something like <a href="https://www.cold-takes.com/transformative-ai-timelines-part-1-of-4-what-kind-of-ai/">PASTA </a>is developed primarily (or first) in country X, then the government of country X could be making a lot of crucial decisions about whether and how to regulate a potential explosion of new technologies. 

</li><li>In addition, the people and organizations leading the way on AI and other technology advancement at that time could be especially influential in such decisions.
</li>
</ul>
<p>
This means it could matter enormously &quot;who leads the way on transformative AI&quot; - which country or countries, which people or organizations.
</p>
<ul>

<li>Will the governments leading the way on transformative AI be authoritarian regimes?

</li><li>Which governments are most likely to (effectively) have a reasonable understanding of the risks and stakes, when making key decisions?

</li><li>Which governments are least likely to try to use advanced technology for entrenching the power and dominance of one group? (Unfortunately, I can&apos;t say there are any that I feel great about here.) Which are most likely to leave the possibility open for something like &quot;avoiding <a href="https://www.cold-takes.com/how-digital-people-could-change-the-world/#lock-in">locked-in</a> outcomes, leaving time for general progress worldwide to raise the odds of a good outcome for everyone possible?&quot;

</li><li>Similar questions apply to the people and organizations leading the way on transformative AI. Which ones are most likely to push things in a positive direction?
</li>
</ul>
<p>
Some people feel that we can make confident statements today about which specific countries, and/or which people and organizations, we should hope lead the way on transformative AI. These people might advocate for actions like:
</p>
<ul>

<li>Increasing the odds that the first PASTA systems are built in countries that are e.g. less authoritarian, which could mean e.g. pushing for more investment and attention to AI development in these countries. 

</li><li>Supporting and trying to speed up AI labs run by people who are likely to make wise decisions (about things like how to engage with governments, what AI systems to publish and deploy vs. keep secret, etc.)
</li>
</ul>
<h2 id="why-i-fear-" competition"-being-overrated-relative-to-"caution"">Why I fear &quot;competition&quot; being overrated, relative to &quot;caution&quot;</h2>


<p>
By default, I expect a lot of people to gravitate toward the &quot;competition&quot; frame rather than the &quot;caution&quot; frame - for reasons that I don&apos;t think are great, such as:
</p>
<ul>

<li>I think people naturally get more animated about &quot;helping the good guys beat the bad guys&quot; than about &quot;helping all of us avoid getting a universally bad outcome, for impersonal reasons such as &apos;we designed sloppy AI systems&apos; or &apos;we created a dynamic in which haste and aggression are rewarded.&apos;&quot;

</li><li>I expect people will tend to be overconfident about which countries, organizations or people they see as the &quot;good guys.&quot; 

</li><li>Embracing the &quot;competition&quot; frame tends to point toward taking actions - such as working to speed up a particular country&apos;s or organization&apos;s AI development - that are lucrative, exciting and naturally easy to feel energy for. Embracing the &quot;caution&quot; frame is much less this way.

</li><li>The biggest concerns that the &quot;caution&quot; frame focuses on - <a href="#worst-misaligned-ai">Misaligned AI</a> and <a href="#next-worst-adversarial-technological-maturity">Adversarial Technological Maturity</a> - are a bit abstract and hard to wrap one&apos;s head around. In many ways they seem to be the highest-stakes risks, but it&apos;s easier to be viscerally scared of &quot;falling behind countries/organizations/people that scare me&quot; than to be viscerally scared of something like &quot;Getting a bad outcome for the long-run future of the galaxy because we rushed things this century.&quot; 
<ul>
 
<li>I think <a href="#worst-misaligned-ai">Misaligned AI</a> is a particularly hard risk for many to take seriously. It sounds wacky and sci-fi-like; people who worry about it tend to be interpreted as picturing something like The Terminator, and it can be hard for their more detailed concerns to be understood.
 
</li><li>I&apos;m hoping to run more posts in the future that help give an intuitive sense for why I think Misaligned AI is a real risk.
</li> 
</ul>
</li> 
</ul>
<p>
So for the avoidance of doubt, I&apos;ll state that I think the &quot;caution&quot; frame has an awful lot going for it. In particular,<strong> <a href="#worst-misaligned-ai">Misaligned AI</a> and <a href="#next-worst-adversarial-technological-maturity">Adversarial Technological Maturity</a> seem a <em>lot</em> worse than other potential transition types,</strong> and both seem like things that have a real chance of making the entire future of our species (and successors) much worse than they could be.
</p>
<p>
I worry that too much of the &quot;competition&quot; frame will lead to downplaying misalignment risk and rushing to deploy unsafe, unpredictable systems, which could have many negative consequences. 
</p>
<p>
With that said, <strong>I put serious weight on both frames. </strong>I remain quite uncertain overall about which frame is more important and helpful (if either is).
</p>
<h2 id="key-open-questions-for-">Key open questions for &quot;caution&quot; vs. &quot;competition&quot;</h2>


<p>
People who take the &quot;caution&quot; frame and people who take the &quot;competition&quot; frame often favor <strong>very different, even contradictory actions</strong>. Actions that look important to people in one frame often look actively harmful to people in the other. 
</p>
<p>
For example, people in the &quot;competition&quot; frame often favor moving forward as fast as possible on developing more powerful AI systems; for people in the &quot;caution&quot; frame, haste is one of the main things to avoid. People in the &quot;competition&quot; frame often favor adversarial foreign relations, while people in the &quot;caution&quot; frame often want foreign relations to be more cooperative.
</p>
<p>
(That said, this dichotomy is a simplification. Many people - including myself - resonate with both frames. And either frame could imply actions normally associated with the other; for example, you might take the &quot;caution&quot; frame but feel that haste is needed now in order to establish one country with a clear enough lead in AI that it can then take its time, prioritize avoiding <a href="#worst-misaligned-ai">misaligned AI</a>, etc.)
</p>
<p>
I wish I could confidently tell you how much weight to put on each frame, and what actions are most likely to be helpful. But I can&apos;t. I think we would have more clarity if we had better answers to some key open questions:
</p>
<h3 id="open-question-how-hard-is-the-alignment-problem">Open question: how hard is the alignment problem?</h3>


<p>
The path to the future that seems worst is <a href="#worst-misaligned-ai">Misaligned AI</a>, in which AI systems end up with non-human-compatible objectives of their own and seek to fill the galaxy according to those objectives. How seriously should we take this risk - how hard will it be to avoid this outcome? <strong>How hard will it be to solve the &quot;alignment problem,&quot; </strong>which essentially means having the technical ability to build systems that won&apos;t do this?<sup id="fnref9"><a href="#fn9" rel="footnote">9</a></sup>
</p>
<ul>

<li>Some people believe that the alignment problem will be formidable; that our only hope of solving it comes in a world where we have enormous amounts of time and aren&apos;t in a race to deploy advanced AI; and that avoiding the &quot;Misaligned AI&quot; outcome should be by far the dominant consideration for the most important century. These people tend to heavily favor the &quot;caution&quot; interventions described above: they believe that rushing toward AI development raises our already-substantial risk of the worst possible outcome.

</li><li>Some people believe it will be easy, and/or that the whole idea of &quot;misaligned AI&quot; is misguided, silly, or even incoherent - planning for an overly specific future event. These people often are more interested in the &quot;competition&quot; interventions described above: they believe that advanced AI will probably be used effectively by whatever country (or in some cases smaller coalition or company) develops it first, and so the question is who will develop it first.

</li><li>And many people are somewhere in between.
</li>
</ul>
<p>
The spread here is extreme. For example, see <a href="https://www.lesswrong.com/posts/QvwSr5LsxyDeaPK5s/existential-risk-from-ai-survey-results">these results</a> from an informal &quot;two-question survey [sent] to ~117 people working on long-term AI risk, asking about the level of existential risk from &apos;humanity not doing enough technical AI safety research&apos; and from &apos;AI systems not doing/optimizing what the people deploying them wanted/intended.&apos;&quot; (As the scatterplot shows, people gave similar answers to the two questions.)
</p>
<p>

<!--<figure class="kg-card kg-image-card kg-width-wide"><img src="https://www.cold-takes.com/content/images/size/w1140/2021/09/ai-alignment-views-chart.png" width=1036><figcaption>&nbsp;</figcaption></figure>-->

    
    <img src="https://www.cold-takes.com/content/images/size/w1140/2021/09/ai-alignment-views-chart.png" alt="How to make the best of the most important century?">

</p>
<p>
We have respondents who think there&apos;s a <5% chance that alignment issues will drastically reduce the goodness of future; respondents who think there's a>95% chance; and just about everything in between.<sup id="fnref10"><a href="#fn10" rel="footnote">10</a></sup> My sense is that this is a fair representation of the situation: even among the few people who have spent the most time thinking about these matters, there is practically no consensus or convergence on how hard the alignment problem will be.
</5%></p>
<p>
I hope that over time, the field of people doing research on AI alignment<sup id="fnref11"><a href="#fn11" rel="footnote">11</a></sup> will grow, and as both AI and AI alignment research advance, we will gain clarity on the difficulty of the AI alignment problem. This, in turn, could give more clarity on prioritizing &quot;caution&quot; vs. &quot;competition.&quot;
</p>
<h3 id="other-open-questions">Other open questions</h3>


<p>
Even if we had clarity on the difficulty of the alignment problem, a lot of thorny questions would remain. 
</p>
<p>
Should we be expecting transformative AI within the next 10-20 years, or much later? Will the leading AI systems go from very limited to very capable quickly (&quot;hard takeoff&quot;) or gradually (&quot;slow takeoff&quot;)?<sup id="fnref12"><a href="#fn12" rel="footnote">12</a></sup> Should we hope that government projects play a major role in AI development, or that transformative AI primarily emerges from the private sector? Are some governments more likely than others to work toward transformative AI being used carefully, inclusively and humanely? What should we hope a government (or company) literally <em>does</em> if it gains the ability to dramatically accelerate scientific and technological advancement via AI?
</p>
<p>
With these questions and others in mind, it&apos;s often very hard to look at some action - like starting a new AI lab, advocating for more caution and safeguards in today&apos;s AI development, etc. - and say whether it raises the likelihood of good long-run outcomes. 
</p>
<h2 id="robustly-helpful-actions">Robustly helpful actions</h2>


<p>
Despite this state of uncertainty, here are a few things that do seem clearly valuable to do today:
</p>
<p>
<strong>Technical research on the alignment problem. </strong>Some researchers work on building AI systems that can get &quot;better results&quot; (winning more board games, classifying more images correctly, etc.) But a smaller set of researchers works on things like:
</p>
<ul>

<li><a href="https://openai.com/blog/learning-to-summarize-with-human-feedback/">Training AI systems to incorporate human feedback into how they perform summarization tasks</a>, so that the AI systems reflect hard-to-define human preferences - something it may be important to be able to do in the future.

</li><li><a href="https://openai.com/blog/microscope/">Figuring out how to understand &quot;what AI systems are thinking and how they&apos;re reasoning,&quot;</a> in order to make them less mysterious.

</li><li><a href="https://ai.googleblog.com/2018/09/introducing-unrestricted-adversarial.html">Figuring out how to stop AI systems from making extremely bad judgments on images designed to fool them</a>, and other work focused on helping avoid the &quot;worst case&quot; behaviors of AI systems. 

</li><li><a href="https://alignmentresearchcenter.org/">Theoretical work </a>on how an AI system might be very advanced, yet not be unpredictable in the wrong ways.
</li>
</ul>
<p>
This sort of work could both reduce the risk of the <a href="#worst-misaligned-ai">Misaligned AI</a> outcome - and/or lead to more clarity on just how big a threat it is. Some takes place in academia, some at AI labs, and some at specialized organizations.
</p>
<p>
<strong>Pursuit of strategic clarity: </strong>doing research that could address other crucial questions (such as those listed <a href="#other-open-questions">above</a>), to help clarify what sorts of immediate actions seem most useful.
</p>
<p>
<strong>Helping governments and societies become, well, nicer. </strong>Helping Country X get ahead of others on AI development could make things better or worse, for reasons given above. But it seems robustly good to work toward a Country X with better, more inclusive values, and a government whose key decision-makers are more likely to make thoughtful, good-values-driven decisions.
</p>
<p>
<strong>Spreading ideas and building communities. </strong>Today, it seems to me that the world is <strong>extremely short on people who share certain basic expectations and concerns</strong>, such as:
</p>
<ul>

<li>Believing that AI research could lead to rapid, radical changes of the <a href="https://www.cold-takes.com/transformative-ai-timelines-part-1-of-4-what-kind-of-ai/#impacts-of-pasta">extreme kind laid out here</a> (well beyond things like e.g. increasing unemployment).

</li><li>Believing that the alignment problem (discussed <a href="#open-question-how-hard-is-the-alignment-problem">above</a>) is at least plausibly a real concern, and taking <a href="#the-caution-frame">the &quot;caution&quot; frame</a> seriously.

</li><li>Looking at the whole situation through a lens of &quot;Let&apos;s get the best outcome possible for the whole world over the long future,&quot; as opposed to more common lenses such as &quot;Let&apos;s try to make money&quot; or &quot;Let&apos;s try to ensure that my home country leads the world in AI research.&quot;
</li>
</ul>
<p>
I think it&apos;s very valuable for there to be more people with this basic lens, particularly working for AI labs and governments. If and when we have more strategic clarity about what actions could maximize the odds of the &quot;most important century&quot; going well, I expect such people to be relatively well-positioned to be helpful. 
</p>
<p>
A number of organizations and people have worked to expose people to the lens above, and help them meet others who share it. I think a good amount of progress (in terms of growing communities) has come from this.
</p>
<p>
<strong>Donating? </strong>One can donate today to places like <a href="https://funds.effectivealtruism.org/funds/far-future">this</a>. But I need to admit that very broadly speaking, there&apos;s no easy translation right now between &quot;money&quot; and &quot;improving the odds that the most important century goes well.&quot; It&apos;s not the case that if one simply sent, say, $1 trillion to the right place, we could all breathe easy about challenges like the alignment problem and <a href="https://www.cold-takes.com/how-digital-people-could-change-the-world/#would-these-impacts-be-a-good-or-bad-thing">risks of digital dystopias</a>.
</p>
<p>
It seems to me that we - as a species - are currently terribly short on people who are paying <em>any</em> attention to the most important challenges ahead of us, and haven&apos;t done the work to have good strategic clarity about what tangible actions to take. <strong>We can&apos;t solve this problem by throwing money at it.</strong><sup id="fnref13"><a href="#fn13" rel="footnote">13</a></sup><strong> First, we need to take it more seriously and understand it better.</strong>
</p>

<p><strong>Next (and last) in series:</strong> <a href="https://www.cold-takes.com/call-to-vigilance/">Call to Vigilance</a></p><!--kg-card-end: html--><!--kg-card-begin: html-->

<p><div style="display:flex; justify-content:center; margin: 0 auto;">

        <span style="margin: 10px;"><a href="https://api.addthis.com/oexchange/0.8/forward/twitter/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fmaking-the-best-of-the-most-important-century%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20How%20to%20make%20the%20best%20of%20the%20most%20important%20century%3F&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-twitter-square.png" border="0" alt="How to make the best of the most important century?"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/facebook/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fmaking-the-best-of-the-most-important-century%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20How%20to%20make%20the%20best%20of%20the%20most%20important%20century%3F&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-facebook-square.png" border="0" alt="How to make the best of the most important century?"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/reddit/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fmaking-the-best-of-the-most-important-century%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20How%20to%20make%20the%20best%20of%20the%20most%20important%20century%3F&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-reddit-square.png" border="0" alt="How to make the best of the most important century?"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/menu/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fmaking-the-best-of-the-most-important-century%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20How%20to%20make%20the%20best%20of%20the%20most%20important%20century%3F&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-addthis-square.png" border="0" alt="How to make the best of the most important century?"></a></span>
        </div>
<center><p id="survey"><a href="https://www.cold-takes.com/making-the-best-of-the-most-important-century/#/portal/signup/free" target="_blank"><button class="button">Subscribe</button></a>&#xA0;<a href="https://www.guidedtrack.com/programs/4kal2ue/run?posttitle=How%20to%20make%20the%20best%20of%20the%20most%20important%20century%3F" target="_blank"><button class="button" id="Survey">Feedback</button></a>&#xA0;<a href="https://forum.effectivealtruism.org/posts/Lbtcjfxhrs8kfKK2M/how-to-make-the-best-of-the-most-important-century-1#comments" target="_blank"><button class="button">Forum</button></a></p><p><em>
Use &quot;Feedback&quot; if you have comments/suggestions you want me to see, or if you&apos;re up for giving some quick feedback about this post (which I greatly appreciate!) Use &quot;Forum&quot; if you want to discuss this post publicly on the Effective Altruism Forum.
</em></p></center>
<!--kg-card-end: html--><!--kg-card-begin: html--><hr>
</p><h2>Footnotes</h2>
<div class="footnotes">
<ol><li id="fn1">
<p>
     From <a href="https://www.cold-takes.com/forecasting-transformative-ai-whats-the-burden-of-proof/#some-rough-probabilities">Forecasting Transformative AI: What&apos;s the Burden of Proof?</a>: &quot;I am forecasting more than a 10% chance transformative AI will be developed within 15 years (by 2036); a ~50% chance it will be developed within 40 years (by 2060); and a ~2/3 chance it will be developed this century (by 2100).&quot;
</p><p>
    Also see <a href="https://www.cold-takes.com/some-additional-detail-on-what-i-mean-by-most-important-century/">Some additional detail on what I mean by &quot;most important century.&quot;</a>&#xA0;<a href="#fnref1" rev="footnote">&#x21A9;</a><li id="fn2">
<p>
     These include the books <a href="https://smile.amazon.com/Superintelligence-Dangers-Strategies-Nick-Bostrom-ebook/dp/B00LOOCGB2/">Superintelligence</a>, <a href="https://smile.amazon.com/Human-Compatible-Artificial-Intelligence-Problem-ebook/dp/B07N5J5FTS">Human Compatible</a>, <a href="https://smile.amazon.com/Life-3-0-Being-Artificial-Intelligence-ebook/dp/B06WGNPM7V">Life 3.0</a>, and <a href="https://smile.amazon.com/Alignment-Problem-Machine-Learning-Values-ebook/dp/B085T55LGK/">The Alignment Problem</a>. The shortest, most accessible presentation I know of is <a href="https://www.vox.com/future-perfect/2018/12/21/18126576/ai-artificial-intelligence-machine-learning-safety-alignment">The case for taking AI seriously as a threat to humanity</a> (Vox article by Kelsey Piper). This <a href="https://www.alignmentforum.org/posts/HduCjmXTBD4xYTegv/draft-report-on-existential-risk-from-power-seeking-ai">report on existential risk from power-seeking AI</a>, by Open Philanthropy&apos;s Joe Carlsmith, lays out a detailed set of premises that would collectively imply the problem is a serious one.&#xA0;<a href="#fnref2" rev="footnote">&#x21A9;</a><li id="fn3">
<p>
     The order of goodness isn&apos;t absolute, of course. There are versions of &quot;Adversarial Technological Maturity&quot; that could be worse than &quot;Misaligned AI&quot; - for example, if the former results in power going to those who deliberately inflict suffering.&#xA0;<a href="#fnref3" rev="footnote">&#x21A9;</a><li id="fn4">
<p>
     Part of the reason for this is that faster-moving, less-careful parties could end up quickly outnumbering others and determining the future of the galaxy. There is also a longer-run risk discussed in Nick Bostrom&apos;s <a href="https://www.nickbostrom.com/fut/evolution.html">The Future of Human Evolution</a>; also see <a href="https://slatestarcodex.com/2014/07/13/growing-children-for-bostroms-disneyland/">this discussion </a>of Bostrom&apos;s ideas on Slate Star Codex, though also see <a href="http://reflectivedisequilibrium.blogspot.com/2012/09/spreading-happiness-to-stars-seems.html">this piece by Carl Shulman</a> arguing that this dynamic is unlikely to result in total elimination of nice things.&#xA0;<a href="#fnref4" rev="footnote">&#x21A9;</a><li id="fn5">
<p>
     See page 191.&#xA0;<a href="#fnref5" rev="footnote">&#x21A9;</a><li id="fn6">
<p>
     E.g., see <a href="https://www.cold-takes.com/how-digital-people-could-change-the-world/#social-science">this section</a> of <a href="https://www.cold-takes.com/how-digital-people-could-change-the-world/">Digital People Would Be An Even Bigger Deal</a>.&#xA0;<a href="#fnref6" rev="footnote">&#x21A9;</a><li id="fn7">
<p>
     One relevant paper: <a href="https://nickbostrom.com/papers/aipolicy.pdf">Public Policy and Superintelligent AI: A Vector Field Approach</a> by Bostrom, Dafoe and Flynn.&#xA0;<a href="#fnref7" rev="footnote">&#x21A9;</a><li id="fn8">
<p>
      <a href="#next-worst-adversarial-technological-maturity">Adversarial Technological Maturity </a>refers to a world in which highly advanced technology has <strong>already been developed,</strong> likely with the help of AI, and different coalitions are vying for influence over the world. By contrast, &quot;Competition&quot; refers to a strategy for how to behave <strong><em>before </em>the development of advanced AI</strong>. One might imagine a world in which some government or coalition takes a &quot;competition&quot; frame, develops advanced AI long before others, and then makes a series of good decisions that <em>prevent</em> Adversarial Technological Maturity. (Or conversely, a world in which failure to do well at &quot;competition&quot; raises the risks of Adversarial Technological Maturity.)&#xA0;<a href="#fnref8" rev="footnote">&#x21A9;</a><li id="fn9">
<p>
     See definitions of this problem at <a href="https://en.wikipedia.org/wiki/AI_control_problem">Wikipedia</a> and <a href="https://ai-alignment.com/clarifying-ai-alignment-cec47cd69dd6">Paul Christiano&apos;s Medium</a>.&#xA0;<a href="#fnref9" rev="footnote">&#x21A9;</a><li id="fn10">
<p>
     A more detailed, private survey done for <a href="https://www.alignmentforum.org/posts/HduCjmXTBD4xYTegv/draft-report-on-existential-risk-from-power-seeking-ai">this report</a>, asking about the probability of &quot;doom&quot; before 2070 due to the type of problem discussed in the report, got answers ranging from <1% to>50%. In my opinion, there are very thoughtful people who have seriously considered these matters at both ends of that range.&#xA0;<a href="#fnref10" rev="footnote">&#x21A9;</a><li id="fn11">
<p>
     Some example technical topics <a href="https://www.openphilanthropy.org/focus/global-catastrophic-risks/potential-risks-advanced-artificial-intelligence/the-open-phil-ai-fellowship#examples">here</a>.&#xA0;<a href="#fnref11" rev="footnote">&#x21A9;</a><li id="fn12">
<p>
     Some discussion of this topic here: <a href="https://www.alignmentforum.org/posts/YgNYA6pj2hPSDQiTE/distinguishing-definitions-of-takeoff">Distinguishing definitions of takeoff - AI Alignment Forum</a>&#xA0;<a href="#fnref12" rev="footnote">&#x21A9;</a><li id="fn13">
<p>
     Some more thought on &quot;when money isn&apos;t enough&quot; at <a href="https://blog.givewell.org/2013/08/29/we-cant-simply-buy-capacity/">this old GiveWell post</a>..&#xA0;<a href="#fnref13" rev="footnote">&#x21A9;</a>

</p></li></p></li></p></li></1%></p></li></p></li></p></li></p></li></p></li></p></li></p></li></p></li></p></li></p></li></ol></div><!--kg-card-end: html-->]]></content:encoded></item><item><title><![CDATA[One Cold Link: “The Past and Future of Economic Growth: A Semi-Endogenous Perspective”]]></title><description><![CDATA[A paper asking big questions about what has powered economic growth over the last 50+ years, and what the long-run prospects look like.]]></description><link>https://www.cold-takes.com/past-and-future-of-economic-growth-paper/</link><guid isPermaLink="false">6139c26f0f5c9a003ee4a410</guid><dc:creator><![CDATA[Holden Karnofsky]]></dc:creator><pubDate>Thu, 09 Sep 2021 18:47:15 GMT</pubDate><content:encoded><![CDATA[<!--kg-card-begin: html-->

<p>
<a href="https://www.nber.org/papers/w29126">The Past and Future of Economic Growth: A Semi-Endogenous Perspective</a> is a growth economics paper by Charles I. Jones, asking big questions about what has powered economic growth<sup id="fnref1"><a href="#fn1" rel="footnote">1</a></sup> over the last 50+ years, and what the long-run prospects for continued economic growth look like. I think the ideas in it will be unfamiliar to most people, but they make a good amount of intuitive sense; and if true, they seem very important for thinking about the long-run future of the economy.
</p>
<p>
Key quotes, selected partly for comprehensibility to laypeople and ordered so that you should be able to pick up the gist of the paper by reading them:
</p>
<p>
&#x201C;Where do ideas come from? The history of innovation is very clear on this point: new ideas are discovered through the hard work and serendipity of people. Just as more autoworkers will produce more cars, more researchers and innovators will produce more new ideas &#x2026; The surprise is that we are now done; that is all we need for the semi-endogenous model of economic growth. People produce ideas and ... those ideas raise everyone&#x2019;s income ... the growth rate of income per person depends on the growth rate of researchers, which is in turn ultimately equal to the growth rate of the population.&#x201D;
</p>
<p>
A key idea not explicitly stated in that quote, but emphasized elsewhere in the paper, is that <a href="https://web.stanford.edu/~chadj/IdeaPF.pdf">ideas get harder to find</a>: so if you want to maintain the same rate of innovation, you need more and more researchers over time. This is a simple model that can potentially help explain some otherwise odd-seeming phenomena, such as the fact that <a href="https://www.theatlantic.com/science/archive/2018/11/diminishing-returns-science/575665/">science seems to be &#x201C;slowing down.&#x201D;</a> Basically, it&#x2019;s possible that how much innovation we get is just a function of how many people are working on innovating - and we need more people over time to keep up the same rate.</p>

<p>So in the short run, you can get more innovation via things like more researcher jobs and better education, but in the long run, the only route is more population.
</p>
<p>
&#x201C;Even in this &#x2026; framework in which population growth is the only potential source of growth in the long run, other factors explain more than 80% of U.S. growth in recent decades: the contribution of population growth is 0.3% out of the 2% growth we observe. In other words, the level effects associated with rising educational attainment, declining misallocation, and rising research intensity have been overwhelmingly important for the past 50+ years.&#x201D;
</p>
<p>
&#x201C;The point to emphasize here is that this framework strongly implies that, unless something dramatic changes, future growth rates will be substantially lower. In particular, all the sources other than population growth are inherently transitory, and once these sources have run their course, all that will remain is the 0.3 percentage point contribution from population growth. In other words &#x2026; the implication is that long-run growth in living standards will be 0.3% per year rather than 2% per year &#x2014; an enormous slowdown!&#x201D;
</p>
<p>
&#x201C;if population growth is negative, these idea-driven models predict that living standards stagnate for a population that vanishes! This is a stunningly negative result, especially when compared to the standard result we have been examining throughout the paper. In the usual case with positive population growth, living standards rise exponentially forever for a population that itself rises exponentially. Whether we live in an &#x201C;expanding cosmos&#x201D; or an &#x201C;empty planet&#x201D; depends, remarkably, on whether the total fertility rate is above or below a number like 2 or 2.1.&#x201D;
</p>
<p>
&#x201C;Peters and Walsh (2021) ... find that declining population growth generates lower entry, reduced creative destruction, increased concentration, rising markups, and lower productivity growth, all facts that we see in the firm-level data.&#x201D;
</p>
<p>
So far, the implication is:<ul>
    <li>In the short run, we&#x2019;ve had high growth for reasons that can&apos;t continue indefinitely. (For example, one such factor is a rising share of the population that has a certain level of education, but that share can&apos;t go above 100%. The high-level point is that if we want more researchers, we can only get that via a higher population or a higher % of people who are researchers, and the latter can only go so high.) 
        </li><li>In the long run, growth (in living standards) basically comes down to population growth.</li></ul>
</p>
<p>
But the paper also gives two reasons that growth could rise instead of falling.
</p>
<p>
Reason one:
</p>
<p>
&#x201C;The world contains more than 7 billion people. However, according to the OECD&#x2019;s Main Science and Technology Indicators, the number of full-time equivalent researchers in the world appears to be less than 10 million. In other words something on the order of one or two out of every thousand people in the world is engaged in research ... There is ample scope for substantially increasing the number of researchers over the next century, even if population growth slows or is negative. I see three ways this &#x2018;finding new Einsteins&#x2019; can occur &#x2026; 
</p>
<p>
&#x201C;<strong>The rise of China, India, and other countries.</strong> The United States, Western Europe, and Japan together have about 1 billion people, or only about 1/7th the world&#x2019;s population. China and India each have this many people. As economic development proceeds in China, India, and throughout the world, the pool from which we may find new talented inventors will multiply. How many Thomas Edisons and Jennifer Doudnas have we missed out on among these billions of people because they lacked education and opportunity?
</p>
<p>
&#x201C;<strong>Finding new Doudnas: women in research.</strong> Another huge pool of underutilized talent is women &#x2026;. Brouillette (2021) uses patent data to document that in 1976 less than 3 percent of U.S. inventors were women. Even as of 2016 the share was less than 12 percent. He estimates that eliminating the barriers that lead to this misallocation of talent could raise economic growth in the United States by up to 0.3 percentage points per year over the next century.
</p>
<p>
&#x201C;<strong>Other sources of within-country talent.</strong> Bell, Chetty, Jaravel, Petkova and Van Reenen (2019) document that the extent to which people are exposed to inventive careers in childhood has a large influence on who becomes an inventor. They show that exposure in childhood is limited for girls, people of certain races, and people in low-income neighborhoods, even conditional on math test scores in grade school, and refer to these missed opportunities as &#x2018;lost Einsteins.&#x2019;&#x201D;
</p>
<p>
The other reason that growth could rise will be familiar to readers of this blog:
</p>
<p>
&#x201C;Another potential reason for optimism about future growth prospects is the possibility of automation, both in the production of goods and in the production of ideas &#x2026; [according to a particular model,] an increase in the automation of tasks in idea production (&#x2191;&#x3B1;) causes the growth rate of the economy to increase &#x2026; if the fraction of tasks that are automated (&#x3B1;) rises to reach the rate at which ideas are getting harder to find (&#x3B2;), we get a singularity! [Caveats follow]&#x201D;
</p>
<p>
Oversimplified recap: innovation comes down to the number of researchers; some key recent sources of growth in this can&apos;t continue indefinitely; if population growth stagnates, eventually so must innovation and living standards; but we could get more researchers via lowering barriers to entry and/or via AI and automation (and/or via more population growth).
</p>
<p>
None of these claims are empirical, settled science. They all are implications of what I believe are the leading simple models of economic growth. But to me they all make good sense, and I think the reason they aren&#x2019;t more &quot;in the water&quot; is because people don&#x2019;t tend to talk about the drivers of the <em>long-run past and future</em> of economic growth (as I have <a href="https://www.cold-takes.com/this-cant-go-on/#neglected-possibilities">complained previously</a>!)
</p>
<p>
Here are Leopold Aschenbrenner&#x2019;s <a href="https://www.forourposterity.com/best-chad-jones-papers/">favorite papers by the same author </a>(including this one). 
</p>

<!--kg-card-end: html--><!--kg-card-begin: html-->

<p><div style="display:flex; justify-content:center; margin: 0 auto;">

        <span style="margin: 10px;"><a href="https://api.addthis.com/oexchange/0.8/forward/twitter/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fpast-and-future-of-economic-growth-paper%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20One%20Cold%20Link%3A%20%E2%80%9CThe%20Past%20and%20Future%20of%20Economic%20Growth%3A%20A%20Semi-Endogenous%20Perspective%E2%80%9D&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-twitter-square.png" border="0" alt="Twitter"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/facebook/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fpast-and-future-of-economic-growth-paper%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20One%20Cold%20Link%3A%20%E2%80%9CThe%20Past%20and%20Future%20of%20Economic%20Growth%3A%20A%20Semi-Endogenous%20Perspective%E2%80%9D&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-facebook-square.png" border="0" alt="Facebook"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/reddit/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fpast-and-future-of-economic-growth-paper%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20One%20Cold%20Link%3A%20%E2%80%9CThe%20Past%20and%20Future%20of%20Economic%20Growth%3A%20A%20Semi-Endogenous%20Perspective%E2%80%9D&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-reddit-square.png" border="0" alt="Reddit"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/menu/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fpast-and-future-of-economic-growth-paper%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20One%20Cold%20Link%3A%20%E2%80%9CThe%20Past%20and%20Future%20of%20Economic%20Growth%3A%20A%20Semi-Endogenous%20Perspective%E2%80%9D&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-addthis-square.png" border="0" alt="More"></a></span>
        </div>
<center><p id="survey"><a href="https://www.cold-takes.com/past-and-future-of-economic-growth-paper/#/portal/signup/free" target="_blank"><button class="button">Subscribe</button></a>&#xA0;<a href="https://www.guidedtrack.com/programs/4kal2ue/run?posttitle=One%20Cold%20Link%3A%20%E2%80%9CThe%20Past%20and%20Future%20of%20Economic%20Growth%3A%20A%20Semi-Endogenous%20Perspective%E2%80%9D" target="_blank"><button class="button" id="Survey">Feedback</button></a><!--No Forum post--></p></center>
<!--kg-card-end: html--><!--kg-card-begin: html--><hr>
</p><h2>Footnotes</h2>
<div class="footnotes">
<ol><li id="fn1">
<p>
     You can try <a href="https://www.cold-takes.com/what-is-economic-growth/">this short explanation</a> if you don&#x2019;t know what economic growth is.&#xA0;<a href="#fnref1" rev="footnote">&#x21A9;</a>

</p></li></ol></div><!--kg-card-end: html--><!--kg-card-begin: html--><p style="font-size:1%">For email filter: florpschmop</p><!--kg-card-end: html-->]]></content:encoded></item><item><title><![CDATA[Medium Lights]]></title><description><![CDATA[It's great when someone is able to walk you through why they're this amazed and impressed with something you wouldn't have noticed.]]></description><link>https://www.cold-takes.com/medium-lights/</link><guid isPermaLink="false">6137d50ea7ed4b003bf82145</guid><dc:creator><![CDATA[Holden Karnofsky]]></dc:creator><pubDate>Wed, 08 Sep 2021 16:26:46 GMT</pubDate><media:content url="https://www.cold-takes.com/content/images/2021/09/floater-annotated-1.png" medium="image"/><content:encoded><![CDATA[<!--kg-card-begin: html--><!--<p><img src="https://www.cold-takes.com/content/images/size/w1140/2021/09/floater-annotated-1.png"></p>-->

<img src="https://www.cold-takes.com/content/images/2021/09/floater-annotated-1.png" alt="Medium Lights"><p>
I like this<a href="https://mediumlights.substack.com/"> basketball substack called Medium Lights</a> that explains all kinds of cool basketball plays the announcers wouldn&apos;t necessarily comment on. Unlike most sites like this, it usually explains in enough detail that I can actually follow what&apos;s cool about the play. It really highlights how much more interesting games could be with good announcers.
</p>
<p>
A few I&apos;ve especially enjoyed:<a href="https://mediumlights.substack.com/p/lebron-baits-kirk-hinrich"> </a>
</p>
<ul>

<li><a href="https://mediumlights.substack.com/p/lebron-baits-kirk-hinrich">LeBron lets people drive past him so he can block their shot from behind</a>.

</li><li><a href="https://mediumlights.substack.com/p/lebron-reads-the-defense">LeBron also will watch the defense on a failed offensive possession, then start a future possession the same way to get the same behavior and exploit it</a>.

</li><li>Daniel Theis&apos;s &quot;<a href="https://mediumlights.substack.com/p/daniel-theis-paint-seal">paint seal</a>,&quot; the sort of thing I&apos;d never notice just by watching.

</li><li><a href="https://mediumlights.substack.com/p/nikola-jokics-water-polo-pump-fake">Nikola Jokic&apos;s water polo pump fake</a>.

</li><li><a href="https://mediumlights.substack.com/p/tyler-herros-high-arching-floater">A collection of absurdly high-arcing shots</a>.

</li><li>&quot;<a href="https://mediumlights.substack.com/p/dwyane-wades-sneaky-baseline-cuts">Dwyane Wade&apos;s sneaky baseline cuts: an appreciation for one of the best off-ball players of all time</a>&quot;

</li><li>A walkthrough of <a href="https://mediumlights.substack.com/p/draymond-guards-five-players-on-one">Draymond Green guarding all five opposing players on one possession</a>.

</li><li>This one is just silly (but top notch silly): <a href="https://mediumlights.substack.com/p/robin-lopez-makes-everyone-proud">Robin Lopez makes everyone proud</a>

</li><li><a href="https://mediumlights.substack.com/p/draymond-green-throws-steph-curry">Draymond &quot;throwing Steph Curry open.&quot;</a> &quot;honestly, my only reaction when i saw it the first time was &apos;what...?&apos; ...  [Curry&apos;s] whole body is turned towards his own basket when the ball is already halfway to him ... i don&apos;t really have any words for this. i guess this is what happens when two guys play together for a long time - the chemistry is there and cool shit just happens.&quot; (To understand &quot;throwing someone open&quot; you can see <a href="https://www.youtube.com/watch?v=UG4DKNasoPU&amp;t=36s">this Reddit post</a>, though I have to say ... watching <a href="https://www.youtube.com/watch?v=UG4DKNasoPU&amp;t=36s">this example</a>, I wondered if it is an &quot;amazing successful pass&quot; that just had a 50-50 chance of being an &quot;embarrassing intercepted pass.&quot;)
</li>
</ul>
<p>
It&apos;s great when someone is able to walk you through why they&apos;re this amazed and impressed with something you wouldn&apos;t have noticed.
</p><!--kg-card-end: html--><!--kg-card-begin: html-->

<p><div style="display:flex; justify-content:center; margin: 0 auto;">

        <span style="margin: 10px;"><a href="https://api.addthis.com/oexchange/0.8/forward/twitter/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fmedium-lights%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20Medium%20Lights&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-twitter-square.png" border="0" alt="Medium Lights"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/facebook/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fmedium-lights%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20Medium%20Lights&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-facebook-square.png" border="0" alt="Medium Lights"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/reddit/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fmedium-lights%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20Medium%20Lights&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-reddit-square.png" border="0" alt="Medium Lights"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/menu/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fmedium-lights%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20Medium%20Lights&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-addthis-square.png" border="0" alt="Medium Lights"></a></span>
        </div>
<center><p id="survey"><a href="#subscribe"><button class="button">Subscribe</button></a>&#xA0;<a href="https://www.guidedtrack.com/programs/4kal2ue/run?posttitle=Medium%20Lights" target="_blank"><button class="button" id="Survey">Feedback</button></a><!--No Forum post--></p></center>
<!--kg-card-end: html--></p>]]></content:encoded></item><item><title><![CDATA[AI Timelines: Where the Arguments, and the "Experts," Stand]]></title><description><![CDATA[What the best available forecasting methods say - and why there's no "expert field" for this topic.]]></description><link>https://www.cold-takes.com/where-ai-forecasting-stands-today/</link><guid isPermaLink="false">6125c5f3e680a1003ebdf596</guid><dc:creator><![CDATA[Holden Karnofsky]]></dc:creator><pubDate>Tue, 07 Sep 2021 17:48:42 GMT</pubDate><media:content url="https://www.cold-takes.com/content/images/2021/08/bio-anchors-probability-chart-7.png" medium="image"/><content:encoded><![CDATA[<!--kg-card-begin: html--><img src="https://www.cold-takes.com/content/images/2021/08/bio-anchors-probability-chart-7.png" alt="AI Timelines: Where the Arguments, and the &quot;Experts,&quot; Stand"><p><figure><iframe title="AI Timelines: Where the Arguments, and the &#x201D;Experts,&#x201D; Stand" allowtransparency="true" height="150" width="100%" style="border: none; min-width: min(100%, 430px);" scrolling="no" data-name="pb-iframe-player" src="https://www.podbean.com/player-v2/?i=fq5wv-10c9f9b-pb&amp;from=pb6admin&amp;share=1&amp;download=1&amp;rtl=0&amp;fonts=Arial&amp;skin=1&amp;font-color=auto&amp;btn-skin=7"></iframe><figcaption><em>Audio also available by searching Stitcher, Spotify, Google Podcasts, etc. for &quot;Cold Takes Audio&quot;</em></figcaption></figure></p>

<!--<p><a href="https://www.cold-takes.com/roadmap-for-the-most-important-century-series/"><img src="https://www.cold-takes.com/content/images/2021/08/mic-diagram---ai-timelines.png"></a></p>-->


<a href="https://www.cold-takes.com/roadmap-for-the-most-important-century-series/"><svg viewbox="0 0 750 178" fill="none" xmlns="http://www.w3.org/2000/svg" id="MasterDiagram">
<g class="Frame 1">
<g id="Today">
<text class="Today&#xE2;&#x20AC;&#x2122;s world" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-weight="bold" letter-spacing="0em"><tspan x="9" y="95.7402">Today&#x2019;s world</tspan></text>
<rect class="Rectangle 1" x="0.5" y="66.5" width="137" height="49" rx="7.5" stroke="var(--darker-gray-color)"/>
</g>
<g id="ArrowToTAI">
<path class="Arrow 7" d="M176.341 86.3655C176.543 86.1771 176.554 85.8607 176.366 85.6588L173.295 82.3691C173.107 82.1672 172.79 82.1563 172.588 82.3447C172.387 82.5331 172.376 82.8495 172.564 83.0514L175.293 85.9756L172.369 88.7049C172.167 88.8933 172.156 89.2097 172.345 89.4116C172.533 89.6135 172.85 89.6244 173.051 89.436L176.341 86.3655ZM146.983 85.4997L175.983 86.4997L176.017 85.5003L147.017 84.5003L146.983 85.4997Z" fill="var(--darker-gray-color)"/>
</g>
<g id="Transformative">
<text class="Transformative AI" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-weight="bold" letter-spacing="0em"><tspan x="192" y="92.6523">Transformative AI</tspan></text>
<rect class="Rectangle 2" x="182.5" y="65.5" width="165" height="47" rx="7.5" stroke="var(--darker-gray-color)"/>
</g>
<g id="ArrowFromTAI" opacity="0.3">
<path class="Arrow 1" d="M389.491 31.0944C389.543 30.8233 389.366 30.5611 389.094 30.509L384.675 29.6592C384.404 29.607 384.142 29.7846 384.09 30.0558C384.038 30.3269 384.215 30.589 384.487 30.6412L388.415 31.3966L387.659 35.3246C387.607 35.5958 387.785 35.8579 388.056 35.91C388.327 35.9622 388.589 35.7846 388.641 35.5135L389.491 31.0944ZM358.28 52.414L389.28 31.414L388.72 30.586L357.72 51.586L358.28 52.414Z" fill="var(--darker-gray-color)"/>
<path class="Arrow 2" d="M385.342 92.3648C385.543 92.1759 385.554 91.8595 385.365 91.658L382.287 88.3751C382.098 88.1737 381.782 88.1634 381.58 88.3523C381.379 88.5412 381.369 88.8576 381.557 89.0591L384.293 91.9772L381.375 94.713C381.174 94.9018 381.163 95.2182 381.352 95.4197C381.541 95.6212 381.858 95.6314 382.059 95.4425L385.342 92.3648ZM353.984 91.4997L384.984 92.4997L385.016 91.5003L354.016 90.5003L353.984 91.4997Z" fill="var(--darker-gray-color)"/>
<path class="Arrow 3" d="M384.955 148.498C385.23 148.523 385.473 148.32 385.498 148.045L385.905 143.564C385.93 143.289 385.728 143.046 385.453 143.021C385.178 142.996 384.934 143.198 384.909 143.473L384.547 147.457L380.564 147.095C380.289 147.07 380.046 147.272 380.021 147.547C379.996 147.822 380.198 148.066 380.473 148.091L384.955 148.498ZM354.616 112.32L384.616 148.32L385.384 147.68L355.384 111.68L354.616 112.32Z" fill="var(--darker-gray-color)"/>
</g>
<g id="DigitalPeople" opacity="0.3">
<text class="Digital people" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-weight="bold" letter-spacing="0em"><tspan x="401" y="36.6523">Digital people</tspan></text>
<text class="World of" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="14" letter-spacing="0em"><tspan x="435.216" y="16.7852">World of</tspan></text>
<rect class="Digital people bubble" x="394.5" y="0.5" width="133" height="45" rx="7.5" stroke="var(--darker-gray-color)"/>
</g>
<g id="MisalignedAndOther" opacity="0.3">
<text class="Misaligned AI" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-weight="bold" letter-spacing="0em"><tspan x="405.436" y="107.652">Misaligned AI</tspan></text>
<text class="World run by" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="14" letter-spacing="0em"><tspan x="422.104" y="87.7852">World run by</tspan></text>
<text class="Something else" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-weight="bold" letter-spacing="0em"><tspan x="398.053" y="164.652">Something else</tspan></text>
<rect class="Rectangle 3" x="395.5" y="68.5" width="132" height="52" rx="7.5" stroke="var(--darker-gray-color)"/>
<rect class="Rectangle 5" x="395.5" y="140.5" width="132" height="37" rx="7.5" stroke="var(--darker-gray-color)"/>
<text class="or" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-style="italic" letter-spacing="0em"><tspan x="455.476" y="61.6523">or</tspan></text>
<text class="or_2" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-style="italic" letter-spacing="0em"><tspan x="455.476" y="133.652">or</tspan></text>
    </g></g>
<g id="SGWC" opacity="0.3">
<text class="Stable, galaxy-wide civilization" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-weight="bold" letter-spacing="0em"><tspan x="583.109" y="89.6523">Stable, galaxy-wide
</tspan><tspan x="618.222" y="110.652">civilization</tspan></text>
<rect class="Rectangle 6" x="575.5" y="69.5" width="174" height="52" rx="7.5" stroke="var(--darker-gray-color)"/></g>
<g id="ArrowToSGWC">
<path opacity="0.3" class="Arrow 4" d="M562 56.5C562.276 56.5 562.5 56.2761 562.5 56L562.5 51.5C562.5 51.2239 562.276 51 562 51C561.724 51 561.5 51.2239 561.5 51.5L561.5 55.5L557.5 55.5C557.224 55.5 557 55.7239 557 56C557 56.2761 557.224 56.5 557.5 56.5L562 56.5ZM534.646 29.3536L561.646 56.3536L562.354 55.6464L535.354 28.6464L534.646 29.3536Z" fill="var(--darker-gray-color)"/>
<path opacity="0.3" class="Arrow 5" d="M561.354 95.3536C561.549 95.1583 561.549 94.8417 561.354 94.6464L558.172 91.4645C557.976 91.2692 557.66 91.2692 557.464 91.4645C557.269 91.6597 557.269 91.9763 557.464 92.1716L560.293 95L557.464 97.8284C557.269 98.0237 557.269 98.3403 557.464 98.5355C557.66 98.7308
                                       557.976 98.7308 558.172 98.5355L561.354 95.3536ZM535 95.5L561 95.5L561 94.5L535 94.5L535 95.5Z" fill="var(--darker-gray-color)"/>
<path opacity="0.3" class="Arrow 6" d="M561.497 117.053C561.526 116.778 561.327 116.532 561.053 116.503L556.578 116.027C556.304 115.998 556.057 116.196 556.028 116.471C555.999 116.746 556.198 116.992 556.472 117.021L560.45 117.444L560.027 121.422C559.998 121.696 560.196 121.943 560.471 121.972C560.746 122.001 560.992 121.802 561.021 121.528L561.497 117.053ZM535.314 138.389L561.314 117.389L560.686 116.611L534.686 137.611L535.314 138.389Z" fill="var(--darker-gray-color)"/>
</g>
    </svg></a>


<p>
<blockquote><p>This piece starts with a summary of when we should expect transformative AI to be developed, based on the multiple angles covered previously in the series. I think this is useful, even if you&apos;ve read all of the previous pieces, but if you&apos;d like to skip it, click <a href="#Part2">here</a>.</p>
</blockquote></p>
<p>
</p><p>I then address the question: &quot;Why isn&apos;t there a robust expert consensus on this topic, and what does that mean for us?&quot;</p>
<p></p>
<p>

</p>
<p>
I estimate that there is <strong>more than a 10% chance we&apos;ll see transformative AI within 15 years (by 2036); a ~50% chance we&apos;ll see it within 40 years (by 2060); and a ~2/3 chance we&apos;ll see it this century (by 2100). </strong>
</p>
<p>
(By &quot;transformative AI,&quot; I mean &quot;AI powerful enough to bring us into a new, qualitatively different future.&quot; I&apos;ve argued that advanced AI <a href="https://www.cold-takes.com/transformative-ai-timelines-part-1-of-4-what-kind-of-ai/#impacts-of-pasta">could</a> be sufficient to make this the <a href="https://www.cold-takes.com/roadmap-for-the-most-important-century-series/">most important century</a>.)
</p>
<p>
This is my overall conclusion based on a number of technical reports approaching AI forecasting from different angles - many of them produced by <a href="https://www.openphilanthropy.org">Open Philanthropy</a> over the past few years as we&apos;ve tried to develop a thorough picture of transformative AI forecasting to inform our longtermist grantmaking.
</p>
<p>
Here&apos;s a <strong>one-table summary </strong>of the different angles on forecasting transformative AI that I&apos;ve discussed, with links to more detailed discussion in <a href="https://www.cold-takes.com/roadmap-for-the-most-important-century-series/#forecasting-transformative-ai-this-century">previous posts</a> as well as to underlying technical reports:
</p>

<figure id="SummaryTable"><table style="border-collapse: collapse;">
      <tr>
   <td style="border: 1px solid; vertical-align: top;"><strong>Forecasting angle</strong>
   </td>
   <td style="border: 1px solid; vertical-align: top;"><strong>Key in-depth pieces (abbreviated titles)</strong>
   </td>
   <td style="border: 1px solid; vertical-align: top;"><strong>My takeaways</strong>
   </td>
  </tr>

    <tr><td style="border: 1px solid; vertical-align: top; text-align: center;" colspan="3"><em>Probability estimates for transformative AI</em></td></tr>
      <tr>
   <td style="border: 1px solid; vertical-align: top;"><strong><a href="https://www.cold-takes.com/are-we-trending-toward-transformative-ai-how-would-we-know/#surveying-experts">Expert survey.</a> </strong>What do AI researchers expect?
   </td>
   <td style="border: 1px solid; vertical-align: top;"><a href="https://arxiv.org/pdf/1705.08807.pdf">Evidence from AI Experts</a>
   </td>
   <td style="border: 1px solid; vertical-align: top;">Expert survey implies<sup id="fnref1"><a href="#fn1" rel="footnote">1</a></sup> a ~20% probability by 2036; ~50% probability by 2060; ~70% probability by 2100. Slightly differently phrased questions (posed to a minority of respondents) have much later estimates.
   </td>
  </tr>
  <tr>
   <td style="border: 1px solid; vertical-align: top;"><strong><a href="https://www.cold-takes.com/forecasting-transformative-ai-the-biological-anchors-method-in-a-nutshell/">Biological anchors framework</a>. </strong>Based on the usual patterns in how much &quot;AI training&quot; costs, how much would it cost to train an AI model as big as a human brain to perform the hardest tasks humans do? And when will this be cheap enough that we can expect someone to do it?
   </td>
   <td style="border: 1px solid; vertical-align: top;"><a href="https://drive.google.com/drive/u/1/folders/15ArhEPZSTYU8f012bs6ehPS6-xmhtBPP">Bio Anchors</a>, drawing on <a href="https://www.openphilanthropy.org/blog/new-report-brain-computation">Brain Computation</a>
   </td>
   <td style="border: 1px solid; vertical-align: top;">&gt;10% probability by 2036; ~50% chance by 2055; ~80% chance by 2100.
   </td>
  </tr>
    <tr><td style="border: 1px solid; vertical-align: top; text-align: center;" colspan="3"><em>Angles on the <a href="https://www.cold-takes.com/forecasting-transformative-ai-whats-the-burden-of-proof/">burden of proof</a></em></td></tr>
  <tr>
   <td style="border: 1px solid; vertical-align: top;">It&apos;s unlikely that any given century would be the &quot;most important&quot; one. (<a href="https://www.cold-takes.com/forecasting-transformative-ai-whats-the-burden-of-proof#most-important-century-skepticism">More</a>)
   </td>
   <td style="border: 1px solid; vertical-align: top;"><a href="https://static1.squarespace.com/static/5506078de4b02d88372eee4e/t/5f36b015d9a3691ba8e1096b/1597419543571/Are+we+living+at+the+hinge+of+history.pdf">Hinge</a>; <a href="https://forum.effectivealtruism.org/posts/j8afBEAa7Xb2R9AZN/thoughts-on-whether-we-re-living-at-the-most-influential">Response to Hinge</a>
   </td>
   <td style="border: 1px solid; vertical-align: top;">We have many reasons to think this century is a &quot;special&quot; one before looking at the details of AI. Many have been covered in previous pieces; another is covered in the next row. 
   </td>
  </tr>
  <tr>
   <td style="border: 1px solid; vertical-align: top;">What would you forecast about transformative AI timelines, based <strong>only</strong> on basic information about (a) how many years people have been trying to build transformative AI; (b) how much they&apos;ve &quot;invested&quot; in it (in terms of the number of AI researchers and the amount of computation used by them); (c) whether they&apos;ve done it yet (so far, they haven&apos;t)? (<a href="https://www.cold-takes.com/forecasting-transformative-ai-whats-the-burden-of-proof#semi-informative-priors">More</a>)
   </td>
   <td style="border: 1px solid; vertical-align: top;"><a href="https://www.openphilanthropy.org/blog/report-semi-informative-priors">Semi-informative Priors</a>
   </td>
   <td style="border: 1px solid; vertical-align: top;">Central estimates: 8% by 2036; 13% by 2060; 20% by 2100.<sup id="fnref2"><a href="#fn2" rel="footnote">2</a></sup> In my view, this report highlights that the history of AI is short, investment in AI is increasing rapidly, and so we shouldn&apos;t be too surprised if transformative AI is developed soon. 
   </td>
  </tr>
  <tr>
   <td style="border: 1px solid; vertical-align: top;">Based on analysis of economic models and economic history, how likely is &apos;explosive growth&apos; - defined as &gt;30% annual growth in the world economy - by 2100? Is this far enough outside of what&apos;s &quot;normal&quot; that we should doubt the conclusion? (<a href="https://www.cold-takes.com/forecasting-transformative-ai-whats-the-burden-of-proof#most-important-century-skepticism">More</a>)
   </td>
   <td style="border: 1px solid; vertical-align: top;"><a href="https://www.openphilanthropy.org/could-advanced-ai-drive-explosive-economic-growth">Explosive Growth</a>, <a href="https://www.openphilanthropy.org/blog/modeling-human-trajectory">Human Trajectory</a>
   </td>
   <td style="border: 1px solid; vertical-align: top;"><a href="https://www.openphilanthropy.org/blog/modeling-human-trajectory">Human Trajectory</a> projects the past forward, implying explosive growth by 2043-2065.
<p>
<a href="https://www.openphilanthropy.org/could-advanced-ai-drive-explosive-economic-growth">Explosive Growth</a> concludes: &quot;I find that economic considerations don&#x2019;t provide a good reason to dismiss the possibility of TAI being developed in this century. In fact, there is a plausible economic perspective from which sufficiently advanced AI systems are expected to cause explosive growth.&quot;
   </p></td>
  </tr>
  <tr>
   <td style="border: 1px solid; vertical-align: top;">&quot;How have people predicted AI ... in the past, and should we adjust our own views today to correct for patterns we can observe in earlier predictions? ... We&#x2019;ve encountered the view that AI has been prone to repeated over-hype in the past, and that we should therefore expect that today&#x2019;s projections are likely to be over-optimistic.&quot; (<a href="https://www.cold-takes.com/forecasting-transformative-ai-whats-the-burden-of-proof#history-of-">More</a>)
   </td>
   <td style="border: 1px solid; vertical-align: top;"><a href="https://www.openphilanthropy.org/focus/global-catastrophic-risks/potential-risks-advanced-artificial-intelligence/what-should-we-learn-past-ai-forecasts">Past AI Forecasts</a>
   </td>
   <td style="border: 1px solid; vertical-align: top;">&quot;The peak of AI hype seems to have been from 1956-1973. Still, the hype implied by some of the best-known AI predictions from this period is commonly exaggerated.&quot; 
   </td>
  </tr>
    </table><figcaption><em>For transparency, note that many of the technical reports are <a href="https://www.openphilanthropy.org/">Open Philanthropy</a> analyses, and I am co-CEO of Open Philanthropy.</em></figcaption></figure>


<p id="Part2">
Having considered the above, I expect some readers to still feel a sense of unease. Even if they think my arguments make sense, they may be wondering: <strong>if this is true, why isn&apos;t it more widely discussed and accepted? What&apos;s the state of expert opinion?</strong>
</p>
<p>
My summary of the state of expert opinion at this time is:
</p>
<p>
 
</p>
<ul>

<li>The claims I&apos;m making do not <em>contradict</em> any particular expert consensus. (In fact, the probabilities I&apos;ve given aren&apos;t too far off from what AI researchers seem to predict, as shown in the first row.) But there are some <a href="https://www.cold-takes.com/are-we-trending-toward-transformative-ai-how-would-we-know/#surveying-experts">signs they aren&apos;t thinking too hard about the matter</a>. 

</li><li>The Open Philanthropy technical reports I&apos;ve relied on have had significant external expert review. Machine learning researchers reviewed <a href="https://drive.google.com/drive/u/1/folders/15ArhEPZSTYU8f012bs6ehPS6-xmhtBPP">Bio Anchors</a>; neuroscientists reviewed <a href="https://www.openphilanthropy.org/blog/new-report-brain-computation">Brain Computation</a>; economists reviewed <a href="https://www.openphilanthropy.org/could-advanced-ai-drive-explosive-economic-growth">Explosive Growth</a>; academics focused on relevant topics in uncertainty and/or probability reviewed <a href="https://www.openphilanthropy.org/blog/report-semi-informative-priors">Semi-informative Priors</a>.<sup id="fnref2"><a href="#fn2" rel="footnote">2</a></sup> (Some of these reviews had significant points of disagreement, but none of these points seemed to be cases where the reports contradicted a clear consensus of experts or literature.)

</li><li>But there is also no active, robust expert consensus supporting claims like <em>&quot;There&apos;s at least a 10% chance of transformative AI by 2036&quot;</em> or <em>&quot;There&apos;s a good chance we&apos;re in the most important century for humanity,&quot;</em> the way that there is supporting e.g. the need to take action against climate change.</li></ul>
<p>
Ultimately, my claims are about <strong>topics that simply have no &quot;field&quot; of experts devoted to studying them. That, in and of itself, is a scary fact, </strong>and something that I hope will eventually change.
</p>
<p>
But should we be willing to act on the &quot;most important century&quot; hypothesis in the meantime?
</p>
<p>
Below, I&apos;ll discuss:
</p>
<ul>

<li>What an &quot;AI forecasting field&quot; might look like.

</li><li>A &quot;skeptical view&quot; that says today&apos;s discussions around these topics are too small, homogeneous and insular (which I agree with) - and that we therefore shouldn&apos;t act on the <a href="https://www.cold-takes.com/roadmap-for-the-most-important-century-series/">&quot;most important century&quot; hypothesis</a> until there is a mature, robust field (which I don&apos;t).

</li><li>Why I think we should take the hypothesis seriously in the meantime, until and unless such a field develops: 
<ul>
 
<li>We don&apos;t have time to wait for a robust expert consensus.
 
</li><li>If there are good rebuttals out there - or potential future experts who could develop such rebuttals - we haven&apos;t found them yet. The more seriously the hypothesis gets taken, the more likely such rebuttals are to appear. (Aka the <a href="https://bigthink.com/david-ryan-polgar/want-the-right-answer-online-dont-ask-questions-just-post-it-wrong">Cunningham&apos;s Law</a> theory: &quot;the best way to get a right answer is to post a wrong answer.&quot;)
 
</li><li>I think that consistently insisting on a robust expert consensus is a dangerous reasoning pattern. In my view, it&apos;s OK to be at some risk of self-delusion and insularity, in exchange for doing the right thing when it counts most.
</li> 
</ul>
</li> 
</ul>
<h2 id="what-kind-of-expertise-is-ai-forecasting-expertise">What kind of expertise is AI forecasting expertise?</h2>


<p>
Questions analyzed in the technical reports listed <a href="#SummaryTable">above</a> include:
</p>
<ul>

<li>Are AI capabilities getting more impressive over time? (AI, history of AI)

</li><li>How can we compare AI models to animal/human brains? (AI, neuroscience)

</li><li>How can we compare AI capabilities to animals&apos; capabilities? (AI, ethology)

</li><li>How can we estimate the expense of training a large AI system for a difficult task, based on information we have about training past AI systems? (AI, curve-fitting)

</li><li>How can we make a minimal-information estimate about transformative AI, based only on how many years/researchers/dollars have gone into the field so far? (Philosophy, probability)

</li><li>How likely is explosive economic growth this century, based on theory and historical trends? (Growth economics, economic history)

</li><li>What has &quot;AI hype&quot; been like in the past? (History)
</li>
</ul>
<p>
When talking about wider implications of transformative AI for the &quot;most important century,&quot; I&apos;ve also discussed things like &quot;How feasible are <a href="https://www.cold-takes.com/digital-people-faq/#feasibility">digital people</a> and <a href="https://www.cold-takes.com/how-digital-people-could-change-the-world/#space-expansion">establishing space settlements throughout the galaxy</a>?&quot; These topics touch physics, neuroscience, engineering, philosophy of mind, and more.
</p>
<p>
<strong>There&apos;s no obvious job or credential that makes someone an expert on the question of when we can expect transformative AI, or the question of whether we&apos;re in the most important century. </strong>
</p>
<p>
(I particularly would disagree with any claim that we should be relying exclusively on AI researchers for these forecasts. In addition to the fact that <a href="https://www.cold-takes.com/are-we-trending-toward-transformative-ai-how-would-we-know/#surveying-experts">they don&apos;t seem to be thinking very hard about the topic</a>, I think that relying on people who specialize in building ever-more powerful AI models to tell us when transformative AI might come is like relying on solar energy R&amp;D companies - or oil extraction companies, depending on how you look at it - to forecast carbon emissions and climate change. They certainly have part of the picture. But forecasting is a distinct activity from innovating or building state-of-the-art systems.)
</p>
<p>
And I&apos;m not even sure these questions have the right shape for an academic field. Trying to forecast transformative AI, or determine the odds that we&apos;re in the most important century, seems:
</p>
<ul>

<li>More similar to the <a href="https://projects.fivethirtyeight.com/2020-election-forecast/">FiveThirtyEight election model</a> (&quot;Who&apos;s going to win the election?&quot;) than to academic political science (&quot;How do governments and constituents interact?&quot;); 

</li><li>More similar to trading financial markets (&quot;Is this price going up or down in the future?&quot;) than to academic economics (&quot;Why do recessions exist?&quot;);<sup id="fnref3"><a href="#fn3" rel="footnote">3</a></sup> 

</li><li>More similar to <a href="https://www.givewell.org/">GiveWell&apos;s</a> research (&quot;Which charity will help people the most, per dollar?&quot;) than to academic development economics (&quot;What causes poverty and what can reduce it?&quot;)<sup id="fnref4"><a href="#fn4" rel="footnote">4</a></sup></li></ul>
<p>
That is, it&apos;s not clear to me what a natural &quot;institutional home&quot; for expertise on transformative AI forecasting, and the &quot;most important century,&quot; would look like. But it seems fair to say there aren&apos;t large, robust institutions dedicated to this sort of question today.
</p>
<h2 id="how-should-we-act-in-the-absence-of-a-robust-expert-consensus">How should we act in the absence of a robust expert consensus?</h2>


<h3 id="the-skeptical-view">The skeptical view</h3>


<p>
Lacking a robust expert consensus, I expect some (really, most) people will be skeptical no matter what arguments are presented.
</p>
<p>
Here&apos;s a version of a very general skeptical reaction I have a fair amount of empathy for:
</p>
<ol>

<li><em>This is all just too <a href="https://www.cold-takes.com/forecasting-transformative-ai-whats-the-burden-of-proof/#formalizing-the-">wild</a>.</em>

</li><li><em>You&apos;re making an over-the-top claim about living in the most important century. This <strong>pattern-matches to self-delusion.</strong></em>

</li><li><em>You&apos;ve argued that the <a href="https://www.cold-takes.com/forecasting-transformative-ai-whats-the-burden-of-proof/">burden of proof </a>shouldn&apos;t be so high, because there are lots of ways in which we live in a <a href="https://www.cold-takes.com/all-possible-views-about-humanitys-future-are-wild/">remarkable</a> and <a href="https://www.cold-takes.com/this-cant-go-on/">unstable </a>time. But ... I don&apos;t trust myself to assess those claims, or your claims about AI, or really anything on these wild topics.</em>

</li><li><em>I&apos;m worried by how few people seem to be engaging these arguments. About how <strong>small, homogeneous and insular</strong> the discussion seems to be. Overall, this feels more like a story smart people are telling themselves - with lots of charts and numbers to rationalize it - about their place in history. It doesn&apos;t feel &quot;real.&quot;</em>

</li><li><em>So call me back when there&apos;s a mature field of perhaps hundreds or thousands of experts, critiquing and assessing each other, and they&apos;ve reached the same sort of consensus that we see for climate change.</em>
</li>
</ol>
<p>
I see how you could feel this way, and I&apos;ve felt this way myself at times - especially on points #1-#4. But I&apos;ll give <strong>three reasons that point #5 doesn&apos;t seem right.</strong>
</p>
<h3 id="reason-1-we-dont-have-time-to-wait-for-a-robust-expert-consensus">Reason 1: we don&apos;t have time to wait for a robust expert consensus</h3>


<p>
I worry that the arrival of transformative AI could play out as a kind of slow-motion, higher-stakes version of the COVID-19 pandemic. The case for expecting something big to happen is there, if you look at the best information and analyses available today. But the situation is broadly unfamiliar; it doesn&apos;t fit into patterns that our institutions regularly handle.  And every extra year of action is valuable.
</p>
<p>
You could also think of it as a sped-up version of the dynamic with climate change. Imagine if greenhouse gas emissions had only started to rise recently<sup id="fnref5"><a href="#fn5" rel="footnote">5</a></sup> (instead of in the <a href="https://ourworldindata.org/co2-and-other-greenhouse-gas-emissions">mid-1800s</a>), and if there were no established field of climate science. It would be a really bad idea to wait decades for a field to emerge, before seeking to reduce emissions.
</p>
<h3 id="reason-2-cunninghams-law">Reason 2: <a href="https://bigthink.com/david-ryan-polgar/want-the-right-answer-online-dont-ask-questions-just-post-it-wrong">Cunningham&apos;s Law</a> (&quot;the best way to get a right answer is to post a wrong answer&quot;) may be our best hope for finding the flaw in these arguments</h3>


<p>
I&apos;m serious, though.
</p>
<p>
Several years ago, some <a href="https://www.cold-takes.com/roadmap-for-the-most-important-century-series/#acknowledgements">colleagues </a>and I suspected that the &quot;most important century&quot; hypothesis could be true. But before acting on it too much, we wanted to see whether we could find fatal flaws in it.
</p>
<p>
One way of interpreting our actions over the last few years is <strong>as if we were doing everything we could to learn that the hypothesis is wrong.</strong>
</p>
<p>
First, we tried talking to people about the key arguments - AI researchers, economists, etc. But:
</p>
<ul>

<li>We had vague ideas of the arguments in this series (mostly or perhaps entirely picked up <a href="https://www.cold-takes.com/roadmap-for-the-most-important-century-series/#acknowledgements">from other people</a>). We weren&apos;t able to state them with good crispness and specificity.

</li><li>There were a lot of key factual points that we thought would probably check out,<sup id="fnref6"><a href="#fn6" rel="footnote">6</a></sup> but hadn&apos;t nailed down and couldn&apos;t present for critique.

</li><li>Overall, we couldn&apos;t even really articulate enough of a concrete case to give the others a fair chance to shoot it down.</li></ul>
<p>
So we put a lot of work into creating technical reports on many of the key arguments. (These are now public, and included in the table at the top of this piece.) This put us in position to publish the arguments, and potentially encounter fatal counterarguments.
</p>
<p>
Then, we commissioned external expert reviews.<sup id="fnref7"><a href="#fn7" rel="footnote">7</a></sup>
</p>
<p>
Speaking only for my own views, the &quot;most important century&quot; hypothesis seems to have survived all of this. Indeed, having examined the many angles and gotten more into the details, I believe it more strongly than before.
</p>
<p>
But let&apos;s say that this is just because the <em>real</em> experts - people we haven&apos;t found yet, with devastating counterarguments - find the whole thing so silly that they&apos;re <a href="https://philiptrammell.com/blog/46/">not bothering to engage</a>. Or, let&apos;s say that there are people out there today who could <em>someday</em> become experts on these topics, and knock these arguments down. What could we do to bring this about?
</p>
<p>
The best answer I&apos;ve come up with is: &quot;If this hypothesis became better-known, more widely accepted, and more influential, it would get more critical scrutiny.&quot; 
</p>
<p>
This series is an attempted step in that direction - to move toward broader credibility for the &quot;most important century&quot; hypothesis. This would be a good thing if the hypothesis were true; it also seems like the best next step if my only goal were to challenge my beliefs and learn that it is false.
</p>
<p>
Of course, I&apos;m not saying to accept or promote the &quot;most important century&quot; hypothesis if it doesn&apos;t seem correct to you. But I think that if your <em>only</em> reservation is about the lack of robust consensus, continuing to ignore the situation seems odd. If people behaved this way generally (ignoring any hypothesis not backed by a robust consensus), I&apos;m not sure I see how any hypothesis - including true ones - would go from fringe to accepted.
</p>
<h3 id="reason-3-skepticism-this-general-seems-like-a-bad-idea">Reason 3: skepticism this general seems like a bad idea</h3>


<p>
Back when I was focused on <a href="http://www.givewell.org">GiveWell</a>, people would occasionally say something along the lines of: &quot;You know, you can&apos;t hold every argument to the standard that GiveWell holds its top charities to - seeking randomized controlled trials, robust empirical data, etc. Some of the best opportunities to do good will be the ones that are less obvious - so this standard risks <a href="https://www.openphilanthropy.org/blog/hits-based-giving#Anti-principles_for_hits-based_giving">ruling out some of your biggest potential opportunities to have impact</a>.&quot; 
</p>
<p>
I think this is right. I think it&apos;s important to check one&apos;s general approach to reasoning and evidentiary standards and ask: &quot;What are some scenarios in which my approach fails, and in which I&apos;d really prefer that it succeed?&quot; In my view, <strong>it&apos;s OK to be at some risk of self-delusion and insularity, in exchange for doing the right thing when it counts most.</strong>
</p>
<p>
I think the lack of a robust expert consensus - and concerns about self-delusion and insularity - provide good reason to <em>dig hard </em>on the &quot;most important century&quot; hypothesis, rather than accepting it immediately.  To ask where there might be an undiscovered flaw, to look for some bias toward inflating our own importance, to research the most questionable-seeming parts of the argument, etc.
</p>
<p>
But if you&apos;ve investigated the matter as much as is reasonable/practical for you - and haven&apos;t found a flaw <em>other</em> than considerations like &quot;There&apos;s no robust expert consensus&quot; and &quot;I&apos;m worried about self-delusion and insularity&quot; - then I think writing off the hypothesis is the sort of thing that essentially <strong>guarantees you won&apos;t be among the earlier people to notice and act on a tremendously important issue, if the opportunity arises.</strong> I think that&apos;s too much of a sacrifice, in terms of giving up potential opportunities to do a lot of good.
</p>
<p><strong>Next in series:</strong> <a href="https://www.cold-takes.com/making-the-best-of-the-most-important-century/">How to make the best of the most important century?</a></p><!--kg-card-end: html--><!--kg-card-begin: html-->

<p><div style="display:flex; justify-content:center; margin: 0 auto;">

        <span style="margin: 10px;"><a href="https://api.addthis.com/oexchange/0.8/forward/twitter/offer?url=https%3A%2F%2Fwhere-ai-forecasting-stands-today&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20AI%20Timelines%3A%20Where%20the%20Arguments%2C%20and%20the%20%22Experts%2C%22%20Stand&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-twitter-square.png" border="0" alt="AI Timelines: Where the Arguments, and the &quot;Experts,&quot; Stand"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/facebook/offer?url=https%3A%2F%2Fwhere-ai-forecasting-stands-today&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20AI%20Timelines%3A%20Where%20the%20Arguments%2C%20and%20the%20%22Experts%2C%22%20Stand&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-facebook-square.png" border="0" alt="AI Timelines: Where the Arguments, and the &quot;Experts,&quot; Stand"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/reddit/offer?url=https%3A%2F%2Fwhere-ai-forecasting-stands-today&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20AI%20Timelines%3A%20Where%20the%20Arguments%2C%20and%20the%20%22Experts%2C%22%20Stand&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-reddit-square.png" border="0" alt="AI Timelines: Where the Arguments, and the &quot;Experts,&quot; Stand"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/menu/offer?url=https%3A%2F%2Fwhere-ai-forecasting-stands-today&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20AI%20Timelines%3A%20Where%20the%20Arguments%2C%20and%20the%20%22Experts%2C%22%20Stand&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-addthis-square.png" border="0" alt="AI Timelines: Where the Arguments, and the &quot;Experts,&quot; Stand"></a></span>
        </div>
<center><p id="survey"><a href="https://where-ai-forecasting-stands-today#/portal/signup/free" target="_blank"><button class="button">Subscribe</button></a>&#xA0;<a href="https://www.guidedtrack.com/programs/4kal2ue/run?posttitle=AI%20Timelines%3A%20Where%20the%20Arguments%2C%20and%20the%20%22Experts%2C%22%20Stand" target="_blank"><button class="button" id="Survey">Feedback</button></a>&#xA0;<a href="https://forum.effectivealtruism.org/posts/7JxsXYDuqnKMqa6Eq/ai-timelines-where-the-arguments-and-the-experts-stand#comments" target="_blank"><button class="button">Forum</button></a></p><p><em>
Use &quot;Feedback&quot; if you have comments/suggestions you want me to see, or if you&apos;re up for giving some quick feedback about this post (which I greatly appreciate!) Use &quot;Forum&quot; if you want to discuss this post publicly on the Effective Altruism Forum.
</em></p></center>
<!--kg-card-end: html--><!--kg-card-begin: html-->
<hr>
</p><h2 id="footnotes">Footnotes</h2>
<div class="footnotes">
<ol><li id="fn1">
<p>
     Technically, these probabilities are for &#x201C;human-level machine intelligence.&#x201D; In general, this chart simplifies matters by presenting one unified set of probabilities. In general, all of these probabilities refer to something at <em>least</em> as capable as <a href="https://www.cold-takes.com/transformative-ai-timelines-part-1-of-4-what-kind-of-ai/">PASTA</a>, so they directionally should be underestimates of the probability of PASTA (though I don&apos;t think this is a major issue).&#xA0;<a href="#fnref1" rev="footnote">&#x21A9;</a><li id="fn2">

<p>
     Reviews of Bio Anchors are <a href="https://drive.google.com/drive/u/1/folders/1XkTYFiZQUT6UAUL2Wyg9wD0qG57KYfjq">here</a>; reviews of Explosive Growth are <a href="https://www.openphilanthropy.org/could-advanced-ai-drive-explosive-economic-growth#AppendixH">here</a>; reviews of Semi-informative Priors are <a href="https://www.openphilanthropy.org/blog/report-semi-informative-priors#LinksToReviewer">here</a>. Brain Computation was reviewed at an earlier time when we hadn&apos;t designed the process to result in publishing reviews, but over 20 conversations with experts that informed the report are available <a href="https://www.openphilanthropy.org/research/conversations">here</a>. Human Trajectory hasn&apos;t been reviewed, although a lot of its analysis and conclusions feature in Explosive Growth, which has been.&#xA0;<a href="#fnref2" rev="footnote">&#x21A9;</a><li id="fn3">

<p>
     The academic fields are quite broad, and I&apos;m just giving example questions that they tackle.&#xA0;<a href="#fnref3" rev="footnote">&#x21A9;</a><li id="fn4">
<p>
     Though climate science is an example of an academic field that invests a lot in forecasting the future.&#xA0;<a href="#fnref4" rev="footnote">&#x21A9;</a><li id="fn5">
<p>
     The field of AI has existed since <a href="https://en.wikipedia.org/wiki/Dartmouth_workshop">1956</a>, but it&apos;s only in the last decade or so that machine learning models have started to get within range of <a href="https://www.cold-takes.com/forecasting-transformative-ai-the-biological-anchors-method-in-a-nutshell/#conclusions-of-bio-anchors">the size of insect brains</a> and perform well on relatively difficult tasks.&#xA0;<a href="#fnref5" rev="footnote">&#x21A9;</a><li id="fn6">

<p>
     Often, we were simply going off of our impressions of what others who had thought about the topic a lot thought.&#xA0;<a href="#fnref6" rev="footnote">&#x21A9;</a><li id="fn7">
<p>
     Reviews of Bio Anchors are <a href="https://drive.google.com/drive/u/1/folders/1XkTYFiZQUT6UAUL2Wyg9wD0qG57KYfjq">here</a>; reviews of Explosive Growth are <a href="https://www.openphilanthropy.org/could-advanced-ai-drive-explosive-economic-growth#AppendixH">here</a>; reviews of Semi-informative Priors are <a href="https://www.openphilanthropy.org/blog/report-semi-informative-priors#LinksToReviewer">here</a>. Brain Computation was reviewed at an earlier time when we hadn&apos;t designed the process to result in publishing reviews, but over 20 conversations with experts that informed the report are available <a href="https://www.openphilanthropy.org/research/conversations">here</a>. Human Trajectory hasn&apos;t been reviewed, although a lot of its analysis and conclusions feature in Explosive Growth, which has been.&#xA0;<a href="#fnref7" rev="footnote">&#x21A9;</a>

</p></li></p></li></p></li></p></li></p></li></p></li></p></li></ol></div>
<!--kg-card-end: html-->]]></content:encoded></item><item><title><![CDATA[Cold Links: Useful]]></title><description><![CDATA[<!--kg-card-begin: html-->
<p>
These are assorted links that I expect readers to maybe find useful on a personal basis. Don&apos;t worry, this won&apos;t be too frequent an occurrence - I aim for 99% of this blog to be about the last 100-10,000 years and the next 100-10 zillion</p>]]></description><link>https://www.cold-takes.com/cold-links-useful/</link><guid isPermaLink="false">6130951aa7ed4b003bf812ee</guid><dc:creator><![CDATA[Holden Karnofsky]]></dc:creator><pubDate>Thu, 02 Sep 2021 19:53:32 GMT</pubDate><content:encoded><![CDATA[<!--kg-card-begin: html-->
<p>
These are assorted links that I expect readers to maybe find useful on a personal basis. Don&apos;t worry, this won&apos;t be too frequent an occurrence - I aim for 99% of this blog to be about the last 100-10,000 years and the next 100-10 zillion years.
</p>

<p>
Ben Kuhn&apos;s <a href="https://www.benkuhn.net/vc/">tips for the most immersive video calls</a> - the most comprehensive guide I&apos;ve seen to good Zooming. 
</p>
<p>
Some<a href="https://thepointsguy.com/2018/03/how-to-find-lodging-with-good-wi-fi/"> tips on how to make sure you have good wifi at your hotel/AirBNB</a>. It shocks me that AirBNB doesn&apos;t do more to help (e.g., nudging hosts to verify and list their speeds and networking equipment, allowing guests to filter by this). Some of the tips here are obvious, some I didn&apos;t know, but it seems like a good checklist.
</p>
<p>
I strongly endorse this<a href="http://sethgodin.typepad.com/seths_blog/2018/04/words-on-slides.html"> advice on PowerPoint presentations</a>.
</p>
<p>
Interesting <a href="https://www.howiechong.com/journal/2014/2/bike-helmets?fbclid=IwAR3KRBESvDwgxEItuRK-0uwo3dEqtjSQXFTpy7Adu_Iu9u41jQdl341eKDY">argument against wearing a bike helmet</a>. I wear one because I don&apos;t want to look reckless and actively enjoy looking dorky, but &#xAF;\_(&#x30C4;)_/&#xAF;
</p>
<p>
I recommend traveling with an <a href="https://smile.amazon.com/Leviton-692-W-Triple-Grounding-Adapter/dp/B000P9SXTG/">octopus plug</a> so you never have to fight over an airport charger (instead, you&apos;re a hero!)
</p>
<p>
<a href="http://fivethirtyeight.com/features/which-diet-will-help-you-lose-the-most-weight/">FiveThirtyEight</a> looks at a recent review of randomized studies on which diets work best. The effect sizes seem pretty good - ~15lbs avg weight loss after a year - compared to what I expected based on previous coverage of this topic. Atkins doesn&apos;t win, Ornish (very traditional approach to dieting) does, though they&apos;re all extremely close, consistent with my general view that &quot;any diet you can stick to will probably help because it will be different from eating whatever you feel like.&quot; Paleo isn&apos;t included in the analysis.
</p>
<p>
Basic take, but important. WSJ:<a href="https://www.wsj.com/articles/the-man-who-wrote-those-password-rules-has-a-new-tip-n3v-r-m1-d-1502124118"> The Man Who Wrote Those Password Rules Has a New Tip: N3v$r M1^d!</a> &quot;Bill Burr&#x2019;s 2003 report recommended using numbers, obscure characters and capital letters and updating regularly&#x2014;he regrets the error.&quot; Not surprising at all - it&apos;s bad advice that websites still enforce today, and<a href="https://xkcd.com/936/"> XKCD&apos;s take is spot on</a>. I recommend LastPass and maximal use of two-factor authentication. 
</p>
<p>
Here&apos;s an<a href="http://www.kalzumeus.com/2017/09/09/identity-theft-credit-reports/"> excruciatingly detailed guide</a> to filing a complaint with a credit reporting agency that actually gets acted on. Key advice is to stay away from online and phone communications and use certified mail for everything, which shows them you&apos;re collecting a paper trail and scares them; key incredible fact is that you&apos;re not allowed to use form letters because ... well, because credit reporting agencies don&apos;t want you to and they&apos;ve lobbied to prohibit it. (Even though they use form letters.) I found this whole long thing weirdly fun to read. Just thinking about confronting one of these awful bureaucracies and knowing how to get good results made me smile. Advice may also have applications for dealing with bureaucracies more generally.
</p>
<p>
<a href="https://slate.com/technology/2013/06/rescuing-drowning-children-how-to-know-when-someone-is-in-trouble-in-the-water.html">How to recognize when a child is drowning</a>. If you donate to effective global health charities to save children&apos;s lives, then you need to know how to do this too in order to be consistent.<sup id="fnref1"><a href="#fn1" rel="footnote">1</a></sup>
</p><!--kg-card-end: html--><!--kg-card-begin: html-->

<p><div style="display:flex; justify-content:center; margin: 0 auto;">

        <span style="margin: 10px;"><a href="https://api.addthis.com/oexchange/0.8/forward/twitter/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fcold-links-useful%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20Cold%20Links%3A%20Useful&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-twitter-square.png" border="0" alt="Twitter"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/facebook/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fcold-links-useful%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20Cold%20Links%3A%20Useful&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-facebook-square.png" border="0" alt="Facebook"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/reddit/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fcold-links-useful%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20Cold%20Links%3A%20Useful&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-reddit-square.png" border="0" alt="Reddit"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/menu/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fcold-links-useful%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20Cold%20Links%3A%20Useful&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-addthis-square.png" border="0" alt="More"></a></span>
        </div>
<center><p id="survey"><a href="https://www.cold-takes.com/cold-links-useful/#/portal/signup/free" target="_blank"><button class="button">Subscribe</button></a>&#xA0;<a href="https://www.guidedtrack.com/programs/4kal2ue/run?posttitle=Cold%20Links%3A%20Useful" target="_blank"><button class="button" id="Survey">Feedback</button></a><!--No Forum post--></p></center>
<!--kg-card-end: html--><!--kg-card-begin: html--><hr>
</p><h2>Footnotes</h2>
<div class="footnotes">
<ol><li id="fn1">
<p>
     This is a joke. If you don&apos;t get it, don&apos;t worry about it.&#xA0;<a href="#fnref1" rev="footnote">&#x21A9;</a>

</p></li></ol></div><!--kg-card-end: html-->]]></content:encoded></item><item><title><![CDATA[Forecasting transformative AI: the "biological anchors" method in a nutshell]]></title><description><![CDATA[My preferred method of forecasting transformative AI, with pros and cons.]]></description><link>https://www.cold-takes.com/forecasting-transformative-ai-the-biological-anchors-method-in-a-nutshell/</link><guid isPermaLink="false">6117514881b1c6003ee57257</guid><dc:creator><![CDATA[Holden Karnofsky]]></dc:creator><pubDate>Tue, 31 Aug 2021 18:18:03 GMT</pubDate><media:content url="https://www.cold-takes.com/content/images/2021/08/bio-anchors-timeline-twitter-4.png" medium="image"/><content:encoded><![CDATA[<!--kg-card-begin: html--><img src="https://www.cold-takes.com/content/images/2021/08/bio-anchors-timeline-twitter-4.png" alt="Forecasting transformative AI: the &quot;biological anchors&quot; method in a nutshell"><p><figure><iframe title="Forecasting transformative AI: the &#x201D;biological anchors&#x201D; method in a nutshell" allowtransparency="true" height="150" width="100%" style="border: none; min-width: min(100%, 430px);" scrolling="no" data-name="pb-iframe-player" src="https://www.podbean.com/player-v2/?i=3mqbg-10c9f95-pb&amp;from=pb6admin&amp;share=1&amp;download=1&amp;rtl=0&amp;fonts=Arial&amp;skin=1&amp;font-color=auto&amp;btn-skin=7"></iframe><figcaption><em>Audio also available by searching Stitcher, Spotify, Google Podcasts, etc. for &quot;Cold Takes Audio&quot;</em></figcaption></figure></p>

<!--<p><a href="https://www.cold-takes.com/roadmap-for-the-most-important-century-series/"><img src="https://www.cold-takes.com/content/images/2021/08/mic-diagram---ai-timelines.png"></a></p>-->


<a href="https://www.cold-takes.com/roadmap-for-the-most-important-century-series/"><svg viewbox="0 0 750 178" fill="none" xmlns="http://www.w3.org/2000/svg" id="MasterDiagram">
<g class="Frame 1">
<g id="Today">
<text class="Today&#xE2;&#x20AC;&#x2122;s world" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-weight="bold" letter-spacing="0em"><tspan x="9" y="95.7402">Today&#x2019;s world</tspan></text>
<rect class="Rectangle 1" x="0.5" y="66.5" width="137" height="49" rx="7.5" stroke="var(--darker-gray-color)"/>
</g>
<g id="ArrowToTAI">
<path class="Arrow 7" d="M176.341 86.3655C176.543 86.1771 176.554 85.8607 176.366 85.6588L173.295 82.3691C173.107 82.1672 172.79 82.1563 172.588 82.3447C172.387 82.5331 172.376 82.8495 172.564 83.0514L175.293 85.9756L172.369 88.7049C172.167 88.8933 172.156 89.2097 172.345 89.4116C172.533 89.6135 172.85 89.6244 173.051 89.436L176.341 86.3655ZM146.983 85.4997L175.983 86.4997L176.017 85.5003L147.017 84.5003L146.983 85.4997Z" fill="var(--darker-gray-color)"/>
</g>
<g id="Transformative">
<text class="Transformative AI" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-weight="bold" letter-spacing="0em"><tspan x="192" y="92.6523">Transformative AI</tspan></text>
<rect class="Rectangle 2" x="182.5" y="65.5" width="165" height="47" rx="7.5" stroke="var(--darker-gray-color)"/>
</g>
<g id="ArrowFromTAI" opacity="0.3">
<path class="Arrow 1" d="M389.491 31.0944C389.543 30.8233 389.366 30.5611 389.094 30.509L384.675 29.6592C384.404 29.607 384.142 29.7846 384.09 30.0558C384.038 30.3269 384.215 30.589 384.487 30.6412L388.415 31.3966L387.659 35.3246C387.607 35.5958 387.785 35.8579 388.056 35.91C388.327 35.9622 388.589 35.7846 388.641 35.5135L389.491 31.0944ZM358.28 52.414L389.28 31.414L388.72 30.586L357.72 51.586L358.28 52.414Z" fill="var(--darker-gray-color)"/>
<path class="Arrow 2" d="M385.342 92.3648C385.543 92.1759 385.554 91.8595 385.365 91.658L382.287 88.3751C382.098 88.1737 381.782 88.1634 381.58 88.3523C381.379 88.5412 381.369 88.8576 381.557 89.0591L384.293 91.9772L381.375 94.713C381.174 94.9018 381.163 95.2182 381.352 95.4197C381.541 95.6212 381.858 95.6314 382.059 95.4425L385.342 92.3648ZM353.984 91.4997L384.984 92.4997L385.016 91.5003L354.016 90.5003L353.984 91.4997Z" fill="var(--darker-gray-color)"/>
<path class="Arrow 3" d="M384.955 148.498C385.23 148.523 385.473 148.32 385.498 148.045L385.905 143.564C385.93 143.289 385.728 143.046 385.453 143.021C385.178 142.996 384.934 143.198 384.909 143.473L384.547 147.457L380.564 147.095C380.289 147.07 380.046 147.272 380.021 147.547C379.996 147.822 380.198 148.066 380.473 148.091L384.955 148.498ZM354.616 112.32L384.616 148.32L385.384 147.68L355.384 111.68L354.616 112.32Z" fill="var(--darker-gray-color)"/>
</g>
<g id="DigitalPeople" opacity="0.3">
<text class="Digital people" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-weight="bold" letter-spacing="0em"><tspan x="401" y="36.6523">Digital people</tspan></text>
<text class="World of" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="14" letter-spacing="0em"><tspan x="435.216" y="16.7852">World of</tspan></text>
<rect class="Digital people bubble" x="394.5" y="0.5" width="133" height="45" rx="7.5" stroke="var(--darker-gray-color)"/>
</g>
<g id="MisalignedAndOther" opacity="0.3">
<text class="Misaligned AI" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-weight="bold" letter-spacing="0em"><tspan x="405.436" y="107.652">Misaligned AI</tspan></text>
<text class="World run by" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="14" letter-spacing="0em"><tspan x="422.104" y="87.7852">World run by</tspan></text>
<text class="Something else" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-weight="bold" letter-spacing="0em"><tspan x="398.053" y="164.652">Something else</tspan></text>
<rect class="Rectangle 3" x="395.5" y="68.5" width="132" height="52" rx="7.5" stroke="var(--darker-gray-color)"/>
<rect class="Rectangle 5" x="395.5" y="140.5" width="132" height="37" rx="7.5" stroke="var(--darker-gray-color)"/>
<text class="or" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-style="italic" letter-spacing="0em"><tspan x="455.476" y="61.6523">or</tspan></text>
<text class="or_2" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-style="italic" letter-spacing="0em"><tspan x="455.476" y="133.652">or</tspan></text>
    </g></g>
<g id="SGWC" opacity="0.3">
<text class="Stable, galaxy-wide civilization" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-weight="bold" letter-spacing="0em"><tspan x="583.109" y="89.6523">Stable, galaxy-wide
</tspan><tspan x="618.222" y="110.652">civilization</tspan></text>
<rect class="Rectangle 6" x="575.5" y="69.5" width="174" height="52" rx="7.5" stroke="var(--darker-gray-color)"/></g>
<g id="ArrowToSGWC">
<path opacity="0.3" class="Arrow 4" d="M562 56.5C562.276 56.5 562.5 56.2761 562.5 56L562.5 51.5C562.5 51.2239 562.276 51 562 51C561.724 51 561.5 51.2239 561.5 51.5L561.5 55.5L557.5 55.5C557.224 55.5 557 55.7239 557 56C557 56.2761 557.224 56.5 557.5 56.5L562 56.5ZM534.646 29.3536L561.646 56.3536L562.354 55.6464L535.354 28.6464L534.646 29.3536Z" fill="var(--darker-gray-color)"/>
<path opacity="0.3" class="Arrow 5" d="M561.354 95.3536C561.549 95.1583 561.549 94.8417 561.354 94.6464L558.172 91.4645C557.976 91.2692 557.66 91.2692 557.464 91.4645C557.269 91.6597 557.269 91.9763 557.464 92.1716L560.293 95L557.464 97.8284C557.269 98.0237 557.269 98.3403 557.464 98.5355C557.66 98.7308
                                       557.976 98.7308 558.172 98.5355L561.354 95.3536ZM535 95.5L561 95.5L561 94.5L535 94.5L535 95.5Z" fill="var(--darker-gray-color)"/>
<path opacity="0.3" class="Arrow 6" d="M561.497 117.053C561.526 116.778 561.327 116.532 561.053 116.503L556.578 116.027C556.304 115.998 556.057 116.196 556.028 116.471C555.999 116.746 556.198 116.992 556.472 117.021L560.45 117.444L560.027 121.422C559.998 121.696 560.196 121.943 560.471 121.972C560.746 122.001 560.992 121.802 561.021 121.528L561.497 117.053ZM535.314 138.389L561.314 117.389L560.686 116.611L534.686 137.611L535.314 138.389Z" fill="var(--darker-gray-color)"/>
</g>
    </svg></a>


<p>
<blockquote>This is one of 4 posts summarizing hundreds of pages of technical reports focused almost entirely on forecasting one number: the year by which transformative AI will be developed.<sup id="fnref1"><a href="#fn1" rel="footnote">1</a></sup><strong> </strong>
</blockquote></p>
<p>
By &quot;transformative AI,&quot; I mean &quot;AI powerful enough to bring us into a new, qualitatively different future.&quot; I specifically focus on what I&apos;m calling <a href="https://www.cold-takes.com/transformative-ai-timelines-part-1-of-4-what-kind-of-ai/">PASTA</a>: AI systems that can essentially automate all of the human activities needed to speed up scientific and technological advancement.
</p>
<p>
The sooner PASTA might be developed, the sooner the world could change <a href="https://www.cold-takes.com/transformative-ai-timelines-part-1-of-4-what-kind-of-ai/#impacts-of-pasta">radically</a>, and the more important it seems to be thinking today about how to make that change go well vs. poorly.
</p>
<p>
This post is a layperson-compatible summary of <a href="https://www.lesswrong.com/posts/KrJfoZzpSDpnrv9va/draft-report-on-ai-timelines">Ajeya Cotra&apos;s &quot;Forecasting Transformative AI with Biological Anchors</a>&quot; (which I&apos;ll abbreviate below as <strong>&quot;Bio Anchors&quot;</strong>), and its pros and cons.<sup id="fnref2"><a href="#fn2" rel="footnote">2</a></sup> It is the forecast I find most informative for transformative AI, with some caveats:
</p>
<ul>
<li>This approach is relatively complex, and it requires a fairly large number of assumptions and uncertain estimates. These qualities make it relatively difficult to explain, and they are also a mark against the method&apos;s reliability. 

</li><li>Hence, as of today, I don&apos;t think this method is as trustworthy as the <a href="https://www.cold-takes.com/are-we-trending-toward-transformative-ai-how-would-we-know/#what-kind-of-forecast-am-i-going-for">examples I gave previously</a> for forecasting a qualitatively different future. It does not have the simplicity and directness of some of those examples, such as modeling COVID-19&apos;s spread. And while climate modeling is also very complex, climate modeling has been worked on by a large number of experts over decades, whereas the Bio Anchors methodology doesn&apos;t have much history.
</li>
</ul>
<p>
Nonetheless, I think it is the best available &quot;best guess estimate&quot; methodology for transformative AI timelines as of today. And as discussed in the <a href="#pros-and-cons-of-the-biological-anchors-method-for-forecasting-transformative-ai-timelines">final section</a>, one can <strong>step back from a lot of the details to see that this century will likely see us hit some of the more &quot;extreme&quot; milestones in the report that strongly suggest the feasibility of transformative AI.</strong>
</p>
<p>
The basic idea is:
</p>
<ul>

<li>Modern AI models can &quot;learn&quot; to do tasks via a (financially costly) process known as &quot;training.&quot; You can think of training as a massive amount of trial-and-error. For example, voice recognition AI models are given an audio file of someone talking, take a guess at what the person is saying, then are given the right answer. By doing this millions of times, they &quot;learn&quot; to reliably translate speech to text. More: <a href="#training">Training</a>

</li><li>The bigger an AI model and the more complex the task, the more the training process costs. Some AI models are bigger than others; to date, none are anywhere near &quot;as big as the human brain&quot; (what this means will be elaborated below). More: <a href="#model-size-and-task-type">Model size and task type</a>

</li><li>The biological anchors method asks: <strong>&quot;Based on the usual patterns in how much training costs, how much would it cost to train an AI model as big as a human brain to perform the hardest tasks humans do? And when will this be cheap enough that we can expect someone to do it?&quot;</strong> More: <a href="#estimating-the-expense">Estimating the expense</a>
</li>
</ul>
<p>
Bio Anchors models a broad variety of different ways of approaching this question, generating estimates in a wide range from &quot;aggressive&quot; (projecting transformative AI sooner) to &quot;conservative&quot; (later). But from essentially all of these angles, it places a high probability on transformative AI this century.
</p>
<figure class="kg-card kg-image-card kg-width-wide"><img src="https://www.cold-takes.com/content/images/size/w1000/2021/08/bio-anchors-probability-chart.png" width="1036" alt="Forecasting transformative AI: the &quot;biological anchors&quot; method in a nutshell"><figcaption>This chart is from the report. You can roughly read the y-axis as the probability that transformative AI is developed by the year in question, although there is some additional nuance in the report. I won&apos;t be explaining what each of the different &quot;Conditional on&quot; models means; it&apos;s enough to know that each represents a different angle on forecasting transformative AI.</figcaption></figure>
<p>
<em></em>
</p>
<p>
<figure class="kg-card kg-image-card kg-width-wide"><img src="https://www.cold-takes.com/content/images/2021/08/bio-anchors-timeline-hack.png" width="1036" alt="Forecasting transformative AI: the &quot;biological anchors&quot; method in a nutshell"><!--<img src="https://www.cold-takes.com/content/images/size/w1000/2021/08/bio-anchors-timeline.png" width=1036><img src="https://www.cold-takes.com/content/images/size/w1000/2021/08/bio-anchors-timeline-2.png" width=1036>--><figcaption>Thanks to Mar&#xED;a Guti&#xE9;rrez Rojas for this graphic. The top timeline gives major milestones for AI computing, past and future (the future ones are projected by Bio Anchors). Below it are (cropped) other timelines showing how significant this few-hundred-year period (more at <a href="https://www.cold-takes.com/this-cant-go-on/">This Can&apos;t Go On</a>), and this era (more at <a href="https://www.cold-takes.com/all-possible-views-about-humanitys-future-are-wild/">All Possible Views About Humanity&apos;s Future Are Wild</a>), appear to be.</figcaption></figure>
</p>
<p>
I&apos;ll now elaborate on each of these a bit more. This is the densest part of this series, and some people might prefer to stick with the above summary and skip to the next post.
</p>
<p>
Note that Bio Anchors uses a number of different approaches (which it calls &quot;anchors&quot;) to estimate transformative AI timelines, and combines them into one aggregate view. In this summary, I&apos;m most focused on a particular set of these - called the &quot;neural net anchors&quot; - which are driving most of the report&apos;s aggregate timelines. Some of what I say applies to all anchors, but some applies only to the &quot;neural net anchors.&quot;
</p>
<h2 id="training">Training</h2>


<p>
As discussed <a href="https://www.cold-takes.com/transformative-ai-timelines-part-1-of-4-what-kind-of-ai/#making-pasta">previously</a>, there are essentially two ways to &quot;teach&quot; a computer to do a task:
</p>
<ol>

<li><strong>&quot;Program&quot; in extremely specific, step-by-step instructions for completing the task.</strong> When this can be done, the computer can generally execute the instructions very quickly, reliably and cheaply. For example, you might program a computer to examine each record in a database and print the ones that match a user&apos;s search terms - you would &quot;instruct&quot; it in exactly how to do this, and it would be able to do the task very well.

</li><li><strong>&quot;Train&quot; an AI to do the task purely by trial and error. </strong>Today, the most common way of doing this is by using a &quot;neural network,&quot; which you might think of sort of like a &quot;digital brain&quot; that starts in a random state: it hasn&apos;t yet been wired to do specific things. For example, say we want an AI to be able to say whether a photo is of a dog or a cat. It&apos;s hard to give fully specific step-by-step instructions for doing this; instead, we can take a neural network and send in a million example images (each one labeled as a &quot;dog&quot; or a &quot;cat&quot;). Each time it sees an example, it will tweak its internal wiring to make it more likely to get the right answer on similar cases in the future. After enough examples, it will be wired to correctly recognize dogs vs. cats.
</li>
</ol>
<p>
(We could maybe also move up another level of meta, and try to &quot;train&quot; models to be able to learn from &quot;training&quot; itself as efficiently as possible. This is called &quot;meta-learning,&quot; but my understanding is that it hasn&apos;t had great success yet.)
</p>
<p>
&quot;Training&quot; is a sort of brute-force, expensive alternative to &quot;programming.&quot; The advantage is that we don&apos;t need to be able to provide specific instructions - we can just give an AI lots of examples of doing the task right, and it will learn to do the task. The disadvantage is that we need a <strong><em>lot</em> of examples, which requires a lot of processing power, which costs money.</strong>
</p>
<p>
How much? This depends on the size of the model (neural network) and the nature of the task itself. For some tasks AIs have learned as of 2021, training a single model could cost millions of dollars. For more complex tasks (such as &quot;<a href="https://www.cold-takes.com/transformative-ai-timelines-part-1-of-4-what-kind-of-ai/#making-pasta">do innovative scientific research</a>&quot;) and bigger models (reaching the size of the human brain), training a model could cost far more than that. 
</p>
<p>
Bio Anchors is interested in the question: <strong>&quot;When will it be affordable to train a model, using a relatively crude trial-and-error-based approach, to do the hardest tasks humans can do?&quot;</strong> 
</p>
<p>
These tasks could include the tasks necessary for <a href="https://www.cold-takes.com/transformative-ai-timelines-part-1-of-4-what-kind-of-ai/">PASTA</a>, such as:
</p>
<ul>

<li>Learn about science from teachers, textbooks and homework as effectively as a human can.

</li><li>Push the frontier of science by asking questions, doing analyses and writing papers, as effectively as a human can.
</li>
</ul>
<p>
The next section will discuss how Bio Anchors fleshes out the idea of the &quot;hardest tasks humans can do&quot; (which it assumes would require a &quot;human-brain-sized&quot; model).
</p>
<h2 id="model-size-and-task-type">Model size and task type</h2>


<p>
Bio Anchors hypothesizes that we can estimate &quot;how expensive it is to train a model&quot; based on two basic parameters: the <strong>model size</strong> and the <strong>task type.</strong>
</p>
<p>
<strong>Model size. </strong>As stated above, you might think of a neural network as a &quot;digital brain&quot; that starts in a random state. In general, a <em>larger</em> &quot;digital brain&quot; - with more digital-versions-of-neurons and digital-versions-of-synapses<sup id="fnref3"><a href="#fn3" rel="footnote">3</a></sup> - can learn more complex tasks. A larger &quot;digital brain&quot; also requires more computations - and is hence more expensive - each time it is used (for example, for each example it is learning from).
</p>
<p>
Drawing on the analysis in <a href="https://www.openphilanthropy.org/brain-computation-report">Joe Carlsmith&apos;s &quot;How Much Computational Power Does It Take to Match the Human Brain?&quot;</a> (abbreviated in this piece as &quot;Brain Computation&quot;), Bio Anchors estimates comparisons between the size of &quot;digital brains&quot; (AI models) and &quot;animal brains&quot; (bee brains, mouse brains, human brains). These estimates imply that <strong>today&apos;s AI systems are sometimes as big as insect brains, but never quite as big as mouse brains</strong> - as of this writing, the largest known language model was the first to come reasonably close<sup id="fnref4"><a href="#fn4" rel="footnote">4</a></sup> - and <strong>not yet even 1% as big as human brains.</strong><sup id="fnref5"><a href="#fn5" rel="footnote">5</a></sup>
</p>
<p>
The bigger the model, the more processing power it takes to train. Bio Anchors assumes that a <strong>transformative AI model would need to be about 10x the size of a human brain,</strong> so a lot bigger than any current AI model. (The 10x is to leave some space for the idea that &quot;digital brains&quot; might be less efficient than human brains; see <a href="https://docs.google.com/document/d/1IJ6Sr-gPeXdSJugFulwIpvavc0atjHGM82QjIfUSBGQ/edit#heading=h.z8ucahktj3ug">this section</a> of the report.) This is one of the reasons it would be very expensive to train.
</p>
<p>
It could turn out that a smaller AI model is still big enough to learn the above sort of tasks. Or it could turn out that the needed model size is bigger than Bio Anchors estimates, perhaps because Bio Anchors has underestimated the effective &quot;size&quot; of the human brain, or because the human brain is better-designed than &quot;digital brains&quot; by more than Bio Anchors has guessed.
</p>
<p>
<strong>Task type. </strong>In order to learn a task, an AI model needs to effectively &quot;try&quot; (or &quot;watch&quot;) the task a large number of times, learning from trial-and-error. The more costly (in processing power, and therefore money) the task is to try/watch, the more costly it will be for the AI model to learn it.
</p>
<p>
It&apos;s hard to quantify how costly a task is to try/watch. Bio Anchors&apos;s attempt to do this is the most contentious part of the analysis, according to the technical reviewers who have reviewed it so far.
</p>
<p>
You can roughly think of the Bio Anchors framework as saying: 
</p>
<ul>

<li>There are some tasks that a human can do with only a second of thought, such as classifying an image as a cat or dog. 

</li><li>There are other tasks that might take a human several minutes of thought, such as solving a logic puzzle.

</li><li>Other tasks could take hours, days, months or even years, and require not just thinking, but interacting with the environment. For example, writing a scientific paper.

</li><li>The tasks on the longer end of this spectrum will be more costly to try/watch, so it will be more costly to train an AI model to do them. For example, it&apos;s more costly (takes more time, and more money) to have a million &quot;tries&quot; at a task that takes an hour than it is to have a million &quot;tries&quot; at a task that takes a second.

</li><li>However, the framework isn&apos;t as simple as this sounds. Many tasks that seem like &quot;long&quot; tasks (such as writing an essay) could in fact be broken into a series of &quot;shorter&quot; tasks (such as writing individual sentences).  
<ul>
 
<li>If an AI model can be trained to do a shorter &quot;sub-task,&quot;, it might be able to do the longer task by simply repeating the shorter sub-task over and over again - without ever needing to be explicitly &quot;trained&quot; to do the longer task. 
 
</li><li>For example, an AI model might get a million &quot;tries&quot; at the task: &quot;Read a partly-finished essay and write a good next sentence.&quot; If it then learns to do this task well, it could potentially write a long essay by simply repeating this task over and over again. It wouldn&apos;t need to go into a separate training process where it gets a million &quot;tries&quot; at the more time-consuming task of writing an entire essay.
 
</li><li>So it becomes crucial whether the hardest and most important tasks (such as those listed above) are the kind that can be &quot;decomposed&quot; into short/easy tasks.
</li> 
</ul>
</li> 
</ul>
<h2 id="estimating-the-expense">Estimating the expense</h2>


<p>
Bio Anchors looks at how expensive existing AI models were to train, depending on model size and task type (as defined above). It then extrapolates this to see how expensive an AI model would be to train if it:
</p>
<ul>

<li>Had a size 10x larger than a human brain.<sup id="fnref6"><a href="#fn6" rel="footnote">6</a></sup>

    </li><li>Trained on a task where each &quot;try&quot; took days, weeks, or months of intensive &quot;thinking.&quot;</li></ul>
<p>
As of today, this sort of training would cost in the ballpark of a million trillion dollars, which is enormously more than total world wealth. So it isn&apos;t surprising that nobody has tried to train such a model. 
</p>
<p>
However, Bio Anchors also projects the following trends out into the future:
</p>
<ul>

<li>Advances in both hardware and software that could make computing power cheaper.

</li><li>A growing economy, and a growing role of AI in the economy, that could increase the amount AI labs are able to spend training large models to $1 trillion and beyond.
</li>
</ul>
<p>
According to these projections, at some point the &quot;amount AI labs are able to spend&quot; becomes equal to the &quot;expense of training a human-brain-sized model on the hardest tasks.&quot; Bio Anchors bases its projections for &quot;when transformative AI will be developed&quot; on when this happens.
</p>
<p>
Bio Anchors also models uncertainty in all of the parameters above, and considers alternative approaches to the &quot;model size and task type&quot; parameters.<sup id="fnref7"><a href="#fn7" rel="footnote">7</a></sup> By doing this, it estimates the probability that transformative AI will be developed by 2030, 2035, etc.
</p>
<h2 id="aggressive-or-conservative">Aggressive or conservative?</h2>


<p>
Bio Anchors involves a number of simplifications that could cause it to be too aggressive (expecting transformative AI to come sooner than is realistic) or too conservative (expecting it to come later than is realistic). 
</p>
<p>
The argument I most commonly hear that it is &quot;<strong>too aggressive</strong>&quot; is along the lines of: &quot;There&apos;s no reason to think that a modern-methods-based AI can learn everything a human does, using trial-and-error training - no matter how big the model is and how much training it does. Human brains can reason in unique ways, unmatched and unmatchable by any AI unless we come up with fundamentally new approaches to AI.&quot; This kind of argument is often accompanied by saying that AI systems don&apos;t &quot;truly understand&quot; what they&apos;re reasoning about, and/or that they are merely imitating human reasoning through pattern recognition. 
</p>
<p>
I think this may turn out to be correct, but I wouldn&apos;t bet on it. A full discussion of why is outside the scope of this post, but in brief:
</p>
<ul>

<li>I am unconvinced that there is a deep or stable distinction between &quot;pattern recognition&quot; and &quot;true understanding&quot; (<a href="https://slatestarcodex.com/2019/02/19/gpt-2-as-step-toward-general-intelligence/">this Slate Star Codex piece</a> makes this point). &quot;True understanding&quot; might just be what really good pattern recognition looks like. Part of my thinking here is an intuition that even when people (including myself) superficially appear to &quot;understand&quot; something, their reasoning often (I&apos;d even say usually) breaks down when considering an unfamiliar context. In other words, I think what we think of as &quot;true understanding&quot; is more of an ideal than a reality.

</li><li>I feel underwhelmed with the track record of those who have made this sort of argument - I don&apos;t feel they have been able to pinpoint what &quot;true reasoning&quot; looks like, such that they could make robust predictions about what would prove difficult for AI systems. (For example, see <a href="https://nostalgebraist.tumblr.com/post/628024664310136832/gary-marcus-has-co-authored-a-brief-critique-of">this discussion of Gary Marcus&apos;s latest critique of GPT3</a>).

</li><li>&quot;Some breakthroughs / fundamental advances are needed&quot; might be true. But for Bio Anchors to be overly aggressive, it isn&apos;t enough that <em>some </em>breakthroughs are needed; the breakthroughs needed have to be <em>more than what AI scientists are capable of in the coming decades</em>, the time frame over which Bio Anchors forecasts transformative AI. It seems hard to be confident that things will play out this way - especially because: 
<ul>
 
<li>Even moderate advances in AI systems could bring more talent and funding into the field (as is already happening<sup id="fnref8"><a href="#fn8" rel="footnote">8</a></sup>). 
 
</li><li>If money, talent and processing power are plentiful, and progress toward PASTA is primarily held up by some particular weakness of how AI systems are designed and trained, a sustained attempt by researchers to fix this weakness could work. When we&apos;re talking about multi-decade timelines, that might be plenty of time for researchers to find whatever is missing from today&apos;s techniques.</li></ul></li></ul>
<p>
More broadly, Bio Anchors could be too aggressive due to its assumption that &quot;computing power is the bottleneck&quot;: 
</p>
<ul>

<li>It assumes that <em>if</em> one could pay for all the computing power to do the brute-force &quot;training&quot; described above for the key tasks (e.g., automating scientific work), this would be enough to develop transformative AI. 

</li><li>But in fact, training an AI model doesn&apos;t just require purchasing computing power. It requires hiring researchers, running experiments, and perhaps most importantly, finding a way to set up the &quot;trial and error&quot; process so that the AI can get a huge number of &quot;tries&quot; at the key task. It may turn out that doing so is prohibitively difficult.
</li>
</ul>
<p>
On the other hand, there are several ways in which Bio Anchors could be <strong>too conservative</strong> (underestimating the likelihood of transformative AI being developed soon). 
</p>
<ul>

<li>Perhaps with enough ingenuity, one could create a transformative AI by &quot;programming&quot; it to do key tasks, rather than having to &quot;train&quot; it (see <a href="#training">above </a>for the distinction). This could require far less computation, and hence be far less expense. Or one could use a combination of &quot;programming&quot; and &quot;training&quot; to achieve better efficiency than Bio Anchors implies, while still not needing to capture everything via &quot;programming.&quot;

</li><li>Or one could find far superior approaches to AI that can be &quot;trained&quot; much more efficiently. One possibility here is &quot;meta-learning&quot;: effectively training an AI system on the &quot;task&quot; of being trained, itself.

</li><li>Or perhaps most likely, over time AI might become a bigger and bigger part of the economy, and there could be a proliferation of different AI systems that have each been customized and invested in to do different real-world tasks. The more this happens, the more opportunity there is for individual ingenuity and luck to result in more innovations, and more capable AI systems in particular economic contexts.  
<ul>
 
<li>Perhaps at some point, it will be possible to integrate many systems with different abilities in order to tackle some particularly difficult task like &quot;automating science,&quot; without needing a dedicated astronomically expensive &quot;training run.&quot;
 
</li><li>Or perhaps AI that falls short of PASTA will still be useful enough to generate a lot of cash, and/or help researchers make compute cheaper and more efficient. This in turn could lead to still bigger AI models that further increase availability of cash and efficiency of compute. That, in turn, could cause a PASTA-level training run to be affordable earlier than Bio Anchors projects.
</li> 
</ul>

</li><li>Additionally, some technical reviewers of Bio Anchors feel that its treatment of <a href="#model-size-and-task-type">task type</a> is too conservative. They believe that the most important tasks (and perhaps all tasks) that AI needs to be trained on will be on the &quot;easier/cheaper&quot; end of the spectrum, compared to what Bio Anchors assumes. (See the <a href="#model-size-and-task-type">above section</a> for what it means for a task to be &quot;easier/cheaper&quot; or &quot;harder/more expensive&quot;). For a related argument, see <a href="https://www.lesswrong.com/posts/rzqACeBGycZtqCfaX/fun-with-12-ooms-of-compute">Fun with +12 OOMs of Compute</a>, which makes the intuitive point that Bio Anchors is imagining a truly massive amount of computation needed to create PASTA, and less could easily be enough.
</li>
</ul>
<p>
I don&apos;t think it is obvious whether, overall, Bio Anchors is too aggressive (expecting transformative AI to come sooner than is realistic) or too conservative (expecting it to come later). The report itself states that it&apos;s likely to be too aggressive over the next few years and too conservative &gt;50 years out, and likely most useful in between.<sup id="fnref9"><a href="#fn9" rel="footnote">9</a></sup>
</p>
<p>
Intellectually, it feels to me as though the report is more likely to be too conservative. I find its <a href="https://docs.google.com/document/d/1cCJjzZaJ7ATbq8N2fvhmsDOUWdm7t3uSSXv6bD0E_GM/edit#heading=h.hu3zy1xzvvo">responses</a> to the &quot;Too aggressive&quot; points above fairly compelling, and I think the &quot;Too conservative&quot; points are more likely to end up being correct. In particular, I think it&apos;s hard to rule out the possibility of ingenuity leading to transformative AI in some far more efficient way than the &quot;brute-force&quot; method contemplated here. And I think the treatment of &quot;task type&quot; is definitely erring in a conservative direction.
</p>
<p>
However, I also have an intuitive preference (which is related to the &quot;burden of proof&quot; analyses given <a href="https://www.cold-takes.com/forecasting-transformative-ai-whats-the-burden-of-proof/">previously</a>) to err on the conservative side when making estimates like this. Overall, my best guesses about transformative AI timelines are similar to those of Bio Anchors.
</p>
<h2 id="conclusions-of-bio-anchors">Conclusions of Bio Anchors</h2>


<p>
Bio Anchors estimates a <strong>&gt;10% chance of transformative AI by 2036, a ~50% chance by 2055, and an ~80% chance by 2100.</strong>
</p>
<p>
It&apos;s also worth noting what the report says about AI systems today. It estimates that:
</p>
<ul>

<li>Today&apos;s largest AI models, such as <a href="https://en.wikipedia.org/wiki/GPT-3">GPT-3</a>, are a <strong>bit smaller than mouse brains, and are starting to get within range (if they were to grow another 100x-1000x) of human brains. </strong>So we might soon be getting close to AI systems that can be trained to do anything that humans can do with ~1 second of thought. Consistent with this, it seems to me that we&apos;re just starting to reach the point where language models <em>sound</em> like humans who are talking without thinking very hard.<sup id="fnref10"><a href="#fn10" rel="footnote">10</a></sup> If anything, &quot;human who puts in no more than 1 second of thought per word&quot; seems somewhat close to what GPT-3 is doing, even though it&apos;s much smaller than a human brain.

</li><li>It&apos;s only very recently that AI models have gotten this big. A &quot;large&quot; AI model before 2020 would be more in the range of a honeybee brain. So for models even in the very recent past, we should be asking whether AI systems seem to be &quot;as smart as insects.&quot; Here&apos;s <a href="https://www.lesswrong.com/posts/yW3Tct2iyBMzYhTw7/how-does-bee-learning-compare-with-machine-learning">one attempt to compare AI and honeybee capabilities</a> (by Open Philanthropy intern Guille Costa), concluding that the most impressive honeybee capabilities the author was able to pinpoint do appear to be doable for AI systems.<sup id="fnref11"><a href="#fn11" rel="footnote">11</a></sup></li></ul>
<p>
I include these notes because:
</p>
<ul>

<li>The Bio Anchors analysis seems fully consistent with what we&apos;re observing from AI systems today (and have over the last decade or two), while also implying that we&apos;re likely to see more transformative abilities in the coming decades.

</li><li>I think it&apos;s particularly noteworthy that we&apos;re getting close to the time when an AI model is &quot;as big as a human brain&quot; (according to the Bio Anchors / <a href="https://www.openphilanthropy.org/brain-computation-report">Brain Computation </a>estimation method). It may turn out that such an AI model is able to &quot;learn&quot; a lot about the world and produce a lot of economic value, even if it can&apos;t yet do the hardest things humans do. And this, in turn, could kick off skyrocketing investment in AI (both money and talent), leading to a lot more innovation and further breakthroughs. This is a simple reason to believe that transformative AI by 2036 is plausible.
</li>
</ul>
<p>
Finally, I note that Bio Anchors includes an &quot;evolution&quot; analysis among the different approaches it considers. This analysis hypothesizes that in order to produce transformative AI, one would need to do about as many computations as all animals in history combined, in order to re-create the progress that was made by natural selection. 
</p>
<p>
I consider the &quot;evolution&quot; analysis to be <em>very</em> conservative, because machine learning is capable of much faster progress than the sort of trial-and-error associated with natural selection. Even if one believes in something along the lines of &quot;Human brains reason in unique ways, unmatched and unmatchable by a modern-day AI,&quot; it seems that whatever is unique about human brains should be re-discoverable if one is able to essentially re-run the whole history of natural selection. And even this very conservative analysis estimates a ~50% chance of transformative AI by 2100.
</p>
<h2 id="pros-and-cons-of-the-biological-anchors-method-for-forecasting-transformative-ai-timelines">Pros and cons of the biological anchors method for forecasting transformative AI timelines</h2>


<p>
<strong>Cons. </strong>I&apos;ll start with what I see as the biggest downside: this is a very complex forecasting framework, which relies crucially on multiple extremely uncertain estimates and assumptions, particularly:
</p>
<ul>

<li>Whether it&apos;s reasonable to believe that an AI system could learn the key tasks listed above (the ones required for PASTA) given enough trial-and-error training.

</li><li>How to compare the size of AI models with the size of animal/human brains.

</li><li>How to characterize &quot;task type,&quot; estimating how &quot;difficult&quot; and expensive a task is to &#x201C;try&#x201D; or &#x201C;watch&#x201D; once.

</li><li>How to use the model size and task type to estimate how expensive it would be to train an AI model to do the key tasks.

</li><li>How to estimate future advances in both hardware and software that could make computing power cheaper.

</li><li>How to estimate future increases in how much AI labs could be able to spend training models.
</li>
</ul>
<p>
This kind of complexity and uncertainty means (IMO) that we shouldn&apos;t consider the forecasts to be highly reliable, especially today when the whole framework is fairly new. If we got to the point where as much scrutiny and effort had gone into AI forecasting as climate forecasting, it might be a different matter.
</p>
<p>
<strong>Pros. </strong>That said, the biological anchors method is essentially the only one I know of that estimates transformative AI timelines from <strong>objective facts </strong>(where possible) <strong>and explicit assumptions </strong>(elsewhere)<strong>.</strong><sup id="fnref12"><a href="#fn12" rel="footnote">12</a></sup><strong> </strong>It does not rely on any concepts as vague and intuitive as &quot;how fast AI systems are getting more impressive&quot; (discussed <a href="https://www.cold-takes.com/are-we-trending-toward-transformative-ai-how-would-we-know/#subjective-extrapolations-and-">previously</a>). Every assumption and estimate in the framework can be explained, discussed, and - over time - tested. 
</p>
<p>
Even in its current early stage, I consider this a valuable property of the biological anchors framework. It means that the framework can give us timelines estimates that aren&apos;t simply rehashes of intuitions about whether it feels as though transformative AI is approaching.<sup id="fnref13"><a href="#fn13" rel="footnote">13</a></sup> 
</p>
<p>
I also think it&apos;s encouraging that even with all the guesswork, the testable &quot;predictions&quot; the framework makes as of today seem reasonable (see previous section). <strong>The framework provides a way of thinking about how it could be simultaneously true that (a) the AI systems of a decade ago didn&apos;t seem very impressive at all; (b) the AI systems of today can do many impressive things but still feel far short of what humans are able to do; (c) the next few decades - or even the next 15 years - could easily see the development of transformative AI.</strong>
</p>
<p>
Additionally, I think it&apos;s worth noting a <strong>couple of high-level points</strong> from Bio Anchors that <strong>don&apos;t depend on quite so many estimates and assumptions:</strong>
</p>
<ul>

<li>In the coming decade or so, we&apos;re likely to see - for the first time - AI models with comparable &quot;size&quot; to the human brain. 

</li><li>If AI models continue to become larger and more efficient at the rates that Bio Anchors estimates, it will probably become <strong>affordable this century to hit some pretty extreme milestones - the &quot;high end&quot; of what Bio Anchors thinks might be necessary. </strong>These are hard to summarize, but see the &quot;long horizon neural net&quot; and &quot;evolution anchor&quot; frameworks in the report. 

</li><li>One way of thinking about this is that the next century will likely see us go from &quot;not enough compute to run a human-sized model at all&quot; to &quot;extremely plentiful compute, as much as even quite conservative estimates of what we might need.&quot; Compute isn&apos;t the only factor in AI progress, but to the extent other factors (algorithms, training processes) became the new bottlenecks, there will likely be powerful incentives (and multiple decades) to resolve them.
</li>
</ul>
<p>
A final advantage of Bio Anchors is that we can continue to watch AI progress over time, and compare what we see to the report&apos;s framework. For example, we can watch for:
</p>
<ul>

<li>Whether there are some tasks that just can&apos;t be learned, even with plenty of trial and error - or whether some tasks require amounts of training very different from what the report estimates.

</li><li>How AI models&apos; capabilities compare to those of animals that we are currently modeling as &quot;similarly sized.&quot; If AI models seem more capable than such animals, we may be overestimating how large a model we would need to be in order to e.g. automate science. If they seem less capable, we may be underestimating it.

</li><li>How hardware and software are progressing, and whether AI models are getting bigger at the rate the report currently projects.
</li>
</ul>
<p>
The next piece will summarize all of the different analyses so far about transformative AI timelines. It will then discuss a remaining reservation: that there is no robust expert consensus on this topic.
</p>

<p><strong>Next in series:</strong> <a href="https://www.cold-takes.com/where-ai-forecasting-stands-today/">AI Timelines: Where the Arguments, and the &quot;Experts,&quot; Stand</a></p><!--kg-card-end: html--><!--kg-card-begin: html-->

<p><div style="display:flex; justify-content:center; margin: 0 auto;">

        <span style="margin: 10px;"><a href="https://api.addthis.com/oexchange/0.8/forward/twitter/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fforecasting-transformative-ai-the-biological-anchors-method-in-a-nutshell%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20Forecasting%20transformative%20AI%3A%20the%20%22biological%20anchors%22%20method%20in%20a%20nutshell&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-twitter-square.png" border="0" alt="Forecasting transformative AI: the &quot;biological anchors&quot; method in a nutshell"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/facebook/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fforecasting-transformative-ai-the-biological-anchors-method-in-a-nutshell%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20Forecasting%20transformative%20AI%3A%20the%20%22biological%20anchors%22%20method%20in%20a%20nutshell&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-facebook-square.png" border="0" alt="Forecasting transformative AI: the &quot;biological anchors&quot; method in a nutshell"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/reddit/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fforecasting-transformative-ai-the-biological-anchors-method-in-a-nutshell%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20Forecasting%20transformative%20AI%3A%20the%20%22biological%20anchors%22%20method%20in%20a%20nutshell&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-reddit-square.png" border="0" alt="Forecasting transformative AI: the &quot;biological anchors&quot; method in a nutshell"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/menu/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fforecasting-transformative-ai-the-biological-anchors-method-in-a-nutshell%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20Forecasting%20transformative%20AI%3A%20the%20%22biological%20anchors%22%20method%20in%20a%20nutshell&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-addthis-square.png" border="0" alt="Forecasting transformative AI: the &quot;biological anchors&quot; method in a nutshell"></a></span>
        </div>
<center><p id="survey"><a href="https://www.cold-takes.com/forecasting-transformative-ai-the-biological-anchors-method-in-a-nutshell/#/portal/signup/free" target="_blank"><button class="button">Subscribe</button></a>&#xA0;<a href="https://www.guidedtrack.com/programs/4kal2ue/run?posttitle=Forecasting%20transformative%20AI%3A%20the%20%22biological%20anchors%22%20method%20in%20a%20nutshell" target="_blank"><button class="button" id="Survey">Feedback</button></a>&#xA0;<a href="https://forum.effectivealtruism.org/posts/vCaEnTbZ5KbypaGsm/forecasting-transformative-ai-the-biological-anchors-method#comments" target="_blank"><button class="button">Forum</button></a></p><p><em>
Use &quot;Feedback&quot; if you have comments/suggestions you want me to see, or if you&apos;re up for giving some quick feedback about this post (which I greatly appreciate!) Use &quot;Forum&quot; if you want to discuss this post publicly on the Effective Altruism Forum.
</em></p></center>
<!--kg-card-end: html--><!--kg-card-begin: html--><hr>
</p><h2>Footnotes</h2>
<div class="footnotes">
<ol><li id="fn1">
<p>
     Of course, the answer could be &quot;A kajillion years from now&quot; or &quot;Never.&quot;&#xA0;<a href="#fnref1" rev="footnote">&#x21A9;</a><li id="fn2">
<p>
     For transparency, note that this is an <a href="https://www.openphilanthropy.org">Open Philanthropy</a> analysis, and I am co-CEO of Open Philanthropy.&#xA0;<a href="#fnref2" rev="footnote">&#x21A9;</a><li id="fn3">
<p>
     I (like Bio Anchors) generally consider the synapse count more important than the neuron count, for reasons I won&apos;t go into here.&#xA0;<a href="#fnref3" rev="footnote">&#x21A9;</a><li id="fn4">
<p>
     <a href="https://en.wikipedia.org/wiki/GPT-3">Wikipedia</a>: &quot;GPT-3&apos;s full version has a capacity of 175 billion machine learning parameters ... Before the release of GPT-3, the largest language model was Microsoft&apos;s Turing NLG, introduced in February 2020, with a capacity of 17 billion parameters.&quot; Wikipedia doesn&apos;t state this, but I don&apos;t believe there are publicly known AI models larger than these language models (with the exception of &quot;<a href="https://arxiv.org/abs/1701.06538">mixture-of-experts models</a>&quot; that I think we should disregard for these purposes, for reasons I won&apos;t go into here). <a href="https://en.wikipedia.org/wiki/List_of_animals_by_number_of_neurons">Wikipedia estimates</a> about 1 trillion synapses for a house mouse&apos;s brain; Bio Anchors&apos;s methodology for brain comparisons (based on <a href="https://www.openphilanthropy.org/brain-computation-report">Brain Computation</a>) essentially equates synapses to parameters.&#xA0;<a href="#fnref4" rev="footnote">&#x21A9;</a><li id="fn5">
<p>
     Bio Anchors estimates about 100 trillion parameters for the human brain, based on the fact that it has about 100 trillion synapses.&#xA0;<a href="#fnref5" rev="footnote">&#x21A9;</a><li id="fn6">

<p>
     As noted above, the 10x is to leave some space for the idea that &quot;digital brains&quot; might be less efficient than human brains. See <a href="https://docs.google.com/document/d/1IJ6Sr-gPeXdSJugFulwIpvavc0atjHGM82QjIfUSBGQ/edit#heading=h.z8ucahktj3ug">this section</a> of the report.&#xA0;<a href="#fnref6" rev="footnote">&#x21A9;</a><li id="fn7">
<p>
     For example, one approach hypothesizes that training could be made cheaper by &quot;meta-learning,&quot; discussed above; another approach hypothesizes that in order to produce transformative AI, one would need to do about as many computations as all animals in history combined, in order to re-create the progress that was made by natural selection.)&#xA0;<a href="#fnref7" rev="footnote">&#x21A9;</a><li id="fn8">

<p>
     See charts from the early sections of the <a href="https://aiindex.stanford.edu/wp-content/uploads/2021/03/2021-AI-Index-Report_Master.pdf">2021 AI Index Report</a>, for example.&#xA0;<a href="#fnref8" rev="footnote">&#x21A9;</a><li id="fn9">
<p>
     See <a href="https://docs.google.com/document/d/1IJ6Sr-gPeXdSJugFulwIpvavc0atjHGM82QjIfUSBGQ/edit#heading=h.y045l51rb826">this section</a>.&#xA0;<a href="#fnref9" rev="footnote">&#x21A9;</a><li id="fn10">

<p>
     For a collection of links to GPT-3 demos, see <a href="https://www.lesswrong.com/posts/6Hee7w2paEzHsD6mn/collection-of-gpt-3-results">this post</a>.&#xA0;<a href="#fnref10" rev="footnote">&#x21A9;</a><li id="fn11">
<p>
     In fact, he estimates that AI systems appear to use about 1000x less compute, which would match the above point in terms of suggesting that AI systems might be more efficient than animal/human brains and that the Bio Anchors estimates might be too conservative. However, he doesn&apos;t address the fact that bees arguably perform a more diverse set of tasks than the AI systems they&apos;re being compared to.&#xA0;<a href="#fnref11" rev="footnote">&#x21A9;</a><li id="fn12">
<p>
     Other than the &quot;semi-informative priors&quot; method discussed <a href="https://www.cold-takes.com/forecasting-transformative-ai-whats-the-burden-of-proof/">previously</a>.&#xA0;<a href="#fnref12" rev="footnote">&#x21A9;</a><li id="fn13">
<p>
     Of course, this isn&apos;t to say the estimates are <em>completely independent</em> of intuitions - intuitions are likely to color our choices of estimates for many of the difficult-to-estimate figures. But the ability to scrutinize and debate each estimate separately is helpful here.&#xA0;<a href="#fnref13" rev="footnote">&#x21A9;</a>

</p></li></p></li></p></li></p></li></p></li></p></li></p></li></p></li></p></li></p></li></p></li></p></li></p></li></ol></div><!--kg-card-end: html-->]]></content:encoded></item><item><title><![CDATA[More on “multiple world-size economies per atom”]]></title><description><![CDATA[A follow up on "This Can't Go On" for the skeptical.]]></description><link>https://www.cold-takes.com/more-on-multiple-world-size-economies-per-atom/</link><guid isPermaLink="false">6127e8d3a7ed4b003bf8053a</guid><dc:creator><![CDATA[Holden Karnofsky]]></dc:creator><pubDate>Fri, 27 Aug 2021 18:07:27 GMT</pubDate><content:encoded><![CDATA[<!--kg-card-begin: html--><p>
In <a href="https://www.cold-takes.com/this-cant-go-on/">This Can&#x2019;t Go On</a>, I <a href="https://www.cold-takes.com/this-cant-go-on/#why-cant-this-go-on">argued</a> that 8200 more years of today&#x2019;s growth rate would require us to sustain &#x201C;multiple economies as big as today&apos;s entire world economy <em>per atom</em>.&#x201D;
</p>
<p>
Feedback on this bit was split between &#x201C;That is so obviously impossible, 8200 years of 2% growth is an absurd idea - growth will have to slow much before then&#x201D; and &#x201C;Why is that impossible? With ever-increasing creativity, we could increase quality of life higher and higher, without needing to keep using more and more material resources.&#x201D;
</p>
<p>
Here I&#x2019;m going to respond to the latter point, which means expanding on why 8200 years of 2% growth doesn&#x2019;t look like a reasonable thing to expect. I&#x2019;m going to make lots of extremely wild assumptions and talk about all kinds of weird possibilities just so that I cover even far-fetched ways for 2% growth to continue. 
</p>
<p>
If you are already on team &#x201C;Yeah, I don&#x2019;t see the world economy growing that much,&#x201D; you should skip this post unless you&apos;d enjoy seeing the case made in a fair amount of detail.
</p>
<h2 id="how-we-could-support">How we COULD support &#x201C;multiple world-size economies per atom&#x201D;</h2>

<p>
I do think it&#x2019;s <em>conceivable</em> that we could support multiple world-size economies per atom. Here&#x2019;s one way: 
</p>
<p>
Say that we discover some new activity, or experience, or drug, that people really, really, REALLY value.
</p>
<p>
Specifically, the market values it at 10^85 of today&#x2019;s US dollars (that&#x2019;s ten trillion trillion trillion trillion trillion trillion trillion dollars). That means it&apos;s valued about 10^71 times as much as everything the world produces in a year right now (combined).<sup id="fnref1"><a href="https://www.cold-takes.com/p/864b10db-9ba3-4e6e-add3-5cad875af1b0#fn1" rel="footnote">1</a></sup>
</p>
<p>
Then, one person having this experience<sup id="fnref2"><a href="https://www.cold-takes.com/p/864b10db-9ba3-4e6e-add3-5cad875af1b0#fn2" rel="footnote">2</a></sup> would mean the size of the economy is at least $10^85. And that would, indeed, be the equivalent of multiple of today&#x2019;s world economies per atom.<sup id="fnref3"><a href="https://www.cold-takes.com/p/864b10db-9ba3-4e6e-add3-5cad875af1b0#fn3" rel="footnote">3</a></sup>
</p>
<p>
To be clear, it&#x2019;s not that we would&#x2019;ve crammed multiple of today&#x2019;s world economies into each atom. It&#x2019;s that we would&#x2019;ve crammed something 10^71 times as valuable as today&#x2019;s world economy into a mere <a href="https://science.howstuffworks.com/atoms-in-person.htm#:~:text=A%20human%20body%20weighing%20154,7%20followed%20by%2027%20zeros.">10^28 atoms</a> that make up a human being.
</p>
<p>
What would it mean, though, to value a single experience 10^71 times as much as today&#x2019;s entire world economy?
</p>
<p>
One way of thinking about it might be: 
</p>
<ul>

<li>&#x201C;A 1 in 10^71 chance of this thing being experienced would be as valuable as all of today&#x2019;s world economy.&#x201D; 

</li><li>Or to make it a bit easier to intuit (while needing to oversimplify), &#x201C;If I were risk-neutral, I&#x2019;d be thrilled to accept a gamble where I would die immediately, with near certainty, in exchange for a 1 in 10^71 chance of getting to have this experience.&#x201D;<sup id="fnref4"><a href="https://www.cold-takes.com/p/864b10db-9ba3-4e6e-add3-5cad875af1b0#fn4" rel="footnote">4</a></sup>

</li><li>How near-certain would death be? Well, for starters, if all the people who have ever lived to date accepted this gamble, it would be approximately certain that they would <em>all</em> lose and end up with immediate death.<sup id="fnref5"><a href="https://www.cold-takes.com/p/864b10db-9ba3-4e6e-add3-5cad875af1b0#fn5" rel="footnote">5</a></sup> 

</li><li>But this really isn&#x2019;t coming anywhere close to communicating how bad the odds would be for this gamble. It&#x2019;s more like: if there were one person for each atom in the galaxy, and each of them took the gamble, they&apos;d probably still <strong>all </strong>lose.<sup id="fnref6"><a href="https://www.cold-takes.com/p/864b10db-9ba3-4e6e-add3-5cad875af1b0#fn6" rel="footnote">6</a></sup>

</li><li>So to personally take a gamble with those kinds of odds &#x2026; the experience had better be REALLY good to compensate.  
<ul>
 
<li>We&#x2019;re not talking about &#x201C;the best experience you&#x2019;ve ever had&#x201D; level here - it wouldn&#x2019;t be sensible to value that more than an entire life, and the idea that it&#x2019;s worth as much as today&#x2019;s world economy seems pretty clearly wrong. 
 
</li><li>We&#x2019;re talking about something just unfathomably beyond anything any human has ever experienced.
</li> 
</ul>
</li> 
</ul>
<h2 id="blowing-out-the-numbers-more">Blowing out the numbers more</h2>


<p>
Imagine the single best second of your life, the kind of thing evoked by <a href="https://www.nickbostrom.com/utopia.html">Letter from Utopia</a>:
</p>
<p>
<blockquote>Have you ever experienced a moment of bliss? On the rapids of inspiration maybe, your mind tracing the shapes of truth and beauty? Or in the pulsing ecstasy of love? Or in a glorious triumph achieved with true friends? Or in a conversation on a vine-overhung terrace one star-appointed night? Or perhaps a melody smuggled itself into your heart, charming it and setting it alight with kaleidoscopic emotions? Or when you prayed, and felt heard?
</blockquote></p>
<p>
If you have experienced such a moment &#x2013; experienced <em>the best type</em> of such a moment &#x2013; then you may have discovered inside it a certain idle but sincere thought: &#x201C;Heaven, yes! I didn&#x2019;t realize it could be like this. This is so right, on whole different level of right; so real, on a whole different level of real. Why can&#x2019;t it be like this always? Before I was sleeping; now I am awake.&#x201D;
</p>
<p>
Yet a little later, scarcely an hour gone by, and the ever-falling soot of ordinary life is already covering the whole thing. The silver and gold of exuberance lose their shine, and the marble becomes dirty.
</p>
<p>
Now imagine, implausibly, that this single second was worth as much as the entire world economy outputs in a year today. (It doesn&#x2019;t seem possible that it could be worth more, since the world economy that year <em>included</em> that second of your life, plus the rest of your year and many other people&#x2019;s years.)
</p>
<p>
And now imagine a <em>full year</em> in which <em>every second</em> is as good as <em>that second. </em>We&#x2019;ll call this the &#x201C;perfect year.&#x201D; According to the assumptions above, the perfect year would be no more than about 3*10^8 times as valuable as the world economy (there are about 3*10^8 seconds in a year).
</p>
<p>
And now imagine that <em>every atom in the galaxy</em> could be a person having the perfect year. This would now be about 10^70 * (3 * 10^8) = 3*10^78 as much value as today&#x2019;s world economy. <strong>2% growth would get us there in 9150 years.</strong>
</p>

<p>(A crucial and perhaps counterintuitive assumption I&apos;m making here, throughout, is that &quot;2% growth&quot; means &quot;2% <em>really real</em> growth&quot; - that whatever is valuable, holistically speaking, about annual world output today, we&apos;ll get 2% more of it each year. I think this is already the kind of assumption many people are making when they say we don&apos;t need more material to have ever-increasing wealth. If you think the 2% growth of the recent past is more &quot;fake&quot; than this and that it will continue in a &quot;fake&quot; way, that would be a debate for another time.)</p>

<p>
And 1200 years after <em>that</em>, if each year still had 2% growth, the economy would be another ~20 billion times bigger. So now, for every atom in the galaxy, there&#x2019;d have to be someone whose year was in some sense ~20 billion times <em>better</em> (or &quot;more valuable&quot;) than the perfect year. 
</p>
<!--<p>
Again, one way to think about this is that they’d accept a near-certain chance of losing the perfect year entirely (simply not existing) in exchange for a 1 in 20 billion chance of having this new, even better kind of year. If everyone alive today took that gamble, what would <em>probably</em> happen is that every single person would lose.<sup id="fnref7"><a href="https://www.cold-takes.com/p/864b10db-9ba3-4e6e-add3-5cad875af1b0#fn7" rel="footnote">7</a></sup>
</p>-->
<p>
We&#x2019;re still only talking about ~10,000 years of 2% growth.
</p>
<h2 id="new-life-forms">New life forms</h2>


<p>
It&#x2019;s still conceivable! Who knows what the future will bring.
</p>
<p>
But at this point it&#x2019;s very intuitive to me that we are not talking about anything that looks like &#x201C;Humans in human bodies having human kinds of fun and fulfillment.&#x201D; An economy of this value seems to require fundamentally re-engineering something about the human experience - finding some way of arranging matter that creates far more happiness, or fulfillment, or something, that we would value so astronomically more than even the heights of human experience today.
</p>
<p>
And I think the most natural way for that to happen is something like: &#x201C;Discovering fundamental principles behind what we value, and fundamental principles of how to arrange matter to get the most of it.&#x201D; Which in turn suggests something more like &#x201C;Once we have that level of understanding, we start to arrange the matter in the galaxy optimally, and quickly get close to the limits of what&#x2019;s possible&#x201D; than like &#x201C;We grow at 2%, every year, for continuing thousands of years, even as (as would happen with e.g. <a href="https://www.cold-takes.com/how-digital-people-could-change-the-world/">digital people</a>) we become beings who can do as much in a year as humans could do in hundreds or thousands of years.&#x201D;
</p>
<h2 id="but-it-could-still-happen">But it could still happen?</h2>


<p>
I guess? This was never meant to be a mathematical proof of the impossibility of 2%/year growth. It&#x2019;s possible in theory.
</p>
<p>
But at this point, seeing what a funky and fundamentally transformed galaxy it would require within 10,000 years, what is the <em>affirmative</em> reason to expect 2%/year growth for that long a period of time? Is it that &#x201C;This is the trendline, and by default I expect the trendline to continue?&#x201D;
</p>
<p>
But that trendline is only a couple hundred years old - why expect it to continue for another 10,000? 
</p>
<p>
Why not, instead, expect the <em><a href="https://www.cold-takes.com/this-cant-go-on/#explosion-and-collapse">longer-term pattern of accelerating economic growth </a></em>to be what continues, until we approach some sort of fundamental limit on how much value we can cram into a given amount of matter? Or expect growth to fall gradually from here and never reach today&apos;s level again?
</p>
<p>
The last couple of centuries have been a wild ride, with wealth and living conditions improving at a historically high rate. But I don&#x2019;t think that gives us reason to think that this trend goes to infinity. I believe the limits are somewhere, and it looks like sometime in the next 10,000 years, we&#x2019;re either going to have to approach those limits, or <a href="https://www.cold-takes.com/this-cant-go-on/#why-cant-this-go-on">stagnate or collapse</a>.
</p>
<p>
Hopefully I&#x2019;ve given a sense for why it seems so unlikely that there will be 10,000 more years in the future that each have 2% or greater growth. Which would imply that <em>each</em> of the last 100+ years will turn out to be one of the fastest-growing 10,000 years of all time.
</p>
<p>If you&apos;d like to comment on this post, <a href="https://forum.effectivealtruism.org/posts/pFHN3nnN9WbfvWKFg/this-can-t-go-on?commentId=K3ESWKD3Kk5ywN7MP">this</a> would be a good place to do so.</p>

<!--kg-card-end: html--><!--kg-card-begin: html-->

<p><div style="display:flex; justify-content:center; margin: 0 auto;">

        <span style="margin: 10px;"><a href="https://api.addthis.com/oexchange/0.8/forward/twitter/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fmore-on-multiple-world-size-economies-per-atom%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20More%20on%20%E2%80%9Cmultiple%20world-size%20economies%20per%20atom%E2%80%9D&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-twitter-square.png" border="0" alt="Twitter"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/facebook/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fmore-on-multiple-world-size-economies-per-atom%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20More%20on%20%E2%80%9Cmultiple%20world-size%20economies%20per%20atom%E2%80%9D&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-facebook-square.png" border="0" alt="Facebook"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/reddit/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fmore-on-multiple-world-size-economies-per-atom%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20More%20on%20%E2%80%9Cmultiple%20world-size%20economies%20per%20atom%E2%80%9D&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-reddit-square.png" border="0" alt="Reddit"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/menu/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fmore-on-multiple-world-size-economies-per-atom%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20More%20on%20%E2%80%9Cmultiple%20world-size%20economies%20per%20atom%E2%80%9D&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-addthis-square.png" border="0" alt="More"></a></span>
        </div>
<center><p id="survey"><a href="https://www.cold-takes.com/more-on-multiple-world-size-economies-per-atom/#/portal/signup/free" target="_blank"><button class="button">Subscribe</button></a>&#xA0;<a href="https://www.guidedtrack.com/programs/4kal2ue/run?posttitle=More%20on%20%E2%80%9Cmultiple%20world-size%20economies%20per%20atom%E2%80%9D" target="_blank"><button class="button" id="Survey">Feedback</button></a><!--No Forum post--></p></center>
<!--kg-card-end: html--><!--kg-card-begin: html--><hr>
</p><h2 id="footnotes">Footnotes</h2>
<div class="footnotes">
<ol><li id="fn1">
<p>
     Today&#x2019;s economy is a bit less than $10^14 per year (<a href="https://en.wikipedia.org/wiki/Gross_world_product">source</a>). $10^85 = $10^14 * 10^71.&#xA0;<a href="#fnref1" rev="footnote">&#x21A9;</a><li id="fn2">
<p>
     (And paying full price for it, in a way that gets recorded by GDP statistics, which could get a bit hairy.)&#xA0;<a href="#fnref2" rev="footnote">&#x21A9;</a><li id="fn3">
<p>
     See <a href="https://www.cold-takes.com/this-cant-go-on/#fn7">previous estimate</a> of 10^70 atoms in the galaxy.&#xA0;<a href="#fnref3" rev="footnote">&#x21A9;</a><li id="fn4">

<p>
     This assumes that one values one&#x2019;s own life not much more than a year of the world economy&#x2019;s output. I do not expect that I will see enough disagreement on this point to want to write another post on the matter, but it&#x2019;s possible.</p>
    
    <p>It is also making an iffy assumption about &quot;risk-neutrality.&quot; In reality, one might personally value this experience much less than 10^71 times as much as one&apos;s own life, while still paying resources for it that would be sufficient to save an extraordinarily large number of <em>other</em> people&apos;s lives. It&apos;s hard to convey the same kind of magnitudes by appealing to impartiality, so I went with this intuition pump anyway; I think it does give the right basic sense of how mind-bogglingly large the value of this experience would be.&#xA0;<a href="#fnref4" rev="footnote">&#x21A9;</a><li id="fn5">
<p>
     The calculation here would be: if there are 10^10 people alive today (this is &quot;rounding up&quot; from ~8 billion to 10 billion), and each has a 10^-71 (1 in 10^71) chance of winning the gamble, then each has a (1-10^-71) chance of losing the gamble. So the probability that they <strong>all</strong> lose the gamble is (1-10^-71)^(10^10), which is almost exactly 100%.&#xA0;<a href="#fnref5" rev="footnote">&#x21A9;</a><li id="fn6">
<p>
     Similar calculation to the previous footnote, but with a population of 10^70 (one for each <a href="https://www.cold-takes.com/this-cant-go-on/#fn7">atom in the galaxy</a>), so the probability that they all lose the gamble is (1-10^-71)^(10^70), which I think is around 90% (Excel can&apos;t actually handle numbers this big but this is what similar calculations imply).&#xA0;<a href="#fnref6" rev="footnote">&#x21A9;</a><li id="fn7">
<p>
     (Footnote deleted)&#xA0;<a href="#fnref7" rev="footnote">&#x21A9;</a>

</p></li></p></li></p></li></p></li></p></li></p></li></p></li></ol></div><!--kg-card-end: html--><!--kg-card-begin: html--><p style="font-size:1%">For email filter: florpschmop</p><!--kg-card-end: html-->]]></content:encoded></item><item><title><![CDATA[The gloves are off, the pants are on]]></title><description><![CDATA[I collect good examples of how much BS is floating around out there unchecked.]]></description><link>https://www.cold-takes.com/the-gloves-are-off-the-pants-are-on/</link><guid isPermaLink="false">60dd57aa6da77c003b2e76fc</guid><dc:creator><![CDATA[Holden Karnofsky]]></dc:creator><pubDate>Thu, 26 Aug 2021 19:06:44 GMT</pubDate><media:content url="https://www.cold-takes.com/content/images/2021/08/winston-churchill-w-meter.png" medium="image"/><content:encoded><![CDATA[<!--kg-card-begin: html-->
<img src="https://www.cold-takes.com/content/images/2021/08/winston-churchill-w-meter.png" alt="The gloves are off, the pants are on"><p>
<em>A lie gets halfway around the world before the truth has a chance to get its pants on. - <a href="https://www.brainyquote.com/quotes/winston_churchill_103564">Winston Churchill</a></em>
</p>
<p>
<em>(...<a href="https://professorbuzzkill.com/twain-lie-travels/">Not really, though</a>)</em>
</p>
<p>
I collect links that are good examples of &quot;how much BS is floating around out there unchecked.&quot; Though in many cases I haven&apos;t run everything to ground, and it could be the debunking that&apos;s BS, or both the original and the debunking ... so I do want the lesson to be &quot;Don&apos;t trust things you&apos;ve heard,&quot; as opposed to &quot;Trust this debunking.&quot; (I will probably make a future post devoted to debunking debunkings.)
</p>
<p>
<a href="http://danluu.com/dunning-kruger/">Nice piece debunking various memes that are supposedly based on studies</a>: &quot;the less someone knows about a subject, the more they think they know&quot; (not actually Dunning-Kruger at all), &quot;money doesn&#x2019;t make people happy&quot; (it seems like it does when making some basic adjustments to the data - I think this point is well known by now), &quot;people bounce back from setbacks (as well as positive events) and return to a fixed level of happiness&quot; (guess not) and &quot;type systems help in programming&quot; (don&apos;t know what this one is about).
</p>
<p>
You may have come across most of these, but here in one place are debunkings (of varying convincingness) of pretty much all of the famous old social psychology experiments that blew my mind when I was in my 20s:
</p>
<ul>

<li><a href="http://journals.sagepub.com/doi/abs/10.1177/0956797618761661?journalCode=pssa&amp;">Disappointing replication of the &quot;marshmallow&quot; experiment</a>.

</li><li>One person argues that<a href="https://www.psychologytoday.com/us/blog/freedom-learn/201310/why-zimbardo-s-prison-experiment-isn-t-in-my-textbook"> the Prison Experiment was a case of subjects behaving as their experimenters clearly wanted them to</a>.

</li><li>The Robbers Cave experiment<a href="https://www.theguardian.com/science/2018/apr/16/a-real-life-lord-of-the-flies-the-troubling-legacy-of-the-robbers-cave-experiment"> backstory sounds particularly dicey</a>: the experimenter tried to get two bands of boys to fight each other ala Lord of the Flies, failed miserably, tried again, succeeded, never mentioned the first time, and gained fame (despite even the 2nd try having the same issue as the Prison Experiment mentioned above).

</li><li>Finally, someone reports that many of the Milgram participants&apos; own reports of why they had administered electric shocks provides more support for &quot;they didn&apos;t think the person was really being harmed&quot; than for Milgram&apos;s theory (<a href="https://digest.bps.org.uk/2017/12/12/interviews-with-milgram-participants-provide-little-support-for-the-contemporary-theory-of-engaged-followership/">link</a>). I think this is the least compelling of critiques listed here, as people will rationalize their behavior with all kinds of stuff.
</li>
</ul>
<p>
    Via an old <a href="https://marginalrevolution.com/marginalrevolution/2018/07/happiness-results-robust.html">Marginal Revolution post</a>, here&apos;s a <a href="https://www.nber.org/system/files/working_papers/w19950/w19950.pdf">study claiming that none of the famous findings about happiness are robust</a>. As far as I can tell, the central claim is basically this general idea: say that Group A consists of two people who each rate their happiness 6/10. And Group B consists of one person rating their happiness 4/10, and another 7/10. In some fundamental sense, we don&apos;t know which group has higher &apos;average&apos; happiness, because for all we know, each increment on the 1-10 scale could represent an extra &apos;10 units&apos; of happiness or an extra &apos;10 times as many units&apos; of happiness, or something else. Now, sometimes we might be able to know which group has higher average happiness despite this issue (for example, a two 7&apos;s and a 6 vs. two 8&apos;s and a 7), but the authors here argue that the famous happiness findings are not robust in this way. Which I think makes sense, though I hope to read this more closely later on.
</p>
<p>
One of the studies I&apos;ve found most mind-blowing (with video evidence!) was <a href="https://twitter.com/BrianRoemmele/status/1213860120058220546">this seeming demonstration that chimpanzees have better working memory than humans</a> (at least for a particular, surprising task). But oops, <a href="https://twitter.com/KrisSabbi/status/1214260415862956042">here is a very believable-sounding debunking</a> claiming that chimpanzees were trained extensively on the task, and similarly trained humans keep up just fine.
</p>
<p>
Finally, here&apos;s a 2008 <a href="https://jamanetwork.com/journals/jamapsychiatry/fullarticle/210112">study</a> questioning the connection between exercise and certain mental health benefits in a convincing-seeming way. &quot;Regular exercise is associated with reduced anxious and depressive symptoms in the population at large, but the association is not because of causal effects of exercise.&quot; The usual result is that exercise does cause such benefits, but this one looked at twin pairs and over-time changes: &quot;Cross-sectional and longitudinal associations were small and were best explained by common genetic factors with opposite effects on exercise behavior and symptoms of anxiety and depression. In genetically identical twin pairs, the twin who exercised more did not display fewer anxious and depressive symptoms than the co-twin who exercised less. Longitudinal analyses showed that increases in exercise participation did not predict decreases in anxious and depressive symptoms.&quot;
</p><!--kg-card-end: html--><!--kg-card-begin: html-->

<p><div style="display:flex; justify-content:center; margin: 0 auto;">

        <span style="margin: 10px;"><a href="https://api.addthis.com/oexchange/0.8/forward/twitter/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fthe-gloves-are-off-the-pants-are-on%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20The%20gloves%20are%20off%2C%20the%20pants%20are%20on&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-twitter-square.png" border="0" alt="The gloves are off, the pants are on"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/facebook/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fthe-gloves-are-off-the-pants-are-on%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20The%20gloves%20are%20off%2C%20the%20pants%20are%20on&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-facebook-square.png" border="0" alt="The gloves are off, the pants are on"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/reddit/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fthe-gloves-are-off-the-pants-are-on%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20The%20gloves%20are%20off%2C%20the%20pants%20are%20on&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-reddit-square.png" border="0" alt="The gloves are off, the pants are on"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/menu/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fthe-gloves-are-off-the-pants-are-on%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20The%20gloves%20are%20off%2C%20the%20pants%20are%20on&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-addthis-square.png" border="0" alt="The gloves are off, the pants are on"></a></span>
        </div>
<center><p id="survey"><a href="https://www.cold-takes.com/the-gloves-are-off-the-pants-are-on//#/portal/signup/free" target="_blank"><button class="button">Subscribe</button></a>&#xA0;<a href="https://www.guidedtrack.com/programs/4kal2ue/run?posttitle=The%20gloves%20are%20off%2C%20the%20pants%20are%20on" target="_blank"><button class="button" id="Survey">Feedback</button></a><!--No Forum post--></p></center>
<!--kg-card-end: html--></p>]]></content:encoded></item><item><title><![CDATA[Are we "trending toward" transformative AI? (How would we know?)]]></title><description><![CDATA[AI progress may not trend in the ways we intutiively expect.]]></description><link>https://www.cold-takes.com/are-we-trending-toward-transformative-ai-how-would-we-know/</link><guid isPermaLink="false">6115eff681b1c6003ee5719a</guid><dc:creator><![CDATA[Holden Karnofsky]]></dc:creator><pubDate>Tue, 24 Aug 2021 17:15:31 GMT</pubDate><media:content url="https://www.cold-takes.com/content/images/2021/08/covid-vs-workplace-closures-2-5.png" medium="image"/><content:encoded><![CDATA[<!--kg-card-begin: html--><!--<p><a href="https://www.cold-takes.com/roadmap-for-the-most-important-century-series/"><img src="https://www.cold-takes.com/content/images/2021/08/mic-diagram---ai-timelines.png"></a></p>-->

<img src="https://www.cold-takes.com/content/images/2021/08/covid-vs-workplace-closures-2-5.png" alt="Are we &quot;trending toward&quot; transformative AI? (How would we know?)"><p><figure><iframe title="Are we &quot;trending toward&quot; transformative AI? (How would we know?)" allowtransparency="true" height="150" width="100%" style="border: none; min-width: min(100%, 430px);" scrolling="no" data-name="pb-iframe-player" src="https://www.podbean.com/player-v2/?i=t68dc-10bec1d-pb&amp;from=pb6admin&amp;share=1&amp;download=1&amp;rtl=0&amp;fonts=Arial&amp;skin=1&amp;font-color=auto&amp;btn-skin=7"></iframe><figcaption><em>Audio also available by searching Stitcher, Spotify, Google Podcasts, etc. for &quot;Cold Takes Audio&quot;</em></figcaption></figure></p>
<!--<p><a href="https://www.cold-takes.com/roadmap-for-the-most-important-century-series/"><img src="https://www.cold-takes.com/content/images/2021/08/mic-diagram---ai-timelines.png"></a></p>-->

<a href="https://www.cold-takes.com/roadmap-for-the-most-important-century-series/"><svg viewbox="0 0 750 178" fill="none" xmlns="http://www.w3.org/2000/svg" id="MasterDiagram">
<g class="Frame 1">
<g id="Today">
<text class="Today&#xE2;&#x20AC;&#x2122;s world" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-weight="bold" letter-spacing="0em"><tspan x="9" y="95.7402">Today&#x2019;s world</tspan></text>
<rect class="Rectangle 1" x="0.5" y="66.5" width="137" height="49" rx="7.5" stroke="var(--darker-gray-color)"/>
</g>
<g id="ArrowToTAI">
<path class="Arrow 7" d="M176.341 86.3655C176.543 86.1771 176.554 85.8607 176.366 85.6588L173.295 82.3691C173.107 82.1672 172.79 82.1563 172.588 82.3447C172.387 82.5331 172.376 82.8495 172.564 83.0514L175.293 85.9756L172.369 88.7049C172.167 88.8933 172.156 89.2097 172.345 89.4116C172.533 89.6135 172.85 89.6244 173.051 89.436L176.341 86.3655ZM146.983 85.4997L175.983 86.4997L176.017 85.5003L147.017 84.5003L146.983 85.4997Z" fill="var(--darker-gray-color)"/>
</g>
<g id="Transformative">
<text class="Transformative AI" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-weight="bold" letter-spacing="0em"><tspan x="192" y="92.6523">Transformative AI</tspan></text>
<rect class="Rectangle 2" x="182.5" y="65.5" width="165" height="47" rx="7.5" stroke="var(--darker-gray-color)"/>
</g>
<g id="ArrowFromTAI" opacity="0.3">
<path class="Arrow 1" d="M389.491 31.0944C389.543 30.8233 389.366 30.5611 389.094 30.509L384.675 29.6592C384.404 29.607 384.142 29.7846 384.09 30.0558C384.038 30.3269 384.215 30.589 384.487 30.6412L388.415 31.3966L387.659 35.3246C387.607 35.5958 387.785 35.8579 388.056 35.91C388.327 35.9622 388.589 35.7846 388.641 35.5135L389.491 31.0944ZM358.28 52.414L389.28 31.414L388.72 30.586L357.72 51.586L358.28 52.414Z" fill="var(--darker-gray-color)"/>
<path class="Arrow 2" d="M385.342 92.3648C385.543 92.1759 385.554 91.8595 385.365 91.658L382.287 88.3751C382.098 88.1737 381.782 88.1634 381.58 88.3523C381.379 88.5412 381.369 88.8576 381.557 89.0591L384.293 91.9772L381.375 94.713C381.174 94.9018 381.163 95.2182 381.352 95.4197C381.541 95.6212 381.858 95.6314 382.059 95.4425L385.342 92.3648ZM353.984 91.4997L384.984 92.4997L385.016 91.5003L354.016 90.5003L353.984 91.4997Z" fill="var(--darker-gray-color)"/>
<path class="Arrow 3" d="M384.955 148.498C385.23 148.523 385.473 148.32 385.498 148.045L385.905 143.564C385.93 143.289 385.728 143.046 385.453 143.021C385.178 142.996 384.934 143.198 384.909 143.473L384.547 147.457L380.564 147.095C380.289 147.07 380.046 147.272 380.021 147.547C379.996 147.822 380.198 148.066 380.473 148.091L384.955 148.498ZM354.616 112.32L384.616 148.32L385.384 147.68L355.384 111.68L354.616 112.32Z" fill="var(--darker-gray-color)"/>
</g>
<g id="DigitalPeople" opacity="0.3">
<text class="Digital people" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-weight="bold" letter-spacing="0em"><tspan x="401" y="36.6523">Digital people</tspan></text>
<text class="World of" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="14" letter-spacing="0em"><tspan x="435.216" y="16.7852">World of</tspan></text>
<rect class="Digital people bubble" x="394.5" y="0.5" width="133" height="45" rx="7.5" stroke="var(--darker-gray-color)"/>
</g>
<g id="MisalignedAndOther" opacity="0.3">
<text class="Misaligned AI" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-weight="bold" letter-spacing="0em"><tspan x="405.436" y="107.652">Misaligned AI</tspan></text>
<text class="World run by" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="14" letter-spacing="0em"><tspan x="422.104" y="87.7852">World run by</tspan></text>
<text class="Something else" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-weight="bold" letter-spacing="0em"><tspan x="398.053" y="164.652">Something else</tspan></text>
<rect class="Rectangle 3" x="395.5" y="68.5" width="132" height="52" rx="7.5" stroke="var(--darker-gray-color)"/>
<rect class="Rectangle 5" x="395.5" y="140.5" width="132" height="37" rx="7.5" stroke="var(--darker-gray-color)"/>
<text class="or" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-style="italic" letter-spacing="0em"><tspan x="455.476" y="61.6523">or</tspan></text>
<text class="or_2" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-style="italic" letter-spacing="0em"><tspan x="455.476" y="133.652">or</tspan></text>
    </g></g>
<g id="SGWC" opacity="0.3">
<text class="Stable, galaxy-wide civilization" fill="var(--darker-gray-color)" xml:space="preserve" style="white-space: pre" font-family="Roboto" font-size="18" font-weight="bold" letter-spacing="0em"><tspan x="583.109" y="89.6523">Stable, galaxy-wide
</tspan><tspan x="618.222" y="110.652">civilization</tspan></text>
<rect class="Rectangle 6" x="575.5" y="69.5" width="174" height="52" rx="7.5" stroke="var(--darker-gray-color)"/></g>
<g id="ArrowToSGWC">
<path opacity="0.3" class="Arrow 4" d="M562 56.5C562.276 56.5 562.5 56.2761 562.5 56L562.5 51.5C562.5 51.2239 562.276 51 562 51C561.724 51 561.5 51.2239 561.5 51.5L561.5 55.5L557.5 55.5C557.224 55.5 557 55.7239 557 56C557 56.2761 557.224 56.5 557.5 56.5L562 56.5ZM534.646 29.3536L561.646 56.3536L562.354 55.6464L535.354 28.6464L534.646 29.3536Z" fill="var(--darker-gray-color)"/>
<path opacity="0.3" class="Arrow 5" d="M561.354 95.3536C561.549 95.1583 561.549 94.8417 561.354 94.6464L558.172 91.4645C557.976 91.2692 557.66 91.2692 557.464 91.4645C557.269 91.6597 557.269 91.9763 557.464 92.1716L560.293 95L557.464 97.8284C557.269 98.0237 557.269 98.3403 557.464 98.5355C557.66 98.7308
                                       557.976 98.7308 558.172 98.5355L561.354 95.3536ZM535 95.5L561 95.5L561 94.5L535 94.5L535 95.5Z" fill="var(--darker-gray-color)"/>
<path opacity="0.3" class="Arrow 6" d="M561.497 117.053C561.526 116.778 561.327 116.532 561.053 116.503L556.578 116.027C556.304 115.998 556.057 116.196 556.028 116.471C555.999 116.746 556.198 116.992 556.472 117.021L560.45 117.444L560.027 121.422C559.998 121.696 560.196 121.943 560.471 121.972C560.746 122.001 560.992 121.802 561.021 121.528L561.497 117.053ZM535.314 138.389L561.314 117.389L560.686 116.611L534.686 137.611L535.314 138.389Z" fill="var(--darker-gray-color)"/>
</g>
    </svg></a>


<p>
<blockquote><p>This is one of 4 posts summarizing hundreds of pages of technical reports focused almost entirely on forecasting one number: the year by which transformative AI will be developed.<sup id="fnref1"><a href="#fn1" rel="footnote">1</a></sup>
</p>
<p>
By &quot;transformative AI,&quot; I mean &quot;AI powerful enough to bring us into a new, qualitatively different future.&quot; I specifically focus on what I&apos;m calling <a href="https://www.cold-takes.com/transformative-ai-timelines-part-1-of-4-what-kind-of-ai/">PASTA</a>: AI systems that can essentially automate all of the human activities needed to speed up scientific and technological advancement.
</p>
<p>
The sooner PASTA might be developed, the sooner the world could change <a href="https://www.cold-takes.com/transformative-ai-timelines-part-1-of-4-what-kind-of-ai/#impacts-of-pasta">radically</a>, and the more important it seems to be thinking today about how to make that change go well vs. poorly.</p></blockquote>
</p>
<p>
In this post and the next, I will talk about the forecasting methods underlying my current view: I believe there&apos;s <strong>more than a 10% chance we&apos;ll see something <a href="https://www.cold-takes.com/transformative-ai-timelines-part-1-of-4-what-kind-of-ai/">PASTA</a>-like enough to qualify as &quot;transformative AI&quot; within 15 years (by 2036); a ~50% chance we&apos;ll see it within 40 years (by 2060); and a ~2/3 chance we&apos;ll see it this century (by 2100).</strong>
</p>
<p>
Below, I will:
</p>
<ul>

<li>Discuss <a href="#what-kind-of-forecast-am-i-going-for">what kind of forecast I&apos;m going for</a>.  
<ul>
 
<li>I&apos;m not sure whether it will feel as though transformative AI is &quot;on the way&quot; long before it arrives. I&apos;m hoping, instead, that we can use trends in key underlying facts about the world (such as AI capabilities, model size, etc.) to forecast a qualitatively unfamiliar future. 
 
</li><li>An analogy for this sort of forecasting would be something like: &quot;This water isn&apos;t bubbling, and there are no signs of bubbling, but the temperature has gone from 70&#xB0; Fahrenheit<sup id="fnref2"><a href="#fn2" rel="footnote">2</a></sup> to 150&#xB0;, and if it hits 212&#xB0;, the water will bubble.&quot; Or: &quot;It&apos;s like forecasting school closures and overbooked hospitals, when there aren&apos;t any yet, based on trends in reported infections.&quot;
</li> 
</ul>

</li><li>Discuss whether we can look for <a href="#subjective-extrapolations-and-" ai-impressiveness"">trends in how &quot;impressive&quot; or &quot;capable&quot; AI systems are</a>. I think this approach is unreliable: (a) AI progress may not &quot;trend&quot; in the way we expect; (b) in my experience, different AI researchers have radically different intuitions about which systems are impressive or capable, and how progress is going. 

</li><li>Briefly discuss <a href="https://arxiv.org/pdf/1705.08807.pdf">Grace et al 2017</a>, the best existing survey of AI researchers on transformative AI timelines. Its conclusions broadly seem in line with my own forecasts, though there are signs the researchers weren&apos;t thinking very hard about the questions.</li></ul>
<p>
The next piece in this series will focus on <a href="https://www.lesswrong.com/posts/KrJfoZzpSDpnrv9va/draft-report-on-ai-timelines">Ajeya Cotra&apos;s &quot;Forecasting Transformative AI with Biological Anchors</a>&quot; (which I&apos;ll abbreviate below as &quot;Bio Anchors&quot;), the forecast I find most informative for transformative AI.
</p>
<h2 id="what-kind-of-forecast-am-i-going-for">What kind of forecast am I going for?</h2>


<p>
There are a couple of ways in which forecasting transformative AI is different from the kind of forecasting we might be used to.
</p>
<p>
First, I&apos;m forecasting over very long time horizons (decades), unlike e.g. a weather forecast (days) or an election forecast (months). This makes the task quite a bit harder,<sup id="fnref3"><a href="#fn3" rel="footnote">3</a></sup> and harder for outsiders to evaluate since I don&apos;t have a clearly relevant <a href="https://www.cold-takes.com/prediction-track-records-i-know-of/">track record</a> of making forecasts on similar topics.
</p>
<p>
Second, I lack rich, clearly relevant data sources, and I can&apos;t look back through a bunch of similar forecasts from the past. FiveThirtyEight&apos;s <a href="https://projects.fivethirtyeight.com/2020-election-forecast/">election</a> forecasts look at hundreds of polls, and they have a model of how well polls have predicted elections in the past. Forecasting transformative AI needs to rely more on intuition, guesswork and judgment, in terms of determining what data is most relevant and how it&apos;s relevant.
</p>
<p>
Finally, I&apos;m trying to forecast a <strong>qualitatively unfamiliar future</strong>. Transformative AI - and the strange future it comes with - doesn&apos;t <em>feel</em> like something we&apos;re &quot;trending toward&quot; year to year.
</p>
<ul>

<li>If I were trying to forecast when the world population would hit 10 billion, I could simply extrapolate <a href="https://ourworldindata.org/world-population-growth#future-population-growth">existing trends</a> of world population. World population itself is known to be growing and can be directly estimated. In my view, extrapolating out a long-running trend is one of the better ways to make a forecast.

</li><li>When FiveThirtyEight makes election forecasts, there&apos;s a background understanding that there&apos;s going to be an election on a certain date, and whoever wins will take office on another date. We all buy into that basic framework, and there&apos;s a general understanding that better polling means a better chance of winning.

</li><li>By contrast, transformative AI - and the strange future it comes with - isn&apos;t something we&apos;re &quot;headed for&quot; in any clearly measurable way. There&apos;s no clear metric like &quot;transformativeness of AI&quot; or &quot;weirdness of the world&quot; that&apos;s going up regularly every year such that we can project it out into the future and get the date that something like <a href="https://www.cold-takes.com/transformative-ai-timelines-part-1-of-4-what-kind-of-ai/">PASTA</a> will be developed. 
</li>
</ul>
<p>
Perhaps for some, these points gives enough reason to ignore the whole possibility of transformative AI, or assume it&apos;s very far away. But I don&apos;t think this is a good idea, for a couple of reasons. 
</p>
<p>
First, I have a background view that something like <a href="https://www.cold-takes.com/transformative-ai-timelines-part-1-of-4-what-kind-of-ai/">PASTA </a>is in a sense &quot;inevitable,&quot; assuming continued advances in society and computing. The basic intuition here - which I could expand on if there&apos;s <a href="https://www.guidedtrack.com/programs/4kal2ue/run?posttitle=Are%20we%20%22trending%20toward%22%20transformative%20AI%3F%20(How%20would%20we%20know%3F)">interest</a> - is that human brains are numerous and don&apos;t seem to need particular rare materials to produce, so it should be possible at some point to synthetically replicate the key parts of their functionality.<sup id="fnref4"><a href="#fn4" rel="footnote">4</a></sup>
</p>
<p>
At the same time, I&apos;m not confident that PASTA will feel qualitatively as though it&apos;s &quot;on the way&quot; well before it arrives. (More on this <a href="#subjective-extrapolations-and-" ai-impressiveness"">below</a>.) So I&apos;m inclined to look for ways to estimate when we can expect this development, despite the challenges, and despite the fact that it doesn&apos;t feel today as though it&apos;s around the corner.
</p>
<p>
I think there are plenty of example cases where a <strong>qualitatively unfamiliar future could be seen in advance by plotting the trend in some underlying, related facts about the world.</strong> A few that come to mind:
</p>
<ul>

<li>When COVID-19 first emerged, a lot of people had trouble taking it seriously because it didn&apos;t feel as though we were &quot;trending toward&quot; or &quot;headed for&quot; a world full of overflowing hospitals, office and school closures, etc. At the time (say, January 2020), there were a relatively small number of cases, an even smaller number of deaths, and no qualitative sense of a global emergency. The only thing alarming about COVID-19, at first, was that case counts were growing at a fast exponential rate (though the overall number of cases was still small). But it was possible to extrapolate from the fast growth in case counts to a risk of a global emergency, and <a href="https://80000hours.org/podcast/episodes/howie-rob-coronavirus-february-3rd/">some people did</a>. (And <a href="https://i.insider.com/5e59596efee23d0fb873eb46?width=750&amp;format=jpeg&amp;auto=webp">some didn&apos;t</a>.)

</li><li>Climatologists forecast a global rise in temperatures that&apos;s significantly more than what we&apos;ve seen over the past few decades, and could have major consequences far beyond what we&apos;re seeing today. They do this by forecasting trends in greenhouse gas emissions and extrapolating <em>from there</em> to temperature and consequences. If you simply tried to ask &quot;How fast is the temperature rising?&quot; or &quot;Are hurricanes getting worse?&quot;, and based all your forecasts of the future on those, you probably wouldn&apos;t be forecasting the same kinds of extreme events around 2100.<sup id="fnref5"><a href="#fn5" rel="footnote">5</a></sup>

</li><li>To give a more long-run example, we can project a date by which the sun will burn out, and conclude that the world will look very different by that date than it does now, even though there&apos;s no trend of things getting colder or darker today.</li></ul>

<figure class="kg-card kg-image-card kg-width-wide"><img src="https://www.cold-takes.com/content/images/2021/08/covid-vs-workplace-closures-3.png" width="1036" alt="Are we &quot;trending toward&quot; transformative AI? (How would we know?)"><figcaption>COVID-19 cases from <a href="https://portal.who.int/report/eios-covid19-counts/#display=Global&amp;nrow=1&amp;ncol=1&amp;arr=row&amp;pg=1&amp;labels=view_who_regions,view_continents&amp;sort=global_code;asc&amp;filter=&amp;sidebar=-1&amp;fv=">WHO</a>. Workplace closures are from <a href="https://ourworldindata.org/grapher/workplace-closures-covid">this OWiD data</a>, simply scored as 1 for &quot;recommended,&quot; 2 for &quot;required for some,&quot; 3 for &quot;required for all but key workers&quot; and summed across all countries.</figcaption></figure>
<p></p>
<p>
An analogy for this sort of forecasting would be something like: &quot;This water isn&apos;t bubbling, and there are no signs of bubbling, but the temperature has gone from 70&#xB0; Fahrenheit<sup id="fnref6"><a href="#fn6" rel="footnote">6</a></sup> to 150&#xB0;, and if it hits 212&#xB0;, the water will bubble.&quot; 
</p>
<p>
Ideally, I can find some underlying factors that are changing regularly enough for us to predict them (such as growth in the <a href="https://openai.com/blog/ai-and-compute/">size and cost of AI models</a>), and then argue that if those factors reach a certain point, the odds of transformative AI will be high. 
</p>
<p>
You can think of this approach as answering the question: &quot;If I think something like PASTA is inevitable, and I&apos;m trying to guess the timing of it using a few different analysis methods, what do I guess?&quot; We can separately ask &quot;And is there reason that this guess is implausible, untrustworthy, or too &apos;wild?&apos;&quot; - this was addressed in the <a href="https://www.cold-takes.com/forecasting-transformative-ai-whats-the-burden-of-proof/">previous piece in this series</a>.
</p>
<h2 id="subjective-extrapolations-and-" ai-impressiveness"">Subjective extrapolations and &quot;AI impressiveness&quot;</h2>


<p>
<em>For a different presentation of some similar content, see <a href="https://docs.google.com/document/d/1cCJjzZaJ7ATbq8N2fvhmsDOUWdm7t3uSSXv6bD0E_GM/edit#heading=h.njuz93bimqty">this section</a> of <a href="https://drive.google.com/drive/u/1/folders/15ArhEPZSTYU8f012bs6ehPS6-xmhtBPP">Bio Anchors</a>.</em>
</p>
<p>
If we&apos;re looking for some underlying factors in the world that predict when transformative AI is coming, perhaps the first thing we should look for is trends in how &quot;impressive&quot; or &quot;capable&quot; AI systems are.
</p>
<p>
The easiest version of this would be if the world happened to shake out such that:
</p>
<ul>

<li>One day, for the first time, an AI system managed to get a passing grade on a 4th-grade science exam.

</li><li>Then we saw the first AI passing (and then acing) a 5th grade exam, then 6th grade exam, etc.

</li><li>Then we saw the first AI earning a PhD, then the first AI writing a published paper, etc. all the way up to the first AI that could do Nobel-Prize-worthy science work.

</li><li>This all was spread out regularly over the decades, so we could clearly see the state of the art advancing from 4th grade to 5th grade to 6th grade, all the way up to &quot;postdoc&quot; and beyond. And all of this happened slowly and regularly enough that we could start putting a date on &quot;full-blown scientist AI&quot; several decades in advance.
</li>
</ul>
<p>
It would be very convenient - I almost want to say &quot;polite&quot; - of AI systems to advance in this manner. It would also be &quot;polite&quot; if AI advanced in the way that some people seem to casually imagine it will: first taking over jobs like &quot;truck driver&quot; and &quot;assembly line worker,&quot; then jobs like &quot;teacher&quot; and &quot;IT support,&quot; and then jobs like &quot;doctor&quot; and &quot;lawyer,&quot; before progressing to &quot;scientist.&quot; 
</p>
<p>
Either of these would give us plenty of lead time and a solid basis to project when science-automating AI is coming. Unfortunately, I don&apos;t think we can count on such a thing. 
</p>
<ul>

<li>AI seems to progress very differently from humans. For example, there were superhuman AI chess players<sup id="fnref7"><a href="#fn7" rel="footnote">7</a></sup> long before there was AI that could reliably tell apart pictures of dogs and cats.<sup id="fnref8"><a href="#fn8" rel="footnote">8</a></sup> 

</li><li>One possibility is that AI systems will be capable of the hardest intellectual tasks insects can do, then of the hardest tasks mice and other small mammals can do, then monkeys, then humans - effectively matching the abilities of larger and larger brains. If this happened, we wouldn&apos;t necessarily see many signs of AI being able to e.g. do science until we were <em>very</em> close. Matching a 4th-grader might not happen until the very end.

</li><li>Another possibility is that AI systems will be able to do anything that a human can do within 1 second, then anything that a human can do within 10 seconds, etc. This could also be quite a confusing progression that makes it non-obvious how to forecast progress.</li></ul>
<p>
Actually, if we didn&apos;t already know how humans tend to mature, we might find a child&apos;s progress to be pretty confusing and hard to extrapolate. <strong>Watching someone progress from birth to age 8 wouldn&apos;t necessarily give you any idea that they were, say, 1/3 of the way to being able to start a business, make an important original scientific discovery, etc.</strong>  (Even <em>knowing </em>the usual course of human development, it&apos;s hard to tell from observing an 8-year-old what professional-level capabilities they could/will end up with in adulthood.)
</p>
<p>
Overall, it&apos;s quite unclear how we should think about the spectrum from &quot;not impressive/capable&quot; to &quot;very impressive/capable&quot; for AI. And indeed, in my experience, different AI researchers have radically different intuitions about which systems are impressive or capable, and how progress is going. I&apos;ve often had the experience of seeing one AI researcher friend point to some new result and say &quot;This is huge, how can anyone not see how close we&apos;re getting to powerful AI?&quot; while another says &quot;This is a minor advance with little significance.&quot;<sup id="fnref9"><a href="#fn9" rel="footnote">9</a></sup>
</p>
<p>
It would be great if we could forecast the year transformative AI will be developed, by using a chart like this (from <a href="https://drive.google.com/drive/u/1/folders/15ArhEPZSTYU8f012bs6ehPS6-xmhtBPP">Bio Anchors</a>; &quot;TAI&quot; means &quot;transformative AI&quot;):
</p>
<p>


<figure class="kg-card kg-image-card kg-width-wide"><img src="https://www.cold-takes.com/content/images/size/w1000/2021/08/impressiveness-extrap-chart.png" alt="Are we &quot;trending toward&quot; transformative AI? (How would we know?)" width="1036"></figure>
</p>

<p>
But as far as I can tell, there&apos;s no way to define the y-axis that wouldn&apos;t be fiercely debated between experts.
</p>
<h2 id="surveying-experts">Surveying experts</h2>


<p>
One way to deal with this uncertainty and confusion would be to survey a large number of experts and simply ask them when they expect transformative AI to be developed. We might hope that each of the experts (or at least, many of them) is doing their own version of the &quot;impressiveness extrapolation&quot; above - or if not, that they&apos;re doing something else that can help them get a reasonable estimate. By averaging many estimates, we might get an aggregate that reflects the &quot;wisdom of crowds.&quot;<sup id="fnref10"><a href="#fn10" rel="footnote">10</a></sup>
</p>
<p>
I think the best version of this exercise is <a href="https://arxiv.org/pdf/1705.08807.pdf">Grace et al 2017</a>, a survey of 352 AI researchers that included a question about &#x201C;when unaided machines can accomplish every task better and more cheaply than human workers&quot; (which would presumably include tasks that advance scientific and technological development, and hence would qualify as <a href="https://www.cold-takes.com/transformative-ai-timelines-part-1-of-4-what-kind-of-ai/">PASTA</a>). The two big takeaways from this survey, according to <a href="https://drive.google.com/drive/u/1/folders/15ArhEPZSTYU8f012bs6ehPS6-xmhtBPP">Bio Anchors</a> and me, are:
</p>
<ul>

<li><strong>A ~20% probability of this sort of AI by 2036; a ~50% probability by 2060; a ~70% probability by 2100. These match the figures I give in the introduction.</strong>

</li><li>Much later estimates for slightly differently phrased questions (posed to a smaller subset of respondents), implying (to me) that the researchers simply weren&apos;t thinking very hard about the questions.<sup id="fnref11"><a href="#fn11" rel="footnote">11</a></sup>
</li>
</ul>
<p>
My bottom line: this evidence is consistent with my current probabilities, though potentially not very informative. The next piece in this series will be entirely focused on <a href="https://www.lesswrong.com/posts/KrJfoZzpSDpnrv9va/draft-report-on-ai-timelines">Ajeya Cotra&apos;s &quot;Forecasting Transformative AI with Biological Anchors,&quot;</a> the forecasting method I find most informative here.
</p>

<p><strong>Next in series:</strong> <a href="https://www.cold-takes.com/forecasting-transformative-ai-the-biological-anchors-method-in-a-nutshell/">Forecasting transformative AI: the &quot;biological anchors&quot; method in a nutshell</a></p>

<!--kg-card-end: html--><!--kg-card-begin: html--><hr>
<h2>Footnotes</h2>
<div class="footnotes">

<ol><li id="fn1">
<p>
     Of course, the answer could be &quot;A kajillion years from now&quot; or &quot;Never.&quot;&#xA0;<a href="#fnref1" rev="footnote">&#x21A9;</a><li id="fn2">

<p>
     Centigrade equivalents for this sentence: 21&#xB0;, 66&#xB0;, 100&#xB0;&#xA0;<a href="#fnref2" rev="footnote">&#x21A9;</a><li id="fn3">
<p>
     Some notes on longer-term forecasting <a href="https://www.openphilanthropy.org/blog/how-feasible-long-range-forecasting#Tetlock_long-range_forecasting_and_questions_of_relevance">here</a>.&#xA0;<a href="#fnref3" rev="footnote">&#x21A9;</a><li id="fn4">
<p>
     See also <a href="https://www.lesswrong.com/posts/HhWhaSzQr6xmBki8F/birds-brains-planes-and-ai-against-appeals-to-the-complexity">this piece</a> for a bit of a more fleshed out argument along these lines, which I don&apos;t agree with fully as stated (I don&apos;t think it presents a strong case for transformative AI soon), but which I think gives a good sense of my intuitions about in-principle feasibility. Also see <a href="https://arxiv.org/abs/1703.10987">On the Impossibility of Supersized Machines</a> for some implicit (joking) responses to many common arguments for why transformative AI might be impossible to create. &#xA0;<a href="#fnref4" rev="footnote">&#x21A9;</a><li id="fn5">

<p>
     For example, see the temperature chart <a href="https://www.climate.gov/news-features/understanding-climate/climate-change-global-temperature-projections#:~:text=Results%20from%20a%20wide%20range,gases%20that%20human%20activities%20produce.">here</a> - the lowest line seems like it would be a reasonable projection, if temperature were the only thing you were looking at.&#xA0;<a href="#fnref5" rev="footnote">&#x21A9;</a><li id="fn6">
<p>
     Centigrade equivalents for this sentence: 21&#xB0;, 66&#xB0;, 100&#xB0;&#xA0;<a href="#fnref6" rev="footnote">&#x21A9;</a><li id="fn7">

<p>
     <a href="https://en.wikipedia.org/wiki/Deep_Blue_(chess_computer)#Deep_Blue_versus_Kasparov">1997</a>.&#xA0;<a href="#fnref7" rev="footnote">&#x21A9;</a><li id="fn8">
<p>
     The Kaggle &quot;dogs vs. cats&quot; challenge was <a href="https://www.kaggle.com/c/dogs-vs-cats/leaderboard">created in 2013</a>.&#xA0;<a href="#fnref8" rev="footnote">&#x21A9;</a><li id="fn9">
<p>
     From <a href="https://drive.google.com/drive/u/1/folders/15ArhEPZSTYU8f012bs6ehPS6-xmhtBPP">Bio Anchors</a>: &quot;We have heard ML experts with relatively short timelines argue that AI systems today can essentially see as well as humans, understand written information, and beat humans at almost all strategy games, and the set of things they can do is expanding rapidly, leading them to expect that transformative AI would be attainable in the next decade or two by training larger models on a broader distribution of ML problems that are more targeted at generating economic value. Conversely, we have heard ML experts with relatively long timelines argue that ML systems require much more data to learn than humans do, are unable to transfer what they learn in one context to a slightly different context, and don&#x2019;t seem capable of much structured logical and causal reasoning; this leads them to believe we would need to make multiple major breakthroughs to develop TAI. At least one Open Philanthropy technical advisor has advanced each of these perspectives.&quot;&#xA0;<a href="#fnref9" rev="footnote">&#x21A9;</a><li id="fn10">
<p>
     <a href="https://en.wikipedia.org/wiki/Wisdom_of_the_crowd">Wikipedia</a>: &quot;The classic wisdom-of-the-crowds finding ... At a 1906 country fair in Plymouth, 800 people participated in a contest to estimate the weight of a slaughtered and dressed ox. Statistician Francis Galton observed that the median guess, 1207 pounds, was accurate within 1% of the true weight of 1198 pounds.&quot;&#xA0;<a href="#fnref10" rev="footnote">&#x21A9;</a><li id="fn11">

<p>
     <a href="https://drive.google.com/drive/u/1/folders/15ArhEPZSTYU8f012bs6ehPS6-xmhtBPP">Bio Anchors</a>: 
<ul>

<li><em>Some researchers were asked to forecast &#x201C;HLMI&#x201D; as defined above [human-level machine intelligence, which I would take to include something like PASTA], while a randomly-selected subset was instead asked to forecast &#x201C;full automation of labor&#x201D;, the time when &#x201C;all occupations are fully automatable.&#x201D; Despite the fact that achieving HLMI seems like it should quickly lead to full automation of labor, the median estimate for full automation of labor was ~2138 while the median estimate for HLMI was ~2061, almost 80 years earlier. </em>

</li><li><em>Random subsets of respondents were asked to forecast when individual milestones (e.g. laundry folding, human-level StarCraft, or human-level math research) would be achieved. The median year by which respondents expected machines to be able to automate AI research was ~2104, while the median estimate for HLMI was ~2061 -- another clear inconsistency because &#x201C;AI research&#x201D; is a task done by human workers.</em>&#xA0;<a href="#fnref11" rev="footnote">&#x21A9;</a></li></ul>

</p></li></p></li></p></li></p></li></p></li></p></li></p></li></p></li></p></li></p></li></p></li></ol></div><!--kg-card-end: html--><!--kg-card-begin: html-->

<p><div style="display:flex; justify-content:center; margin: 0 auto;">

        <span style="margin: 10px;"><a href="https://api.addthis.com/oexchange/0.8/forward/twitter/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fare-we-trending-toward-transformative-ai-how-would-we-know%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20Are%20we%20%22trending%20toward%22%20transformative%20AI%3F%20(How%20would%20we%20know%3F)&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-twitter-square.png" border="0" alt="Are we &quot;trending toward&quot; transformative AI? (How would we know?)"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/facebook/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fare-we-trending-toward-transformative-ai-how-would-we-know%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20Are%20we%20%22trending%20toward%22%20transformative%20AI%3F%20(How%20would%20we%20know%3F)&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-facebook-square.png" border="0" alt="Are we &quot;trending toward&quot; transformative AI? (How would we know?)"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/reddit/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fare-we-trending-toward-transformative-ai-how-would-we-know%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20Are%20we%20%22trending%20toward%22%20transformative%20AI%3F%20(How%20would%20we%20know%3F)&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-reddit-square.png" border="0" alt="Are we &quot;trending toward&quot; transformative AI? (How would we know?)"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/menu/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fare-we-trending-toward-transformative-ai-how-would-we-know%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20Are%20we%20%22trending%20toward%22%20transformative%20AI%3F%20(How%20would%20we%20know%3F)&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-addthis-square.png" border="0" alt="Are we &quot;trending toward&quot; transformative AI? (How would we know?)"></a></span>
        </div>
<center><p id="survey"><a href="https://www.cold-takes.com/are-we-trending-toward-transformative-ai-how-would-we-know/#/portal/signup/free" target="_blank"><button class="button">Subscribe</button></a>&#xA0;<a href="https://www.guidedtrack.com/programs/4kal2ue/run?posttitle=Are%20we%20%22trending%20toward%22%20transformative%20AI%3F%20(How%20would%20we%20know%3F)" target="_blank"><button class="button" id="Survey">Feedback</button></a>&#xA0;<a href="https://forum.effectivealtruism.org/posts/SueEfksxDQm4Q97oK/are-we-trending-toward-transformative-ai-how-would-we-know#comments" target="_blank"><button class="button">Forum</button></a></p><p><em>
Use &quot;Feedback&quot; if you have comments/suggestions you want me to see, or if you&apos;re up for giving some quick feedback about this post (which I greatly appreciate!) Use &quot;Forum&quot; if you want to discuss this post publicly on the Effective Altruism Forum.
</em></p></center>
<!--kg-card-end: html--></p>]]></content:encoded></item><item><title><![CDATA[Cold Links: heartwarming sports stuff]]></title><description><![CDATA[Apparently Marcus Smart is really good at "taking charges," which means positioning himself so that an offensive player will run into him and foul him.]]></description><link>https://www.cold-takes.com/cold-links-heartwarming-sports-stuff/</link><guid isPermaLink="false">6116068781b1c6003ee571e7</guid><dc:creator><![CDATA[Holden Karnofsky]]></dc:creator><pubDate>Thu, 19 Aug 2021 17:44:51 GMT</pubDate><media:content url="https://www.cold-takes.com/content/images/2021/08/smart-takes-charge-1.gif" medium="image"/><content:encoded><![CDATA[<!--kg-card-begin: html--><img src="https://www.cold-takes.com/content/images/2021/08/smart-takes-charge-1.gif" alt="Cold Links: heartwarming sports stuff"><p>
Apparently Marcus Smart is <a href="https://www.espn.com/blog/boston/celtics/post/_/id/4725417/marcus-smart-taking-charge-for-boston-celtics">really good at &quot;taking charges,&quot;</a> which means positioning himself so that an offensive player will run into him and foul him. This requires getting into the right position and then specifically not moving for a second (the latter is necessary to trigger the ref&apos;s classification as an offensive foul). It&apos;s very funny to watch him do this (he&apos;s #36, the guy who takes a crotch to the face):
</p>
<p>

<img src="https://www.cold-takes.com/content/images/2021/08/smart-takes-charge.gif" alt="Cold Links: heartwarming sports stuff">

</p>
<p>
Marcus Smart has become my new central example of someone who enthusiastically takes one for the team.
</p>
<p>
A friend summarizes something he saw on ESPN (see also the <a href="https://en.wikipedia.org/wiki/1994_John_Tyler_vs._Plano_East_high_school_football_game">Wikipedia entry</a>, and <a href="https://www.youtube.com/watch?v=ZHkABO0VwCg&amp;ab_channel=jodyvancevids">video</a>):
</p>
<p>

    <blockquote>this is an amazing sports story i never knew about. Texas HS football. It was on ESPN b/c it&apos;s the 20th anniversary.
</blockquote></p>
<p>

    &quot;... giving the Lions a seemingly insurmountable 41&#x2013;17 lead with only 3:03 remaining.
</p>
<p>

    However, on a two-play 70-yard drive, the Panthers scored a touchdown to bring the score to 41&#x2013;23 (after a failed two-point conversion) with 2:36 on the clock. The Panthers then successfully executed three onside kicks in a row, recovering the ball each time and then driving down the field for a touchdown on each occasion.
</p>
<p>

    ... giving the Panthers a 44&#x2013;41 comeback lead with only 24 seconds remaining.
</p>
<p>

    In a final twist, however, after the Panthers did a regular kickoff, the Lions&apos; returner Roderick Dunn caught the ball at his own three-yard line and took it 97 yards for a touchdown at 0:11 and a 48&#x2013;44 Lions victory.
</p>
<p>

    He was the very same player who had muffed the reception of the final two onside kicks.&apos;
</p>
<p>

    ... interviews with the players from today [were on ESPN, not the Wikipedia page]:
</p>
<p>

    -- the guys from the team that lost were still crying about it
</p>
<p>

    -- the guy that ran back the kick said it was one of the greatest moments in his life and he still thinks about when he&apos;s down. the lesson, he says, is &apos;never give up.&apos;
</p>
<p>
I wasn&apos;t able to easily verify a lot of this, but here is a <a href="https://www.mansworldindia.com/sports/story-of-nav-bhatia-nba-hall-of-fame/">very short, sweet story about Nav Bhatia</a>, perhaps the first person inducted into the NBA Hall of Fame and given a championship ring for ... being a really dedicated fan? Apparently he hasn&apos;t missed a Toronto Raptors home game in 25 years. 
</p>
<p>
Watch this kid&apos;s <a href="https://i.imgur.com/GSt0HWL.gifv">reaction to getting a spare racket from tennis legend Novak Djokovich</a>. Sports!
</p>
<p>
<a href="https://deadspin.com/super-nice-soccer-guy-rewarded-for-his-compassion-with-1830382771">Super Nice Soccer Guy Rewarded For His Compassion With Easiest Goal Of His Life</a>.
</p>
<p>
<a href="http://deadspin.com/andrew-luck-is-an-affable-weirdo-1671351814">Very funny article on now-retired former star NFL quarterback Andrew Luck</a>: apparently he sincerely congratulated people who tackle him hard, and this was completely unique for a quarterback and was seen as extremely unnerving by the defenders. Doesn&apos;t mean he enjoyed the tackles though - he <a href="https://en.wikipedia.org/wiki/Andrew_Luck#Retirement">retired at 29</a>.
</p>
<p>
The article completely delivers on the headline:<a href="https://www.npr.org/sections/thetwo-way/2018/03/30/598263399/36-year-old-accountant-called-in-as-emergency-nhl-goalie-and-he-crushed-it?utm_source=facebook.com&amp;utm_medium=social&amp;utm_campaign=npr&amp;utm_term=nprnews&amp;utm_content=20180330"> 36-Year-Old Accountant Called In As Emergency NHL Goalie &#x2014; And He Crushed It</a>.
</p>
<p>
That <a href="https://www.aljazeera.com/sports/2021/8/2/best-moment-of-tokyo-2020-world-reacts-to-shared-olympic-gold">story about two athletes (friends) who shared the gold medal</a> is too recent and too popular for Cold Links, so maybe I&apos;ll link to it in a year or two. I did manage to find someone <a href="https://www.foxsports.com.au/tokyo-olympics-2021/tokyo-olympics-2021-high-jump-final-shared-gold-medal-mutaz-essa-barshim-and-gianmarco-tamberi-win-explained-rule/news-story/e5917258d01197da640822c55b2f420a">complaining about it</a>.
</p>

<p>Readers sent in cool new links on &quot;intensity&quot; in sports, which I&apos;ll put out another time (only heartwarming stuff allowed in this one).</p><!--kg-card-end: html--><!--kg-card-begin: html-->

<p><div style="display:flex; justify-content:center; margin: 0 auto;">

        <span style="margin: 10px;"><a href="https://api.addthis.com/oexchange/0.8/forward/twitter/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fcold-links-heartwarming-sports-stuff%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20Cold%20Links%3A%20heartwarming%20sports%20stuff&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-twitter-square.png" border="0" alt="Cold Links: heartwarming sports stuff"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/facebook/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fcold-links-heartwarming-sports-stuff%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20Cold%20Links%3A%20heartwarming%20sports%20stuff&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-facebook-square.png" border="0" alt="Cold Links: heartwarming sports stuff"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/reddit/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fcold-links-heartwarming-sports-stuff%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20Cold%20Links%3A%20heartwarming%20sports%20stuff&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-reddit-square.png" border="0" alt="Cold Links: heartwarming sports stuff"></a></span><span style="margin: 10px;">
        <a href="https://api.addthis.com/oexchange/0.8/forward/menu/offer?url=https%3A%2F%2Fwww.cold-takes.com%2Fcold-links-heartwarming-sports-stuff%2F&amp;pubid=ra-60a178324cffc42e&amp;title=Cold%20Takes%20-%20Cold%20Links%3A%20heartwarming%20sports%20stuff&amp;ct=1" target="_blank"><img width="32" src="https://www.cold-takes.com/content/images/2021/06/ct-addthis-square.png" border="0" alt="Cold Links: heartwarming sports stuff"></a></span>
        </div>
<center><p id="survey"><a href="https://www.cold-takes.com/cold-links-heartwarming-sports-stuff/#/portal/signup/free" target="_blank"><button class="button">Subscribe</button></a>&#xA0;<a href="https://www.guidedtrack.com/programs/4kal2ue/run?posttitle=Cold%20Links%3A%20heartwarming%20sports%20stuff" target="_blank"><button class="button" id="Survey">Feedback</button></a><!--No Forum post--></p></center>
<!--kg-card-end: html--></p>]]></content:encoded></item></channel></rss>