{"id":11866,"date":"2025-07-21T22:22:08","date_gmt":"2025-07-21T22:22:08","guid":{"rendered":"https:\/\/www.newsbeep.com\/au\/11866\/"},"modified":"2025-07-21T22:22:08","modified_gmt":"2025-07-21T22:22:08","slug":"human-level-ai-is-not-inevitable-we-have-the-power-to-change-course-garrison-lovely","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/au\/11866\/","title":{"rendered":"Human-level AI is not inevitable. We have the power to change course | Garrison Lovely"},"content":{"rendered":"<p> Illustration: Petra P\u00e9terffy\/The Guardian<\/p>\n<p class=\"dcr-16w5gq9\">\u201cTechnology happens because it is possible,\u201d OpenAI CEO, Sam Altman, <a href=\"https:\/\/www.nytimes.com\/2023\/03\/31\/technology\/sam-altman-open-ai-chatgpt.html#:~:text=Technology%20happens%20because%20it%20is%20possible%2C%E2%80%9D\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">told<\/a> the New York Times in 2019, consciously paraphrasing Robert Oppenheimer, the father of the atomic bomb.<\/p>\n<p class=\"dcr-16w5gq9\">Altman captures a Silicon Valley mantra: technology marches forward inexorably.<\/p>\n<p class=\"dcr-16w5gq9\">Another widespread techie conviction is that the first human-level AI \u2013 also known as artificial general intelligence (AGI) \u2013 will lead to one of two futures: a post-scarcity <a href=\"https:\/\/waitbutwhy.com\/2015\/01\/artificial-intelligence-revolution-2.html\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">techno-utopia<\/a> or the <a href=\"https:\/\/www.cold-takes.com\/ai-could-defeat-all-of-us-combined\/#the-standard-argument-superintelligence-and-advanced-technology\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">annihilation of humanity<\/a>.<\/p>\n<p class=\"dcr-16w5gq9\">For <a href=\"https:\/\/www.theguardian.com\/environment\/2017\/jul\/10\/earths-sixth-mass-extinction-event-already-underway-scientists-warn\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">countless other species<\/a>, the arrival of humans <a href=\"https:\/\/www.likevillepodcast.com\/articles\/2021\/1\/25\/what-happened-to-the-megafauna-a-selection-from-joseph-henrichs-the-secret-of-our-success-2017\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">spelled<\/a> doom. We weren\u2019t tougher, faster or stronger \u2013 just smarter and better coordinated. In many cases, extinction was an accidental byproduct of some other goal we had. A true AGI would amount to creating a <a href=\"https:\/\/yoshuabengio.org\/2023\/06\/24\/faq-on-catastrophic-ai-risks\/#:~:text=If%20we%20design,we%20drove%20extinct.\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">new species<\/a>, which might quickly <a href=\"https:\/\/arxiv.org\/pdf\/2310.17688.pdf#page=2\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">outsmart<\/a> or <a href=\"https:\/\/www.planned-obsolescence.org\/continuous-doesnt-mean-slow\/\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">outnumber<\/a> us. It could see humanity as a minor obstacle, like an <a href=\"https:\/\/www.vox.com\/future-perfect\/2018\/10\/16\/17978596\/stephen-hawking-ai-climate-change-robots-future-universe-earth#:~:text=%E2%80%9CYou%E2%80%99re%20probably%20not%20an%20evil%20ant%2Dhater%20who%20steps%20on%20ants%20out%20of%20malice%2C%20but%20if%20you%E2%80%99re%20in%20charge%20of%20a%20hydroelectric%20green%2Denergy%20project%20and%20there%E2%80%99s%20an%20anthill%20in%20the%20region%20to%20be%20flooded%2C%20too%20bad%20for%20the%20ants.%20Let%E2%80%99s%20not%20place%20humanity%20in%20the%20position%20of%20those%20ants%2C%E2%80%9D%20Hawking%20writes.\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">anthill<\/a> in the way of a planned hydroelectric dam, or a <a href=\"https:\/\/www.vox.com\/the-highlight\/23777171\/ai-animals-rights-cruelty-transhumanism-bostrom\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">resource to exploit<\/a>, like the billions of animals confined in factory farms.<\/p>\n<p class=\"dcr-16w5gq9\">Altman, along with the heads of the other top AI labs, believes that AI-driven extinction is a <a href=\"https:\/\/www.safe.ai\/work\/statement-on-ai-risk\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">real possibility<\/a> (joining hundreds of leading AI researchers and prominent figures).<\/p>\n<p class=\"dcr-16w5gq9\">Given all this, it\u2019s natural to ask: should we really try to build a technology that may kill us all if it goes wrong?<\/p>\n<p class=\"dcr-16w5gq9\">Perhaps the most common reply says: AGI is inevitable. It\u2019s just too useful not to build. After all, AGI would be the ultimate technology \u2013 what a colleague of Alan Turing <a href=\"https:\/\/en.wikipedia.org\/wiki\/I._J._Good#:~:text=Thus%20the%20first%20ultraintelligent%20machine,to%20keep%20it%20under%20control.\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">called<\/a> \u201cthe last invention that man need ever make\u201d. Besides, the reasoning goes within AI labs, if we don\u2019t, someone else will do it \u2013 less responsibly, of course.<\/p>\n<p class=\"dcr-16w5gq9\">A new ideology out of Silicon Valley, <a href=\"https:\/\/en.wikipedia.org\/wiki\/Effective_accelerationism\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">effective accelerationism<\/a> (e\/acc), <a href=\"https:\/\/effectiveacceleration.tech\/\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">claims<\/a> that AGI\u2019s inevitability is a consequence of the second law of thermodynamics and that its engine is \u201ctechnocapital\u201d. The e\/acc <a href=\"https:\/\/effectiveacceleration.tech\/\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">manifesto<\/a> asserts: \u201cThis engine cannot be stopped. The ratchet of progress only ever turns in one direction. Going back is not an option.\u201d<\/p>\n<p class=\"dcr-16w5gq9\">For <a href=\"https:\/\/twitter.com\/sama\/status\/1540227243368058880?lang=en\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">Altman<\/a> and e\/accs, technology takes on a mystical quality \u2013 the march of invention is treated as a fact of nature. But it\u2019s not. Technology is the product of deliberate human choices, motivated by myriad powerful forces. We have the agency to shape those forces, and history shows that we\u2019ve done it before.<\/p>\n<p class=\"dcr-16w5gq9\">No technology is inevitable, not even something as tempting as AGI.<\/p>\n<p class=\"dcr-16w5gq9\">Some AI worriers like to <a href=\"https:\/\/worldspiritsockpuppet.substack.com\/i\/91988109\/extremely-valuable-technologies\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">point out<\/a> <a href=\"https:\/\/verfassungsblog.de\/paths-untaken\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">the times<\/a> humanity resisted and restrained valuable technologies.<\/p>\n<p class=\"dcr-16w5gq9\">Fearing novel risks, biologists initially <a href=\"https:\/\/www.nature.com\/articles\/455290a\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">banned<\/a> and then successfully <a href=\"https:\/\/intelligence.org\/files\/TheAsilomarConference.pdf#page=5\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">regulated<\/a> experiments on recombinant DNA in the 1970s.<\/p>\n<p class=\"dcr-16w5gq9\">No human has been reproduced via <a href=\"https:\/\/www.genome.gov\/about-genomics\/fact-sheets\/Cloning-Fact-Sheet#:~:text=Despite%20several%20highly,had%20been%20retracted.\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">cloning<\/a>, even though it\u2019s been <a href=\"https:\/\/www.statnews.com\/2020\/02\/21\/human-reproductive-cloning-curious-incident-of-the-dog-in-the-night-time\/#:~:text=But%20in%20January,never%20banned%20it.\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">technically possible<\/a> for over a decade, and the only scientist to genetically engineer humans was <a href=\"https:\/\/www.theguardian.com\/world\/2019\/dec\/30\/gene-editing-chinese-scientist-he-jiankui-jailed-three-years\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">imprisoned<\/a> for his efforts.<\/p>\n<p class=\"dcr-16w5gq9\">Nuclear power <a href=\"https:\/\/ourworldindata.org\/nuclear-energy#safety-of-nuclear-energy\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">can provide<\/a> consistent, carbon-free energy, but vivid fears of catastrophe have <a href=\"https:\/\/wiki.aiimpacts.org\/responses_to_ai\/technological_inevitability\/incentivized_technologies_not_pursued\/nuclear_power\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">motivated<\/a> stifling regulations and <a href=\"https:\/\/www.politico.eu\/article\/nuclear-reactors-germany-invest-gas-power-plants-energy\/#:~:text=Germany%20shut%20down%20its%20final%20three%20nuclear%20reactors%20last%20April%2C%20despite%20warnings%20that%20it%20would%20cause%20more%20fossil%20fuel%20to%20be%20burned.%20Last%20year%2C%20a%20report%20from%20Berlin%27s%20own%20climate%20agency%20said%20the%20country%20was%20likely%20to%20miss%20its%20target%20of%20cutting%20greenhouse%20emissions%20by%2065%20percent%20by%202030.\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">outright bans<\/a>.<\/p>\n<p class=\"dcr-16w5gq9\">And if Altman were <a href=\"https:\/\/dgardner.substack.com\/p\/technology-is-not-inevitable\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">more familiar<\/a> with the history of the Manhattan Project, he might realize that the creation of nuclear weapons in 1945 <a href=\"https:\/\/ethos.lps.library.cmu.edu\/article\/id\/35\/#:~:text=The%20development%20of%20nuclear,a%20program%20with%20vigor.\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">was actually<\/a> a highly contingent and unlikely outcome, <a href=\"https:\/\/ethos.lps.library.cmu.edu\/article\/id\/35\/#:~:text=The%20recommendation%20was%20based%20on%20technical%20promise%2C%20but%20also%20on%20the%20strong%20and%2C%20at%20the%20time%20authentic%2C%20belief%20that%20the%20Germans%20could%20be%20even%20farther%20ahead%20at%20that%20point%20and%20that%20they%20were%20in%20a%20genuine%20%E2%80%9Crace%E2%80%9D%20for%20the%20bomb\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">motivated<\/a> by a mistaken belief that the Germans were ahead in a \u201crace\u201d for the bomb. Philip Zelikow, the historian who led the 9\/11 Commission, <a href=\"https:\/\/youtu.be\/IPaofrAiP5s?si=_2a-zZUKjPibrez_&amp;t=456\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">said<\/a>: \u201cI think had the United States not built an atomic bomb during the Second World War, it\u2019s actually not clear to me when or possibly even if an atomic bomb ever is built.\u201d<\/p>\n<p class=\"dcr-16w5gq9\">It\u2019s now hard to imagine a world without nuclear weapons. But in a little-known episode, then president Ronald Reagan and Soviet leader Mikhail Gorbachev <a href=\"https:\/\/www.theatlantic.com\/politics\/archive\/2016\/01\/ronald-reagans-disarmament-dream\/422244\/\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">nearly agreed<\/a> to ditch all their bombs (a misunderstanding over the \u201c<a href=\"https:\/\/en.wikipedia.org\/wiki\/Strategic_Defense_Initiative\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">Star Wars<\/a>\u201d satellite defense system dashed these hopes). Even though the dream of full disarmament remains just that, nuke counts are <a href=\"https:\/\/ourworldindata.org\/grapher\/nuclear-warhead-stockpiles-lines?country=~OWID_WRL\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">less than<\/a> 20% of their 1986 peak, thanks largely to <a href=\"https:\/\/armscontrolcenter.org\/strategic-arms-reduction-treaty-start-i\/#:~:text=It%20entered%20into%20force%20in,strategic%20nuclear%20weapons%20were%20dismantled.\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">international agreements<\/a>.<\/p>\n<p class=\"dcr-16w5gq9\">These choices weren\u2019t made in a vacuum. Reagan was a <a href=\"https:\/\/www.armscontrol.org\/act\/2010_12\/LookingBack#18:~:text=Superficially%2C%20the%20Reagan,the%20late%201980s.\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">staunch opponent<\/a> of disarmament before the <a href=\"https:\/\/www.armscontrol.org\/act\/2010_12\/LookingBack\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">millions-strong<\/a> Nuclear Freeze movement got to him. In 1983, he <a href=\"https:\/\/www.armscontrol.org\/act\/2010_12\/LookingBack#18:~:text=In%20the%20fall%20of%201983,for%20a%20Soviet%20negotiating%20partner.\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">commented<\/a> to his secretary of state : \u201cIf things get hotter and hotter and arms control remains an issue, maybe I should go see [Soviet leader Yuri] Andropov and propose eliminating all nuclear weapons.\u201d<\/p>\n<p class=\"dcr-16w5gq9\">There are extremely strong economic incentives to keep burning fossil fuels, but climate advocacy has <a href=\"https:\/\/earthbound.report\/2019\/05\/02\/how-extinction-rebellion-shifted-the-overton-window\/\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">pried open<\/a> the Overton window and significantly accelerated our decarbonization efforts.<\/p>\n<p class=\"dcr-16w5gq9\">In April 2019, the young climate group <a href=\"https:\/\/rebellion.global\/\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">Extinction Rebellion<\/a> (XR) <a href=\"https:\/\/www.theguardian.com\/environment\/2019\/apr\/16\/extinction-rebellion-climate-protesters-disrupt-london-rail-tube-lines-blockade-landmarks\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">brought<\/a> London to a halt, demanding the UK target net-zero carbon emissions by 2025. Their controversial civil disobedience prompted parliament to <a href=\"https:\/\/www.cnn.com\/2019\/05\/01\/europe\/uk-climate-emergency-scn-intl\/index.html\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">declare<\/a> a climate emergency and the Labour party to adopt a 2030 <a href=\"https:\/\/www.ft.com\/content\/04f1255c-8c34-11e9-a24d-b42f641eca37\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">target<\/a> to decarbonize the UK\u2019s <a href=\"https:\/\/www.theguardian.com\/environment\/2024\/mar\/18\/sunak-and-ministers-stoking-division-over-uk-net-zero-target-warns-ed-miliband#:~:text=Labour%20did%20not%20claim%20it%20would%20decarbonise%20the%20economy%20by%202030%20%E2%80%93%20rather%2C%20that%20is%20the%20target%20for%20decarbonising%20electricity%2C%20which%20is%20less%20than%20a%20fifth%20of%20overall%20UK%20emissions%2C%20and%20the%20former%20%C2%A328bn%20investment%20target%20covered%20the%20whole%20of%20the%20green%20economy.\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">electricity production<\/a>.<\/p>\n<p class=\"dcr-16w5gq9\">The Sierra Club\u2019s <a href=\"https:\/\/coal.sierraclub.org\/\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">Beyond Coal<\/a> campaign was lesser-known but <a href=\"https:\/\/www.politico.com\/agenda\/story\/2015\/05\/inside-war-on-coal-000002\/\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">wildly effective<\/a>. In just its first five years, the campaign helped shutter more than one-third of US coal plants. <a href=\"https:\/\/www.carbonbrief.org\/analysis-why-us-carbon-emissions-have-fallen-14-since-2005\/#:~:text=Coal%2Dto%2Dgas%20switching%20in%20the%20power%20sector%20is%20the%20largest%20driver%2C%20accounting%20for%2033%25%20of%20the%20emissions%20reduction%20in%202016.\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">Thanks<\/a> <a href=\"https:\/\/sustainability.stanford.edu\/news\/global-carbon-emissions-fossil-fuels-reached-record-high-2023#:~:text=The%20United%20States,nearly%20a%20decade.\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">primarily<\/a> to its move from coal, US per capita carbon emissions <a href=\"https:\/\/ourworldindata.org\/grapher\/co-emissions-per-capita?tab=chart&amp;country=~USA\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">are now<\/a> lower than they were in 1913.<\/p>\n<p class=\"dcr-16w5gq9\">In many ways, the challenge of regulating efforts to build AGI is much smaller than that of decarbonizing. <a href=\"https:\/\/ourworldindata.org\/grapher\/fossil-fuels-share-energy?tab=chart&amp;country=~OWID_WRL\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">Eighty-two percent<\/a> of global energy production comes from fossil fuels. Energy is what makes civilization work, but we\u2019re not dependent on a hypothetical AGI to make the world go round.<\/p>\n<p class=\"dcr-16w5gq9\">Further, slowing and guiding the development of future systems doesn\u2019t mean we\u2019d need to stop using existing systems or developing specialist AIs to tackle important problems in medicine, climate and elsewhere.<\/p>\n<p class=\"dcr-16w5gq9\">It\u2019s obvious why so many capitalists are AI enthusiasts: they foresee a technology that can achieve their long-time dream of cutting workers out of the loop (and the balance sheet).<\/p>\n<p class=\"dcr-16w5gq9\">But governments are not profit maximizers. Sure, they care about economic growth, but they also care about things like employment, social stability, market concentration, and, occasionally, democracy.<\/p>\n<p class=\"dcr-16w5gq9\">It\u2019s far less clear how AGI would affect these domains overall. Governments aren\u2019t prepared for a world where most people are technologically unemployed.<\/p>\n<p class=\"dcr-16w5gq9\">Capitalists often get what they want, particularly in recent decades, and the boundless pursuit of profit may undermine any regulatory effort to slow the speed of AI development. But capitalists don\u2019t always get what they want.<\/p>\n<p class=\"dcr-16w5gq9\">At a bar in San Francisco in February, a longtime <a href=\"https:\/\/www.theguardian.com\/technology\/openai\" data-link-name=\"in body link\" data-component=\"auto-linked-tag\" rel=\"nofollow noopener\" target=\"_blank\">OpenAI<\/a> safety researcher pronounced to a group that the e\/accs shouldn\u2019t be worried about the \u201cextreme\u201d AI safety people, because they\u2019ll never have power. The boosters should actually be afraid of AOC and Senator Josh Hawley because they \u201ccan really fuck things up for you\u201d.<\/p>\n<p class=\"dcr-16w5gq9\">Assuming humans stick around for many millennia, there\u2019s no way to know we won\u2019t eventually build AGI. But this isn\u2019t really what the inevitabilists are saying. Instead, the message tends to be: AGI is imminent. Resistance is futile.<\/p>\n<p class=\"dcr-16w5gq9\">But whether we build AGI in five, 20 or 100 years really matters. And the timeline is far more in our control than the boosters will admit. Deep down, I suspect many of them realize this, which is why they spend so much effort trying to convince others that there\u2019s no point in trying. Besides, if you think AGI is inevitable, why bother convincing anybody?<\/p>\n<p class=\"dcr-16w5gq9\">We actually <a href=\"https:\/\/dynomight.net\/gpt-2\/\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">had<\/a> the computing power required to train GPT-2 more than a decade before OpenAI actually did it, but people didn\u2019t know whether it was worth doing.<\/p>\n<p class=\"dcr-16w5gq9\">But right now, the top AI labs are locked in such a fierce race that they <a href=\"https:\/\/time.com\/6898967\/ai-extinction-national-security-risks-report\/\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">aren\u2019t implementing<\/a> all the precautions that even their own safety teams want. (One OpenAI employee <a href=\"https:\/\/twitter.com\/GarrisonLovely\/status\/1781009052660531362\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">announced<\/a> recently that he quit \u201cdue to losing confidence that it would behave responsibly around the time of AGI\u201d.) There\u2019s a \u201csafety tax\u201d that labs can\u2019t afford to pay if they hope to stay competitive; testing <a href=\"https:\/\/www.alignmentforum.org\/posts\/tmyTb4bQQi7C47sde\/safety-capabilities-tradeoff-dials-are-inevitable-in-agi\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">slows<\/a> product releases and consumes company resources.<\/p>\n<p class=\"dcr-16w5gq9\">Governments, on the other hand, aren\u2019t subject to the same financial pressures.<\/p>\n<p class=\"dcr-16w5gq9\">An inevitabilist tech entrepreneur recently <a href=\"https:\/\/twitter.com\/PeterDiamandis\/status\/1779133938268610637\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">said<\/a> regulating AI development is impossible \u201cunless you control every line of written code\u201d. That might be true if anyone could spin up an AGI on their laptop. But it turns out that building advanced, general AI models <a href=\"https:\/\/epochai.org\/trends#compute-trends-section\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">requires<\/a> enormous arrays of supercomputers, with chips produced by an <a href=\"https:\/\/www.governance.ai\/post\/computing-power-and-the-governance-of-ai#:~:text=Figure%203%3A%20An%20illustration%20of%20the%20compute%20supply%20chain.%C2%A0\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">absurdly monopolistic<\/a> industry. Because of this, many AI safety advocates see \u201c<a href=\"https:\/\/www.governance.ai\/post\/computing-power-and-the-governance-of-ai\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">compute governance<\/a>\u201d as a promising approach. Governments could compel cloud computing providers to halt next generation training runs that don\u2019t comply with established <a href=\"https:\/\/www.nist.gov\/itl\/ai-risk-management-framework\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">guardrails<\/a>. Far from locking out upstarts or requiring Orwellian levels of surveillance, thresholds could be chosen to only affect players who can afford to spend <a href=\"https:\/\/www.wired.com\/story\/openai-ceo-sam-altman-the-age-of-giant-ai-models-is-already-over\/#:~:text=At%20the%20MIT%20event%2C%20Altman%20was%20asked%20%C2%A0if%20training%20GPT%2D4%20cost%20%24100%20million%3B%20he%20replied%2C%20%E2%80%9CIt%E2%80%99s%20more%20than%20that.%E2%80%9D\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">more than $100m<\/a> on a single training run.<\/p>\n<p class=\"dcr-16w5gq9\">Governments do have to worry about international competition and the risk of unilateral disarmament, so to speak. But <a href=\"https:\/\/aitreaty.org\/\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">international treaties<\/a> can be negotiated to widely share the benefits from cutting-edge AI systems while ensuring that labs aren\u2019t blindly scaling up systems they <a href=\"https:\/\/arxiv.org\/pdf\/2310.17688.pdf#page=2\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">don\u2019t understand<\/a>.<\/p>\n<p class=\"dcr-16w5gq9\">And while the world may feel fractious, rival nations have cooperated to surprising degrees.<\/p>\n<p class=\"dcr-16w5gq9\">The Montreal Protocol <a href=\"https:\/\/worksinprogress.co\/issue\/how-we-fixed-the-ozone-layer\/\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">fixed<\/a> the ozone layer by banning chlorofluorocarbons. Most of the world has agreed to ethically motivated bans on militarily useful weapons, such as <a href=\"https:\/\/en.wikipedia.org\/wiki\/Biological_Weapons_Convention\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">biological<\/a> and <a href=\"https:\/\/en.wikipedia.org\/wiki\/Chemical_Weapons_Convention\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">chemical weapons<\/a>, <a href=\"https:\/\/en.wikipedia.org\/wiki\/Protocol_on_Blinding_Laser_Weapons\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">blinding laser weapons<\/a>, and \u201c<a href=\"https:\/\/en.wikipedia.org\/wiki\/Environmental_Modification_Convention\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">weather warfare<\/a>\u201d.<\/p>\n<p class=\"dcr-16w5gq9\">In the 1960s and 70s, many analysts <a href=\"https:\/\/asteriskmag.com\/issues\/03\/the-puzzle-of-non-proliferation#alarm-and-response\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">feared<\/a> that every country that could build nukes, would. But <a href=\"https:\/\/www.researchgate.net\/publication\/264542682_Forgoing_the_nuclear_option_States_that_could_build_nuclear_weapons_but_chose_not_to_do_so\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">most<\/a> of the world\u2019s roughly three-dozen nuclear programs <a href=\"https:\/\/www.belfercenter.org\/sites\/default\/files\/files\/publication\/When%20Did%20%28and%20Didn%27t%29%20States%20Proliferate%3F_1.pdf\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">were abandoned<\/a>. This wasn\u2019t the result of happenstance, but rather the creation of a global nonproliferation norm through deliberate statecraft, like the 1968 <a href=\"https:\/\/en.wikipedia.org\/wiki\/Treaty_on_the_Non-Proliferation_of_Nuclear_Weapons\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">Non-Proliferation Treaty<\/a>.<\/p>\n<p class=\"dcr-16w5gq9\">On the few occasions when Americans were asked if they wanted superhuman AI, <a href=\"https:\/\/acrobat.adobe.com\/id\/urn:aaid:sc:VA6C2:a01a156b-36de-4eec-929e-f085673c5b51\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">large<\/a> <a href=\"https:\/\/www.vox.com\/future-perfect\/2023\/9\/19\/23879648\/americans-artificial-general-intelligence-ai-policy-poll#:~:text=It%20reveals%20that%2063%20percent%20of%20voters%20say%20regulation%20should%20aim%20to%20actively%20prevent%20AI%20superintelligence.\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">majorities<\/a> said \u201cno\u201d. Opposition to AI <a href=\"https:\/\/wiki.aiimpacts.org\/responses_to_ai\/public_opinion_on_ai\/surveys_of_public_opinion_on_ai\/surveys_of_us_public_opinion_on_ai#:~:text=YouGov%27s%20AI%20and,13th%20(56%25%20concerned).\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">has grown<\/a> as the technology has become more prevalent. When people argue that AGI is inevitable, what they\u2019re really saying is that the popular will shouldn\u2019t matter. The boosters see the masses as provincial neo-Luddites who don\u2019t know what\u2019s good for them. That\u2019s why inevitability holds such rhetorical allure for them; it lets them avoid making their real argument, which they know is a loser in the court of public opinion.<\/p>\n<p class=\"dcr-16w5gq9\">The draw of AGI is strong. But the risks involved are potentially civilization-ending. A civilization-scale effort is needed to compel the necessary powers to resist it.<\/p>\n<p class=\"dcr-16w5gq9\">Technology happens because people make it happen. We can choose otherwise.<\/p>\n<p><script async src=\"https:\/\/platform.twitter.com\/widgets.js\" charset=\"utf-8\"><\/script><\/p>\n","protected":false},"excerpt":{"rendered":"Illustration: Petra P\u00e9terffy\/The Guardian \u201cTechnology happens because it is possible,\u201d OpenAI CEO, Sam Altman, told the New York&hellip;\n","protected":false},"author":2,"featured_media":11867,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[20],"tags":[256,254,255,64,63,105],"class_list":{"0":"post-11866","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-ai","9":"tag-artificial-intelligence","10":"tag-artificialintelligence","11":"tag-au","12":"tag-australia","13":"tag-technology"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts\/11866","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/comments?post=11866"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts\/11866\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/media\/11867"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/media?parent=11866"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/categories?post=11866"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/tags?post=11866"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}