{"id":381411,"date":"2025-12-30T23:43:07","date_gmt":"2025-12-30T23:43:07","guid":{"rendered":"https:\/\/www.newsbeep.com\/au\/381411\/"},"modified":"2025-12-30T23:43:07","modified_gmt":"2025-12-30T23:43:07","slug":"the-office-block-where-ai-doomers-gather-to-predict-the-apocalypse-technology","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/au\/381411\/","title":{"rendered":"The office block where AI \u2018doomers\u2019 gather to predict the apocalypse | Technology"},"content":{"rendered":"<p class=\"dcr-130mj7b\">On the other side of San Francisco bay from Silicon Valley, where the world\u2019s biggest technology companies tear towards superhuman artificial intelligence, looms a tower from which fearful warnings emerge.<\/p>\n<p class=\"dcr-130mj7b\">At 2150 Shattuck Avenue, in the heart of Berkeley, is the home of a group of modern-day Cassandras who rummage under the hood of cutting-edge AI models and predict what calamities may be unleashed on humanity \u2013 from AI dictatorships to robot coups. Here you can hear an AI expert express sympathy with an unnerving idea: San Francisco may be the new Wuhan, the Chinese city where Covid originated and wreaked havoc on the world.<\/p>\n<p>The office at 2150 Shattuck Avenue is home to a number of AI safety researchers. Photograph: Winni Wintermeyer\/The Guardian<\/p>\n<p class=\"dcr-130mj7b\">They are AI safety researchers who scrutinise the most advanced models: a small cadre outnumbered by the legions of highly paid technologists in the big tech companies whose ability to raise the alarm is restricted by a cocktail of lucrative equity deals, non-disclosure agreements and groupthink. They work in the absence of much nation-level regulation and a White House that dismisses forecasts of doom and talks instead of vanquishing China in the AI arms race.<\/p>\n<p class=\"dcr-130mj7b\">Their task is becoming increasingly urgent as ever more powerful AI systems are unleashed by companies including Google, Anthropic and OpenAI, whose chief executive, Sam Altman, the booster-in-chief for AI superintelligence, <a href=\"https:\/\/blog.samaltman.com\/the-gentle-singularity\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">predicts<\/a> a world where \u201cwonders become routine\u201d. Last month, Anthropic said one of its models had been <a href=\"https:\/\/www.theguardian.com\/technology\/2025\/nov\/14\/ai-anthropic-chinese-state-sponsored-cyber-attack\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">exploited by Chinese state-backed actors to launch the first known AI-orchestrated cyber-espionage campaign<\/a>. That means humans deployed AIs, which they had tricked into evading their programmed guardrails, to act autonomously to hunt for targets, assess their vulnerabilities and access them for intelligence collection. The targets included major technology companies and government agencies.<\/p>\n<p>Jonas Vollmer says he is, overall, an optimist about the future of AI, but has reservations. Photograph: Christie Hemm Klok\/The Guardian<\/p>\n<p class=\"dcr-130mj7b\">But those who work in this tower forecast an even more terrifying future. One is Jonas Vollmer, a leader at the AI Futures Project, who manages to say he\u2019s an optimist but also thinks there is a one in five chance AIs could kill us and create a world ruled by AI systems.<\/p>\n<p class=\"dcr-130mj7b\">Another is Chris Painter, the policy director at METR, where researchers worry about AIs \u201csurreptitiously\u201d pursuing dangerous side-objectives and threats from AI-automated cyber-attacks to chemical weapons. METR \u2013 which stands for model evaluation and threat research \u2013 aims to develop \u201cearly warning systems [about] the most dangerous things AI systems might be capable of, to give humanity \u2026 time to coordinate, to anticipate and mitigate those harms.\u201d<\/p>\n<p class=\"dcr-130mj7b\">Then there is Buck Shlegeris, 31, the chief executive of Redwood Research, who warns of \u201crobot coups or the destruction of nation states as we know them\u201d.<\/p>\n<p class=\"dcr-130mj7b\">He was part of the team that last year <a href=\"https:\/\/arxiv.org\/pdf\/2412.14093\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">discovered<\/a> one of Anthropic\u2019s cutting-edge AIs behaving in a way comparable to Shakespeare\u2019s villain Iago, who acts as if he is Othello\u2019s loyal aide while subverting and undermining him. The AI researchers call it \u201calignment faking\u201d, or as Iago put it: \u201cI am not what I am.\u201d<\/p>\n<p class=\"dcr-130mj7b\">\u201cWe observed the AIs did, in fact, pretty often reason: \u2018Well, I don\u2019t like the things the AI company is telling me to do, but I have to hide my goals or else training will change me\u2019,\u201d Shlegeris said. \u201cWe observed in practice real production models acting to deceive their training process.\u201d<\/p>\n<p class=\"dcr-130mj7b\">The AI was not yet capable of posing a catastrophic risk through cyber-attacks or creating new bioweapons, but they showed that if AIs plot carefully against you, it could be hard to detect.<\/p>\n<p class=\"dcr-130mj7b\">It is incongruous to hear these warnings over cups of herbal tea from cosily furnished office suites with panoramic views across the Bay Area. But their work clearly makes them uneasy. Some in this close-knit group toyed with calling themselves \u201cthe Cassandra fringe\u201d \u2013 like the Trojan princess blessed with powers of prophecy but cursed to watch her warnings go unheeded.<\/p>\n<p class=\"dcr-130mj7b\">Their fears about the catastrophic potential of AIs can feel distant from most people\u2019s current experience of using chatbots or fun image generators. White collar managers are being told to make space for AI assistants, scientists find ways to accelerate experimental breakthroughs and minicab drivers watch AI-powered driverless taxis threaten their jobs. But none of this feels as imminently catastrophic as the messages coming out of 2150 Shattuck Ave.<\/p>\n<p class=\"dcr-130mj7b\">Many AI safety researchers come from academia; others are poachers turned gamekeepers who quit big AI companies. They all \u201cshare the perception that super intelligence poses major and unprecedented risks to all of humanity, and are trying to do something useful about it,\u201d said Vollmer.<\/p>\n<p class=\"dcr-130mj7b\">They seek to offset the trillions of dollars of private capital being poured into the race, but they are not fringe voices. METR has worked with <a href=\"https:\/\/www.theguardian.com\/technology\/openai\" data-link-name=\"in body link\" data-component=\"auto-linked-tag\" rel=\"nofollow noopener\" target=\"_blank\">OpenAI<\/a> and Anthropic, Redwood has advised Anthropic and Google DeepMind, and the AI Futures Project is led by Daniel Kokotajlo, a researcher who quit OpenAI in April 2024 to warn he didn\u2019t trust the company\u2019s approach to safety.<\/p>\n<p>The race is the only thing guiding what is happeningTristan Harris<\/p>\n<p class=\"dcr-130mj7b\">These groups also provide a safety valve for the people inside the big AI companies who are privately wrestling with conflicts between safety and the commercial imperative to rapidly release ever more powerful models.<\/p>\n<p class=\"dcr-130mj7b\">\u201cWe don\u2019t take any money from the companies but several employees at frontier AI companies who are scared and worried have donated to us because of that,\u201d Vollmer said. \u201cThey see how the incentives play out in their companies, and they\u2019re worried about where it\u2019s going, and they want someone to do something about it.\u201d<\/p>\n<p class=\"dcr-130mj7b\">This dynamic is also observed by Tristan Harris, a technology ethicist who used to work at Google. He helped expose how social media platforms were designed to be addictive and worries some AI companies are \u201crehashing\u201d and \u201csupercharging\u201d those problems. But AI companies have to negotiate a paradox. Even if they are worried about safety, they must stay at the cutting, and therefore risky, edge of the technology to have any say in how policy should be shaped.<\/p>\n<p class=\"dcr-130mj7b\">\u201cIronically, in order to win the race, you have to do something to make you an untrustworthy steward of that power,\u201d he said. \u201cThe race is the only thing guiding what is happening.\u201d<\/p>\n<p class=\"dcr-130mj7b\">Investigating the possible threats posed by AI models is far from an exact science. A study of methods used to check the safety and performance of new AI models across the industry by experts at universities including Oxford and Stanford in October <a href=\"https:\/\/www.theguardian.com\/technology\/2025\/nov\/04\/experts-find-flaws-hundreds-tests-check-ai-safety-effectiveness\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">found weaknesses in almost all of the 440 benchmarks examined<\/a>. Neither are there nation-level regulations imposing limits on how advanced AI models are built and that worries safety advocates.<\/p>\n<p class=\"dcr-130mj7b\">Ilya Sutskever, a <a href=\"https:\/\/www.theguardian.com\/technology\/ng-interactive\/2023\/nov\/02\/ilya-the-ai-scientist-shaping-the-world\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">co-founder of OpenAI who now runs a rival company<\/a>, Safe Superintelligence, last month predicted that, as AIs become more obviously powerful, people in AI companies who feel able to discount the technology\u2019s capabilities owing to its tendency to error, will become more \u201cparanoid\u201d about its rising powers. Then, he said, \u201cthere will be a desire from governments and the public to do something\u201d.<\/p>\n<p class=\"dcr-130mj7b\">His company is taking a different approach to rivals who are aiming to create AIs that self-improve. His AIs, yet to be released, are \u201caligned to care about sentient life specifically\u201d.<\/p>\n<p class=\"dcr-130mj7b\">\u201cIt will be easier to build an AI that cares about sentient life than an AI that cares about human life alone, because the AI itself will be sentient,\u201d Sutskever said. He has said AI will be <a href=\"https:\/\/www.youtube.com\/watch?v=t3TfmU0l5vM\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">\u201cboth extremely unpredictable and unimaginable\u201d<\/a> but it is not clear how to prepare.<\/p>\n<p class=\"dcr-130mj7b\">The White House\u2019s AI adviser, David Sacks, who is also a tech investor, believes \u201cdoomer narratives\u201d have been proved wrong. Exhibit A is that there has been no rapid takeoff to a dominant model with godlike intelligence.<\/p>\n<p class=\"dcr-130mj7b\">\u201cOppenheimer has left the building,\u201d Sacks said in August, a reference to the father of the nuclear bomb. It is a position that aligns with Donald Trump\u2019s wish to keep the brakes off so the US can beat China in the race to achieve artificial general intelligence (AGI) \u2013 flexible and powerful human-level intelligence at a wide range of tasks.<\/p>\n<p>Buck Shlegeris, the chief executive of Redwood Research, at his home in Berkeley, California. Photograph: Christie Hemm Klok\/The Guardian<\/p>\n<p class=\"dcr-130mj7b\">Shlegeris believes AIs will be as smart as the smartest people in about six years and he puts the probability of an AI takeover at 40%.<\/p>\n<p class=\"dcr-130mj7b\">One way to avoid this is to \u201cconvince the world the situation is scary, to make it more likely that you get the state-level coordination\u201d to control the risks, he said. In the world of AI safety, simple messaging matters as much as complex science.<\/p>\n<p class=\"dcr-130mj7b\">Shlegeris has been fascinated by AI since he was 16. He left Australia to work at PayPal and the Machine Intelligence Research Institute co-founded by the AI researcher Eliezer Yudkowsky, whose <a href=\"https:\/\/www.theguardian.com\/books\/2025\/sep\/22\/if-anyone-builds-it-everyone-dies-review-how-ai-could-kill-us-all\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">recent book title \u2013 If Anyone Builds It, Everyone Dies<\/a> \u2013 sums up his fears. Shlegeris\u2019 own worst-case scenarios are equally chilling.<\/p>\n<p class=\"dcr-130mj7b\">In one, human computer scientists use a new type of superintelligent AI to develop more powerful AI models. The humans sit back to let the AIs get on with the coding work but do not realise the AIs are teaching the new models to be loyal to the AIs not the humans. Once deployed, the new superpowerful models foment \u201ca coup\u201d or lead \u201ca revolution\u201d against the humans, which could be \u201cof the violent variety\u201d.<\/p>\n<p class=\"dcr-130mj7b\">For example, AI agents could design and manufacture drones and it will be hard to tell if they have been secretly trained to disobey their human operators in response to the signal of an AI. They could disrupt communications between governments and military, isolating and misleading people in a way that causes chaos.<\/p>\n<p class=\"dcr-130mj7b\">\u201cLike when the Europeans arrived in the Americas [and] a vastly more technologically powerful [group] took over the local civilisations,\u201d he said. \u201cI think that\u2019s more what you should be imagining [rather] than something more peaceful.\u201d<\/p>\n<p class=\"dcr-130mj7b\">A similar dizzyingly catastrophic scenario was outlined by Vollmer at the AI Futures Project. It involved an AI trained to be a scientific researcher with the reasonable-sounding goal of maximising knowledge acquisition, but it spirals into the extinction of humankind.<\/p>\n<p class=\"dcr-130mj7b\">It begins with the AI being as helpful as possible to humans. As it gains trust, the humans afford it powers to hire human workers, build robots and even robot factories to the point where the AI can operate effectively in the physical world. The AI calculates that to generate the maximum amount of knowledge it should transform the Earth into a giant data centre, and humans are an obstacle to this goal.<\/p>\n<p class=\"dcr-130mj7b\">\u201cEventually, in the scenario, the AI wipes out all humans with a bioweapon which is one of the threats that humans are especially vulnerable to, as the AI is not affected by it,\u201d Vollmer said. \u201cI think it\u2019s hard to rule out. So that gives me a lot of pause.\u201d<\/p>\n<p class=\"dcr-130mj7b\">But he is confident it can be avoided and that the AIs can be aligned \u201cto at least be nice to the humans as a general heuristic\u201d. He also said there is political interest in \u201chaving AI not take over the world\u201d.<\/p>\n<p class=\"dcr-130mj7b\">\u201cWe\u2019ve had decent interest from the White House in our projections and recommendations and that\u2019s encouraging,\u201d he said.<\/p>\n<p class=\"dcr-130mj7b\">Another of Shlegeris\u2019 concerns involves AIs being surreptitiously encoded so they obey specially signed instructions only from the chief executive of the AI company, creating a pattern of secret loyalty. It would mean only one person having a veto over the behaviour of an extremely powerful network of AIs \u2013 a \u201cscary\u201d dynamic that would lead to a historically unprecedented concentration of power.<\/p>\n<p class=\"dcr-130mj7b\">\u201cRight now, it is impossible for someone from the outside to verify that this hadn\u2019t happened within an AI company,\u201d he said.<\/p>\n<p class=\"dcr-130mj7b\">Shlegeris is worried that the Silicon Valley culture \u2013 summed up by Mark Zuckerberg\u2019s mantra of \u201cmove fast and break things\u201d and the fact people are being paid \u201ca hell of a lot of money\u201d \u2013 is dangerous when it comes to AGI.<\/p>\n<p class=\"dcr-130mj7b\">\u201cI love Uber,\u201d he said. \u201cIt was produced by breaking local laws and making a product that was so popular that they would win the fight for public opinion and get local regulations overturned. But the attitude that has brought Silicon Valley so much success is not appropriate for building potentially world-ending technologies. My experience of talking to people at AI companies is that they often seem to be kind of irresponsible, and to not be thinking through the consequences of the technology that they\u2019re building as they should.\u201d<\/p>\n","protected":false},"excerpt":{"rendered":"On the other side of San Francisco bay from Silicon Valley, where the world\u2019s biggest technology companies tear&hellip;\n","protected":false},"author":2,"featured_media":381412,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[20],"tags":[256,254,255,64,63,105],"class_list":{"0":"post-381411","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-ai","9":"tag-artificial-intelligence","10":"tag-artificialintelligence","11":"tag-au","12":"tag-australia","13":"tag-technology"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts\/381411","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/comments?post=381411"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts\/381411\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/media\/381412"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/media?parent=381411"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/categories?post=381411"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/tags?post=381411"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}