{"id":185814,"date":"2025-10-08T16:02:07","date_gmt":"2025-10-08T16:02:07","guid":{"rendered":"https:\/\/www.newsbeep.com\/uk\/185814\/"},"modified":"2025-10-08T16:02:07","modified_gmt":"2025-10-08T16:02:07","slug":"the-flawed-silicon-valley-consensus-on-ai","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/uk\/185814\/","title":{"rendered":"The flawed Silicon Valley consensus on AI"},"content":{"rendered":"<p>Stay informed with free updates<\/p>\n<p class=\"article__content-sign-up-topic-description o3-type-body-base\">Simply sign up to the Technology sector myFT Digest &#8212; delivered directly to your inbox.<\/p>\n<p>Oscar Wilde once defined fox hunting as the unspeakable in pursuit of the uneatable. Were he alive today he might describe the quest for artificial general intelligence as the unfathomable in pursuit of the indefinable.<\/p>\n<p>Many hundreds of billions of dollars are currently being pumped into building generative AI models; it\u2019s a race to achieve human-level intelligence. But not even the developers fully understand how their\u00a0models work or agree exactly what AGI means.\u00a0<\/p>\n<p>Instead of hyperventilating about AI ushering in a new era of abundance, wouldn\u2019t it be better to drop the rhetoric and build AI systems for more defined, realisable goals? That was certainly the view at a conference hosted by the University of Southampton and the Royal Society last week. \u201cWe should stop asking: is the machine intelligent? And ask: what exactly does the machine do?\u201d said Shannon Vallor, a professor at the University of Edinburgh. She has a point.<\/p>\n<p>The term AGI was first popularised in computing in the 2000s to describe how AI systems might one day perform general-purpose, human-level reasoning (as opposed to narrow AI that excels at one thing). Since then, it has become the holy grail for the industry, used to justify a colossal spending spree.<\/p>\n<p>The leading AI research labs, OpenAI and Google DeepMind, both have an explicit corporate mission to achieve AGI, albeit with varying definitions. OpenAI\u2019s is: \u201ca highly autonomous system that outperforms humans at most economically valuable work\u201d. But even Sam Altman, its chief executive, who has signed <a href=\"https:\/\/www.ft.com\/content\/5f6f78af-aed9-43a5-8e31-2df7851ceb67\" data-trackable=\"link\" rel=\"nofollow noopener\" target=\"_blank\">$1tn in deals this year<\/a> to boost computing power, <a href=\"https:\/\/www.cnbc.com\/2025\/08\/11\/sam-altman-says-agi-is-a-pointless-term-experts-agree.html\" data-trackable=\"link\" rel=\"nofollow noopener\" target=\"_blank\">concedes<\/a> this is not a \u201csuper-useful term\u201d.<\/p>\n<p>Even if we accept its use case, there remain two concerns. What happens if we achieve AGI. And what happens if we don\u2019t.<\/p>\n<p>The Silicon Valley consensus, as it has been called, suggests that AGI, however defined, is within reach this decade. That goal is being pursued with missionary zeal by the leading AI labs in the belief that it will unleash massive productivity gains and generate huge returns for investors.\u00a0<\/p>\n<p>Indeed, some west coast tech leaders have founded <a href=\"https:\/\/www.washingtonpost.com\/technology\/2025\/08\/26\/silicon-valley-ai-super-pac\/\" data-trackable=\"link\" rel=\"nofollow noopener\" target=\"_blank\">a $100mn political action committee<\/a> to support \u201cpro-AI\u201d candidates in the 2026 midterm elections and squash unhelpful regulation. They point to the <a href=\"https:\/\/ig.ft.com\/ai-personal-assistant\/\" data-trackable=\"link\" rel=\"nofollow noopener\" target=\"_blank\">astonishingly rapid adoption of AI-powered chatbots<\/a> and scoff at the doomers, or decelerationists, who want to slow progress and hobble the US in its technological race with China.<\/p>\n<p>But AGI is not going to be an instant, miraculous blessing. OpenAI itself acknowledges that it would also come \u201cwith serious risk of misuse, drastic accidents, and societal disruption\u201d. This helps explain why<a href=\"https:\/\/www.ft.com\/content\/0211e603-7da6-45a7-909a-96ec28bf6c5a\" data-trackable=\"link\" rel=\"nofollow noopener\" target=\"_blank\"> insurers are now balking<\/a> at providing comprehensive cover to the industry.<\/p>\n<p>Some experts, such as Eliezer Yudkowsky and Nate Soares, go way further, warning that a rogue superintelligence could pose an existential threat to humanity. The title of their recent book, If Anyone Builds It, Everyone Dies, pretty much sums up the argument.<\/p>\n<p>Yet not everyone is convinced that AGI\u2019s arrival is imminent. Sceptics doubt that the industry\u2019s favourite party trick of scaling computing power to produce smarter large language models will take us there. \u201cWe\u2019re still a few conceptual breakthroughs short of AGI,\u201d says one top researcher.\u00a0<\/p>\n<p>In a survey conducted by <a href=\"https:\/\/aaai.org\/wp-content\/uploads\/2025\/03\/AAAI-2025-PresPanel-Report-FINAL.pdf\" data-trackable=\"link\" rel=\"nofollow noopener\" target=\"_blank\">the Association for the Advancement of Artificial Intelligence<\/a> this year, 76 per cent of the 475 (mostly academic) respondents thought it unlikely or very unlikely that current approaches would yield AGI. That could be a problem: the US stock market appears to rest on <a href=\"https:\/\/www.ft.com\/content\/6cc87bd9-cb2f-4f82-99c5-c38748986a2e\" data-trackable=\"link\" rel=\"nofollow noopener\" target=\"_blank\">the opposite conviction.<\/a><\/p>\n<p>Many of the attendees at last week\u2019s event contested the Silicon Valley framing of AI. The world was not on a predestined technological trajectory with only one outcome. Other approaches were worth pursuing rather than betting so much on the deep learning behind generative AI models. AI companies could not just shrug off the problems they were creating today with promises of a more glorious tomorrow. The rest of society should resist being treated \u201cas passengers at the back of a bus\u201d hoping AI takes us somewhere nice, said Vallor.<\/p>\n<p>The 85-year-old computer pioneer Alan Kay, a revered figure in the industry, offered some perspective. He argued that AI could undoubtedly bring real benefits. Indeed, it had helped detect his cancer in an MRI scan. \u201cAI is a lifesaver,\u201d he said.<\/p>\n<p>However, Kay worried that humans are easily fooled and that AI companies cannot always explain how their models produce the results. Software engineers, like aeroplane designers or bridge builders, he said, had a duty of care to ensure their systems did not cause harm or fail. The main theme of this century should be safety.<\/p>\n<p>The best way forward would be to harness humanity\u2019s collective intelligence, steadily amassed over generations. \u201cWe already have artificial superhuman intelligence,\u201d Kay said. \u201cIt is science.\u201d AI has already produced some exquisite breakthroughs, such as Google DeepMind\u2019s AlphaFold model that predicted the structures of over 200mn proteins \u2014 winning the researchers a Nobel.<\/p>\n<p>Kay highlighted his particular concerns about the vulnerabilities of AI-generated code. Citing his fellow computer scientist Butler Lampson, he said: \u201cStart the genies off in bottles and keep them there.\u201d That\u2019s not a bad adage for our AI age.<\/p>\n<p><a href=\"https:\/\/www.ft.com\/content\/mailto:john.thornhill@ft.com\" data-trackable=\"link\" rel=\"nofollow noopener\" target=\"_blank\">john.thornhill@ft.com<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"Stay informed with free updates Simply sign up to the Technology sector myFT Digest &#8212; delivered directly to&hellip;\n","protected":false},"author":2,"featured_media":185815,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[2],"tags":[49,50,51,47,52,48],"class_list":{"0":"post-185814","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-headlines","8":"tag-headlines","9":"tag-news","10":"tag-top-news","11":"tag-top-stories","12":"tag-topnews","13":"tag-topstories"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/posts\/185814","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/comments?post=185814"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/posts\/185814\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/media\/185815"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/media?parent=185814"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/categories?post=185814"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/tags?post=185814"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}