{"id":203,"date":"2025-07-17T10:06:27","date_gmt":"2025-07-17T10:06:27","guid":{"rendered":"https:\/\/www.newsbeep.com\/au\/203\/"},"modified":"2025-07-17T10:06:27","modified_gmt":"2025-07-17T10:06:27","slug":"ai-firms-unprepared-for-dangers-of-building-human-level-systems-report-warns-artificial-intelligence-ai","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/au\/203\/","title":{"rendered":"AI firms \u2018unprepared\u2019 for dangers of building human-level systems, report warns | Artificial intelligence (AI)"},"content":{"rendered":"<p class=\"dcr-16w5gq9\">Artificial intelligence companies are \u201cfundamentally unprepared\u201d for the consequences of creating systems with human-level intellectual performance, according to a <a href=\"https:\/\/www.theguardian.com\/technology\/2023\/mar\/31\/ai-research-pause-elon-musk-chatgpt\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">leading AI safety<\/a> group.<\/p>\n<p class=\"dcr-16w5gq9\">The Future of Life Institute (FLI) said none of the firms on its <a href=\"https:\/\/futureoflife.org\/ai-safety-index-summer-2025\/\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">AI safety index<\/a> scored higher than a D for \u201cexistential safety planning\u201d.<\/p>\n<p class=\"dcr-16w5gq9\">One of the five reviewers of the FLI\u2019s report said that, despite aiming to develop artificial general intelligence (AGI), none of the companies scrutinised had \u201canything like a coherent, actionable plan\u201d to ensure the systems remained safe and controllable.<\/p>\n<p class=\"dcr-16w5gq9\">AGI refers to a theoretical stage of AI development at which a system is capable of matching a human in carrying out any intellectual task. OpenAI, the developer of ChatGPT, has said its mission is to ensure AGI \u201cbenefits all of humanity\u201d. Safety campaigners have warned that AGI could pose an existential threat by evading human control and <a href=\"https:\/\/www.theguardian.com\/technology\/2023\/oct\/25\/ai-dangers-must-be-faced-head-on-rishi-sunak-to-tell-tech-summit\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">triggering a catastrophic event<\/a>.<\/p>\n<p class=\"dcr-16w5gq9\">The FLI\u2019s report said: \u201cThe industry is fundamentally unprepared for its own stated goals. Companies claim they will achieve artificial general intelligence (AGI) within the decade, yet none scored above D in existential safety planning.\u201d<\/p>\n<p class=\"dcr-16w5gq9\">The index evaluates seven AI developers \u2013 Google DeepMind, OpenAI, Anthropic, Meta, xAI and China\u2019s Zhipu AI and DeepSeek \u2013 across six areas including \u201ccurrent harms\u201d and \u201cexistential safety\u201d.<\/p>\n<p class=\"dcr-16w5gq9\">Anthropic received the highest overall safety score with a C+, followed by OpenAI with a C and Google DeepMind with a C-.<\/p>\n<p class=\"dcr-16w5gq9\">The FLI is a US-based non-profit that campaigns for safer use of cutting-edge technology and is able to operate independently due to an \u201cunconditional\u201d donation from crypto entrepreneur <a href=\"https:\/\/www.theguardian.com\/technology\/2022\/sep\/15\/ethereum-cryptocurrency-completes-move-to-cut-co2-output-by-99\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">Vitalik Buterin<\/a>.<\/p>\n<p class=\"dcr-16w5gq9\">SaferAI, another safety-focused non-profit, <a href=\"https:\/\/ratings.safer-ai.org\/\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">also released a report on Thursday<\/a> warning that advanced AI companies have \u201cweak to very weak risk management practices\u201d and labelled their current approach \u201cunacceptable\u201d.<\/p>\n<p class=\"dcr-16w5gq9\">The FLI safety grades were assigned and reviewed by a panel of AI experts, including British computer scientist Stuart Russell, and Sneha Revanur, founder of AI regulation campaign group Encode Justice.<\/p>\n<p class=\"dcr-16w5gq9\">Max Tegmark, a co-founder of FLI and a professor at Massachusetts Institute of Technology, said it was \u201cpretty jarring\u201d that cutting-edge AI firms were aiming to build super-intelligent systems without publishing plans to deal with the consequences.<\/p>\n<p class=\"dcr-16w5gq9\">He said: \u201cIt\u2019s as if someone is building a gigantic nuclear power plant in New York City and it is going to open next week \u2013 but there is no plan to prevent it having a meltdown.\u201d<\/p>\n<p class=\"dcr-16w5gq9\">Tegmark said the technology was continuing to <a href=\"https:\/\/www.theguardian.com\/technology\/2025\/jan\/28\/former-openai-safety-researcher-brands-pace-of-ai-development-terrifying\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">outpace expectations<\/a>, citing a previously held belief that experts would have decades to address the challenges of AGI. \u201cNow the companies themselves are saying it\u2019s a few years away,\u201d he said.<\/p>\n<p class=\"dcr-16w5gq9\">He added that progress in AI capabilities had been \u201cremarkable\u201d since the <a href=\"https:\/\/www.theguardian.com\/technology\/2025\/feb\/14\/global-disunity-energy-concerns-and-the-shadow-of-musk-key-takeaways-from-the-paris-ai-summit\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">global AI summit in Paris in February<\/a>, with new models such as xAI\u2019s Grok 4, Google\u2019s Gemini 2.5, and its video generator Veo3, all showing improvements on their forebears.<\/p>\n<p class=\"dcr-16w5gq9\">A Google DeepMind spokesperson said the reports did not take into account \u201call of Google DeepMind\u2019s AI safety efforts\u201d. They added: \u201cOur comprehensive approach to AI safety and security extends well beyond what\u2019s captured.\u201d<\/p>\n<p class=\"dcr-16w5gq9\">OpenAI, Anthropic, Meta, xAI, Zhipu AI and DeepSeek have also been approached for comment.<\/p>\n","protected":false},"excerpt":{"rendered":"Artificial intelligence companies are \u201cfundamentally unprepared\u201d for the consequences of creating systems with human-level intellectual performance, according to&hellip;\n","protected":false},"author":2,"featured_media":204,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[20],"tags":[256,254,255,64,63,105],"class_list":{"0":"post-203","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-ai","9":"tag-artificial-intelligence","10":"tag-artificialintelligence","11":"tag-au","12":"tag-australia","13":"tag-technology"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts\/203","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/comments?post=203"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts\/203\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/media\/204"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/media?parent=203"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/categories?post=203"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/tags?post=203"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}