{"id":173071,"date":"2025-12-03T15:28:08","date_gmt":"2025-12-03T15:28:08","guid":{"rendered":"https:\/\/www.newsbeep.com\/ie\/173071\/"},"modified":"2025-12-03T15:28:08","modified_gmt":"2025-12-03T15:28:08","slug":"top-ai-companies-safety-practices-fall-short-says-new-report","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/ie\/173071\/","title":{"rendered":"Top AI companies\u2019 safety practices fall short, says new report"},"content":{"rendered":"<p id=\"anchor-4da6f2\" class=\"body-graf\">As leading artificial intelligence companies release increasingly capable AI systems, a new report is sounding the alarm about what it says are some of those companies\u2019 lagging safety practices.<\/p>\n<p id=\"anchor-3821b9\" class=\"body-graf\">The Winter 2025 AI Safety Index, which examines the safety protocols of eight leading AI companies, found that their approaches \u201clack the concrete safeguards, independent oversight and credible long-term risk-management strategies that such powerful systems demand.\u201d<\/p>\n<p id=\"anchor-5ab6e6\" class=\"body-graf\">Sabina Nong, an AI safety investigator at the nonprofit Future of Life Institute (FLI), which organized the report and works to address large-scale risks from technologies like nuclear weapons and AI, said in an interview at the <a href=\"https:\/\/www.far.ai\/events\/event-list\/san-diego-alignment-workshop\" target=\"_blank\" rel=\"nofollow noopener\">San Diego Alignment Workshop<\/a> that the analysis revealed a divide in organizations\u2019 approaches to safety. <\/p>\n<p id=\"anchor-025c4b\" class=\"body-graf\">\u201cWe see two clusters of companies in terms of their safety promises and practices,\u201d Nong said. \u201cThree companies are leading: Anthropic, OpenAI, Google DeepMind, in that order, and then five other companies are on the next tier.\u201d <\/p>\n<p id=\"anchor-079859\" class=\"body-graf\">The lower tier of five companies includes xAI and Meta, along with the Chinese AI companies Z.ai, DeepSeek and Alibaba Cloud. Chinese models have been<a href=\"https:\/\/www.nbcnews.com\/tech\/innovation\/silicon-valley-building-free-chinese-ai-rcna242430\" target=\"_blank\" rel=\"nofollow noopener\"> increasingly adopted<\/a> in Silicon Valley as their capabilities have quickly advanced, and they are readily available because they are largely open source.<\/p>\n<p id=\"anchor-c0ac9a\" class=\"body-graf\">Anthropic, the highest-ranked company on the list, got a C+ grade, while Alibaba Cloud, the lowest-ranked, received a D-.<\/p>\n<p id=\"anchor-768566\" class=\"body-graf\">The index examined 35 safety indicators across six domains, including companies\u2019 risk-assessment practices, information sharing protocols and whistleblowing protections, in addition to support for AI safety research. <\/p>\n<p id=\"anchor-deb529\" class=\"body-graf\">Eight independent AI experts, including Massachusetts Institute of Technology professor Dylan Hadfield-Menell and Yi Zeng, a professor at the Chinese Academy of Sciences, graded companies\u2019 fulfillment of the safety indicators.<\/p>\n<p id=\"anchor-a02016\" class=\"body-graf\">FLI President Max Tegmark, an MIT professor, said the report provided clear evidence that AI companies are speeding toward a dangerous future, partly because of a lack of regulations around AI. <\/p>\n<p id=\"anchor-da5af3\" class=\"body-graf\">\u201cThe only reason that there are so many C\u2019s and D\u2019s and F\u2019s in the report is because there are fewer regulations on AI than on making sandwiches,\u201d Tegmark told NBC News, referring to the <a href=\"https:\/\/www.nbcnews.com\/tech\/tech-news\/steve-bannon-elizabeth-warren-bipartisan-backlash-erupts-push-block-st-rcna245040\" target=\"_blank\" rel=\"nofollow noopener\">continued lack of adequate AI laws<\/a> and the established nature of food-safety regulation.<\/p>\n<p id=\"anchor-f2dc63\" class=\"body-graf\">The report recommended that AI companies share more information about their internal processes and assessments, use independent safety evaluators, increase efforts to prevent AI psychosis and harm and reduce lobbying, among other measures.<\/p>\n<p id=\"anchor-1fb050\" class=\"body-graf\">Tegmark, Nong and FLI are particularly concerned about the potential for AI systems to cause catastrophic harm, especially given <a href=\"https:\/\/time.com\/7205596\/sam-altman-superintelligence-agi\/?utm_source=chatgpt.com\" target=\"_blank\" rel=\"nofollow noopener\">calls from AI leaders like Sam Altman<\/a>, the CEO of OpenAI, to build AI systems that are smarter than humans \u2014 also called artificial superintelligence.<\/p>\n<p id=\"anchor-e66eb9\" class=\"body-graf\">\u201cI don\u2019t think companies are prepared for the existential risk of the superintelligent systems that they are about to create and are so ambitious to march towards,\u201d Nong said.<\/p>\n<p id=\"anchor-5cd913\" class=\"body-graf\">The report, released Wednesday morning, comes on the heels of several boundary-pushing AI model launches. <a href=\"https:\/\/blog.google\/products\/gemini\/gemini-3\/\" target=\"_blank\" rel=\"nofollow noopener\">Google\u2019s Gemini 3 model<\/a>, released at the end of November, has <a href=\"https:\/\/artificialanalysis.ai\/articles\/gemini-3-pro-everything-you-need-to-know\" target=\"_blank\" rel=\"nofollow noopener\">set records for performance<\/a> on a series of tests designed to measure AI systems\u2019 capabilities. <\/p>\n<p id=\"anchor-0bb12f\" class=\"body-graf\">On Monday, one of China\u2019s leading AI companies, DeepSeek, released a <a href=\"https:\/\/api-docs.deepseek.com\/news\/news251201\" target=\"_blank\" rel=\"nofollow noopener\">cutting-edge model<\/a> that <a href=\"https:\/\/www.bloomberg.com\/news\/articles\/2025-12-01\/deepseek-debuts-new-ai-models-to-rival-google-and-openai\" target=\"_blank\" rel=\"nofollow noopener\">appears to match Gemini 3\u2019s capabilities<\/a> in several domains. <\/p>\n<p id=\"anchor-925454\" class=\"body-graf\">Though AI capability tests <a href=\"https:\/\/www.nbcnews.com\/tech\/tech-news\/ai-chatgpt-test-smart-capabilities-may-exaggerated-flawed-study-rcna241969\" target=\"_blank\" rel=\"nofollow noopener\">are increasingly criticized as flawed<\/a>, partly because of the potential for AI systems to become hyper-focused on passing a specific series of unrealistic challenges, the record-breaking scores from new models signal systems\u2019 relative performance above competitors.<\/p>\n<p id=\"anchor-ca5205\" class=\"body-graf\">Even though DeepSeek\u2019s new model performs at or near the frontier of AI capabilities, Wednesday\u2019s Safety Index report says DeepSeek fails on many key safety considerations.<\/p>\n<p id=\"anchor-fcf330\" class=\"body-graf\">The report scored DeepSeek second-to-last out of the eight companies on an overall safety metric. The report\u2019s independent panel found that, unlike all leading American companies, DeepSeek does not publish any framework outlining its safety-minded evaluations or mitigations and does not disclose a whistleblowing policy that could help identify key risks from AI models.<\/p>\n<p id=\"anchor-1fed58\" class=\"body-graf\">Frameworks outlining company safety policies and testing mechanisms are <a href=\"https:\/\/www.nbcnews.com\/tech\/tech-news\/ai-law-california-ca-companies-regulation-newsom-rcna234562\" target=\"_blank\" rel=\"nofollow noopener\">now required for companies operating in California<\/a>. Those <a href=\"https:\/\/www.frontiermodelforum.org\/updates\/issue-brief-components-of-frontier-ai-safety-frameworks\/\" target=\"_blank\" rel=\"nofollow noopener\">frameworks can help companies<\/a> avoid severe risks, like the potential for AI products to be used in cybersecurity attacks or bioweapon design.  <\/p>\n<p id=\"anchor-8b29f6\" class=\"body-graf\">The report classifies DeepSeek in the lower tier of safety-minded companies. \u201cThe lower tier companies continue to fall short on basic elements such as safety frameworks, governance structures, and comprehensive risk assessment,\u201d the report says.<\/p>\n<p id=\"anchor-f391aa\" class=\"body-graf\">Tegmark said, \u201cSecond-tier companies have been completely obsessed by catching up to the technical frontier, but now that they have, they no longer have an excuse to not also prioritize safety.\u201d <\/p>\n<p id=\"anchor-b76664\" class=\"body-graf\">Advances in AI capabilities have <a href=\"https:\/\/www.businessinsider.com\/sora-video-app-openai-pros-cons-explained-why-2025-10\" target=\"_blank\" rel=\"nofollow noopener\">recently grabbed headlines<\/a> as AI systems <a href=\"https:\/\/www.theverge.com\/news\/831760\/openai-google-rate-limit-sora-nano-banana-pro\" target=\"_blank\" rel=\"nofollow noopener\">are increasingly applied to consumer-facing products<\/a> like <a href=\"https:\/\/openai.com\/sora\/\" target=\"_blank\" rel=\"nofollow noopener\">OpenAI\u2019s Sora video-generation app<\/a> and <a href=\"https:\/\/gemini.google\/overview\/image-generation\/\" target=\"_blank\" rel=\"nofollow noopener\">Google\u2019s Nano Banana image-generation model<\/a>.<\/p>\n<p id=\"anchor-53875c\" class=\"body-graf\">However, Wednesday\u2019s report argues that the steady increase in capabilities is severely outpacing any expansion of safety-focused efforts. \u201cThis widening gap between capability and safety leaves the sector structurally unprepared for the risks it is actively creating,\u201d it says.<\/p>\n<p id=\"anchor-94d873\" class=\"endmark body-graf\">This reporter is a Tarbell Fellow, funded through the Tarbell Center for AI Journalism, a nonprofit devoted to supporting the news coverage of artificial intelligence. The Tarbell Center has received funding from the Future of Life Institute, which is a subject of this article. The Tarbell Center had no input in NBC News\u2019 reporting.<\/p>\n","protected":false},"excerpt":{"rendered":"As leading artificial intelligence companies release increasingly capable AI systems, a new report is sounding the alarm about&hellip;\n","protected":false},"author":2,"featured_media":173072,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[20],"tags":[220,218,219,61,60,80],"class_list":{"0":"post-173071","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-ai","9":"tag-artificial-intelligence","10":"tag-artificialintelligence","11":"tag-ie","12":"tag-ireland","13":"tag-technology"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/posts\/173071","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/comments?post=173071"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/posts\/173071\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/media\/173072"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/media?parent=173071"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/categories?post=173071"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/tags?post=173071"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}