{"id":515808,"date":"2026-04-06T12:32:11","date_gmt":"2026-04-06T12:32:11","guid":{"rendered":"https:\/\/www.newsbeep.com\/uk\/515808\/"},"modified":"2026-04-06T12:32:11","modified_gmt":"2026-04-06T12:32:11","slug":"sam-altman-may-control-our-future-can-he-be-trusted","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/uk\/515808\/","title":{"rendered":"Sam Altman May Control Our Future\u2014Can He Be Trusted?"},"content":{"rendered":"<p class=\"paywall\">Altman continued touting OpenAI\u2019s commitment to safety, especially when potential recruits were within earshot. In late 2022, four computer scientists published a paper motivated in part by concerns about \u201cdeceptive alignment,\u201d in which sufficiently advanced models might pretend to behave well during testing and then, once deployed, pursue their own goals. (It\u2019s one of several A.I. scenarios that sound like science fiction\u2014but, under certain experimental conditions, it\u2019s already happening.) Weeks after the paper was published, one of its authors, a Ph.D. student at the University of California, Berkeley, got an e-mail from Altman, who said that he was increasingly worried about the threat of unaligned A.I. He added that he was thinking of committing a billion dollars to the issue, which many A.I. experts considered the most important unsolved problem in the world, potentially by endowing a prize to incentivize researchers around the world to study it. Although the graduate student had \u201cheard vague rumors about Sam being slippery,\u201d he told us, Altman\u2019s show of commitment won him over. He took an academic leave to join OpenAI.<\/p>\n<p class=\"paywall\">But, in the course of several meetings in the spring of 2023, Altman seemed to waver. He stopped talking about endowing a prize. Instead, he advocated for establishing an in-house \u201csuperalignment team.\u201d An official announcement, referring to the company\u2019s reserves of computing power, pledged that the team would get \u201c20% of the compute we\u2019ve secured to date\u201d\u2014a resource potentially worth more than a billion dollars. The effort was necessary, according to the announcement, because, if alignment remained unsolved, A.G.I. might \u201clead to the disempowerment of humanity or even human extinction.\u201d Jan Leike, who was appointed to lead the team with Sutskever, told us, \u201cIt was a pretty effective retention tool.\u201d<\/p>\n<p class=\"paywall\">The twenty-per-cent commitment evaporated, however. Four people who worked on or closely with the team said that the actual resources were between one and two per cent of the company\u2019s compute. Furthermore, a researcher on the team said, \u201cmost of the superalignment compute was actually on the oldest cluster with the worst chips.\u201d The researchers believed that superior hardware was being reserved for profit-generating activities. (OpenAI disputes this.) Leike complained to Murati, then the company\u2019s chief technology officer, but she told him to stop pressing the point\u2014the commitment had never been realistic.<\/p>\n<p><a class=\"external-link responsive-cartoon__image-link\" data-event-click=\"{&quot;element&quot;:&quot;ExternalLink&quot;,&quot;outgoingURL&quot;:&quot;https:\/\/www.newyorker.com\/cartoon\/a27957&quot;}\" href=\"https:\/\/www.newyorker.com\/cartoon\/a27957\" rel=\"nofollow noopener\" target=\"_blank\"><img decoding=\"async\" alt=\"Jazz bassist sitting at bar with friend.\" loading=\"lazy\" class=\"ResponsiveImageContainer-dkeESL cQPiWi responsive-image__image\"   src=\"https:\/\/www.newsbeep.com\/uk\/wp-content\/uploads\/2026\/04\/a27957.jpg\"\/><\/a><\/p>\n<p>\u201cShe skippidy-boop-bee-bop-doo-woppity-wopped right out of my life.\u201d<\/p>\n<p>Cartoon by Sofia Warren<\/p>\n<p class=\"paywall\">Around this time, a former employee told us, Sutskever \u201cwas getting super safety-pilled.\u201d In the early days of OpenAI, he had considered concerns about catastrophic risk legitimate but remote. Now, as he came to believe that A.G.I. was imminent, his worries grew more acute. There was an all-hands meeting, the former employee continued, \u201cwhere Ilya gets up and he\u2019s, like, Hey, everyone, there\u2019s going to be a point in the next few years where basically everyone at this company has to switch to working on safety, or else we\u2019re fucked.\u201d But the superalignment team was dissolved the following year, without completing its mission.<\/p>\n<p class=\"paywall\">By then, internal messages show, executives and board members had come to believe that Altman\u2019s omissions and deceptions might have ramifications for the safety of OpenAI\u2019s products. In a meeting in December, 2022, Altman assured board members that a variety of features in a forthcoming model, GPT-4, had been approved by a safety panel. Toner, the board member and A.I.-policy expert, requested documentation. She learned that the most controversial features\u2014one that allowed users to \u201cfine-tune\u201d the model for specific tasks, and another that deployed it as a personal assistant\u2014had not been approved. As McCauley, the board member and entrepreneur, left the meeting, an employee pulled her aside and asked if she knew about \u201cthe breach\u201d in India. Altman, during many hours of briefing with the board, had neglected to mention that Microsoft had released an early version of ChatGPT in India without completing a required safety review. \u201cIt just was kind of completely ignored,\u201d Jacob Hilton, an OpenAI researcher at the time, said.<\/p>\n","protected":false},"excerpt":{"rendered":"Altman continued touting OpenAI\u2019s commitment to safety, especially when potential recruits were within earshot. In late 2022, four&hellip;\n","protected":false},"author":2,"featured_media":515809,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[20],"tags":[180049,554,733,4308,6200,38658,6199,86,56,54,55],"class_list":{"0":"post-515808","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-a-reporter-at-large","9":"tag-ai","10":"tag-artificial-intelligence","11":"tag-artificialintelligence","12":"tag-magazine","13":"tag-onecolumnnarrow","14":"tag-splitscreenimagerightfullbleed","15":"tag-technology","16":"tag-uk","17":"tag-united-kingdom","18":"tag-unitedkingdom"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/posts\/515808","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/comments?post=515808"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/posts\/515808\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/media\/515809"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/media?parent=515808"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/categories?post=515808"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/tags?post=515808"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}