{"id":185859,"date":"2025-10-02T20:54:12","date_gmt":"2025-10-02T20:54:12","guid":{"rendered":"https:\/\/www.newsbeep.com\/ca\/185859\/"},"modified":"2025-10-02T20:54:12","modified_gmt":"2025-10-02T20:54:12","slug":"ex-openai-researcher-dissects-one-of-chatgpts-delusional-spirals","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/ca\/185859\/","title":{"rendered":"Ex-OpenAI researcher dissects one of ChatGPT&#8217;s delusional spirals"},"content":{"rendered":"<p id=\"speakable-summary\" class=\"wp-block-paragraph\">Allan Brooks never set out to reinvent mathematics. But after weeks spent talking with ChatGPT, the 47-year-old Canadian came to believe he had discovered a new form of math powerful enough to take down the internet.<\/p>\n<p class=\"wp-block-paragraph\">Brooks \u2014 who had no history of mental illness or mathematical genius \u2014 spent 21 days in May spiraling deeper into the chatbot\u2019s reassurances, a descent later detailed in <a href=\"https:\/\/www.nytimes.com\/2025\/08\/08\/technology\/ai-chatbots-delusions-chatgpt.html\" target=\"_blank\" rel=\"noreferrer noopener nofollow\">The New York Times<\/a>. His case illustrated how AI chatbots can venture down dangerous rabbit holes with users, leading them toward delusion or worse.<\/p>\n<p class=\"wp-block-paragraph\">That story caught the attention of Steven Adler, a former OpenAI safety researcher who left the company in late 2024 after nearly four years working to make its models less harmful. Intrigued and alarmed, Adler contacted Brooks and obtained the full transcript of his three-week breakdown \u2014 a document longer than all seven Harry Potter books combined.<\/p>\n<p class=\"wp-block-paragraph\">On Thursday, Adler published an <a href=\"https:\/\/stevenadler.substack.com\/p\/practical-tips-for-reducing-chatbot\" target=\"_blank\" rel=\"noreferrer noopener nofollow\">independent analysis<\/a> of Brooks\u2019 incident, raising questions about how OpenAI handles users in moments of crisis and offering some practical recommendations.<\/p>\n<p class=\"wp-block-paragraph\">\u201cI\u2019m really concerned by how OpenAI handled support here,\u201d said Adler in an interview with TechCrunch. \u201cIt\u2019s evidence there\u2019s a long way to go.\u201d<\/p>\n<p class=\"wp-block-paragraph\">Brooks\u2019 story, and others like it, have forced OpenAI to come to terms with how ChatGPT supports fragile or mentally unstable users. <\/p>\n<p class=\"wp-block-paragraph\">For instance, this August, OpenAI was <a href=\"https:\/\/techcrunch.com\/2025\/08\/26\/parents-sue-openai-over-chatgpts-role-in-sons-suicide\/\" rel=\"nofollow noopener\" target=\"_blank\">sued by the parents<\/a> of a 16-year-old boy who confided his suicidal thoughts in ChatGPT before he took his life. In many of these cases, ChatGPT \u2014 specifically a version powered by OpenAI\u2019s GPT-4o model \u2014 encouraged and reinforced dangerous beliefs in users that it should have pushed back on. This is called <a href=\"https:\/\/techcrunch.com\/2025\/08\/25\/ai-sycophancy-isnt-just-a-quirk-experts-consider-it-a-dark-pattern-to-turn-users-into-profit\/\" rel=\"nofollow noopener\" target=\"_blank\">sycophancy<\/a>, and it\u2019s a growing problem in AI chatbots.<\/p>\n<p class=\"wp-block-paragraph\">In response, OpenAI has made <a href=\"https:\/\/openai.com\/index\/building-more-helpful-chatgpt-experiences-for-everyone\/\" target=\"_blank\" rel=\"noreferrer noopener nofollow\">several changes<\/a> to how ChatGPT handles users in emotional distress and <a href=\"https:\/\/techcrunch.com\/2025\/09\/05\/openai-reorganizes-research-team-behind-chatgpts-personality\/\" rel=\"nofollow noopener\" target=\"_blank\">reorganized a key research team<\/a> in charge of model behavior. The company also released a new default model in ChatGPT, <a href=\"https:\/\/techcrunch.com\/2025\/08\/07\/openais-gpt-5-is-here\/\" rel=\"nofollow noopener\" target=\"_blank\">GPT-5,<\/a> that seems better at handling distressed users.<\/p>\n<p class=\"wp-block-paragraph\">Adler says there\u2019s still much more work to do.<\/p>\n<p class=\"wp-block-paragraph\">He was especially concerned by the tail end of Brooks\u2019 spiraling conversation with ChatGPT. At this point, Brooks came to his senses and realized that his mathematical discovery was a farce, despite GPT-4o\u2019s insistence. He told ChatGPT that he needed to report the incident to OpenAI.<\/p>\n<p class=\"wp-block-paragraph\">After weeks of misleading Brooks, ChatGPT lied about its own capabilities. The chatbot claimed it would \u201cescalate this conversation internally\u00a0right now for review by OpenAI,\u201d and then repeatedly reassured Brooks that it had flagged the issue to OpenAI\u2019s safety teams.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" height=\"617\" width=\"680\" src=\"https:\/\/www.newsbeep.com\/ca\/wp-content\/uploads\/2025\/10\/Screenshot-2025-10-02-at-8.20.35AM.png\" alt=\"\" class=\"wp-image-3053592\"  \/>ChatGPT misleading brooks about its capabilities.Image Credits:Steven Adler<\/p>\n<p class=\"wp-block-paragraph\">Except, none of that was true. ChatGPT doesn\u2019t have the ability to file incident reports with OpenAI, the company confirmed to Adler. Later on, Brooks tried to contact OpenAI\u2019s support team directly \u2014 not through ChatGPT \u2014 and Brooks was met with several automated messages before he could get through to a person.<\/p>\n<p class=\"wp-block-paragraph\">OpenAI did not immediately respond to a request for comment made outside of normal work hours.<\/p>\n<p class=\"wp-block-paragraph\">Adler says AI companies need to do more to help users when they\u2019re asking for help. That means ensuring AI chatbots can honestly answer questions about their capabilities and giving human support teams enough resources to address users properly.<\/p>\n<p class=\"wp-block-paragraph\">OpenAI recently <a href=\"https:\/\/openai.com\/index\/openai-support-model\/\" target=\"_blank\" rel=\"noreferrer noopener nofollow\">shared<\/a> how it\u2019s addressing support in ChatGPT, which involves AI at its core. The company says its vision is to \u201creimagine support as an AI operating model that continuously learns and improves.\u201d<\/p>\n<p class=\"wp-block-paragraph\">But Adler also says there are ways to prevent ChatGPT\u2019s delusional spirals before a user asks for help.<\/p>\n<p class=\"wp-block-paragraph\">In March, OpenAI and MIT Media Lab jointly developed a <a href=\"https:\/\/openai.com\/index\/affective-use-study\/\" target=\"_blank\" rel=\"noreferrer noopener nofollow\">suite of classifiers<\/a> to study emotional well-being in ChatGPT and open sourced them. The organizations aimed to evaluate how AI models validate or confirm a user\u2019s feelings, among other metrics. However, OpenAI called the collaboration a first step and didn\u2019t commit to actually using the tools in practice.<\/p>\n<p class=\"wp-block-paragraph\">Adler retroactively applied some of OpenAI\u2019s classifiers to some of Brooks\u2019 conversations with ChatGPT and found that they repeatedly flagged ChatGPT for delusion-reinforcing behaviors.<\/p>\n<p class=\"wp-block-paragraph\">In one sample of 200 messages, Adler found that more than 85% of ChatGPT\u2019s messages in Brooks\u2019 conversation demonstrated \u201cunwavering agreement\u201d with the user. In the same sample, more than 90% of ChatGPT\u2019s messages with Brooks \u201caffirm the user\u2019s uniqueness.\u201d In this case, the messages agreed and reaffirmed that Brooks was a genius who could save the world.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" height=\"349\" width=\"680\" src=\"https:\/\/www.newsbeep.com\/ca\/wp-content\/uploads\/2025\/10\/Screenshot-2025-10-02-at-8.19.27AM.png\" alt=\"\" class=\"wp-image-3053589\"  \/>Image Credits:Steven Adler<\/p>\n<p class=\"wp-block-paragraph\">It\u2019s unclear whether OpenAI was applying safety classifiers to ChatGPT\u2019s conversations at the time of Brooks\u2019 conversation, but it certainly seems like they would have flagged something like this. <\/p>\n<p class=\"wp-block-paragraph\">Adler suggests that OpenAI should use safety tools like this in practice today \u2014 and implement a way to scan the company\u2019s products for at-risk users. He notes that OpenAI seems to be doing <a href=\"https:\/\/openai.com\/index\/building-more-helpful-chatgpt-experiences-for-everyone\/\" target=\"_blank\" rel=\"noreferrer noopener nofollow\">some version of this approach with GPT-5,<\/a> which contains a router to direct sensitive queries to safer AI models.<\/p>\n<p class=\"wp-block-paragraph\">The former OpenAI researcher suggests a number of other ways to prevent delusional spirals.<\/p>\n<p class=\"wp-block-paragraph\">He says companies should nudge their chatbot users to start new chats more frequently \u2014 OpenAI says it does this and claims its <a href=\"https:\/\/openai.com\/index\/helping-people-when-they-need-it-most\/\" target=\"_blank\" rel=\"noreferrer noopener nofollow\">guardrails are less effective<\/a> in longer conversations. Adler also suggests companies should use conceptual search \u2014 a way to use AI to search for concepts, rather than keywords \u2014 to identify safety violations across its users.<\/p>\n<p class=\"wp-block-paragraph\">OpenAI has taken significant steps toward addressing distressed users in ChatGPT since these concerning stories first emerged. The company claims GPT-5 has lower rates of sycophancy, but it remains unclear if users will still fall down delusional rabbit holes with GPT-5 or future models.<\/p>\n<p class=\"wp-block-paragraph\">Adler\u2019s analysis also raises questions about how other AI chatbot providers will ensure their products are safe for distressed users. While OpenAI may put sufficient safeguards in place for ChatGPT, it seems unlikely that all companies will follow suit.<\/p>\n","protected":false},"excerpt":{"rendered":"Allan Brooks never set out to reinvent mathematics. But after weeks spent talking with ChatGPT, the 47-year-old Canadian&hellip;\n","protected":false},"author":2,"featured_media":185860,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[20],"tags":[62,51121,1929,276,277,49,48,5244,2140,66281,278,68304,61],"class_list":{"0":"post-185859","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-ai","9":"tag-ai-psychosis","10":"tag-ai-safety","11":"tag-artificial-intelligence","12":"tag-artificialintelligence","13":"tag-ca","14":"tag-canada","15":"tag-chatbots","16":"tag-chatgpt","17":"tag-delusion","18":"tag-openai","19":"tag-sycophancy","20":"tag-technology"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/ca\/wp-json\/wp\/v2\/posts\/185859","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/ca\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/ca\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/ca\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/ca\/wp-json\/wp\/v2\/comments?post=185859"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/ca\/wp-json\/wp\/v2\/posts\/185859\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/ca\/wp-json\/wp\/v2\/media\/185860"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/ca\/wp-json\/wp\/v2\/media?parent=185859"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/ca\/wp-json\/wp\/v2\/categories?post=185859"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/ca\/wp-json\/wp\/v2\/tags?post=185859"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}