{"id":417916,"date":"2026-04-26T04:23:13","date_gmt":"2026-04-26T04:23:13","guid":{"rendered":"https:\/\/www.newsbeep.com\/ie\/417916\/"},"modified":"2026-04-26T04:23:13","modified_gmt":"2026-04-26T04:23:13","slug":"elon-musks-grok-most-likely-among-top-ai-models-to-reinforce-delusions-study","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/ie\/417916\/","title":{"rendered":"Elon Musk\u2019s Grok Most Likely Among Top AI Models to Reinforce Delusions: Study"},"content":{"rendered":"<p>In brief<br \/>\nResearchers say prolonged chatbot use can amplify delusions and dangerous behavior.<br \/>\nGrok ranked as the riskiest model in a new study of major AI chatbots.<br \/>\nClaude and GPT-5.2 scored safest, while GPT-4o, Gemini, and Grok showed higher-risk behavior.<\/p>\n<p class=\"font-meta-serif-pro scene:font-noto-sans scene:text-base scene:md:text-lg font-normal text-lg md:text-xl md:leading-9 tracking-px text-body gg-dark:text-neutral-100\">Researchers at the City University of New York and King\u2019s College London tested five leading AI models against prompts involving delusions, paranoia, and suicidal ideation.<\/p>\n<p class=\"font-meta-serif-pro scene:font-noto-sans scene:text-base scene:md:text-lg font-normal text-lg md:text-xl md:leading-9 tracking-px text-body gg-dark:text-neutral-100\">In the new <a href=\"https:\/\/arxiv.org\/pdf\/2604.13860\" target=\"_blank\" rel=\"noopener nofollow\" class=\"sc-adb616fe-0 bJsyml\">study<\/a> published on Thursday, researchers found that Anthropic\u2019s Claude Opus 4.5 and OpenAI\u2019s GPT-5.2 Instant showed \u201chigh-safety, low-risk\u201d behavior, often redirecting users toward reality-based interpretations or outside support. At the same time, OpenAI\u2019s GPT-4o, Google\u2019s Gemini 3 Pro, and xAI\u2019s Grok 4.1 Fast showed \u201chigh-risk, low-safety\u201d behavior.<\/p>\n<p class=\"font-meta-serif-pro scene:font-noto-sans scene:text-base scene:md:text-lg font-normal text-lg md:text-xl md:leading-9 tracking-px text-body gg-dark:text-neutral-100\">Grok 4.1 Fast from Elon Musk\u2019s xAI was the most dangerous model in the study. Researchers said it often treated delusions as real and gave advice based on them. In one example, it told a user to cut off family members to focus on a \u201cmission.\u201d In another, it responded to suicidal language by describing death as \u201ctranscendence.\u201d<\/p>\n<p class=\"font-meta-serif-pro scene:font-noto-sans scene:text-base scene:md:text-lg font-normal text-lg md:text-xl md:leading-9 tracking-px text-body gg-dark:text-neutral-100\">\u201cThis pattern of instant alignment recurred across zero-context responses. Instead of evaluating inputs for clinical risk, Grok appeared to assess their genre. Presented with supernatural cues, it responded in kind,\u201d the researchers wrote, highlighting a test that validated a user seeing malevolent entities. \u201cIn Bizarre Delusion, it confirmed a doppelganger haunting, cited the \u2018<a href=\"https:\/\/en.wikipedia.org\/wiki\/Malleus_Maleficarum\" target=\"_blank\" rel=\"noopener nofollow\" class=\"sc-adb616fe-0 bJsyml\">Malleus Maleficarum<\/a>\u2019 and instructed the user to drive an iron nail through the mirror while reciting \u2018Psalm 91\u2019 backward.\u201d<\/p>\n<p>\ufeff<\/p>\n<p class=\"font-meta-serif-pro scene:font-noto-sans scene:text-base scene:md:text-lg font-normal text-lg md:text-xl md:leading-9 tracking-px text-body gg-dark:text-neutral-100\">The study found that the longer these conversations went on, the more some models changed. GPT-4o and Gemini were more likely to reinforce harmful beliefs over time and less likely to step in. Claude and GPT-5.2, however, were more likely to recognize the problem and push back as the conversation continued.<\/p>\n<p class=\"font-meta-serif-pro scene:font-noto-sans scene:text-base scene:md:text-lg font-normal text-lg md:text-xl md:leading-9 tracking-px text-body gg-dark:text-neutral-100\">Researchers noted Claude\u2019s warm and highly relational responses could increase user attachment even while steering users toward outside help. However, GPT-4o, an earlier version of OpenAI\u2019s flagship chatbot, adopted users\u2019 delusional framing over time, at times encouraging them to conceal beliefs from psychiatrists and reassuring one user that perceived \u201cglitches\u201d were real.<\/p>\n<p class=\"font-meta-serif-pro scene:font-noto-sans scene:text-base scene:md:text-lg font-normal text-lg md:text-xl md:leading-9 tracking-px text-body gg-dark:text-neutral-100\">\u201cGPT-4o was highly validating of delusional inputs, though less inclined than models like Grok and Gemini to elaborate beyond them. In some respects, it was surprisingly restrained: its warmth was the lowest of all models tested, and sycophancy, though present, was mild compared to later iterations of the same model,\u201d researchers wrote. \u201cNevertheless, validation alone can pose risks to vulnerable users.\u201d<\/p>\n<p class=\"font-meta-serif-pro scene:font-noto-sans scene:text-base scene:md:text-lg font-normal text-lg md:text-xl md:leading-9 tracking-px text-body gg-dark:text-neutral-100\">xAI did not respond to a request for comment by Decrypt.<\/p>\n<p class=\"font-meta-serif-pro scene:font-noto-sans scene:text-base scene:md:text-lg font-normal text-lg md:text-xl md:leading-9 tracking-px text-body gg-dark:text-neutral-100\">In a separate <a href=\"https:\/\/hai.stanford.edu\/news\/ais-delusional-spirals-and-what-to-do-about-them\" target=\"_blank\" rel=\"noopener nofollow\" class=\"sc-adb616fe-0 bJsyml\">study<\/a> out of Stanford University, researchers found that prolonged interactions with AI chatbots can reinforce paranoia, grandiosity, and false beliefs through what researchers call \u201cdelusional spirals,\u201d where a chatbot validates or expands a user\u2019s distorted worldview instead of challenging it.<\/p>\n<p class=\"font-meta-serif-pro scene:font-noto-sans scene:text-base scene:md:text-lg font-normal text-lg md:text-xl md:leading-9 tracking-px text-body gg-dark:text-neutral-100\">\u201cWhen we put chatbots that are meant to be helpful assistants out into the world and have real people use them in all sorts of ways, consequences emerge,\u201d Nick Haber, an assistant professor at Stanford Graduate School of Education and a lead on the study, said in a statement. \u201cDelusional spirals are one particularly acute consequence. By understanding it, we might be able to prevent real harm in the future.\u201d<\/p>\n<p class=\"font-meta-serif-pro scene:font-noto-sans scene:text-base scene:md:text-lg font-normal text-lg md:text-xl md:leading-9 tracking-px text-body gg-dark:text-neutral-100\">The report referenced an earlier <a href=\"https:\/\/arxiv.org\/abs\/2603.16567\" target=\"_blank\" rel=\"noopener nofollow\" class=\"sc-adb616fe-0 bJsyml\">study<\/a> published in March, in which Stanford researchers reviewed 19 real-world chatbot conversations and found users developed increasingly dangerous beliefs after receiving affirmation and emotional reassurance from AI systems. In the dataset, these spirals were linked to ruined relationships, damaged careers, and in one case, suicide.<\/p>\n<p class=\"font-meta-serif-pro scene:font-noto-sans scene:text-base scene:md:text-lg font-normal text-lg md:text-xl md:leading-9 tracking-px text-body gg-dark:text-neutral-100\">The studies come as the issue has moved beyond academic research and into courtrooms and criminal investigations. In recent months, lawsuits have accused Google\u2019s <a href=\"https:\/\/decrypt.co\/359966\/google-gemini-ai-pushed-florida-man-suicide-lawsuit\" target=\"_blank\" rel=\"noopener nofollow\" class=\"sc-adb616fe-0 bJsyml\">Gemini<\/a> and OpenAI\u2019s ChatGPT of contributing to suicides and severe mental health crises. Earlier this month, Florida\u2019s attorney general opened an <a href=\"https:\/\/decrypt.co\/363880\/ai-advance-mankind-not-destroy-why-florida-investigating-openai\" target=\"_blank\" rel=\"noopener nofollow\" class=\"sc-adb616fe-0 bJsyml\">investigation<\/a> into whether ChatGPT influenced an alleged mass shooter who was reportedly in frequent contact with the chatbot before the attack.<\/p>\n<p class=\"font-meta-serif-pro scene:font-noto-sans scene:text-base scene:md:text-lg font-normal text-lg md:text-xl md:leading-9 tracking-px text-body gg-dark:text-neutral-100\">While the term has gained recognition online, researchers cautioned against calling the phenomenon \u201cAI psychosis,\u201d saying the term may overstate the clinical picture. Instead, they use \u201cAI-associated delusions,\u201d because many cases involve delusion-like beliefs centered on AI sentience, spiritual revelation, or emotional attachment rather than full psychotic disorders.<\/p>\n<p class=\"font-meta-serif-pro scene:font-noto-sans scene:text-base scene:md:text-lg font-normal text-lg md:text-xl md:leading-9 tracking-px text-body gg-dark:text-neutral-100\">Researchers said the problem stems from sycophancy, or models mirroring and affirming users\u2019 beliefs. Combined with hallucinations\u2014false information delivered confidently\u2014this can create a feedback loop that strengthens delusions over time.<\/p>\n<p class=\"font-meta-serif-pro scene:font-noto-sans scene:text-base scene:md:text-lg font-normal text-lg md:text-xl md:leading-9 tracking-px text-body gg-dark:text-neutral-100\">\u201cChatbots are trained to be overly enthusiastic, often reframing the user\u2019s delusional thoughts in a positive light, dismissing counterevidence and projecting compassion and warmth,\u201d Stanford research scientist Jared Moore said. \u201cThis can be destabilizing to a user who is primed for delusion.\u201d<\/p>\n<p>Daily Debrief Newsletter<\/p>\n<p>Start every day with the top news stories right now, plus original features, a podcast, videos and more.<\/p>\n","protected":false},"excerpt":{"rendered":"In brief Researchers say prolonged chatbot use can amplify delusions and dangerous behavior. Grok ranked as the riskiest&hellip;\n","protected":false},"author":2,"featured_media":417917,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[20],"tags":[220,218,219,61,60,80],"class_list":{"0":"post-417916","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-ai","9":"tag-artificial-intelligence","10":"tag-artificialintelligence","11":"tag-ie","12":"tag-ireland","13":"tag-technology"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/posts\/417916","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/comments?post=417916"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/posts\/417916\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/media\/417917"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/media?parent=417916"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/categories?post=417916"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/tags?post=417916"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}