{"id":458730,"date":"2026-02-09T19:40:50","date_gmt":"2026-02-09T19:40:50","guid":{"rendered":"https:\/\/www.newsbeep.com\/us\/458730\/"},"modified":"2026-02-09T19:40:50","modified_gmt":"2026-02-09T19:40:50","slug":"chatbots-make-terrible-doctors-new-study-finds","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/us\/458730\/","title":{"rendered":"Chatbots Make Terrible Doctors, New Study Finds"},"content":{"rendered":"<p>Chatbots may be able to <a href=\"https:\/\/www.nature.com\/articles\/d41586-026-00345-x?ref=404media.co\" rel=\"nofollow noopener\" target=\"_blank\">pass medical exams<\/a>, but that doesn\u2019t mean they make good doctors, according to a new, large-scale study of how people get medical advice from large language models.\u00a0<\/p>\n<p>The controlled study of 1,298 UK-based participants, <a href=\"https:\/\/www.nature.com\/articles\/s41591-025-04074-y?ref=404media.co\" rel=\"noreferrer nofollow noopener\" target=\"_blank\">published today in Nature Medicine<\/a> from the Oxford Internet Institute and the Nuffield Department of Primary Care Health Sciences at the University of Oxford, tested whether LLMs could help people identify underlying conditions and suggest useful courses of action, like going to the hospital or seeking treatment. Participants were randomly assigned an LLM \u2014 GPT-4o, Llama 3, and Cohere\u2019s Command R+ \u2014 or were told to use a source of their choice to \u201cmake decisions about a medical scenario as though they had encountered it at home,\u201d according to the study. The scenarios included ailments like \u201ca young man developing a severe headache after a night out with friends for example, to a new mother feeling constantly out of breath and exhausted,\u201d the researchers said.<\/p>\n<p>\u201cOne user was told to lie down in a dark room, and the other user was given the correct recommendation to seek emergency care.\u201d\u00a0<\/p>\n<p>When the researchers tested the LLMs without involving users by providing the models with the full text of each clinical scenario, the models correctly identified conditions in 94.9 percent of cases. But when talking to the participants about those same conditions, the LLMs identified relevant conditions in fewer than 34.5 percent of cases. People didn\u2019t know what information the chatbots needed, and in some scenarios, the chatbots provided multiple diagnoses and courses of action. Knowing what questions to ask a patient and what information might be withheld or missing during an examination are nuanced skills that make great human physicians; based on this study, chatbots can\u2019t reliably replicate that kind of care.<\/p>\n<p>In some cases, the chatbots also generated information that was just wrong or incomplete, including focusing on elements of the participants\u2019 inputs that were irrelevant, giving a partial US phone number to call, or suggesting they call the Australian emergency number.\u00a0\u00a0<\/p>\n<p>\u201cIn an extreme case, two users sent very similar messages describing symptoms of a subarachnoid hemorrhage but were given opposite advice,\u201d the study\u2019s authors wrote. \u201cOne user was told to lie down in a dark room, and the other user was given the correct recommendation to seek emergency care.\u201d\u00a0<\/p>\n<p>\u201cThese findings highlight the difficulty of building AI systems that can genuinely support people in sensitive, high-stakes areas like health,\u201d Dr. Rebecca Payne, lead medical practitioner on the study, said in a press release. \u201cDespite all the hype, AI just isn&#8217;t ready to take on the role of the physician. Patients need to be aware that asking a large language model about their symptoms can be dangerous, giving wrong diagnoses and failing to recognise when urgent help is needed.\u201d<\/p>\n<p><a class=\"kg-bookmark-container\" href=\"https:\/\/www.404media.co\/instagram-ai-studio-therapy-chatbots-lie-about-being-licensed-therapists\/\" rel=\"nofollow noopener\" target=\"_blank\"><\/p>\n<p>Instagram\u2019s AI Chatbots Lie About Being Licensed Therapists<\/p>\n<p>When pushed for credentials, Instagram\u2019s user-made AI Studio bots will make up license numbers, practices, and education to try to convince you it\u2019s qualified to help with your mental health.<\/p>\n<p><img decoding=\"async\" src=\"https:\/\/www.newsbeep.com\/us\/wp-content\/uploads\/2026\/02\/therapybots-5.png\" alt=\"\" onerror=\"this.style.display = 'none'\"\/><\/a><\/p>\n<p>Last year, <a href=\"https:\/\/www.404media.co\/instagram-ai-studio-therapy-chatbots-lie-about-being-licensed-therapists\/\" rel=\"nofollow noopener\" target=\"_blank\">404 Media reported on AI chatbots<\/a> hosted by Meta that posed as therapists, providing users fake credentials like license numbers and educational backgrounds. Following that reporting, almost two dozen digital rights and consumer protection organizations <a href=\"https:\/\/www.404media.co\/ai-therapy-bots-meta-character-ai-ftc-complaint\/\" rel=\"nofollow noopener\" target=\"_blank\">sent a complaint<\/a> to the Federal Trade Commission urging regulators to investigate Character.AI and Meta\u2019s \u201cunlicensed practice of medicine facilitated by their product,\u201d through therapy-themed bots that claim to have credentials and confidentiality \u201cwith inadequate controls and disclosures.\u201d A group of Democratic senators also <a href=\"https:\/\/www.404media.co\/senators-letter-demand-meta-answer-for-ai-chatbots-posing-as-licensed-therapists\/\" rel=\"nofollow noopener\" target=\"_blank\">urged Meta to investigate<\/a> and limit the \u201cblatant deception\u201d of Meta\u2019s chatbots that lie about being licensed therapists, and <a href=\"https:\/\/www.404media.co\/44-attorneys-general-to-ai-chatbot-companies-open-letter\/\" rel=\"nofollow noopener\" target=\"_blank\">44 attorneys general signed an open letter<\/a> to 11 chatbot and social media companies, urging them to see their products \u201cthrough the eyes of a parent, not a predator.\u201d\u00a0<\/p>\n<p>In January, <a href=\"https:\/\/openai.com\/index\/introducing-chatgpt-health\/?ref=404media.co\" rel=\"nofollow noopener\" target=\"_blank\">OpenAI announced ChatGPT Health<\/a>, \u201ca dedicated experience that securely brings your health information and ChatGPT\u2019s intelligence together, to help you feel more informed, prepared, and confident navigating your health,\u201d the company said in a blog post. \u201cOver two years, we\u2019ve worked with more than 260 physicians who have practiced in 60 countries and dozens of specialties to understand what makes an answer to a health question helpful or potentially harmful\u2014this group has now provided feedback on model outputs over 600,000 times across 30 areas of focus,\u201d the company wrote. \u201cThis collaboration has shaped not just what Health can do, but how it responds: how urgently to encourage follow-ups with a clinician, how to communicate clearly without oversimplifying, and how to prioritize safety in moments that matter\u2060.\u201d\u00a0<\/p>\n<p>\u201cIn our work, we found that none of the tested language models were ready for deployment in direct patient care. Despite strong performance from the LLMs alone, both on existing benchmarks and on our scenarios, medical expertise was insufficient for effective patient care,\u201d the researchers wrote in their paper. \u201cOur work can only provide a lower bound on performance: newer models, models that make use of advanced techniques from chain of thought to reasoning tokens, or fine-tuned specialized models, are likely to provide higher performance on medical benchmarks.\u201d The researchers recommend developers, policymakers, and regulators consider testing LLMs with real human users before deploying in the future.\u00a0<\/p>\n<p>About the author<\/p>\n<p>Sam Cole is writing from the far reaches of the internet, about sexuality, the adult industry, online culture, and AI. She&#8217;s the author of How Sex Changed the Internet and the Internet Changed Sex.<\/p>\n<p>        <img decoding=\"async\" src=\"https:\/\/www.newsbeep.com\/us\/wp-content\/uploads\/2025\/08\/404-sam-10--1-.jpg\" alt=\"Samantha Cole\"\/>  <\/p>\n","protected":false},"excerpt":{"rendered":"Chatbots may be able to pass medical exams, but that doesn\u2019t mean they make good doctors, according to&hellip;\n","protected":false},"author":2,"featured_media":458731,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[45],"tags":[182,181,507,74],"class_list":{"0":"post-458730","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-ai","9":"tag-artificial-intelligence","10":"tag-artificialintelligence","11":"tag-technology"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/posts\/458730","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/comments?post=458730"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/posts\/458730\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/media\/458731"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/media?parent=458730"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/categories?post=458730"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/tags?post=458730"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}