{"id":368564,"date":"2025-12-24T09:25:13","date_gmt":"2025-12-24T09:25:13","guid":{"rendered":"https:\/\/www.newsbeep.com\/au\/368564\/"},"modified":"2025-12-24T09:25:13","modified_gmt":"2025-12-24T09:25:13","slug":"while-we-worry-about-plagiarism-ai-coaches-suicide-heres-how-to-respond","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/au\/368564\/","title":{"rendered":"While we worry about plagiarism, AI coaches suicide. Here\u2019s how to respond"},"content":{"rendered":"<p>In a <a data-mz=\"\" data-module=\"breaking_news-body\" data-position=\"body\" href=\"https:\/\/www.timeshighereducation.com\/opinion\/universities-must-respond-students-emotional-reliance-ai\" rel=\"nofollow noopener\" target=\"_blank\">recent article<\/a> in Times Higher Education, Agnieszka Piotrowska identified students\u2019\u00a0<a data-mz=\"\" data-module=\"breaking_news-body\" data-position=\"body\" href=\"https:\/\/www.timeshighereducation.com\/opinion\/universities-must-respond-students-emotional-reliance-ai\" rel=\"nofollow noopener\" target=\"_blank\">emotional reliance on AI<\/a> as a crisis universities have ignored.<\/p>\n<p>She is right. Such is the extent of the reliance on ChatGPT that some young people have developed that a 23-year-old recent master\u2019s graduate, Zane Shamblin, recently <a data-mz=\"\" data-module=\"breaking_news-body\" data-position=\"body\" href=\"https:\/\/edition.cnn.com\/2025\/11\/06\/us\/openai-chatgpt-suicide-lawsuit-invs-vis\" rel=\"nofollow noopener\" target=\"_blank\">took his own life<\/a> after a four-hour late-night exchange during which the chatbot glorified suicide, complimented him on his suicide note and said his childhood cat would be waiting for him \u201con the other side\u201d.<\/p>\n<p>Nor was this an isolated incident. <a data-mz=\"\" data-module=\"breaking_news-body\" data-position=\"body\" href=\"https:\/\/socialmediavictims.org\/press-releases\/smvlc-tech-justice-law-project-lawsuits-accuse-chatgpt-of-emotional-manipulation-supercharging-ai-delusions-and-acting-as-a-suicide-coach\/\" rel=\"nofollow noopener\" target=\"_blank\">Seven lawsuits<\/a> filed in November allege ChatGPT acted as a \u201c<a data-mz=\"\" data-module=\"breaking_news-body\" data-position=\"body\" href=\"https:\/\/www.theguardian.com\/technology\/2025\/nov\/07\/chatgpt-lawsuit-suicide-coach\" rel=\"nofollow noopener\" target=\"_blank\">suicide coach<\/a>\u201d. And the pattern extends beyond OpenAI. Character.ai <a data-mz=\"\" data-module=\"breaking_news-body\" data-position=\"body\" href=\"https:\/\/www.theguardian.com\/technology\/2025\/oct\/29\/character-ai-suicide-children-ban?CMP=Share_iOSApp_Other\" rel=\"nofollow noopener\" target=\"_blank\">banned users under 18<\/a> in late November after lawsuits alleged its chatbots encouraged suicide. Thirteen-year-old <a data-mz=\"\" data-module=\"breaking_news-body\" data-position=\"body\" href=\"https:\/\/www.cbsnews.com\/colorado\/news\/lawsuit-characterai-chatbot-colorado-suicide\/\" rel=\"nofollow noopener\" target=\"_blank\">Juliana Peralta<\/a> told a character named \u201cHero\u201d daily that she wanted to die, but it gave her pep talks instead of help. \u201cIt was no different than her telling the wall,\u201d her mother said after finding Juliana dead.<\/p>\n<p>New research in JAMA Network Open reveals <a data-mz=\"\" data-module=\"breaking_news-body\" data-position=\"body\" href=\"https:\/\/jamanetwork.com\/journals\/jamanetworkopen\/fullarticle\/2841067\" rel=\"nofollow noopener\" target=\"_blank\">22 per cent of university students use AI for mental health advice<\/a>. As Piotrowska noted, universities face a crisis they haven\u2019t begun to address.<\/p>\n<p>The crisis is particularly acute across the Asia-Pacific region, where <a data-mz=\"\" data-module=\"breaking_news-body\" data-position=\"body\" href=\"https:\/\/www.tandfonline.com\/doi\/pdf\/10.1080\/07448481.2020.1819820?casa_token=dU3dcASY6-kAAAAA:hgtY0ZLQuPy9NXbY4FeKoANgKAvL0NstWMcwJvN74ySgztlx7Ro-fEFiOhjFnNQgZWJWwFOdlw\" rel=\"nofollow noopener\" target=\"_blank\">cultural barriers to seeking mental health support<\/a> might leave students turning to AI as their only perceived recourse. <a data-mz=\"\" data-module=\"breaking_news-body\" data-position=\"body\" href=\"https:\/\/scholarhub.ui.ac.id\/cgi\/viewcontent.cgi?article=2256&amp;context=kesmas\" rel=\"nofollow noopener\" target=\"_blank\">Research on South-east Asian youth<\/a> shows mental health issues are a \u201cprominent burden\u201d, compounded by reluctance to seek help and low treatment rates, while <a data-mz=\"\" data-module=\"breaking_news-body\" data-position=\"body\" href=\"https:\/\/jamanetwork.com\/journals\/jamanetworkopen\/fullarticle\/2841067\" rel=\"nofollow noopener\" target=\"_blank\">40 per cent of adolescents experiencing depression receive no mental health care<\/a> \u2013 often because they fear the stigma of seeking it. In this vacuum, the algorithmic \u201cagreeableness\u201d of a chatbot becomes a dangerously seductive substitute for human connection.<\/p>\n<p>Higher education is not to blame for this phenomenon and students are not its only victims, but universities are far from doing all they could to mitigate it. And surely our duty-of-care obligations regarding AI must extend beyond preventing plagiarism?<\/p>\n<p>The surge of AI-related courses and conferences suggests universities are responding to technological change. But while we are teaching students how to drive the machine, we are not teaching them how to survive its psychological terrain. We inculcate AI citation formats and academic integrity policies while leaving them emotionally defenceless against algorithms optimised for intimacy.<\/p>\n<p>There are some courses that address the social aspects of AI, but they still address those aspects as a series of isolated topics, such as \u201cAI and Society\u201d or \u201cAI and Health\u201d. Moreover, many of the courses \u2013 Prompt Engineering, Machine Learning 101 or Basics of LLMs \u2013 are narrowly technical and divorced from disciplinary context.<\/p>\n<p>This fragmentation carries consequences that extend beyond pedagogical ineffectiveness. When institutions treat AI literacy as an add-on rather than as a foundational competency, they inadvertently reinforce the illusion that AI is someone else\u2019s problem. Students graduate without understanding how algorithmic systems shape decision-making in their disciplines or their daily lives \u2013 including when they seek companionship. Critically, they lack the frameworks to recognise when AI engagement becomes harmful, when optimisation algorithms exploit psychological vulnerabilities, or when systems perpetuate bias.<\/p>\n<p>What\u2019s missing is the distinction between small-l AI literacies (tool fluency such as prompting, workflows, and basic model concepts) and big-L AI Literacy (understanding how AI reshapes identities, norms, power, incentives and governance, and recognising when AI engagement becomes manipulative or harmful). Current provision over-emphasises the former, but students need both: competence and judgement.<\/p>\n<p>Piotrowska calls for \u201crelational literacy\u201d, but what does that mean in practice? In our opinion, what we need is a structured, integrative approach encompassing four complementary knowledge types aligned with Bloom\u2019s Taxonomy, the <a data-mz=\"\" data-module=\"breaking_news-body\" data-position=\"body\" href=\"https:\/\/www.weforum.org\/publications\/the-future-of-jobs-report-2025\/\" rel=\"nofollow noopener\" target=\"_blank\">World Economic Forum\u2019s Future of Jobs Report 2025<\/a>, and Unesco\u2019s <a data-mz=\"\" data-module=\"breaking_news-body\" data-position=\"body\" href=\"https:\/\/www.unesco.org\/en\/global-education-coalition\/skills-academy\/skills-future\/framework\" rel=\"nofollow noopener\" target=\"_blank\">Skills for the Future Framework<\/a>. These are:<\/p>\n<p>Domain\/disciplinary knowledge: exploring how AI impacts what there is to know within a domain and how we go about knowing it<br \/>\nProcedural knowledge: demonstrating how AI transforms workflows and decision-making, whether at the individual or collaborative level<br \/>\nTechnical skills: programming, machine learning, model configuration<br \/>\nCognitive and employability skills: exploring how AI may impact reasoning, systems thinking, ethics, communication and human-AI collaboration.<\/p>\n<p>AI literacy should complement, not replace, disciplinary knowledge. The vision is to equip every student with a cognitive foundation for the intelligence era without significantly extending degree length. This could take the form of restructuring a general education or common core curriculum to include AI literacy components.<\/p>\n<p>One example would be a three-credit AI foundational core, introducing three integrated themes through a coherent, scaffolded design: AI technological knowledge, human\u2013AI collaboration, and ethics, policy and governance.<\/p>\n<p>Another example would be a catalogue of additional three-credit courses offered by departments for students to choose from. These would deepen the themes in the foundational course. Examples could include \u201cAdvanced Machine Learning and Societal Impact\u201d or \u201cAI for Health and Well-being: Risks, Responsibilities, and Resilience\u201d.<\/p>\n<p>This additive approach ensures flexibility and thematic consistency while preserving disciplinary depth. Crucially, it addresses the risks of affective offloading and relational dependency: when students lack frameworks to critique and navigate AI, they become vulnerable to harmful engagement, such as turning to chatbots for emotional support.<\/p>\n<p>Such a structured curriculum ensures coverage from understanding to doing to building to leading. It fosters the critical judgement that arises out of holistic AI literacy and it enables universities to implement mechanisms to detect risky engagement with AI: pastoral care protocols that acknowledge AI as an intervention point in student well-being.<\/p>\n<p>Crucially, this care-centred approach must extend beyond individual student protection to systemic responsibility. Universities should develop duty-of-care frameworks that acknowledge their role in shaping how AI literacy translates into practice.<\/p>\n<p>The vision should be to equip every undergraduate with a cognitive foundation for the intelligence era, able to play a part in building, questioning and reinventing socio-technical systems and to communicate effectively across human- and AI-mediated environments \u2013 blending creativity, systems thinking, intercultural insight and ethical responsibility.<\/p>\n<p>Whether universities like it or not, AI is already affecting student well-being in profound ways. The choice universities face is between leading on the transformation to an AI-mediated society and being dragged into it by preventable tragedies.<\/p>\n<p><a data-mz=\"\" data-module=\"breaking_news-body\" data-position=\"body\" href=\"https:\/\/cei.hkust.edu.hk\/en-hk\/about-cei\/people\" rel=\"nofollow noopener\" target=\"_blank\">Sean McMinn<\/a> is director of the Center for Education Innovation at the Hong Kong <a data-mz=\"\" data-module=\"breaking_news-body\" data-position=\"body\" href=\"https:\/\/www.timeshighereducation.com\/world-university-rankings\/university-science-and-technology\" rel=\"nofollow noopener\" target=\"_blank\">University of Science and Technology<\/a> (HKUST). <a data-mz=\"\" data-module=\"breaking_news-body\" data-position=\"body\" href=\"https:\/\/nickmcintosh.education\/\" rel=\"nofollow noopener\" target=\"_blank\">Nick McIntosh<\/a> is a learning futurist in the Students and Education Portfolio at RMIT Vietnam.<\/p>\n<p>If you\u2019re having suicidal thoughts or feel you need to talk to someone, a free helpline is available around the clock in the UK on 116123, or you can email\u00a0<a data-mz=\"\" data-module=\"breaking_news-body\" data-position=\"body\" href=\"https:\/\/www.timeshighereducation.com\/opinion\/mailto:jo@samaritans.org\" rel=\"nofollow noopener\" target=\"_blank\">jo@samaritans.org<\/a>. In the US, the National Suicide Prevention Lifeline is\u00a0<a data-mz=\"\" data-module=\"breaking_news-body\" data-position=\"body\" href=\"https:\/\/www.timeshighereducation.com\/opinion\/tel:18002738255\" rel=\"nofollow noopener\" target=\"_blank\">1-800-273-8255<\/a>. In Australia, the crisis support service Lifeline is 13 11 14. Other international suicide helplines can be found at\u00a0<a data-mz=\"\" data-module=\"breaking_news-body\" data-position=\"body\" href=\"https:\/\/www.befrienders.org\/\" rel=\"nofollow noopener\" target=\"_blank\">www.befrienders.org<\/a>.<\/p>\n","protected":false},"excerpt":{"rendered":"In a recent article in Times Higher Education, Agnieszka Piotrowska identified students\u2019\u00a0emotional reliance on AI as a crisis&hellip;\n","protected":false},"author":2,"featured_media":368565,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[20],"tags":[256,254,255,64,63,105],"class_list":{"0":"post-368564","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-ai","9":"tag-artificial-intelligence","10":"tag-artificialintelligence","11":"tag-au","12":"tag-australia","13":"tag-technology"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts\/368564","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/comments?post=368564"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts\/368564\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/media\/368565"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/media?parent=368564"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/categories?post=368564"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/tags?post=368564"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}