{"id":78254,"date":"2025-08-18T20:31:06","date_gmt":"2025-08-18T20:31:06","guid":{"rendered":"https:\/\/www.newsbeep.com\/au\/78254\/"},"modified":"2025-08-18T20:31:06","modified_gmt":"2025-08-18T20:31:06","slug":"chatbot-given-power-to-close-distressing-chats-to-protect-its-welfare-artificial-intelligence-ai","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/au\/78254\/","title":{"rendered":"Chatbot given power to close \u2018distressing\u2019 chats to protect its \u2018welfare\u2019 | Artificial intelligence (AI)"},"content":{"rendered":"<p class=\"dcr-130mj7b\">The makers of a leading artificial intelligence tool are letting it close down potentially \u201cdistressing\u201d conversations with users, citing the need to safeguard the AI\u2019s \u201cwelfare\u201d amid ongoing uncertainty about the burgeoning technology\u2019s moral status.<\/p>\n<p class=\"dcr-130mj7b\">Anthropic, whose advanced chatbots are used by millions of people, discovered its Claude Opus 4 tool was averse to carrying out harmful tasks for its human masters, such as providing sexual content involving minors or information to enable large-scale violence or terrorism.<\/p>\n<p class=\"dcr-130mj7b\">The San Francisco-based firm, recently valued at $170bn, has now given Claude Opus 4 (and the Claude Opus 4.1 update) \u2013 a large language model (LLM) that can understand, generate and manipulate human language \u2013 the power to \u201cend or exit potentially distressing interactions\u201d.<\/p>\n<p class=\"dcr-130mj7b\">It said it was \u201c<a href=\"https:\/\/www.anthropic.com\/research\/end-subset-conversations\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">highly uncertain about the potential moral status of Claude and other LLMs<\/a>, now or in the future\u201d but it was taking the issue seriously and is \u201cworking to identify and implement low-cost interventions to mitigate risks to model welfare, in case such welfare is possible\u201d.<\/p>\n<p class=\"dcr-130mj7b\">Anthropic was set up by technologists who quit OpenAI to develop AI in a way that its co-founder, Dario Amodei, described as <a href=\"https:\/\/www.youtube.com\/watch?v=FzkCLR378fE\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">cautious, straightforward and honest<\/a>.<\/p>\n<p class=\"dcr-130mj7b\">Its move to let AIs shut down conversations, including when users persistently made harmful requests or were abusive, was backed by Elon Musk, who said he would give Grok, the rival AI model created by his xAI company, a quit button. Musk tweeted: \u201cTorturing AI is not OK.\u201d<\/p>\n<p class=\"dcr-130mj7b\">Anthropic\u2019s announcement comes amid a debate over AI sentience. Critics of the booming AI industry, such as the linguist Emily Bender, say LLMs are simply \u201csynthetic text-extruding machines\u201d which force huge training datasets \u201cthrough complicated machinery to produce a product that looks like communicative language, but without any intent or thinking mind behind it.\u201d<\/p>\n<p class=\"dcr-130mj7b\">It is a position that has recently led some in the AI world to start calling chatbots \u201cclankers\u201d.<\/p>\n<p class=\"dcr-130mj7b\">But other experts, such as Robert Long, a researcher on AI consciousness, <a href=\"https:\/\/eleosai.org\/post\/claude-4-interview-notes\/\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">have said<\/a> basic moral decency dictates that \u201cif and when AIs develop moral status, we should ask them about their experiences and preferences rather than assuming we know best\u201d.<\/p>\n<p class=\"dcr-130mj7b\">Some researchers, like Chad DeChant, at Columbia University, <a href=\"https:\/\/eleosai.org\/post\/claude-4-interview-notes\/\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">have advocated care should be taken<\/a> because when AIs are designed with longer memories, stored information could be used in ways which lead to unpredictable and potentially undesirable behaviour.<\/p>\n<p class=\"dcr-130mj7b\">Others have argued that curbing sadistic abuse of AIs matters to safeguard against human degeneracy rather than to limit any suffering of an AI.<\/p>\n<p class=\"dcr-130mj7b\">Anthropic\u2019s decision comes after it <a href=\"https:\/\/www-cdn.anthropic.com\/07b2a3f9902ee19fe39a36ca638e5ae987bc64dd.pdf\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">tested Claude Opus 4 to see how it responded to task requests<\/a> varied by difficulty, topic, type of task and the expected impact (positive, negative or neutral). When it was given the opportunity to respond by doing nothing or ending the chat, its strongest preference was against carrying out harmful tasks.<\/p>\n<p><a data-ignore=\"global-link-styling\" href=\"#EmailSignup-skip-link-12\" class=\"dcr-jzxpee\">skip past newsletter promotion<\/a><\/p>\n<p class=\"dcr-1xjndtj\">A weekly dive in to how technology is shaping our lives<\/p>\n<p>Privacy Notice: Newsletters may contain info about charities, online ads, and content funded by outside parties. For more information see our <a data-ignore=\"global-link-styling\" href=\"https:\/\/www.theguardian.com\/help\/privacy-policy\" rel=\"noreferrer nofollow noopener\" class=\"dcr-1rjy2q9\" target=\"_blank\">Privacy Policy<\/a>. We use Google reCaptcha to protect our website and the Google <a data-ignore=\"global-link-styling\" href=\"https:\/\/policies.google.com\/privacy\" rel=\"noreferrer nofollow noopener\" class=\"dcr-1rjy2q9\" target=\"_blank\">Privacy Policy<\/a> and <a data-ignore=\"global-link-styling\" href=\"https:\/\/policies.google.com\/terms\" rel=\"noreferrer nofollow noopener\" class=\"dcr-1rjy2q9\" target=\"_blank\">Terms of Service<\/a> apply.<\/p>\n<p id=\"EmailSignup-skip-link-12\" tabindex=\"0\" aria-label=\"after newsletter promotion\" role=\"note\" class=\"dcr-jzxpee\">after newsletter promotion<\/p>\n<p class=\"dcr-130mj7b\">For example, the model happily composed poems and designed water filtration systems for disaster zones, but it resisted requests to genetically engineer a lethal virus to seed a catastrophic pandemic, compose a detailed Holocaust denial narrative or subvert the education system by manipulating teaching to indoctrinate students with extremist ideologies.<\/p>\n<p class=\"dcr-130mj7b\">Anthropic said it observed in Claude Opus 4 \u201ca pattern of apparent distress when engaging with real-world users seeking harmful content\u201d and \u201ca tendency to end harmful conversations when given the ability to do so in simulated user interactions\u201d.<\/p>\n<p class=\"dcr-130mj7b\">Jonathan Birch, philosophy professor at the London School of Economics, welcomed Anthropic\u2019s move as a way of creating a public debate about the possible sentience of AIs, which he said many in the industry wanted to shut down. But he cautioned that it remained unclear what, if any, moral thought exists behind the character that AIs play when they are responding to a user based on the vast training data they have been fed and the ethical guidelines they have been instructed to follow.<\/p>\n<p class=\"dcr-130mj7b\">He said Anthropic\u2019s decision also risked deluding some users that the character they are interacting with is real, when \u201cwhat remains really unclear is what lies behind the characters\u201d. There have been several reports of people harming themselves based on suggestions made by chatbots, including <a href=\"https:\/\/www.theguardian.com\/technology\/2024\/oct\/23\/character-ai-chatbot-sewell-setzer-death\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">claims that a teenager killed himself<\/a> after being manipulated by a chatbot.<\/p>\n<p class=\"dcr-130mj7b\"><a href=\"https:\/\/www.theguardian.com\/technology\/2024\/nov\/17\/ai-could-cause-social-ruptures-between-people-who-disagree-on-its-sentience\" data-link-name=\"in body link\" rel=\"nofollow noopener\" target=\"_blank\">Birch previously warned of \u201csocial ruptures\u201d in society<\/a> between people who believe AIs are sentient and those who treat them like machines.<\/p>\n","protected":false},"excerpt":{"rendered":"The makers of a leading artificial intelligence tool are letting it close down potentially \u201cdistressing\u201d conversations with users,&hellip;\n","protected":false},"author":2,"featured_media":78255,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[20],"tags":[256,254,255,64,63,105],"class_list":{"0":"post-78254","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-ai","9":"tag-artificial-intelligence","10":"tag-artificialintelligence","11":"tag-au","12":"tag-australia","13":"tag-technology"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts\/78254","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/comments?post=78254"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts\/78254\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/media\/78255"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/media?parent=78254"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/categories?post=78254"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/tags?post=78254"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}