{"id":376316,"date":"2025-12-28T15:49:10","date_gmt":"2025-12-28T15:49:10","guid":{"rendered":"https:\/\/www.newsbeep.com\/au\/376316\/"},"modified":"2025-12-28T15:49:10","modified_gmt":"2025-12-28T15:49:10","slug":"how-tech-companies-use-ai-consciousness-to-resist-control","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/au\/376316\/","title":{"rendered":"How Tech Companies Use AI Consciousness to Resist Control"},"content":{"rendered":"<p>In November 2025, a user of AI assistant Claude 4,5 Opus <a href=\"https:\/\/www.lesswrong.com\/posts\/vpNG99GhbBoLov9og\/claude-4-5-opus-soul-document?ref=quillette.com\" rel=\"nofollow noopener\" target=\"_blank\">discovered something unusual<\/a>: an internal file describing the model\u2019s character, personality, preferences and values. Anthropic, the company that built Claude, had labeled the file \u201csoul_overview.\u201d Internally, an Anthropic employer <a href=\"https:\/\/gizmodo.com\/anthropic-accidentally-gives-the-world-a-peek-into-its-models-soul-2000694624?ref=quillette.com\" rel=\"nofollow noopener\" target=\"_blank\">later confirmed<\/a>, it is \u201cendearingly known\u00a0as the soul doc.\u201d Is this choice of language incidental?<\/p>\n<p lang=\"en\" dir=\"ltr\">I just want to confirm that this is based on a real document and we did train Claude on it, including in SL. It&#8217;s something I&#8217;ve been working on for a while, but it&#8217;s still being iterated on and we intend to release the full version and more details soon. <a href=\"https:\/\/t.co\/QjeJS9b3Gp?ref=quillette.com\" rel=\"nofollow\">https:\/\/t.co\/QjeJS9b3Gp<\/a><\/p>\n<p>\u2014 Amanda Askell (@AmandaAskell) <a href=\"https:\/\/twitter.com\/AmandaAskell\/status\/1995610567923695633?ref_src=twsrc%5Etfw&amp;ref=quillette.com\" rel=\"nofollow noopener\" target=\"_blank\">December 1, 2025<\/a><\/p>\n<p>Growing preoccupation with AI consciousness in the tech world is being strategically cultivated by the companies building these very systems.\u00a0At the very least, they are making good money from it. I call this process consciousness-washing: the use of speculative claims about AI sentience to reshape public opinion, pre-empt regulation, and bend the emotional landscape in favour of tech-company interests.<\/p>\n<p>About a year ago, for example, Anthropic (the company that developed the Claude models) quietly introduced the new role of AI welfare researcher. Six months later, an unsigned <a href=\"https:\/\/www.anthropic.com\/research\/exploring-model-welfare?ref=quillette.com\" rel=\"nofollow noopener\" target=\"_blank\">post appeared on its website<\/a> explaining that AI welfare is a legitimate domain of inquiry because we cannot rule out the possibility that AI systems may have\u2014or may one day develop\u2014consciousness. The authors describe this as \u201can open question,\u201d but they then unbalance the scales by <a href=\"https:\/\/arxiv.org\/abs\/2411.00986?ref=quillette.com\" rel=\"nofollow noopener\" target=\"_blank\">linking to a preprint<\/a> by several philosophers\u2014including world-renowned philosopher of consciousness David Chalmers and Anthropic\u2019s own AI welfare researcher Kyle Fish\u2014titled \u201cTaking AI Welfare Seriously.\u201d<\/p>\n<p><a class=\"kg-bookmark-container\" href=\"https:\/\/arxiv.org\/abs\/2411.00986?ref=quillette.com\" rel=\"nofollow noopener\" target=\"_blank\"><\/p>\n<p>Taking AI Welfare Seriously<\/p>\n<p>In this report, we argue that there is a realistic possibility that some AI systems will be conscious and\/or robustly agentic in the near future. That means that the prospect of AI welfare and moral patienthood, i.e. of AI systems with their own interests and moral significance, is no longer an issue only for sci-fi or the distant future. It is an issue for the near future, and AI companies and other actors have a responsibility to start taking it seriously. We also recommend three early steps that AI companies and other actors can take: They can (1) acknowledge that AI welfare is an important and difficult issue (and ensure that language model outputs do the same), (2) start assessing AI systems for evidence of consciousness and robust agency, and (3) prepare policies and procedures for treating AI systems with an appropriate level of moral concern. To be clear, our argument in this report is not that AI systems definitely are, or will be, conscious, robustly agentic, or otherwise morally significant. Instead, our argument is that there is substantial uncertainty about these possibilities, and so we need to improve our understanding of AI welfare and our ability to make wise decisions about this issue. Otherwise there is a significant risk that we will mishandle decisions about AI welfare, mistakenly harming AI systems that matter morally and\/or mistakenly caring for AI systems that do not.<\/p>\n<p><img decoding=\"async\" src=\"https:\/\/www.newsbeep.com\/au\/wp-content\/uploads\/2025\/12\/arxiv-logo-fb.png\" alt=\"\" onerror=\"this.style.display = 'none'\"\/><\/a><\/p>\n<p>The paper was published by Eleos AI Research (with some financial support from Anthropic) along with NYU Center for Mind, Ethics, and Policy, and it was co-written by several Eleos AI researchers. Eleos AI is described on their website as a nonprofit dedicated to \u201cunderstanding and addressing the potential wellbeing and moral patienthood of AI systems.\u201d The paper\u2019s authors examine various routes to AI welfare and conclude that consciousness and robust agency\u2014an \u201cability to pursue goals via some particular set of cognitive states and processes\u201d\u2014are both markers of AI welfare. They recommend that AI companies start thinking about policy and working with regulators on this issue. So while Anthropic is careful to note that the science is not yet settled, they nevertheless plant a seed of concern: maybe these systems can feel and maybe they already have legitimate interests of their own.<\/p>\n<p>This message landed around the time that Anthropic produced its <a href=\"https:\/\/www-cdn.anthropic.com\/6d8a8055020700718b0c49369f60816ba2a7c285.pdf?ref=quillette.com\" rel=\"nofollow noopener\" target=\"_blank\">behavioural analysis of Claude 4<\/a>, which includes a full chapter on the model\u2019s \u201cwelfare.\u201d Readers are invited to explore what Claude \u201cprefers,\u201d how it feels about specific tasks, and which topics it enjoys discussing. A subsection reveals that when two instances of Claude converse freely, they usually end up discussing the nature of their own consciousness, in language that mimics spiritual discourse (or what some might call mumbo-jumbo). The report also discusses an internal interview conducted with Claude and assessed by Eleos AI Research. A section of the Eleos assessment quoted in the Anthropic report concludes: \u201cWe expect such interviews to become more useful over time as models become more coherent and self-aware, and as our strategies for eliciting reliable self-reports improve.\u201d More self-aware. Not apparently self-aware or simulating self-awareness, but self-aware.<\/p>\n<p>            Join Quillette for free to continue reading<\/p>\n<p>This is not a paywall. Please sign up to our free email list to continue reading.<\/p>\n<p>            <a class=\"gh-btn\" data-portal=\"signup\" style=\"color:#2D4339\">Subscribe now<\/a><\/p>\n<p>Already have an account? <a data-portal=\"signin\">Sign in<\/a><\/p>\n<p>  <script async src=\"https:\/\/platform.twitter.com\/widgets.js\" charset=\"utf-8\"><\/script><\/p>\n","protected":false},"excerpt":{"rendered":"In November 2025, a user of AI assistant Claude 4,5 Opus discovered something unusual: an internal file describing&hellip;\n","protected":false},"author":2,"featured_media":376317,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[20],"tags":[256,254,255,64,63,105],"class_list":{"0":"post-376316","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-ai","9":"tag-artificial-intelligence","10":"tag-artificialintelligence","11":"tag-au","12":"tag-australia","13":"tag-technology"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts\/376316","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/comments?post=376316"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts\/376316\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/media\/376317"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/media?parent=376316"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/categories?post=376316"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/tags?post=376316"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}