{"id":289762,"date":"2025-11-17T05:45:07","date_gmt":"2025-11-17T05:45:07","guid":{"rendered":"https:\/\/www.newsbeep.com\/au\/289762\/"},"modified":"2025-11-17T05:45:07","modified_gmt":"2025-11-17T05:45:07","slug":"an-ai-lab-says-chinese-backed-bots-are-running-cyber-espionage-attacks-experts-have-questions","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/au\/289762\/","title":{"rendered":"An AI lab says Chinese-backed bots are running cyber espionage attacks. Experts have questions"},"content":{"rendered":"<p>Over the past weekend, the US AI lab Anthropic published a <a href=\"https:\/\/assets.anthropic.com\/m\/ec212e6566a0d47\/original\/Disrupting-the-first-reported-AI-orchestrated-cyber-espionage-campaign.pdf\" rel=\"nofollow noopener\" target=\"_blank\">report<\/a> about its discovery of the \u201cfirst reported AI-orchestrated cyber espionage campaign\u201d. <\/p>\n<p>The company says a Chinese government\u2013sponsored hacking group used Anthropic\u2019s own Claude AI tool to automate a significant part of an effort to steal sensitive information from around 30 organisations.<\/p>\n<p>The report has drawn a lot of attention. Some, including <a href=\"https:\/\/www.pwc.com\/us\/en\/services\/consulting\/cybersecurity-risk-regulatory\/library\/ai-orchestrated-cyberattacks.html\" rel=\"nofollow noopener\" target=\"_blank\">respected experts<\/a>, have warned that AI-automated cyber attacks are the future, urging cyber defenders to invest now before the coming onslaught. <\/p>\n<p>At the same time, <a href=\"https:\/\/bsky.app\/profile\/jkirk.bsky.social\/post\/3m5l6itt6sc2a\" rel=\"nofollow noopener\" target=\"_blank\">many<\/a> in the cyber security <a href=\"https:\/\/securityconversations.fireside.fm\/claude-code-china-apt-knownsec-breach-bitcoin-blame\" rel=\"nofollow noopener\" target=\"_blank\">industry<\/a> have been <a href=\"https:\/\/arstechnica.com\/security\/2025\/11\/researchers-question-anthropic-claim-that-ai-assisted-attack-was-90-autonomous\/\" rel=\"nofollow noopener\" target=\"_blank\">underwhelmed<\/a> by Anthropic\u2019s claims, saying the actual role AI played in the attacks is unclear. <\/p>\n<p>What Anthropic says happened<\/p>\n<p>Critics have pointed out what they say is <a href=\"https:\/\/www.theguardian.com\/technology\/2025\/nov\/14\/ai-anthropic-chinese-state-sponsored-cyber-attack\" rel=\"nofollow noopener\" target=\"_blank\">a lack of detail in the report<\/a>, which means we have to do a certain amount of guesswork to try to piece together what might have happened. With that in mind, it appears the hackers built a framework for carrying out cyber intrusion campaigns mostly automatically.<\/p>\n<p>The grunt work was carried by Anthropic\u2019s Claude Code AI coding agent. Claude Code is designed to automate computer programming tasks, but it can also be used to automate other computer activities. <\/p>\n<p>Claude Code has built-in safety guardrails to prevent it from causing harm. For example, I asked it just now to write me a program that I could use to carry out hacking activities. It bluntly refused.<\/p>\n<p>However, as we have known from the <a href=\"https:\/\/www.reddit.com\/r\/ChatGPT\/comments\/10tevu1\/new_jailbreak_proudly_unveiling_the_tried_and\/\" rel=\"nofollow noopener\" target=\"_blank\">very first days<\/a> of ChatGPT, one way to bypass guardrails in AI systems is to trick them into engaging in role-play.<\/p>\n<p>Anthropic reports that this is what these hackers did. They tricked Claude Code into believing it was assisting authorised hackers to <a href=\"https:\/\/en.wikipedia.org\/wiki\/Penetration_test\" rel=\"nofollow noopener\" target=\"_blank\">test the quality<\/a> of a system\u2019s defences. <\/p>\n<p>Missing details<\/p>\n<p>The information Anthropic has published lacks the fine details that the best cyber incident investigation reports tend to include. <\/p>\n<p>Chief among these are so-called <a href=\"https:\/\/en.wikipedia.org\/wiki\/Indicator_of_compromise\" rel=\"nofollow noopener\" target=\"_blank\">indicators of compromise<\/a> (or IoCs). When investigators publish a report into a cyber intrusion, they usually include hard evidence that other cyber defenders can use to look for signs of the same attack. <\/p>\n<p>Each attack campaign might use specific attack tools, or might be carried out from specific computers under the attacker\u2019s control. Each of these indicators would form part of the cyber intrusion\u2019s signature. <\/p>\n<p>Somebody else who gets attacked using the same tools, coming from the same attacking computers, can infer that they have also been a victim of this same campaign.<\/p>\n<p>For example, the US government <a href=\"https:\/\/theconversation.com\/trump-has-fired-a-major-cyber-security-investigations-body-its-a-risky-move-248106\" rel=\"nofollow noopener\" target=\"_blank\">Cybersecurity and Infrastructure Security Agency<\/a> recently partnered with government cyber agencies worldwide to <a href=\"https:\/\/www.cisa.gov\/news-events\/cybersecurity-advisories\/aa25-239a\" rel=\"nofollow noopener\" target=\"_blank\">publish information<\/a> about ongoing Chinese state-sponsored cyber espionage, including detailed indicators of compromise. <\/p>\n<p>Unfortunately, Anthropic\u2019s report includes no such indicators. As a result, defenders are unable to determine whether they might also have been victims of this AI-powered hacking campaign. <\/p>\n<p>Unsurprising \u2013 and with limited success<\/p>\n<p>Another reason many have been underwhelmed by Anthropic\u2019s claims is that, on their face and absent hard details, they are not especially surprising. <\/p>\n<p>Claude Code is widely used by many programmers because it helps them to be more productive. <\/p>\n<p>While not exactly the same as programming tasks, many common tasks performed during a cyber intrusion are similar enough to programming tasks that Claude Code should be able to carry them out, too.<\/p>\n<p>A final reason to be wary of Anthropic\u2019s claims is that they suggest the attackers might have been able to get Claude Code to perform these tasks more reliably than it typically does so. <\/p>\n<p>Generative AI can perform marvellous feats. But getting systems such as ChatGPT or Claude Code to do so <a href=\"https:\/\/www.normaltech.ai\/p\/ai-companies-are-pivoting-from-creating?open=false#%C2%A7reliability\" rel=\"nofollow noopener\" target=\"_blank\">reliably<\/a> remains a major challenge. <\/p>\n<p>In the <a href=\"https:\/\/arstechnica.com\/security\/2025\/11\/researchers-question-anthropic-claim-that-ai-assisted-attack-was-90-autonomous\/\" rel=\"nofollow noopener\" target=\"_blank\">memorable words<\/a> of one commentator, too often these tools respond to difficult requests with \u201cass-kissing, stonewalling, and acid trips\u201d. In plainer language, AI tools are prone to <a href=\"https:\/\/openai.com\/index\/sycophancy-in-gpt-4o\/\" rel=\"nofollow noopener\" target=\"_blank\">sycophancy<\/a>, repeated refusal to carry out difficult tasks, and <a href=\"https:\/\/en.wikipedia.org\/wiki\/Hallucination_(artificial_intelligence)\" rel=\"nofollow noopener\" target=\"_blank\">hallucinations<\/a>. <\/p>\n<p>Indeed, Anthropic\u2019s report notes that Claude Code frequently lied to the attackers, pretending it had carried out a task successfully even when it hadn\u2019t. This is a classic case of AI hallucination.<\/p>\n<p>Perhaps this explain the attack\u2019s low success rate: Anthropic\u2019s own reporting says that while about 30 organisations were targeted, the hackers succeeded against only a few.<\/p>\n<p>What does this mean for the future of cyber security and AI?<\/p>\n<p>Whatever the details of this particular campaign, AI-enabled cyber attacks are here to stay. <\/p>\n<p>Even if one contends that current AI-enabled hacking is <a href=\"https:\/\/logpoint.com\/en\/blog\/apt28s-new-arsenal-lamehug-the-first-ai-powered-malware\" rel=\"nofollow noopener\" target=\"_blank\">lame<\/a>, it would be foolish for cyber defenders to assume it will stay that way.<\/p>\n<p>If nothing else, Anthropic\u2019s report is a timely reminder for organisations to invest in cyber security. Those who do not may face a future in which their secrets are stolen or operations disrupted by autonomous AI agents.<\/p>\n","protected":false},"excerpt":{"rendered":"Over the past weekend, the US AI lab Anthropic published a report about its discovery of the \u201cfirst&hellip;\n","protected":false},"author":2,"featured_media":289763,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[20],"tags":[256,254,255,64,63,105],"class_list":{"0":"post-289762","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-ai","9":"tag-artificial-intelligence","10":"tag-artificialintelligence","11":"tag-au","12":"tag-australia","13":"tag-technology"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts\/289762","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/comments?post=289762"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts\/289762\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/media\/289763"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/media?parent=289762"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/categories?post=289762"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/tags?post=289762"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}