{"id":114229,"date":"2025-08-27T17:22:12","date_gmt":"2025-08-27T17:22:12","guid":{"rendered":"https:\/\/www.newsbeep.com\/us\/114229\/"},"modified":"2025-08-27T17:22:12","modified_gmt":"2025-08-27T17:22:12","slug":"anthropics-auto-clicking-ai-chrome-extension-raises-browser-hijacking-concerns","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/us\/114229\/","title":{"rendered":"Anthropic\u2019s auto-clicking AI Chrome extension raises browser-hijacking concerns"},"content":{"rendered":"<p>The company tested 123 cases representing 29 different attack scenarios and found a 23.6 percent attack success rate when browser use operated without safety mitigations.<\/p>\n<p>One example involved a malicious email that instructed Claude to delete a user&#8217;s emails for &#8220;mailbox hygiene&#8221; purposes. Without safeguards, Claude followed these instructions and deleted the user&#8217;s emails without confirmation.<\/p>\n<p>Anthropic says it has implemented several defenses to address these vulnerabilities. Users can grant or revoke Claude&#8217;s access to specific websites through site-level permissions. The system requires user confirmation before Claude takes high-risk actions like publishing, purchasing, or sharing personal data. The company has also blocked Claude from accessing websites offering financial services, adult content, and pirated content by default.<\/p>\n<p>These safety measures reduced the attack success rate from 23.6 percent to 11.2 percent in autonomous mode. On a specialized test of four browser-specific attack types, the new mitigations reportedly reduced the success rate from 35.7 percent to 0 percent.<\/p>\n<p>Independent AI researcher Simon Willison, who has extensively written about AI security risks and <a href=\"https:\/\/arstechnica.com\/information-technology\/2022\/09\/twitter-pranksters-derail-gpt-3-bot-with-newly-discovered-prompt-injection-hack\/\" rel=\"nofollow noopener\" target=\"_blank\">coined<\/a> the term &#8220;prompt injection&#8221; in 2022, called the remaining 11.2 percent attack rate &#8220;catastrophic,&#8221; <a href=\"https:\/\/simonwillison.net\/2025\/Aug\/26\/piloting-claude-for-chrome\/\" rel=\"nofollow noopener\" target=\"_blank\">writing<\/a> on his blog that &#8220;in the absence of 100% reliable protection I have trouble imagining a world in which it&#8217;s a good idea to unleash this pattern.&#8221;<\/p>\n<p>By &#8220;pattern,&#8221; Willison is referring to the recent trend of integrating AI agents into web browsers. &#8220;I strongly expect that the entire concept of an agentic browser extension is fatally flawed and cannot be built safely,&#8221; he <a href=\"https:\/\/simonwillison.net\/2025\/Aug\/25\/agentic-browser-security\/\" rel=\"nofollow noopener\" target=\"_blank\">wrote<\/a> in an earlier post on similar prompt-injection security issues recently found in Perplexity Comet.<\/p>\n<p>The security risks are no longer theoretical. Last week, Brave&#8217;s security team <a href=\"https:\/\/brave.com\/blog\/comet-prompt-injection\/\" rel=\"nofollow noopener\" target=\"_blank\">discovered<\/a> that Perplexity&#8217;s Comet browser could be tricked into accessing users&#8217; Gmail accounts and triggering password recovery flows through malicious instructions hidden in Reddit posts. When users asked Comet to summarize a Reddit thread, attackers could embed invisible commands that instructed the AI to open Gmail in another tab, extract the user&#8217;s email address, and perform unauthorized actions. Although Perplexity attempted to fix the vulnerability, Brave later confirmed that its mitigations were defeated and the security hole remained.<\/p>\n<p>For now, Anthropic plans to use its new research preview to identify and address attack patterns that emerge in real-world usage before making the Chrome extension more widely available. In the absence of good protections from AI vendors, the burden of security falls on the user, who is taking a large risk by using these tools on the open web. As Willison noted in his post about Claude for Chrome, &#8220;I don&#8217;t think it&#8217;s reasonable to expect end users to make good decisions about the security risks.&#8221;<\/p>\n","protected":false},"excerpt":{"rendered":"The company tested 123 cases representing 29 different attack scenarios and found a 23.6 percent attack success rate&hellip;\n","protected":false},"author":2,"featured_media":114230,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[45],"tags":[182,181,507,74],"class_list":{"0":"post-114229","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-ai","9":"tag-artificial-intelligence","10":"tag-artificialintelligence","11":"tag-technology"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/posts\/114229","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/comments?post=114229"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/posts\/114229\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/media\/114230"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/media?parent=114229"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/categories?post=114229"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/tags?post=114229"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}