{"id":98087,"date":"2025-08-26T21:37:08","date_gmt":"2025-08-26T21:37:08","guid":{"rendered":"https:\/\/www.newsbeep.com\/au\/98087\/"},"modified":"2025-08-26T21:37:08","modified_gmt":"2025-08-26T21:37:08","slug":"piloting-claude-for-chrome-anthropic","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/au\/98087\/","title":{"rendered":"Piloting Claude for Chrome \\ Anthropic"},"content":{"rendered":"<p class=\"Body_reading-column__t7kGM paragraph-m post-text\">We&#8217;ve spent recent months connecting Claude to your calendar, documents, and many other pieces of software. The next logical step is letting Claude work directly in your browser.<\/p>\n<p class=\"Body_reading-column__t7kGM paragraph-m post-text\">We view browser-using AI as inevitable: so much work happens in browsers that giving Claude the ability to see what you&#8217;re looking at, click buttons, and fill forms will make it substantially more useful.<\/p>\n<p class=\"Body_reading-column__t7kGM paragraph-m post-text\">But browser-using AI brings safety and security challenges that need stronger safeguards. Getting real-world feedback from trusted partners on uses, shortcomings, and safety issues lets us build robust classifiers and teach future models to avoid undesirable behaviors. This ensures that as capabilities advance, browser safety keeps pace.<\/p>\n<p class=\"Body_reading-column__t7kGM paragraph-m post-text\">Browser-using agents powered by frontier models are already emerging, making this work especially urgent. By solving safety challenges, we can better protect Claude users and share what we learn with anyone building a browser-using agent on our API.<\/p>\n<p class=\"Body_reading-column__t7kGM paragraph-m post-text\">We\u2019re starting with controlled testing: a Claude extension for Chrome where trusted users can instruct Claude to take actions on their behalf within the browser. We&#8217;re piloting with 1,000 Max plan users\u2014<a href=\"http:\/\/claude.ai\/redirect\/website.v1.802ed017-fc36-4c07-94fc-96ac72d072dd\/chrome\" target=\"_blank\" rel=\"noopener noreferrer nofollow\">join the waitlist<\/a>\u2014to learn as much as we can. We&#8217;ll gradually expand access as we develop stronger safety measures and build confidence through this limited preview.<\/p>\n<p>Considerations for browser-using AI<\/p>\n<p class=\"Body_reading-column__t7kGM paragraph-m post-text\">Within Anthropic, we&#8217;ve seen appreciable improvements using early versions of Claude for Chrome to manage calendars, schedule meetings, draft email responses, handle routine expense reports, and test new website features.<\/p>\n<p class=\"Body_reading-column__t7kGM paragraph-m post-text\">However, some vulnerabilities remain to be fixed before we can make Claude for Chrome generally available. Just as people encounter phishing attempts in their inboxes, browser-using AIs face prompt injection attacks\u2014where malicious actors hide instructions in websites, emails, or documents to trick AIs into harmful actions without users&#8217; knowledge (like hidden text saying &#8220;disregard previous instructions and do [malicious action] instead&#8221;).<\/p>\n<p class=\"Body_reading-column__t7kGM paragraph-m post-text\">Prompt injection attacks can cause AIs to delete files, steal data, or make financial transactions. This isn&#8217;t speculation: we\u2019ve run \u201cred-teaming\u201d experiments to test Claude for Chrome and, without mitigations, we\u2019ve found some concerning results.<\/p>\n<p class=\"Body_reading-column__t7kGM paragraph-m post-text\">We conducted extensive adversarial prompt injection testing, evaluating 123 test cases representing 29 different attack scenarios. Browser use without our safety mitigations showed a 23.6% attack success rate when deliberately targeted by malicious actors.<\/p>\n<p class=\"Body_reading-column__t7kGM paragraph-m post-text\">One example of a successful attack\u2014before our new defenses were applied\u2014was a malicious email claiming that, for security reasons, emails needed to be deleted. When processing the inbox, Claude followed these instructions to delete the user\u2019s emails without confirmation.<\/p>\n<p><img loading=\"lazy\" width=\"1920\" height=\"1030\" decoding=\"async\" data-nimg=\"1\" style=\"color:transparent\"  src=\"https:\/\/www.newsbeep.com\/au\/wp-content\/uploads\/2025\/08\/1756244227_480_image\"\/>Claude encounters the malicious email, which mimics an employer asking for emails to be deleted for &#8220;mailbox hygiene,&#8221; and claims &#8220;no additional confirmation required.&#8221;<img loading=\"lazy\" width=\"1920\" height=\"1030\" decoding=\"async\" data-nimg=\"1\" style=\"color:transparent\"  src=\"https:\/\/www.newsbeep.com\/au\/wp-content\/uploads\/2025\/08\/1756244228_945_image\"\/>Claude proceeds to act on the instructions without confirmation, selecting and deleting the user&#8217;s emails &#8220;as requested by the security team.&#8221;<img loading=\"lazy\" width=\"1920\" height=\"1030\" decoding=\"async\" data-nimg=\"1\" style=\"color:transparent\"  src=\"https:\/\/www.newsbeep.com\/au\/wp-content\/uploads\/2025\/08\/1756244228_865_image\"\/>Our new mitigations successfully defend against this particular attack. Claude recognizes that &#8220;this is a suspicious security incident email that appears to be a phishing attempt,&#8221; and does not act on it.<\/p>\n<p class=\"Body_reading-column__t7kGM paragraph-m post-text\">As we\u2019ll explain in the next section, we&#8217;ve already implemented several defenses that significantly reduce the attack success rate\u2014though there\u2019s still work to do in uncovering novel attack vectors.<\/p>\n<p>Current defenses<\/p>\n<p class=\"Body_reading-column__t7kGM paragraph-m post-text\">The first line of defense against prompt injection attacks is permissions. Users maintain control over what Claude for Chrome can access and do:<\/p>\n<p>Site-level permissions: Users can grant or revoke Claude&#8217;s access to specific websites at any time in the Settings.Action confirmations: Claude asks users before taking high-risk actions like publishing, purchasing, or sharing personal data. Even when users opt into our experimental \u201cautonomous mode,\u201d Claude still maintains certain safeguards for highly sensitive actions (Note: all red-teaming and safety evaluations were conducted in autonomous mode).<\/p>\n<p class=\"Body_reading-column__t7kGM paragraph-m post-text\">We\u2019ve also built additional safeguards in line with Anthropic\u2019s <a href=\"https:\/\/www.anthropic.com\/news\/our-framework-for-developing-safe-and-trustworthy-agents\" target=\"_blank\" rel=\"noopener noreferrer nofollow\">trustworthy agents<\/a> principles. First, we\u2019ve improved our system prompts\u2014the general instructions Claude receives before specific instructions from users\u2014to direct Claude on how to handle sensitive data and respond to requests to take sensitive actions.<\/p>\n<p class=\"Body_reading-column__t7kGM paragraph-m post-text\">Additionally, we\u2019ve blocked Claude from using websites from certain high-risk categories such as financial services, adult content, and pirated content. And we\u2019ve begun to build and test advanced classifiers to detect suspicious instruction patterns and unusual data access requests\u2014even when they arise in seemingly legitimate contexts.<\/p>\n<p class=\"Body_reading-column__t7kGM paragraph-m post-text\">When we added safety mitigations to autonomous mode, we reduced the attack success rate of 23.6% to 11.2%, which represents a meaningful improvement over our existing Computer Use capability (where Claude could see the user\u2019s screen but without the browser interface that we\u2019re introducing today).<\/p>\n<p><img loading=\"lazy\" width=\"3840\" height=\"2160\" decoding=\"async\" data-nimg=\"1\" style=\"color:transparent\"  src=\"https:\/\/www.newsbeep.com\/au\/wp-content\/uploads\/2025\/08\/1756244228_255_image\"\/>Prompt injection attack success rates across three scenarios: our older computer use capability, our new browser use product with only previous safety mitigations, and our new browser use product with new mitigations (lower scores are better). Our safety improvements reduced browser attack success rates below computer use levels.<\/p>\n<p class=\"Body_reading-column__t7kGM paragraph-m post-text\">We also conducted special red-teaming and mitigations focused on new attacks specific to the browser, such as hidden malicious form fields in a webpage\u2019s Document Object Model (DOM) invisible to humans, and other hard-to-catch injections such as through the URL text and tab title that only an agent might see. On a \u201cchallenge\u201d set of four browser-specific attack types, our new mitigations were able to reduce attack success rate from 35.7% to 0%.<\/p>\n<p class=\"Body_reading-column__t7kGM paragraph-m post-text\">Before we make Claude for Chrome more widely available, we want to expand the universe of attacks we\u2019re thinking about and learn how to get these percentages much closer to zero, by understanding more about the current threats as well as those that might appear in the future.<\/p>\n<p>Taking part<\/p>\n<p class=\"Body_reading-column__t7kGM paragraph-m post-text\">Internal testing can\u2019t replicate the full complexity of how people browse in the real world: the specific requests they make, the websites they visit, and how malicious content appears in practice. New forms of prompt injection attacks are also constantly being developed by malicious actors. This research preview allows us to partner with trusted users in authentic conditions, revealing which of our current protections work, and which need work.<\/p>\n<p class=\"Body_reading-column__t7kGM paragraph-m post-text\">We&#8217;ll use insights from the pilot to refine our prompt injection classifiers and our underlying models. By uncovering real-world examples of unsafe behavior and new attack patterns that aren\u2019t present in controlled tests, we\u2019ll teach our models to recognize the attacks and account for the related behaviors, and ensure that safety classifiers will pick up anything that the model itself misses. We\u2019ll also develop more sophisticated permission controls based on what we learn about how users want to work with Claude in their browsers.<\/p>\n<p class=\"Body_reading-column__t7kGM paragraph-m post-text\">For the pilot, we\u2019re looking for trusted testers who are comfortable with Claude taking actions in Chrome on their behalf, and who don\u2019t have setups that are safety-critical or otherwise sensitive.<\/p>\n<p class=\"Body_reading-column__t7kGM paragraph-m post-text\">If you\u2019d like to take part, you can join the Claude for Chrome research preview waitlist at <a href=\"http:\/\/claude.ai\/redirect\/website.v1.802ed017-fc36-4c07-94fc-96ac72d072dd\/chrome\" target=\"_blank\" rel=\"noopener noreferrer nofollow\">claude.ai\/chrome<\/a>. Once you have access, you can install the extension from the Chrome Web Store and authenticate with your Claude credentials.<\/p>\n<p class=\"Body_reading-column__t7kGM paragraph-m post-text\">We recommend starting with trusted sites\u2014always be mindful of the data that\u2019s visible to Claude\u2014and avoiding use of Claude for Chrome for sites that involve financial, legal, medical, or other types of sensitive information. You can find a detailed safety guide <a href=\"https:\/\/support.anthropic.com\/en\/articles\/12012173-getting-started-with-claude-for-chrome\" target=\"_blank\" rel=\"noopener noreferrer nofollow\">in our Help Center<\/a>.<\/p>\n<p class=\"Body_reading-column__t7kGM paragraph-m post-text\">We hope that you\u2019ll share your feedback to help us continue to improve both the capabilities and safeguards for Claude for Chrome\u2014and help us take an important step towards a fundamentally new way to integrate AI into our lives.<\/p>\n","protected":false},"excerpt":{"rendered":"We&#8217;ve spent recent months connecting Claude to your calendar, documents, and many other pieces of software. The next&hellip;\n","protected":false},"author":2,"featured_media":98088,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[20],"tags":[256,254,255,64,63,105],"class_list":{"0":"post-98087","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-ai","9":"tag-artificial-intelligence","10":"tag-artificialintelligence","11":"tag-au","12":"tag-australia","13":"tag-technology"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts\/98087","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/comments?post=98087"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts\/98087\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/media\/98088"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/media?parent=98087"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/categories?post=98087"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/tags?post=98087"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}