{"id":118420,"date":"2025-11-02T23:33:10","date_gmt":"2025-11-02T23:33:10","guid":{"rendered":"https:\/\/www.newsbeep.com\/ie\/118420\/"},"modified":"2025-11-02T23:33:10","modified_gmt":"2025-11-02T23:33:10","slug":"tech-groups-step-up-efforts-to-solve-ais-big-security-flaw","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/ie\/118420\/","title":{"rendered":"Tech groups step up efforts to solve AI\u2019s big security flaw"},"content":{"rendered":"<p>The world\u2019s top artificial intelligence groups are stepping up efforts to solve a critical security flaw in their large language models that can be exploited by cyber criminals. <\/p>\n<p><a href=\"https:\/\/www.ft.com\/stream\/9c55489a-6ba2-4360-91d9-7e96db22af73\" data-trackable=\"link\" rel=\"nofollow noopener\" target=\"_blank\">Google DeepMind<\/a>, <a href=\"https:\/\/www.ft.com\/stream\/15c0cb45-8892-46cd-a086-1d2716ae7246\" data-trackable=\"link\" rel=\"nofollow noopener\" target=\"_blank\">Anthropic<\/a>, <a href=\"https:\/\/www.ft.com\/stream\/e3402603-d253-4aa1-ac4d-fc9bdbf4ccb8\" data-trackable=\"link\" rel=\"nofollow noopener\" target=\"_blank\">OpenAI<\/a> and <a href=\"https:\/\/www.ft.com\/stream\/4f447b5d-53f5-41bd-ab42-3c0cfc161699\" data-trackable=\"link\" rel=\"nofollow noopener\" target=\"_blank\">Microsoft<\/a> are among those trying to prevent so-called indirect prompt injection attacks, where a third party hides commands in websites or emails designed to trick the AI model into revealing unauthorised information, such as confidential data. <\/p>\n<p>\u201cAI is being used by cyber actors at every chain of the attack right now,\u201d said Jacob Klein, who leads the threat intelligence team at AI start-up Anthropic.\u00a0<\/p>\n<p>AI groups are using a variety of techniques, including hiring external testers and using AI-powered tools, to detect and reduce <a href=\"https:\/\/www.ft.com\/content\/14a2c98b-c8d5-4e5b-a7b0-30f0a05ec432\" data-trackable=\"link\" rel=\"nofollow noopener\" target=\"_blank\">malicious uses<\/a> of their powerful technologies. But experts warned the industry had not yet solved how to stop indirect prompt injection attacks. <\/p>\n<p>Part of the problem is LLMs are designed to follow instructions, and currently do not distinguish between legitimate commands from users and input that should not be trusted. This is also the reason why AI models are prone to jailbreaking, where users can prompt LLMs to disregard their safeguards. <\/p>\n<p>Klein said Anthropic works with external testers to make its <a href=\"https:\/\/www.ft.com\/content\/b9c6b9d0-c418-45b1-a266-72605ccc8aaa\" data-trackable=\"link\" rel=\"nofollow noopener\" target=\"_blank\">Claude model<\/a> more resistant to indirect prompt injection attacks. They also have AI tools to detect when they might be happening.\u00a0<\/p>\n<p>\u201cWhen we find a malicious use, depending on confidence levels, we may automatically trigger some intervention or it may send it to human review,\u201d he added.\u00a0<\/p>\n<p>Google DeepMind uses a technique called automated red teaming, where the company\u2019s internal researchers constantly attack its Gemini model in a realistic way to uncover potential security weaknesses.\u00a0<\/p>\n<p>In May, the UK\u2019s National Cyber Security Centre warned that this flaw posed an increased threat, as it risks exposing millions of companies and individuals that use LLMs and chatbots to <a href=\"https:\/\/www.ft.com\/content\/602e10e7-00b3-4c9b-bb27-6480d7246f37\" data-trackable=\"link\" rel=\"nofollow noopener\" target=\"_blank\">sophisticated phishing attacks and scams<\/a>.\u00a0<\/p>\n<p>LLMs also have another major vulnerability, where outsiders can create back doors and cause the models to misbehave by inserting malicious material into data that is then used in AI training. <\/p>\n<p>These so-called \u201cdata poisoning attacks\u201d are easier to conduct than scientists previously believed, according to new <a href=\"https:\/\/arxiv.org\/abs\/2510.07192\" data-trackable=\"link\" rel=\"nofollow noopener\" target=\"_blank\">research<\/a> published last month by AI start-up Anthropic, the UK\u2019s AI Security Institute and the Alan Turing Institute. <\/p>\n<p>While these vulnerabilities pose big risks, experts argue that AI is also helping to boost company\u2019s defences against cyber attacks. <\/p>\n<p>For years, attackers have had a slight advantage, in that they only needed to find one weakness, while defenders had to protect everything, said Microsoft\u2019s corporate vice-president and deputy chief information security officer Ann Johnson. <\/p>\n<p>\u201cDefensive systems are learning faster, adapting faster, and moving from reactive to proactive,\u201d she added. <\/p>\n<p>The race to solve flaws in AI models comes as cyber security is emerging as one of the top concerns for companies seeking to adopt AI tools into their business.<\/p>\n<p>A recent <a href=\"https:\/\/www.ft.com\/content\/e93e56df-dd9b-40c1-b77a-dba1ca01e473\" data-trackable=\"link\" rel=\"nofollow noopener\" target=\"_blank\">Financial Times analysis<\/a> of hundreds of corporate filings and executive transcripts at S&amp;P 500 companies last year found the most commonly cited worry was cyber security, which was mentioned as a risk by more than half of the S&amp;P 500 in 2024.<\/p>\n<p>Hacking experts said the advancement of AI in recent years has already boosted the multibillion-dollar cyber crime industry. It has provided amateur hackers with cheap tools to write harmful software, as well as systems for professional criminals to better automate and scale up their operations.\u00a0<\/p>\n<p>LLMs allow hackers to quickly generate new malicious code that has not been detected yet, which makes it harder to defend against, said Jake Moore, global cyber security adviser at cyber security group ESET. <\/p>\n<p>A recent <a href=\"https:\/\/mitsloan.mit.edu\/ideas-made-to-matter\/80-ransomware-attacks-now-use-artificial-intelligence\" data-trackable=\"link\" rel=\"nofollow noopener\" target=\"_blank\">study<\/a> by researchers at MIT found that 80 per cent of ransomware attacks they examined used AI, and in 2024, phishing scams and deepfake-related fraud linked to the technology saw a 60 per cent increase.<\/p>\n<p>AI tools are also being used by hackers to <a href=\"https:\/\/www.ft.com\/content\/d60fb4fb-cb85-4df7-b246-ec3d08260e6f\" data-trackable=\"link\" rel=\"nofollow noopener\" target=\"_blank\">collect information<\/a> on victims online. LLMs can scour the web efficiently for personal data on someone\u2019s public accounts, images or even find audio clips of someone speaking. <\/p>\n<p>These could be used to conduct sophisticated social engineering attacks for financial crimes, said Paul Fabara, Visa\u2019s chief risk and client services officer.\u00a0<\/p>\n<p>Vijay Balasubramaniyan, chief executive and co-founder of Pindrop, a cyber security firm specialising in voice fraud, said generative AI has made creating realistic sounding deepfakes much easier and quicker than before.\u00a0<\/p>\n<p class=\"n-content-recommended__title o3-type-body-highlight\">Recommended<\/p>\n<p><a href=\"https:\/\/www.ft.com\/content\/d3119d3f-97bd-4ff4-905d-b471a8828beb\" data-trackable=\"image-link\" data-trackable-context-story-link=\"image-link\" tabindex=\"-1\" aria-hidden=\"true\" rel=\"nofollow noopener\" target=\"_blank\"><img decoding=\"async\" class=\"o-teaser__image\" src=\"https:\/\/www.newsbeep.com\/ie\/wp-content\/uploads\/2025\/11\/https:\/\/images.ft.com\/v3\/image\/raw\/https%3A%2F%2Fd1e00ek4ebabms.cloudfront.net%2Fproduction%2Fd69475.jpeg\" alt=\"An illustration of a skull on a white square with squiggly lines coming out of it, all on a blue background\"\/><\/a><\/p>\n<p>\u201cBack in 2023, we\u2019d see one deepfake attack per month across the entire customer base. Now we\u2019re seeing seven per day per customer,\u201d he added.\u00a0<\/p>\n<p>Companies are particularly vulnerable to these kinds of attacks, said ESET\u2019s Moore. AI systems can collate information from the public internet, such as employees\u2019 LinkedIn posts, to find out what kind of programs and software companies use day to day, and then use that to find vulnerabilities.\u00a0<\/p>\n<p>Anthropic recently intercepted a sophisticated actor using the company\u2019s language models for \u201cvibe hacking\u201d, where the person had automated many of the processes of a large-scale attack.<\/p>\n<p>The bad actor used Claude Code to automate reconnaissance, harvest victims\u2019 credentials and infiltrate systems. The person had targeted 17 organisations to extort up to $500,000 from them.\u00a0<\/p>\n<p>Cyber experts said companies need to stay vigilant in monitoring for new threats and consider restricting how many people have access to sensitive datasets and AI tools that are prone to attacks. \u00a0<\/p>\n<p>\u201cIt doesn\u2019t take much to be a crook nowadays,\u201d said Visa\u2019s Fabara. \u201cYou get a laptop, $15 to download the cheap bootleg version of gen AI in the dark web and off you go.\u201d <\/p>\n","protected":false},"excerpt":{"rendered":"The world\u2019s top artificial intelligence groups are stepping up efforts to solve a critical security flaw in their&hellip;\n","protected":false},"author":2,"featured_media":118421,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[20],"tags":[220,218,219,61,60,80],"class_list":{"0":"post-118420","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-ai","9":"tag-artificial-intelligence","10":"tag-artificialintelligence","11":"tag-ie","12":"tag-ireland","13":"tag-technology"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/posts\/118420","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/comments?post=118420"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/posts\/118420\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/media\/118421"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/media?parent=118420"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/categories?post=118420"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/ie\/wp-json\/wp\/v2\/tags?post=118420"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}