{"id":391912,"date":"2026-01-26T21:11:08","date_gmt":"2026-01-26T21:11:08","guid":{"rendered":"https:\/\/www.newsbeep.com\/uk\/391912\/"},"modified":"2026-01-26T21:11:08","modified_gmt":"2026-01-26T21:11:08","slug":"ai-agents-are-mathematically-incapable-of-doing-functional-work-paper-finds","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/uk\/391912\/","title":{"rendered":"AI Agents Are Mathematically Incapable of Doing Functional Work, Paper Finds"},"content":{"rendered":"<p>\t<img decoding=\"async\" class=\"archive-post-thumb article-featured-image w-full h-auto mb-3\" src=\"https:\/\/www.newsbeep.com\/uk\/wp-content\/uploads\/2026\/01\/ai-agents-incapable-math.jpg\"   fetchpriority=\"high\" width=\"2048\" height=\"1297\" alt=\"A paper claims to mathematically prove that AI agents have a hard ceiling to their capabilities that they will never surpass.\"\/><\/p>\n<p>\t\t\tGaston Paris \/ Contributor\n\t<\/p>\n<p class=\"pw-incontent-excluded article-paragraph skip\">A months-old but until now overlooked study <a href=\"https:\/\/www.wired.com\/story\/ai-agents-math-doesnt-add-up\/\" rel=\"noreferrer nofollow noopener\" target=\"_blank\">recently featured in Wired<\/a> claims to mathematically prove that large language models \u201care incapable of carrying out computational and agentic tasks beyond a certain complexity\u201d \u2014 that level of complexity being, crucially, pretty low.<\/p>\n<p class=\"article-paragraph skip\">The paper, which has not been peer reviewed, was written by Vishal Sikka, a former CTO at the German software giant SAP, and his son Varin Sikka. Sikka senior knows a thing or two about AI: he studied under John McCarthy, the Turing Award-winning computer scientist who literally founded the entire field of artificial intelligence, and in fact helped coin the very term.<\/p>\n<p class=\"article-paragraph skip\">\u201cThere is no way they can be reliable,\u201d Vishal Sikka told Wired.\u00a0<\/p>\n<p class=\"article-paragraph skip\">When asked by the interviewer, Sikka also agreed that we should forget about AI agents running nuclear power plants and other strident promises thrown around by AI boosters.<\/p>\n<p class=\"article-paragraph skip\">Ignore the rhetoric that tech CEOs spew onstage and pay attention to what the researchers that work for them are finding, and you\u2019ll find that even the AI industry agrees that the tech has some <a href=\"https:\/\/futurism.com\/openai-mistake-hallucinations\" rel=\"nofollow noopener\" target=\"_blank\">fundamental limitations baked into its architecture<\/a>. In September, for example, OpenAI scientists admitted that AI <a href=\"https:\/\/futurism.com\/fixing-hallucinations-destroy-chatgpt\" rel=\"nofollow noopener\" target=\"_blank\">hallucinations<\/a>, in which LLMs confidently make up facts, were still a pervasive problem even in increasingly advanced systems, and that model accuracy would \u201cnever\u201d reach 100 percent.<\/p>\n<p class=\"article-paragraph skip\">That would seemingly put a big dent in the feasibility of so-called AI agents, which are models designed to autonomously carry out tasks without human intervention, and which the industry universally decided last year would be its next big thing. Some companies that embraced AI agents to downsize their workforces quickly realized that the agents they <a href=\"https:\/\/futurism.com\/klarna-ai-automation-engineers\" rel=\"nofollow noopener\" target=\"_blank\">weren\u2019t anywhere near good enough to replace the outgoing humans<\/a>, perhaps because they hallucinated so much and could <a href=\"https:\/\/futurism.com\/artificial-intelligence\/paper-tested-ai-online-freelance-work\" rel=\"nofollow noopener\" target=\"_blank\">barely complete any of the tasks given to them<\/a>.<\/p>\n<p class=\"article-paragraph skip\">AI leaders insist that stronger guardrails external to the AI models can filter out the hallucinations. They may always be prone to hallucinating, but if these slip-ups are rare enough, then eventually companies will trust them to start doing tasks that they once entrusted to flesh and blood grunts. In the same paper that OpenAI researchers conceded that the models would never reach perfect accuracy, they also dismissed the idea that hallucinations are \u201cinevitable,\u201d because LLMs \u201ccan abstain when uncertain.\u201d (Despite that, you\u2019d be hard-pressed to find a single popular chatbot that actually does that, almost certainly because it would make the chatbots seem less impressive and less engaging to use.)<\/p>\n<p class=\"article-paragraph skip\">Even though he\u2019s adamant LLMs have a hard ceiling, Sikka agrees with figures in the AI industry who insist that hallucinations can be reined in.<\/p>\n<p class=\"article-paragraph skip\">\u201cOur paper is saying that a pure LLM has this inherent limitation \u2014 but at the same time it is true that you can build components around LLMs that overcome those limitations,\u201d he told Wired.<\/p>\n<p class=\"article-paragraph skip\">More on AI: <a href=\"https:\/\/futurism.com\/artificial-intelligence\/onlyfans-rival-ai-psychosis\" rel=\"nofollow noopener\" target=\"_blank\">OnlyFans Rival Seemingly Succumbs to AI Psychosis, Which We Dare You to Try Explain to Your Parents<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"Gaston Paris \/ Contributor A months-old but until now overlooked study recently featured in Wired claims to mathematically&hellip;\n","protected":false},"author":2,"featured_media":391913,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[20],"tags":[554,733,4308,86,56,54,55],"class_list":{"0":"post-391912","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-ai","9":"tag-artificial-intelligence","10":"tag-artificialintelligence","11":"tag-technology","12":"tag-uk","13":"tag-united-kingdom","14":"tag-unitedkingdom"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/posts\/391912","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/comments?post=391912"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/posts\/391912\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/media\/391913"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/media?parent=391912"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/categories?post=391912"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/uk\/wp-json\/wp\/v2\/tags?post=391912"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}