{"id":376287,"date":"2026-04-12T19:06:12","date_gmt":"2026-04-12T19:06:12","guid":{"rendered":"https:\/\/www.newsbeep.com\/nz\/376287\/"},"modified":"2026-04-12T19:06:12","modified_gmt":"2026-04-12T19:06:12","slug":"even-more-good-news-for-the-future-of-neurosymbolic-ai","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/nz\/376287\/","title":{"rendered":"Even more good news for the future of neurosymbolic AI"},"content":{"rendered":"<p>Remember how <a href=\"https:\/\/open.substack.com\/pub\/garymarcus\/p\/a-knockout-blow-for-llms?r=8tdk6&amp;utm_medium=ios\" rel=\"nofollow noopener\" target=\"_blank\">that infamous Apple reasoning paper, The Illusion of Thinking showed that LLMs had trouble on the Tower of Hanoi<\/a>, working fine versions with small numbers of disk but breaking down altogether with larger numbers of disks?<\/p>\n<p><a target=\"_blank\" href=\"https:\/\/substackcdn.com\/image\/fetch\/$s_!aM7U!,f_auto,q_auto:good,fl_progressive:steep\/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59e36b79-818a-429f-92b0-fe1cd47254c9_677x298.jpeg\" data-component-name=\"Image2ToDOM\" rel=\"nofollow noopener\" class=\"image-link image2 is-viewable-img can-restack\"><img decoding=\"async\" src=\"https:\/\/www.newsbeep.com\/nz\/wp-content\/uploads\/2026\/04\/https:\/\/substack-post-media.s3.amazonaws.com\/public\/images\/59e36b79-818a-429f-92b0-fe1cd47254c9_677x.jpeg\" width=\"677\" height=\"298\" data-attrs=\"{&quot;src&quot;:&quot;https:\/\/substack-post-media.s3.amazonaws.com\/public\/images\/59e36b79-818a-429f-92b0-fe1cd47254c9_677x298.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:298,&quot;width&quot;:677,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:129308,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image\/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https:\/\/garymarcus.substack.com\/i\/193969990?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F59e36b79-818a-429f-92b0-fe1cd47254c9_677x298.jpeg&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}\" alt=\"\"   fetchpriority=\"high\" class=\"sizing-normal\"\/><\/a><\/p>\n<p>I wrote it up (quoting Josh Wolfe\u2019s legendary tweet about how \u201cApple [had] just GaryMarcus\u2019s LLM reasoning ability\u201d) in an essay called <a href=\"https:\/\/garymarcus.substack.com\/p\/a-knockout-blow-for-llms\" rel=\"nofollow noopener\" target=\"_blank\">A knockout blow for LLMs<\/a>, If you have forgotten the situation you can go there for a refresher. It was (and remains) the most popular essay I have posted here.<\/p>\n<p>In part I concluded<\/p>\n<p>What the Apple paper shows, most fundamentally, regardless of how you define AGI, is that LLMs are no substitute for good well-specified conventional algorithms. (They also can\u2019t play chess as well as conventional algorithms, can\u2019t fold proteins like special-purpose neurosymbolic hybrids, can\u2019t run databases as well as conventional databases, etc.)<\/p>\n<p>\u2026<\/p>\n<p>Worse, as the latest Apple papers shows, LLMs may well work on your easy test set (like Hanoi with 4 discs) and seduce you into thinking it has built a proper, generalizable solution when it does not.<\/p>\n<p>At least for the next decade, LLMs (with and without inference time \u201creasoning\u201d) will continue have their uses, \u2026. But anybody who thinks LLMs are a direct route to the sort AGI that could fundamentally transform society for the good is kidding themselves. This does not mean that the field of neural networks is dead, or that deep learning is dead. LLMs are just one form of deep learning, and maybe others \u2014 especially those that play nicer with symbols \u2013 will eventually thrive. <\/p>\n<p>At the time many LLM fans were hopping mad, convinced that the Apple paper was biased and must be flawed. <\/p>\n<p>Hundreds of thousands of them (maybe more) circulated a critique of the Apple paper called \u201cThe illusion of the illusion\u201d that turned out to be AI-generated and  <a href=\"https:\/\/x.com\/blancheminerva\/status\/1933835315594104859?s=61\" rel=\"nofollow\">error-ridden<\/a>, literally written as <a href=\"https:\/\/x.com\/lxrjl\/status\/1934228650120872012?s=61\" rel=\"nofollow\">joke<\/a>, in the tradition of the <a href=\"https:\/\/en.wikipedia.org\/wiki\/Sokal_affair\" rel=\"nofollow noopener\" target=\"_blank\">Sokal hoax<\/a>.  <\/p>\n<p>In reality,  <a href=\"https:\/\/garymarcus.substack.com\/p\/five-quick-updates-about-that-apple?utm_source=publication-search\" rel=\"nofollow noopener\" target=\"_blank\">the replies to the Apple reasoning paper were weak sauce<\/a> (see also <a href=\"https:\/\/garymarcus.substack.com\/p\/seven-replies-to-the-viral-apple?utm_source=publication-search\" rel=\"nofollow noopener\" target=\"_blank\">this<\/a>). <\/p>\n<p>Another paper later last summer on \u201cthe mirage of reasoning\u201d made similar arguments around <a href=\"https:\/\/arxiv.org\/pdf\/2508.01191\" rel=\"nofollow noopener\" target=\"_blank\">chain of reasoning models<\/a>, and several other papers, like the recent <a href=\"https:\/\/open.substack.com\/pub\/garymarcus\/p\/breaking-llm-reasoning-continues?utm_campaign=post-expanded-share&amp;utm_medium=web\" rel=\"nofollow noopener\" target=\"_blank\">Stanford taxonomy of LLM reasoning errors,<\/a> have since extended the point.  LLMs have persistent weaknesses around planning, reasoning and generalization.  <\/p>\n<p>Which is a key part of why many of us have always urged for neurosymbolic hybrids, where neural networks do pattern recognition, and classical AI techniques help out around planning and reasoning.<\/p>\n<p><a href=\"https:\/\/arxiv.org\/abs\/2602.19260\" rel=\"nofollow noopener\" target=\"_blank\">A new paper from Tufts<\/a> (released in late February, but just shared with me yesterday) picks up where the Apple reasoning paper left off, and does three interesting things. <\/p>\n<p>First it (conceptually) replicates and extends the Apple reasoning paper, by showing that a newer variant on LLMs, called VLAs \u2013Vision-Language-Action models, which are increasingly popular in robotics \u2014 also suffer from the same generalization problems on the Tower Hanoi, similarly working fine on small versions of the problem but breaking down as more rings get added. (Humans who master Hanoi develop more robust solutions.)<\/p>\n<p>Second, they show that a neurosymbolic hybrid model, which integrates both a neural networks (for pattern recognition) and a symbolic planner generalizes far better.: \u201cOn the 3-block task, the neuro-symbolic model achieves 95% success compared to 34% for the best-performing VLA. The neuro-symbolic model also generalizes to an unseen 4-block variant (78% success), whereas both VLAs fail to complete the task\u201d.<\/p>\n<p>Third, they show that the neurosymbolic hybrid was vastly more efficent in terms of energy use &#8211; by nearly two orders of magnitude. LLMs are an efficient way to pattern recognition where perfect results are not required, but an inefficient way to reason a plan. Different tools for different jobs. A good hybrid neurosymbolic system is about picking the right constellation of tools for a given job.<\/p>\n<p>You can read the full paper by Timothy Duggan, Pierrick Lorang, Hong Lu, and Matthias Scheutz, which very much vindicates the Apple reasoning paper, <a href=\"https:\/\/arxiv.org\/abs\/2602.19260\" rel=\"nofollow noopener\" target=\"_blank\">here<\/a>.<\/p>\n<p><a target=\"_blank\" href=\"https:\/\/substackcdn.com\/image\/fetch\/$s_!qHLQ!,f_auto,q_auto:good,fl_progressive:steep\/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff85529f5-92bb-4758-94f4-60b1940564e7_1873x837.png\" data-component-name=\"Image2ToDOM\" rel=\"nofollow noopener\" class=\"image-link image2 is-viewable-img can-restack\"><img decoding=\"async\" src=\"https:\/\/www.newsbeep.com\/nz\/wp-content\/uploads\/2026\/04\/https:\/\/substack-post-media.s3.amazonaws.com\/public\/images\/f85529f5-92bb-4758-94f4-60b1940564e7_1873.jpeg\" width=\"1873\" height=\"837\" data-attrs=\"{&quot;src&quot;:&quot;https:\/\/substack-post-media.s3.amazonaws.com\/public\/images\/f85529f5-92bb-4758-94f4-60b1940564e7_1873x837.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:837,&quot;width&quot;:1873,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:247010,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image\/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https:\/\/garymarcus.substack.com\/i\/193969990?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F44328d28-d85d-41e6-a0b3-c8f2d960aa45_1977x837.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}\" alt=\"\"   loading=\"lazy\" class=\"sizing-normal\"\/><\/a><\/p>\n<p>\u00a7<\/p>\n<p>Still, as I wrote yesterday (and also six years ago, in <a href=\"https:\/\/arxiv.org\/abs\/2002.06177\" rel=\"nofollow noopener\" target=\"_blank\">The Next Decade in AI<\/a>) taking a neurosymbolic approach is not a panacea. It\u2019s just a starting point. <\/p>\n<p>For example, the specific model here is purpose-built; we would like a general purpose system that can induce what is needed for the particular occasion; that\u2019s not this. Claude Code (in particular version 4.6), which does seem genuinely <a href=\"https:\/\/garymarcus.substack.com\/p\/the-biggest-advance-in-ai-since-the?utm_source=publication-search\" rel=\"nofollow noopener\" target=\"_blank\">impressive<\/a> to me and a number of friends I trust,  still makes plenty of mistakes, and still can\u2019t be trusted. As one of those friend, deeply positive but still measured, put it, Claude Code is still best treated as a tool, rather than a complete solution. It\u2019s obviously not AGI.<\/p>\n<p>What we have now are lights into the future, not full answers. <\/p>\n<p>But there is more and more reason to think that neurosymbolic AI is a stronger starting point, with potential for better generalization and more efficient performance. <\/p>\n<p>As a society we could spend another trillion dollars pursue LLMs, or a trillion dollars seeking new, hybrid approaches. I for one know which I would pick.<\/p>\n","protected":false},"excerpt":{"rendered":"Remember how that infamous Apple reasoning paper, The Illusion of Thinking showed that LLMs had trouble on the&hellip;\n","protected":false},"author":2,"featured_media":376288,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[20],"tags":[365,363,364,111,139,69,145],"class_list":{"0":"post-376287","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-ai","9":"tag-artificial-intelligence","10":"tag-artificialintelligence","11":"tag-new-zealand","12":"tag-newzealand","13":"tag-nz","14":"tag-technology"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/nz\/wp-json\/wp\/v2\/posts\/376287","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/nz\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/nz\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/nz\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/nz\/wp-json\/wp\/v2\/comments?post=376287"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/nz\/wp-json\/wp\/v2\/posts\/376287\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/nz\/wp-json\/wp\/v2\/media\/376288"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/nz\/wp-json\/wp\/v2\/media?parent=376287"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/nz\/wp-json\/wp\/v2\/categories?post=376287"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/nz\/wp-json\/wp\/v2\/tags?post=376287"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}