{"id":99943,"date":"2025-08-27T15:25:07","date_gmt":"2025-08-27T15:25:07","guid":{"rendered":"https:\/\/www.newsbeep.com\/au\/99943\/"},"modified":"2025-08-27T15:25:07","modified_gmt":"2025-08-27T15:25:07","slug":"scientists-just-developed-a-new-ai-modeled-on-the-human-brain-its-outperforming-llms-like-chatgpt-at-reasoning-tasks","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/au\/99943\/","title":{"rendered":"Scientists just developed a new AI modeled on the human brain \u2014 it&#8217;s outperforming LLMs like ChatGPT at reasoning tasks"},"content":{"rendered":"<p id=\"2248aeb0-dd99-4251-9536-dc5790db5eca\">Scientists have developed a new type of <a data-analytics-id=\"inline-link\" href=\"https:\/\/www.livescience.com\/technology\/artificial-intelligence\/what-is-artificial-intelligence-ai\" data-before-rewrite-localise=\"https:\/\/www.livescience.com\/technology\/artificial-intelligence\/what-is-artificial-intelligence-ai\" rel=\"nofollow noopener\" target=\"_blank\">artificial intelligence<\/a> (AI) model that can reason differently from most large language models (LLMs) like ChatGPT, resulting in much better performance in key benchmarks.<\/p>\n<p>The new reasoning AI, called a hierarchical reasoning model (HRM), is inspired by the <a data-analytics-id=\"inline-link\" href=\"https:\/\/pmc.ncbi.nlm.nih.gov\/articles\/PMC11665873\/\" target=\"_blank\" data-url=\"https:\/\/pmc.ncbi.nlm.nih.gov\/articles\/PMC11665873\/\" referrerpolicy=\"no-referrer-when-downgrade\" data-hl-processed=\"none\" rel=\"nofollow noopener\">hierarchical and multi-timescale processing<\/a> in the human brain \u2014 the way different brain regions integrate information over varying durations (from milliseconds to minutes).<\/p>\n<p><a id=\"elk-seasonal\" data-url=\"\" href=\"\" data-hl-processed=\"none\"\/><\/p>\n<p id=\"2248aeb0-dd99-4251-9536-dc5790db5eca-2\">Scientists at Sapient, an AI company in Singapore, say this reasoning model can achieve better performance and can work more efficiently. This is thanks to the model requiring fewer parameters and training examples.<\/p>\n<p>You may like<\/p>\n<p id=\"81a6a9c0-bebd-4a05-9b21-04c3757e7477\">The HRM model has 27 million parameters while using 1,000 training samples, the scientists said in a study uploaded June 26 to the preprint <a data-analytics-id=\"inline-link\" href=\"https:\/\/arxiv.org\/abs\/2506.21734\" target=\"_blank\" data-url=\"https:\/\/arxiv.org\/abs\/2506.21734\" referrerpolicy=\"no-referrer-when-downgrade\" data-hl-processed=\"none\" rel=\"nofollow noopener\">arXiv<\/a> database (which has yet to be peer-reviewed). In comparison, most advanced LLMs have billions or even trillions of parameters. Although an exact figure has not been made public, <a data-analytics-id=\"inline-link\" href=\"https:\/\/lifearchitect.ai\/gpt-5\/#:~:text=It&#039;s%20incredible.,around%203.5T%20parameters%20dense.\" target=\"_blank\" data-url=\"https:\/\/lifearchitect.ai\/gpt-5\/#:~:text=It&#039;s%20incredible.,around%203.5T%20parameters%20dense.\" referrerpolicy=\"no-referrer-when-downgrade\" data-hl-processed=\"none\" rel=\"nofollow noopener\">some estimates<\/a> suggest that the newly released GPT-5 has between 3 trillion and 5 trillion parameters.<\/p>\n<p><a id=\"elk-6f9308d0-a44a-4c9a-913d-aca34a89e603\" data-url=\"\" href=\"\" data-hl-processed=\"none\"\/>A new way of thinking for AI<\/p>\n<p id=\"c88f6c31-d10b-4618-825d-c5439edbdff1\">When the researchers tested HRM in the <a data-analytics-id=\"inline-link\" href=\"https:\/\/arcprize.org\/\" target=\"_blank\" data-url=\"https:\/\/arcprize.org\/\" referrerpolicy=\"no-referrer-when-downgrade\" data-hl-processed=\"none\" rel=\"nofollow noopener\">ARC-AGI benchmark<\/a> \u2014 a notoriously tough examination that aims to test how close models are to achieving <a data-analytics-id=\"inline-link\" href=\"https:\/\/www.livescience.com\/technology\/artificial-intelligence\/what-is-artificial-general-intelligence-agi\" data-before-rewrite-localise=\"https:\/\/www.livescience.com\/technology\/artificial-intelligence\/what-is-artificial-general-intelligence-agi\" rel=\"nofollow noopener\" target=\"_blank\">artificial general intelligence<\/a> (AGI) \u2014 the system achieved impressive results, according to the study.<\/p>\n<p>HRM scored 40.3% in ARC-AGI-1, compared with 34.5% for OpenAI&#8217;s o3-mini-high, 21.2% for Anthropic&#8217;s Claude 3.7 and 15.8% for Deepseek R1. In the tougher ARC-AGI-2 test, HRM scored 5% versus o3-mini-high&#8217;s 3%, Deepseek R1&#8217;s 1.3% and Claude 3.7&#8217;s 0.9%.<\/p>\n<p>Most advanced LLMs use chain-of-thought (CoT) reasoning, in which a complex problem is broken down into multiple, much simpler intermediate steps that are expressed in natural language. It emulates the human thought process by breaking down elaborate problems into digestible chunks.<\/p>\n<p class=\"newsletter-form__strapline\">Get the world\u2019s most fascinating discoveries delivered straight to your inbox.<\/p>\n<p>Related: <a data-analytics-id=\"inline-link\" href=\"https:\/\/www.livescience.com\/technology\/artificial-intelligence\/ai-is-entering-an-unprecedented-regime-should-we-stop-it-and-can-we-before-it-destroys-us\" data-before-rewrite-localise=\"https:\/\/www.livescience.com\/technology\/artificial-intelligence\/ai-is-entering-an-unprecedented-regime-should-we-stop-it-and-can-we-before-it-destroys-us\" rel=\"nofollow noopener\" target=\"_blank\">AI is entering an &#8216;unprecedented regime.&#8217; Should we stop it \u2014 and can we \u2014 before it destroys us?<\/a><\/p>\n<p>But the Sapient scientists argue in the study that CoT has key shortcomings \u2014 namely &#8220;brittle task decomposition, extensive data requirements, and high latency.&#8221;<\/p>\n<p>Instead, HRM executes sequential reasoning tasks in a single forward pass, without any explicit supervision of the intermediate steps, through two modules. One high-level module is responsible for slow, abstract planning, while a low-level module handles rapid and detailed computations. This is similar to the way in which the <a data-analytics-id=\"inline-link\" href=\"https:\/\/www.livescience.com\/29365-human-brain.html\" data-before-rewrite-localise=\"https:\/\/www.livescience.com\/29365-human-brain.html\" rel=\"nofollow noopener\" target=\"_blank\">human brain<\/a> processes information in different regions.<\/p>\n<p>It operates by applying iterative refinement \u2014 a computing technique that improves the accuracy of a solution by repeatedly refining an initial approximation \u2014 over several short bursts of &#8220;thinking.&#8221; Each burst considers whether the process of thinking should continue or be submitted as a &#8220;final&#8221; answer to the initial prompt.<\/p>\n<p id=\"dd09fd3d-5a36-4242-9b80-ddf4eda33228\">HRM achieved near-perfect performance on challenging tasks like complex Sudoku puzzles \u2014 which conventional LLMs could not accomplish \u2014 as well as excelling at optimal path-finding in mazes.<\/p>\n<p>The paper has not been peer-reviewed, but the organizers of the ARC-AGI benchmark attempted to recreate the results for themselves after the study scientists <a data-analytics-id=\"inline-link\" href=\"https:\/\/github.com\/sapientinc\/HRM?tab=readme-ov-file\" target=\"_blank\" data-url=\"https:\/\/github.com\/sapientinc\/HRM?tab=readme-ov-file\" referrerpolicy=\"no-referrer-when-downgrade\" data-hl-processed=\"none\" rel=\"nofollow noopener\">open-sourced their model on GitHub<\/a>.<\/p>\n<p>Although they reproduced the numbers, representatives said in a <a data-analytics-id=\"inline-link\" href=\"https:\/\/arcprize.org\/blog\/hrm-analysis\" target=\"_blank\" data-url=\"https:\/\/arcprize.org\/blog\/hrm-analysis\" referrerpolicy=\"no-referrer-when-downgrade\" data-hl-processed=\"none\" rel=\"nofollow noopener\">blog post<\/a>, they made some surprising findings, including that the hierarchical architecture had minimal performance impact \u2014 instead, there was an under-documented refinement process during training that drove substantial performance gains.<\/p>\n","protected":false},"excerpt":{"rendered":"Scientists have developed a new type of artificial intelligence (AI) model that can reason differently from most large&hellip;\n","protected":false},"author":2,"featured_media":99944,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[20],"tags":[256,254,255,64,63,105],"class_list":{"0":"post-99943","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-ai","9":"tag-artificial-intelligence","10":"tag-artificialintelligence","11":"tag-au","12":"tag-australia","13":"tag-technology"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts\/99943","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/comments?post=99943"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts\/99943\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/media\/99944"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/media?parent=99943"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/categories?post=99943"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/tags?post=99943"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}