{"id":28629,"date":"2025-07-28T23:27:08","date_gmt":"2025-07-28T23:27:08","guid":{"rendered":"https:\/\/www.newsbeep.com\/au\/28629\/"},"modified":"2025-07-28T23:27:08","modified_gmt":"2025-07-28T23:27:08","slug":"ai-progress-shifts-beyond-raw-data-and-computing-power","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/au\/28629\/","title":{"rendered":"AI Progress Shifts Beyond Raw Data and Computing Power"},"content":{"rendered":"<p>How an artificial intelligence (AI) model performs is critically important to the enterprises investing in and deploying the technology. But there\u2019s debate whether advancements in large language models are slowing.<\/p>\n<p>This debate centers around AI scaling laws.<\/p>\n<p>Popularized by OpenAI, the concept behind AI scaling laws is simply this: Larger models trained on more compute will yield better performance. OpenAI\u2019s 2020 <a class=\"editor-rtfLink\" href=\"https:\/\/arxiv.org\/pdf\/2001.08361\" target=\"_blank\" rel=\"noopener nofollow\">paper<\/a>, \u201cScaling Laws for Neural Language Models\u201d\u00a0was the first influential paper to demonstrate scaling laws.<\/p>\n<p>Google DeepMind\u2019s 2022 <a class=\"editor-rtfLink\" href=\"https:\/\/arxiv.org\/abs\/2203.15556\" target=\"_blank\" rel=\"noopener nofollow\">paper<\/a>, \u201cTraining Compute-Optimal Large Language Models\u201d\u00a0added a crucial insight: It demonstrated that data, instead of model size, and compute were the two key factors influencing a model\u2019s performance. Its Chinchilla model, which was less than half the size of GPT-3 but with four times more data, outperformed GPT-3.<\/p>\n<p>\u201cOver the past few years, AI labs have hit on what feels like a winning strategy: scaling more parameters, more data, more compute,\u201d said <a href=\"https:\/\/www.linkedin.com\/in\/garrytan\/\" target=\"_blank\" rel=\"noopener nofollow\">Garry Tan<\/a>, president of startup accelerator <a href=\"https:\/\/www.ycombinator.com\/\" target=\"_blank\" rel=\"noopener nofollow\">Y Combinator<\/a>, on its <a class=\"editor-rtfLink\" href=\"https:\/\/www.youtube.com\/watch?v=d6Ed5bZAtrM\" target=\"_blank\" rel=\"noopener nofollow\">video series<\/a> YC Decoded. \u201cKeep scaling your models and they keep improving.\u201d<\/p>\n<p>But there are signs that initial leaps in performance are decelerating.<\/p>\n<p>The two main fuels for scaling \u2014 data and computing \u2014 are becoming scarcer and more expensive, wrote <a href=\"https:\/\/www.linkedin.com\/in\/adnano\/\" target=\"_blank\" rel=\"noopener nofollow\">Adnan Masood<\/a>, <a href=\"https:\/\/www.ust.com\/\" target=\"_blank\" rel=\"noopener nofollow\">UST\u2019s<\/a> chief architect of AI and machine learning, in a <a class=\"editor-rtfLink\" href=\"https:\/\/medium.com\/%40adnanmasood\/is-there-a-wall-34d02dfd85f3\" target=\"_blank\" rel=\"noopener nofollow\">blog post<\/a>. \u201cThese trends strongly indicate a plateau in the current trajectory of large language models.\u201d<\/p>\n<p>For example, in knowledge quizzes, math word problems and coding tests, improvements are \u201cflattening,\u201d Masood said. He noted that in the knowledge quiz MMLU benchmark, GPT-3 recorded 43.9% and doubled to 86.4% for GPT-4 in 2023 but has since plateaued at 90% in 2024.<\/p>\n<p>\u201cIf the old scaling laws are beginning to lose their edge, what comes next?\u201d Tan asked.<\/p>\n<p>The answer from both Tan and Masood is that scaling laws are changing. AI model performance is still advancing, but it\u2019s now due to new techniques and not just making data and computing power larger.<\/p>\n<p>That\u2019s why OpenAI introduced its o1 and o3 AI reasoning models after the GPT-series, Tan said. For the o, or omni, models, OpenAI used \u201cchain of thought\u201d techniques to get the model to think through its answers. The result was improved performance. (OpenAI skipped naming the model o2 because it was the name of a telecom provider.)<\/p>\n<p>\u201cOpenAI researchers found that the longer o1 was able to think, the better it performed,\u201d Tan said. \u201cNow, with the recent release of its successor, o3, the sky seems to be the limit for this new paradigm of scaling LLMs.\u201d<\/p>\n<p>Tan said o3 \u201csmashed benchmarks that were previously considered far out of reach for AI.\u201d<\/p>\n<p>In the U.S., leading AI models retain the usage top spot for only about three weeks before being overtaken \u2014\u00a0especially by open\u2011source competitors, according to a June 2025 <a class=\"editor-rtfLink\" href=\"https:\/\/foundationmodelreport.ai\/2025.pdf\" target=\"_blank\" rel=\"noopener nofollow\">report<\/a> from Innovation Endeavors. The model cycle release remains fast \u2014\u00a0if no longer exponential. <\/p>\n<p>Scaling laws are not yet dying, but the AI community is preparing for a future that emphasizes smarter architectures, reasoning\u2011driven models, and use of distributed data sources.<\/p>\n<p>Read more: <\/p>\n<p><a class=\"editor-rtfLink\" href=\"https:\/\/www.pymnts.com\/news\/artificial-intelligence\/2025\/what-businesses-should-ask-when-choosing-an-ai-model\/\" target=\"_blank\" rel=\"noopener nofollow\">What Businesses Should Ask When Choosing an AI Model<\/a><\/p>\n<p><a class=\"editor-rtfLink\" href=\"https:\/\/www.pymnts.com\/artificial-intelligence-2\/2025\/court-rules-anthropic-doesnt-need-permission-to-train-ai-with-books\/\" target=\"_blank\" rel=\"noopener nofollow\">Court Rules Anthropic Doesn\u2019t Need Permission to Train AI With Books<\/a><\/p>\n<p><a class=\"editor-rtfLink\" href=\"https:\/\/www.pymnts.com\/artificial-intelligence-2\/2025\/meta-large-language-models-will-not-get-to-human-level-intelligence\/\" target=\"_blank\" rel=\"noopener nofollow\">Meta Chief AI Scientist Slams Quest for Human-Level Intelligence<\/a><\/p>\n<p>\u00a0<\/p>\n","protected":false},"excerpt":{"rendered":"How an artificial intelligence (AI) model performs is critically important to the enterprises investing in and deploying the&hellip;\n","protected":false},"author":2,"featured_media":28630,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[21],"tags":[256,27412,64,63,27413,257,19941,44,5044,6628,105],"class_list":{"0":"post-28629","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-computing","8":"tag-ai","9":"tag-ai-scaling-laws","10":"tag-au","11":"tag-australia","12":"tag-chain-of-thought","13":"tag-computing","14":"tag-large-language-models","15":"tag-news","16":"tag-openai","17":"tag-pymnts-news","18":"tag-technology"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts\/28629","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/comments?post=28629"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts\/28629\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/media\/28630"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/media?parent=28629"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/categories?post=28629"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/tags?post=28629"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}