{"id":312935,"date":"2025-11-27T21:14:14","date_gmt":"2025-11-27T21:14:14","guid":{"rendered":"https:\/\/www.newsbeep.com\/au\/312935\/"},"modified":"2025-11-27T21:14:14","modified_gmt":"2025-11-27T21:14:14","slug":"can-bigger-is-better-scaling-laws-keep-ai-improving-forever-history-says-we-cant-be-too-sure","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/au\/312935\/","title":{"rendered":"Can bigger-is-better \u2018scaling laws\u2019 keep AI improving forever? History says we can\u2019t be too sure"},"content":{"rendered":"<p>OpenAI chief executive Sam Altman \u2013 perhaps the most prominent face of the artificial intelligence (AI) boom that accelerated with the launch of ChatGPT in 2022 \u2013 loves scaling laws. <\/p>\n<p>These widely admired rules of thumb linking the size of an AI model with its capabilities inform much of the headlong rush among the AI industry to buy up powerful computer chips, build unimaginably large data centres, and <a href=\"https:\/\/theconversation.com\/ai-is-consuming-more-power-than-the-grid-can-handle-nuclear-might-be-the-answer-258677\" rel=\"nofollow noopener\" target=\"_blank\">re-open shuttered nuclear plants<\/a>.<\/p>\n<p>As Altman argued in <a href=\"https:\/\/blog.samaltman.com\/three-observations\" rel=\"nofollow noopener\" target=\"_blank\">a blog post earlier this year<\/a>, the thinking is that the \u201cintelligence\u201d of an AI model \u201croughly equals the log of the resources used to train and run it\u201d \u2013 meaning you can steadily produce <a href=\"https:\/\/www.commerce.senate.gov\/services\/files\/6B937B74-31EE-4777-%20B004-3D6DC0DC3FBA\" rel=\"nofollow noopener\" target=\"_blank\">better performance<\/a> by exponentially increasing the scale of data and computing power involved.<\/p>\n<p>First <a href=\"https:\/\/arxiv.org\/abs\/2001.08361\" rel=\"nofollow noopener\" target=\"_blank\">observed in 2020<\/a> and <a href=\"https:\/\/mbrenndoerfer.com\/writing\/chinchilla-scaling-laws-compute-optimal-training-resource-allocation\" rel=\"nofollow noopener\" target=\"_blank\">further refined<\/a> in 2022, the scaling laws for large language models (LLMs) come from drawing lines on charts of experimental data. For engineers, they give a simple formula that tells you how big to build the next model and what performance increase to expect.<\/p>\n<p>Will the scaling laws keep on scaling as AI models get bigger and bigger? AI companies are betting hundreds of billions of dollars that they will \u2013 but history suggests it is not always so simple.<\/p>\n<p>Scaling laws aren\u2019t just for AI<\/p>\n<p>Scaling laws can be wonderful. Modern aerodynamics is built on them, for example.<\/p>\n<p>Using an elegant piece of mathematics called the <a href=\"https:\/\/en.wikipedia.org\/wiki\/Buckingham_%CF%80_theorem\" rel=\"nofollow noopener\" target=\"_blank\">Buckingham \u03c0 theorem<\/a>, engineers discovered how to compare small models in wind tunnels or test basins with full-scale planes and ships by making sure some <a href=\"https:\/\/en.wikipedia.org\/wiki\/Similitude\" rel=\"nofollow noopener\" target=\"_blank\">key numbers<\/a> matched up. <\/p>\n<p>Those scaling ideas inform the design of almost everything that flies or floats, as well as industrial fans and pumps. <\/p>\n<p>Another famous scaling idea underpinned the boom decades of the silicon chip revolution.  Moore\u2019s law \u2013 the idea that the number of the tiny switches called transistors on a microchip would double every two years or so \u2013 helped designers create the small, powerful computing technology we have today.<\/p>\n<p>But there\u2019s a catch: not all \u201cscaling laws\u201d are laws of nature. Some are purely mathematical and can hold indefinitely. Others are just lines fitted to data that work beautifully until you stray too far from the circumstances where they were measured or designed.<\/p>\n<p>When scaling laws break down<\/p>\n<p>History is littered with painful reminders of scaling laws that broke. A classic example is the collapse of the <a href=\"https:\/\/www.livescience.com\/technology\/engineering\/science-history-the-%20tacoma-narrows-bridge-collapses-forcing-a-complete-rethink-in-structural-%20engineering-nov-7-1940\" rel=\"nofollow noopener\" target=\"_blank\">Tacoma Narrows Bridge<\/a> in 1940. <\/p>\n<p>The bridge was designed by scaling up what had worked for smaller bridges to something longer and slimmer. Engineers assumed the same scaling arguments would hold: if a certain ratio of stiffness to bridge length worked before, it should work again. <\/p>\n<p>Instead, moderate winds set off an unexpected instability called aeroelastic flutter. The bridge deck tore itself apart, collapsing just four months after opening.<\/p>\n<p>Likewise, even the \u201claws\u201d of microchip manufacturing had an expiry date. For decades, Moore\u2019s law (transistor counts doubling every couple of years) and Dennard scaling (a larger number of smaller transistors running faster while using the same amount of power) were astonishingly reliable guides for chip design and industry roadmaps. <\/p>\n<p>As transistors became small enough to be measured in nanometres, however, those neat scaling rules began to <a href=\"https:\/\/cap.csail.mit.edu\/death-moores-law-what-it-means-and-what-might-%20fill-gap-going-forward\" rel=\"nofollow noopener\" target=\"_blank\">collide<\/a> with hard physical limits.<\/p>\n<p>When transistor gates shrank to just a few atoms thick, they started leaking current and behaving unpredictably. The operating voltages could also no longer be reduced with being lost in background noise. <\/p>\n<p>Eventually, shrinking was no longer the way forward. Chips have still grown more powerful, but now through new designs rather than just scaling down.<\/p>\n<p>Laws of nature or rules of thumb?<\/p>\n<p>The language-model scaling curves that Altman celebrates are real, and so far they\u2019ve been extraordinarily useful. <\/p>\n<p>They told researchers that models would keep getting better if you fed them enough data and computing power. They also showed earlier systems were <a href=\"https:\/\/arxiv.org\/abs\/2001.08361\" rel=\"nofollow noopener\" target=\"_blank\">not fundamentally limited<\/a> \u2013 they just hadn\u2019t had enough resources thrown at them.<\/p>\n<p>But these are undoubtedly curves that have been fit to data. They are less like the derived mathematical scaling laws used in aerodynamics and more like the useful rules of thumb used in microchip design \u2013 and that means they likely won\u2019t work forever.<\/p>\n<p>The language model scaling rules don\u2019t necessarily encode real-world problems such as limits to the availability of high-quality data for training, or the difficulty of getting AI to deal with novel tasks \u2013 let alone safety constraints or the economic difficulties of building data centres and power grids. There is no law of nature or theorem guaranteeing that \u201cintelligence scales\u201d forever.<\/p>\n<p>Investing in the curves<\/p>\n<p>So far, the scaling curves for AI look pretty smooth \u2013 but the financial curves are a different story.<\/p>\n<p>Deutsche Bank <a href=\"https:\/\/aimagazine.com\/news\/deutsche-bank-why-the-ai-boom-risks-a-us-800bn-shortfall\" rel=\"nofollow noopener\" target=\"_blank\">recently warned<\/a> of an AI \u201cfunding gap\u201d based on Bain Capital estimates of a US$800 billion mismatch between projected AI revenues and the investment in chips, data centres and power that would be needed to keep current growth going. <\/p>\n<p>JP Morgan, for their part, has <a href=\"https:\/\/www.tomshardware.com\/tech-industry\/artificial-%20intelligence\/usd650-billion-in-annual-revenue-required-to-deliver-10-percent-%20return-on-ai-buildout-investment-j-p-morgan-claims-equivalent-to-usd35-%20payment-from-every-iphone-user-or-usd180-from-every-netflix-subscriber-in-%20perpetuity\" rel=\"nofollow noopener\" target=\"_blank\">estimated<\/a> that the broader AI sector might need around US$650 billion in annual revenue just to earn a modest 10% return on the planned build-out of AI infrastructure.<\/p>\n<p>We\u2019re still finding out which kind of law governs frontier LLMs. The realities may keep playing along with the current scaling rules; or new bottlenecks \u2013 data, energy, users\u2019 willingness to pay \u2013 may bend the curve. <\/p>\n<p>Altman\u2019s bet is that the LLM scaling laws will continue. If that\u2019s so, it may be worth building enormous amounts of computing power because the gains are predictable. On the other hand, the banks\u2019 growing unease is a reminder that some scaling stories can turn out to be Tacoma Narrows: beautiful curves in one context, hiding a nasty surprise in the next.<\/p>\n","protected":false},"excerpt":{"rendered":"OpenAI chief executive Sam Altman \u2013 perhaps the most prominent face of the artificial intelligence (AI) boom that&hellip;\n","protected":false},"author":2,"featured_media":312936,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[20],"tags":[256,254,255,64,63,105],"class_list":{"0":"post-312935","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-artificial-intelligence","8":"tag-ai","9":"tag-artificial-intelligence","10":"tag-artificialintelligence","11":"tag-au","12":"tag-australia","13":"tag-technology"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts\/312935","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/comments?post=312935"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts\/312935\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/media\/312936"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/media?parent=312935"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/categories?post=312935"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/tags?post=312935"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}