{"id":121543,"date":"2025-09-05T20:41:03","date_gmt":"2025-09-05T20:41:03","guid":{"rendered":"https:\/\/www.newsbeep.com\/au\/121543\/"},"modified":"2025-09-05T20:41:03","modified_gmt":"2025-09-05T20:41:03","slug":"ai-growth-is-about-to-hit-a-memory-wall","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/au\/121543\/","title":{"rendered":"AI growth is about to hit a (memory) wall"},"content":{"rendered":"<p>                  AI chip and infrastructure orders continue to flood inBut memory bandwidth could be a limiting factoring in putting all that new compute power to efficient useTools can help improve available memory bandwidth, but there&#8217;s no cure all yet<\/p>\n<p>Orders for AI gear have been pouring in and not just at chip giant Nvidia either. AMD bagged deals with Meta, Microsoft, OpenAI and Oracle. Cisco <a href=\"https:\/\/www.fierce-network.com\/broadband\/cisco-bets-big-ai-profit-still-way\" target=\"_blank\" rel=\"nofollow noopener\">already has<\/a> $2 billion worth of AI orders this year. And just this week Broadcom <a href=\"https:\/\/www.investing.com\/news\/transcripts\/earnings-call-transcript-broadcom-q3-2025-sees-strong-earnings-stock-rises-93CH-4225766\" target=\"_blank\" rel=\"nofollow noopener\">revealed it landed<\/a> a $10 billion contract for AI racks based on its XPU chips. But the AI boom is about to hit a serious speed bump.<\/p>\n<p>We\u2019ve already written about the power, cooling and networking challenges AI data centers face. But there\u2019s another major problem that could hinder AI growth: memory.<\/p>\n<p>While GPUs get all the love, AI requires more than pure processing power. As we <a href=\"https:\/\/www.fierce-network.com\/cloud\/cpus-are-unsung-heroes-ai\" target=\"_blank\" rel=\"nofollow noopener\">recently pointed out<\/a>, CPUs will play a vital role in orchestrating workloads and data pipelines for AI applications. But memory bandwidth allows data to flow, and right now there\u2019s not nearly enough of it.<\/p>\n<p>\u201cMemory is very much a limiting factor in AI scale-out and performance,\u201d J. Gold Associates Founder Jack Gold explained. \u201cGPUs are often restricted in performance by the need to connect to external memory over interconnects that slow things down. So, anything that can bring memory closer\/faster to the GPU has a big performance improvement.\u201d<\/p>\n<p>How did this happen?<\/p>\n<p>JB Baker, VP of Products at ScaleFlux, told Fierce that while both processor and memory capabilities have grown exponentially in recent years, the latter hasn\u2019t grown at the same multiplier as the former.<\/p>\n<p>The result is a gap between how many calculations a chip can process in one second and how much memory bandwidth is available to send data to the chips to be processed. In other words, memory has become a bottleneck. (There&#8217;s a nifty little chart you can <a href=\"https:\/\/ayarlabs.com\/glossary\/memory-wall\/\" target=\"_blank\" rel=\"nofollow noopener\">check out here<\/a>, and an IEEE <a href=\"https:\/\/ieeexplore.ieee.org\/document\/10477550\" target=\"_blank\" rel=\"nofollow noopener\">paper on the Memory Wall here<\/a> if you&#8217;re feeling extra nerdy.)<\/p>\n<p>Baker said it\u2019s a bit like having a sprawling crop field next to a massive lake full of water but only having a tiny garden hose to get the water to the field.\u00a0<\/p>\n<p>\u201cThere\u2019s a lot of potential compute capacity that is going unutilized and at the same time it\u2019s burning power,\u201d he said. \u201cSo, it\u2019s not only that I lost out on things that those processors could have done, but they burn power sitting idle.\u201d<\/p>\n<p>Gold noted the issue isn\u2019t unique to GPUs \u2013 CPUs have run up against this wall as well. But the issue is coming to a head in the AI era, particularly in light of power constraints and a desire to massively scale AI deployments.<\/p>\n<p>Can it be fixed?<\/p>\n<p>Baker said right now there\u2019s no cure-all solution, but things are moving in the right direction.<\/p>\n<p>He pointed to compute express link (CXL) technology and solid state drive (SSD) advancements as tools that can help scale memory bandwidth. (Of course, both are products ScaleFlux offers, but others are working on the problem as well.<\/p>\n<p>Nvidia is tackling the issue through its NVLink and <a href=\"https:\/\/www.linkedin.com\/posts\/stephen-bates-8791263_gtc25-gtc25-gtc-activity-7309963278640631809-ffxG?utm_source=share&amp;utm_medium=member_desktop&amp;rcm=ACoAAATj4lMBESQfwvvNIfR7y15s_6RXelgwiA8\" target=\"_blank\" rel=\"nofollow noopener\">Storage-Next initiatives<\/a>, and Gold noted other players \u2013 including Intel spin-off Cornelis \u2013 are trying to speed things up in this space as well.<\/p>\n<p>In terms of what it all means for telcos and enterprises, Baker argued scaling AI will require a rebalancing of the capex equation to focus on more than just GPUs for on-prem deployments.<\/p>\n<p>\u201cIf you don\u2019t put the right amount of dollars into your memory, into your storage and your networking, you\u2019re going to waste the ones that you spend on GPUs,\u201d he concluded.<\/p>\n","protected":false},"excerpt":{"rendered":"AI chip and infrastructure orders continue to flood inBut memory bandwidth could be a limiting factoring in putting&hellip;\n","protected":false},"author":2,"featured_media":121544,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[21],"tags":[282,64,63,32754,281,2575,257,85701,18370,40170,2291,105],"class_list":{"0":"post-121543","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-computing","8":"tag-artificial-intelligence-ai","9":"tag-au","10":"tag-australia","11":"tag-broadcom","12":"tag-cloud","13":"tag-cloud-computing","14":"tag-computing","15":"tag-electrical-power-consumption","16":"tag-fierce-network-homepage","17":"tag-gpu","18":"tag-nvidia","19":"tag-technology"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts\/121543","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/comments?post=121543"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts\/121543\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/media\/121544"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/media?parent=121543"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/categories?post=121543"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/tags?post=121543"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}