Sunday, April 19, 2026

AI Growth Fuels DRAM Scarcity and Worth Surge

If it feels nowadays as if the whole lot in know-how is about AI, that’s as a result of it’s. And nowhere is that extra true than available in the market for pc reminiscence. Demand, and profitability, for the kind of DRAM used to feed GPUs and different accelerators in AI information facilities is so enormous that it’s diverting away provide of reminiscence for different makes use of and inflicting costs to skyrocket. In keeping with Counterpoint Analysis, DRAM costs have risen 80-90 precent to this point this quarter.

The most important AI {hardware} firms say they’ve secured their chips out so far as 2028, however that leaves everyone else—makers of PCs, client gizmos, and the whole lot else that should briefly retailer a billion bits—scrambling to take care of scarce provide and inflated costs.

How did the electronics trade get into this mess, and extra importantly, how will it get out? IEEE Spectrum requested economists and reminiscence specialists to elucidate. They are saying at this time’s scenario is the results of a collision between the DRAM trade’s historic increase and bust cycle and an AI {hardware} infrastructure build-out that’s with out precedent in its scale. And, barring some main collapse within the AI sector, it should take years for brand spanking new capability and new know-how to convey provide in keeping with demand. Costs may keep excessive even then.

To grasp each ends of the story, you must know the primary wrongdoer within the provide and demand swing, high-bandwidth reminiscence, or HBM.

What’s HBM?

HBM is the DRAM trade’s try to short-circuit the slowing tempo of Moore’s Legislation by utilizing 3D chip packaging know-how. Every HBM chip is made up of as many as 12 thinned-down DRAM chips known as dies. Every die comprises quite a few vertical connections known as by silicon vias (TSVs). The dies are piled atop one another and related by arrays of microscopic solder balls aligned to the TSVs. This DRAM tower—nicely, at about 750 micrometers thick, it’s extra of a brutalist office-block than a tower—is then stacked atop what’s known as the bottom die, which shuttles bits between the reminiscence dies and the processor.

This complicated piece of know-how is then set inside a millimeter of a GPU or different AI accelerator, to which it’s linked by as many as 2,048 micrometer-scale connections. HBMs are connected on two sides of the processor, and the GPU and reminiscence are packaged collectively as a single unit.

The thought behind such a decent, highly-connected squeeze with the GPU is to knock down what’s known as the reminiscence wall. That’s the barrier in power and time of bringing the terabytes per second of information wanted to run giant language fashions into the GPU. Reminiscence bandwidth is a key limiter to how briskly LLMs can run.

As a know-how, HBM has been round for greater than 10 years, and DRAM makers have been busy boosting its functionality.

As the dimensions of AI fashions has grown, so has HBM’s significance to the GPU. However that’s come at a price. SemiAnalysis estimates that HBM typically prices thrice as a lot as different sorts of reminiscence and constitutes 50 % or extra of the price of the packaged GPU.

Origins of the reminiscence chip scarcity

Reminiscence and storage trade watchers agree that DRAM is a extremely cyclical trade with enormous booms and devastating busts. With new fabs costing US $15 billion or extra, companies are extraordinarily reluctant to broaden and will solely have the money to take action throughout increase occasions, explains Thomas Coughlin, a storage and reminiscence skilled and president of Coughlin Associates. However constructing such a fab and getting it up and working can take 18 months or extra, virtually making certain that new capability arrives nicely previous the preliminary surge in demand, flooding the market and miserable costs.

The origins of at this time’s cycle, says Coughlin, go all the way in which again to the chip provide panic surrounding the COVID-19 pandemic . To keep away from supply-chain stumbles and help the fast shift to distant work, hyperscalers—information heart giants like Amazon, Google, and Microsoft—purchased up enormous inventories of reminiscence and storage, boosting costs, he notes.

However then provide turned extra common and information heart growth fell off in 2022, inflicting reminiscence and storage costs to plummet. This recession continued into 2023, and even resulted in massive reminiscence and storage firms resembling Samsung slicing manufacturing by 50 % to attempt to hold costs from going under the prices of producing, says Coughlin. It was a uncommon and pretty determined transfer, as a result of firms sometimes must run crops at full capability simply to earn again their worth.

After a restoration started in late 2023, “all of the reminiscence and storage firms had been very cautious of accelerating their manufacturing capability once more,” says Coughlin. “Thus there was little or no funding in new manufacturing capability in 2024 and thru most of 2025.”

chart visualization

The AI information heart increase

That lack of recent funding is colliding headlong with an enormous increase in demand from new information facilities. Globally, there are practically 2,000 new information facilities both deliberate or beneath development proper now, in keeping with Information Heart Map. In the event that they’re all constructed, it could symbolize a 20 % soar within the world provide, which stands at round 9,000 services now.

If the present build-out continues at tempo, McKinsey predicts firms will spend $7 trillion by 2030, with the majority of that—$5.2 trillion—going to AI-focused information facilities. Of that chunk, $3.3 billion will go towards servers, information storage, and community gear, the agency predicts.

The most important beneficiary to this point of the AI information heart increase is definitely GPU-maker Nvidia. Income for its information heart enterprise went from barely a billion within the ultimate quarter of 2019 to $51 billion within the quarter that resulted in October 2025. Over this era, its server GPUs have demanded not simply increasingly gigabytes of DRAM however an growing variety of DRAM chips. The not too long ago launched B300 makes use of eight HBM chips, every of which is a stack of 12 DRAM dies. Opponents’ use of HBM has largely mirrored Nvidia’s. AMD’s MI350 GPU, for instance, additionally makes use of eight, 12-die chips.

chart visualization

With a lot demand, an growing fraction of the income for DRAM makers comes from HBM. Micron—the quantity three producer behind SK Hynix and Samsung—reported that HBM and different cloud-related reminiscence went from being 17 % of its DRAM income in 2023 to just about 50 % in 2025.

Micron predicts the entire marketplace for HBM will develop from $35 billion in 2025 to $100 billion by 2028—a determine bigger than your complete DRAM market in 2024, CEO Sanjay Mehrotra instructed analysts in December. It’s reaching that determine two years sooner than Micron had beforehand anticipated. Throughout the trade, demand will outstrip provide “considerably… for the foreseeable future,” he mentioned.

chart visualization

Future DRAM provide and know-how

“There are two methods to deal with provide points with DRAM: with innovation or with constructing extra fabs,” explains Mina Kim, an economist with the Mkecon Insights. “As DRAM scaling has turn into harder, the trade has turned to superior packaging… which is simply utilizing extra DRAM.”

Micron, Samsung, and SK Hynix mixed make up the overwhelming majority of the reminiscence and storage markets, and all three have new fabs and services within the works. Nevertheless, these are unlikely to contribute meaningfully to bringing down costs.

Micron is within the means of constructing an HBM fab in Singapore that must be in manufacturing in 2027. And it’s retooling a fab it bought from PSMC in Taiwan that may start manufacturing within the second half of 2027. Final month, Micron broke floor on what will likely be a DRAM fab complicated in Onondaga County, N.Y. It is not going to be in full manufacturing till 2030.

Samsung plans to begin producing at a brand new plant in Pyeongtaek, South Korea in 2028.

SK Hynix is constructing HBM and packaging services in West Lafayette, Indiana set to start manufacturing by the top of 2028, and an HBM fab it’s constructing in Cheongju must be full in 2027.

Talking of his sense of the DRAM market, Intel CEO Lip-Bu Tan instructed attendees on the Cisco AI Summit final week: “There’s no reduction till 2028.”

With these expansions unable to contribute for a number of years, different components will likely be wanted to extend provide. “Aid will come from a mixture of incremental capability expansions by current DRAM leaders, yield enhancements in superior packaging, and a broader diversification of provide chains,” says Shawn DuBravac , chief economist for the World Electronics Affiliation (previously the IPC). “New fabs will assist on the margin, however the quicker beneficial properties will come from course of studying, higher [DRAM] stacking effectivity, and tighter coordination between reminiscence suppliers and AI chip designers.”

So, will costs come down as soon as a few of these new crops come on line? Don’t wager on it. “Usually, economists discover that costs come down far more slowly and reluctantly than they go up. DRAM at this time is unlikely to be an exception to this normal statement, particularly given the insatiable demand for compute,” says Kim.

Within the meantime, applied sciences are within the works that might make HBM an excellent larger client of silicon. The usual for HBM4 can accommodate 16 stacked DRAM dies, regardless that at this time’s chips solely use 12 dies. Attending to 16 has so much to do with the chip stacking know-how. Conducting warmth by the HBM “layer cake” of silicon, solder, and help materials is a key limiter to going greater and in repositioning HBM contained in the package deal to get much more bandwidth.

SK Hynix claims a warmth conduction benefit by a producing course of known as superior MR-MUF (mass reflow molded underfill). Additional out, another chip stacking know-how known as hybrid bonding may assist warmth conduction by lowering the die-to-die vertical distance primarily to zero. In 2024, researchers at Samsung proved they might produce a 16-high stack with hybrid bonding, they usually advised that 20 dies was not out of attain.

From Your Web site Articles

Associated Articles Across the Net

Related Articles

LEAVE A REPLY

Please enter your comment!
Please enter your name here

Latest Articles