Eugene Mymrin | Second | Getty Photos
All computing gadgets require an element referred to as reminiscence, or RAM, for short-term information storage, however this 12 months, there will not be sufficient of those important elements to fulfill worldwide demand.
That is as a result of firms like Nvidia, Superior Micro Gadgets and Google want a lot RAM for his or her synthetic intelligence chips, and people firms are the primary ones in line for the elements.
Three main reminiscence distributors — Micron, SK Hynix and Samsung Electronics — make up almost the complete RAM market, and their companies are benefitting from the surge in demand.
“We have now seen a really sharp, vital surge in demand for reminiscence, and it has far outpaced our means to produce that reminiscence and, in our estimation, the availability functionality of the entire reminiscence trade,” Micron enterprise chief Sumit Sadana informed CNBC this week on the CES commerce present in Las Vegas.
Micron’s inventory is up 247% over the previous 12 months, and the corporate reported that internet revenue almost tripled in the newest quarter. Samsung this week mentioned that it expects its December quarter working revenue to almost triple as effectively. In the meantime, SK Hynix is contemplating a U.S. itemizing as its inventory worth in South Korea surges, and in October, the corporate mentioned it had secured demand for its complete 2026 RAM manufacturing capability.
Now, costs for reminiscence are rising.
TrendForce, a Taipei-based researcher that carefully covers the reminiscence market, this week mentioned it expects common DRAM reminiscence costs to rise between 50% and 55% this quarter versus the fourth quarter of 2025. TrendForce analyst Tom Hsu informed CNBC that sort of improve for reminiscence costs was “unprecedented.”
Three-to-one foundation
Chipmakers like Nvidia encompass the a part of the chip that does the computation — the graphics processing unit, or GPU — with a number of blocks of a quick, specialised part referred to as high-bandwidth reminiscence, or HBM, Sadana mentioned. HBM is usually seen when chipmakers maintain up their new chips. Micron provides reminiscence to each Nvidia and AMD, the 2 main GPU makers.
Nvidia’s Rubin GPU, which lately entered manufacturing, comes with as much as 288 gigabytes of next-generation HBM4 reminiscence per chip. HBM is put in in eight seen blocks above and under the processor, and that GPU might be bought as a part of single server rack referred to as NVL72, which fittingly combines 72 of these GPUs right into a single system. By comparability, smartphones usually include 8 or 12GB of lower-powered DDR reminiscence.
Nvidia founder and CEO Jensen Huang introduces the Rubin GPU and the Vera CPU as he speaks throughout Nvidia Dwell at CES 2026 forward of the annual Shopper Electronics Present in Las Vegas, Nevada, on Jan. 5, 2026.
Patrick T. Fallon | AFP | Getty Photos
However the HBM reminiscence that AI chips want is far more demanding than the RAM used for shoppers’ laptops and smartphones. HBM is designed for high-bandwidth specs required by AI chips, and it is produced in a sophisticated course of the place Micron stacks 12 to 16 layers of reminiscence on a single chip, turning it right into a “dice.”
When Micron makes one little bit of HBM reminiscence, it has to forgo making three bits of extra typical reminiscence for different gadgets.
“As we improve HBM provide, it leaves much less reminiscence left over for the non-HBM portion of the market, due to this three-to-one foundation,” Sadana mentioned.
Hsu, the TrendForce analyst, mentioned that reminiscence makers are favoring server and HBM functions over different shoppers as a result of there’s increased potential for development in demand in that enterprise and cloud service suppliers are much less price-sensitive.
In December, Micron mentioned it might discontinue part of its enterprise that aimed to supply reminiscence for shopper PC builders so the corporate might save provide for AI chips and servers.
Some contained in the tech trade are marveling at how a lot and the way rapidly the value of RAM for shoppers has elevated.
Dean Beeler, co-founder and tech chief at Juice Labs, mentioned that a couple of months in the past, he loaded up his pc with 256GB of RAM, the utmost quantity that present shopper motherboards assist. That price him about $300 on the time.
“Who knew that will find yourself being ~$3,000 of RAM just some months later,” he posted on Fb on Monday.
‘Reminiscence wall’
AI researchers began to see reminiscence as a bottleneck simply earlier than OpenAI’s ChatGPT hit the market in late 2022, mentioned Majestic Labs co-founder Sha Rabii, an entrepreneur who beforehand labored on silicon at Google and Meta.
Prior AI programs have been designed for fashions like convolutional neural networks, which require much less reminiscence than giant language fashions, or LLMs, which might be well-liked right this moment, Rabii mentioned.
Whereas AI chips themselves have been getting a lot quicker, reminiscence has not, he mentioned, which results in highly effective GPUs ready round to get the information wanted to run LLMs.
“Your efficiency is proscribed by the quantity of reminiscence and the velocity of the reminiscence that you’ve got, and for those who maintain including extra GPUs, it isn’t a win,” Rabii mentioned.
The AI trade refers to this because the “reminiscence wall.”
Erik Isakson | Digitalvision | Getty Photos
“The processor spends extra time simply twiddling its thumbs, ready for information,” Micron’s Sadana mentioned.
Extra and quicker reminiscence signifies that AI programs can run larger fashions, serve extra clients concurrently and add “context home windows” that enable chatbots and different LLMs to recollect earlier conversations with customers, which provides a contact of personalization to the expertise.
Majestic Labs is designing an AI system for inference with 128 terabytes of reminiscence, or about 100 instances extra reminiscence than some present AI programs, Rabii mentioned, including that the corporate plans to eschew HBM reminiscence for lower-cost choices. Rabii mentioned the extra RAM and structure assist within the design will allow its computer systems to assist considerably extra customers on the similar time than different AI servers whereas utilizing much less energy.
Offered out for 2026
Wall Avenue has been asking firms within the shopper electronics enterprise, like Apple and Dell Applied sciences, how they are going to deal with the reminiscence scarcity and in the event that they is perhaps compelled to lift costs or minimize margins. As of late, reminiscence accounts for about 20% of the {hardware} prices of a laptop computer, Hsu mentioned. That is up from between 10% and 18% within the first half of 2025.
In October, Apple finance chief Kevan Parekh informed analysts that his firm was seeing a “slight tailwind” on reminiscence costs however he downplayed it as “nothing actually to notice there.”
However in November, Dell mentioned it anticipated its price foundation for all of its merchandise to go up because of the reminiscence scarcity. COO Jefferey Clarke informed analysts that Dell deliberate to alter its mixture of configurations to attenuate the value impacts, however he mentioned the scarcity will seemingly have an effect on retail costs for gadgets.
“I do not see how this is not going to make its approach into the shopper base,” Clarke mentioned. “We’ll do all the things we are able to to mitigate that.”
Even Nvidia, which has emerged as the largest buyer within the HBM market, is dealing with questions on its ravenous reminiscence wants — particularly, about its shopper merchandise.
At a press convention Tuesday at CES, Nvidia CEO Jensen Huang was requested if he was involved that the corporate’s gaming clients is perhaps resentful of AI expertise due to rising recreation console and graphics playing cards costs which might be being pushed by the reminiscence scarcity.
Huang mentioned Nvidia is a really giant buyer of reminiscence and has lengthy relationships with the businesses within the house however that, in the end, there would should be extra reminiscence factories as a result of the wants of AI are so excessive.
“As a result of our demand is so excessive, each manufacturing unit, each HBM provider, is gearing up, they usually’re all doing nice,” Huang mentioned.
At most, Micron can solely meet two-thirds of the medium-term reminiscence necessities for some clients, Sadana mentioned. However the firm is presently constructing two large factories referred to as fabs in Boise, Idaho, that can begin producing reminiscence in 2027 and 2028, he mentioned. Micron can also be going to interrupt floor on a fab within the city of Clay, New York, that he mentioned is anticipate to come back on-line in 2030.
However for now, “we’re bought out for 2026,” Sadana mentioned.