The memory shortage is structural, and the question sitting on top of every infrastructure budget is the same: when does it end? The people who should know can’t agree. Intel’s CEO says no relief until 2028. Counterpoint Research reckons Q4 2027. IDC says it’s not a cyclical blip at all — it’s a permanent structural reset.
Why can’t they agree? Because they’re making different bets on two things: how fast AI demand will keep growing and how quickly HBM yields will improve. This article maps every announced fab timeline against demand, explains why prices will come down a lot more slowly than they went up, and gives you a planning horizon you can actually work with. For the full picture of how AI created this mess, see the full AI memory shortage story.
What Do the Analysts Actually Disagree About — and Why Does It Matter?
Intel CEO Lip-Bu Tan told the Cisco AI Summit in February 2026 there is “no relief until 2028” — the most pessimistic mainstream forecast from a named executive. Counterpoint Research analyst Tarun Pathak puts Q4 2027 as the earliest point where supply and demand curves could cross.
Micron CEO Sanjay Mehrotra projects the HBM market hitting $100 billion by 2028. Micron is sold out for 2026, with demand having “far outpaced our ability to supply that memory.” TrendForce projects the memory market surging to $842.7 billion in 2027. And IDC’s Nabilia Popal doesn’t mince words: “This is not just a temporary situation. This is going to result in a structural reset of the entire industry.”
So here’s where the disagreement lands. It’s all about two variables: how fast AI demand will grow and how quickly HBM yields will improve. Get both wrong in the same direction and you’re off by years.
The spread between Q4 2027 and 2028+ is the window you need to plan around. For the current price data, see our analysis of why DRAM prices have doubled.
What Is the Fab Construction Timeline — and Why Can’t Money Simply Speed It Up?
A new fab takes at least 18 months to build — and that’s just the physical construction. Getting yields up to meaningful volume adds more months on top. Each facility costs $15 billion or more. The physics don’t compress, no matter how much capital you throw at them.
The timing problem goes back to the 2022-2023 memory bust. Prices tanked, Samsung cut production by 50%, and once things recovered the industry was gun-shy about expanding. Then AI demand exploded and caught everyone short.
HBM production makes it worse. Each HBM chip stacks up to 12 thinned-down DRAM dies. When Micron makes one bit of HBM, it has to forgo making three bits of conventional memory — the “three-to-one basis.” And even when DRAM wafers are available, packaging capacity at TSMC (CoWoS) and SK Hynix (MR-MUF) puts a hard cap on how many HBM units actually get assembled.
Shawn DuBravac at the Global Electronics Association reckons yield improvement is the faster path: better stacking efficiency and tighter coordination between memory suppliers and AI chip designers will deliver gains before new fabs do. For more on the packaging bottleneck, see how AI broke the memory supply chain.
When Will New Capacity Actually Arrive — and Will It Be Enough?
The 2027 wave brings three facilities: Micron Singapore (HBM fab), Micron Taiwan (retooled from a PSMC acquisition, H2 2027), and SK Hynix’s $13 billion HBM packaging facility at Cheongju — the world’s largest HBM assembly plant.
The 2028 wave adds SK Hynix West Lafayette, Indiana (CHIPS Act-funded) and Samsung Pyeongtaek with a new production line.
Then there’s the Micron Clay, New York megafab — the single largest planned capacity addition. It broke ground in January 2026, but first production will not arrive until 2030. The facility that would have made the biggest near-term difference is now a 2030 story.
Here’s the catch: all three 2027 facilities are purpose-built for HBM. They don’t directly relieve conventional DRAM supply. And new capacity has to outpace growing demand, not just match it. NVIDIA’s B300 uses eight HBM chips, each stacking 12 DRAM dies. HBM4 goes to 16. IDC expects just 16% year-over-year DRAM supply growth in 2026, well below AI-driven demand growth. These new fabs may merely keep pace rather than creating any surplus.
Why Do Memory Prices Fall More Slowly Than They Rise?
Most shortage coverage skips this bit, and it’s the part that matters most for your budget.
Kim at Mkecon Insights puts it plainly: “In general, economists find that prices come down much more slowly and reluctantly than they go up. DRAM today is unlikely to be an exception.”
It’s called price asymmetry. Three things drive it.
First, contract reset cycles. OEMs like Dell and HP purchase memory in bulk about a year in advance. When spot conditions ease, your contracted pricing stays elevated until the next reset window. You’re locked in.
Second, vendor inventory protection. Pricing floors and volume commitments slow the pass-through of falling costs. Procurement leverage hinges on strategic alignment, not volume. Hyperscalers lock in supply. Everyone else fights over what’s left.
Third, vendor pricing discipline. Manufacturers who sank $15 billion into each fab need to recoup that capital. Analysts assess that these price increases appear more durable than temporary. They’re not going to race to the bottom.
For your budgeting: even when supply relief arrives, prices won’t snap back. Model a gradual 12-18 month decline from the inflection point, not a step function.
What Planning Horizon Should Enterprises Use for Hardware Refresh and Cloud Commitment Decisions?
Here’s the concrete planning guidance.
Upside scenario (Q4 2027): The earliest possible inflection, if HBM yield ramps exceed expectations and AI demand moderates. This is the Counterpoint Research view. Don’t build your base budget on this.
Base case (gradual improvement through 2028): New fabs come online, yield improves, but price asymmetry means prices decline slowly. Meaningful relief arrives mid-2028. Use this for your 2026-2028 infrastructure roadmap.
Downside scenario (2029-2030): If HBM4 demand accelerates before capacity catches up, or if the Micron Clay NY delay cascades, full normalisation doesn’t arrive until 2030.
For hardware refresh: if servers need replacing in 2026, waiting for price relief is not viable. The earliest meaningful supply improvement is 18+ months away. Budget at current elevated pricing and move on.
For cloud commitments: multi-year contracts signed in 2026 lock in shortage pricing. Consider shorter terms with renegotiation windows aligned to the Q4 2027-2028 relief window. Give yourself room to move.
Even after normalisation, memory pricing may settle at a higher equilibrium than pre-2025 levels. The permanent wafer reallocation toward HBM means conventional DRAM carries a structural premium until manufacturers reverse course. For more on timing these decisions, see managing infrastructure costs during the shortage.
Is There a Scenario Where the Shortage Ends Faster — or Slower — Than the Base Forecast?
Yes, on both sides.
The fastest path to relief is a demand contraction. If enterprise AI deployments don’t prove ROI — a risk Deloitte‘s 2026 semiconductor outlook flags — spending could contract before new capacity arrives. Deloitte notes the industry “has placed all its eggs in the AI basket”. If those eggs don’t hatch, the surplus comes early.
On the supply side, HBM yield improvements could accelerate relief faster than new construction. SK Hynix’s MR-MUF process and hybrid bonding could enable denser stacks without waiting for new fabs.
The downside scenarios deserve equal weight. HBM4 goes to 16 stacked dies versus 12 today, consuming more wafer area before new capacity arrives. McKinsey predicts $7 trillion in data centre spending by 2030. If that accelerates, demand could outrun even the most aggressive fab plans.
Unlike the COVID chip shortage, which was cyclical, this one is structural. Resolution requires active vendor decisions to reverse wafer allocation, not just the passage of time.
Your planning assumption: gradual improvement through 2028, with demand-side wildcards as the main variables to watch. For the full story, see what the memory supply crisis means for tech companies and what comes next.
FAQ
When will the memory shortage actually end?
The credible range is Q4 2027 (Counterpoint Research) to 2028+ (Intel CEO’s forecast). Full price normalisation may not come until 2030, given the Micron Clay NY delay and price asymmetry. Plan for gradual improvement through 2028 as your base case.
Why is the 2025-2026 memory shortage different from the COVID chip shortage?
The COVID shortage was cyclical — a temporary demand spike that normalised, and supply caught up. This one is structural — IDC frames it as a permanent reallocation of wafer capacity toward HBM. It doesn’t self-correct when demand eases. Manufacturers have to deliberately reverse it, and that requires them to be confident conventional DRAM demand will recover.
Is it a bad time to buy servers because of the AI memory crunch?
If servers need replacing in 2026, waiting isn’t an option — the earliest meaningful supply improvement is 18+ months out. Budget at current elevated pricing. For discretionary upgrades, consider deferring to late 2027 or early 2028.
How long is the AI chip shortage going to last?
The AI-driven memory shortage is forecast to persist through at least 2027, with most analysts projecting gradual improvement in 2028. It specifically affects DRAM and HBM supply; broader chip shortages vary by segment.
Will RAM prices ever go back to normal?
Price asymmetry means memory prices fall more slowly than they rise. Contract reset cycles and vendor pricing strategy slow the decline. Pre-2025 pricing levels may not return because of the structural premium from permanent wafer reallocation toward HBM.
Why can’t memory manufacturers just build more fabs faster?
A new fab takes 18+ months to build and costs $15 billion or more. Getting yields up to meaningful volume adds further months. The physics of cleanroom construction, equipment installation, and yield optimisation simply can’t be compressed with more money.
What is HBM and why does it cause shortages of regular memory?
HBM (high-bandwidth memory) stacks multiple DRAM dies vertically to provide the bandwidth AI GPUs need. Each HBM unit consumes three times the wafer area of conventional DRAM. When manufacturers allocate cleanroom capacity to HBM, they directly reduce conventional DRAM output. It’s a zero-sum game.
Which memory fab will make the biggest difference to supply?
The Micron Clay, New York megafab was the largest planned capacity addition, but it’s been delayed to 2030. SK Hynix Cheongju (2027) and Micron Singapore (2027) are the nearest major additions, though both are HBM-focused.
What is the CHIPS Act doing about the memory shortage?
The CHIPS Act funds US-based fab construction, including Micron’s Clay NY facility and SK Hynix’s West Lafayette plant. But these are targeting 2028-2030 production starts, so CHIPS Act investment doesn’t speed up near-term relief.
Should I sign a multi-year cloud contract during the memory shortage?
Consider shorter commitment terms with renegotiation windows aligned to Q4 2027-2028. Multi-year contracts signed in 2026 lock in shortage pricing. Shorter terms give you the flexibility to renegotiate when supply improves.
What does “structural reset” mean for memory pricing?
IDC uses “structural reset” to describe the permanent reallocation of wafer capacity from conventional DRAM to HBM. Unlike a cyclical shortage, this shift doesn’t self-correct when demand eases — manufacturers have to deliberately reverse it, and that requires confident forecasts of conventional DRAM demand recovery.
Could the memory shortage end sooner if AI demand drops?
Yes. If enterprise AI deployments fail to prove ROI (a risk Deloitte flags), AI infrastructure spending could contract before new capacity arrives. That would reduce HBM demand, free up wafer capacity for conventional DRAM, and potentially create oversupply after 2027.