Samsung has made a deal to supply NVIDIA with its cutting-edge HBM3E 12-Hi memory. The news confirms Samsung will deliver 30,000 to 50,000 units of this high-speed memory to NVIDIA in the near future. This partnership could reshape the AI chip market, and here’s why it matters.
The HBM3E 12-Hi memory is designed for NVIDIA’s liquid-cooled AI servers, likely powering the upcoming Blackwell Ultra GPUs. It’s a beast, offering up to 96GB per stack and over 1.2TB/s of bandwidth, perfect for heavy AI tasks like training large language models.
Samsung’s memory uses a 10nm-class process for better efficiency and performance, with improved thermal control and signal quality. While Samsung didn’t confirm its exclusive use in AI servers, the deal marks a major step forward after earlier struggles to meet NVIDIA’s strict standards.
Samsung’s Big Comeback in AI Memory
Samsung faced setbacks with its earlier HBM3 memory due to thermal issues, losing ground to rivals like SK hynix, NVIDIA’s main supplier. But this new agreement shows Samsung’s back in the game.
The company already supplies HBM3E to AMD for its Instinct MI350 AI accelerators, and now NVIDIA’s approval puts Samsung alongside SK hynix and Micron in the race for AI memory dominance. Reports suggest Samsung may have cut prices to secure this deal, giving it a competitive edge in the tight HBM market.
The partnership could shift market dynamics, easing the tight supply of HBM and potentially lowering costs for AI server builders. Samsung’s massive production capacity and its work on next-gen HBM4, set for 2026, make it a key player to watch. This deal is a lifeline for Samsung’s chip division, which has struggled with losses in the DRAM market.
What’s Next for Samsung and NVIDIA
Samsung’s HBM3E supply is expected to start soon, with volume production possibly kicking off in Q1 2026. This aligns with NVIDIA’s rollout of Blackwell Ultra GPUs, which will rely on high-performance memory for AI workloads.
The tech world is buzzing, with posts on X calling this a “pivotal recovery” for Samsung. However, SK hynix still leads with its advanced 12-layer HBM3E, and the race for HBM4 is heating up.
This partnership is great news for NVIDIA’s customers, who get more supply options, and for Samsung, which is clawing back market share. Fans are excited about what this means for faster, more efficient AI tech.