What are you looking for ?
Advertise with us
RAIDON

Micron Ships Samples of HBM4 to Key Customers to Power Next-Gen AI Platforms

HBM4 36GB 12-high products lead industry in power efficiency for data center and cloud AI acceleration

The importance of high-performance memory has never been greater, fueled by its crucial role in supporting the growing demands of AI training and inference workloads in data centers.

Built on its well-established 1ß (1-beta) DRAM process, proven 12-high advanced packaging technology and highly capable memory built-in self-test (MBIST) feature, Micron HBM4 provides integration for customers and partners developing next-gen AI platforms.

Micron Hbm4 2025 06

Micron Technology, Inc. announced the shipment of HBM4 36GB 12-high samples to multiple key customers.

This milestone extends Micron’s leadership in memory performance and power efficiency for AI applications. Built on its well-established 1ß (1-beta) DRAM process, proven 12-high advanced packaging technology and highly capable memory built-in self-test (MBIST) feature, the firm’s HBM4 provides integration for customers and partners developing next-generation AI platforms.

Leap forward
As use of GenAI continues to grow, the ability to effectively manage inference becomes more important. The company’s HBM4 features a 2,048-bit interface, achieving speeds greater than 2.0TB/s per memory stack and more than 60% better performance over the previous-gen. (1) This expanded interface facilitates rapid communication and a high-throughput design that accelerates the inference performance of large language models and chain-of-thought reasoning systems. Simply put, HBM4 will help AI accelerators respond faster and reason more effectively.

Additionally, Micron HBM4 features over 20% better power efficiency compared to the firm’s previous-gen HBM3E products, which first established new, unrivaled benchmarks in HBM power efficiency in the industry. (2) This improvement provides maximum throughput with the lowest power consumption to maximize data center efficiency. (2)

GenAI use cases continue to multiply, and this transformative technology is poised to deliver significant benefits to society. HBM4 is a crucial enabler, driving quicker insights and discoveries that will foster innovation in diverse fields such as healthcare, finance and transportation.

“Micron HBM4’s performance, higher bandwidth and industry-leading power efficiency are a testament to our memory technology and product leadership,” said Raj Narasimhan, SVP and GM, cloud memory business unit, Micron. “Building on the remarkable milestones achieved with our HBM3E deployment, we continue to drive innovation with HBM4 and our robust portfolio of AI memory and storage solutions. Our HBM4 production milestones are aligned with our customers’ next-generation AI platform readiness to ensure seamless integration and volume ramp.”

Intelligence Accelerated: Micron’s role in the AI revolution
For nearly 5 decades, Micron has pushed the boundaries of memory and storage innovation. Today, the company continues to accelerate AI by delivering a broad portfolio of solutions that turn data into intelligence, fueling breakthroughs from the data center to the edge. With HBM4, the firm reinforces its position as a critical catalyst for AI innovation and a reliable partner for our customers’ most demanding solutions.

Micron plans to ramp HBM4 in calendar year 2026, aligned to the ramp of customers’ next-gen AI platforms.

(1) Based on internal Micron HBM4 testing and published HBM3E specifications (2.0TB/s vs. 1.2TB/s).
(2) Based on internal Micron simulation projections in comparison to Micron HBM3E 36GB 12-high and similar competitive products.

Read also :
Articles_bottom
ExaGrid
AIC
Teledyne
ATTO
OPEN-E