Get all your news in one place.
100’s of premium titles.
One app.
Start reading
TechRadar
TechRadar
Efosa Udinmwen

Micron launches 36GB HBM3E memory as it plays catch up with Samsung and SK Hynix as archrivals frantically rush towards the next big thing — HBM4 with its 16 layers, 1.65TBps bandwidth, and 48GB SKUs

Micron HBM3E.

Micron has officially launched its 36GB HBM3E 12-high memory, signaling the company's entry into the competitive landscape of high-performance memory solutions for AI and data-driven systems.

As AI workloads grow more complex and data-heavy, the need for energy-efficient memory solutions becomes paramount. Micron's HBM3E aims to strike that balance, providing faster processing speeds without the high energy demands typically associated with such powerful systems.

Micron's 36GB HBM3E 12-high memory stands out for its increased capacity as it offers a 50% increase in capacity compared to current HBM3E offerings. This makes it a crucial component for AI accelerators and data centers managing large workloads.

36GB HBM3E 12-High memory for AI acceleration

Micron says its offering delivers more than 1.2 terabytes per second (TB/s) of memory bandwidth, with a pin speed greater than 9.2 gigabits per second (Gb/s), ensuring rapid data access for AI applications. While Micron's new memory addresses the increasing demand for larger AI models and more efficient data processing, it also lowers power consumption by 30% compared to its rivals.

Although Micron’s HBM3E 12-high memory brings noteworthy improvements in terms of capacity and power efficiency, it enters a field where both Samsung and SK Hynix have already established dominance. These two rivals are aggressively pursuing the next big thing in high-bandwidth memory—HBM4.

(Image credit: Micron)

HBM4 is expected to feature 16 layers of DRAM, offering more than 1.65TBps of bandwidth—far surpassing the capabilities of HBM3E. Additionally, with configurations reaching up to 48GB per stack, HBM4 will provide even greater memory capacity, enabling AI systems to handle increasingly complex workloads.

Despite the pressure from its competitors, Micron’s HBM3E 12-high memory remains a critical player in the AI ecosystem.

The company has already begun shipping production-capable units to key industry partners for qualification, allowing them to incorporate the memory into their AI accelerators and data center infrastructures. Micron’s robust support network and ecosystem partnerships ensure that its memory solutions are integrated seamlessly into existing systems, driving performance improvements in AI workloads.

A notable collaboration is Micron’s partnership with TSMC’s 3DFabric Alliance, which helps optimize AI system manufacturing. This alliance supports the development of Micron’s HBM3E memory and ensures that it can be integrated into advanced semiconductor designs, further enhancing the capabilities of AI accelerators and supercomputers.

More from TechRadar Pro

Sign up to read this article
Read news from 100’s of titles, curated specifically for you.
Already a member? Sign in here
Related Stories
Top stories on inkl right now
One subscription that gives you access to news from hundreds of sites
Already a member? Sign in here
Our Picks
Fourteen days free
Download the app
One app. One membership.
100+ trusted global sources.