DESIGN TOOLS

Invalid input. Special characters are not supported.

High-bandwidth memory

Memory that matches the scale of human ambition

HBM4 gives artificial intelligence and scientific computing the memory to amplify what we can achieve. Picture reasoning models designing new medicines through millions of logical steps or researchers modeling planetary storms to understand Earth’s climate. HBM4 streams terabytes of data at high speeds, allowing us to engage with questions about technology and science that will define our future.

MEMORY MEETS AMBITION

Technical specs that match the scale of tomorrow’s biggest challenges

Micron's HBM4 features a wider 2048-pin bus interface operating at speeds greater than 11.0 Gbps, delivering greater than 2.8 TB/s of bandwidth per stack, more than double that of the previous generation. This expanded bandwidth addresses emerging AI workload requirements, from ultra-long context windows spanning millions of tokens to steady real-time responses from multimodal AI systems.

More than double the bandwidth*

HBM4 specs - double the bandwidth - 2.8TB/s

Twice the bus width*


HBM4 specs - Twice the bus width - 2048 I/O

Improved power efficiency*


HBM4 specs - 20% improved efficiency

* Comparing HBM4 12-high to HBM3E 12-high. Power efficiency is measured in picojoules per bit (pJ/bit) at similar speeds.

HBM4 applications: from reasoning to scientific discovery

HBM technology opens new pathways for AI and scientific computing by solving key problems: room for a lot of data and to access it quickly. As these fields flourish, they need both expanded memory capacity and very fast data access. Our latest HBM4 delivers both high capacity and >2.8 TB/s bandwidth, enabling humanity to pursue more ambitious problems.

Advanced reasoning

Reasoning models are like scientists working through complex problems step by step. These AI systems take time to evaluate problems, building elaborate chains of logic in memory and exploring many paths. This process is memory-intensive because models must remember large amounts of context as they reason. HBM4 ensures models can access and update this data quickly to keep the reasoning flowing.

Advanced reasoning image

Multimodal AI

Multimodal AI systems process different types of data all at once much like we do. That includes text, images, video, sound and sensor data. These systems don’t just look at one format at a time; they layer them together to understand context that a “single-mode” AI cannot. HBM supports the AI system in keeping all these formats in memory together so that it can find connections between them.

Robot picking tomatoes image

AI agents

Think of AI agents as AI systems given tools to take actions to complete a task.1 They must hold large datasets in memory while also accessing that data at high speeds to coordinate complex tasks. They excel at connecting the right data to the right processes. HBM4 provides the high bandwidth for these agents to work together in multi-agent systems, creating a network of shared intelligence.

Manufacturing image

Scientific & high-performance computing

HPC systems enable us to run simulations that would be impossible with traditional computing. For example, supercomputers hold enormous data sets in memory while they model complex systems like Earth’s atmosphere and planetary phenomena like Jupiter’s ancient storms. HBM4 capacity determines how much of the problem fits in memory, while its bandwidth determines how quickly the system can solve it.

Earth curvature image

Frequently asked questions

Here you will find answers to frequently asked questions about HBM4, including how it is manufactured, when it should be used alongside other types of memory, and other important details about our latest HBM technology.

HBM stands for High Bandwidth Memory. The “4” indicates that this is the fourth architectural generation of HBM. Unlike traditional memory, where chips sit next to one another, HBM stacks DRAM dies on top of each other and creates direct electrical pathways through the silicon using through-silicon vias (TSVs). This means data can flow directly through the silicon stack instead of traveling around the edges of chips.

The bandwidth of HBM4, over 2.8 TB/s, matters most for applications in AI and HPC that need to move terabytes of data at high speeds. Advanced reasoning models, for example, must evaluate hundreds of intermediate logical steps as they work through a problem. This requires terabytes of data to flow between main memory and processors each second as calculations progress.

HBM4 works alongside other memory types rather than replacing them. For example, in modern systems, CPUs use LPDDR5 or DDR5 to coordinate the system while GPUs use HBM4 for demanding math (i.e., complex algorithms).

HBM4 takes everything that worked well in HBM3 and HBM3E and makes it more powerful. A wider interface operates at speeds greater than 11.0 Gbps to deliver more than double the bandwidth of the previous gen. This matters because it addresses emerging requirements from AI workloads with long context windows spanning millions of tokens to scientific simulations running on next-gen supercomputers.

Traditional DRAM such as DDR memory handles general computing tasks while HBM supports AI and HPC applications that require continuous terabyte streams of data. The architecture of HBM stacks ultra-thin DRAM dies and connects them with thousands of through-silicon vias (TSVs). This vertical design requires higher precision in manufacturing, making HBM one of the more challenging memory products to produce.

HBM4 12-high provides 36GB of memory capacity per stack (the same as the previous generation) but with more than 2.8 TB/s. The increase in bandwidth (more than twice that of HBM3E) means the processor can access this capacity much faster, supporting more demanding AI workloads and scientific simulations than the previous gen at the same capacity could handle.

Capacity is how much data memory can hold, while bandwidth is how much of that data can flow each second. An HBM4 12-high stack can hold 36GB of data. And 2.8 TB/s means that in one second, the equivalent of 2.8 terabytes (TB) of data can flow between HBM and processor. Capacity determines what data you can fit in memory and bandwidth determines how quickly you can access that data.

Manufacturing HBM starts by fabricating three types of silicon wafers. One creates dies with through-silicon vias (TSVs) for electrical connections. Another produces thicker top dies without TSVs. The third fabricates logic dies with TSVs to interface with the system.

Only dies that pass testing continue further to assembly. Specialized equipment then stacks multiple DRAM dies on the logic die. The thicker top DRAM die completes the stack and provides memory as well as structural integrity. Once assembled, the complete HBM cube undergoes final testing to verify all connections work properly.

Yes, HBM4 works with both GPUs and custom ASICs (application-specific integrated circuits). The memory connects to any processor that can handle its high bandwidth interface and has the appropriate packaging.

High-end computing systems solve scientific problems (e.g., supercomputers) and train AI models on exabytes of data. To do this efficiently, memory must move data fast enough to keep thousands of processor cores busy. With a bandwidth of more than 2.8 TB/s, HBM4 speeds up AI training, lowers inference latency through faster KV cache access, and enables more detailed scientific simulations.

1 Anthropic. (2026, February 18). Measuring AI agent autonomy in practice. https://www.anthropic.com/research/measuring-agent-autonomy