Accelerate AI Workloads with Rambus HBM4E Memory Controller

By Chad Cox

Production Editor

Embedded Computing Design

March 05, 2026

News

Accelerate AI Workloads with Rambus HBM4E Memory Controller
Image Credit: Rambus

Rambus Inc. announced its HBM4E Memory Controller IP capable of facilitating innovative HBM memory deployments for next-generation AI accelerators, graphics, and HPC applications. It operates at up to 16 Gigabits per second (Gbps) per pin delivering a throughput of 4.1 Terabytes per second (TB/s) to each memory device. For an AI accelerator with eight attached HBM4E devices, this translates to over 32 TB/s of memory bandwidth for demanding AI workloads.

The solution is compatible with third-party standard or TSV PHY solutions to integrate a complete HBM4E memory subsystem into a 2.5D or 3D package for AI SoC or custom die solutions.

“Given the insatiable bandwidth demands of AI, it’s imperative for the memory ecosystem to continue aggressively advancing memory performance,” said Simon Blake-Wilson, SVP and general manager of Silicon IP, at Rambus. “As a leading silicon IP provider for AI applications, we are bringing the industry’s leading HBM4E Controller IP solution to the market as a key enabler for breakthrough performance in next-generation AI processors and accelerators.”

Per the press release, the HBM4E Controller is available for licensing, and early access design customers can engage today.

For more information, visit rambus.com/interface-ip/hbm/.

Chad Cox. Production Editor, Embedded Computing Design, has responsibilities that include handling the news cycle, newsletters, social media, and advertising. Chad graduated from the University of Cincinnati with a B.A. in Cultural and Analytical Literature.

More from Chad