High Bandwidth Memory (HBM) is the commonly used type of DRAM for data center GPUs like NVIDIA's H200 and AMD's MI325X. High Bandwidth Flash (HBF) is a stack of flash chips with an HBM interface. What ...
TL;DR: SK hynix will showcase its AI memory technologies at CES 2025, featuring solutions for on-device AI and next-generation AI memories. The company aims to highlight its technological ...
As SK hynix leads and Samsung lags, Micron positions itself as a strong contender in the high-bandwidth memory market for generative AI. Micron Technology (Nasdaq:MU) has started shipping samples of ...
TL;DR: SK hynix CEO Kwak Noh-Jung unveiled the "Full Stack AI Memory Creator" vision at the SK AI Summit 2025, emphasizing collaboration to overcome AI memory challenges. SK hynix aims to lead AI ...
Forbes contributors publish independent expert analyses and insights. This article discusses memory and chip and system design talks at the 2025 AI Infra Summit in Santa Clara, CA by Kove, Pliops and ...
The next generation of high-bandwidth memory, HBM4, was widely expected to require hybrid bonding to unlock a 16-high memory stack. A JEDEC move made that unnecessary with this generation, but it’s ...
The US government has imposed fresh export controls on the sale of high tech memory chips used in artificial intelligence (AI) applications to China. The rules apply to US-made high bandwidth memory ...
JEDEC’s HBM4 and the emerging SPHBM4 standard boost bandwidth and expand packaging options, helping AI and HPC systems push past the memory and I/O walls.
With doubled I/O interfaces and refined low voltage TSV design, HBM4 reshapes how memory stacks sustain throughput under data ...