SPHBM4 cuts pin counts dramatically while preserving hyperscale-class bandwidth performanceOrganic substrates reduce ...
SK Hynix and Taiwan’s TSMC have established an ‘AI Semiconductor Alliance’. SK Hynix has emerged as a strong player in the high-bandwidth memory (HBM) market due to the generative artificial ...
High Bandwidth Memory (HBM) is the commonly used type of DRAM for data center GPUs like NVIDIA's H200 and AMD's MI325X. High Bandwidth Flash (HBF) is a stack of flash chips with an HBM interface. What ...
Explosive growth of generative artificial intelligence (AI) applications in recent quarters has spurred demand for AI servers and skyrocketing demand for AI processors. Most of these processors — ...
With the goal of increasing system performance per watt, the semiconductor industry is always seeking innovative solutions that go beyond the usual approaches of increasing memory capacity and data ...
As SK hynix leads and Samsung lags, Micron positions itself as a strong contender in the high-bandwidth memory market for generative AI. Micron Technology (Nasdaq:MU) has started shipping samples of ...
TL;DR: SK hynix unveiled next-gen HBM4 memory with up to 16-Hi stacks, 48GB capacity, and 2TB/sec bandwidth, targeting NVIDIA's Vera Rubin AI GPUs in 2026. Mass production begins in late 2025, ...
Micron has now entered the HBM3 race by introducing a “second-generation” HBM3 DRAM memory stack, fabricated with the company’s 1β semiconductor memory process technology, which the company announced ...
High-bandwidth memory (HBM) is an important part of AI processors, handling massive data processing and computations. In response, Teradyne, Inc. has launched the Magnum 7H, a next-generation memory ...