Samsung Now Producing 16Gbit GDDR6 for Upcoming GPUs
Samsung is throwing its hat into the GDDR6 ring and joining Micron in ramping the new memory technology for upcoming GPU products. It’s not a surprising move, but it does suggest that GDDR6 will be more widely adopted than its predecessor, GDDR5X.
Samsung is touting the new memory as being built on a 10nm* process at double the density of its previous RAM. 16Gbit chips means scaling up to 2GB of RAM per GDDR6 chip. This also clears the way for much higher amounts of RAM onboard GPUs over time, though I doubt we’ll see many 24GB GPUs in the near future. Even advanced 4K titles with HDR and other bells and whistles don’t push that kind of envelope (for now).
One potential advantage of the GDDR6 push is that we should finally see 2GB cards dropping off the map this generation. With Intel now fielding 4GB GPUs on its Radeon-integrated hardware, hopefully we’ll see a shift to larger RAM buffers across the board.
Samsung is claiming its GDDR6 can scale up to 72GB/s per channel (18Gbps per pin), which is more than twice as fast as the old GDDR5 standard and its 8Gbps performance. This ignores GDDR5X, of course, but since Samsung never built that type of RAM it can get away with skipping it as a point of comparison. Early GPUs are likely to opt for lower-clocked RAM, but a 72GB/s channel transfer rate is impressive, implying that a 256-bit GPU could hit 576GB/s of memory bandwidth. The next generation of midrange cards from AMD and Nvidia should be potent competitors for this reason alone.
The fact that Samsung is picking up GDDR6 also suggests robust demand from multiple companies. GDDR5X was an Nvidia-Micron play, but never saw wider adoption in the market. With multiple companies ramping GDDR6, it’s clear it’ll be more popular.
One big question is how GDDR6 will compare with HBM2 at high clocks and wide channels. HBM2’s higher cost and more difficult manufacturing process are balanced, to an extent, by its significantly lower power consumption and smaller, simpler board layouts. This allows for GPUs like AMD’s Radeon Nano, and improves overall efficiency. If GDDR6 can match or approach these outcomes, we may see HBM2 shrink back or vanish altogether. On the other hand, Samsung is touting its 2.4Gbps HBM2 stack as well, which would give GPUs based on it a substantial performance kick of their own.
Samsung claims its new standard offers a 35 percent improvement in power consumption compared with GDDR5, with 30 percent higher yields per wafer on GDDR6 compared with GDDR5 thanks to smaller process geometries. There’s no word on product introductions, but 2018 launches seem likely.
* – Samsung calls its 10nm manufacturing for GDDR6 “10nm-class” rather than “10nm.” The term denotes a process node between 10nm – 19nm, which is to say, it doesn’t define a process node at all. “10nm-class” or equivalent terms from all vendors should be considered fictional labels, not meaningful product classifications.
Nvidia’s RTX 4080 16GB Could be Even Harder to Buy than the RTX 4090
FPS-starved gamers in the Nvidia camp might find themselves forced to upgrade to Ampere this holiday shopping season.
Some Nvidia RTX 4080 16GB GPUs Cost Almost as Much as an RTX 4090
if you were hoping to get an RTX 4080 16GB near MSRP, your odds just got a bit worse.
Hynix Demos First 16Gb DDR5-5200 DIMMs
Hynix has demonstrated the first 16Gb DDR5-5200 DIMMs with commercialization planned for 2020.
Western Digital Plans First 16GB MAMR HDDs in 2019
Western Digital is planning a new 16TB HDD for 2019, this one based on a new type of recording. Say hello to MAMR.