New Intel Compute Express Link Boosts Accelerator, CPU Bandwidth

New Intel Compute Express Link Boosts Accelerator, CPU Bandwidth

Intel and a consortium of leading technology companies (Alibaba, Cisco, Dell EMC, Facebook, Google, HPE, Huawei, and Microsoft) announced the release of the Common Express Link (CXL) 1.0 standard today, with the goal of creating a new product ecosystem and hardware standard. While it’s built squarely on PCIe (PCIe Gen 5, specifically), CXL offers features that the standard PCI Express bus lacks, such as maintaining memory coherency between the CPU and various attached accelerators.

New Intel Compute Express Link Boosts Accelerator, CPU Bandwidth

Intel has worked on CXL for at least four years before deciding to open the standard development process to a larger group of companies. Other firms will be able to join if they wish. Devices that support CXL should be capable of operating in ‘CXL Mode’ when inserted into a compatible PCIe slot (backward compatibility between PCIe mode and CXL mode should be seamless).

Initial deployments of the technology will focus on links between FPGAs and GPUs, the two most common types of ‘accelerator’ that a customer is likely to use. With that said, there are some significant names not on the CXL list, including AMD, ARM, Xilinx, and Nvidia (apparently pleased enough with its own NVLink work with IBM not to feel the need). Companies like Amazon and Baidu are nowhere to be seen, either. This could change if the industry standardizes on CXL, of course, but multiple firms that belong to CXL are also part of other initiatives. Dell, HP, and Huawei are also part of the Gen-Z consortium. Huawei is also a member of the CCIX consortium. Some firms are clearly supporting more than one technological effort to create next-generation standards.

Intel expects to complete the 1.0 standard and make it available to members in the first half of this year, with supporting products available in 2021. Expect to see a fair bit of technical discussion on these issues, particularly given how critical it is to minimize the cost and latency of moving data when working with accelerators. One of the barriers standing in the way between us and higher compute performance is the fundamental power cost of moving data in the first place. The standard with the best chance of adoption will be one that can minimize power costs without sacrificing performance to do it.

Continue reading

Nvidia Unveils Ampere A100 80GB GPU With 2TB/s of Memory Bandwidth
Nvidia Unveils Ampere A100 80GB GPU With 2TB/s of Memory Bandwidth

Nvidia announced an 80GB Ampere A100 GPU this week, for AI software developers who really need some room to stretch their legs.

Why Latency Impacts SSD Performance More Than Bandwidth Does
Why Latency Impacts SSD Performance More Than Bandwidth Does

Bandwidth matters to SSD performance, but latency matters more. It's also why SSDs are such good upgrades for such a wide range of hardware, including machines that are 15-20 years old.

JEDEC Extends HBM2 Standard to 24GB, 307GB/s Bandwidth Per Stack
JEDEC Extends HBM2 Standard to 24GB, 307GB/s Bandwidth Per Stack

JEDEC has updated the HBM2 standard for higher transfer speeds and increased density.

Google Stadia Will Eat 1TB Bandwidth Caps for Breakfast
Google Stadia Will Eat 1TB Bandwidth Caps for Breakfast

Google's Stadia service comes with bandwidth consumption requirements that could put a serious strain on 1TB data caps across the country. 4K isn't going to be accessible to a lot of folks — not without overage fees.