August 28th, 2024

Copper's reach is shrinking so Broadcom is strapping optics directly to GPUs

Broadcom is enhancing AI cluster performance by integrating optical interconnects with GPUs, using co-packaged optics to increase bandwidth to 1.6 TB/sec and reduce power consumption, amid competition from Intel and AMD.

Read original articleLink Icon
Copper's reach is shrinking so Broadcom is strapping optics directly to GPUs

Broadcom is advancing its optical interconnect technology to enhance the performance of AI clusters by integrating optics directly with GPUs. Traditional copper connections are becoming inadequate due to their limited range and bandwidth, especially as data rates exceed 200 Gbit/sec. Broadcom's VP, Manish Mehta, highlighted that copper can only effectively transmit signals for about three to five meters before degradation occurs. To address this, Broadcom is exploring co-packaged optics (CPO), which combines optical engines with GPUs, significantly increasing interconnect bandwidth to approximately 1.6 TB/sec. This innovation could enable large-scale systems with hundreds of GPUs to function as a single unit, contrasting with current scale-out systems that rely on slower Ethernet or InfiniBand networks. Broadcom's experiments have demonstrated promising results, achieving error-free performance with a test chip designed to emulate a GPU. This approach not only enhances bandwidth but also reduces power consumption per port. While other companies like Intel and AMD are also investigating similar technologies, Broadcom's focus on CPO could position it as a leader in the next generation of AI computing infrastructure.

- Broadcom is integrating optical interconnects directly with GPUs to improve AI cluster performance.

- Traditional copper connections are limited in range and bandwidth, prompting the shift to optical solutions.

- Co-packaged optics (CPO) could allow hundreds of GPUs to operate as a single system.

- Broadcom's tests show significant increases in bandwidth and reductions in power consumption.

- Other tech companies are exploring similar optical technologies, indicating a competitive landscape.

Related

Intel CPU with Optical OCI Chiplet Demoed with 4Tbps of Bandwidth and 100M Reach

Intel CPU with Optical OCI Chiplet Demoed with 4Tbps of Bandwidth and 100M Reach

Intel unveiled a new CPU with an Optical Compute Interconnect chiplet, offering 4Tbps bandwidth and 100m reach. The chiplet merges Photonic and Electronic Integrated Circuits, featuring built-in lasers for improved server communication, especially in AI servers. Intel plans to scale the technology up to 32Tbps, focusing on power savings and enhanced bandwidth. This advancement aligns with the industry's move towards co-packaged optics for better efficiency and performance, marking a significant shift in server architecture towards power efficiency and high-speed data transfer.

Nvidia NVLink Switch Chips Change to the HGX B200

Nvidia NVLink Switch Chips Change to the HGX B200

NVIDIA introduced the HGX B200 board at Computex 2024, featuring two NVLink Switch chips instead of four, aiming to enhance performance and efficiency in high-performance computing applications by optimizing GPU configurations.

GPUs can now use PCIe-attached memory or SSDs to boost VRAM capacity

GPUs can now use PCIe-attached memory or SSDs to boost VRAM capacity

Major companies like AMD, Intel, and Nvidia are considering supporting Panmnesia's CXL IP technology for GPU memory expansion using PCIe-attached memory or SSDs. The low-latency CXL IP addresses increasing memory needs for AI training datasets, offering improved GPU performance for AI and HPC applications. Adoption by GPU manufacturers is uncertain, potentially shaping future GPU memory expansion technologies.

Unleashing 100 Mpps with Fd.io VPP on GCP x86

Unleashing 100 Mpps with Fd.io VPP on GCP x86

The article explores high-performance networking on Google Cloud Platform with DPDK, gVNIC, and FD.io VPP. It discusses the evolution of network technologies, NFV, DPDK's impact, and system requirements for efficient packet processing.

Nvidia NVLink and Nvidia NVSwitch Supercharge Large Language Model Inference

Nvidia NVLink and Nvidia NVSwitch Supercharge Large Language Model Inference

NVIDIA's NVLink and NVSwitch technologies enhance multi-GPU performance for large language model inference, enabling efficient communication and real-time processing, while future innovations aim to improve bandwidth and scalability.

Link Icon 0 comments