Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

In-Network Collective Acceleration for AI Fabrics

Open Compute Project via YouTube

Overview

Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Learn about In Network Collective (INC) acceleration solutions for AI and machine learning workloads in this 24-minute conference talk from the Open Compute Project. Discover how network switches can offload critical collective operations like All Reduce, ReduceScatter, and AllGather to overcome communication bandwidth limitations in GPU-based systems. Explore how performing reduction operations directly in the switch fabric can reduce network bandwidth requirements by half compared to traditional GPU-based approaches, while enabling higher Model FLOPS Utilization (MFU) and lower memory footprint at GPU endpoints. Examine a practical INC offload solution implemented in high-performance, low-latency Ethernet switches, including detailed performance measurements and real-world application benefits. Get updates on ongoing standardization efforts within the Ultra Ethernet Consortium (UEC) and OCP SAI communities that are shaping the future of network-accelerated AI infrastructure.

Syllabus

In Network Collective acceleration for AI Fabrics

Taught by

Open Compute Project

Reviews

Start your review of In-Network Collective Acceleration for AI Fabrics

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.