UC San Diego Product Management Certificate — AI-Powered PM Training
Stuck in Tutorial Hell? Learn Backend Dev the Right Way
Overview
Google, IBM & Meta Certificates — All 10,000+ Courses at 40% Off
One annual plan covers every course and certificate on Coursera. 40% off for a limited time.
Get Full Access
Learn about In Network Collective (INC) acceleration solutions for AI and machine learning workloads in this 24-minute conference talk from the Open Compute Project. Discover how network switches can offload critical collective operations like All Reduce, ReduceScatter, and AllGather to overcome communication bandwidth limitations in GPU-based systems. Explore how performing reduction operations directly in the switch fabric can reduce network bandwidth requirements by half compared to traditional GPU-based approaches, while enabling higher Model FLOPS Utilization (MFU) and lower memory footprint at GPU endpoints. Examine a practical INC offload solution implemented in high-performance, low-latency Ethernet switches, including detailed performance measurements and real-world application benefits. Get updates on ongoing standardization efforts within the Ultra Ethernet Consortium (UEC) and OCP SAI communities that are shaping the future of network-accelerated AI infrastructure.
Syllabus
In Network Collective acceleration for AI Fabrics
Taught by
Open Compute Project