Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Accelerating Distributed Graph Learning by Using Collaborative In-Network Multicast and Aggregation

USENIX via YouTube

Overview

Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Learn how to accelerate distributed graph neural network (GNN) training through innovative in-network processing techniques in this 16-minute conference presentation from USENIX ATC '25. Discover the challenges of distributed GNN training systems, including memory limitations, redundant traffic, and bandwidth bottlenecks that occur when partitioning large graphs across multiple workers. Explore how traditional approaches suffer from complex dependencies among graph data and limited switch-aggregator resources that lead to performance degradation. Understand the proposed SwitchGNN solution that addresses these issues through coordinated in-network multicast and aggregation, featuring a graph-aware multicast reordering algorithm that prioritizes vertices with higher neighbor counts to reduce communication time. Examine the multi-level graph partitioning mechanism that prevents aggregator overflow by partitioning boundary vertices into independent blocks for batch processing while maintaining graph propagation correctness. Review the implementation details using P4 programmable switches and DPDK host stack, along with experimental results from real testbed and NS3 simulations demonstrating up to 74% reduction in training time through effective communication overhead reduction.

Syllabus

USENIX ATC '25 - Accelerating Distributed Graph Learning by Using Collaborative In-Network...

Taught by

USENIX

Reviews

Start your review of Accelerating Distributed Graph Learning by Using Collaborative In-Network Multicast and Aggregation

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.