Tutorial - Unlock the Future of Kubernetes and Accelerators With Dynamic Resource Allocation (DRA)
CNCF [Cloud Native Computing Foundation] via YouTube
Save 40% on 3 months of Coursera Plus
AI Engineer - Learn how to integrate AI into software applications
Overview
Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Learn how to leverage Dynamic Resource Allocation (DRA) in Kubernetes to optimize GPU and specialized hardware utilization for AI workloads in this comprehensive tutorial. Discover how DRA addresses the limitations of traditional device plugin APIs by enabling precise control over resource sharing between Pods, supporting multiple GPU models per node, and handling dynamic allocation of multi-instance GPU (MIG) configurations. Explore the architecture and inner workings of DRA within Kubernetes clusters while gaining hands-on experience through practical demonstrations of requesting GPU resources and network-attached devices like edge IP cameras. Master the techniques for maximizing hardware utilization across diverse workloads and understand how DRA extends beyond GPUs to support any specialized hardware that Pods may require, positioning yourself at the forefront of cloud-native AI infrastructure management.
Syllabus
Tutorial: Unlock the Future of Kubernetes and Accelerators With Dynamic Resource Alloc... Rey Lejano
Taught by
CNCF [Cloud Native Computing Foundation]