Is Sharing GPU to Multiple Containers Feasible?
CNCF [Cloud Native Computing Foundation] via YouTube
Google, IBM & Microsoft Certificates — All in One Plan
2,000+ Free Courses with Certificates: Coding, AI, SQL, and More
Overview
Google, IBM & Meta Certificates — All 10,000+ Courses at 40% Off
One annual plan covers every course and certificate on Coursera. 40% off for a limited time.
Get Full Access
Explore the feasibility of sharing GPUs across multiple containers in this 27-minute conference talk from the Cloud Native Computing Foundation (CNCF). Dive into cost-effective strategies for maximizing GPU utilization in ML workloads, particularly for inference tasks. Learn how to provision and attach GPUs using Kubernetes device plugins, and discover techniques for extending the NVIDIA device plugin to schedule multiple ML workloads on a single GPU. Gain insights into collecting GPU information with Prometheus and understand the native GPU sharing capabilities in Kubernetes without relying on additional technologies like VMware's vGPUs.
Syllabus
Is Sharing GPU to Multiple Containers Feasible? - Samed Güner, SAP
Taught by
CNCF [Cloud Native Computing Foundation]