Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Optimizing AI Workloads in Kubernetes - Pruning for Efficiency and Scale

Platform Engineering via YouTube

Overview

Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Learn how to optimize AI workloads in Kubernetes environments through model pruning techniques in this 14-minute conference talk. Explore the growing importance of resource efficiency and cost management as AI adoption accelerates in cloud-native environments. Discover model pruning as an optimization technique and understand its integration with Kubernetes-native tools for enhanced performance. Master strategies for resource scheduling and autoscaling configurations specifically designed for AI workloads. Examine best practices for deploying pruned AI models within Kubernetes clusters while maintaining optimal performance levels. Understand the benefits, trade-offs, and technical considerations of model pruning for AI inference in cloud environments. Gain practical insights into scaling AI applications more efficiently while significantly reducing resource usage and associated operational costs. Acquire valuable knowledge for platform teams looking to implement cost-effective AI optimization strategies in production Kubernetes environments.

Syllabus

Optimizing AI workloads in Kubernetes: Pruning for efficiency and scale

Taught by

Platform Engineering

Reviews

Start your review of Optimizing AI Workloads in Kubernetes - Pruning for Efficiency and Scale

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.