Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

The Optimization Trap in Edge AI

EDGE AI FOUNDATION via YouTube

Overview

Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Learn how to avoid performance pitfalls when deploying AI models on edge devices by understanding the critical relationship between model architecture and hardware optimization. Discover why theoretical efficiency metrics like FLOPs and parameters often fail to predict real-world performance, using the surprising example of how MobileNet V2 runs slower than ResNet18 on GPUs despite being "more efficient." Explore hardware selection strategies where NPUs can outperform GPUs despite lower TOPS ratings due to factors like operator support, kernel fusion, and memory behavior. Master a four-step framework for hardware-aware development: profiling on real devices from day one, verifying operator compatibility early, automating bottleneck discovery in CI pipelines, and optimizing with hardware-specific techniques like targeted pruning and mixed precision. Examine a practical case study of Llama 3.2-1B optimization on Snapdragon Gen 3, achieving 31% faster token generation, 25% faster prompt processing, and 126% faster initialization with minimal accuracy loss through strategic hardware-aware optimization techniques.

Syllabus

The Optimization Trap in Edge AI

Taught by

EDGE AI FOUNDATION

Reviews

Start your review of The Optimization Trap in Edge AI

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.