Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Heterogeneous Memory Opportunity with Agentic AI and Memory Centric Computing

Open Compute Project via YouTube

Overview

Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Explore the critical memory bottlenecks limiting large language model (LLM) inference services on GPU systems in this 16-minute conference talk from the Open Compute Project. Learn how limited HBM capacity creates severe constraints on model size, KV cache, and RAG vector databases, resulting in an imbalance with available GPU compute and memory bandwidth resources. Discover emerging second-tier memory solutions including SOCAMM, MRDIMM, and CXL memory as potential solutions to overcome the memory wall, with experimental validation results from DGX GPU systems demonstrating their effectiveness. Examine how CPU-GPU interconnects become the next performance bottleneck and understand Samsung's vision for Processing-Near-Memory (PNM) technology with Processing-in-Memory (PIM) as a fundamental solution to data transfer challenges in memory-centric computing architectures for agentic AI applications.

Syllabus

Heterogeneous Memory Opportunity with Agentic AI and Memory Centric Computing

Taught by

Open Compute Project

Reviews

Start your review of Heterogeneous Memory Opportunity with Agentic AI and Memory Centric Computing

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.