Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Co-Designing for Scale - CXL Based Memory Solution for Data Centric Workloads

Open Compute Project via YouTube

Overview

Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Explore hardware-software co-optimized solutions for memory-bound workloads in this 30-minute conference talk from the Open Compute Project. Learn how to leverage disaggregated and composable CXL resources to augment memory-constrained AI accelerators through flexible system architectures designed for data-intensive applications including billion-scale vector search, cloud-native and in-memory databases, data analytics, and large-scale AI inferencing. Discover innovative approaches featuring switch-based disaggregated memory, Near Memory Compute accelerators, and memory pooling appliances that improve efficiency, reduce total cost of ownership, and enhance system performance. Examine how advanced open-source frameworks such as NVIDIA Dynamo, along with KV-cache-focused solutions like Mooncake and LMCache, address software challenges introduced by heterogeneity while lowering adoption barriers for cost-effective, scalable, and energy-efficient AI infrastructure. The presentation features insights from industry experts including Gaurav Agarwal from Marvell, Anil Godbole from Intel, Jianping Jiang from Xconn Technologies Holdings, and Xinjun Yang from Alibaba, providing comprehensive perspectives on cutting-edge CXL-based composable memory systems.

Syllabus

Co Designing for Scale CXL Based Memory Solution for Data Centric Workloads

Taught by

Open Compute Project

Reviews

Start your review of Co-Designing for Scale - CXL Based Memory Solution for Data Centric Workloads

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.