MangoBoost Full Stack AI Infrastructure Solutions - MLPerf Inference, Training, and Storage Case Study
Open Compute Project via YouTube
Free courses from frontend to fullstack and AI
AI Adoption - Drive Business Value and Organizational Impact
Overview
Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Explore MangoBoost's comprehensive AI infrastructure solutions in this 15-minute conference talk that addresses the critical challenges of deploying large-scale AI systems with multiple GPUs, high-speed NICs, and storage devices. Learn how MangoBoost's holistic approach tackles the key challenge of efficiently orchestrating data movement across GPUs, storage, and networks while maximizing the massive compute capabilities of modern GPUs. Discover the LLMBoost software's full-stack solution covering inference, training, RAG (Retrieval-Augmented Generation), and management capabilities. Examine the innovative DPU hardware that fully accelerates GPU-to-GPU communication through GPUBoost and GPU-to-storage access via StorageBoost, operating at line rate over standard ethernet protocols including RoCEv2, NVME-of, and UEC. Review MangoBoost's top-ranked performance results across MLPerf Inference, Training, and Storage benchmarks using their solutions on modern GPUs such as AMD MI300X. Understand the capabilities of MangoBoost DPU cards that deliver 400 Gbps network performance and explore the DPU-enabled JBOF (Just a Bunch of Flash) and JBOD (Just a Bunch of Disks) headless storage solutions specifically designed for AI systems.
Syllabus
MangoBoost Full Stack AI Infrastructure Solutions MLPerf Inference, Training, Storage Case St
Taught by
Open Compute Project