Overview
Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Learn how to unlock the hidden potential of local NVMe storage within GPU and CPU-based compute clusters to dramatically improve AI workload performance and reduce costs. Discover how the increasing complexity and extended context lengths of AI inferencing workloads create costly I/O bottlenecks that require expensive storage and network upgrades, while a high-performance solution already exists within your existing infrastructure. Explore how local NVMe storage, typically underutilized in compute platforms, offers superior cost-performance characteristics compared to external storage solutions. Master the implementation of a standards-based approach using Linux pNFS v4.2 with Flex Files through Hammerspace technology to activate local NVMe as shared, protected Tier 0 storage within a global namespace. Understand how this solution transforms idle storage capacity into a seamless tier that integrates multiple storage types from any vendor while providing automated data orchestration and data protection without proprietary software requirements. Examine how this approach enables longer context windows in AI inferencing to be sustained directly on local storage, significantly reducing costly and latency-inducing data transfers to external storage while utilizing existing infrastructure without additional power or network upgrades. Analyze the substantial cost and complexity reductions achievable in AI project implementations, and discover how activating local NVMe within cloud compute nodes provides unmatched performance that dramatically outperforms traditional cloud storage bandwidth constraints and accelerates time-to-value for AI workloads.
Syllabus
SNIA SDC 2025 - Activating Tier 0 Storage Within GPU - and CPU-based Compute Clusters
Taught by
SNIAVideo