How We Save $900 per Day with Self-Hosted AI - Building Scalable Local LLM Infrastructure
CNCF [Cloud Native Computing Foundation] via YouTube
Overview
Coursera Spring Sale
40% Off Coursera Plus Annual!
Grab it
Learn how to dramatically reduce AI infrastructure costs by building self-hosted LLM solutions in this keynote presentation from KubeCon + CloudNativeCon. Discover the practical implementation strategies used by a leading e-commerce website to deploy customized LLM applications on private cloud infrastructure, achieving 90% cost reduction compared to SaaS AI providers like OpenAI. Explore the technology stack for orchestrating inference workloads on cloud GPUs and understand how to overcome the limitations of external AI services including high costs, lack of customization, privacy concerns, and usage throttling. Gain insights into building stable, scalable, and high-performance AI applications while maintaining complete control over scalability and quality of service through private cloud infrastructure deployment.
Syllabus
Keynote: How We Save $900 per Day with Self-Hosted AI: Building Scalable Local... Vivian Hu & Lv Yi
Taught by
CNCF [Cloud Native Computing Foundation]