Scaling Generative AI - Building Production-Ready LLM Applications
CNCF [Cloud Native Computing Foundation] via YouTube
Overview
Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Explore the critical aspects of developing production-ready Large Language Model (LLM) applications using Java in this 32-minute conference talk from CNCF's KubeCon + CloudNativeCon. Learn how to leverage Java's strengths to build scalable and efficient LLM systems while addressing key challenges such as performance optimization, resource management, and seamless integration with existing infrastructures. Gain practical knowledge on handling massive datasets, optimizing model inference, and fine-tuning LLMs for optimal performance. Discover strategies for ensuring the reliability and scalability of your LLM deployments, empowering you to create robust and high-performing AI applications. Whether you're a seasoned Java developer or new to the AI domain, acquire valuable insights and guidance for your LLM development journey, equipping yourself with the tools and knowledge to navigate the complexities of building production-grade LLM systems.
Syllabus
Scaling Generative AI: Building Production-Ready LLM Applications - Daniel Oh, Red Hat
Taught by
CNCF [Cloud Native Computing Foundation]