Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Demystifying Self-Hosted LLMs: A Beginner's Guide to Self-Hosting on Kubernetes with Podman Desktop

DevConf via YouTube

Overview

Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
This conference talk by Praveen Kumar, Ramakrishna Yekulla, and Shardul Inamdar addresses the challenges of deploying large language models (LLMs) in production environments and offers a beginner-friendly approach to self-hosting on Kubernetes. Learn how organizations can achieve enhanced data privacy, greater flexibility in model training, and potential cost savings through self-hosted LLM solutions. Discover the capabilities of the Podman Desktop AI Lab extension for streamlining LLM workload development, deployment, and management on Kubernetes. The 45-minute presentation covers essential topics including strategic selection and containerization of open-source LLM models, creation of Kubernetes deployment manifests for LLM workloads, resource provisioning to meet computational demands, and detailed exploration of the Podman Desktop AI Lab extension's integration with Kubernetes. Gain insights into how this technology enables organizations to maintain control over their data while building trust in AI technologies, particularly important for enterprises prioritizing data governance and compliance.

Syllabus

Demystifying Self Hosted LLMs: A Beginner's Guide to Self Hosting on Kubernetes with Podman Desktop

Taught by

DevConf

Reviews

Start your review of Demystifying Self-Hosted LLMs: A Beginner's Guide to Self-Hosting on Kubernetes with Podman Desktop

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.