Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Deploy a Llama 3 AI Chatbot on Kubernetes - Full Walkthrough with Hugging Face and GPUs

Linode via YouTube

Overview

Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Learn to deploy a production-ready AI chatbot using Llama 3 model from Hugging Face on Kubernetes infrastructure with GPU acceleration in this 14-minute practical tutorial. Master the deployment of chatbots using the Llama 3 model from Hugging Face while leveraging Akamai App Platform with LKE (Linode Kubernetes Engine) for simplified deployment processes. Discover why inference at the edge matters and how to make AI applications scalable through hands-on implementation. Connect all essential components including the model, user interface, GPU resources, and APIs using open source tools like KServe and Knative for a complete deployment pipeline. Address common AI chatbot challenges such as hallucinations and implement proper guardrails for reliable performance. Gain practical experience with cloud native AI architecture, real-time deployment steps, and the tools needed to create secure, production-ready chatbots that can be deployed quickly without requiring advanced cloud expertise or specialized AI knowledge.

Syllabus

Deploy a Llama 3 AI Chatbot on Kubernetes (Full Walkthrough, Hugging Face + GPUs)

Taught by

Linode

Reviews

Start your review of Deploy a Llama 3 AI Chatbot on Kubernetes - Full Walkthrough with Hugging Face and GPUs

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.