The Llama Stack Tutorial - Llama Stack and RAG - Chat with your documents - Episode 3
-
27
-
- Write review
Our career paths help you become job ready faster
AI Product Expert Certification - Master Generative AI Skills
Overview
Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Learn to implement Retrieval Augmented Generation (RAG) using Meta's open-source Llama Stack project in this 13-minute tutorial video. Discover how to enable large language models to reference external knowledge beyond their training data by setting up a local Llama Stack server with Podman, creating and ingesting documents into a vector database, and building a RAG agent that selectively retrieves context from your data. Master the process of chatting with real documents like PDFs, invoices, or project files using Agentic RAG, and understand how RAG integrates unique data into AI workflows while leveraging Llama Stack's scalability from local development to production deployment on Kubernetes.
Syllabus
The Llama Stack Tutorial: Episode Three - Llama Stack & RAG: Chat with your documents
Taught by
Red Hat Developer