The Llama Stack Tutorial - Llama Stack and RAG - Chat with your documents - Episode 3
-
27
-
- Write review
Master AI and Machine Learning: From Neural Networks to Applications
Get 20% off all career paths from fullstack to AI
Overview
Google, IBM & Meta Certificates — All 10,000+ Courses at 40% Off
One annual plan covers every course and certificate on Coursera. 40% off for a limited time.
Get Full Access
Learn to implement Retrieval Augmented Generation (RAG) using Meta's open-source Llama Stack project in this 13-minute tutorial video. Discover how to enable large language models to reference external knowledge beyond their training data by setting up a local Llama Stack server with Podman, creating and ingesting documents into a vector database, and building a RAG agent that selectively retrieves context from your data. Master the process of chatting with real documents like PDFs, invoices, or project files using Agentic RAG, and understand how RAG integrates unique data into AI workflows while leveraging Llama Stack's scalability from local development to production deployment on Kubernetes.
Syllabus
The Llama Stack Tutorial: Episode Three - Llama Stack & RAG: Chat with your documents
Taught by
Red Hat Developer