Smart Recall - Enhance Local LLM Conversations with Embedding-Aware Context Retrieval
OpenSource Connections via YouTube
AI Engineer - Learn how to integrate AI into software applications
Cybersecurity: Ethical Hacking Fundamentals - Self Paced Online
Overview
Google, IBM & Meta Certificates — All 10,000+ Courses at 40% Off
One annual plan covers every course and certificate on Coursera. 40% off for a limited time.
Get Full Access
Learn how to enhance local large language model conversations through embedding-aware context retrieval in this 41-minute conference talk from Haystack EU 2025. Discover a practical service architecture for improving contextual continuity in chat applications by leveraging locally stored conversation history. Explore a Python-based approach that dynamically retrieves and rewrites prior conversation turns based on semantic similarity, utilizing embeddings, token limits, and summarization techniques to provide relevant memory windows to your model. Master the techniques for structuring past interactions, filtering for importance, and integrating efficient recall mechanisms to ensure your local LLMs maintain coherence, conciseness, and contextual awareness throughout extended conversations. Gain practical insights into solving the common problem of forgetful AI assistants by implementing smart memory systems that enhance the user experience in local LLM deployments.
Syllabus
Haystack EU 2025: Smart Recall: Enhance Local LLM Conversations w/ Embedding-Aware Context Retrieval
Taught by
OpenSource Connections