Smart Recall - Enhance Local LLM Conversations with Embedding-Aware Context Retrieval
OpenSource Connections via YouTube
AI Adoption - Drive Business Value and Organizational Impact
35% Off Finance Skills That Get You Hired - Code CFI35
Overview
Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Learn how to enhance local large language model conversations through embedding-aware context retrieval in this 41-minute conference talk from Haystack EU 2025. Discover a practical service architecture for improving contextual continuity in chat applications by leveraging locally stored conversation history. Explore a Python-based approach that dynamically retrieves and rewrites prior conversation turns based on semantic similarity, utilizing embeddings, token limits, and summarization techniques to provide relevant memory windows to your model. Master the techniques for structuring past interactions, filtering for importance, and integrating efficient recall mechanisms to ensure your local LLMs maintain coherence, conciseness, and contextual awareness throughout extended conversations. Gain practical insights into solving the common problem of forgetful AI assistants by implementing smart memory systems that enhance the user experience in local LLM deployments.
Syllabus
Haystack EU 2025: Smart Recall: Enhance Local LLM Conversations w/ Embedding-Aware Context Retrieval
Taught by
OpenSource Connections