Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Going Beyond RAG - Extended Mind Transformers

AI Engineer via YouTube

Overview

Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Explore a revolutionary approach to transformer architecture that challenges traditional Retrieval Augmented Generation methods in this 16-minute conference talk. Learn about Extended Mind Transformers (EMT), an innovative transformer variant that leverages the model's inherent key/query mechanism to dynamically select and attend to relevant information during each generation step, rather than relying on embedding-based document retrieval. Discover how this architecture achieves state-of-the-art performance in long context applications while addressing fundamental limitations of RAG systems. Examine the key design decisions behind EMT implementation, including extended mind attention mechanisms, evaluation methodologies, and strategies for reducing hallucinations. Gain insights into the mathematical foundations and practical applications of this approach, with access to implementation resources through GitHub repositories and Hugging Face model collections for hands-on experimentation.

Syllabus

Introduction
Long Context vs RAG
Extended Mind Attention
Evaluations
Results
Citations
Reduce hallucinations
Parameters

Taught by

AI Engineer

Reviews

Start your review of Going Beyond RAG - Extended Mind Transformers

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.