Overview
Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Explore a comprehensive paper analysis examining Titans, a novel neural architecture that combines short-term attention mechanisms with long-term memory modules to overcome the quadratic scaling limitations of traditional Transformers. Discover how this innovative approach addresses the fundamental trade-off between recurrent models that compress data into fixed-size hidden states and attention mechanisms that capture direct dependencies across entire context windows but suffer from computational constraints. Learn about the dual-memory framework where attention serves as short-term memory for accurate dependency modeling within limited contexts, while neural memory acts as persistent long-term storage for historical information. Examine three architectural variants of Titans and understand how they enable parallelizable training while maintaining fast inference speeds. Analyze experimental results demonstrating superior performance compared to Transformers and modern linear recurrent models across diverse domains including language modeling, common-sense reasoning, genomics, and time series analysis. Investigate the architecture's remarkable ability to scale effectively to context windows exceeding 2 million tokens while achieving higher accuracy in needle-in-haystack retrieval tasks, representing a significant advancement in handling extremely long sequences for practical applications.
Syllabus
Titans: Learning to Memorize at Test Time (Paper Analysis)
Taught by
Yannic Kilcher