Attention and Transformers in Advanced NLP - Lecture 4
Graham Neubig via YouTube
The Fastest Way to Become a Backend Developer Online
Future-Proof Your Career: AI Manager Masterclass
Overview
Google, IBM & Meta Certificates — All 10,000+ Courses at 40% Off
One annual plan covers every course and certificate on Coursera. 40% off for a limited time.
Get Full Access
Dive into the intricacies of advanced natural language processing techniques in this comprehensive lecture from CMU's CS 11-711 course. Explore the fundamental concepts of attention mechanisms and the revolutionary Transformer architecture. Gain a deep understanding of multi-head attention, positional encodings, and layer normalization. Delve into optimizers and training strategies for large language models. Examine the LLaMa architecture and its significance in the field. This 1-hour 19-minute session, led by Graham Neubig, provides a thorough exploration of cutting-edge NLP technologies that form the backbone of modern language understanding and generation systems.
Syllabus
CMU Advanced NLP Fall 2024 (4): Attention and Transformers
Taught by
Graham Neubig