Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

MIT OpenCourseWare

Deep Learning for Natural Language - Transformers - Lecture 7

MIT OpenCourseWare via YouTube

Overview

Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Explore the revolutionary Transformer architecture in this comprehensive lecture from MIT's Hands-On Deep Learning course. Learn how Transformers work through a practical airline travel-related example that demonstrates the model's ability to process and understand natural language sequences. Discover the key components of Transformer models including attention mechanisms, encoder-decoder structures, and how they revolutionized natural language processing tasks. Understand the mathematical foundations behind self-attention and multi-head attention that enable Transformers to capture long-range dependencies in text. Examine how these models process sequential data differently from traditional RNNs and CNNs, making them more efficient and effective for language tasks. Gain insights into the architecture that powers modern language models and has become the foundation for breakthrough applications in machine translation, text generation, and language understanding.

Syllabus

7: Deep Learning for Natural Language – Transformers

Taught by

MIT OpenCourseWare

Reviews

Start your review of Deep Learning for Natural Language - Transformers - Lecture 7

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.