Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

MIT OpenCourseWare

Deep Learning for Natural Language - Transformers, Self-Supervised Learning - Lecture 8

MIT OpenCourseWare via YouTube

Overview

Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Explore the architecture and applications of transformers in natural language processing through this comprehensive lecture from MIT's Hands-On Deep Learning course. Delve into the fundamental concepts of transformer models, understanding their revolutionary impact on NLP tasks and how they have become the backbone of modern language models. Learn about self-supervised learning techniques and how they enable transformers to learn meaningful representations from large amounts of unlabeled text data. Examine the attention mechanism that allows transformers to process sequences more effectively than traditional recurrent neural networks. Discover practical implementation strategies for using pre-trained transformer models and fine-tuning them for specific NLP applications. Gain insights into the mathematical foundations underlying transformer architectures, including multi-head attention, positional encoding, and layer normalization. Understand how self-supervised learning paradigms like masked language modeling and next sentence prediction enable transformers to capture complex linguistic patterns and semantic relationships.

Syllabus

8: Deep Learning for Natural Language – Transformers, Self-Supervised Learning

Taught by

MIT OpenCourseWare

Reviews

Start your review of Deep Learning for Natural Language - Transformers, Self-Supervised Learning - Lecture 8

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.