Recurrent Neural Networks, Transformers, and Attention - MIT 6.S191 Lecture 2

Recurrent Neural Networks, Transformers, and Attention - MIT 6.S191 Lecture 2

https://www.youtube.com/@AAmini/videos via YouTube Direct link

​ - Introduction

1 of 19

1 of 19

​ - Introduction

Class Central Classrooms beta

YouTube videos curated by Class Central.

Classroom Contents

Recurrent Neural Networks, Transformers, and Attention - MIT 6.S191 Lecture 2

Automatically move to the next video in the Classroom when playback concludes

  1. 1 ​ - Introduction
  2. 2 ​ - Sequence modeling
  3. 3 ​ - Neurons with recurrence
  4. 4 - Recurrent neural networks
  5. 5 - RNN intuition
  6. 6 ​ - Unfolding RNNs
  7. 7 - RNNs from scratch
  8. 8 - Design criteria for sequential modeling
  9. 9 - Word prediction example
  10. 10 ​ - Backpropagation through time
  11. 11 - Gradient issues
  12. 12 ​ - Long short term memory LSTM
  13. 13 ​ - RNN applications
  14. 14 - Attention fundamentals
  15. 15 - Intuition of attention
  16. 16 - Attention and search relationship
  17. 17 - Learning attention with neural networks
  18. 18 - Scaling attention and applications
  19. 19 - Summary

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.