Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

MIT: Recurrent Neural Networks

Alexander Amini and Massachusetts Institute of Technology via YouTube

Overview

Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Explore deep sequence modeling with recurrent neural networks in this lecture from MIT's Introduction to Deep Learning course. Dive into the challenges of modeling long-term dependencies in sequences and learn about various RNN architectures, including standard RNNs and Long Short-Term Memory (LSTM) networks. Discover techniques for addressing gradient flow issues, such as exploding and vanishing gradients. Examine practical applications of RNNs in music generation, sentiment classification, and machine translation. Gain insights into attention mechanisms and their role in improving sequence modeling performance. Enhance your understanding of deep learning techniques for processing and generating sequential data.

Syllabus

Intro
Sequences in the wild
A sequence modeling problem: predict the next word
use a fixed window
can't model long-term dependencies
use entire sequence as set of counts
counts don't preserve order
use a really big fixed window
no parameter sharing
Sequence modeling: design criteria
Standard feed-forward neural network
Recurrent neural networks: sequence modeling
A standard "vanilla" neural network
A recurrent neural network (RNN)
RNN state update and output
RNNs: computational graph across time
Recall: backpropagation in feed forward models
RNNs: backpropagation through time
Standard RNN gradient flow: exploding gradients
Standard RNN gradient flow:vanishing gradients
The problem of long-term dependencies
Trick #1: activation functions
Trick #2: parameter initialization
Standard RNN In a standard RNN repeating modules contain a simple computation node
Long Short Term Memory (LSTMs)
LSTMs: forget irrelevant information
LSTMs: output filtered version of cell state
LSTM gradient flow
Example task: music generation
Example task: sentiment classification
Example task: machine translation
Attention mechanisms
Recurrent neural networks (RNNs)

Taught by

https://www.youtube.com/@AAmini/videos

Reviews

5.0 rating, based on 2 Class Central reviews

Start your review of MIT: Recurrent Neural Networks

  • This MIT lecture series on RNNs is a brilliant deep dive into the subject! The instructor explains complex concepts with clarity, covering everything from basic RNNs to advanced architectures like LSTMs and GRUs. The mathematical foundations are well-presented, making it ideal for those who want a rigorous yet accessible understanding. The examples and visualizations enhance comprehension, and the pacing keeps you engaged. While some prior ML knowledge helps, the explanations are thorough enough for motivated learners. A fantastic resource for students and professionals alike—concise, high-quality, and packed with insights. Highly recommended for anyone serious about mastering RNNs! 9.5/10!

  • The course is very good, I really liked the content covered, the teacher is excellent and the explanations are great.

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.