All Courses
5 lessons
RNN & LSTM
Sequence modeling before attention — and the problems that motivated it
Lessons
- 01
Recurrent Neural Network
Hidden state, shared weights, sequential processing.
MediumOpen - 02
Backprop Through Time
Unroll the loop to compute gradients across a sequence.
HardOpen - 03
Vanishing Gradient Problem
Why long sequences kill plain RNNs — analytically.
HardOpen - 04
LSTM
Gates, cell state, and the first real fix for long memory.
HardOpen - 05
GRU
A lighter LSTM that often matches it.
MediumOpen