Input → [Hidden State] → Output
⬆️ ⬇️
(Recurrent loop)
Imagine you’re reading a sentence. You don’t just look at one word at a time and forget everything before it—you remember previous words to understand the whole sentence.
That’s exactly what Recurrent Neural Networks (RNNs) do! They’re a type of neural network designed to remember past information, making them great for working with sequential data (like sentences, speech, and time-series data).
Most traditional neural networks don’t remember past inputs. For example, a regular neural network sees only one word at a time and doesn't recall the words before it.
🔴 Regular Neural Network (No memory):
Each input is treated independently.
"I" → "am" → "learning" → "RNNs"
(It doesn’t know that "I" and "am" came before "learning".)
🟢 RNN (Has memory!):
Each word influences the next prediction because it remembers previous words.
("I" → "am") → ("am" → "learning") → ("learning" → "RNNs")
(Now, it remembers context from previous words.)
An RNN loops through data, keeping a memory of past inputs.
Each step consists of:
Mathematically, at each time step $t$:
$$ h_t = f(W_h h_{t-1} + W_x x_t) $$