Input → [Update Gate] → [Hidden State] → Output
⬆️ ⬇️
[Reset Gate] → (Controls memory update)
Imagine you are reading a story.
You don’t need to remember every word, just the important details.
💡 GRUs are like LSTMs but simpler and faster!
They remember important things and forget unnecessary ones, helping computers understand speech, text, and time-series data better.
🔹 GRUs = A Simpler, Faster LSTM!
🔹 GRUs use 2 gates:
Gate | What It Does | Real-Life Example |
---|---|---|
Reset Gate 🔄 | Decides how much of the past to forget. | "Forget past minor details, just focus on key parts." |
Update Gate ✅ | Decides how much of the past to keep. | "Keep the important parts and carry them forward." |
💡 GRUs decide what to forget and what to keep dynamically.
At every step tt:
1️⃣ Reset Gate (Forget Old Info)
$$ r_t = \sigma(W_r \cdot [h_{t-1}, x_t]) $$