Input → [Update Gate] → [Hidden State] → Output
        ⬆️           ⬇️
       [Reset Gate] → (Controls memory update)

Imagine you are reading a story.

You don’t need to remember every word, just the important details.

💡 GRUs are like LSTMs but simpler and faster!

They remember important things and forget unnecessary ones, helping computers understand speech, text, and time-series data better.

1. Why Do We Need GRUs?

Problem with RNNs: They Forget!

🔹 GRUs = A Simpler, Faster LSTM!

2. How Do GRUs Work?

🔹 GRUs use 2 gates:

Gate What It Does Real-Life Example
Reset Gate 🔄 Decides how much of the past to forget. "Forget past minor details, just focus on key parts."
Update Gate Decides how much of the past to keep. "Keep the important parts and carry them forward."

💡 GRUs decide what to forget and what to keep dynamically.

3. How GRUs Work Mathematically (Simplified)

At every step tt:

1️⃣ Reset Gate (Forget Old Info)

$$ r_t = \sigma(W_r \cdot [h_{t-1}, x_t]) $$