Input → [Hidden State] → Output
         ⬆️    ⬇️  
      (Recurrent loop)

Imagine you’re reading a sentence. You don’t just look at one word at a time and forget everything before it—you remember previous words to understand the whole sentence.

That’s exactly what Recurrent Neural Networks (RNNs) do! They’re a type of neural network designed to remember past information, making them great for working with sequential data (like sentences, speech, and time-series data).

1. What Makes RNNs Special?

Most traditional neural networks don’t remember past inputs. For example, a regular neural network sees only one word at a time and doesn't recall the words before it.

🔴 Regular Neural Network (No memory):

Each input is treated independently.

"I" → "am" → "learning" → "RNNs"

(It doesn’t know that "I" and "am" came before "learning".)

🟢 RNN (Has memory!):

Each word influences the next prediction because it remembers previous words.

("I" → "am") → ("am" → "learning") → ("learning" → "RNNs")

(Now, it remembers context from previous words.)

2. How Does an RNN Work?

An RNN loops through data, keeping a memory of past inputs.

Each step consists of:

  1. Taking input (e.g., a word in a sentence).
  2. Updating memory (remembering past words).
  3. Predicting output (the next word, for example).

Mathematically, at each time step $t$:

$$ h_t = f(W_h h_{t-1} + W_x x_t) $$