Recurrent Neural Networks (RNN)

Reference

Why recurrent

RNNs are called recurrent because they perform the same task for every element of a sequence, with the output being depended on the previous computations.

Another way to think about RNNs is that they have a “memory” which captures information about what has been calculated so far.

Architecture and components

RNN-1.png|400

h(t)=f(Ux(t)+Wh(t1))

Forward propagation


where the tanh can be ReLu too and the softmax can be sigmoid too.

Back propagation

The gradient computation involves performing a forward propagation pass moving left to right through the graph shown above followed by a backward propagation pass moving right to left through the graph.

Different types of RNN

Pasted image 20240825204844.png|500

Variations of RNN

Pros & Cons