Credit: Deep Learning A-Z™: Hands-On Artificial Neural Networks
Plan of attack:
The whole concept behind deep learning is to try to mimic the human brain.
ANN
CNN
RNN
We have a simple Artificial Neural Network: 1 input, 1 output, and 1 hidden layer.
We can represent the RNN by squashing the ANN. Think of it as we are looking the ANN from underneath - as a new dimension.
To simplify things, we change multiple arrows to 2, then we twist the ANN to make it vertical.
We change the color of the hidden layer from green to blue and add another line, which represents the temporal loop. This means the hidden layer not only gives an output but also feeds back to itself.
The official representation of RNN is to unroll the temporal loop, and put RNN in the following manner.
Note that we are looking in the new dimension, so each one circle is not 1 neuron but actually a whole layer of neurons.
This means we have inputs coming into the neurons then you have output but also the neurons connect to themselves through time. That is the whole concept when they have short-term memory that they remember what was in that neuron just previously.
Here we have $w_{rec}$, stands for weight recurring, that is the weight used to connect the hidden layers to themselves in the unrolled temporal loop.
In summary
Solutions:
Additional Reading