Learn Before
Concept

Math behind the simple RNNs

For each neuron in the sequence, the output \hat y^{} and the hidden state h^{} at time step t can be calculated as follows. h^{}=g_1(Wh^{} \oplus Ux^{}+b_h) \hat y^{}=g_2(Vh^{}+b_y) , where g_1, g_2 are activation functions, U,V,W are trainable parameter matrixes, b_h and b_y are bias terms, and \oplus denotes that it needs to be summed up element-wise.

When the network is trained, it not only assigns weights U to each neuron’s inputs, but also discovers the weight parameters W of the hidden state function. These parameters define how much of the information from the previous steps should be carried forward to each subsequent step.

0

2

Updated 2020-10-03

Tags

Data Science

Related