Learn Before
Concept
When are exploding/vanishing gradients a problem?
Exploding/vanshing gradients is a common problem in reccurrent neural networks. Consider a network where an input is multiplied by a matrix times. Let have eigendecomposition .
We can see that if , our result will approach as gets large. This is an exploding gradient. Also, if , the result will approach $0t$ gets large. This is a vanishing gradient.
0
2
Updated 2021-06-24
Tags
Data Science