Towards better understanding and improving optimization in recurrent neural networks
Recurrent neural networks (RNN) are known for their notorious exploding and vanishing gradient problem (EVGP). This problem becomes more evident in tasks where the information needed to correctly solve them exist over long time scales, because it prevents important gradient components from being bac...
Main Author: | |
---|---|
Other Authors: | |
Language: | English |
Published: |
2021
|
Subjects: | |
Online Access: | http://hdl.handle.net/1866/24319 |