Title
A Simple Way to Initialize Recurrent Networks of Rectified Linear Units.
Abstract
Learning long term dependencies in recurrent networks is difficult due to vanishing and exploding gradients. To overcome this difficulty, researchers have developed sophisticated optimization techniques and network architectures. In this paper, we propose a simpler solution that use recurrent neural networks composed of rectified linear units. Key to our solution is the use of the identity matrix or its scaled version to initialize the recurrent weight matrix. We find that our solution is comparable to LSTM on our four benchmarks: two toy problems involving long-range temporal structures, a large language modeling problem and a benchmark speech recognition problem.
Year
Venue
Field
2015
CoRR
Rectifier (neural networks),Computer science,Matrix (mathematics),Network architecture,Recurrent neural network,Artificial intelligence,Identity matrix,Language model,Machine learning
DocType
Volume
Citations 
Journal
abs/1504.00941
94
PageRank 
References 
Authors
3.55
16
3
Name
Order
Citations
PageRank
Quoc V. Le18501366.59
Navdeep Jaitly22988166.08
geoffrey e hinton3404354751.69