Tiven Wang
Wang Tiven August 01, 2018
425 favorite favorites
bookmark bookmark
share share

Recurrent Neural Networks

Markov chain model

https://colah.github.io/posts/2015-08-Understanding-LSTMs/

https://www.tensorflow.org/tutorials/sequences/recurrent

在使用神经网络时,很重要的一点是正确的处理数据维度。Recurrent Neural Networks 输入数据一般使用 3D 容积输入

  • Mini-batch Size 一批多少个数据
  • Number of columns in our vector per time-step 每一步向量里多少列
  • Number of time-steps 多少步

IMAGE: recurrent neural networks input

LSTMs

梯度消失问题 (vanishing gradient problem)

Long short-term memory (LSTM)

Gated recurrent unit (GRUs) are a gating mechanism in recurrent neural networks, introduced in 2014 by Kyunghyun Cho et al. Their performance on polyphonic music modeling and speech signal modeling was found to be similar to that of long short-term memory (LSTM). However, GRUs have been shown to exhibit better performance on smaller datasets. They have fewer parameters than LSTM, as they lack an output gate.

RMSprop optimizer

https://www.quora.com/Why-is-it-said-that-RMSprop-optimizer-is-recommended-in-training-recurrent-neural-networks-What-is-the-explanation-behind-it

http://ruder.io/optimizing-gradient-descent/

Hyperparameters

hidden_layer_size 如何选择?

BPTT

https://distill.pub/2016/augmented-rnns/

References

Similar Posts

Comments

Back to Top