Jump to content

Gated recurrent unit

From Wikipedia, the free encyclopedia

This is an old revision of this page, as edited by Selichm (talk | contribs) at 09:41, 23 November 2016. The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

Gated recurrent units are a gating mechanism in recurrent neural networks, introduced in 2014. Their performance on polyphonic music modeling and speech signal modeling was found to be similar to that of long short-term memory.[1] They have fewer parameters than LSTM, as they lack an output gate.[2]

Architecture

. denotes the Hadamard product.

Variables

  • : input vector
  • : output vector
  • : update gate vector
  • : reset gate vector
  • , and : parameter matrices and vector

Activation functions

References

  1. ^ Chung, Junyoung; Gulcehre, Caglar; Cho, KyungHyun; Bengio, Yoshua (2014). "Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling". arXiv:1412.3555 [cs.NE].
  2. ^ "Recurrent Neural Network Tutorial, Part 4 – Implementing a GRU/LSTM RNN with Python and Theano – WildML". Wildml.com. Retrieved May 18, 2016.