Reservoir computing: Difference between revisions
Added a brief sub-section including recent advances of the Reservoir Computing paradigm for learning in Structured Domains. |
|||
Line 51: | Line 51: | ||
* [http://www.opticsinfobase.org/oe/abstract.cfm?uri=oe-20-3-3241 Optoelectronic Reservoir Computing], Optics Express 2012 |
* [http://www.opticsinfobase.org/oe/abstract.cfm?uri=oe-20-3-3241 Optoelectronic Reservoir Computing], Optics Express 2012 |
||
* [http://www.nature.com/ncomms/journal/v4/n1/full/ncomms2368.html All-optical Reservoir Computing], Nature Communications 2013 |
* [http://www.nature.com/ncomms/journal/v4/n1/full/ncomms2368.html All-optical Reservoir Computing], Nature Communications 2013 |
||
* [http://www.mitpressjournals.org/doi/10.1162/NECO_a_00694#.WL4P9iHyvIo Memristor Models for Machine learning], Neural Computation 2014 [https://arxiv.org/abs/1406.2210 arxiv] |
|||
* [[Deep Learning]] |
* [[Deep Learning]] |
||
* [[Extreme Learning Machines]] |
* [[Extreme Learning Machines]] |
Revision as of 01:42, 7 March 2017
Reservoir computing is a framework for computation that may be viewed as an extension of neural networks.[1] Typically an input signal is fed into a fixed (random) dynamical system called a reservoir and the dynamics of the reservoir map the input to a higher dimension. Then a simple readout mechanism is trained to read the state of the reservoir and map it to the desired output. The main benefit is that the training is performed only at the readout stage and the reservoir is fixed. Liquid-state machines [2] and echo state networks [3] are two major types of reservoir computing.[4]
Reservoir
The reservoir consists of a collection of recurrently connected units. The connectivity structure is usually random, and the units are usually non-linear. The overall dynamics of the reservoir is driven by the input, and also affected by the past. A rich collection of dynamical input-output mapping is a crucial advantage over simple time delay neural networks.
Readout
The readout is carried out using a linear transformation of the reservoir output. This transformation is adapted to the task of interest by using a linear regression or a Ridge regression using a teaching signal.
Types
Context reverberation network
An early example of reservoir computing was the context reverberation network .[5] In this architecture, an input layer feeds into a high dimensional dynamical system which is read out by a trainable single-layer perceptron. Two kinds of dynamical system were described: a recurrent neural network with fixed random weights, and a continuous reaction-diffusion system inspired by Alan Turing’s model of morphogenesis. At the trainable layer, the perceptron associates current inputs with the signals that reverberate in the dynamical system; the latter were said to provide a dynamic "context" for the inputs. In the language of later work, the reaction-diffusion system served as the reservoir.
Echo state network
Backpropagation-decorrelation
Backpropagation-Decorrelation (BPDC)
Liquid-state machine
Reservoir Computing for Structured Data
The Tree Echo State Network [6] (TreeESN) model represents a generalization of the Reservoir Computing framework to tree structured data.
See also
- Reservoir Computing using delay systems, Nature Communications 2011
- Optoelectronic Reservoir Computing, Scientific Reports February 2012
- Optoelectronic Reservoir Computing, Optics Express 2012
- All-optical Reservoir Computing, Nature Communications 2013
- Memristor Models for Machine learning, Neural Computation 2014 arxiv
- Deep Learning
- Extreme Learning Machines
References
- ^ Schrauwen, Benjamin, David Verstraeten, and Jan Van Campenhout. "An overview of reservoir computing: theory, applications, and implementations." Proceedings of the European Symposium on Artificial Neural Networks ESANN 2007, pp. 471-482.
- ^ Mass, Wolfgang, T. Nachtschlaeger, and H. Markram. "Real-time computing without stable states: A new framework for neural computation based on perturbations." Neural Computation 14(11): 2531–2560 (2002).
- ^ Jaeger, Herbert, "The echo state approach to analyzing and training recurrent neural networks." Technical Report 154 (2001), German National Research Center for Information Technology.
- ^ Echo state network, Scholarpedia
- ^ Kirby, Kevin. "Context dynamics in neural sequential learning." Proceedings of the Florida Artificial Intelligence Research Symposium FLAIRS (1991), 66-70.
- ^ Gallicchio, Claudio; Micheli, Alessio (2013). "Tree Echo State Networks". Neurocomputing. 101: 319–337.