Winner-take-all (computing): Difference between revisions

From Wikipedia, the free encyclopedia
Content deleted Content added
→‎Other uses: First-mover advantage
Tags: Mobile edit Mobile web edit Advanced mobile edit
Kegell (talk | contribs)
Inline citations controlled and added, template message removed
Line 1: Line 1:
{{Other uses|Winner takes all (disambiguation){{!}}Winner takes all}}'''Winner-take-all''' is a computational principle applied in computational [[models of neural network]]s by which [[neuron]]s compete with each other for activation. In the classical form, only the neuron with the highest activation stays active while all other neurons shut down; however, other variations allow more than one neuron to be active, for example the soft winner take-all, by which a power function is applied to the neurons.
{{Short description|Computational principle applied in neural networks}}
{{Other uses|Winner takes all (disambiguation){{!}}Winner takes all}}
{{no footnotes|date=October 2012}}

'''Winner-take-all''' is a computational principle applied in computational [[models of neural network]]s by which [[neuron]]s in a layer compete with each other for activation. In the classical form, only the neuron with the highest activation stays active while all other neurons shut down; however, other variations allow more than one neuron to be active, for example the soft winner take-all, by which a power function is applied to the neurons.


==Neural networks==
==Neural networks==
In the theory of [[artificial neural network]]s, '''winner-take-all''' networks are a case of [[competitive learning]] in [[recurrent neural network]]s. Output nodes in the network mutually inhibit each other, while simultaneously activating themselves through reflexive connections. After some time, only one node in the output layer will be active, namely the one corresponding to the strongest input. Thus the network uses nonlinear inhibition to pick out the largest of a set of inputs. Winner-take-all is a general computational primitive that can be implemented using different types of neural network models, including both continuous-time and spiking networks (Grossberg, 1973; Oster et al. 2009).
In the theory of [[artificial neural network]]s, '''winner-take-all''' networks are a case of [[competitive learning]] in [[recurrent neural network]]s. Output nodes in the network mutually inhibit each other, while simultaneously activating themselves through reflexive connections. After some time, only one node in the output layer will be active, namely the one corresponding to the strongest input. Thus the network uses nonlinear inhibition to pick out the largest of a set of inputs. Winner-take-all is a general computational primitive that can be implemented using different types of neural network models, including both continuous-time and spiking networks<ref>{{Citation |last=Grossberg |first=Stephen |title=Contour Enhancement, Short Term Memory, and Constancies in Reverberating Neural Networks |date=1982 |url=http://link.springer.com/10.1007/978-94-009-7758-7_8 |work=Studies of Mind and Brain |volume=70 |pages=332–378 |place=Dordrecht |publisher=Springer Netherlands |doi=10.1007/978-94-009-7758-7_8 |isbn=978-90-277-1360-5 |access-date=2022-11-05}}</ref><ref name=":0">{{Cite journal |last=Oster |first=Matthias |last2=Rodney |first2=Douglas |last3=Liu |first3=Shih-Chii |date=2009 |title=Computation with Spikes in a Winner-Take-All Network |url=https://direct.mit.edu/neco/article-abstract/21/9/2437/7472/Computation-with-Spikes-in-a-Winner-Take-All |journal=Neural computation |volume=21 |issue=9 |pages=2437-2465 |doi=10.1162/neco.2009.07-08-829}}</ref>.


Winner-take-all networks are commonly used in computational models of the brain, particularly for distributed decision-making or [[Winner-take-all in action selection|action selection]] in the [[cortex (anatomy)|cortex]]. Important examples include hierarchical models of vision (Riesenhuber et al. 1999), and models of selective attention and recognition (Carpenter and Grossberg, 1987; Itti et al. 1998). They are also common in artificial neural networks and neuromorphic analog VLSI circuits. It has been formally proven that the winner-take-all operation is computationally powerful compared to other nonlinear operations, such as thresholding (Maass 2000).
Winner-take-all networks are commonly used in computational models of the brain, particularly for distributed decision-making or [[Winner-take-all in action selection|action selection]] in the [[cortex (anatomy)|cortex]]. Important examples include hierarchical models of vision<ref>{{Cite journal |last=Riesenhuber |first=Maximilian |last2=Poggio |first2=Tomaso |date=1999-11-01 |title=Hierarchical models of object recognition in cortex |url=https://www.nature.com/articles/nn1199_1019 |journal=Nature Neuroscience |language=en |volume=2 |issue=11 |pages=1019–1025 |doi=10.1038/14819 |issn=1097-6256}}</ref>, and models of selective attention and recognition<ref>{{Cite journal |last=Carpenter |first=Gail A. |date=1987 |title=A massively parallel architecture for a self-organizing neural pattern recognition machine |url=https://www.sciencedirect.com/science/article/abs/pii/S0734189X87800142 |journal=Computer Vision, Graphics, and Image Processing |volume=37 |issue=1 |pages=54-115}}</ref><ref>{{Cite journal |last=Itti |first=Laurent |last2=Koch |first2=Christof |date=1998 |title=A Model of Saliency-Based Visual Attention for Rapid Scene Analysis |url=https://dl.acm.org/doi/10.1109/34.730558 |journal=IEEE Transactions on Pattern Analysis and Machine Intelligence |volume=20 |issue=11 |pages=1254–1259 |doi=10.1109/34.730558}}</ref>. They are also common in artificial neural networks and neuromorphic analog VLSI circuits. It has been formally proven that the winner-take-all operation is computationally powerful compared to other nonlinear operations, such as thresholding<ref>{{Cite journal |last=Maass |first=Wolfgang |date=2000-11-01 |title=On the Computational Power of Winner-Take-All |url=https://direct.mit.edu/neco/article/12/11/2519-2535/6425 |journal=Neural Computation |language=en |volume=12 |issue=11 |pages=2519–2535 |doi=10.1162/089976600300014827 |issn=0899-7667}}</ref>.


In many practical cases, there is not only a single neuron which becomes the only active one but there are exactly ''k'' neurons which become active for a fixed number ''k''. This principle is referred to as '''k-winners-take-all'''.
In many practical cases, there is not only one single neuron which becomes active but there are exactly ''k'' neurons which become active for a fixed number ''k''. This principle is referred to as '''k-winners-take-all'''.


==Circuit example==
==Circuit example==
[[Image:Wta 2input.png|thumb|200px|right|A two-input CMOS winner-take-all circuit]]
[[Image:Wta 2input.png|thumb|200px|right|A two-input CMOS winner-take-all circuit]]


A simple, but popular [[CMOS]] winner-take-all circuit is shown on the right. This circuit was originally proposed by Lazzaro et al. (1989) using MOS transistors biased to operate in the weak-inversion or subthreshold regime. In the particular case shown there are only two inputs (''I''<sub>''IN'',1</sub> and ''I''<sub>''IN'',2</sub>), but the circuit can be easily extended to multiple inputs in a straightforward way. It operates on continuous-time input signals (currents) in parallel, using only two transistors per input. In addition, the bias current ''I''<sub>''BIAS''</sub> is set by a single global transistor that is common to all the inputs.
A simple, but popular [[CMOS]] winner-take-all circuit is shown on the right. This circuit was originally proposed by Lazzaro et al. (1989)<ref>{{Cite journal |last=Lazzaro |first=J. |last2=Ryckebusch |first2=S. |last3=Mahowald |first3=M. A. |last4=Mead |first4=C. A. |date=1988-01-01 |title=Winner-Take-All Networks of O(N) Complexity: |url=http://www.dtic.mil/docs/citations/ADA451466 |location=Fort Belvoir, VA |doi=10.21236/ada451466}}</ref> using [[MOSFET|MOS transistors]] biased to operate in the weak-inversion or subthreshold regime. In the particular case shown there are only two inputs (''I''<sub>''IN'',1</sub> and ''I''<sub>''IN'',2</sub>), but the circuit can be easily extended to multiple inputs in a straightforward way. It operates on continuous-time input signals (currents) in parallel, using only two transistors per input. In addition, the bias current ''I''<sub>''BIAS''</sub> is set by a single global transistor that is common to all the inputs.


The largest of the input currents sets the common potential ''V''<sub>''C''</sub>. As a result, the corresponding output carries almost all the bias current, while the other outputs have currents that are close to zero. Thus, the circuit selects the larger of the two input currents, i.e., if ''I''<sub>''IN'',1</sub> > ''I''<sub>''IN'',2</sub>, we get ''I''<sub>''OUT'',1</sub> = ''I''<sub>''BIAS''</sub> and ''I''<sub>''OUT'',2</sub> = 0. Similarly, if ''I''<sub>''IN'',2</sub> > ''I''<sub>''IN'',1</sub>, we get ''I''<sub>''OUT'',1</sub> = 0 and ''I''<sub>''OUT'',2</sub> = ''I''<sub>''BIAS''</sub>.
The largest of the input currents sets the common potential ''V''<sub>''C''</sub>. As a result, the corresponding output carries almost all the bias current, while the other outputs have currents that are close to zero. Thus, the circuit selects the larger of the two input currents, i.e., if ''I''<sub>''IN'',1</sub> > ''I''<sub>''IN'',2</sub>, we get ''I''<sub>''OUT'',1</sub> = ''I''<sub>''BIAS''</sub> and ''I''<sub>''OUT'',2</sub> = 0. Similarly, if ''I''<sub>''IN'',2</sub> > ''I''<sub>''IN'',1</sub>, we get ''I''<sub>''OUT'',1</sub> = 0 and ''I''<sub>''OUT'',2</sub> = ''I''<sub>''BIAS''</sub>.
Line 24: Line 20:


==Other uses==
==Other uses==
In '''stereo matching''' algorithms, following the taxonomy proposed by Scharstein et al. (IJCV 2002), winner-take-all is a local method for disparity computation. Adopting a winner-take-all strategy, the disparity associated with the minimum or maximum cost value is selected at each pixel.
In '''stereo matching''' algorithms, following the taxonomy proposed by Scharstein and Szelliski<ref>{{Cite journal |last=Scharstein |first=Daniel |last2=Szeliski |first2=Richard |date=2002 |title=A Taxonomy and Evaluation of Dense Two-Frame Stereo Correspondence Algorithms |url=http://link.springer.com/10.1023/A:1014573219977 |journal=International Journal of Computer Vision |volume=47 |issue=1/3 |pages=7–42 |doi=10.1023/A:1014573219977}}</ref>, winner-take-all is a local method for disparity computation. Adopting a winner-take-all strategy, the disparity associated with the minimum or maximum cost value is selected at each pixel.


It is axiomatic that in the electronic commerce market, early dominant players such as [[AOL]] or [[Yahoo!]] get most of the rewards. By 1998, one study{{Clarifyme|date=February 2009}} found the top 5% of all web sites garnered more than 74% of all traffic.
It is axiomatic that in the electronic commerce market, early dominant players such as [[AOL]] or [[Yahoo!]] get most of the rewards. By 1998, one study{{Clarifyme|date=February 2009}} found the top 5% of all web sites garnered more than 74% of all traffic.


The '''[[winner take all hypothesis]]''' suggests that once a technology or a firm gets ahead, it will do better and better over time, whereas lagging technology and firms will fall further behind. See [[First-mover advantage]].
The winner-take-all hypothesis in economics suggests that once a technology or a firm gets ahead, it will do better and better over time, whereas lagging technology and firms will fall further behind. See [[First-mover advantage|First-mover adva hypothntage]].


==See also==
==See also==
Line 38: Line 34:
* G.A. Carpenter and S. Grossberg, [http://techlab.bu.edu/resources/articles/] A massively parallel architecture for a self-organizing neural pattern recognition machine, "Computer Vision, Graphics, and Image Processing", "'37:54"', 1987.
* G.A. Carpenter and S. Grossberg, [http://techlab.bu.edu/resources/articles/] A massively parallel architecture for a self-organizing neural pattern recognition machine, "Computer Vision, Graphics, and Image Processing", "'37:54"', 1987.
* S. Grossberg, [http://cns.bu.edu/~steve] Contour enhancement, short-term memory, and constancies in reverberating neural networks, "Studies in Applied Mathematics", "'52:213"', 1973.
* S. Grossberg, [http://cns.bu.edu/~steve] Contour enhancement, short-term memory, and constancies in reverberating neural networks, "Studies in Applied Mathematics", "'52:213"', 1973.
* M. Oster, R. Douglas and S.-C. Liu, [http://www.ini.ch/~shih/papers/OsterLiuNC2009.pdf Computation with spikes in a winner-take-all network], ''Neural Computation'', '''21:9''', 2009.
* M. Oster, R. Douglas and S.-C. Liu, [http://www.ini.ch/~shih/papers/OsterLiuNC2009.pdf Computation with spikes in a winner-take-all Oster et al. 2009network], ''Neural Computation'', '''21:9''', 2009.
* M. Riesenhuber and T. Poggio, [https://www.cs.cmu.edu/afs/cs/academic/class/15883-f15/readings/riesenhuber-1999.pdf Hierarchical models of object recognition in cortex], ''Nature Neuroscience'', '''2:11''', 1999.
* M. Riesenhuber and T. Poggio, [https://www.cs.cmu.edu/afs/cs/academic/class/15883-f15/readings/riesenhuber-1999.pdf Hierarchical models of object recognition in cortex], ''Nature Neuroscience'', '''2:11''', 1999.
* L. Itti, C. Koch and E. Niebur, [http://portal.acm.org/citation.cfm?id=297870 A model of saliency-based visual attention for rapid scene analysis], ''IEEE Transactions on Pattern Analysis and Machine Intelligence'', '''20:11''', 1998.
* L. Itti, C. Koch and E. Niebur, [http://portal.acm.org/citation.cfm?id=297870 A mod]<ref name=":0" />[http://portal.acm.org/citation.cfm?id=297870 el of saliency-based visual attention for rapid scene analysis], ''IEEE Transactions on Pattern Analysis and Machine Intelligence'', '''20:11''', 1998.
* W. Maass, [http://portal.acm.org/citation.cfm?id=1121905 On the computational power of winner-take-all], ''Neural Computation'', '''12:11''', 2000.
* W. Maass, [http://portal.acm.org/citation.cfm?id=1121905 On the computational power of winner-take-all], ''Neural Computation'', '''12:11''', 2000.
* J. Lazzaro, S. Ryckebusch, M. A. Mahowald and C. A. Mead, [http://portal.acm.org/citation.cfm?id=89944 Winner-take-all networks of O(N) complexity], in Advances in Neural Information Processing Systems 1, Morgan Kaufmann Publishers, San Francisco, CA, 1989. Also available online at [http://www.cs.berkeley.edu/~lazzaro/biblio/wta-tech.pdf John Lazzaro's website].
* J. Lazzaro, S. Ryckebusch, M. A. Mahowald and C. A. Mead, [http://portal.acm.org/citation.cfm?id=89944 Winner-take-all networks of O(N) complexity], in Advances in Neural Information Processing Systems 1, Morgan Kaufmann Publishers, San Francisco, CA, 1989. Also available online at [http://www.cs.berkeley.edu/~lazzaro/biblio/wta-tech.pdf John Lazzaro's website].

Revision as of 17:39, 5 November 2022

Winner-take-all is a computational principle applied in computational models of neural networks by which neurons compete with each other for activation. In the classical form, only the neuron with the highest activation stays active while all other neurons shut down; however, other variations allow more than one neuron to be active, for example the soft winner take-all, by which a power function is applied to the neurons.

Neural networks

In the theory of artificial neural networks, winner-take-all networks are a case of competitive learning in recurrent neural networks. Output nodes in the network mutually inhibit each other, while simultaneously activating themselves through reflexive connections. After some time, only one node in the output layer will be active, namely the one corresponding to the strongest input. Thus the network uses nonlinear inhibition to pick out the largest of a set of inputs. Winner-take-all is a general computational primitive that can be implemented using different types of neural network models, including both continuous-time and spiking networks[1][2].

Winner-take-all networks are commonly used in computational models of the brain, particularly for distributed decision-making or action selection in the cortex. Important examples include hierarchical models of vision[3], and models of selective attention and recognition[4][5]. They are also common in artificial neural networks and neuromorphic analog VLSI circuits. It has been formally proven that the winner-take-all operation is computationally powerful compared to other nonlinear operations, such as thresholding[6].

In many practical cases, there is not only one single neuron which becomes active but there are exactly k neurons which become active for a fixed number k. This principle is referred to as k-winners-take-all.

Circuit example

A two-input CMOS winner-take-all circuit

A simple, but popular CMOS winner-take-all circuit is shown on the right. This circuit was originally proposed by Lazzaro et al. (1989)[7] using MOS transistors biased to operate in the weak-inversion or subthreshold regime. In the particular case shown there are only two inputs (IIN,1 and IIN,2), but the circuit can be easily extended to multiple inputs in a straightforward way. It operates on continuous-time input signals (currents) in parallel, using only two transistors per input. In addition, the bias current IBIAS is set by a single global transistor that is common to all the inputs.

The largest of the input currents sets the common potential VC. As a result, the corresponding output carries almost all the bias current, while the other outputs have currents that are close to zero. Thus, the circuit selects the larger of the two input currents, i.e., if IIN,1 > IIN,2, we get IOUT,1 = IBIAS and IOUT,2 = 0. Similarly, if IIN,2 > IIN,1, we get IOUT,1 = 0 and IOUT,2 = IBIAS.

Simulation of the two-input CMOS winner-take-all circuit

A SPICE-based DC simulation of the CMOS winner-take-all circuit in the two-input case is shown on the right. As shown in the top subplot, the input IIN,1 was fixed at 6nA, while IIN,2 was linearly increased from 0 to 10nA. The bottom subplot shows the two output currents. As expected, the output corresponding to the larger of the two inputs carries the entire bias current (10nA in this case), forcing the other output current nearly to zero.

Other uses

In stereo matching algorithms, following the taxonomy proposed by Scharstein and Szelliski[8], winner-take-all is a local method for disparity computation. Adopting a winner-take-all strategy, the disparity associated with the minimum or maximum cost value is selected at each pixel.

It is axiomatic that in the electronic commerce market, early dominant players such as AOL or Yahoo! get most of the rewards. By 1998, one study[clarification needed] found the top 5% of all web sites garnered more than 74% of all traffic.

The winner-take-all hypothesis in economics suggests that once a technology or a firm gets ahead, it will do better and better over time, whereas lagging technology and firms will fall further behind. See First-mover adva hypothntage.

See also

References

  1. ^ Grossberg, Stephen (1982), "Contour Enhancement, Short Term Memory, and Constancies in Reverberating Neural Networks", Studies of Mind and Brain, vol. 70, Dordrecht: Springer Netherlands, pp. 332–378, doi:10.1007/978-94-009-7758-7_8, ISBN 978-90-277-1360-5, retrieved 2022-11-05
  2. ^ a b Oster, Matthias; Rodney, Douglas; Liu, Shih-Chii (2009). "Computation with Spikes in a Winner-Take-All Network". Neural computation. 21 (9): 2437–2465. doi:10.1162/neco.2009.07-08-829.
  3. ^ Riesenhuber, Maximilian; Poggio, Tomaso (1999-11-01). "Hierarchical models of object recognition in cortex". Nature Neuroscience. 2 (11): 1019–1025. doi:10.1038/14819. ISSN 1097-6256.
  4. ^ Carpenter, Gail A. (1987). "A massively parallel architecture for a self-organizing neural pattern recognition machine". Computer Vision, Graphics, and Image Processing. 37 (1): 54–115.
  5. ^ Itti, Laurent; Koch, Christof (1998). "A Model of Saliency-Based Visual Attention for Rapid Scene Analysis". IEEE Transactions on Pattern Analysis and Machine Intelligence. 20 (11): 1254–1259. doi:10.1109/34.730558.
  6. ^ Maass, Wolfgang (2000-11-01). "On the Computational Power of Winner-Take-All". Neural Computation. 12 (11): 2519–2535. doi:10.1162/089976600300014827. ISSN 0899-7667.
  7. ^ Lazzaro, J.; Ryckebusch, S.; Mahowald, M. A.; Mead, C. A. (1988-01-01). "Winner-Take-All Networks of O(N) Complexity:". Fort Belvoir, VA. doi:10.21236/ada451466. {{cite journal}}: Cite journal requires |journal= (help)
  8. ^ Scharstein, Daniel; Szeliski, Richard (2002). "A Taxonomy and Evaluation of Dense Two-Frame Stereo Correspondence Algorithms". International Journal of Computer Vision. 47 (1/3): 7–42. doi:10.1023/A:1014573219977.