To install click the Add extension button. That's it.

The source code for the WIKI 2 extension is being checked by specialists of the Mozilla Foundation, Google, and Apple. You could also do it yourself at any point in time.

4,5
Kelly Slayton
Congratulations on this excellent venture… what a great idea!
Alexander Grigorievskiy
I use WIKI 2 every day and almost forgot how the original Wikipedia looks like.
Live Statistics
English Articles
Improved in 24 Hours
Added in 24 Hours
Languages
Recent
Show all languages
What we do. Every page goes through several hundred of perfecting techniques; in live mode. Quite the same Wikipedia. Just better.
.
Leo
Newton
Brights
Milds

Competitive learning

From Wikipedia, the free encyclopedia

Competitive learning is a form of unsupervised learning in artificial neural networks, in which nodes compete for the right to respond to a subset of the input data.[1][2] A variant of Hebbian learning, competitive learning works by increasing the specialization of each node in the network. It is well suited to finding clusters within data.

Models and algorithms based on the principle of competitive learning include vector quantization and self-organizing maps (Kohonen maps).

YouTube Encyclopedic

  • 1/3
    Views:
    17 979
    15 285
    1 447
  • COMPETITIVE LEARNING
  • Learning Rules | Competitive Learning | Basic Concepts | Neural Networks
  • Competitive learning

Transcription

Principles

There are three basic elements to a competitive learning rule:[3][4]

  • A set of neurons that are all the same except for some randomly distributed synaptic weights, and which therefore respond differently to a given set of input patterns
  • A limit imposed on the "strength" of each neuron
  • A mechanism that permits the neurons to compete for the right to respond to a given subset of inputs, such that only one output neuron (or only one neuron per group), is active (i.e. "on") at a time. The neuron that wins the competition is called a "winner-take-all" neuron.

Accordingly, the individual neurons of the network learn to specialize on ensembles of similar patterns and in so doing become 'feature detectors' for different classes of input patterns.

The fact that competitive networks recode sets of correlated inputs to one of a few output neurons essentially removes the redundancy in representation which is an essential part of processing in biological sensory systems.[5][6]

Architecture and implementation

Competitive neural network architecture

Competitive Learning is usually implemented with Neural Networks that contain a hidden layer which is commonly known as “competitive layer”.[7] Every competitive neuron is described by a vector of weights and calculates the similarity measure between the input data and the weight vector .

For every input vector, the competitive neurons “compete” with each other to see which one of them is the most similar to that particular input vector. The winner neuron m sets its output and all the other competitive neurons set their output .

Usually, in order to measure similarity the inverse of the Euclidean distance is used: between the input vector and the weight vector .

Example algorithm

Here is a simple competitive learning algorithm to find three clusters within some input data.

1. (Set-up.) Let a set of sensors all feed into three different nodes, so that every node is connected to every sensor. Let the weights that each node gives to its sensors be set randomly between 0.0 and 1.0. Let the output of each node be the sum of all its sensors, each sensor's signal strength being multiplied by its weight.

2. When the net is shown an input, the node with the highest output is deemed the winner. The input is classified as being within the cluster corresponding to that node.

3. The winner updates each of its weights, moving weight from the connections that gave it weaker signals to the connections that gave it stronger signals.

Thus, as more data are received, each node converges on the centre of the cluster that it has come to represent and activates more strongly for inputs in this cluster and more weakly for inputs in other clusters.

See also

References

  1. ^ Rumelhart, David; David Zipser; James L. McClelland; et al. (1986). Parallel Distributed Processing, Vol. 1. MIT Press. pp. 151–193.
  2. ^ Grossberg, Stephen (1987-01-01). "Competitive learning: From interactive activation to adaptive resonance" (PDF). Cognitive Science. 11 (1): 23–63. doi:10.1016/S0364-0213(87)80025-3. ISSN 0364-0213.
  3. ^ Rumelhart, David E., and David Zipser. "Feature discovery by competitive learning." Cognitive science 9.1 (1985): 75-112.
  4. ^ Haykin, Simon, "Neural Network. A comprehensive foundation." Neural Networks 2.2004 (2004).
  5. ^ Barlow, Horace B. "Unsupervised learning." Neural computation 1.3 (1989): 295-311.
  6. ^ Edmund T.. Rolls, and Gustavo Deco. Computational neuroscience of vision. Oxford: Oxford university press, 2002.
  7. ^ Salatas, John (24 August 2011). "Implementation of Competitive Learning Networks for WEKA". ICT Research Blog. Retrieved 28 January 2012.

Further information and software

This page was last edited on 10 January 2024, at 17:05
Basis of this page is in Wikipedia. Text is available under the CC BY-SA 3.0 Unported License. Non-text media are available under their specified licenses. Wikipedia® is a registered trademark of the Wikimedia Foundation, Inc. WIKI 2 is an independent company and has no affiliation with Wikimedia Foundation.