Spreading activation

From Wikipedia, the free encyclopedia
Jump to: navigation, search

Spreading activation is a method for searching associative networks, neural networks, or semantic networks. The search process is initiated by labeling a set of source nodes (e.g. concepts in a semantic network) with weights or "activation" and then iteratively propagating or "spreading" that activation out to other nodes linked to the source nodes. Most often these "weights" are real values that decay as activation propagates through the network. When the weights are discrete this process is often referred to as marker passing. Activation may originate from alternate paths, identified by distinct markers, and terminate when two alternate paths reach the same node. However brain studies show that several different brain areas play an important role in semantic processing, [1]

Spreading activation models are used in cognitive psychology[2][3] to model the fan out effect.

Spreading activation can also be applied in information retrieval,[4][5] by means of a network of nodes representing documents and terms contained in those documents.

Algorithm[edit]

A directed graph is populated by Nodes[ 1...N ] each having an associated activation value A [ i ] which is a real number in the range [ 0.0 ... 1.0]. A Link[ i, j ] connects source node[ i ] with target node[ j ]. Each link has an associated weight W [ i, j ] usually a real number in the range [0.0 ... 1.0].[6]

Parameters:

  • Firing threshold F, a real number in the range [0.0 ... 1.0]
  • Decay factor D, a real number in the range [0.0 ... 1.0]

Steps:

  1. Initialize the graph setting all activation values A [ i ] to zero. Set one or more origin nodes to an initial activation value greater than the firing threshold F. A typical initial value is 1.0.
  2. For each unfired node [ i ] in the graph having an activation value A [ i ] greater than the node firing threshold F:
  3. For each Link [ i, j ] connecting the source node [ i ] with target node [ j ], adjust A [ j ] = A [ j ] + (A [ i ] * W [ i, j ] * D) where D is the decay factor.
  4. If a target node receives an adjustment to its activation value so that it would exceed 1.0, then set its new activation value to 1.0. Likewise maintain 0.0 as a lower bound on the target node's activation value should it receive an adjustment to below 0.0.
  5. Once a node has fired it may not fire again, although variations of the basic algorithm permit repeated firings and loops through the graph.
  6. Nodes receiving a new activation value that exceeds the firing threshold F are marked for firing on the next spreading activation cycle.
  7. If activation originates from more than one node, a variation of the algorithm permits marker passing to distinguish the paths by which activation is spread over the graph
  8. The procedure terminates when either there are no more nodes to fire or in the case of marker passing from multiple origins, when a node is reached from more than one path. Variations of the algorithm that permit repeated node firings and activation loops in the graph, terminate after a steady activation state, with respect to some delta, is reached, or when a maximum number of iterations is exceeded.

Examples[edit]

In this example, spreading activation originated at node 1 which has an initial activation value of 1.0 (100%). Each link has the same weight value of 0.9. The decay factor was 0.85. Four cycles of spreading activation have occurred. Color hue and saturation indicate different activation values.

See also[edit]

Notes[edit]

  1. ^ Karalyn Patterson, Peter J. Nestor & Timothy T. Rogers: "Where do you know what you know? The representation of semantic knowledge in the human brain" [1]
  2. ^ Collins, Allan M.; Loftus, Elizabeth F., "A spreading-activation theory of semantic processing", Psychological Review. 1975 Nov Vol 82(6) 407-428 [2]
  3. ^ John R. Anderson. "A spreading activation theory of memory." Journal of Verbal Learning and Verbal Behavior, 1983
  4. ^ S. Preece, A spreading activation network model for information retrieval. PhD thesis, University of Illinois, Urbana-Champaign, 1981.
  5. ^ Fabio Crestani. "Application of Spreading Activation Techniques in Information Retrieval". Artificial Intelligence Review, 1997
  6. ^ Boosting item keyword search with spreading activation Aswath, D.; Ahmed, S.T.; Dapos;cunha, J.; Davulcu, H., Web Intelligence, 2005. Proceedings. The 2005 IEEE/WIC/ACM International Conference on Volume , Issue , 19-22 Sept. 2005 Page(s): 704 - 707

References[edit]

  • Nils J. Nilsson. "Artificial Intelligence: A New Synthesis". Morgan Kaufmann Publishers, Inc., San Francisco, California, 1998, pages 121-122
  • Rodriguez, M.A., " Grammar-Based Random Walkers in Semantic Networks", Knowledge-Based Systems, 21(7), 727-739, doi:10.1016/j.knosys.2008.03.030, 2008.
  • Karalyn Patterson, Peter J. Nestor & Timothy T. Rogers "Where do you know what you know? The representation of semantic knowledge in the human brain", Nature Reviews Neuroscience 8, 976-987 (December 2007)

External links[edit]

  • JMaPSS The Java Marker-Passing Search Service, a relevance search engine employing a family of marker-passing algorithms based on spreading activation theory.
  • Texai An open source project to create artificial intelligence that provides a Java spreading activation class library.