Semantic neural network

From Wikipedia, the free encyclopedia
Jump to: navigation, search

Semantic neural network (SNN) is based on John von Neumann's neural network [von Neumann, 1966] and Nikolai Amosov M-Network. There are limitations to a link topology for the von Neumann’s network but SNN accept a case without these limitations. Only logical values can be processed, but SNN accept that fuzzy values can be processed too. All neurons into the von Neumann network are synchronized by tacts. For further use of self-synchronizing circuit technique SNN accepts neurons can be self-running or synchronized.

In contrast to the von Neumann network there are no limitations for topology of neurons for semantic networks. It leads to the impossibility of relative addressing of neurons as it was done by von Neumann. In this case an absolute readdressing should be used. Every neuron should have a unique identifier that would provide a direct access to another neuron. Of course, neurons interacting by axons-dendrites should have each other's identifiers. An absolute readdressing can be modulated by using neuron specificity as it was realized for biological neural networks.

There’s no description for self-reflectiveness and self-modification abilities into the initial description of semantic networks [Dudar Z.V., Shuklin D.E., 2000]. But in [Shuklin D.E. 2004] a conclusion had been drawn about the necessity of introspection and self-modification abilities in the system. For maintenance of these abilities a concept of pointer to neuron is provided. Pointers represent virtual connections between neurons. In this model, bodies and signals transferring through the neurons connections represent a physical body, and virtual connections between neurons are representing an astral body. It is proposed to create models of artificial neuron networks on the basis of virtual machine supporting the opportunity for paranormal effects.

SNN is generally used for natural language processing.

References[edit]

  • Dudar Z.V., Shuklin D.E., 2000. Implementation of neurons for semantic neural nets that’s understanding texts in natural language. In Radio-electronika i informatika KhTURE, 2000. No 4. Р. 89-96.

  • Shuklin D.E. The Structure of a Semantic Neural Network Extracting the Meaning from a Text, In Cybernetics and Systems Analysis, Volume 37, Number 2, 4 March 2001, pp. 182–186(5) [1]
  • Shuklin D.E. The Structure of a Semantic Neural Network Realizing Morphological and Syntactic Analysis of a Text, In Cybernetics and Systems Analysis, Volume 37, Number 5, September 2001, pp. 770–776(7)
  • Shuklin D.E. Realization of a Binary Clocked Linear Tree and Its Use for Processing Texts in Natural Languages, In Cybernetics and Systems Analysis, Volume 38, Number 4, July 2002, pp. 503–508(6)