New approach for a biological programming language
New findings by researchers led by TU Graz computer scientists Wolfgang Maass and Robert Legenstein on neural information processing in the brain could enable more efficient AI methods.
Specifically, the researchers have succeeded in mathematically modeling the emergence and interaction between so-called “assemblies”. These are neuron groups in the brain that form the basis for higher cognitive abilities such as thinking, imagining, arguing, planning or speech processing.
Better understanding of how the brain works
The Canadian neuroscientist Donald H. Hebb postulated as early as 1949 that neurons form such groups, in other words that they act together to encode individual words or symbols, as well as holistic “concepts”. “However, the existence of assemblies has only become more consolidated in recent years, and our models are based on the latest results from brain research,” explains Maass.
Assemblies are flowing entities that constantly reorganize themselves to process environmental stimuli, give them a symbolic meaning, structure them and transform them into knowledge. This adaptability—also called plasticity—provides the brain with the ability to get over its limited processing capacity and form an “unlimited” number of patterns.
The results not only contribute to a better understanding of the brain, they could also lead to efficient new AI methods as they combine the advantages of two main approaches to AI research: the symbolic and the connectionist.
Symbolic vs. connectionist information processing
Algorithms in symbol systems are based on defined rules (if/then commands) and logical formulas, and are persuasive through their ability to abstract: i.e. the ability to generalize and to apply general contexts to concrete facts. For this reason they are optimally suited for easy application to completely new situations. However, symbol-based systems must be programmed in a complex way and cannot be trained for demanding applications by means of large amounts of data, as is possible with neural networks. The latter consist of small, networked and adaptive computing units that are self-organizing and can quickly solve complex problems when working together. The learning ability of neural networks has made the connectionist approach more attractive for current AI research and for modern AI applications. However, neural networks have difficulties with tasks that did not occur in their training set.
Human brain architecture for machines
The assembly models presented now aim to combine the ability to abstract with the ability to learn. “These are neural networks that work symbolically with their assemblies. The paradigm we use is the human brain, which also combines both,” said Legenstein.
Source: Read Full Article