New learning algorithm should significantly expand the possible applications of AI

New learning algorithm should significantly expand the possible applications of AI
TU Graz computer scientists Robert Legenstein and Wolfgang Maass (from left) are working on energy-efficient AI systems and are inspired by the functioning of the human brain. Credit: Lunghammer—TU Graz

The high energy consumption of artificial neural networks' learning activities is one of the biggest hurdles for the broad use of Artificial Intelligence (AI), especially in mobile applications. One approach to solving this problem can be gleaned from knowledge about the human brain.

Although it has the computing power of a supercomputer, it only needs 20 watts, which is only a millionth of the of a supercomputer.

One of the reasons for this is the efficient transfer of information between in the brain. Neurons send short electrical impulses (spikes) to other neurons—but, to save energy, only as often as absolutely necessary.

Event-based information processing

A working group led by two computer scientists Wolfgang Maass and Robert Legenstein of TU Graz has adopted this principle in the development of the new machine learning algorithm e-prop (short for e-propagation).

Researchers at the Institute of Theoretical Computer Science, which is also part of the European lighthouse project Human Brain Project, use spikes in their model for communication between neurons in an artificial neural network.

The spikes only become active when they are needed for information processing in the network. Learning is a particular challenge for such less active networks, since it takes longer observations to determine which neuron connections improve network performance.

Previous methods achieved too little learning success or required enormous storage space. E-prop now solves this problem by means of a decentralized method copied from the , in which each neuron documents when its connections were used in a so-called e-trace (eligibility trace). The method is roughly as powerful as the best and most elaborate other known learning methods. Details have now been published in the scientific journal Nature Communications.

Online instead of offline

With many of the machine learning techniques currently in use, all network activities are stored centrally and offline in order to trace every few steps how the connections were used during the calculations.

However, this requires a constant data transfer between the memory and the processors—one of the main reasons for the excessive energy consumption of current AI implementations. e-prop, on the other hand, works completely online and does not require separate memory even in real operation—thus making learning much more energy efficient.

Driving force for neuromorphic hardware

Maass and Legenstein hope that e-prop will drive the development of a new generation of mobile learning computing systems that no longer need to be programmed but learn according to the model of the and thus adapt to constantly changing requirements.

The goal is to no longer have these computing systems learn energy-intensively exclusively via a cloud, but to efficiently integrate the greater part of the learning ability into mobile hardware components and thus save energy.

First steps to bring e-prop into the application have already been made. For example, the TU Graz team is working together with the Advanced Processor Technologies Research Group (APT) of the University of Manchester in the Human Brain Project to integrate e-prop into the neuromorphic SpiNNaker system, which has been developed there. At the same time, TU Graz is working with researchers from the semiconductor manufacturer Intel to integrate the algorithm into the next version of Intel's neuromorphic chip Loihi.

This research work is anchored in the Fields of Expertise "Human and biotechnology" and "Information, Communication & Computing", two of the five Fields of Expertise of TU Graz.


Explore further

New approach for a biological programming language

More information: A solution to the learning dilemma for recurrent networks of spiking neurons, Nature Communications (2020). DOI: 10.1038/s41467-020-17236-y
Journal information: Nature Communications

Citation: New learning algorithm should significantly expand the possible applications of AI (2020, July 17) retrieved 26 October 2020 from https://techxplore.com/news/2020-07-algorithm-significantly-applications-ai.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.
392 shares

Feedback to editors

User comments