This article has been reviewed according to Science X's editorial process and policies. Editors have highlighted the following attributes while ensuring the content's credibility:

fact-checked

peer-reviewed publication

trusted source

proofread

An 'introspective' AI finds diversity improves performance

An 'Introspective' AI finds diversity improves performance
Progression from conventional artificial neural network to diverse neural network to learned diverse neural network. Line thicknesses represent weights W, circle thicknesses represent biases b, and sketches inside circles represent activation functions σ. Credit: Scientific Reports (2023). DOI: 10.1038/s41598-023-40766-6

An artificial intelligence with the ability to look inward and fine tune its own neural network performs better when it chooses diversity over lack of diversity, a new study finds. The resulting diverse neural networks were particularly effective at solving complex tasks.

"We created a with a non-human intelligence, an (AI), to see if the AI would choose diversity over the lack of diversity and if its choice would improve the performance of the AI," says William Ditto, professor of physics at North Carolina State University, director of NC State's Nonlinear Artificial Intelligence Laboratory (NAIL) and co-corresponding author of the work. "The key was giving the AI the ability to look inward and learn how it learns."

Neural networks are an advanced type of AI loosely based on the way that our brains work. Our natural neurons exchange according to the strengths of their connections. Artificial neural networks create similarly by adjusting numerical weights and biases during training sessions.

For example, a can be trained to identify photos of dogs by sifting through a large number of photos, making a guess about whether the photo is of a dog, seeing how far off it is and then adjusting its weights and biases until they are closer to reality.

Conventional AI uses to solve problems, but these networks are typically composed of large numbers of identical artificial neurons. The number and strength of connections between those identical neurons may change as it learns, but once the network is optimized, those static neurons are the network.

Ditto's team, on the other hand, gave its AI the ability to choose the number, shape and connection strength between neurons in its neural network, creating sub-networks of different neuron types and connection strengths within the network as it learns.

"Our real brains have more than one type of neuron," Ditto says. "So we gave our AI the ability to look inward and decide whether it needed to modify the composition of its neural network. Essentially, we gave it the control knob for its own brain. So it can solve the problem, look at the result, and change the type and mixture of artificial neurons until it finds the most advantageous one. It's meta-learning for AI.

"Our AI could also decide between diverse or homogenous neurons," Ditto says. "And we found that in every instance the AI chose diversity as a way to strengthen its performance."

The team tested the AI's accuracy by asking it to perform a standard numerical classifying exercise, and saw that its accuracy increased as the number of neurons and neuronal diversity increased. A standard, homogenous AI could identify the numbers with 57% accuracy, while the meta-learning, diverse AI was able to reach 70% accuracy.

According to Ditto, the diversity-based AI is up to 10 times more accurate than conventional AI in solving more complicated problems, such as predicting a pendulum's swing or the motion of galaxies.

"We have shown that if you give an AI the ability to look inward and learn how it learns it will change its internal structure—the structure of its artificial neurons—to embrace diversity and improve its ability to learn and solve problems efficiently and more accurately," Ditto says. "Indeed, we also observed that as the problems become more complex and chaotic the performance improves even more dramatically over an AI that does not embrace ."

The work is published in the journal Scientific Reports.

More information: Choudhary, A. et al. Neuronal diversity can improve machine learning for physics and beyond. Scientific Reports (2023). DOI: 10.1038/s41598-023-40766-6 www.nature.com/articles/s41598-023-40766-6

Journal information: Scientific Reports
Citation: An 'introspective' AI finds diversity improves performance (2023, August 31) retrieved 27 April 2024 from https://techxplore.com/news/2023-08-introspective-ai-diversity.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

Teaching physics to neural networks removes 'chaos blindness'

122 shares

Feedback to editors