Artificial intelligence makes 'smart' apps faster, more efficient

Artificial intelligence makes 'smart' apps faster, more efficient
USask post-doctoral fellow Hao Zhang has developed a new artificial intelligence computer model that holds promise for making “smart” apps such as Siri safer, faster and more energy efficient. Credit: Dave Stobbe for the University of Saskatchewan

Tired of Siri or Google Assistant draining your phone battery?

A new University of Saskatchewan (USask) artificial intelligence holds promise for making "smart" apps such as Amazon, Apple, and Google's virtual assistants safer, faster and more energy efficient.

"Smart" services such as , weather forecasting, virtual assistants, and language translators rely on an artificial intelligence (AI) technology called "deep learning" to predict user patterns.

But these AI processes often require too much storage to be run locally on mobiles, so the data is sent to external servers over the Internet, which requires lots of power, drains the , and may increase a user's privacy risk.

"My method breaks down the AI computational processes in smaller 'chunks' and this helps run the 'smart' apps locally on the , rather than relying on external servers, while reducing ," said Hao Zhang, a USask electrical and computer engineering post-doctoral fellow.

"This research may lead to a different way to design apps and for our such as tablets, phones and computers."

Zhang ran accurate simulations to compare his AI model with those used on current phone systems, and found that his AI model can simultaneously run multiple apps more efficiently than devices currently on the market. His model performed 20 percent faster—twice as fast in some cases—and showed that batteries can last twice as long. The results are published in the journal IEEE Transactions on Computers.

Zhang found that AI processes can handle data efficiently using smaller four-bit sequences with variable length, so he built his model with these shorter "bit chunks." Current devices use a fixed 32-bit sequence to process data more accurately. As a result, phones or computers are not as fast and require more memory space to store the data.

"Large bit sequences are not always required to process data," said Seok-Bum Ko, a USask electrical and computer engineering professor and Zhang's supervisor. "Shorter sequences can be used to save power and increase speed performance, but can still guarantee enough accuracy for the app to function."

While the results are promising, Zhang and Ko are working to integrate their AI model with larger computer and phone systems, and to test how the model will function in real-world processors.

"If all goes well with our research, we may have our model integrated with apps and systems in three or four years," said Ko.

Zhang, whose research is funded by the federal agency NSERC, decided USask was the right place to pursue a Ph.D. degree to study deep learning. As a master's student at the City University of Hong Kong, he had a life-changing experience during his exchange internship at USask under Ko's supervision.

"The two universities have a good research collaboration and the internship experience here was great," he said. "We have good research equipment at USask that can support me to do a lot of experiments and to study many research topics, especially deep learning."

More information: Hao Zhang et al. New Flexible Multiple-Precision Multiply-Accumulate Unit for Deep Neural Network Training and Inference, IEEE Transactions on Computers (2019). DOI: 10.1109/TC.2019.2936192

Citation: Artificial intelligence makes 'smart' apps faster, more efficient (2020, November 13) retrieved 18 April 2024 from
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

Better apps might improve health, study finds


Feedback to editors