This article has been reviewed according to Science X's editorial process and policies. Editors have highlighted the following attributes while ensuring the content's credibility:

fact-checked

peer-reviewed publication

trusted source

proofread

Transforming the future of media with artificial intelligence

Transforming the future of media with artificial intelligence
The Hourglass of Emotions, a comprehensive and multidimensional framework for interpreting emotions, is inspired by neuroscience and motivated by psychology. Positive emotions are at the top of the hourglass, while their corresponding negative emotions are at the bottom. Credit: NTU Singapore

With the ability to analyze large datasets to identify patterns and predict outcomes, all at the click of a button, artificial intelligence (AI) is revolutionizing how we live and work. From offering personalized recommendations to automating tedious tasks, AI can help us make better decisions, work smarter and reduce the likelihood of errors.

Chatbots powered by AI, such as ChatGPT, have transformed the media landscape. They can now have human-like conversations, generate content and analyze emotions from text—abilities once thought to be uniquely human.

Given the sheer amount of social media posts and information on the Internet, AI's ability to decode emotions from words could be a game-changer for applications such as in media monitoring and blocking malicious content.

Decoding emotions

However, AI is still not as effective as humans at recognizing emotions from text. Understanding emotional tones from written words involves understanding the world and social norms that humans learn through experience, which AI cannot do.

An AI platform, SenticNet, has been devised to address the challenges faced by AI when making sense of human languages. Developed by Prof Erik Cambria from NTU's School of Computer Science and Engineering (SCSE), SenticNet integrates human learning modes with traditional learning approaches that machines use to improve the algorithm's ability to analyze emotions.

SenticNet follows a logical process to infer the sentiments expressed in a sentence by categorizing word meanings in a framework resembling commonsense reasoning. Unlike conventional sentiment analysis models, which are often '' that do not provide any insights into their internal reasoning process, the processes by which SenticNet derives its results are transparent, and the results are reproducible and reliable.

"AI systems are becoming less and less transparent, and we hope SenticNet will be able to extract sentiments from text in an explainable manner without compromising performance," said Prof Cambria.

The researchers have demonstrated that combining commonsense reasoning with machine-learning approaches improves performance. When tested, SenticNet outperformed other machine-learning models.

Transforming the future of media with artificial intelligence
Illustration of how a long video is split into shorter segments of multiple scales for searching. Some clips contain ideal matches, while others contain non-ideal matches. The model derives answers from all segments with potential matches. Credit: NTU Singapore

The latest version of SenticNet was reported in the Proceedings of the 13th Conference on Language Resources and Evaluation in 2022.

Prof Cambria is also working on improving SenticNet's ability to encode and decode the meaning behind —a major challenge for AI systems as they do not possess the rich sensory experiences that humans have of the real world.

Searching videos

With moving visuals and sound, videos are an engaging way to convey messages and teach concepts. To allow users to better engage with for education and entertainment, a method developed by Assoc Prof Sun Aixin at SCSE makes video content searchable by matching keywords with on-screen images.

Conventional computer vision techniques can do this but are not as efficient when searching for images in long videos.

Assoc Prof Sun and his colleagues developed an algorithm that treats a video as a text passage so that people can search for specific moments in the clip. Using the method, a long video can thus be split into multiple shorter clips for searching.

"This simple and effective strategy enables images in long videos to be searched efficiently, addressing the issue of performance degradation commonly encountered by conventional computer vision techniques when searching long videos," said Assoc Prof Sun.

The findings were published in IEEE Transactions on Pattern Analysis and Machine Intelligence in 2022.

The researchers are currently working to enhance the algorithm's search accuracy and explore its usage on visual content in medical education and surveillance videos.

Transforming the future of media with artificial intelligence
Seq-DeepFake can detect fake images and recover original images from altered ones. Credit: NTU Singapore

Detecting fake images

Like any new technology, AI is a double-edged sword. Unfortunately, new threats, such as fake images designed to fool or scam audiences, have surfaced alongside the advancement of AI tools.

For instance, facial manipulation technologies can create photorealistic faces that may be used nefariously to mislead people.

Work fronted by Asst Prof Liu Ziwei at SCSE has resulted in an algorithm called Seq-DeepFake, which flags doctored images by recognizing digital fingerprints left by facial manipulation.

Unlike conventional deep fake detection methods that only predict whether images are real or fake, Seq-DeepFake detects the traces left behind by the manipulation sequentially. The alterations are detected within seconds.

The algorithm can also recover the original face from the manipulated face by reversing the manipulation sequence.

"Seq-DeepFake is a powerful tool that can potentially help everyone from to individual users verify the authenticity of visual information in the digital age to combat misinformation," said Asst Prof Liu.

In the future, Asst Prof Liu plans to expand the capabilities of Seq-DeepFake to detect other forms of doctored media such as text and videos.

The findings were published on the European Conference on Computer Vision in 2022.

Citation: Transforming the future of media with artificial intelligence (2024, February 9) retrieved 27 April 2024 from https://techxplore.com/news/2024-02-future-media-artificial-intelligence.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

Creating realistic 'talking heads' with an AI-powered program

4 shares

Feedback to editors