'Deepfake challenge' aims to find tools to fight manipulation

Deepfake videos can make it appear that people are doing or saying fictional things in an effort to spread misinformation
Deepfake videos can make it appear that people are doing or saying fictional things in an effort to spread misinformation

Technology firms and academics have joined together to launch a "deepfake challenge" to improve tools to detect videos and other media manipulated by artificial intelligence.

The initiative announced Thursday includes $10 million from Facebook and aims to curb what is seen as a major threat to the integrity of online information.

The effort is being supported by Microsoft and the industry-backed Partnership on AI and includes academics from the Massachusetts Institute of Technology, Cornell University, University of Oxford, University of California-Berkeley, University of Maryland and University at Albany.

It represents a broad effort to combat the dissemination of manipulated or audio as part of a misinformation campaign.

"The goal of the challenge is to produce technology that everyone can use to better detect when AI has been used to alter a video in order to mislead the viewer," said Facebook chief technical officer Mike Schroepfer.

Schroepfer said deepfake techniques, which present realistic AI-generated videos of people doing and saying fictional things, "have significant implications for determining the legitimacy of information presented online. Yet the industry doesn't have a great data set or benchmark for detecting them."

The challenge is the first project of a committee on AI and integrity created by the Partnership on AI, a group whose mission is to promote beneficial uses of and is backed by Apple, Amazon, IBM and other tech firms and non-governmental organizations.

Terah Lyons, executive director of the Partnership, said the new project is part of an effort to stem AI-generated fakes, which "have significant, global implications for the legitimacy of information online, the quality of public discourse, the safeguarding of human rights and , and the health of democratic institutions."

Facebook said it was offering funds for research collaborations and prizes for the , and would also enter the competition, but not accept any of the .

Oxford professor Philip Torr, one of the academics participating, said new tools are "urgently needed to detect these types of manipulated media.

"Manipulated media being put out on the internet, to create bogus conspiracy theories and to manipulate people for political gain, is becoming an issue of global importance, as it is a fundamental threat to democracy," Torr said in a statement.


Explore further

This deep neural network fights deepfakes

© 2019 AFP

Citation: 'Deepfake challenge' aims to find tools to fight manipulation (2019, September 6) retrieved 21 September 2019 from https://techxplore.com/news/2019-09-deepfake-aims-tools.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.
189 shares

Feedback to editors

User comments

Please sign in to add a comment. Registration is free, and takes less than a minute. Read more