This article has been reviewed according to Science X's editorial process and policies. Editors have highlighted the following attributes while ensuring the content's credibility:


trusted source


Neural network trained using a diverse dataset outperforms conventionally trained algorithms

Neural network trained using a diverse dataset outperforms conventionally trained algorithms
An image from a low-resource population used in the dataset: A stove in Vietnam from a household that earns $237 USD/month. Credit: Victrixia Montes for Dollar Street 2015

Artificially intelligent neural networks, trained by images and videos available on the internet, can recognize faces, objects, and more. But there's a serious drawback. Teaching machine learning algorithms how to identify people or items by relying solely on the visual library of faces and objects found online underrepresents socioeconomic and demographic groups.

A Harvard University machine learning researcher and collaborators from MLCommons and Coactive AI created a more diverse dataset using pictures of objects found in households around the world and trained a to sort objects based on that dataset. Their findings—presented at the Conference on Neural Information Processing Systems—reveal the use of images from low-resource populations can dramatically boost the object recognition performance of machine learning systems.

"There hasn't yet been a strong incentive for equity and equal representation to be built into machine learning systems," says Vijay Janapa Reddi, associate professor at Harvard's John A. Paulson School of Engineering and Applied Sciences (SEAS) and a senior author of the paper. "That's the big picture we're trying to capture with this research."

Reddi, who is also a vice president and board member at ML Commons, a consortium of academic and industry AI leaders, teamed up with colleagues to train a neural network using a dataset of 38,479 images of household objects. The collection of photographs taken in 404 homes across 63 countries in Africa, America, Asia, and Europe is known as "Dollar Street," and was first developed by the Gapminder Foundation. The Swedish-based entity sent photographers around the world to amass images of toothbrushes, toilets, TVs, stoves, beds, lamps, and other objects found in the homes of families with monthly incomes between the U.S. equivalent of $26.99 and $19,671.

"We need to be cognizant of deeper biases in our machine learning systems," Reddi says. "The same word might be given to describe stoves around the world, but if you look at what is called a stove in underrepresented areas versus what's found in wealthy homes, those objects can look and function completely differently."

In their paper, the researchers describe another striking example: in some poor homes around the world, a person might use their hand to brush their teeth. In the Dollar Street dataset, then, a picture of someone's hand might be labeled as both "hand palm" and "toothbrush."

Neural network trained using a diverse dataset outperforms conventionally trained algorithms
A stove in Burundi from a household earning $37 USD/month. Credit: Johan Eriksson for Dollar Street 2015

Using the Dollar Street image collection—which was developed by MLCommons into a robust containing object names/tags, geographic data, and household monthly income—the team found that their trained neural network performed drastically better than leading-edge systems at accurately classifying household items, especially objects found in homes with lower incomes. Their machine learning algorithm correctly identified objects 65% more frequently compared to commonly used neural networks—including ImageNet and Open Images—trained on less diverse datasets sourced from the internet.

"It's shocking to see what state-of-the-art machine learning models take for granted and how poorly they perform at correctly identifying objects from lower-resource settings," Reddi says.

As industry and government rely increasingly on machine learning systems to process information and make decisions, Reddi says this proof-of-concept research demonstrates the danger of neural networks trained without inclusive data representing low-resource populations.

"Dollar Street has been a powerful tool for combating human misconceptions and , and we believe it has the potential to do the same for machines," says Cody Coleman, co-senior author of the paper, who is CEO and co-founder of Coactive AI.

"Dollar Street demonstrates the importance of data in machine learning in a general sense, and specifically the ability of carefully selected data to have an outsized impact on bias," says David Kanter, a co-author on the paper, who is founder and executive director of MLCommons. "My hope is that by hosting and maintaining Dollar Street, we will empower the and industry to develop techniques so that machine learning benefits everyone across the globe, particularly in less developed regions."

"Artificially intelligent systems, if not built equitably and inclusively, will accelerate the divide between the high-resource communities and low-resource ones," Reddi says. "When you're building datasets to train machine learning systems, and you're building that data from a high-resource place and not going out of your way to acquire and include data from lower-resource areas, the implications for learned bias become even bigger. Responsible AI means making machine learning globally accessible, and globally representative."

More information: The Dollar Street Dataset: Images Representing the Geographic and Socioeconomic Diversity of the World.

Provided by Harvard University
Citation: Neural network trained using a diverse dataset outperforms conventionally trained algorithms (2023, February 8) retrieved 24 May 2024 from
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

Can machine-learning models overcome biased datasets?


Feedback to editors