This article has been reviewed according to Science X's editorial process and policies. Editors have highlighted the following attributes while ensuring the content's credibility:

fact-checked

peer-reviewed publication

trusted source

proofread

When AI seems to disclose personal information, users may empathize more

When A.I. discloses personal information, users may empathize more
An example of agents interacting with humans in the study. Credit: Takahiro Tsumura, CC-BY 4.0 (https://creativecommons.org/licenses/by/4.0/)

In a new study, participants showed more empathy for an online anthropomorphic artificial intelligence (AI) agent when it seemed to disclose personal information about itself while chatting with the participants. Takahiro Tsumura of The Graduate University for Advanced Studies, SOKENDAI in Tokyo, Japan, and Seiji Yamada of the National Institute of Informatics, also in Tokyo, present these findings in the open-access journal PLOS ONE on May 10, 2023.

The use of AI in daily life is increasing, raising interest in factors that might contribute to the level of trust and acceptance people feel towards AI agents. Prior research has suggested that people are more likely to accept artificial objects if the objects elicit . For instance, people may empathize with cleaning robots, robots that mimic pets, and anthropomorphic chat tools that provide assistance on websites.

Earlier research has also highlighted the importance of disclosing in building . Stemming from those findings, Tsumura and Yamada hypothesized that self-disclosure by an anthropomorphic AI agent might boost people's empathy toward those agents.

To test this idea, the researchers conducted online experiments in which participants had a text-based chat with an online AI agent that was visually represented by either a human-like illustration or an illustration of an anthropomorphic robot. The chat involved a scenario in which the participant and agent were colleagues on a lunch break at the agent's workplace. In each conversation, the agent seemed to self-disclose either highly work-relevant personal information, less-relevant information about a hobby, or no personal information.

The final analysis included data from 918 participants whose empathy for the AI agent was evaluated using a standard empathy questionnaire. The researchers found that, compared to less-relevant self-disclosure, highly work-relevant self-disclosure from the AI agent was associated with greater empathy from participants. A lack of self-disclosure was associated with suppressed empathy. The agent's appearance as either a human or anthropomorphic robot did not have a significant association with empathy levels.

These findings suggest that self-disclosure by AI agents may, indeed, elicit empathy from humans, which could help inform future development of AI tools.

The authors add, "This study investigates whether self-disclosure by anthropomorphic agents affects human empathy. Our research will change the negative image of artifacts used in society and contribute to future social relationships between humans and anthropomorphic agents."

More information: Influence of agent's self-disclosure on human empathy, PLoS ONE (2023). DOI: 10.1371/journal.pone.0283955

Journal information: PLoS ONE
Citation: When AI seems to disclose personal information, users may empathize more (2023, May 10) retrieved 19 March 2024 from https://techxplore.com/news/2023-05-ai-disclose-personal-users-empathize.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

Empathy can be taught at school—and it can lead to more creative thinking

71 shares

Feedback to editors