Personality-based approach may be key to partnering with robots

Personality-based approach may be key to partnering with robots
Soldiers' facial expressions could help inform more effective teaming with robots. Credit: U.S. Army photo

Human facial expressions could be one of the keys in building trust between Soldiers and autonomous agents.

"We wanted to characterize and quantify factors that impact the that humans have with in automated driving," said Dr. Catherine Neubauer, a researcher at the U.S. Army Combat Capabilities Development Command's Army Research Laboratory. "With this information, we want to develop a robust way to predict decision errors in use to eventually enable active, online mitigation strategies and effective calibration techniques when humans and agents are teaming in real-time."

In typical studies, researchers define groups of people, and either specify and examine variables within those groups or apply variables as measures that tend to represent entire population samples by simple group tendencies. In these traditional constructs, researchers treat sample variability within the experimenter-specified groups as noise.

The Army researchers proposed a new approach, using flexible mixture modeling, which inferred and sorted individuals into four sub-groups by observed differences in traits (e.g., age, personality and maladaptive coping strategies associated with uncertainty) and states (e.g., initial trust, stress and workload perceptions). From these clustered groups, researchers analyzed facial recognition-based measures of emotional expression and subjective self-report measures of trust under different experimental conditions of automation levels and reliability—supporting validity of this clustering approach to identify key factors related to over-trust, under-trust, appropriately calibrated trust and distrust.

"We believe this approach extends the state-of-the-art by explicitly evaluating as a way to quantify and calibrate affect-based trust in response to automation level capability and reliability," Neubauer said. "It could also provide a method to understand the continuous variations in trust during a human-agent interaction, as opposed to the standard approach of participants self-reporting changes in trust after an interaction has occurred."

Personality-based approach may be key to partnering with robots
Army researchers used automated driving simulations to study human emotions and trust. Credit: U.S. Army image

This approach could provide more information than traditional designs, as it can give insight into how manipulations in experiments affect different subgroups of participants.

"It is often stated that for appropriate trust to be developed and effectively calibrated, an individual's expectations must match the system's actual behaviors," Neubauer said. "This research shows that calibration metrics will vary across people; certain groups may be more prone to over-trust automation, while others may mistrust the automation from the start. From a research and development perspective, this approach provides a way to tailor human automation interaction modalities to individuals without needing to define and validate specific models for each person."

The research team published their findings, Analysis of Facial Expressions: Explaining Affective State and Trust-Based Decisions during Interaction with Automation, and provided the following recommendations for researchers when identifying and assessing to calibrate trust in human-autonomy teams:

  1. Researchers should not expect the same requirements for interventions for all individuals
  2. Researchers can use multivariate methodologies focused on subgroups to identify clusters of individuals, which may inform expectations of varying responses and therefore directly influence selection of intervention strategies
  3. Though often neglected in this literature base, researchers can use estimates of emotional expression (i.e., using ) for additional insights to improve trust calibration strategies

"Results showed that the group of individuals who were prone to over-trust also expressed strong outward changes in facial expression and were more likely to need interventions that directed their expectations and encouraged appropriate take-overs," Neubauer said. "Whereas a group with an inherent bias against automation had limited facial expression and may even be more negatively influenced by a similar type of intervention. Online mitigation strategies are unlikely to be robustly effective without accounting for this kind of inter-individual variability using these kinds of methods."

Researchers at the lab will use these methods in future human-autonomy teaming studies to improve prediction of Soldier responses to automation, specifically using modalities such as facial expression analysis to help cue the critical moments in which trust-based interventions are needed. They will also study how group-based interventions can improve trust and team cohesion within human-agent team contexts, e.g., when Soldiers must jointly perform high-consequence tasks with automated agents, such as the Next Generation Combat Vehicle.

Citation: Personality-based approach may be key to partnering with robots (2020, August 6) retrieved 25 March 2023 from
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

It's easier to trust automated vehicles when we know what they plan to do ahead of time


Feedback to editors