Robots deciding their next move need help prioritizing

Robots deciding their next move need help prioritizing
Adversarial and multi-agent environment where two teams play a game of Capture the Flag. New dynamics can be added to the environment to test adaptation. Credit: University of Illinois at Urbana-Champaign

As robots replace humans in dangerous situations such as search and rescue missions, they need to be able to quickly assess and make decisions—to react and adapt like a human being would. Researchers at the University of Illinois at Urbana-Champaign used a model based on the game Capture the Flag to develop a new take on deep reinforcement learning that helps robots evaluate their next move.

The team of researchers chose Capture the Flag because it's played with two teams, each with multiple teammates, where the opposing team is also making decisions.

"Robots can learn how to react in an environment like a competitive game by using a kind of trial and error process, called . They learn what actions to take in a given situation by playing the game," said Huy Tran, a researcher in the Department of Aerospace Engineering at UIUC. "The challenge is to figure out how to create agents that can also adapt to unexpected situations."

Tran said his team realized that the robots needed help in prioritizing tasks.

"Given the overall task of capturing the flag, there are actually sub tasks to accomplish the along the way, which we model in a . What we wanted to explore was whether or not this type of hierarchy would help with the ability to adapt."

With hierarchical , Tran said tasks are split up—to capture the flag or to tag a member of the opposing team to eliminate them—so the model can handle more complex problems.

"By breaking the task into sub tasks, we can improve adaptation. We trained a high-level decision maker who assigns a sub for each agent to focus on." Tran said. The hierarchical structure helps by making updates to the model simpler, Tran said. Only the hierarchical controller would need to be updated rather than each of the agents.

"This approach has the potential to solve interesting and challenging problems, but there are a lot of issues that we still need to address before we can deploy these systems in real-world situations. For example, we learned that this framework can help with adaptation," Tran said, "but we recognize that in this study we decided what the sub tasks should be based on our own intuition of how the game works. That is not ideal because it has our own biases. What we're doing now is looking at new techniques to allow agents to figure out what those sub goals should be on their own."

The study, "Evaluating Adaptation Performance of Hierarchical Deep Reinforcement Learning," was written by Neale Van Stralen, Seung Hyun Kim, Huy T. Tran, and Girish Chowdhary. The research was funded by the Defense Advanced Research Projects Agency and was presented at the 2020 IEEE International Conference on Robotics and Automation (ICRA) and published in the conference proceedings. A short video illustrates the work that includes the hierarchical controller in action.

More information: Neale Van Stolen et al. Evaluating Adaptation Performance of Hierarchical Deep Reinforcement Learning, 2020 IEEE International Conference on Robotics and Automation (ICRA) (2020). DOI: 10.1109/ICRA40945.2020.9197052

Citation: Robots deciding their next move need help prioritizing (2020, October 21) retrieved 22 February 2024 from
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

Researchers exploit weaknesses of master game bots


Feedback to editors