This article has been reviewed according to Science X's editorial process and policies. Editors have highlighted the following attributes while ensuring the content's credibility:


trusted source


Training robotic arms with a hands-off approach

Training robotic arms with a hands-off approach
Credit: Carnegie Mellon University Mechanical Engineering

Humans often take their fine motor abilities for granted. Recreating the mechanical precision of the human body is no easy task—one that graduate students in CMU's Mechanical Engineering Department hope to simplify through artificial intelligence.

Ph.D. candidates Alison Bartch and Abraham George, under the guidance of Amir Barati Farimani, assistant professor of mechanical engineering, first recreated a simple , such as picking up a block, using a . Using this example, they then were able to augment different "-like" examples of the movements to aid the 's learning.

"If I want to show you how to do a task, I just have to do it once or twice before you pick up on it," George said. "So it's very promising that now we can get a robot to replicate our actions after just one or two demos. We have created a control structure where it can watch us, extract what it needs to know, and then perform that action."

The team found that the augmented examples helped to significantly decrease the robot's learning time for the block pick-and-place task compared to a machine learning architecture alone. This, paired with the collection of human data through a VR headset simulation, means this research method has the potential to produce promising results with "under a minute of human input."

This research was presented at the 2023 International Conference on Robotics and Automation.

Credit: Carnegie Mellon University Mechanical Engineering

George described the challenge of creating augmented examples that were both reliable and "novel" for the AI to learn from, so that it could recognize more nuanced differences in the same movements.

"A good analogy for this is when you are trying to train a computer to recognize a picture of a dog, you might show it hundreds of images of dogs versus cats." George said. "But we are trying to train the computer to identify a dog based on just one augmented image. So if we show it a different breed of dog, the computer is going to struggle to identify it as a dog."

Already looking forward, Bartch plans to use similar methods to teach the robot how to interact with more malleable material—particularly clay—and predict how it will shape them.

"For the end goal of having robots in the world, they need to be able to predict how different materials are going to behave," Bartch said. "If you think about an assistant robot at home, the materials they will interact with are deformable: food is deformable, sponges are deformable, clothes are deformable."

More information: Abraham George et al, Minimizing Human Assistance: Augmenting a Single Demonstration for Deep Reinforcement Learning (2023).

Citation: Training robotic arms with a hands-off approach (2023, August 14) retrieved 6 December 2023 from
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

New technique helps user understand why a robot failed, then fine-tune it to perform task


Feedback to editors