Mind-controlled robots now one step closer

robot
Credit: CC0 Public Domain

Tetraplegic patients are prisoners of their own bodies, unable to speak or perform the slightest movement. Researchers have been working for years to develop systems that can help these patients carry out some tasks on their own. "People with a spinal cord injury often experience permanent neurological deficits and severe motor disabilities that prevent them from performing even the simplest tasks, such as grasping an object," says Prof. Aude Billard, the head of EPFL's Learning Algorithms and Systems Laboratory. "Assistance from robots could help these people recover some of their lost dexterity, since the robot can execute tasks in their place." 

Prof. Billard carried out a study with Prof. José del R. Millán, who at the time was the head of EPFL's Brain-Machine Interface laboratory but has since moved to the University of Texas. The two research groups have developed a that can control a robot using emitted by a patient's brain. No or touch function is needed; patients can move the robot simply with their thoughts. The study has been published in Communications Biology

Avoiding obstacles

To develop their system, the researchers started with a that had been developed several years ago. This arm can move back and forth from right to left, reposition objects in front of it and get around objects in its path. "In our study we programmed a robot to avoid obstacles, but we could have selected any other kind of task, like filling a glass of water or pushing or pulling an object," says Prof. Billard. 

The engineers began by improving the robot's mechanism for avoiding obstacles so that it would be more precise. "At first, the robot would choose a path that was too wide for some obstacles, taking it too far away, and not wide enough for others, keeping it too close," says Carolina Gaspar Pinto Ramos Correia, a Ph.D. student at Prof. Billard's lab. "Since the goal of our robot was to help paralyzed patients, we had to find a way for users to be able to communicate with it that didn't require speaking or moving." 

An algorithm that can learn from thoughts 

This entailed developing an algorithm that could adjust the robot's movements based only on a patient's thoughts. The algorithm was connected to a headcap equipped with electrodes for running electroencephalogram (EEG) scans of a patient's . To use the system, all the patient needs to do is look at the robot. If the robot makes an incorrect move, the patient's brain will emit an "error message" through a clearly identifiable signal, as if the patient is saying "No, not like that." The robot will then understand that what it's doing is wrong—but at first it won't know exactly why. For instance, did it get too close to, or too far away from, the object? To help the robot find the right answer, the is fed into the algorithm, which uses an inverse reinforcement learning approach to work out what the patient wants and what actions the robot needs to take. This is done through a trial-and-error process whereby the robot tries out different movements to see which one is correct. The process goes pretty quickly—only three to five attempts are usually needed for the robot to figure out the right response and execute the patient's wishes. "The robot's AI program can learn rapidly, but you have to tell it when it makes a mistake so that it can correct its behavior," says Prof. Millán. "Developing the detection technology for error signals was one of the biggest technical challenges we faced." Iason Batzianoulis, the study's lead author, adds: "What was particularly difficult in our study was linking a patient's brain activity to the robot's control system—or in other words, 'translating' a patient's brain signals into actions performed by the robot. We did that by using machine learning to link a given brain signal to a specific task. Then we associated the tasks with individual robot controls so that the robot does what the patient has in mind."

Next step: A mind-controlled wheelchair

The researchers hope to eventually use their algorithm to control wheelchairs. "For now there are still a lot of engineering hurdles to overcome," says Prof. Billard. "And wheelchairs pose an entirely new set of challenges, since both the patient and the robot are in motion." The team also plans to use their algorithm with a that can read several different kinds of signals and coordinate data received from the brain with those from visual motor functions. 


Explore further

Teaching robots to think like us: Brain cells, electrical impulses steer robot though maze

More information: Iason Batzianoulis et al, Customizing skills for assistive robotic manipulators, an inverse reinforcement learning approach with error-related potentials, Communications Biology (2021). DOI: 10.1038/s42003-021-02891-8
Journal information: Communications Biology

Citation: Mind-controlled robots now one step closer (2021, December 16) retrieved 27 May 2022 from https://techxplore.com/news/2021-12-mind-controlled-robots-closer.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.
418 shares

Feedback to editors