The advance from MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL) could lead to robot-human interaction in manufacturing and construction settings.
CSAIL’s RoboRaise involves placing electromyography (EMG) sensors on a human’s biceps and triceps to monitor muscle activity. Its algorithms then continuously detect changes to the person’s arm level, as well as discrete up-and-down hand gestures the user might make for finer motor control.
The team used the system for a series of tasks involving picking up and assembling mock airplane components. In experiments, users worked on these tasks with the robot and were able to control it to within a few inches of the desired heights by lifting and then tensing their arm. It was more accurate when gestures were used, and the robot is said to have responded correctly to roughly 70 per cent of all gestures.
“Our approach to lifting objects with a robot aims to be intuitive and similar to how you might lift something with another person - roughly copying each other's motions while inferring helpful adjustments,” said graduate student Joseph DelPreto, lead author on a new paper about the project with MIT Professor and CSAIL director Daniela Rus. “The key insight is to use nonverbal cues that encode instructions for how to coordinate, for example to lift a little higher or lower. Using muscle signals to communicate almost makes the robot an extension of yourself that you can fluidly control.”
According to CSAIL, the project builds on the team’s existing system that allows users to instantly correct robot mistakes with brainwaves and hand gestures. “We aim to develop human-robot interaction where the robot adapts to the human, rather than the other way around. This way the robot becomes an intelligent tool for physical work,” said Rus.
According to CSAIL, EMG signals are often very noisy and it can be difficult to predict exactly how a limb is moving based on muscle activity. Even if it is possible to estimate how a person is moving, the desired response of the robot may be unclear.
RoboRaise is said to get around this by putting the human in control. The team’s system uses non-invasive, on-body sensors that detect the firing of neurons as muscles tense or relax. Using wearables also gets around problems of occlusions or ambient noise, which can complicate tasks involving vision or speech.
RoboRaise’s algorithm then processes biceps activity to estimate how the person’s arm is moving so the robot can roughly mimic it, and the person can slightly tense or relax their arm to move the robot up or down. If a user needs the robot to move farther away from their own position or hold a pose for a while, they can just gesture up or down for finer control; a neural network detects these gestures at any time based on biceps and triceps activity.
CSAIL claim that new users can start using the system very quickly, with minimal calibration. After putting on the sensors, they just need to tense and relax their arm a few times then lift a light weight to a few heights. The neural network that detects gestures is only trained on data from previous users.
The team tested the system with 10 users through a series of three lifting experiments: one where the robot didn’t move at all, another where the robot moved in response to their muscles but didn’t help lift the object, and a third where the robot and person lifted an object together.
When the person had feedback from the robot - when they could see it moving or when they were lifting something together - the achieved height was said to be significantly more accurate compared to having no feedback.
The team also tested RoboRaise on assembly tasks, such as lifting a rubber sheet onto a base structure. It was able to successfully lift rigid and flexible objects onto the bases. RoboRaise was implemented on the team’s Baxter humanoid robot, but the team says it could be adapted for any robotic platform.
In the future, the team hopes that adding more muscles or different types of sensors to the system will increase the degrees of freedom, with the ultimate goal of doing even more complex tasks. The team presented their work at the 2019 International Conference on Robotics and Automation
Promoted content: Does social media work for engineers – and how can you make it work for you?
So in addition to doing their own job, engineers are expected to do the marketing department´s work for them as well? Sorry, wait a minute, I know the...