Robots could learn to recognise human emotions, study finds

Their findings suggest robots could learn to use the same movements, alongside facial expressions and tone of voice, to recognise human internal states.

Robots could be taught to recognise human emotions from our movements, a new study shows.

Researchers found humans could recognise excitement, sadness, aggression, and boredom from the way people moved, even if they could not see their facial expressions or hear their voice.
Their findings suggest robots could learn to use the same movements, alongside facial expressions and tone of voice, to recognise human internal states.

It raises the prospect that robots already used to teach second languages could recognise when students are bored and customer service robots could identify when people feel angry or stressed.
Dr Charlotte Edmunds, from Warwick Business School, said: "One of the main goals in the field of human-robot interaction is to create machines that can recognise human emotions and respond accordingly.

"Our results suggest it is reasonable to expect a machine learning algorithm, and consequently a robot, to recognise a range of emotions and social interactions using movements, poses, and facial expressions. The potential applications are huge."

The study was conducted by researchers form Warwick Business School, University of Plymouth, Donders Centre for Cognition at Radboud University in the Netherlands, and the Bristol Robotics Lab at the University of the West of England. It is published in the journal Frontiers in Robotics and AI.
The team of psychologists and computer scientists filmed pairs of children playing with a robot and a computer built into a table with a touchscreen top.

The videos were shown to 284 study participants, who were asked to decide whether the children were excited, bored, or sad. They were also asked if the children were cooperating, competing, or if one of the children had assumed a dominant role in the relationship.

Some participants watched the original videos. A second group saw the footage reduced to ‘stickman' figures that showed exactly the same movements.

Members of both groups agreed on the same emotional label for the children, more often than would be expected if they were guessing.

The researchers then trained a machine-learning algorithm to label the clips, identifying the type of social interaction, the emotions on display, and the strength of each child's internal state, allowing it to compare which child felt more sad or excited.

Dr Edmunds said: "Robot delivery services are already being trialled, but people tend to attack or vandalise them, often because feel threatened.

"The aim is to create a robot that can react to human emotions in difficult situations and get itself out of trouble without having to be monitored or told what to do.

"That doesn't happen at the moment, because robots tend to follow scripts."
Madeleine Bartlett, from the University of Plymouth, said: "It is important that a social robot can recognise human internal states, such as stress, and their strength or severity.

"Different levels of stress require different responses. Low level stress might just require the robot to back away from the human, whilst a high level of stress might be best addressed by having the robot apologise and leave the interaction."

Featured Product

3D Vision: Ensenso B now also available as a mono version!

3D Vision: Ensenso B now also available as a mono version!

This compact 3D camera series combines a very short working distance, a large field of view and a high depth of field - perfect for bin picking applications. With its ability to capture multiple objects over a large area, it can help robots empty containers more efficiently. Now available from IDS Imaging Development Systems. In the color version of the Ensenso B, the stereo system is equipped with two RGB image sensors. This saves additional sensors and reduces installation space and hardware costs. Now, you can also choose your model to be equipped with two 5 MP mono sensors, achieving impressively high spatial precision. With enhanced sharpness and accuracy, you can tackle applications where absolute precision is essential. The great strength of the Ensenso B lies in the very precise detection of objects at close range. It offers a wide field of view and an impressively high depth of field. This means that the area in which an object is in focus is unusually large. At a distance of 30 centimetres between the camera and the object, the Z-accuracy is approx. 0.1 millimetres. The maximum working distance is 2 meters. This 3D camera series complies with protection class IP65/67 and is ideal for use in industrial environments.