Led by Dr. Lola Cañamero at Hertfordshire University and in collaboration with a consortium of universities and robotic companies across Europe, these robots differ from others in the way that they form attachments, interact and express emotion through bodily expression.
Developed as part of the interdisciplinary project FEELIX GROWING (Feel, interact, express: a global approach to development with interdisciplinary grounding), funded by the Sixth Framework Programme of the European Commission and coordinated by Dr Cañamero, the robots have been developed so that they learn to interact with and respond to humans in a similar way to how children learn to do it, and use the same types of expressive and behavioural cues that babies use to learn to interact socially and emotionally with others.
The robots have been created through modelling the early attachment process that human and chimpanzee infants undergo with their caregivers when they develop a preference for a primary caregiver.
They are programmed to learn to adapt to the actions and mood of their human caregivers, and to become particularly attached to an individual who interacts with the robot in a way that is particularly suited to its personality profile and learning needs. The more they interact and are given the appropriate feedback and level of engagement from the human caregiver, the stronger the bond developed and the amount learned.
The robots are said to be capable of expressing anger, fear, sadness, happiness, excitement and pride, and will demonstrate distress if the caregiver fails to provide them comfort when confronted by a stressful situation that they cannot cope with or to interact with them when they need it.
’This behaviour is modelled on what a young child does,’ said Dr Cañamero. ’This is also very similar to the way chimpanzees and other non-human primates develop affective bonds with their caregivers.’
This is the first time that early attachment models of human and non-human primates have been used to program robots that develop emotions in interaction with humans.
’We are working on non-verbal cues and the emotions are revealed through physical postures, gestures and movements of the body, rather than facial or verbal expression,’ Dr Cañamero added.
The researchers, led by Dr Cañamero at Hertfordshire University, are now extending the prototype further and adapting it as part of the EU project ALIZ-E, which will develop robots that learn to be a carer/companion for diabetic children in hospital settings.
Within this project, coordinated by Dr Tony Belpaeme of Plymouth University, the Hertfordshire group will lead research related to the emotions and non-linguistic behaviour of the robots.
The future robot companions will combine non-linguistic and linguistic communication to interact with the children and become increasingly adapted to their individual profiles in order to support both the therapeutic aspects of their treatment and their social and emotional wellbeing.
Oxa launches autonomous Ford E-Transit for van and minibus modes
I'd like to know where these are operating in the UK. The report is notably light on this. I wonder why?