The research found that people’s attention levels and how engrossed they are in on-screen activities can be detected from their eye movements. The team’s findings are published in Cognitive Research: Principles and Implications.
The research suggests a new way to determine the readiness of drivers using auto-pilot mode to respond to real-world signals, such as takeover requests from the car.
Cars with a ‘driverless’ auto-pilot mode are available for commercial private use in some locations, including Germany and certain US states. When using the auto-pilot mode, drivers can take their hands off the wheel and participate in other activities, but may require the driver to take back control of the car when necessary.
The researchers tested whether it was possible to detect if a person was too engrossed in another task to respond swiftly to such a “takeover” signal.
To do this, the team tested 42 participants across two experiments using a procedure that mimicked a “takeover” scenario used in some advanced models of cars with an auto-pilot mode.
Participants were required to search a computer screen with coloured shapes for some target items and linger their gaze on the targets to show they had found them.
The search tasks were either easy (requiring participants to spot an odd ‘L’ shape amongst ‘T’ shapes), or more demanding (requiring participants to spot a specific arrangement of the shape parts and their colour).
At later points in their search task, a tone would then sound and the participants were required to stop watching the screen as fast as they could, pressing a button in response to it.
Researchers monitored the time it took between the tone sounding and the participants pressing the button, alongside analysing how their eyes moved across the screen during their search, to see if attention levels to the task could be detected from a change in their gaze.
They found that when the task demanded more attention, participants took a longer time to stop watching the screen and respond to the tone.
The analysis showed that it was possible to detect participants’ attention levels from their eye movements. An eye movement pattern involving longer fixations and shorter distance of eye travel between all items indicated that the task was more demanding on attention.
According to UCL, the researchers also trained a machine learning model on this data and found that they could predict whether the participants were engaged in the easy or demanding task based on their eye movement patterns.
In a statement, senior author Professor Nilli Lavie, UCL Institute of Cognitive Neuroscience, said: “Driverless car technology is fast advancing and promises a more enjoyable and productive driving experience, where drivers can use their commuting time for other non-driving tasks.
“However, the big question is whether the driver will be able to return to driving swiftly upon receiving a takeover signal if they are fully engaged in another activity.
“Our findings show that it is possible to detect the attention levels of a driver and their readiness to respond to a warning signal, just from monitoring their gaze pattern.
“It is striking that people can get so consumed with their on-screen activity that they ignore the rest of the world around them. Even when they are aware that they should be ready to stop their task and respond to tones as quickly as they can, they take longer to do it when their attention is engrossed in the screen.
“Our research shows that warning signals may not be noticed quickly enough in such cases.”
UCL said larger datasets are required in order to train the machine learning and make it more accurate.
The research was funded by JLR and EPSRC as part of the jointly funded Towards Autonomy: Smart and Connected Control (TASCC) programme.
Onshore wind and grid queue targeted in 2030 energy plan
NESO is expecting the gas powered turbines (all of them) to run for 5% of the time!. I did not realise that this was in the actual plan - but not...