The technique lets users perceive differences between musical instruments, a significant improvement from what standard cochlear implants can offer, said lead researcher Les Atlas, a University of Washington professor of electrical engineering.
‘Right now, cochlear-implant subjects do well when it’s quiet and there is a single person talking, but with music, noisy rooms or multiple people talking, it’s difficult to hear,’ Atlas said in a statement. ‘We are on the way to solving the issue with music.’
Atlas and other researchers believe that hearing music has possible links to hearing speech better in noisy settings, another goal of this research.
Atlas and collaborator Jay Rubinstein, a UW professor of otolaryngology and of bioengineering, and members of their labs recently published their initial findings in the IEEE Transactions on Neural Systems and Rehabilitation Engineering. A study on eight cochlear-implant users showed that using this new coding strategy let them distinguish between musical instruments much more accurately than with the standard devices.
The researchers hope to fine-tune the signal processing to make it compatible with cochlear implants already on the market so users can improve their music perception right away. They also are working on algorithms to better support device users’ perception of pitch and melody.
A cochlear implant is a small, electronic device that lets a person who is profoundly deaf or hard of hearing perceive sound. One piece is placed on the skin behind a person’s ear, while another portion is surgically inserted under the skin. The implant works by directly stimulating the auditory nerve, bypassing damaged portions of the ear. The implant’s signals are sent to the brain, which recognizes the signals as sounds.
The UW scientists developed a new way to process the sounds of musical melodies and notes, which tend to be more complex than speech. Specifically, they tried to improve the ability of cochlear-implant users to detect pitch and timbre in songs.
Pitch is associated with the melody of a song and intonation when speaking. Timbre relates most closely to the varying sounds that different instruments make when playing the same note.
People who use cochlear implants usually perceive words by their syllables and rhythms, not through tone or inflection. The researchers tested their new processing technique on cochlear-implant users by playing common melodies such as ‘Twinkle, Twinkle, Little Star’ with the rhythms removed. They found that timbre recognition – the ability to distinguish between instruments – increased significantly, but the ability to perceive a melody was still difficult for most people.
‘This is the first time anyone has demonstrated increased timbre perception using a different signal-processing system,’ said Rubinstein, a physician at the UW Medical Center and Seattle Children’s hospital and director of the Virginia Merrill Bloedel Hearing Research Center. ‘With cochlear implants, we’ve always been oriented more toward speech sounds. This strategy represents a different way of thinking about signal processing for music.’
This research was funded by the National Institutes of Health, the National Science Foundation, the U.S. Army Research Office, the Institute of Translational Health Sciences at the UW and a Virginia Merrill Bloedel Scholar Award.
Oxa launches autonomous Ford E-Transit for van and minibus modes
I'd like to know where these are operating in the UK. The report is notably light on this. I wonder why?