We've updated our Privacy Policy to make it clearer how we use your personal data. We use cookies to provide you with a better experience. You can read our Cookie Policy here.

Advertisement

Future Hearing Aids Could Read Brainwaves To Listen To Specific Speakers

Future Hearing Aids Could Read Brainwaves To Listen To Specific Speakers content piece image
Credit: Pete Linforth/ Pixabay
Listen with
Speechify
0:00
Register for free to listen to this article
Thank you. Listen to this article using the player above.

Want to listen to this article for FREE?

Complete the form below to unlock access to ALL audio articles.

Read time: 1 minute

In a noisy room with many speakers, hearing aids can suppress background noise, but they have difficulties isolating one voice - that of the person you're talking to at a party, for instance. KU Leuven researchers have now addressed that issue with a technique that uses brainwaves to determine within one second whom you're listening to.

Having a casual conversation at a cocktail party is a challenge for someone with a hearing aid, says Professor Tom Francart from the Department of Neurosciences at KU Leuven: "A hearing aid may select the loudest speaker in the room, for instance, but that is not necessarily the person you're listening to. Alternatively, the system may take into account your viewing direction, but when you're driving a car, you can't look at the passenger sitting next to you."

Researchers have been working on solutions that take into account what the listener wants. "An electroencephalogram (EEG) can measure brainwaves that develop in response to sounds. This technique allows us to determine which speaker someone wants to listen to. The system separates the sound signals produced by different speakers and links them to the brainwaves. The downside is that you have to take into account a delay of ten to twenty seconds to get it right with reasonable certainty."

Artificial intelligence to speed up the process

A new technique makes it possible to step up the pace, Professor Alexander Bertrand from the Department of Electrical Engineering at KU Leuven continues: "Using artificial intelligence, we found that it is possible to directly decode the listening direction from the brainwaves alone, without having to link them to the actual sounds."

"We trained our system to determine whether someone is listening to a speaker on their left or their right. Once the system has identified the direction, the acoustic camera redirects its aim, and the background noise is suppressed. On average, this can now be done within less than one second. That's a big leap forward, as one second constitutes a realistic timespan to switch from one speaker to the other."

From lab to real life

However, it will take at least another five years before we have smart hearing aids that work with brainwaves, Professor Francart continues. "To measure someone's brainwaves in the lab, we make them wear a cap with electrodes. This method is obviously not feasible in real life. But research is already being done into hearing aids with built-in electrodes."

The new technique will be further improved as well, PhD student Simon Geirnaert adds. "We're already conducting further research, for instance into the problem of combining multiple speaker directions at once. The current system simply chooses between two directions. While first experiments show that we can expand that to other possible directions, we need to refine our artificial intelligence system by feeding the system with more brainwave data from users who are also listening to speakers from other directions."

Reference: Geirnaert S, Francart T, Bertrand A. Fast EEG-based decoding of the directional focus of auditory attention using common spatial patterns. IEEE. 2020. doi:10.1109/TBME.2020.3033446

This article has been republished from the following materials. Note: material may have been edited for length and content. For further information, please contact the cited source.