Time to read: 10 minutes
In terms of hearing technology, hearing aids have become smaller and better at providing better speech processing across a wide range of circumstances over the past 20 or so years. Today’s hearing aids offer more programs and user friendly features including Bluetooth connectivity and streaming. On top of that, rechargeable batteries offer long lasting operation without the waste of disposable batteries.
This brings us to the fourth item on the list - Artificial Intelligence. Let’s start with giving it a definition.
Francois Chollet, AI researcher at Google and creator of the machine-learning software library Keras, has said intelligence is tied to a system's ability to adapt and improvise in a new environment, to generalise its knowledge and apply it to unfamiliar scenarios.
"Intelligence is the efficiency with which you acquire new skills at tasks you didn't previously prepare for," he said. "Intelligence is not skill itself, it's not what you can do, it's how well and how efficiently you can learn new things."
No surprise that if after reading that you have visions of HAL from 2001: A Space Odyssey or Skynet from The Terminator films.
The good news (for now) is that Artificial Intelligence is far from reaching ‘self-awareness stage’
It is in the area of Narrow AI where machines excel. That means they are very, very good - better than people - at working in a very narrow set of parameters. Still, that’s nothing to sneeze at. Siri, Alexa and Cortana are all examples of Narrow AI. So too are those message bots that pop up on some web sites to ask if they can help you.
We encounter Narrow AI everywhere. It shapes what we see on search engines, it makes movie recommendations on Netflix, it recognises our faces at passport control at airports.
Now it’s being integrated into hearing aids and the opportunities to make your life more convenient and safe are growing just as fast as the technology itself.
Let’s take a look at hearing aid manufacturers who are promoting the use of AI in their hearing aids and associated apps.
Starkey’s Livio Edge AI uses artificial intelligence in three ways:
Hearing performance - which is the first and most important thing for a hearing aid to do. Here, artificial intelligence is used to provide superior sound quality. Starkey's Edge Mode conducts an AI-based analysis of the environment to make immediate adjustments designed to improve speech audibility issues caused by background noise and even face masks.
(Value Hearing Audiologist Emma Russell looks at the Starkey Livio in this video here.)
Starkey also uses AI in what they call 'healthable technology' which, through the Thrive app features:
The Intelligent Assistant feature offers a wide range of helpful little features including a Siri-like assistant to provide quick answers about your hearing aid and to set personal reminders. Another helpful feature is the ability for your clinician to perform remote support adjustments on your hearing aids as you're wearing them, no matter where you are.
Other features include language translation, speech-to-text transcription, find my phone and find my hearing aid features.
SoundSense Learn uses AI to personalise hearing in the moment using Machine Learning, to calculate the best possible hearing outcome for a given situation in just 20 comparisons. And every time this AI is used, it stores the information it’s given in the cloud, so that it can improve hearing for other users of this feature.
To reach the same result, a user would have to do almost 2,500,000 comparisons for the same result.
(Value Hearing Audiologist Emma Russell reviews Widex Moment here).
Signia also uses machine learning in its hearing aid range.
Many people with hearing loss cite the sound of their own voice as a downside to using hearing aids. Signia addressed this issue by developing Own Voice Processing (OVP) for a more natural-sounding own voice.
In order to function properly, the artificial intelligence involved in OVP has to “learn” the wearer’s voice. While machine learning could take weeks in previous hearing aid models, the voice recognition program in OVP takes mere seconds.
When combined with the Signia Assistant on your mobile phone, it uses AI to tailor your hearing aid settings to your personal preferences for an even clearer sound and best speech comprehension in every situation. It also answers your questions about how to handle your hearing aids so you can easily recall every detail discussed during your appointment with your hearing care professional.
Phonak uses AutoSense OS 4.0 to personalise experience in any sound environment. AutoSense OS 4.0 blends the right combination of sound processing features to support a variety of end-user communication needs and lifestyles. New features include: Speech Enhancer, Dynamic Noise Cancellation and Motion Sensor Hearing.
Oticon has trained the Deep Neural Network (DNN) with 12 million real-life sound scenes. It’s our most advanced hearing aid technology ever.
The DNN has learned to recognise all types of sounds, their details, and how they should ideally sound.
This makes it so much more than a standard artificial intelligence software. It’s a unique and dedicated hearing aid technology, developed for real-time operation in everyday life.