When Alexa Can’t Understand You
When Whitney Bailey bought an Amazon Echo, she wanted to use the hands-free calling feature in case she fell and couldn鈥檛 reach her phone. She hoped that it would offer her family some peace of mind and help make life a little easier. In some ways, she says, it does. But because she has cerebral palsy, her voice is strained when she talks, and she struggles to get Alexa to understand her. To make matters worse, having to repeat commands strains her voice even more.
Thanks to technologies like Google Assistant and Amazon Alexa, we鈥檙e living in an increasingly . With a simple 鈥淎lexa,鈥 we ask a personal voice assistant to tell us the weather, make reservations, or play ambient music.* For most people, this is a pleasant convenience. For the and the , this technology means much more: With a single command, they can ,, or send text messages to loved ones. But for those who struggle to vocalize their speech, it鈥檚 a different story. If voice is the , tech companies need to prioritize developing software that is inclusive of all speech.
In the United States, have trouble using their voice and more than stutter, which can make it difficult for them to fully realize voice-enabled technology. Speech disabilities can stem from a wide variety of causes, including Parkinson鈥檚 disease, cerebral palsy, traumatic brain injuries, even age. In many cases, those with speech disabilities also have limited mobility and motor skills. This makes voice-enabled technology especially beneficial for them, as it doesn鈥檛 involve pushing buttons or tapping a screen. For disabled people, this technology can provide independence, making speech recognition that works for everyone all the more important.
As Americans age, the need to develop better software is becoming more pressing.
Yet voice-enabled tech developers struggle to meet their needs. People with speech disabilities use the same language and grammar that others do. But their speech musculature鈥攖hings like their tongue and jaw鈥攊s affected, resulting in consonants becoming slurred and vowels blending together, says Frank Rudzicz, an associate professor of computer science at the University of Toronto who studies speech and machine learning. These differences present a challenge in developing voice-enabled technologies.
Tech companies rely on user input to . The machine learning that makes voice-enabled tech possible requires massive amounts of data, which comes from the devices. Most of these data points come from younger abled users, says Rudzicz. This means that it can be challenging to use machine-learning techniques to develop inclusive voice-enabled technology that works consistently for populations whose speech varies widely, such as children, the elderly, and the disabled.
But as Americans age, the need to develop better software is becoming more pressing. Today, there are roughly Americans over the age of 65. By 2035, that number will be , meaning more people will be at risk of stroke and degenerative conditions that impair speech. At the same time, voice-enabled tech is becoming even more integrated into our lives, with some experts predicting that by 2020 nearly will have a voice assistant and will be done via voice. As voice-enabled technology becomes more ubiquitous, companies must adapt to the changing populations鈥攁nd their speech.
Andy Theyers, who鈥檚 about his struggle to use voice assistants due to his stutter, says that this is, in part, a reflection of an industry that doesn鈥檛 always prioritize accessibility from the beginning of a product鈥檚 development. Sean Lewis, a motivational speaker with cerebral palsy, agrees. 鈥淯nless [tech developers] personally know someone with a disability,鈥 Lewis says, they 鈥渉ave no idea how a lack of technology affects people鈥檚 lives.鈥
Lewis is grateful for his Samsung phone, which came with Google Voice Assistant. Before his Samsung, he wasn鈥檛 able to send emails or texts on his own. Now, he can do it all through voice commands. But he finds that he often has to repeat himself at least once or twice before the device understands him (a problem that many people without speech disabilities face as well, to be sure). Though voice-enabled tech has improved, he says, 鈥渨e鈥檙e not where we need to be.鈥
Twenty-some years ago Steven Salmon, an author with cerebral palsy, began using DragonDictate voice recognition software to write his books, spelling out words letter by letter. It was a time-consuming process that required his pronunciation to be perfectly consistent. 鈥淚f I had a cold,鈥 he says, 鈥淚 couldn鈥檛 write.鈥 When Salmon received an iPhone in 2015, he tried to get the device to respond to his commands. Voice-enabled tech was more accurate than ever鈥擲iri鈥檚 error rate was . Yet, his phone couldn鈥檛 understand his commands, so he ended up returning the device.
For Theyers, the biggest problem with using his Alexa is triggering it to listen to him. Until 2017, when Amazon 鈥渃omputer鈥 as a wake word, voice-enabled devices required wake phrases that began with a hard vowel鈥攖hink, 鈥淎lexa.鈥 The hard vowel triggers Theyers鈥 stutter, and often by the time he鈥檚 said the trigger word, the device has stopped listening.
He wishes that there was a way to better personalize devices鈥攁nd that technologists would seek out the opinions of people with voice impairments. Whitney Bailey agrees: 鈥淚t can be disheartening when a person has trouble using [technology] because they have a unique speech pattern.鈥
Rudzicz predicts that in the future, technology will be more individualized. Moreover, as tech companies collect massive amounts of data from millions of users, he anticipates that we鈥檒l begin to look at individual user data to see how it differs from the general population, allowing us to adapt models to the individual.
Some companies are already working to develop more individualized software. a startup, is currently beta-testing a speech-recognition app that translates nonstandard speech to standard speech in real time using a closed dictionary. Users, often with the help of a speech therapist or caregiver, create their own dictionary by reading short phrases or sentences. After they create the dictionary鈥攁 process that can take 30 minutes to three hours鈥攖hey can begin to use the app. Voiceitt鈥檚 goal for its first iteration is to help people vocalize their wants and needs, says Sara Smolley, Voiceitts鈥 vice president of strategy. But as the company collects more data, it is exploring how it might be able to find commonalities within demographics that would allow it to develop more tailored software鈥攕uch as special algorithms for native English-speaking, 40-year-old males or native Spanish-speaking 20-year-old females.
Larger tech firms also recognize their role in developing inclusive assistive technologies that are both widely available and relatively inexpensive. A spokesperson from Amazon said the company frequently receives positive feedback from 鈥渁ging-in-place鈥 customers who use Alexa鈥檚 smart-home features as an alternative to going up and down stairs. Amazon did not comment on its future plans regarding accessibility, but it pointed to Amazon鈥檚 Echo Show鈥攚hich offers users , a screen interface that lets users who are deaf and hard of hearing to tap common commands鈥攁s well as Alexa Captioning, which allows users to read Alexa鈥檚 responses. And Microsoft recently an A.I. for accessibility program to create inclusive, affordable tech. These features bring us closer to better technology, but they still have barriers for people with limited mobility and poor fine motor skills, who may be unable to easily walk over to a screen or tap small buttons.
As tech developers continue designing voice-enabled products, the key will be scaling up solutions and supporting their integration into existing technologies. Rudzicz predicts that we鈥檒l see better technologies as the population ages and companies try to cater to people with degenerative conditions. And as these technologies are developed with aging populations in mind, people with congenital disorders like cerebral palsy will benefit, too鈥攑ushing us one step closer to truly inclusive voice-enabled tech.
This article in , a collaboration among , , and .