AI for Social Good – Google AI Weblog

on

|

views

and

comments


Google’s AI for Social Good crew consists of researchers, engineers, volunteers, and others with a shared deal with constructive social impression. Our mission is to show AI’s societal profit by enabling real-world worth, with tasks spanning work in public well being, accessibility, disaster response, local weather and power, and nature and society. We imagine that one of the simplest ways to drive constructive change in underserved communities is by partnering with change-makers and the organizations they serve.

On this weblog submit we focus on work achieved by Mission Euphonia, a crew inside AI for Social Good, that goals to enhance automated speech recognition (ASR) for folks with disordered speech. For folks with typical speech, an ASR mannequin’s phrase error price (WER) might be lower than 10%. However for folks with disordered speech patterns, comparable to stuttering, dysarthria and apraxia, the WER might attain 50% and even 90% relying on the etiology and severity. To assist handle this downside, we labored with greater than 1,000 individuals to gather over 1,000 hours of disordered speech samples and used the information to point out that ASR personalization is a viable avenue for bridging the efficiency hole for customers with disordered speech. We have proven that personalization might be profitable with as little as 3-4 minutes of coaching speech utilizing layer freezing strategies.

This work led to the event of Mission Relate for anybody with atypical speech who may benefit from a customized speech mannequin. In-built partnership with Google’s Speech crew, Mission Relate allows individuals who discover it exhausting to be understood by different folks and expertise to coach their very own fashions. Folks can use these customized fashions to speak extra successfully and achieve extra independence. To make ASR extra accessible and usable, we describe how we fine-tuned Google’s Common Speech Mannequin (USM) to higher perceive disordered speech out of the field, with out personalization, to be used with digital assistant applied sciences, dictation apps, and in conversations.

Addressing the challenges

Working intently with Mission Relate customers, it turned clear that customized fashions might be very helpful, however for a lot of customers, recording dozens or a whole lot of examples might be difficult. As well as, the customized fashions didn’t at all times carry out nicely in freeform dialog.

To handle these challenges, Euphonia’s analysis efforts have been specializing in speaker impartial ASR (SI-ASR) to make fashions work higher out of the field for folks with disordered speech in order that no further coaching is critical.

Prompted Speech dataset for SI-ASR

Step one in constructing a sturdy SI-ASR mannequin was to create consultant dataset splits. We created the Prompted Speech dataset by splitting the Euphonia corpus into prepare, validation and take a look at parts, whereas guaranteeing that every cut up spanned a spread of speech impairment severity and underlying etiology and that no audio system or phrases appeared in a number of splits. The coaching portion consists of over 950k speech utterances from over 1,000 audio system with disordered speech. The take a look at set accommodates round 5,700 utterances from over 350 audio system. Speech-language pathologists manually reviewed the entire utterances within the take a look at set for transcription accuracy and audio high quality.

Actual Dialog take a look at set

Unprompted or conversational speech differs from prompted speech in a number of methods. In dialog, folks converse sooner and enunciate much less. They repeat phrases, restore misspoken phrases, and use a extra expansive vocabulary that’s particular and private to themselves and their neighborhood. To enhance a mannequin for this use case, we created the Actual Dialog take a look at set to benchmark efficiency.

The Actual Dialog take a look at set was created with the assistance of trusted testers who recorded themselves talking throughout conversations. The audio was reviewed, any personally identifiable info (PII) was eliminated, after which that knowledge was transcribed by speech-language pathologists. The Actual Dialog take a look at set accommodates over 1,500 utterances from 29 audio system.

Adapting USM to disordered speech

We then tuned USM on the coaching cut up of the Euphonia Prompted Speech set to enhance its efficiency on disordered speech. As an alternative of fine-tuning the complete mannequin, our tuning was based mostly on residual adapters, a parameter-efficient tuning strategy that provides tunable bottleneck layers as residuals between the transformer layers. Solely these layers are tuned, whereas the remainder of the mannequin weights are untouched. We’ve beforehand proven that this strategy works very nicely to adapt ASR fashions to disordered speech. Residual adapters have been solely added to the encoder layers, and the bottleneck dimension was set to 64.

Outcomes

To guage the tailored USM, we in contrast it to older ASR fashions utilizing the 2 take a look at units described above. For every take a look at, we evaluate tailored USM to the pre-USM mannequin greatest suited to that job: (1) For brief prompted speech, we evaluate to Google’s manufacturing ASR mannequin optimized for brief type ASR; (2) for longer Actual Dialog speech, we evaluate to a mannequin skilled for lengthy type ASR. USM enhancements over pre-USM fashions might be defined by USM’s relative measurement improve, 120M to 2B parameters, and different enhancements mentioned within the USM weblog submit.

Mannequin phrase error charges (WER) for every take a look at set (decrease is healthier).

We see that the USM tailored with disordered speech considerably outperforms the opposite fashions. The tailored USM’s WER on Actual Dialog is 37% higher than the pre-USM mannequin, and on the Prompted Speech take a look at set, the tailored USM performs 53% higher.

These findings recommend that the tailored USM is considerably extra usable for an finish consumer with disordered speech. We are able to show this enchancment by taking a look at transcripts of Actual Dialog take a look at set recordings from a trusted tester of Euphonia and Mission Relate (see under).

Audio1    Floor Reality    Pre-USM ASR    Tailored USM
                    
   I now have an Xbox adaptive controller on my lap.    i now have lots and that guide on my mouth    i now had an xbox adapter controller on my lamp.
                    
   I have been speaking for fairly some time now. Let’s have a look at.    fairly some time now    i have been speaking for fairly some time now.
Instance audio and transcriptions of a trusted tester’s speech from the Actual Dialog take a look at set.

A comparability of the Pre-USM and tailored USM transcripts revealed some key benefits:

  • The primary instance reveals that Tailored USM is healthier at recognizing disordered speech patterns. The baseline misses key phrases like “XBox” and “controller” which can be vital for a listener to know what they’re making an attempt to say.
  • The second instance is an effective instance of how deletions are a main challenge with ASR fashions that aren’t skilled with disordered speech. Although the baseline mannequin did transcribe a portion accurately, a big a part of the utterance was not transcribed, shedding the speaker’s meant message.

Conclusion

We imagine that this work is a vital step in the direction of making speech recognition extra accessible to folks with disordered speech. We’re persevering with to work on bettering the efficiency of our fashions. With the fast developments in ASR, we purpose to make sure folks with disordered speech profit as nicely.

Acknowledgements

Key contributors to this challenge embody Fadi Biadsy, Michael Brenner, Julie Cattiau, Richard Cave, Amy Chung-Yu Chou, Dotan Emanuel, Jordan Inexperienced, Rus Heywood, Pan-Pan Jiang, Anton Kast, Marilyn Ladewig, Bob MacDonald, Philip Nelson, Katie Seaver, Joel Shor, Jimmy Tobin, Katrin Tomanek, and Subhashini Venugopalan. We gratefully acknowledge the assist Mission Euphonia acquired from members of the USM analysis crew together with Yu Zhang, Wei Han, Nanxin Chen, and lots of others. Most significantly, we wished to say an enormous thanks to the two,200+ individuals who recorded speech samples and the numerous advocacy teams who helped us join with these individuals.


1Audio quantity has been adjusted for ease of listening, however the authentic recordsdata could be extra in keeping with these utilized in coaching and would have pauses, silences, variable quantity, and so on. 

Share this
Tags

Must-read

Nvidia CEO reveals new ‘reasoning’ AI tech for self-driving vehicles | Nvidia

The billionaire boss of the chipmaker Nvidia, Jensen Huang, has unveiled new AI know-how that he says will assist self-driving vehicles assume like...

Tesla publishes analyst forecasts suggesting gross sales set to fall | Tesla

Tesla has taken the weird step of publishing gross sales forecasts that recommend 2025 deliveries might be decrease than anticipated and future years’...

5 tech tendencies we’ll be watching in 2026 | Expertise

Hi there, and welcome to TechScape. I’m your host, Blake Montgomery, wishing you a cheerful New Yr’s Eve full of cheer, champagne and...

Recent articles

More like this

LEAVE A REPLY

Please enter your comment!
Please enter your name here