Leveraging Linguistic Context in Dyadic Interactions to Improve Automatic Speech Recognition for Children. (September 2020)
- Record Type:
- Journal Article
- Title:
- Leveraging Linguistic Context in Dyadic Interactions to Improve Automatic Speech Recognition for Children. (September 2020)
- Main Title:
- Leveraging Linguistic Context in Dyadic Interactions to Improve Automatic Speech Recognition for Children
- Authors:
- Kumar, Manoj
Kim, So Hyun
Lord, Catherine
Lyon, Thomas D.
Narayanan, Shrikanth - Abstract:
- Highlights: Participating adult's speech during child-adult dyadic conversation helps improving child speech recognition. Long-term semantic context from adult speech is captured by semantic response generation using seq2seq model. Context-adapted language models result in 10.71% WER (absolute) and 60.55% perplexity (relative) improvement. Abstract: Automatic speech recognition for child speech has been long considered a more challenging problem than for adult speech. Various contributing factors have been identified such as larger acoustic speech variability including mispronunciations due to continuing biological changes in growth, developing vocabulary and linguistic skills, and scarcity of training corpora. A further challenge arises when dealing with spontaneous speech of children involved in a conversational interaction, and especially when the child may have limited or impaired communication ability. This includes health applications, one of the motivating domains of this paper, that involve goal-oriented dyadic interactions between a child and clinician/adult social partner as a part of behavioral assessment. In this work, we use linguistic context information from the interaction to adapt speech recognition models for children speech. Specifically, spoken language from the interacting adult speech provides the context for the child's speech. We propose two methods to exploit this context: lexical repetitions and semantic response generation. For the latter, we makeHighlights: Participating adult's speech during child-adult dyadic conversation helps improving child speech recognition. Long-term semantic context from adult speech is captured by semantic response generation using seq2seq model. Context-adapted language models result in 10.71% WER (absolute) and 60.55% perplexity (relative) improvement. Abstract: Automatic speech recognition for child speech has been long considered a more challenging problem than for adult speech. Various contributing factors have been identified such as larger acoustic speech variability including mispronunciations due to continuing biological changes in growth, developing vocabulary and linguistic skills, and scarcity of training corpora. A further challenge arises when dealing with spontaneous speech of children involved in a conversational interaction, and especially when the child may have limited or impaired communication ability. This includes health applications, one of the motivating domains of this paper, that involve goal-oriented dyadic interactions between a child and clinician/adult social partner as a part of behavioral assessment. In this work, we use linguistic context information from the interaction to adapt speech recognition models for children speech. Specifically, spoken language from the interacting adult speech provides the context for the child's speech. We propose two methods to exploit this context: lexical repetitions and semantic response generation. For the latter, we make use of sequence-to-sequence models that learn to predict the target child utterance given context adult utterances. Long-term context is incorporated in the model by propagating the cell-state across the duration of conversation. We use interpolation techniques to adapt language models at the utterance level, and analyze the effect of length and direction of context (forward and backward). Two different domains are used in our experiments to demonstrate the generalized nature of our methods - interactions between a child with ASD and an adult social partner in a play-based, naturalistic setting, and in forensic interviews between a child and a trained interviewer. In both cases, context-adapted models yield significant improvement (upto 10.71% in absolute word error rate) over the baseline and perform consistently across context windows and directions. Using statistical analysis, we investigate the effect of source-based (adult) and target-based (child) factors on adaptation methods. Our results demonstrate the applicability of our modeling approach in improving child speech recognition by employing information transfer from the adult interlocutor. … (more)
- Is Part Of:
- Computer speech & language. Volume 63(2020)
- Journal:
- Computer speech & language
- Issue:
- Volume 63(2020)
- Issue Display:
- Volume 63, Issue 2020 (2020)
- Year:
- 2020
- Volume:
- 63
- Issue:
- 2020
- Issue Sort Value:
- 2020-0063-2020-0000
- Page Start:
- Page End:
- Publication Date:
- 2020-09
- Subjects:
- Child speech -- Automatic speech recognition -- Autism spectrum disorder -- Forensic interviews
Speech processing systems -- Periodicals
Automatic speech recognition -- Periodicals
Computers -- Periodicals
Linguistics -- Periodicals
Speech-Language Pathology -- Periodicals
Traitement automatique de la parole -- Périodiques
Reconnaissance automatique de la parole -- Périodiques
Automatic speech recognition
Speech processing systems
Electronic journals
Periodicals
006.454 - Journal URLs:
- http://www.journals.elsevier.com/computer-speech-and-language/ ↗
http://www.elsevier.com/journals ↗ - DOI:
- 10.1016/j.csl.2020.101101 ↗
- Languages:
- English
- ISSNs:
- 0885-2308
- Deposit Type:
- Legaldeposit
- View Content:
- Available online (eLD content is only available in our Reading Rooms) ↗
- Physical Locations:
- British Library DSC - 3394.276600
British Library DSC - BLDSS-3PM
British Library HMNTS - ELD Digital store - Ingest File:
- 13576.xml