Files in this item

FilesDescriptionFormat

application/pdf

application/pdfYANG-DISSERTATION-2019.pdf (4MB)Restricted Access
(no description provided)PDF

Description

Title:Dealing with linguistic mismatches for automatic speech recognition
Author(s):Yang, Xuesong
Director of Research:Hasegawa-Johnson, Mark
Doctoral Committee Chair(s):Hasegawa-Johnson, Mark
Doctoral Committee Member(s):Huang, Thomas S; Smaragdis, Paris; Shih, Chilin
Department / Program:Graduate College Programs
Discipline:Informatics
Degree Granting Institution:University of Illinois at Urbana-Champaign
Degree:Ph.D.
Genre:Dissertation
Subject(s):Automatic Speech Recognition
Acoustic Modeling
Multi-Accents
Multi-Lingual
Acoustic Phonetics
Distinctive Features
Acoustic Landmarks
End-to-End
Multi-Task Learning
Model Compression
Deep Learning
Pronunciation Error Detection
Connectionist Temporal Classification
Abstract:Recent breakthroughs in automatic speech recognition (ASR) have resulted in a word error rate (WER) on par with human transcribers on the English Switchboard benchmark. However, dealing with linguistic mismatches between the training and testing data is still a significant challenge that remains unsolved. Under the monolingual environment, it is well-known that the performance of ASR systems degrades significantly when presented with the speech from speakers with different accents, dialects, and speaking styles than those encountered during system training. Under the multi-lingual environment, ASR systems trained on a source language achieve even worse performance when tested on another target language because of mismatches in terms of the number of phonemes, lexical ambiguity, and power of phonotactic constraints provided by phone-level n-grams. In order to address the issues of linguistic mismatches for current ASR systems, my dissertation investigates both knowledge-gnostic and knowledge-agnostic solutions. In the first part, classic theories relevant to acoustics and articulatory phonetics that present capability of being transferred across a dialect continuum from local dialects to another standardized language are re-visited. Experiments demonstrate the potentials that acoustic correlates in the vicinity of landmarks could help to build a bridge for dealing with mismatches across difference local or global varieties in a dialect continuum. In the second part, we design an end-to-end acoustic modeling approach based on connectionist temporal classification loss and propose to link the training of acoustics and accent altogether in a manner similar to the learning process in human speech perception. This joint model not only performed well on ASR with multiple accents but also boosted accuracies of accent identification task in comparison to separately-trained models.
Issue Date:2019-04-15
Type:Thesis
URI:http://hdl.handle.net/2142/105187
Rights Information:Copyright 2019 Xuesong Yang
Date Available in IDEALS:2019-08-23
Date Deposited:2019-05


This item appears in the following Collection(s)

Item Statistics