Regularized Estimation of Gaussian Mixture Models for SVM Based Speaker Recognition
Qian, Kaizhi
Loading…
Permalink
https://hdl.handle.net/2142/55486
Description
Title
Regularized Estimation of Gaussian Mixture Models for SVM Based Speaker Recognition
Author(s)
Qian, Kaizhi
Contributor(s)
Hasegawa-Johnson, Mark
Issue Date
2014-05
Keyword(s)
speaker recognition
elastic net
sparsity
Gaussian mixture model
supervector
Abstract
Speaker adaptation based on the Universal Background Model (UBM) has become a standard approach for speaker recognition. A GMM supervector is constructed by normalizing and stacking the means of the adapted mixture components, which provides a compact representation of the speaker-dependent model in speaker recognition tasks. The estimation of the unknown GMM parameters is usually obtained by the method of maximum a posteriori estimation (MAP), which can be regularized to increase the model interpretability with insufficient training data. In this thesis, the speaker-adapted models are estimated using the MAP with L1-regularization, referred to as the elastic net, based on the assumption that the distinctions between any two speakers are sparse. Experiments on the NIST2008 speaker recognition evaluation task show error rate reduction with the elastic net.
Use this login method if you
don't
have an
@illinois.edu
email address.
(Oops, I do have one)
IDEALS migrated to a new platform on June 23, 2022. If you created
your account prior to this date, you will have to reset your password
using the forgot-password link below.