Exploiting glottal and prosodic information for robust speaker verification

Yuan Fu Liao, Zhi Ren Zeng, Zi He Chen, Yau Tarng Juang

研究成果: Conference article同行評審

摘要

In this paper, three different levels of speaker cues including the glottal, prosodic and spectral information are integrated together to build a robust speaker verification system. The major purpose is to resist the distortion of channels and handsets. Especially, the dynamic behavior of normalized amplitude quotient (NAQ) and prosodic feature contours are modeled using Gaussian of mixture models (GMMs) and two latent prosody analyses (LPAs)-based approaches, respectively. The proposed methods are evaluated on the standard one speaker detection task of the 2001 NIST Speaker Recognition Evaluation Corpus where only one 2-minute training and 30-second trial speech (in average) are available. Experimental results have shown that the proposed approach could improve the equal error rates (EERs) of maximum a priori-adapted (MAP)-GMMs and GMMs+T-norm approaches from 12.4% and 9.5% to 10.3% and 8.3% and finally to 7.8%, respectively.

原文English
期刊Proceedings of the International Conference on Speech Prosody
出版狀態Published - 2006
事件3rd International Conference on Speech Prosody, SP 2006 - Dresden, 德國
持續時間: 2 5月 20065 5月 2006

指紋

深入研究「Exploiting glottal and prosodic information for robust speaker verification」主題。共同形成了獨特的指紋。

引用此