Exploiting glottal and prosodic information for robust speaker verification

Yuan Fu Liao, Zhi Ren Zeng, Zi He Chen, Yau Tarng Juang

研究成果: 書貢獻/報告類型會議論文篇章同行評審

摘要

In this paper, three different levels of speaker cues including the glottal, prosodic and spectral information are integrated together to build a robust speaker verification system. The major purpose is to resist the distortion of channels and handsets. Especially, the dynamic behavior of normalized amplitude quotient (NAQ) and prosodic feature contours are modeled using Gaussian of mixture models (GMMs) and two latent prosody analyses (LPAs)-based approaches, respectively. The proposed methods are evaluated on the standard one speaker detection task of the 2001 NIST Speaker Recognition Evaluation Corpus where only one 2-minute training and 30-second trial speech (in average) are available. Experimental results have shown that the proposed approach could improve the equal error rates (EERs) of maximum a priori-adapted (MAP)-GMMs and GMMs+T-norm approaches from 12.4% and 9.5% to 10.3% and 8.3% and finally to 7.8%, respectively.

原文???core.languages.en_GB???
主出版物標題3rd International Conference on Speech Prosody 2006
編輯R. Hoffmann, H. Mixdorff
發行者International Speech Communications Association
ISBN(電子)9780000000002
出版狀態已出版 - 2006
事件3rd International Conference on Speech Prosody, SP 2006 - Dresden, Germany
持續時間: 2 5月 20065 5月 2006

出版系列

名字Proceedings of the International Conference on Speech Prosody
ISSN(列印)2333-2042

???event.eventtypes.event.conference???

???event.eventtypes.event.conference???3rd International Conference on Speech Prosody, SP 2006
國家/地區Germany
城市Dresden
期間2/05/065/05/06

指紋

深入研究「Exploiting glottal and prosodic information for robust speaker verification」主題。共同形成了獨特的指紋。

引用此