Difference between revisions of "Publications/dehak.14.odyssey"

From LRDE

(Created page with "{{Publication | published = true | date = 2014-06-01 | authors = N Dehak, O Plchot, M H Bahari, L Burget, H Van hamme, R Dehak | title = GMM Weights Adaptation Based on Subspa...")
 
Line 1: Line 1:
 
{{Publication
 
{{Publication
 
| published = true
 
| published = true
| date = 2014-06-01
+
| date = 2014-06-16
 
| authors = N Dehak, O Plchot, M H Bahari, L Burget, H Van hamme, R Dehak
 
| authors = N Dehak, O Plchot, M H Bahari, L Burget, H Van hamme, R Dehak
 
| title = GMM Weights Adaptation Based on Subspace Approaches for Speaker Verification
 
| title = GMM Weights Adaptation Based on Subspace Approaches for Speaker Verification
Line 7: Line 7:
 
| address = Joensuu, Finland
 
| address = Joensuu, Finland
 
| project = SpeakerId
 
| project = SpeakerId
  +
| lrdenewsdate = 2014-06-16
| urllrde = 201406-ODYSSEY
 
 
| abstract = In this paper, we explored the use of Gaussian Mixture Model (GMM) weights adaptation for speaker verifica- tion. We compared two different subspace weight adap- tation approaches: Subspace Multinomial Model (SMM) and Non-Negative factor Analysis (NFA). Both techniques achieved similar results and seemed to outperform the retraining maximum likelihood (ML) weight adaptation. However, the training process for the NFA approach is substantially faster than the SMM technique. The i-vector fusion between each weight adaptation approach and the classical i-vector yielded slight improvements on the tele- phone part of the NIST 2010 Speaker Recognition Eval- uation dataset.
 
| abstract = In this paper, we explored the use of Gaussian Mixture Model (GMM) weights adaptation for speaker verifica- tion. We compared two different subspace weight adap- tation approaches: Subspace Multinomial Model (SMM) and Non-Negative factor Analysis (NFA). Both techniques achieved similar results and seemed to outperform the retraining maximum likelihood (ML) weight adaptation. However, the training process for the NFA approach is substantially faster than the SMM technique. The i-vector fusion between each weight adaptation approach and the classical i-vector yielded slight improvements on the tele- phone part of the NIST 2010 Speaker Recognition Eval- uation dataset.
 
| pages = 48 to 53
 
| pages = 48 to 53
Line 13: Line 13:
 
| id = dehak.14.odyssey
 
| id = dehak.14.odyssey
 
| bibtex =
 
| bibtex =
@InProceedings<nowiki>{</nowiki> dehak.14.odyssey,
 
author = <nowiki>{</nowiki>N. Dehak and O. Plchot and M.H. Bahari and L. Burget and
 
H. Van hamme and R. Dehak<nowiki>}</nowiki>,
 
title = <nowiki>{</nowiki>GMM Weights Adaptation Based on Subspace Approaches for
 
Speaker Verification<nowiki>}</nowiki>,
 
booktitle = <nowiki>{</nowiki>Odyssey 2014, The Speaker and Language Recognition
 
Workshop<nowiki>}</nowiki>,
 
year = 2014,
 
address = <nowiki>{</nowiki>Joensuu, Finland<nowiki>}</nowiki>,
 
month = jun,
 
project = <nowiki>{</nowiki>SpeakerId<nowiki>}</nowiki>,
 
abstract = <nowiki>{</nowiki>In this paper, we explored the use of Gaussian Mixture
 
Model (GMM) weights adaptation for speaker verifica- tion.
 
We compared two different subspace weight adap- tation
 
approaches: Subspace Multinomial Model (SMM) and
 
Non-Negative factor Analysis (NFA). Both techniques
 
achieved similar results and seemed to outperform the
 
retraining maximum likelihood (ML) weight adaptation.
 
However, the training process for the NFA approach is
 
substantially faster than the SMM technique. The i-vector
 
fusion between each weight adaptation approach and the
 
classical i-vector yielded slight improvements on the tele-
 
phone part of the NIST 2010 Speaker Recognition Eval-
 
uation dataset.<nowiki>}</nowiki>,
 
pages = <nowiki>{</nowiki>48--53<nowiki>}</nowiki>
 
<nowiki>}</nowiki>
 
 
 
}}
 
}}

Revision as of 09:00, 27 June 2014

Abstract

In this paper, we explored the use of Gaussian Mixture Model (GMM) weights adaptation for speaker verifica- tion. We compared two different subspace weight adap- tation approaches: Subspace Multinomial Model (SMM) and Non-Negative factor Analysis (NFA). Both techniques achieved similar results and seemed to outperform the retraining maximum likelihood (ML) weight adaptation. However, the training process for the NFA approach is substantially faster than the SMM technique. The i-vector fusion between each weight adaptation approach and the classical i-vector yielded slight improvements on the tele- phone part of the NIST 2010 Speaker Recognition Eval- uation dataset.