Difference between revisions of "Publications/dehak.14.odyssey"
From LRDE
(5 intermediate revisions by the same user not shown) | |||
Line 2: | Line 2: | ||
| published = true |
| published = true |
||
| date = 2014-06-16 |
| date = 2014-06-16 |
||
− | | authors = |
+ | | authors = Najim Dehak, O Plchot, M H Bahari, L Burget, H Van hamme, Réda Dehak |
| title = GMM Weights Adaptation Based on Subspace Approaches for Speaker Verification |
| title = GMM Weights Adaptation Based on Subspace Approaches for Speaker Verification |
||
| booktitle = Odyssey 2014, The Speaker and Language Recognition Workshop |
| booktitle = Odyssey 2014, The Speaker and Language Recognition Workshop |
||
| address = Joensuu, Finland |
| address = Joensuu, Finland |
||
− | | |
+ | | lrdeprojects = SpeakerId |
| lrdenewsdate = 2014-06-16 |
| lrdenewsdate = 2014-06-16 |
||
| abstract = In this paper, we explored the use of Gaussian Mixture Model (GMM) weights adaptation for speaker verifica- tion. We compared two different subspace weight adap- tation approaches: Subspace Multinomial Model (SMM) and Non-Negative factor Analysis (NFA). Both techniques achieved similar results and seemed to outperform the retraining maximum likelihood (ML) weight adaptation. However, the training process for the NFA approach is substantially faster than the SMM technique. The i-vector fusion between each weight adaptation approach and the classical i-vector yielded slight improvements on the tele- phone part of the NIST 2010 Speaker Recognition Eval- uation dataset. |
| abstract = In this paper, we explored the use of Gaussian Mixture Model (GMM) weights adaptation for speaker verifica- tion. We compared two different subspace weight adap- tation approaches: Subspace Multinomial Model (SMM) and Non-Negative factor Analysis (NFA). Both techniques achieved similar results and seemed to outperform the retraining maximum likelihood (ML) weight adaptation. However, the training process for the NFA approach is substantially faster than the SMM technique. The i-vector fusion between each weight adaptation approach and the classical i-vector yielded slight improvements on the tele- phone part of the NIST 2010 Speaker Recognition Eval- uation dataset. |
||
Line 13: | Line 13: | ||
| id = dehak.14.odyssey |
| id = dehak.14.odyssey |
||
| bibtex = |
| bibtex = |
||
+ | @InProceedings<nowiki>{</nowiki> dehak.14.odyssey, |
||
+ | author = <nowiki>{</nowiki>Najim Dehak and O. Plchot and M.H. Bahari and L. Burget |
||
+ | and H. Van hamme and R\'eda Dehak<nowiki>}</nowiki>, |
||
+ | title = <nowiki>{</nowiki><nowiki>{</nowiki>GMM<nowiki>}</nowiki> Weights Adaptation Based on Subspace Approaches for |
||
+ | Speaker Verification<nowiki>}</nowiki>, |
||
+ | booktitle = <nowiki>{</nowiki>Odyssey 2014, The Speaker and Language Recognition |
||
+ | Workshop<nowiki>}</nowiki>, |
||
+ | year = 2014, |
||
+ | address = <nowiki>{</nowiki>Joensuu, Finland<nowiki>}</nowiki>, |
||
+ | month = jun, |
||
+ | abstract = <nowiki>{</nowiki>In this paper, we explored the use of Gaussian Mixture |
||
+ | Model (GMM) weights adaptation for speaker verifica- tion. |
||
+ | We compared two different subspace weight adap- tation |
||
+ | approaches: Subspace Multinomial Model (SMM) and |
||
+ | Non-Negative factor Analysis (NFA). Both techniques |
||
+ | achieved similar results and seemed to outperform the |
||
+ | retraining maximum likelihood (ML) weight adaptation. |
||
+ | However, the training process for the NFA approach is |
||
+ | substantially faster than the SMM technique. The i-vector |
||
+ | fusion between each weight adaptation approach and the |
||
+ | classical i-vector yielded slight improvements on the tele- |
||
+ | phone part of the NIST 2010 Speaker Recognition Eval- |
||
+ | uation dataset.<nowiki>}</nowiki>, |
||
+ | pages = <nowiki>{</nowiki>48--53<nowiki>}</nowiki> |
||
+ | <nowiki>}</nowiki> |
||
+ | |||
}} |
}} |
Latest revision as of 17:00, 27 May 2021
- Authors
- Najim Dehak, O Plchot, M H Bahari, L Burget, H Van hamme, Réda Dehak
- Where
- Odyssey 2014, The Speaker and Language Recognition Workshop
- Place
- Joensuu, Finland
- Type
- inproceedings
- Projects
- SpeakerId"SpeakerId" is not in the list (Vaucanson, Spot, URBI, Olena, APMC, Tiger, Climb, Speaker ID, Transformers, Bison, ...) of allowed values for the "Related project" property.
- Date
- 2014-06-16
Abstract
In this paper, we explored the use of Gaussian Mixture Model (GMM) weights adaptation for speaker verifica- tion. We compared two different subspace weight adap- tation approaches: Subspace Multinomial Model (SMM) and Non-Negative factor Analysis (NFA). Both techniques achieved similar results and seemed to outperform the retraining maximum likelihood (ML) weight adaptation. However, the training process for the NFA approach is substantially faster than the SMM technique. The i-vector fusion between each weight adaptation approach and the classical i-vector yielded slight improvements on the tele- phone part of the NIST 2010 Speaker Recognition Eval- uation dataset.
Bibtex (lrde.bib)
@InProceedings{ dehak.14.odyssey, author = {Najim Dehak and O. Plchot and M.H. Bahari and L. Burget and H. Van hamme and R\'eda Dehak}, title = {{GMM} Weights Adaptation Based on Subspace Approaches for Speaker Verification}, booktitle = {Odyssey 2014, The Speaker and Language Recognition Workshop}, year = 2014, address = {Joensuu, Finland}, month = jun, abstract = {In this paper, we explored the use of Gaussian Mixture Model (GMM) weights adaptation for speaker verifica- tion. We compared two different subspace weight adap- tation approaches: Subspace Multinomial Model (SMM) and Non-Negative factor Analysis (NFA). Both techniques achieved similar results and seemed to outperform the retraining maximum likelihood (ML) weight adaptation. However, the training process for the NFA approach is substantially faster than the SMM technique. The i-vector fusion between each weight adaptation approach and the classical i-vector yielded slight improvements on the tele- phone part of the NIST 2010 Speaker Recognition Eval- uation dataset.}, pages = {48--53} }