dc.contributor.author | Chettri, B | en_US |
dc.contributor.author | Stoller, D | en_US |
dc.contributor.author | Morfi, V | en_US |
dc.contributor.author | Martinez Ramirez, M | en_US |
dc.contributor.author | Benetos, E | en_US |
dc.contributor.author | Sturm, B | en_US |
dc.contributor.author | 20th Annual Conference of the International Speech Communication Association (INTERSPEECH 2019) | en_US |
dc.date.accessioned | 2019-07-11T09:38:02Z | |
dc.date.available | 2019-06-17 | en_US |
dc.date.issued | 2019-09-15 | en_US |
dc.identifier.uri | https://qmro.qmul.ac.uk/xmlui/handle/123456789/58459 | |
dc.description.abstract | Detecting spoofing attempts of automatic speaker verification (ASV) systems is challenging, especially when using only one modelling approach. For robustness, we use both deep neural networks and traditional machine learning models and combine them as ensemble models through logistic regression. They are trained to detect logical access (LA) and physical access (PA) attacks on the dataset released as part of the ASV Spoofing and Countermeasures Challenge 2019. We propose dataset partitions that ensure different attack types are present during training and validation to improve system robustness. Our ensemble model outperforms all our single models and the baselines from the challenge for both attack types. We investigate why some models on the PA dataset strongly outperform others and find that spoofed recordings in the dataset tend to have longer silences at the end than genuine ones. By removing them, the PA task becomes much more challenging, with the tandem detection cost function (t-DCF) of our best single model rising from 0.1672 to 0.5018 and equal error rate (EER) increasing from 5.98% to 19.8% on the development set. | en_US |
dc.format.extent | 1018 - 1022 | en_US |
dc.publisher | International Speech Communication Association (ISCA) | en_US |
dc.title | Ensemble Models for Spoofing Detection in Automatic Speaker Verification | en_US |
dc.type | Conference Proceeding | |
dc.rights.holder | © The Author(s) 2019 | |
pubs.notes | Not known | en_US |
pubs.publication-status | Accepted | en_US |
pubs.publisher-url | https://www.interspeech2019.org/ | en_US |
dcterms.dateAccepted | 2019-06-17 | en_US |
rioxxterms.funder | Default funder | en_US |
rioxxterms.identifier.project | Default project | en_US |
qmul.funder | A Machine Learning Framework for Audio Analysis and Retrieval::Royal Academy of Engineering | en_US |