dc.creator | Ferrer, Luciana | |
dc.creator | Nandwana, Mahesh Kumar | |
dc.creator | McLaren, Mitchell | |
dc.creator | Castan, Diego | |
dc.creator | Lawson, Aaron | |
dc.date.accessioned | 2021-01-21T15:57:38Z | |
dc.date.accessioned | 2022-10-15T14:56:10Z | |
dc.date.available | 2021-01-21T15:57:38Z | |
dc.date.available | 2022-10-15T14:56:10Z | |
dc.date.created | 2021-01-21T15:57:38Z | |
dc.date.issued | 2019-01 | |
dc.identifier | Ferrer, Luciana; Nandwana, Mahesh Kumar; McLaren, Mitchell; Castan, Diego; Lawson, Aaron; Toward Fail-Safe Speaker Recognition: Trial-Based Calibration with a Reject Option; Institute of Electrical and Electronics Engineers; IEEE/ACM Transactions on Audio Speech and Language Processing; 27; 1; 1-2019; 140-153 | |
dc.identifier | 2329-9290 | |
dc.identifier | http://hdl.handle.net/11336/123318 | |
dc.identifier | CONICET Digital | |
dc.identifier | CONICET | |
dc.identifier.uri | https://repositorioslatinoamericanos.uchile.cl/handle/2250/4399384 | |
dc.description.abstract | The output scores of most of the speaker recognition systems are not directly interpretable as stand-alone values. For this reason, a calibration step is usually performed on the scores to convert them into proper likelihood ratios, which have a clear probabilistic interpretation. The standard calibration approach transforms the system scores using a linear function trained using data selected to closely match the evaluation conditions. This selection, though, is not feasible when the evaluation conditions are unknown. In previous work, we proposed a calibration approach for this scenario called trial-based calibration (TBC). TBC trains a separate calibration model for each test trial using data that is dynamically selected from a candidate training set to match the conditions of the trial. In this work, we extend the TBC method, proposing: 1) a new similarity metric for selecting training data that result in significant gains over the one proposed in the original work; 2) a new option that enables the system to reject a trial when not enough matched data are available for training the calibration model; and 3) the use of regularization to improve the robustness of the calibration models trained for each trial. We test the proposed algorithms on a development set composed of several conditions and on the Federal Bureau of Investigation multi-condition speaker recognition dataset, and we demonstrate that the proposed approach reduces calibration loss to values close to 0 for most of the conditions when matched calibration data are available for selection, and that it can reject most of the trials for which relevant calibration data are unavailable. | |
dc.language | eng | |
dc.publisher | Institute of Electrical and Electronics Engineers | |
dc.relation | info:eu-repo/semantics/altIdentifier/url/https://ieeexplore.ieee.org/document/8490592 | |
dc.relation | info:eu-repo/semantics/altIdentifier/doi/http://dx.doi.org/10.1109/TASLP.2018.2875794 | |
dc.rights | https://creativecommons.org/licenses/by-nc-sa/2.5/ar/ | |
dc.rights | info:eu-repo/semantics/openAccess | |
dc.subject | FORENSIC VOICE COMPARISON | |
dc.subject | SPEAKER RECOGNITION | |
dc.subject | TRIAL-BASED CALIBRATION | |
dc.title | Toward Fail-Safe Speaker Recognition: Trial-Based Calibration with a Reject Option | |
dc.type | info:eu-repo/semantics/article | |
dc.type | info:ar-repo/semantics/artículo | |
dc.type | info:eu-repo/semantics/publishedVersion | |