Adversarial optimization for dictionary attacks on speaker verification

Marras Mirko;Fenu Gianni
2019-01-01

Abstract

In this paper, we assess vulnerability of speaker verification systems to dictionary attacks. We seek master voices, i.e., adversarial utterances optimized to match against a large number of users by pure chance. First, we perform menagerie analysis to identify utterances which intrinsically hold this property. Then, we propose an adversarial optimization approach for generating master voices synthetically. Our experiments show that, even in the most secure configuration, on average, a master voice can match approx. 20% of females and 10% of males without any knowledge about the population. We demonstrate that dictionary attacks should be considered as a feasible threat model for sensitive and high-stakes deployments of speaker verification.
2019
Adversarial Examples; Authentication; Biometrics; Dictionary Attacks; Speaker Verification
Files in This Item:
File Size Format  
2430.pdf

Solo gestori archivio

Type: versione editoriale
Size 1.44 MB
Format Adobe PDF
1.44 MB Adobe PDF & nbsp; View / Open   Request a copy

Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

Questionnaire and social

Share on:
Impostazioni cookie