Interspeech 2019 2019
DOI: 10.21437/interspeech.2019-1353
|View full text |Cite
|
Sign up to set email alerts
|

Universal Adversarial Perturbations for Speech Recognition Systems

Abstract: In this work, we demonstrate the existence of universal adversarial audio perturbations that cause mis-transcription of audio signals by automatic speech recognition (ASR) systems. We propose an algorithm to find a single quasi-imperceptible perturbation, which when added to any arbitrary speech signal, will most likely fool the victim speech recognition model. Our experiments demonstrate the application of our proposed technique by crafting audio-agnostic universal perturbations for the state-of-the-art ASR s… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
36
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
3
3

Relationship

0
10

Authors

Journals

citations
Cited by 83 publications
(36 citation statements)
references
References 40 publications
0
36
0
Order By: Relevance
“…A closely related topic are adversarial attacks, first investigated by Szegedy et al (2013) and Goodfellow et al (2015) in computer vision and later extended to text classification (Papernot et al, 2016;Ebrahimi et al, 2018b;Li et al, 2018;Hosseini et al, 2017) and translation (Ebrahimi et al, 2018a;Michel et al, 2019). Of particular relevance to our work is the concept of universal adversarial perturbations (Moosavi-Dezfooli et al, 2017;Wallace et al, 2019;Neekhara et al, 2019), perturbations that are applicable to a wide range of examples. Specifically the adversarial triggers from Wallace et al (2019) are reminiscent of the attack proposed here, with the crucial difference that their attack fixes the model's weights and finds a specific trigger, whereas the attack we explore fixes the trigger and changes the model's weights to introduce a specific response.…”
Section: Related Workmentioning
confidence: 94%
“…A closely related topic are adversarial attacks, first investigated by Szegedy et al (2013) and Goodfellow et al (2015) in computer vision and later extended to text classification (Papernot et al, 2016;Ebrahimi et al, 2018b;Li et al, 2018;Hosseini et al, 2017) and translation (Ebrahimi et al, 2018a;Michel et al, 2019). Of particular relevance to our work is the concept of universal adversarial perturbations (Moosavi-Dezfooli et al, 2017;Wallace et al, 2019;Neekhara et al, 2019), perturbations that are applicable to a wide range of examples. Specifically the adversarial triggers from Wallace et al (2019) are reminiscent of the attack proposed here, with the crucial difference that their attack fixes the model's weights and finds a specific trigger, whereas the attack we explore fixes the trigger and changes the model's weights to introduce a specific response.…”
Section: Related Workmentioning
confidence: 94%
“…Such a technique is often referred to as perturbations. Further, a single noise input can cause false recognition of any speech input, and it is termed as universal perturbation [76], [77]. Adversarial attacks can be non-targeted and targeted.…”
Section: ) Adversarial Perturbationsmentioning
confidence: 99%
“…However, all the aforementioned ASR adversarial attacks are individual attack through solving an optimization problem for each individual input audio, which needs high run-time requirements (e.g., several hours) to compute the adversarial examples per input audio. Alternatively, a more recent work [12] produces a single universal perturbation which can fool ASR systems causing an error in transcription. This work is in the case of untargeted attack, in which the adversary cannot specify the expected speech transcription during the phase of adversary example generation.…”
Section: Related Workmentioning
confidence: 99%