Audio Adversarial Examples: Attacks Using Vocal Masks
Tay, Kai Yuan, Ng, Lynnette, Chua, Wei Han, Loke, Lucerne, Ye, Danqi, Chua, Melissa
–arXiv.org Artificial Intelligence
We construct audio adversarial examples on automatic Speech-To-Text systems . Given any audio waveform, we produce an another by overlaying an audio vocal mask generated from the original audio. We apply our audio adversarial attack to five SOTA STT systems: DeepSpeech, Julius, Kaldi, wav2letter@anywhere and CMUSphinx. In addition, we engaged human annotators to transcribe the adversarial audio. Our experiments show that these adversarial examples fool State-Of-The-Art Speech-To-Text systems, yet humans are able to consistently pick out the speech. The feasibility of this attack introduces a new domain to study machine and human perception of speech.
arXiv.org Artificial Intelligence
Feb-5-2021
- Country:
- Asia (0.04)
- Europe > Italy
- Calabria > Catanzaro Province > Catanzaro (0.04)
- North America > United States
- New York > New York County > New York City (0.04)
- Genre:
- Research Report (0.82)
- Industry:
- Information Technology > Security & Privacy (1.00)
- Technology: