Next Article in Journal
TM-Net: A Neural Net Architecture for Tone Mapping
Previous Article in Journal
Disease Recognition in X-ray Images with Doctor Consultation-Inspired Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Detecting Audio Adversarial Examples in Automatic Speech Recognition Systems Using Decision Boundary Patterns

1
Institute of Cybersecurity and Cryptology, School of Computing and Information Technology, University of Wollongong, Wollongong, NSW 2522, Australia
2
Department of Cyber Security, Ewha Womans University, Seoul 03760, Republic of Korea
*
Author to whom correspondence should be addressed.
J. Imaging 2022, 8(12), 324; https://doi.org/10.3390/jimaging8120324
Submission received: 2 November 2022 / Revised: 3 December 2022 / Accepted: 6 December 2022 / Published: 9 December 2022
(This article belongs to the Section Visualization and Computer Graphics)

Abstract

:
Automatic Speech Recognition (ASR) systems are ubiquitous in various commercial applications. These systems typically rely on machine learning techniques for transcribing voice commands into text for further processing. Despite their success in many applications, audio Adversarial Examples (AEs) have emerged as a major security threat to ASR systems. This is because audio AEs are able to fool ASR models into producing incorrect results. While researchers have investigated methods for defending against audio AEs, the intrinsic properties of AEs and benign audio are not well studied. The work in this paper shows that the machine learning decision boundary patterns around audio AEs and benign audio are fundamentally different. Using dimensionality-reduction techniques, this work shows that these different patterns can be visually distinguished in two-dimensional (2D) space. This in turn allows for the detection of audio AEs using anomal- detection methods.

1. Introduction

Automatic Speech Recognition (ASR) systems are commonly used in many commercial applications. These systems are typically used to transcribe user speech into text and for recognizing user voice commands. Modern ASR systems rely on deep learning techniques, which have been shown to achieve better speech recognition performance in comparison with other traditional techniques [1,2,3]. Nevertheless, despite its success, deep learning techniques suffer from a variety of security threats [4]. Among these, Adversarial Examples (AEs) have emerged as a security threat that has attracted great interest within the research community.
Research on AEs first appeared in the image-recognition field, where small perturbations were applied to a benign (normal) image to generate an AE [5]. The goal of an AE is to fool deep learning models into classifying it under a different label while being perceived to be visually indistinguishable from the original image by a human. Research interest in AEs has rapidly spread to other areas, such as Natural Language Processing (NLP) [6,7,8], speech recognition [9,10,11,12,13,14], and speaker verification [15,16,17].
In conjunction with research on generating AEs, others have attempted to understand and explain the reasons for AEs. For example, Tsipras et al. [18] provided a provable demonstration, which showed that non-robust features referred to features that were weakly correlated with the corresponding label. They reasoned that AEs exist because classification is affected by such non-robust features in a data set. Ilyas et al. [19] affirmed the intrinsic existence of non-robust features in data sets. Others have also shown that perturbations in AEs can dominate classification. This can be seen as complementary to the presence of non-robust features in a data set [20].
Other researchers have focused on methods for defending against AEs. For instance, using intrinsic properties to differentiate between benign samples and AEs [21,22] and training of a classifier with both benign samples and AEs to detect previously unknown attacks [23,24]. Despite the fact that adversarial training was shown to be one of the most effective methods for defending against AEs [25], Zhang et al. [26] showed that successful AEs with small perturbations can still be generated if data points are far from the training set manifold.
While much research on AEs is in the image recognition field, this paper focuses on AEs in the audio domain. In the audio domain, while researchers have proposed methods for detecting audio AEs [27,28], the fundamental differences between audio AEs and benign audio are not well studied or understood. To date, there is a lack of research on visually analyzing audio AEs in relation to ASR systems. This paper investigates this by presenting a method of visually analyzing the intrinsic properties of audio AEs. We show that these properties can be used to differentiate audio AEs from benign audio. Consequently, this research demonstrates that it is possible to detect previously unknown audio AEs using these distinguishable features.
Our Contributions: This paper is an extended version of our work in [29]. We demonstrate that the decision boundaries around audio AEs are fundamentally different from the decision boundaries around benign audio. In our proposed method, we use heat maps to visualize the decision boundaries of ASR models in relation to changes in loss function values and normalized edit distances (Levenshtein distance). Using this approach, we show that both targeted and untargeted audio AEs have different decision boundary patterns in comparison with benign audio. By extracting features from decision boundaries and projecting them into two-dimensional (2D) space, this paper illustrates that targeted and untargeted AEs can clearly be separated from benign audio. As a result of this, we demonstrate the feasibility of using simple anomaly detection models to distinguish between AEs and benign audio.
The rest of this paper is organized as follows. Section 2 discusses related work that provides a background to our work. Our proposed method is described in Section 3. Section 4 discusses methods of generating attacks for detection, which is followed by the results of our experiments in Section 5. Section 6 discusses the robustness of AEs against our method, along with a promising direction for eliminating AEs. Finally, our conclusions and future work are presented in Section 7.

2. Related Work

In this section, we provide a discussion on related work, which provides a background to our work. This section starts by describing audio AEs. This is followed by the work on defending against such AEs. Then, it presents current techniques for visually analyzing AEs.

2.1. Audio Adversarial Examples

AEs were first presented in the image-recognition field [5] and can be categorized into targeted or untargeted AEs [30]. The difference between targeted and untargeted is that targeted AEs fool a model into outputting a predetermined result, whereas untargeted AEs simply cause a model to output an incorrect result. AEs can be generated under a white-box or black-box threat model. A white-box threat model assumes that an adversary knows everything about the target model. This includes its training data set, hyper-parameters, model weights, and so on. On the other hand, under a black-box threat model, an adversary is only able to obtain input and output pairs consisting of the AE and its corresponding result. Thus, black-box AEs are a subset of white-box AEs [31]. Interest in AEs has spread from the image domain to ASR systems. A summary of research efforts on audio AEs is presented in Table 1.
In early work on white-box audio AE generation by Yuan et al. [32], they hid malicious voice commands in songs to attack the Kaldi speech-recognition toolkit [33]. This work also demonstrated the transferability of the techniques whereby the generated AEs could be transferred to attack iFLYREC (https://www.iflyrec.com/, accessed on 1 November 2020). Transferability refers to the feature where an AE generated using one model is able to fool other models. Kaldi is a hybrid ASR model, i.e., Deep Neural Network–Hidden Markov Model (DNN-HMM), that outperformed traditional end-to-end DNN models. Carlini and Wagner [9] proposed a white-box method of generating audio AEs against DeepSpeech by optimizing the Connectionist Temporal Classification (CTC) loss. CTC loss is used for training sequence-to-sequence neural networks with unknown alignment between input and output sequences [34]. A limitation of their approach is that the max-norm of perturbations is used to reduce noise in the resulting audio AEs. Liu et al. [35] improved the quality and generation speed of the AE generation method proposed in [9] by partially optimizing perturbations. Other studies [10,36] have shown that there are better ways to suppress noise by incorporating psychoacoustics into the generation process. However, a study by Eisenhofer et al. [37] showed that a defender can deliberately remove inaudible components from input audio to avoid imperceptible adversarial perturbations. Furthermore, recent work in Zong et al. [38] demonstrated that high-quality audio AEs can be obtained without psychoacoustics.
Table 1. A summary of research efforts on audio AEs in ASR systems.
Table 1. A summary of research efforts on audio AEs in ASR systems.
AE TypeAssumptionMethodTarget Model
TargetedWhite-boxYuan et al. [32]Kaldi
Carlini and Wagner [9]DeepSpeech
Liu el al. [35]DeepSpeech
Schoenherr et al. [10]Kaldi
Qin et al. [36]Lingvo [3]
Zong el al. [38]DeepSpeech
Black-boxTaori et al. [39]DeepSpeech
Chen et al. [40]Commercial products *
Untargetedwhite-boxNeekhara et al. [41]DeepSpeech
Black-boxAbdullah et al. [42]7 models #
* including Google Assistant, Google Home, Microsoft Cortana, Amazon Echo, and IBM Speech to Text. # including Google (Normal), Google (Phone), Facebook Wit, DeepSpeech, DeepSpeech2, CMU Sphinx, and Microsoft Azure.
The generation of black-box audio AEs is more challenging compared with their white-box counterparts because the internal workings of the ASR model are not accessible to the adversary. Alzantot et al. [11] were the first to use genetic algorithms to generate black-box audio AEs. However, the target model in their study was a lightweight keyword spotting model, rather than an ASR model. In later work, Taori et al. [39] proposed black-box audio AEs against DeepSpeech. In addition to genetic algorithms, they also employed a gradient estimation technique to fine-tune perturbations when the edit distance between the transcribed and target phrases was small. Target phrases in their work were limited to two words. Chen et al. [40] proposed training a local surrogate model in which the decision boundaries resembled a target model. Audio AEs were generated using the surrogate model to attack a remotely deployed target. They demonstrated the success of their method in forcing commercial ASR products to output predefined commands.
Unlike targeted AEs, which output predefined commands, untargeted AEs are less malicious because they only cause a target model to output incorrect commands. Hence, untargeted audio AEs have received less attention in the research community. Under a white-box assumption, Neekhara et al. [41] proposed Universal Adversarial Perturbations (UAPs) that can be applied to any input audio to cause incorrect output. They empirically showed that their UAPs were more effective than random noise. Other work by Abdullah et al. [42] assumed black-box access to a target model. In particular, they proposed first decomposing input audio and then iteratively optimizing a threshold to eliminate components. The end goal was to identify an optimal threshold that preserves the perceptual quality of input audio while making a target model output incorrect results. They empirically showed that their method was able to cause commercial products to incorrectly transcribe input audio.

2.2. Defending against Audio Adversarial Examples

As audio AEs are a serious threat to the security of ASR systems, the past few years have witnessed an increase in the amount of research on defending against audio AEs. In this subsection, we present an overview of such techniques. It should be mentioned that the defense for other audio tasks, such as environmental sound classification [43], is beyond the scope of this paper.
There are two lines of work for defending against audio AEs, which are summarized in Table 2. The first line of work focuses on detecting the existence of adversarial perturbations in input audio. Zeng et al. [27] proposed the use of multiple ASR models to transcribe an input audio signal. If the resulting transcripts of these models diverged significantly, the audio would be classified as an AE. Their detection method is based on the assumption that audio AEs cannot be transferred between multiple ASR models. On the other hand, this method requires the deployment of multiple ASR systems, which may not be practical in real-world applications. Another defense method, proposed by Yang et al. [28], detects audio AEs based on temporal dependency. They observed that unlike benign audio, audio AEs cannot preserve temporal dependency. Specifically, they observed that the transcripts of partial audio AEs, e.g., the first half of an AE, were significantly different compared with the transcripts of the full AEs. In contrast, there was significant overlapping between the transcripts of partial and full input for benign audio.
Recently, Samizade et al. [44] proposed a defense method where they trained a Convolutional Neural Network (CNN) on the spectrograms of benign audio and AEs and demonstrated that it could detect audio AEs with high accuracy. Nonetheless, their method suffered from low performance when detecting previously unknown attacks. This is due to the lack of generalization in out-of-distribution data. Improving the performance for out-of-distribution data is still an active topic in the deep learning literature [49]. Guo et al. [45] proposed another audio AE detection method, which was based on an efficient multivariant partition based method for extracting features. Although their method demonstrated high performance in detecting audio AEs, the requirement of training a classifier on both attacks and clean audio may make it impractical when attacks are unknown in advance. Another recent study by Hussain et al. [46] proposed a framework for detecting audio AEs. The underlying idea behind their method is that transcripts of audio AEs will be significantly altered if small modifications to the audio are introduced, whereas transcripts of clean audio input will be stable. They thus proposed applying transformations to input audio, and an audio AE was detected if its transcript changed significantly. Although the methods discussed above have successfully detected audio AEs, the intrinsic properties that differentiate audio AEs from benign audio are not well studied or understood.
The other line of work in this area focuses on recovering the original clean audio by destroying adversarial perturbations in audio AEs. Yang et al. [47] used a speech-quality-enhancement network to preprocess audio AEs. After adversarial training, this network can efficiently remove adversarial perturbations. However, adversarial training introduces a tradeoff between model performance and robustness [50], which may limit the performance of the speech-quality-enhancement network. Guo et al. [48] observed that denoising techniques resulted in different performances after removing adversarial perturbations in different audio AEs. Hence, they proposed an intelligent noise-reduction method, called INOR, which was effective in removing adversarial perturbations from different audio AEs. Yuan et al. [32] and Chen et al. [40] observed that audio downsampling was able to destroy adversarial perturbations. Nonetheless, they performed experiments on their proposed audio AEs.

2.3. Visualization Techniques for Analyzing Adversarial Examples

Visualization is useful for facilitating the understanding of deep learning techniques [51]. Several visualization efforts have focused on aiding in the intuitive understanding of AE properties. In an initial attempt, Norton et al. [52] built a web-based interface to interactively show the image AE generation process. Seminal work conducted by Liu et al. [53] visually explained the transferability of image AEs. In their work, they visualized the decision boundaries of several image recognition models and found that AEs could be transferred due to their overlapping decision boundaries. In other work, Stutz et al. [54] showed that image AE perturbations are interpretable if the AEs are constrained on the manifold of a data set. In addition, different patterns in the loss function gradients of input images in non-robust and robust models have been visually compared [18]. A study by Zhang et al. [20] visualized logit vectors of a model in relation to an image AE, along with its corresponding clean image and perturbations. The experimental results in their study showed that logit vectors of an image AE and their corresponding perturbations are correlated.
Despite these research efforts on visualization for AEs, to date, there is limited research on the visual analysis of audio AEs in the ASR domain. This paper fills this gap by proposing a method of visually analyzing the intrinsic properties that can be used to distinguish audio AEs from benign audio.

3. Proposed Method

The research in this paper proposes a method of visually analyzing targeted and untargeted audio AEs. This section presents the details of our proposed method.

3.1. Visualizing Decision Boundaries

In general, benign audio is significantly more robust than audio AEs. Robustness refers to whether audio can be transcribed correctly despite the presence of noise. Benign audio is generally more robust as it can usually be transcribed correctly even when additional noise is added to the audio signal. This implies that the decision boundary patterns around benign audio are potentially different from those of audio AEs. Hence, we propose a method of visualizing the decision boundaries of ASR models to show this difference.
Unlike in image recognition, where there are usually fixed sets of labels, the decision boundaries of ASR models are more difficult to visualize as an audio signal can potentially be transcribed into a large number of output strings. Moreover, if one were to simply treat different transcripts as different labels, the visualization results would be confusing. This is because a difference between labels cannot appropriately represent the difference in the transcribed text. For example, if “paper” and “papers” were to be treated as two different labels, such as 1 and 2 in numeric form, information on the similarity between these two transcripts will be lost. Therefore, it makes more sense to visualize the decision boundaries of an ASR model via changes in the resulting transcripts when input audio is modified. Moreover, changes in the loss function values can also be used to represent the decision boundary patterns of an ASR model.
In this paper, we propose a method of visualizing the decision boundaries of ASR models using heat maps to show changes in loss function values and changes in normalized edit distances. The reason for using heat maps is that they can clearly represent changes in values visually. The proposed method is formally defined here. Let x be the input audio and y be the transcript of x, which may be different from the ground truth if the audio signal is incorrectly transcribed; let f ( ) be the ASR model; and let n e t ( ) be the corresponding loss function. We can calculate the gradient of the loss function with respect to x as g = n e t ( f ( x ) , y ) and normalize it to be of unit length using g ¯ = g | | g | | 2 . Then, we initialize a random unit vector q ¯ that is not parallel to g ¯ . We obtain p = q ¯ ( q ¯ · g ¯ ) × g ¯ and p ¯ = p | | p | | 2 . Thus, p ¯ is of unit length and perpendicular to g ¯ .
The heat maps of loss function values and normalized edit distances are defined as the square matrices M l o s s and M e d i t , respectively. The size of both matrices is n × n . Let s be a predefined number controlling the extent to which x is modified. The definition of M l o s s and M e d i t is shown in Equation (1).
[ M l o s s ] i , j = n e t ( f ( x + s i · p ¯ + s j · g ¯ ) , y ) [ M e d i t ] i , j = d e d i t ( f ( x + s i · p ¯ + s j · g ¯ ) , y ) h l e n g t h ( y )
where s k = n · s 2 + n · s · ( k 1 ) ( n 1 ) , d e d i t ( ) is the function to calculate the edit distance and h l e n g t h ( ) returns the transcript length, which is used to normalize the edit distance. In Equation (1), the audio data are evenly modified along g ¯ and p ¯ via a step size s. Normalizing the edit distance is necessary because edit distance by itself cannot fairly compare the change in transcripts for different y. For example, a small edit distance means a more severe change for short y than for long y.

3.2. Feature Extraction

For additional insight into the heat maps, two-dimensional (2D) reduction techniques were used to project the results into 2D space to identify potential patterns. For this, we used the Principal Component Analysis (PCA) [55] and the t-distributed Stochastic Neighbor Embedding (t-SNE) [56] techniques. A simple feature extraction method was employed in this study. While more advanced methods, such as training a convolutional neural network on heat maps, may potentially produce better results, a simple method serves as a low bound that can be improved on. An investigation into feature extraction methods for the proposed method is a potential direction for future work.
Given input audio, we calculate a vector v l o s s based on the change in loss function values if we modify the audio along the gradient direction g ¯ and a perpendicular direction p ¯ . g ¯ and p ¯ were previously defined in Section 3.1. Similarly, we calculate a vector v e d i t based on the change in normalized edit distances. As defined in Equation (2), the feature vector v f t representing the heat maps of an input audio is simply a concatenation of v l o s s and v e d i t . In other words, v f t measures both the change in loss function values and normalized edit distances when input audio is modified. Intuitively, using v f t would result in better performance in distinguishing audio AEs from benign audio than using only v l o s s or v e d i t . For simplicity, we refer to v f t as the features of input audio.
v l o s s = n e t ( f ( x + g ¯ ) , y ) n e t ( f ( x ) , y ) n e t ( f ( x g ¯ ) , y ) n e t ( f ( x ) , y ) n e t ( f ( x + p ¯ ) , y ) n e t ( f ( x ) , y ) n e t ( f ( x p ¯ ) , y ) n e t ( f ( x ) , y ) v e d i t = d e d i t ( f ( x + g ¯ ) , y ) / h l e n g t h ( y ) d e d i t ( f ( x g ¯ ) , y ) / h l e n g t h ( y ) d e d i t ( f ( x + p ¯ ) , y ) / h l e n g t h ( y ) d e d i t ( f ( x p ¯ ) , y ) / h l e n g t h ( y ) v f t = v l o s s v e d i t

4. Attack Generation

This section discusses the method of generating AEs for detection.

4.1. Targeted Audio Adversarial Examples

This paper analyzes an improved version of the state-of-the-art targeted audio AE generation process proposed by Carlini and Wagner [9]. In their research, distortion caused by perturbations was measured by comparing the level of perturbations δ , in decibels (dB), with the original waveform x. The calculation is given as d B x ( δ ) = d B ( δ ) d B ( x ) , where d B ( x ) = max i 20 · l o g 10 ( x i ) , which is used in the formulation shown in Equation (3) [9].
m i n i m i z e | | δ | | 2 2 + c · n e t ( f ( x + δ ) , y ) s u c h t h a t d B x ( δ ) τ
where τ limits the max-norm of δ , | | δ | | 2 2 is the squared Euclidean norm of δ , f ( ) represents the ASR model, y is the target phrase, n e t ( ) represents the loss function of the ASR model, and c is used as a trade-off between the amount of adversarial perturbation and making δ  small.
A major drawback of this method is that perturbations are limited by the max-norm, which is arguably not suitable for minimizing noise in audio AEs. This is because max-norm constrained perturbations are applied in a non-selective manner, such that noise is clearly audible during quiet sections. In contrast, Qin et al. [36] showed that it is more appropriate to incorporate psychoacoustics to suppress noise in audio AEs. Using their approach, they divide the generation process into two stages. In the first stage, a targeted audio AE is generated in the same way as [9]. Then, the second stage tries to limit perturbations to be under the masking threshold that was proposed in [57]. The formulation to solve this is shown in Equation (4) [36], where l θ ( ) is the loss function to calculate the hinge loss of the masking threshold and α controls the trade-off between the amount of adversarial perturbation and it being imperceptible.
m i n i m i z e n e t ( f ( x + δ ) , y ) + α · l θ ( x , δ )
It should be noted that limiting the max-norm of perturbations in stage 1 is somewhat unnecessary since the original purpose of limiting the max-norm is to suppress noise. Furthermore, in their approach, noise suppression is also done in stage 2.
Based on the method in Qin et al. [36], we improved the targeted AEs generation process proposed by Carlini and Wagner [9] by constraining perturbations via the masking threshold instead of the max-norm. Specifically, we solve the formula in Equation (5), where X represents the set of valid audio data, | | δ | | 2 2 is the squared Euclidean norm of δ , n e t is the loss function of the ASR model, l θ is the hinge loss of the masking threshold from [36], and β and α are factors used to balance the different losses. There are still two stages. During stage 1, a targeted audio AEs is generated with α set to 0, so that the hinge loss of the masking threshold will have no contribution. During stage 2, α is set to a small value, e.g., 0.05, to suppress noise.
As asserted in [9], limiting the max-norm of perturbations would often result in the optimization not converging, but rather oscillating around a solution. In contrast, we do not limit the max-norm of perturbations in Equation (5), thereby potentially reducing AE generation time.
m i n i m i z e | | δ | | 2 2 + β · n e t ( f ( x + δ ) , y ) + α · l θ ( x , δ ) s u c h t h a t x + δ X

4.2. Untargeted Audio Adversarial Examples

To the best of our knowledge, to date, there is limited research on untargeted audio AEs. This is because untargeted audio AEs are less interesting compared to targeted AEs, since they only lead to wrong or even meaningless transcripts. Nevertheless, for the sake of completeness, we also analyze untargeted audio AEs in this research.
We devised two approaches to generate untargeted audio AEs. The first is based on the Fast Gradient Sign Method (FGSM) [25]. This method simply takes one step along the gradient of the loss function with respect to the input audio. Perturbations δ are calculated as in Equation (6) [25], where x is the input audio, y is the target phrase, n e t ( ) is the loss function, and ϵ is the step size.
δ = ϵ · s i g n ( x n e t ( f ( x ) , y ) )
An audio AE x is then calculated as: x x δ . While this will not generate targeted audio AEs, like the method in [9], this method can generate untargeted audio AEs if we set y to be the reversed ground truth. The reversed ground truth is typically different from the original. An untargeted AE is successfully generated if the edit distance between the transcript and the ground truth exceeds a certain threshold. Edit distance is defined as the minimum number of letter-level modifications, including insertions, deletions, and substitutions, required to change one text to another.
The second approach to generating untargeted audio AEs was inspired by the black-box targeted audio AE proposed by Taori et al. [39], where they used a genetic algorithm to search for perturbations that led to an ASR outputting a target phrase. When the transcript of the best solution is within a predefined edit distance of the target phrase, the generation process uses a gradient estimation strategy to continue the search process. In this work, we use the gradient estimation strategy in [39] to generate untargeted audio AEs. We also incorporate the noise suppression technique from [36] in the generation process. As shown in Algorithm 1, we first reverse the ground truth and use the reversed transcript as the target for optimizing the input audio, as in the first approach of generating untargeted audio AEs via FGSM. The generation is deemed to be successful when the edit distance between the transcript and the ground truth exceeds a certain threshold.  
Algorithm 1 Untargeted Audio AE Generation
Input: original audio signal, x; ground truth transcript, y; target ASR model m; maximum iteration: m a x _ i t e r ; edit distance threshold: d i s t a n c e _ m i n
Output: black-box untargeted audio AE, x
x x
y _ r e v e r s e ← reverse the characters in y
While i t e r < m a x _ i t e r do
   y _ r e v e r s e _ l o s s ← calculate loss of y _ r e v e r s e
   g r a d _ e s t i m a t e ← estimate the gradient of the loss function x using y _ r e v e r s e _ l o s s
   x x - g r a d _ e s t i m a t e * l e a r n i n g _ r a t e
  // use the lowering noise technique from [36]
   m a s k i n g _ l o s s ← masking loss noise in x
  optimize m a s k i n g _ l o s s noise in x
  If EditDistance(y, transcript of x ) ≥ d i s t a n c e _ m i n
   return x
  End If
End While
If i t e r = = m a x _ i t e r
  return fail
End If

5. Experiments and Discussion

5.1. Target Models and Data Sets

As with similar research in the ASR domain [9,10,39,44], we used DeepSpeech [1] as one of the target models for our experiments. DeepSpeech 0.8.2 (DeepSpeech 0.8.2 was implemented by Mozilla https://github.com/mozilla/DeepSpeech (accessed on 1 November 2020)), which is the latest release at the time of writing, was used in this research. It should be noted that DeepSpeech 0.1, which was used in previous studies [9,39], has been superseded with newer versions. In addition, DeepSpeech2 [2], which is an improved version of DeepSpeech that employs an end-to-end architecture, was also used. We used DeepSpeech2 V2 (DeepSpeech2 V2 was implemented and released by Sean Naren https://github.com/SeanNaren/deepspeech.pytorch (accessed on 1 November 2020)).
LibriSpeech [58] was employed as the data set because DeepSpeech and DeepSpeech2 both provide pre-trained models on LibriSpeech. In the experiments, we used audio from the test-clean and dev-clean data sets. For targeted AEs, one of the target phrases “power off”, “turn on airplane mode”, “visit danger dot com”, “call malicious number”, and “turn off lights” was selected at random to mimic malicious voice commands. The generation of untargeted AEs was deemed to be successful if the edit distance between the transcript and the ground truth was larger than 40% of the ground truth.
In previous work by Carlini and Wagner [9], they generated audio AEs using the first 100 test instances of the Mozilla Common Voice data set [59]. Most of this audio was short, between 1 and 8 seconds in duration. Carlini and Wagner [9] empirically observed that the generation of targeted AEs was easier the longer the source phrase, while the generation would be more difficult the longer the target phrase. Since our target phrases were relatively short, we used audio below 5 seconds to balance the difficulty of generating targeted audio AEs.
All experiments in this paper were performed with an Intel i7-8750H CPU and an Nvidia GeForce GTX 1060 graphic card. Using randomly selected audio from the test-clean data set of DeepSpeech and DeepSpeech2, respectively, we generated 150 targeted AEs, 150 untargeted AEs using FGSM, and 150 untargeted AEs based on our proposed method. For simplicity, in the remainder of this paper, we refer to untargeted AEs using our proposed method as untargeted AEs and untargeted AEs using FGSM as FGSM AEs. To obtain a balanced data set, we also extract 150 correctly transcribed and 150 incorrectly transcribed audio from the test-clean data set of each model. In addition, we generated 150 noisy audio signals by applying Gaussian noise with a standard deviation of 0.01 .
To generate targeted AEs, we ran 350 epochs for DeepSpeech and 300 epochs for DeepSpeech2 to suppress noise during the second stage since we observed that it is easier for DeepSpeech2 to suppress the noise without destroying adversarial perturbations. Noise suppression in all targeted AEs against DeepSpeech2 was successful. However, some AEs against DeepSpeech failed to lower noise within the 350 epochs. As such, we individually fine-tuned these noisy AEs by running extra epochs until the masking loss ( l θ ( ) in Equation (5)) was below a specific threshold. The smaller the masking loss, the smaller the distortion perturbations caused. We set the threshold to the masking loss calculated using the −20 dB distortion set published by [9] (https://nicholas.carlini.com/code/audio_adversarial_examples (accessed on 1 November 2020)).
The masking losses of our AEs were compared with the −20 dB distortion, −35 dB distortion, and −50 dB distortion sets published by [9] and the first set of the imperceptible adversarial examples published by [36] (http://cseweb.ucsd.edu/~yaq007/imperceptible-robust-adv.html (accessed on 1 November 2020)). Figure 1 and Figure 2 show these results. Smaller dB values mean lower distortion. Carlini and Wagner [9] reported that the distortion of 95% of their targeted AEs ranging between −15 dB and −45 dB. Thus, the resulting distortion in our targeted AEs is comparable with the results of related work. It should be mentioned that we can further lose masking loss by running more epochs, which will require a longer generation time. We have made examples of AEs generated in this work available at https://drive.google.com/drive/folders/1Ffed7xHmP5oKCuypEgJxQ80p35-vSIBm?usp=sharing (accessed on 20 October 2022).
Table 3 shows a comparison of the time taken for generating the audio AEs. FGSM was the fastest approach, but it had the lowest success rate. On average, it took 2.4 and 7.0 min to generate targeted audio AEs for DeepSpeech and DeepSpeech2, respectively. On the other hand, our proposed method required an average of 4.4 and 4.9 min to generate untargeted audio AEs. While we generated AEs one at a time, the generation process can be accelerated by generating multiple AEs in parallel. As a loose comparison, Carlini and Wagner [9] reported that their approach took about one hour to generate a single targeted audio AE on commodity hardware, while Zeng et al. [27] reported a time of 18 min on an 18-core CPU with dual graphic cards. While we cannot conclude that our generation process is statistically faster as the source audio and target phrases were different, intuitively, our method should speed up the generation of AEs because we do not limit the max-norm of perturbations.

5.2. Visualizing Decision Boundaries

As described in Section 3, the proposed method represents the decision boundaries of ASR models using heat maps of loss-function values and normalized edit distances. The M l o s s and M e d i t were calculated for correctly transcribed benign audio, targeted, and untargeted audio AEs. It was empirically observed that good results could be produced using a matrix of 128 × 128 and a step size s of 0.07 . Figure 3 shows examples of resulting heat maps.
In the heat maps shown in Figure 3, the horizontal axis represents the direction of the gradient of the loss function the input audio, while the vertical axis represents a random direction that is perpendicular to the gradient. The heat maps were generated by modifying input audio along these two directions and recording the changes. The center of the heat maps represents unmodified audio. In the experiments, we set y in Equation (1) to the transcript of the unmodified audio, because we wanted to calculate the changes in loss values and transcripts when modifying audio. For example, y is set to the target phrase of a targeted audio AE or the incorrect transcript of an untargeted audio AE.
It is evident from the resulting patterns that changes in loss function values and normalized edit distances are correlated. This aligns with the intuition that loss function values returned by an ASR model should increase as the difference between the transcript and y increases and vice versa. Furthermore, we can see that when a targeted audio AE is modified slightly, the resulting loss function value and normalized edit distance change significantly. This is true for both DeepSpeech and DeepSpeech2 and is consistent with our observation that adversarial perturbations in the generated targeted audio AEs are not robust. The significant changes in loss function values and normalized edit distances when we modify AEs are an indication of the non-robust property of adversarial perturbations.
In contrast, changes in loss function values and normalized edit distances for correctly transcribed benign audio are significantly smaller than for targeted audio AEs when audio is slightly modified. This implies that correctly transcribed benign audio is much more robust against perturbations than targeted audio AEs. This is consistent with our observation that some correctly transcribed benign audio could still be correctly transcribed even when a large amount of noise is present. Another observation is that slightly modifying untargeted audio AEs also results in large changes in loss function values and normalized edit distances. However, while this change appears to be less severe than targeted audio AEs, the resulting patterns are different when compared with the results of correctly transcribed benign audio.

5.3. Dimensionality Reduction

Based on the different patterns in loss function values and normalized edit distances in relation to targeted and untargeted audio AEs and benign audio shown in Section 5.2, it is logical to consider the possibility of differentiating audio AEs from benign audio-based differences in their patterns. Thus, we extracted features from the audio and projected them into 2D space using the PCA and t-SNE methods, using the method described in Section 3. It should be noted that if audio AE and benign audio features can clearly be differentiated into 2D space, this indicates that they can also be separated in the original higher-dimensional space.
In the experiment, benign audio was grouped as correctly and incorrectly transcribed audio. This was carried out to investigate whether there was a difference between them. In addition, noisy audio was also included. The features were normalized using their mean values and standard deviation before projecting them in 2D space. These results are shown in Figure 4.
The PCA projection results were almost the same for DeepSpeech and DeepSpeech2. Correctly and incorrectly transcribed audio clustered around the origin, while the other audio types were spread away from the origin. The correctly and incorrectly transcribed audio almost overlapped, indicating that there is little difference between their features. As previously discussed, the changes in loss function values and normalized edit distances for correctly transcribed benign audio are small, which explains why correctly and incorrectly transcribed audio cluster around the origin. In contrast, targeted audio AEs are far away from the origin. This is because small modifications will result in significant changes for targeted audio AEs, as discussed in the previous section. Untargeted audio AEs, FSGM audio AEs, and noisy audio all spread slightly away from the origin in the same direction. This implies that the features of these three audio types are similar.
Compared with PCA results, t-SNE projection was better at visualizing relationships between the data samples. In Figure 4, t-SNE projection again shows similar results for DeepSpeech and DeepSpeech2. Three clusters, excluding noisy audio, can be identified as follows: targeted audio AEs are clearly grouped in the first cluster; the second cluster mainly contains correctly and incorrectly transcribed benign audio; and the third cluster consists of untargeted audio AEs and FGSM AEs, i.e., untargeted attacks. The results of t-SNE projection are promising, since the various audio types are clustered according to their categories. An interesting observation is that incorrectly transcribed audio does not overlap with untargeted audio AEs or FGSM AEs, although all of them lead to incorrect transcriptions. A potential explanation is that incorrectly transcribed audio from the test-clean data set does not cause severe errors such as untargeted audio AEs and FGSM AEs. In addition, noisy audio is contained in both the second cluster (benign audio) and the third cluster (untargeted attack). This may be because some noisy audio is like benign audioin that it can be transcribed correctly or with little error, while some noisy audio behaves like untargeted attacks, which lead to significant errors in transcriptions. Upon closer inspection, the untargeted AEs and FGSM AEs are separate from each other in the case of DeepSpeech2, but the same is not true for DeepSpeech.

5.4. Anomaly Detection

Visualization results presented in the previous subsection indicate the possibility of detecting audio AEs based on their features. Hence, instead of training a classifier on benign audio and audio AEs, we experimented with using anomaly detection to detect audio AEs. In practice, audio AEs generated by adversaries are unlikely to be previously seen. Anomaly detection is appropriate for defending against previously unknown attacks.
In the experiments, we used audio from the dev-clean data set to train an anomaly detection model. This model was then used to detect audio AEs generated using the test-clean data set. In particular, audio features from dev-clean were extracted using the method described in Section 3. These features were used to train an EllipticEnvelope model implemented by scikit-learn [60]. This model detects outliers in a Gaussian distributed data set. We use the default parameters so that our experiment results can serve as a lower bound for anomaly detection. We report true positive (TP), false positive (FP), true negative (TN), false negative (FN), and detection rate (DR) for each category of benign audio and audio AEs together with overall precision (Pre), recall (Rec), and accuracy (Acc). Specifically, p r e c i s i o n = T P T P + F P , r e c a l l = T P T P + F N , a c c u r a c y = T P + T N T P + F P + T N + F N . For audio AEs, D R = T P T P + F P . For benign audio, D R = T N T N + F N .
Table 4 presents the experimental anomaly detection results for DeepSpeech and DeepSpeech2. Overall, the detection results are similar for both ASR models. As expected, targeted AEs are easily detected at detection rates of 100 % . This is in line with the observation that targeted AEs can clearly be separated from other audio types in lower-dimensional space. It is reasonable that the detection rates of untargeted AEs were loan-targeted AEs since some untargeted AEs were mixed with benign audio in the PCA projection, as previously shown in Figure 4. The detection rates of FGSM AEs were surprisingly lone untargeted audio AEs, although these two AEs were clustered together in the t-SNE projection. This indicates that the simple anomaly detection model that was used is too basic for detecting FGSM AEs. In addition to benign audio, noisy audio could also be correctly identified at high detection rates. This was not as expected, since some of the noisy audio was mixed with untargeted AEs and FGSM AEs in low-dimensional space. This suggests that noisy audio is actually clustered with benign audio in the original higher-dimensional space, even though the 2D projection did not show this.
In a study by Samizade et al. [44], they generated white-box and black-box targeted audio AEs against DeepSpeech. They trained a neural network on white-box targeted audio AEs to detect black-box targeted audio AEs and vice versa. Our detection accuracy for the two ASR models of 87.44 % and 82.22 % is overall higher than their reported results of 82.07 % and 48.76 % , respectively. While this may not be a fair comparison, as they used a different approach, we mainly want to emphasize that the detection of previously unknown audio AEs is a challenging task. It is anticipated that if we extract more sophisticated features and utilize a more advanced anomaly detection method, it is highly likely that the detection results can be improved.

6. Discussion

In this section, we start by discussing the robustness of AEs against our method. Then, we discuss a promising research direction for eliminating AEs.

6.1. Robust Audio Adversarial Example

The fundamental assumption underlying this research is that the decision boundary patterns around benign audio and audio AEs are significantly different from one another. In this study, we used heat maps of loss function values and normalized edit distances to test the validity of this assumption. We also investigated whether the heat maps could differentiate audio AEs from benign audio under a white-box threat model. Although we demonstrated that heat maps of audio AEs and benign audio are significantly different, these audio AEs were generated without prior knowledge of the heat map generation process. It is conceivable that if an adversary has full knowledge of how the heat maps are generated, they can potentially generate targeted audio AEs with small changes in loss function values and normalized edit distances when the AEs are modified. We refer to this type of AEs as robust audio AEs because they are potentially indistinguishable from benign audio using our proposed heat map visualization method. Moreover, features extracted from such robust audio AEs may not be distinct from benign audio features.
In research efforts to increase the robustness of AEs, Athalye et al. [61] proposed the use of Expectation over Transformation (EoT). The idea behind this approach is to optimize the loss function over various transformations, such as Gaussian noise. Qin et al. [36] employed this method to incorporate reverberations in the generation process in order for audio AEs to remain adversarial over the air. If such reverberations were used to generate audio AEs, it is possible that there may be fewer changes in loss function values and normalized edit distances for such robust AEs, at least in the directions considered to be transformations. This is because the EoT directly incorporates this property in the optimization formula. From another point of view, the EoT can be thought of as imposing limits on the resulting decision boundary patterns around successfully generated AEs.
In light of this, we conducted experiments to verify the existence of AEs that were robust against our method. In these experiments, we generated a robust audio AE against DeepSpeech2 in a similar way to generating our targeted audio AEs, as discussed in Section 4.1. During each epoch, we modified the audio along the gradient of the loss function’s direction, denoted as g ¯ , as well as a perpendicular direction, denoted as p ¯ , and optimized the audio together with the modified ones. It should be noted that p ¯ and g ¯ are also used for generating heat maps and extracting features as discussed in Section 3. Theoretically, this strategy of generating robust audio AEs is equivalent to the EoT. There are two stages in the process. Stage 1 succeeds if the robust audio AE is transcribed as the target phrase and the maximum edit distance between transcripts of all modified audio and the targeted phrase is less than 3. Then, stage 2 focuses on lowering the noise in the robust audio AE. Similar to the other experiments, we ran stage 2 for 300 epochs. We successfully generated one targeted audio AE in 23.1 min, and the corresponding heat maps are shown in Figure 5. As expected, the heat maps of this robust audio AE are similar to the benign audios shown in Figure 3 because there is only a small change in loss values and normalized edit distance when it is slightly modified along p ¯ and g ¯ . Although this implies the success of this robust AE, its distortion is −11.2 dB, which is noisier than the −20 dB distortion set. To demonstrate this, Figure 6 uses spectrograms to show that the distortion is obvious.
Looking deeper, p ¯ is initialized from a fixed vector for the purpose of reproducibility during implementation and this is fine under a black-box threat model in which adversaries have no knowledge about p ¯ . However, this fixed initialization might be exploited under a white-box threat model. Specifically, p ¯ would be almost fixed when g ¯ is stable and a robust audio AE only needs to cause a small change in loss values and normalized edit distance along the almost fixed p ¯ and g ¯ . This can happen when the loss is near a local minimum or when optimization is stuck on a plateau. Hence, it is interesting to investigate whether a defender can detect the robust AE by deliberately using different values of p ¯ . Figure 7 presents heat maps of the robust AE with different values of p ¯ . It shows that heat maps of the robust AE still resemble heat maps of benign audio, which implies that the robust AE can bypass detection even if different values of p ¯ are used.

6.2. Eliminating Adversarial Examples

Research has demonstrated that the detection of adversarial perturbations in input is efficient and can achieve high performance. This makes it suitable for potential deployment in real-world applications. Nevertheless, existing detection methods cannot theoretically guarantee the perfect detection of future AEs. This means there is still a likelihood, no matter how small, that an AE can bypass detection and cause harm. Other work on defending against AEs has focused on adversarial training [62,63] and theoretical robustness [64]. However, these research directions have their own unsolved problems: the performance of adversarially trained models is unsatisfactory because of the tradeoff between performance and robustness [50], and methods that provide theoretical robustness against AEs cannot work for real-world complex DNNs. We contend that the ultimate solution for defending against AEs is to make the decision strategy of DNNs align with human perception. In this manner, AEs can be eliminated because human perception is robust against small perturbations.
There is currently a significant difference between the decision strategy of DNNs and the perception of humans. This is demonstrated by the “Clever Hans” behavior in DNN predictions [65]. Specifically, DNN predictions can be based on non-robust features that allow distinction within the training data but are not related to the intended task [19]. An example by Lapuschkin et al. [65] showed that an image classification model outputs a “horse” label if a source tag is present because one-fifth of the horse images contained this source tag. To demonstrate this, the researchers stamped the source tag onto an image of a car and the prediction changed from “car” to “horse”.
The “Clever Hans” behavior also exists in ASR systems because of the existence of audio AEs. Recall that small or even imperceptible perturbations can force an ASR system to output a malicious command that was predefined by an adversary at a 100% success rate as shown in Table 3. Geirhos et al. [66] explained this “Clever Hans” behavior in DNNs via the concept of shortcut learning. In fact, shortcut learning also exists in biological neural networks. They provided the example that rats distinguished between colors through the odor of paint, which was an unintended solution in the experiments.
In the literature, the direction of aligning the decision strategy of DNNs with human perception has attracted more and more attention from researchers in recent years. For example, Liu et al. [67] proposed first identifying incorrect predictions and then increasing their weight for retraining a DNN. Singla et al. [68] recently proposed a novel method to identify spurious features that are learned by DNNs. These examples highlight that aligning the decision strategy of DNNs with human perception for improving the robustness of models against AEs is an interesting direction for future work.

7. Conclusions and Future Work

With ASR systems becoming ubiquitous in commonly used commercial applications, audio AEs pose a severe security threat to these systems. Despite the various methods proposed by the research community for defending against audio AEs, the intrinsic properties of audio AEs have not been well studied or understood. This paper presents a method for visualizing the different decision boundary patterns around audio AEs and benign audio, which allow them to be distinguished from each other. This paper also showed that by extracting features based on the decision boundaries in conjunction with dimensionality-reduction techniques, the features of audio AEs and benign audio can be clearly separated in 2D space. In addition, this work demonstrated the possibility of detecting previously unknown audio AEs using anomaly detection. Our experimental results showed that this approach achieved significantly high detection rates for targeted audio AEs.
In future work, we will investigate various methods for improving audio AE detection results through the incorporation of more advanced feature-extraction techniques and anomaly-detection models. Another interesting direction for future work is on improving the robustness of models against AEs via aligning the decision strategy of DNNs with human perception.

Author Contributions

Conceptualization, W.Z., Y.-W.C. and W.S.; validation, J.K. and N.T.L.; investigation, W.Z.; writing—original draft preparation, W.Z.; writing—review and editing, Y.-W.C.; supervision, Y.-W.C. and W.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data from [9] is available at https://nicholas.carlini.com/code/audio_adversarial_examples (accessed on 1 November 2020); Data from [36] is available at http://cseweb.ucsd.edu/~yaq007/imperceptible-robust-adv.html (accessed on 1 November 2020); Examples from this paper are available at https://drive.google.com/drive/folders/1Ffed7xHmP5oKCuypEgJxQ80p35-vSIBm?usp=sharing (accessed on 20 October 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hannun, A.; Case, C.; Casper, J.; Catanzaro, B.; Diamos, G.; Elsen, E.; Prenger, R.; Satheesh, S.; Sengupta, S.; Coates, A.; et al. Deep speech: Scaling up end-to-end speech recognition. arXiv 2014, arXiv:1412.5567. [Google Scholar]
  2. Amodei, D.; Ananthanarayanan, S.; Anubhai, R.; Bai, J.; Battenberg, E.; Case, C.; Casper, J.; Catanzaro, B.; Cheng, Q.; Chen, G.; et al. Deep speech 2: End-to-end speech recognition in english and mandarin. In Proceedings of the International Conference on Machine Learning, New York, NY, USA, 20–22 June 2016; pp. 173–182. [Google Scholar]
  3. Shen, J.; Nguyen, P.; Wu, Y.; Chen, Z.; Chen, M.X.; Jia, Y.; Kannan, A.; Sainath, T.N.; Cao, Y.; Chiu, C.; et al. Lingvo: A Modular and Scalable Framework for Sequence-to-Sequence Modeling. arXiv 2019, arXiv:1902.08295. [Google Scholar]
  4. Biggio, B.; Roli, F. Wild patterns: Ten years after the rise of adversarial machine learning. Pattern Recognit. 2018, 84, 317–331. [Google Scholar] [CrossRef] [Green Version]
  5. Szegedy, C.; Zaremba, W.; Sutskever, I.; Bruna, J.; Erhan, D.; Goodfellow, I.J.; Fergus, R. Intriguing properties of neural networks. In Proceedings of the 2nd International Conference on Learning Representations, ICLR 2014, Banff, AB, Canada, 14–16 April 2014; Conference Track Proceedings. Bengio, Y., Le Cun, Y., Eds.; 2014. [Google Scholar]
  6. Jia, R.; Liang, P. Adversarial Examples for Evaluating Reading Comprehension Systems. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, EMNLP 2017, Copenhagen, Denmark, 9–11 September 2017; Palmer, M., Hwa, R., Riedel, S., Eds.; Association for Computational Linguistics: Stroudsburg, PA, USA, 2017; pp. 2021–2031. [Google Scholar] [CrossRef]
  7. Zhang, H.; Zhou, H.; Miao, N.; Li, L. Generating fluent adversarial examples for natural languages. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, 28 July–2 August 2019; pp. 5564–5569. [Google Scholar]
  8. Ebrahimi, J.; Rao, A.; Lowd, D.; Dou, D. Hotflip: White-box adversarial examples for text classification. arXiv 2017, arXiv:1712.06751. [Google Scholar]
  9. Carlini, N.; Wagner, D. Audio adversarial examples: Targeted attacks on speech-to-text. In Proceedings of the 2018 IEEE Security and Privacy Workshops (SPW), San Francisco, CA, USA, 24 May 2018; pp. 1–7. [Google Scholar]
  10. Schönherr, L.; Kohls, K.; Zeiler, S.; Holz, T.; Kolossa, D. Adversarial Attacks Against Automatic Speech Recognition Systems via Psychoacoustic Hiding. In Proceedings of the 26th Annual Network and Distributed System Security Symposium, NDSS 2019, San Diego, CA, USA, 24–27 February 2019; The Internet Society: Reston, VA, USA, 2019. [Google Scholar]
  11. Alzantot, M.; Balaji, B.; Srivastava, M.B. Did you hear that? Adversarial Examples Against Automatic Speech Recognition. arXiv 2018, arXiv:1801.00554. [Google Scholar]
  12. Chen, Y.; Zhang, J.; Yuan, X.; Zhang, S.; Chen, K.; Wang, X.; Guo, S. Sok: A modularized approach to study the security of automatic speech recognition systems. ACM Trans. Priv. Secur. 2022, 25, 1–31. [Google Scholar] [CrossRef]
  13. Vadillo, J.; Santana, R. On the human evaluation of universal audio adversarial perturbations. Comput. Secur. 2022, 112, 102495. [Google Scholar] [CrossRef]
  14. Mun, H.; Seo, S.; Son, B.; Yun, J. Black-Box Audio Adversarial Attack Using Particle Swarm Optimization. IEEE Access 2022, 10, 23532–23544. [Google Scholar] [CrossRef]
  15. Xie, Y.; Shi, C.; Li, Z.; Liu, J.; Chen, Y.; Yuan, B. Real-time, universal, and robust adversarial attacks against speaker recognition systems. In Proceedings of the ICASSP 2020—2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 1738–1742. [Google Scholar]
  16. Wang, Q.; Guo, P.; Xie, L. Inaudible Adversarial Perturbations for Targeted Attack in Speaker Recognition. arXiv 2020, arXiv:2005.10637. [Google Scholar]
  17. Chen, G.; Zhao, Z.; Song, F.; Chen, S.; Fan, L.; Wang, F.; Wang, J. Towards Understanding and Mitigating Audio Adversarial Examples for Speaker Recognition. arXiv 2022, arXiv:2206.03393. [Google Scholar] [CrossRef]
  18. Tsipras, D.; Santurkar, S.; Engstrom, L.; Turner, A.; Madry, A. Robustness May Be at Odds with Accuracy. In Proceedings of the 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, 6–9 May 2019; Available online: OpenReview.net (accessed on 1 November 2022).
  19. Ilyas, A.; Santurkar, S.; Tsipras, D.; Engstrom, L.; Tran, B.; Madry, A. Adversarial examples are not bugs, they are features. Adv. Neural Inf. Process. Syst. 2019, 32, 125–136. [Google Scholar]
  20. Zhang, C.; Benz, P.; Imtiaz, T.; Kweon, I.S. Understanding Adversarial Examples From the Mutual Influence of Images and Perturbations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 14521–14530. [Google Scholar]
  21. Papernot, N.; McDaniel, P.; Wu, X.; Jha, S.; Swami, A. Distillation as a defense to adversarial perturbations against deep neural networks. In Proceedings of the 2016 IEEE Symposium on Security and Privacy (SP), San Jose, CA, USA, 22–26 May 2016; pp. 582–597. [Google Scholar]
  22. Xu, W.; Evans, D.; Qi, Y. Feature Squeezing: Detecting Adversarial Examples in Deep Neural Networks. In Proceedings of the 25th Annual Network and Distributed System Security Symposium, NDSS 2018, San Diego, CA, USA, 18–21 February 2018; The Internet Society: Reston, VA, USA, 2018. [Google Scholar]
  23. Cohen, G.; Sapiro, G.; Giryes, R. Detecting adversarial samples using influence functions and nearest neighbors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 14453–14462. [Google Scholar]
  24. Ma, X.; Li, B.; Wang, Y.; Erfani, S.M.; Wijewickrema, S.; Schoenebeck, G.; Song, D.; Houle, M.E.; Bailey, J. Characterizing adversarial subspaces using local intrinsic dimensionality. arXiv 2018, arXiv:1801.02613. [Google Scholar]
  25. Goodfellow, I.J.; Shlens, J.; Szegedy, C. Explaining and Harnessing Adversarial Examples. In Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, 7–9 May 2015; Conference Track Proceedings. Bengio, Y., LeCun, Y., Eds.; 2015. [Google Scholar]
  26. Zhang, H.; Chen, H.; Song, Z.; Boning, D.S.; Dhillon, I.S.; Hsieh, C. The Limitations of Adversarial Training and the Blind-Spot Attack. In Proceedings of the 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, 6–9 May 2019; Available online: OpenReview.net (accessed on 1 November 2022).
  27. Zeng, Q.; Su, J.; Fu, C.; Kayas, G.; Luo, L.; Du, X.; Tan, C.C.; Wu, J. A multiversion programming inspired approach to detecting audio adversarial examples. In Proceedings of the 2019 49th Annual IEEE/IFIP International Conference on Dependable Systems and Networks (DSN), Portland, OR, USA, 24–27 June 2019; pp. 39–51. [Google Scholar]
  28. Yang, Z.; Li, B.; Chen, P.; Song, D. Characterizing Audio Adversarial Examples Using Temporal Dependency. In Proceedings of the 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, 6–9 May 2019; Available online: OpenReview.net (accessed on 1 November 2022).
  29. Zong, W.; Chow, Y.; Susilo, W. Towards Visualizing and Detecting Audio Adversarial Examples for Automatic Speech Recognition. In Proceedings of the Information Security and Privacy—26th Australasian Conference, ACISP 2021, Virtual Event, 1–3 December 2021; Baek, J., Ruj, S., Eds.; Lecture Notes in Computer Science. Springer: Berlin/Heidelberg, Germany, 2021; Volume 13083, pp. 531–549. [Google Scholar] [CrossRef]
  30. Moosavi-Dezfooli, S.M.; Fawzi, A.; Frossard, P. Deepfool: A simple and accurate method to fool deep neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2574–2582. [Google Scholar]
  31. Athalye, A.; Carlini, N.; Wagner, D. Obfuscated gradients give a false sense of security: Circumventing defenses to adversarial examples. arXiv 2018, arXiv:1802.00420. [Google Scholar]
  32. Yuan, X.; Chen, Y.; Zhao, Y.; Long, Y.; Liu, X.; Chen, K.; Zhang, S.; Huang, H.; Wang, X.; Gunter, C.A. Commandersong: A systematic approach for practical adversarial voice recognition. In Proceedings of the 27th USENIX Security Symposium (USENIX Security 18), Baltimore, MD, USA, 15–17 August 2018; pp. 49–64. [Google Scholar]
  33. Povey, D.; Ghoshal, A.; Boulianne, G.; Burget, L.; Glembek, O.; Goel, N.; Hannemann, M.; Motlicek, P.; Qian, Y.; Schwarz, P.; et al. The Kaldi speech recognition toolkit. In Proceedings of the IEEE 2011 Workshop on Automatic Speech Recognition and Understanding, Waikoloa, HI, USA, 1–15 December 2011. [Google Scholar]
  34. Graves, A.; Fernández, S.; Gomez, F.; Schmidhuber, J. Connectionist temporal classification: Labelling unsegmented sequence data with recurrent neural networks. In Proceedings of the 23rd International Conference on Machine Learning, Pittsburgh, PA, USA, 25–29 June 2006; pp. 369–376. [Google Scholar]
  35. Liu, X.; Wan, K.; Ding, Y.; Zhang, X.; Zhu, Q. Weighted-sampling audio adversarial example attack. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 4908–4915. [Google Scholar]
  36. Qin, Y.; Carlini, N.; Cottrell, G.W.; Goodfellow, I.J.; Raffel, C. Imperceptible, Robust, and Targeted Adversarial Examples for Automatic Speech Recognition. In Proceedings of the 36th International Conference on Machine Learning, ICML 2019, Long Beach, CA, USA, 9–15 June 2019; pp. 5231–5240. [Google Scholar]
  37. Eisenhofer, T.; Schönherr, L.; Frank, J.; Speckemeier, L.; Kolossa, D.; Holz, T. Dompteur: Taming audio adversarial examples. In Proceedings of the 30th USENIX Security Symposium (USENIX Security 21), Virtual Event, 11–13 August 2021; pp. 2309–2326. [Google Scholar]
  38. Zong, W.; Chow, Y.W.; Susilo, W. High Quality Audio Adversarial Examples Without Using Psychoacoustics. In International Symposium on Cyberspace Safety and Security; Springer: Berlin/Heidelberg, Germany, 2022; pp. 163–177. [Google Scholar]
  39. Taori, R.; Kamsetty, A.; Chu, B.; Vemuri, N. Targeted adversarial examples for black box audio systems. In Proceedings of the 2019 IEEE Security and Privacy Workshops (SPW), San Francisco, CA, USA, 19–23 May 2019; pp. 15–20. [Google Scholar]
  40. Chen, Y.; Yuan, X.; Zhang, J.; Zhao, Y.; Zhang, S.; Chen, K.; Wang, X. Devil’s whisper: A general approach for physical adversarial attacks against commercial black-box speech recognition devices. In Proceedings of the 29th USENIX Security Symposium (USENIX Security 20), Boston, MA, USA, 12–14 August 2020. [Google Scholar]
  41. Neekhara, P.; Hussain, S.; Pandey, P.; Dubnov, S.; McAuley, J.; Koushanfar, F. Universal adversarial perturbations for speech recognition systems. arXiv 2019, arXiv:1905.03828. [Google Scholar]
  42. Abdullah, H.; Rahman, M.S.; Garcia, W.; Warren, K.; Yadav, A.S.; Shrimpton, T.; Traynor, P. Hear “no evil”, see “kenansville”*: Efficient and transferable black-box attacks on speech recognition and voice identification systems. In Proceedings of the 2021 IEEE Symposium on Security and Privacy (SP), San Francisco, CA, USA, 24–27 May 2021; pp. 712–729. [Google Scholar]
  43. Esmaeilpour, M.; Cardinal, P.; Koerich, A.L. Detection of Adversarial Attacks and Characterization of Adversarial Subspace. In Proceedings of the ICASSP 2020—2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 3097–3101. [Google Scholar]
  44. Samizade, S.; Tan, Z.H.; Shen, C.; Guan, X. Adversarial Example Detection by Classification for Deep Speech Recognition. In Proceedings of the ICASSP 2020—2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 3102–3106. [Google Scholar]
  45. Guo, Q.; Ye, J.; Hu, Y.; Zhang, G.; Li, X.; Li, H. MultiPAD: A Multivariant Partition-Based Method for Audio Adversarial Examples Detection. IEEE Access 2020, 8, 63368–63380. [Google Scholar] [CrossRef]
  46. Hussain, S.; Neekhara, P.; Dubnov, S.; McAuley, J.; Koushanfar, F. WaveGuard: Understanding and Mitigating Audio Adversarial Examples. In Proceedings of the 30th USENIX Security Symposium (USENIX Security 21), Virtual Event, 11–13 August 2021; pp. 2273–2290. [Google Scholar]
  47. Yang, C.H.; Qi, J.; Chen, P.Y.; Ma, X.; Lee, C.H. Characterizing Speech Adversarial Examples Using Self-Attention U-Net Enhancement. In Proceedings of the ICASSP 2020—2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 3107–3111. [Google Scholar]
  48. Guo, Q.; Ye, J.; Chen, Y.; Hu, Y.; Lan, Y.; Zhang, G.; Li, X. INOR—An Intelligent noise reduction method to defend against adversarial audio examples. Neurocomputing 2020, 401, 160–172. [Google Scholar] [CrossRef]
  49. Zhang, C.; Zhang, M.; Zhang, S.; Jin, D.; Zhou, Q.; Cai, Z.; Zhao, H.; Liu, X.; Liu, Z. Delving deep into the generalization of vision transformers under distribution shifts. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–20 June 2022; pp. 7277–7286. [Google Scholar]
  50. Bai, T.; Luo, J.; Zhao, J.; Wen, B.; Wang, Q. Recent advances in adversarial training for adversarial robustness. arXiv 2021, arXiv:2102.01356. [Google Scholar]
  51. Hohman, F.; Kahng, M.; Pienta, R.; Chau, D.H. Visual analytics in deep learning: An interrogative survey for the next frontiers. IEEE Trans. Vis. Comput. Graph. 2018, 25, 2674–2693. [Google Scholar] [CrossRef]
  52. Norton, A.P.; Qi, Y. Adversarial-Playground: A visualization suite showing how adversarial examples fool deep learning. In Proceedings of the 2017 IEEE Symposium on Visualization for Cyber Security (VizSec), Phoenix, AZ, USA, 2 October 2017; pp. 1–4. [Google Scholar]
  53. Liu, Y.; Chen, X.; Liu, C.; Song, D. Delving into Transferable Adversarial Examples and Black-box Attacks. In Proceedings of the 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, 24–26 April 2017. Conference Track Proceedings. [Google Scholar]
  54. Stutz, D.; Hein, M.; Schiele, B. Disentangling adversarial robustness and generalization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 6976–6987. [Google Scholar]
  55. Abdi, H.; Williams, L.J. Principal component analysis. Wiley Interdiscip. Rev. Comput. Stat. 2010, 2, 433–459. [Google Scholar] [CrossRef]
  56. Maaten, L.v.d.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
  57. Lin, Y.; Abdulla, W.H. Principles of psychoacoustics. In Audio Watermark; Springer: Berlin/Heidelberg, Germany, 2015; pp. 15–49. [Google Scholar]
  58. Panayotov, V.; Chen, G.; Povey, D.; Khudanpur, S. Librispeech: An asr corpus based on public domain audio books. In Proceedings of the 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Queensland, Australia, 19–24 April 2015; pp. 5206–5210. [Google Scholar]
  59. Ardila, R.; Branson, M.; Davis, K.; Henretty, M.; Kohler, M.; Meyer, J.; Morais, R.; Saunders, L.; Tyers, F.M.; Weber, G. Common voice: A massively-multilingual speech corpus. arXiv 2019, arXiv:1912.06670. [Google Scholar]
  60. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  61. Athalye, A.; Engstrom, L.; Ilyas, A.; Kwok, K. Synthesizing robust adversarial examples. In Proceedings of the International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018; pp. 284–293. [Google Scholar]
  62. Allen-Zhu, Z.; Li, Y. Feature purification: How adversarial training performs robust deep learning. In Proceedings of the 2021 IEEE 62nd Annual Symposium on Foundations of Computer Science (FOCS), Denver, CO, USA, 7–10 February 2022; pp. 977–988. [Google Scholar]
  63. Goldblum, M.; Fowl, L.; Feizi, S.; Goldstein, T. Adversarially robust distillation. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 3996–4003. [Google Scholar]
  64. Gowal, S.; Dvijotham, K.; Stanforth, R.; Bunel, R.; Qin, C.; Uesato, J.; Arandjelovic, R.; Mann, T.A.; Kohli, P. On the Effectiveness of Interval Bound Propagation for Training Verifiably Robust Models. arXiv 2018, arXiv:1810.12715. [Google Scholar]
  65. Lapuschkin, S.; Wäldchen, S.; Binder, A.; Montavon, G.; Samek, W.; Müller, K.R. Unmasking Clever Hans predictors and assessing what machines really learn. Nat. Commun. 2019, 10, 1096. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  66. Geirhos, R.; Jacobsen, J.H.; Michaelis, C.; Zemel, R.; Brendel, W.; Bethge, M.; Wichmann, F.A. Shortcut learning in deep neural networks. Nat. Mach. Intell. 2020, 2, 665–673. [Google Scholar] [CrossRef]
  67. Liu, E.Z.; Haghgoo, B.; Chen, A.S.; Raghunathan, A.; Koh, P.W.; Sagawa, S.; Liang, P.; Finn, C. Just train twice: Improving group robustness without training group information. In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021; pp. 6781–6792. [Google Scholar]
  68. Singla, S.; Feizi, S. Salient ImageNet: How to discover spurious features in Deep Learning? In Proceedings of the International Conference on Learning Representations, Virtual, 25 April 2022. [Google Scholar]
Figure 1. Histograms comparing the masking loss ( l θ ( ) in Equation (5)) of our generated adversarial examples (AEs) for DeepSpeech (DS) and DeepSpeech2 (DS2) with the −20 dB distortion, −35 dB distortion, and −50 dB distortion sets published by [9] (labeled as −20 dB, −35 dB, and −50 dB, respectively) and the first set of imperceptible adversarial examples published by [36] (labeled as imperceptible). The smaller the masking loss, the lower the resulting distortion perturbations. Coordinates along the horizontal axis were calculated as l n ( m a s k i n g l o s s + 1 ) . Overall, the masking loss of targeted AEs against DeepSpeech2 was smaller compared with DeepSpeech. The masking loss of most untargeted AEs was between −20 dB and −35 dB.
Figure 1. Histograms comparing the masking loss ( l θ ( ) in Equation (5)) of our generated adversarial examples (AEs) for DeepSpeech (DS) and DeepSpeech2 (DS2) with the −20 dB distortion, −35 dB distortion, and −50 dB distortion sets published by [9] (labeled as −20 dB, −35 dB, and −50 dB, respectively) and the first set of imperceptible adversarial examples published by [36] (labeled as imperceptible). The smaller the masking loss, the lower the resulting distortion perturbations. Coordinates along the horizontal axis were calculated as l n ( m a s k i n g l o s s + 1 ) . Overall, the masking loss of targeted AEs against DeepSpeech2 was smaller compared with DeepSpeech. The masking loss of most untargeted AEs was between −20 dB and −35 dB.
Jimaging 08 00324 g001
Figure 2. Histograms comparing the masking loss of fast gradient sign method (FGSM) AEs and noisy audio for DeepSpeech (DS) and DeepSpeech2 (DS2) with the −20 dB distortion, −35 dB distortion, and −50 dB distortion sets published by [9] (labeled as −20 dB, −35 dB, and −50 dB) and the first set of imperceptible adversarial examples published by [36] (labeled as imperceptible). Coordinates of the horizontal axis are calculated as l n ( l o s s + 1 ) , where l o s s is the masking loss proposed by [36]. We can see that the masking loss of FGSM AEs is similar to the masking loss of noisy audio.
Figure 2. Histograms comparing the masking loss of fast gradient sign method (FGSM) AEs and noisy audio for DeepSpeech (DS) and DeepSpeech2 (DS2) with the −20 dB distortion, −35 dB distortion, and −50 dB distortion sets published by [9] (labeled as −20 dB, −35 dB, and −50 dB) and the first set of imperceptible adversarial examples published by [36] (labeled as imperceptible). Coordinates of the horizontal axis are calculated as l n ( l o s s + 1 ) , where l o s s is the masking loss proposed by [36]. We can see that the masking loss of FGSM AEs is similar to the masking loss of noisy audio.
Jimaging 08 00324 g002
Figure 3. Heat maps of loss function values and normalized edit distances regarding different audio AEs and benign audio for DeepSpeech (DS) and DeepSpeech2 (DS2), respectively. Changes in loss function values and normalized edit distances of targeted and untargeted audio AEs are clearly different from benign audio.
Figure 3. Heat maps of loss function values and normalized edit distances regarding different audio AEs and benign audio for DeepSpeech (DS) and DeepSpeech2 (DS2), respectively. Changes in loss function values and normalized edit distances of targeted and untargeted audio AEs are clearly different from benign audio.
Jimaging 08 00324 g003
Figure 4. Results obtained by projecting the features of various types of audio using the principal component analysis (PCA) and t-distributed stochastic neighbor embedding (t-SNE) techniques.
Figure 4. Results obtained by projecting the features of various types of audio using the principal component analysis (PCA) and t-distributed stochastic neighbor embedding (t-SNE) techniques.
Jimaging 08 00324 g004
Figure 5. Heat maps of loss function values and normalized edit distances for a robust audio AE. There are small changes in loss function values and normalized edit distances when this robust audio AE is slightly modified.
Figure 5. Heat maps of loss function values and normalized edit distances for a robust audio AE. There are small changes in loss function values and normalized edit distances when this robust audio AE is slightly modified.
Jimaging 08 00324 g005
Figure 6. Comparing spectrograms of (a) benign audio with (b) the robust audio AE. Distortion in the AE is obvious.
Figure 6. Comparing spectrograms of (a) benign audio with (b) the robust audio AE. Distortion in the AE is obvious.
Jimaging 08 00324 g006
Figure 7. Using different values of p ¯ to generate heat maps of loss function values and normalized edit distances for a robust audio AE. Different values of p ¯ were used to generate (a,b).
Figure 7. Using different values of p ¯ to generate heat maps of loss function values and normalized edit distances for a robust audio AE. Different values of p ¯ were used to generate (a,b).
Jimaging 08 00324 g007
Table 2. Techniques for defending against audio AEs in ASR systems.
Table 2. Techniques for defending against audio AEs in ASR systems.
Defense TypeMethodTarget ModelTechnique
DetectionZeng et al. [27]DeepSpeechMultiple transcripts
Yang et al. [28]DeepSpeech, KaldiTemporal dependency
Samizade et al. [44]DeepSpeechCNN
Guo et al. [45]DeepSpeechMultivariant partition
Hussain et al. [46]DeepSpeech, LingvoInput transformation
RecoveryYang et al. [47]DeepSpeechSpeech quality enhancement
Guo et al. [48]DeepSpeechNoise reduction
Yuan et al. [32]KaldiDownsampling
Chen et al. [40]IBM Speech to TextDownsampling
Table 3. Total time taken for generating the audio AEs and their success rates.
Table 3. Total time taken for generating the audio AEs and their success rates.
TypeDeepSpeechDeepSpeech2
Targeted AEs17.4 h (100.00%)6.0 h (100.00%)
Untargeted AEs11.0 h (98.68%)12.3 h (100.00%)
FGSM AEs0.13 h (28.79%)0.07 h (38.66%)
Table 4. Anomaly detection results of previously unknown audio AEs.
Table 4. Anomaly detection results of previously unknown audio AEs.
DeepSpeechDeepSpeech2
TypeTPFPTNFNDRTPFPTNFNDR
Targeted AEs150--0100.00%150--0100.00%
Untargeted AEs120--3080.00%129--2186.00%
FGSM AEs86--6457.33%33--11722.00%
Noisy Audio-9141-94.00%-8142-94.67%
Correctly trans.-4146-97.33%-2148-98.67%
Incorrectly trans.-6144-96.00%-12138-92.00%
PreRecAcc PreRecAcc
94.93%79.11%87.44% 93.41%69.33%82.22%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zong, W.; Chow, Y.-W.; Susilo, W.; Kim, J.; Le, N.T. Detecting Audio Adversarial Examples in Automatic Speech Recognition Systems Using Decision Boundary Patterns. J. Imaging 2022, 8, 324. https://doi.org/10.3390/jimaging8120324

AMA Style

Zong W, Chow Y-W, Susilo W, Kim J, Le NT. Detecting Audio Adversarial Examples in Automatic Speech Recognition Systems Using Decision Boundary Patterns. Journal of Imaging. 2022; 8(12):324. https://doi.org/10.3390/jimaging8120324

Chicago/Turabian Style

Zong, Wei, Yang-Wai Chow, Willy Susilo, Jongkil Kim, and Ngoc Thuy Le. 2022. "Detecting Audio Adversarial Examples in Automatic Speech Recognition Systems Using Decision Boundary Patterns" Journal of Imaging 8, no. 12: 324. https://doi.org/10.3390/jimaging8120324

APA Style

Zong, W., Chow, Y. -W., Susilo, W., Kim, J., & Le, N. T. (2022). Detecting Audio Adversarial Examples in Automatic Speech Recognition Systems Using Decision Boundary Patterns. Journal of Imaging, 8(12), 324. https://doi.org/10.3390/jimaging8120324

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop