Acoustic-Based Machine Condition Monitoring—Methods and Challenges
Abstract
:1. Introduction
2. Methods—Acoustic-Based Machine Condition Monitoring
2.1. Detection of Anomalous Sound
2.1.1. Autoencoder-Based Anomaly Detection
2.1.2. Gaussian Mixture Model-Based Anomaly Detection
2.1.3. Outlier Exposure-Based Anomaly Detection
2.1.4. Signal Processing Methods
2.2. Classification of Anomalous Sound
2.2.1. Feature-Based Machine Learning Methods
- (1)
- Feature Extraction
- (a)
- Time domain-based feature extraction
- (b)
- Frequency domain-based feature extraction
- (c)
- Time-frequency domain-based feature extraction
SN | Signal Analysis Domain | Features | Summary |
---|---|---|---|
1 | Time Domain | Zero Crossing Rate | The rate of sign-changes along a signal within a frame length. |
2 | Frequency Domain | Short-time Energy | The sum of squares of the signal values normalised by frame length. |
3 | Frequency Domain | Entropy of Energy | Shannon entropy of the normalised energies within a frame length. |
4 | Frequency Domain | Spectral Centroid | The centre of mass of the spectrum of a frame. Determined by the weighted mean of the frequencies present within the spectrum of a frame length. |
5 | Frequency Domain | Spectral Spread | The second central moment of the spectrum of a frame length |
6 | Frequency Domain | Spectral Entropy | Shannon entropy of the normalised spectral energies within the spectrum of a frame length. |
7 | Frequency Domain | Spectral Flux | The squared difference between the normalised magnitudes of the spectra of the two successive frame length. |
8 | Frequency Domain | Spectral Roll-off | This is the frequency below which 90% of the spectral distribution for the frame is concentrated. |
9 | Frequency Domain | MFCC | Mel-Frequency Cepstrum Coefficient (MFCC) provide an effective representation of sound which closely mimics the sound perception of the human ear. MFCC are determined by taking the linear Discrete Cosine Transform (DCT) of the log power spectrum on the nonlinear Mel scale. |
10 | Frequency Domain | Chroma Vector | A representation of the spectrum projected onto 12 bins representing the 12 distinct semitones (or chroma) of the musical octave. |
11 | Frequency Domain | Chroma Deviation | Standard deviation of the chroma vector. |
12 | Frequency Domain | Band-power ratio | Normalised spectral peaks within fault frequency band |
- (2)
- Classifier Learning Algorithms
- (a)
- K-Nearest Neighbors (KNN): KNN is a non-parametric and instance-based machine learning algorithm which can be used for both classification and regression [39,41]. It is classed as a non-parametric method because it makes no explicit assumption about the underlying distribution of the training data and an instance-based method because it does not learn a discriminative function from the training data but memorises it instead [39,41]. When KNN is used for classification, its input consists of the K closest training instances to the unknown instance in the feature space based on a similarity distance metric, e.g., Euclidean distance, hamming distance, Chebyshev distance, Minkowski distance, etc. The output class membership of the unknown instance is determined by a majority vote of its K nearest neighbors. Although KNN is a simplistic classifier model, it is very versatile (i.e., used in many applications), robust (i.e., tight error bounds) and often used as a benchmark for comparison with more complex classifiers [42,43].
- (b)
- Linear Support Vector Machine (SVM): SVM can be viewed as a discriminative classifier model defined by a separating hyperplane [39]. In a nutshell, when an SVM is given labeled training data, the algorithm outputs an optimal hyperplane which classifies new unseen data. The optimal hyperplane is determined by maximising the margin or distance between the nearest points (support vectors) to the hyperplane. Sometimes, the data are not linearly separable, SVM circumvents this by adopting either a soft margin parameter in the optimisation loss or using kernel tricks to transform the feature set into a higher dimensional space.
- (c)
- Random Forest: Random Forest is an ensemble method of learning based on contribution from multiple decision trees [39]. A decision tree is a simple model to classify a dataset, where the data is continuously split based on parameters such as information gain, Gini index, etc. When random forest is used as a classifier, each decision tree in the ensemble, makes a class prediction, and the class with the most vote is the model prediction. A key aspect of the random forest classifier model is that the decision trees are uncorrelated. To achieve uncorrelated decision trees, several techniques such as bagging and feature randomness during tree split are used. Bagging ensures that each individual tree, randomly sample from the dataset with replacement, thus producing different trees in the ensemble.
- (d)
- Decision Tree: Decision tree is used for solving classification problems by crafting a tree-structure where internal nodes represent data attributes, branches represent decision rules and end leaf nodes represent outcomes. It applies a hierarchical structure in determining patterns within data with the intent of creating decision-making rules and predicting regression relationships between dependent and independent variables [39,40]. Optimising the decision tree model, relevant hyperparameters are minimum leaf size, maximum number of split and split criteria, e.g., Gini index, information gain, etc.
- (e)
- Naive Bayes: Naive Bayes classifier rely on Bayes theorem for solving classification problems [39]. Bayes theorem provides a means to formalise the relationship of conditional probabilities or likelihoods of statistical variables. In Naive Bayes classifier, the interest lies in determining the posterior probability of a class label (Y) given some observed features, i.e., . Using Bayes theorem, this posterior probability is expressed as:
- (f)
- Artificial Neural Network (ANN)/Multi-Layer Perceptron (MLP): ANN or MLP is inspired by the brain biological neural system. It uses the means of simulating the electrical activity of the brain and nervous system interaction to learn a data-driven model. The structure of an ANN comprises of an input layer, one or more hidden layers and an output layer as shown in Figure 3 [39]. Each layer is made up of nodes or neurons and is fully connected to every node in the subsequent layers through weights (w), biases (b), and threshold/activation function. Information in the ANN move in two directions: feed forward propagation (i.e., operating normally) and backward propagation (i.e., during training). In the feedforward propagation, information arrives at the input layer neurons to trigger the connected hidden neurons in subsequent layer. All the neurons in the subsequent layer do not fire at the same time. The node would receive the input from previous node, this is multiplied by the weight of the connection between the neurons; all such inputs from connected previous neurons are summed at each neuron in the next layer. If these values at each neuron is above a threshold value based on chosen activation function, e.g., sigmoid function, hyperbolic tangent (tanh), rectified linear unit (ReLU), etc. the node would fire and pass on the output, or if less than the threshold value, it would not fire. This process is continued for all the layers and nodes in the ANN operating in the feedforward mode from the input layer to the output layer. The backward propagation is used to train the ANN network. Starting from the output layer, this process compares the predicted output with actual output per layer and updates the weights of each neuron connection in the layer by minimize the error using a technique such as gradient descent amongst others as shown in Figure 3. This way, the ANN model learns the relationship between the input and output.
SN | Classifier Learning Algorithms | Features | Application | Ref. |
---|---|---|---|---|
1 | SVM | Frequency domain signal analysis: Band-power ratio | Detection of air leaks between grate bars lined sinter strand pallets in a sintering plant | [36] |
2 | Decision Tree (J48/C4.5 Algorithm) | Frequency domain signal analysis: Band-power ratio | Detection of air leaks between grate bars lined sinter strand pallets in a sintering plant | [36] |
3 | Deep Neural Network (DNN) | Frequency domain signal analysis: Short-Term Fourier Transform (STFT) | Detecting changes in electric motor operational states such as supply voltage and load | [14] |
4 | Decision tree, Naive Bayes, kNN, SVM, Discriminant Analysis, Ensemble classifier, with Bayesian Optimisation | Frequency domain signal analysis: Wavelet packet transform, with Principal Component Analysis (PCA) | Detecting of internal combustion engine fault | [40] |
5 | kNN, SVM, and Multi-layer Perceptron (MLP) | Frequency domain signal analysis: Wavelet packet transform with various mother wavelets | Detecting of internal combustion engine fault | [44] |
6 | Artificial Neural Network (ANN) | Frequency domain signal analysis: Spectral peaks from the fast Fourier Transform of acoustic signal (0–2996.25 Hz) | Detecting loose stator coils in induction electric motors | [6] |
2.2.2. Acoustic Image-Based Deep Learning Methods
- (1)
- Acoustic Image Representation
- (a)
- Spectrogram: This is a two-dimensional representation of the frequency characteristics of a time-domain signal as it changes over time as shown in Figure 4. Spectrogram is generated using Fourier transform of the time-domain signal; the time-domain signal is first divided into smaller segments of equal length with some overlap; then, fast Fourier transform (FFT) is applied to each segment to determine its frequency spectrum; the resulting spectrogram becomes a side-by-side overlay of the frequency spectrum of each segment over time. FFT represents an algorithm to compute the discrete Fourier transform (DFT) of the windowed time-domain signal, represented as [16]:The spectrogram is obtained as the logarithm of the DFT, as such [16]:
- (b)
- (c)
- Cochleagram: A cochleagram is a time-frequency representation of the frequency filtering response of the cochlea (in the inner ear) as simulated by a bank of Gammatone filters [48]. The Gammatone filter represents a pure sinusoidal tone that is modulated by a Gamma distribution function; the impulse response of the Gammatone filter is expressed as [16]:
- (2)
- Deep Learning Methods
- (a)
- Convolution Neural Network (CNN): CNN is inspired from the operation of the mammalian visual cortex. As shown in Figure 7, CNN is a multi-stage neural network made up of key stages: filter stage (i.e., convolution layer, pooling layer, normalisation layer and activation layer) and classification stage (i.e., fully connected layer of multilayer perceptron) [51]. The convolution layer functions to extract feature set from acoustic image representation into a feature map, pooling layer reduces the dimensionality of the feature map, and the classification stage performs the classification task using the multilayer perceptron. [47] has applied CNN with a combination of log-spectrogram, short-time Fourier transform and log-Mel spectrogram features to classify rolling-element bearing cage fault based on acoustics signals. Implemented CNN model consisted of three stage feature extraction layers: fully connected layer (shape = 16 × 16, rectified linear unit (ReLU) activation function, max. pooling = 2 × 2), fully connected layer (shape = 32 × 32, ReLU, max. pooling = 2 × 2), and fully connected layer (shape = 64 × 64, ReLU, max. pooling = 2 × 2) and a final classification stage based on multi-layer perception with 512 hidden nodes, ReLU and sigmoid activation function. Dataset was very sparse, and model was not optimized; therefore, impacting model performance on training accuracy. Table 6 highlights other applications of acoustic image-based classifiers of anomalous machine operating sounds.
- (b)
- Recurrent Neural Network (RNN): RNN is a type of neural network which uses sequential data or time series data to learn. Unlike CNN, RNN have internal memory state (i.e., can be trained to hold knowledge about the past); this is possible as inputs and outputs are not independent of each other, prior inputs influence the current input and output; simply put, output from previous layer state are feed back to the input of the next layer state. As shown in Figure 8, x is input layer, h is middle layer (i.e., consist of multiple hidden layers) and y is output layer. W, V and U are the parameters of the network such as weights and biases. At any given time (t), the current input is constituted from the input x(t) and previous x(t − 1); as such the output from x(t − 1) is feedback into the input x(t) to improve the network output. This way, information cycles through a loop within the hidden layers in the middle layer. RNN uses the same network parameters for every hidden layer, such as: activation function, weights, and biases (W, V, U). Despite the flexibility of the basic RNN model to learning sequential data, they suffer from the vanishing gradient problem (i.e., difficulty training the model when the weights get too small, and the model stops learning) and exploding gradient problem (i.e., difficulty training the model due to very high weight assignment). To overcome these challenges, the long short-term memory (LSTM) network variant of RNN is normally used. LSTM has the capability to learn long-term dependencies between time steps of sequential data. LSTM can read, write and delete information from its memory. It achieves this via a gating process made up of three stages: forget gate, update/input gate and output gate which interacts with is long-term memory and short-term memory pathways used to feedback its memory states amongst hidden layers. As shown in Figure 9, “c” represents the cell state and long-term memory, “h” represents the hidden state and short-term memory, and “x” represent the sequential data input. The forget gate determines how much of the cell state “c” is thrown away or forgotten. The update gate determines how much of new information is going to be stored in the cell state, and output gate determines what is going to be outputted. [52] has applied LSTM RNN with cochleagram features to classify varying rolling-element bearing faults based on 60 s acoustics signals. Implemented model consisted of an input feature set based on 128 gammatone filter bank cochleagram; Considering a 1 s. duration as a frame, the 60 s dataset generated 60-time frames. Each frame is represented as a cochleagram. 67% of the dataset was used to train the LSTM RNN model and 33% for testing. Model accuracy on fault classification task was 94.7%. Table 6 highlights other applications of acoustic image-based classifiers of anomalous machine operating sounds.
- (c)
- Spiking Neural Network (SNN): SNN is a brain-inspired neural network where information is represented as binary events (spikes). It shares similarity with concepts such as event potentials in the brain. SNN incorporates time into its propagation model for information; SNN only transmit information when neuronal potential exceeds a threshold value. Working only with discrete timed events, SNS accepts as input spike train and outputs spike train. As such, information is required to be encoded into the spikes which is achieved via different encoding means: binary coding (i.e., all-or-nothing encoding with neurons active or inactive per time, rate coding, fully temporal codes (i.e., precise timing of spikes), latency coding, amongst others [53]. As shown in Figure 10, SNN is trained with the margin maximization technique, described in [54]. During first epoch, SNN hidden layer is developed based on neuron addition scheme. In subsequent epochs, the weights and biases of the hidden layer neurons are updated further using the margin maximization technique. Here, weights of the winner neuron are strengthened, while those of the others are inhibited; this reflects the Hebbian learning rule of the natural neurons; as a result, neurons are only connected to their local neurons, so they process the relevant input patterns together. This approach maximizes the margin among the classes which lends itself to training the spike patterns. Ref. [48] has applied SNN with cochleagram features to classify varying rolling-element bearing faults based on 10 s acoustics signals. Implemented model consisted of an input feature set based on 128 gammatone filter bank cochleagram; later reduced to 50 using principal component analysis (PCA). Considering a 10 ms duration as a frame, the 10 s dataset generated 1000-time frames. Each frame was encoded into a spike train using the population coding method. 90% of the dataset was used to train the SNN model and 10% for testing. Model accuracy was above 85%. Table 6 highlights other applications of acoustic image-based classifiers of anomalous machine operating sounds.
SN | Acoustic Image Representation | Deep Learning Methods | Application | Ref. |
---|---|---|---|---|
1 | Spectrogram/Log-Spectrogram | CNN * | Detection of rolling-element bearing fault such as cage defect | [47] |
RNN * | Detection of air leaks between grate bars lined sinter strand pallets in a sintering plant | [36] | ||
2 | Cochleagram | RNN * | Detection of rolling-element bearing fault such as inner race defect, outer race defect, rolling-element defect, combined defect, and heavily worn bearing | [52] |
3 | Cochleagram | SNN * | Detection of rolling-element bearing fault such as inner race defect, outer race defect, rolling-element defect, combined defect, and heavily worn bearing | [48] |
4 | Spectrogram (from STFT) | CNN * | Detection of rolling-element bearing fault such as cage defect | [47] |
5 | Log-Mel Spectrogram | CNN * | Detection of rolling-element bearing cage fault | [47,55] |
3. Datasets for Detection and Classification of Anomalous Machine Sound (DCAMS)
3.1. ToyADMOS Dataset
3.2. MIMII Dataset
3.3. DCASE Dataset
3.4. IDMT-ISA-ELECTRIC-ENGINE Dataset
3.5. MIMII DUE Dataset
3.6. ToyADMOS2 Dataset
3.7. MIMII DG Dataset
4. Challenges
4.1. Sound Mixtures with Background Noise
4.2. Domain Shift with Changes in Machine Operation and Background Noise
4.3. Domain Generalisation Invariant to Changes in Machine Operation and Background Noise
4.4. Effect of Measurement Distance, Measurement Device and Sampling Parameters
4.4.1. Measurement Distance (Microphones Positions)
4.4.2. Single Microphone Measurement Device and Sampling Parameters
4.4.3. Microphone Array Measurement (Acoustic Camera)
- (1)
- Delay and Sum Beamforming in the Time-Domain: This is demonstrated in Figure 18 as follows, considering only two sound sources as an example (i.e., source 1 and source 2). For each sound source, the travel path of emitted sound to the microphone array would be different; as such, captured signals by the microphone array would show different delays and phases for the measured signals from both sources. As both parameters, delay, and phase, are proportional to the travelled distance between microphone array and source; with the knowledge of the speed of sound in the medium (e.g., air), the runtime delay is estimated for the signal of interest (source 1) reaching all the microphone locations. The measured signal for every microphone in the array is then shifted by the calculated runtime delay for that channel, creating an alignment in phase in the time-domain for the signal of interest (source 1). The resulting signals from every microphone channel are summed and normalised by the number of microphones in the array; As shown in Figure 18, the signal of interest (source 1) is amplified due to constructive interference while source 2 is minimized due to destructive interference. To create the final acoustic scene representation, for each microphone channel, the root mean square (RMS) amplitude value or the maximum amplitude value of the time-domain acoustic signal can be evaluated for visualization as an acoustic map.
- (2)
- Delay and Sum Beamforming in the Frequency Domain: This is demonstrated in Figure 19 as follows, considering only two sound sources as an example (i.e., source 1 and source 2). For each sound source, the travel path of emitted sound to the microphone array would be different; as such, captured signals by the microphone array would show different delays and phases for the measured signals from both sources. The delay for the signal of interest can be determined using information such as, distance between source and microphone and the speed of sound in the medium. Fourier transform is performed at all microphone channel resulting in a complex spectrum for amplitude and phase. To eliminate the delay in phase for the signal of interest at all microphone location, the complex spectra is multiplied by a complex phase term as shown in Figure 19, bringing the interested acoustic source in phase without impacting the amplitude of the spectra. The resulting complex spectra from all the microphone channels are summed and normalised by the number of microphone channels. The interest sound source signal (source 1) is enhanced due to constructive interference, while source 2 is diminished due to destructive interference.
5. Outlook
5.1. Addressing Pitfalls in Acoustic Data Collection
5.2. Addressing Measurement Artifacts (i.e., Background Noise, and Distance Effect)
5.3. Improving Data-Driven Model Accuracy for Application: Domain Adaptation versus Domain Generalisation
5.4. Addressing Multi-Fault Diagnosis
5.5. Improving Acoustic Camera Spatial Detection of Machine Faults
6. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Gemmeke, J.F.; Ellis, D.P.W.; Freedman, D.; Jansen, A.; Lawrence, W.; Moore, R.C.; Plakal, M.; Ritter, M. Audio Set: An Ontology and Human-Labeled Dataset for Audio Events. In Proceedings of the 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), New Orleans, LA, USA, 5–9 March 2017; pp. 776–780. [Google Scholar]
- Cramer, J.; Wu, H.-H.; Salamon, J.; Bello, J.P. Look, Listen, and Learn More: Design Choices for Deep Audio Embeddings. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 3852–3856. [Google Scholar]
- Arandjelović, R.; Zisserman, A. Look, Listen and Learn. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017. [Google Scholar]
- Kong, Q.; Cao, Y.; Iqbal, T.; Wang, Y.; Wang, W.; Plumbley, M. PANNs: Large-Scale Pretrained Audio Neural Networks for Audio Pattern Recognition (Pretrained Models). IEEE/ACM Trans. Audio Speech Lang. Process. 2020, 28, 2880–2894. [Google Scholar] [CrossRef]
- Hershey, S.; Chaudhuri, S.; Ellis, D.P.W.; Gemmeke, J.F.; Jansen, A.; Moore, R.C.; Plakal, M.; Platt, D.; Saurous, R.A.; Seybold, B.; et al. CNN architectures for large-scale audio classification. In Proceedings of the 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), New Orleans, LA, USA, 5–9 March 2017; pp. 131–135. [Google Scholar] [CrossRef] [Green Version]
- Gaylard, A.; Meyer, A.; Landy, C. Acoustic Evaluation of Faults in Electrical Machines. In Proceedings of the 1995 Seventh International Conference on Electrical Machines and Drives (Conf. Publ. No. 412), Durham, UK, 11–13 September 1995; pp. 147–150. [Google Scholar]
- Kawaguchi, Y.; Endo, T. How Can We Detect Anomalies from Subsampled Audio Signals? In Proceedings of the 2017 IEEE 27th International Workshop on Machine Learning for Signal Processing (MLSP), Tokyo, Japan, 25–28 September 2017; pp. 1–6. [Google Scholar]
- Koizumi, Y.; Saito, S.; Uematsu, H.; Harada, N. Optimizing Acoustic Feature Extractor for Anomalous Sound Detection Based on Neyman-Pearson Lemma. In Proceedings of the 2017 25th European Signal Processing Conference (EUSIPCO), Kos, Greece, 28 August–2 September 2017; pp. 698–702. [Google Scholar]
- Koizumi, Y.; Saito, S.; Uematsu, H.; Harada, N.; Imoto, K. ToyADMOS: A Dataset of Minia-ture-Machine Operating Sounds for Anomalous Sound Detection. In Proceedings of the 2019 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics (WASPAA), New Paltz, NY, USA, 20–23 October 2019; pp. 313–317. [Google Scholar]
- Purohit, H.; Tanabe, R.; Ichige, T.; Endo, T.; Nikaido, Y.; Suefusa, K.; Kawaguchi, Y. MIMII Dataset: Sound Dataset for Malfunctioning Industrial Machine Investigation and Inspection. In Proceedings of the Detection and Classification of Acoustic Scenes and Events 2019 Workshop (DCASE2019), New York, NY, USA, 25–26 October 2019; pp. 209–213. [Google Scholar]
- Kawaguchi, Y.; Imoto, K.; Koizumi, Y.; Harada, N.; Niizumi, D.; Dohi, K.; Tanabe, R.; Purohit, H.; En-do, T. Description and Discussion on DCASE 2021 Challenge Task 2: Unsupervised Anomalous Sound Detection for Machine Condition Monitoring under Domain Shifted Conditions. arXiv 2021, arXiv:2106.04492. [Google Scholar]
- Dohi, K.; Imoto, K.; Harada, N.; Niizumi, D.; Koizumi, Y.; Nishida, T.; Purohit, H.; Endo, T.; Yamamoto, M.; Kawaguchi, Y. Description and Discussion on DCASE 2022 Challenge Task 2: Unsupervised Anomalous Sound Detection for Machine Condition Monitoring Applying Domain Generalization Tech-niques. arXiv 2022, arXiv:2206.05876. [Google Scholar]
- Koizumi, Y.; Kawaguchi, Y.; Imoto, K.; Nakamura, T.; Nikaido, Y.; Tanabe, R.; Purohit, H.; Suefusa, K.; Endo, T.; Yasuda, M.; et al. Description and Discussion on DCASE2020 Challenge Task2: Unsupervised Anomalous Sound Detection for Machine Condition Monitorin. In Proceedings of the Detection and Classification of Acoustic Scenes and Events 2020 Workshop (DCASE2020), Tokyo, Japan, 2–4 November 2020; pp. 81–85. [Google Scholar]
- Grollmisch, S.; Abeßer, J.; Liebetrau, J.; Lukashevich, H. Sounding Industry: Challenges and Datasets for Industrial Sound Analysis. In Proceedings of the 2019 27th European Signal Processing Conference (EUSIPCO), A Coruña, Spain, 2–6 September 2019; pp. 1–5. [Google Scholar]
- Koizumi, Y.; Saito, S.; Uematsu, H.; Kawachi, Y.; Harada, N. Unsupervised Detection of Anomalous Sound Based on Deep Learning and the Neyman–Pearson Lemma. IEEE/ACM Trans Audio Speech Lang Process 2019, 27, 212–224. [Google Scholar] [CrossRef] [Green Version]
- Sharan, R.; Xiong, H.; Berkovsky, S. Benchmarking Audio Signal Representation Techniques for Classification with Convolutional Neural Networks. Sensors 2021, 21, 3434. [Google Scholar] [CrossRef]
- Roche, F.; Hueber, T.; Limier, S.; Girin, L. Autoencoders for Music Sound Modeling: A Com-parison of Linear, Shallow, Deep, Recurrent and Variational Models. In Proceedings of the 16th Sound & Music Computing Conference (SMC 2019), Malaga, Spain, 28–31 May 2019. [Google Scholar]
- Bai, J.; Chen, C.; Chen, J. Feature Based Fusion System for Anomalous Sounds Monitoring. In Proceedings of the 5th Workshop on Detection and Classification of Acoustic Scenes and Events (DCASE), Tokyo, Japan, 2–4 November 2020. [Google Scholar]
- Ahmed, F.; Nguyen, P.; Courville, A. An Ensemble Approach for Detecting Machine Failure from Sound. In Proceedings of the 5th Workshop on Detection and Classification of Acoustic Scenes and Events (DCASE), Tokyo, Japan, 2–4 November 2020. [Google Scholar]
- Alam, J.; Boulianne, G.; Gupta, V.; Fathan, A. An Ensemble Approach to Unsupervised Anomalous Sound Detection. In Proceedings of the 5th Workshop on Detection and Classification of Acoustic Scenes and Events (DCASE), Tokyo, Japan, 2–4 November 2020. [Google Scholar]
- Morita, K.; Yano, T.; Tran, K.Q. Anomalous Sound Detection by Using Local Outlier Factor and Gaussian Mixture Model. In Proceedings of the 5th Workshop on Detection and Classification of Acoustic Scenes and Events (DCASE), Tokyo, Japan, 2–4 November 2020. [Google Scholar]
- Hendrycks, D.; Mazeika, M.; Dietterich, T. Deep Anomaly Detection with Outlier Exposure. In Proceedings of the Seventh International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
- Sandler, M.; Howard, A.G.; Zhu, M.; Zhmoginov, A.; Chen, L.-C. Inverted Residuals and Lin-ear Bottlenecks: Mobile Networks for Classification, Detection and Segmentation. In Proceedings of the IEEE conference on computer vision and pattern recognition, Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
- Heng, R.B.W.; Nor, M.J.M. Statistical Analysis of Sound and Vibration Signals for Monitor-ing Rolling Element Bearing Condition. Appl. Acoust. 1998, 53, 211–226. [Google Scholar] [CrossRef]
- Van Riesen, D.; Schlensok, C.; Henrotte, F.; Hameyer, K. Acoustic Measurement for Detecting Manufacturing Faults in Electrical Machines. In Proceedings of the 17th International Conference on Electrical Machines (ICEM), Chania, Greece, 2–5 September 2006. [Google Scholar]
- Benko, U.; Petrovcic, J.; Juričić, Đ.; Tavčar, J.; Rejec, J. An Approach to Fault Diagnosis of Vac-uum Cleaner Motors Based on Sound Analysis. Mech. Syst. Signal Process. 2005, 19, 427–445. [Google Scholar] [CrossRef]
- Mishra, R.; Gu, F.; Fazenda, B.; Stubbs, C.; Ball, A. Measurement and Characterisation of Faults in the Intake System of a Turbocharged Engine Using a Directional Acoustic Probe. In Proceedings of the COMADEM 2009, San Sebastian, Spain, 9–11 June 2009. [Google Scholar]
- Wu, Z.; Huang, N.E. Ensemble Emprical Mode Decomposition: A Noise-Assisted Data Analysis Method. Adv. Adapt. Data Anal. 2009, 1, 1–41. [Google Scholar] [CrossRef]
- Fazenda, B.M. Acoustic Based Condition Monitoring of Turbine Blades. In Proceedings of the 18th International Congress on Sound and Vibration, Rio de Janeiro, Brazil, 10–14 July 2011. [Google Scholar]
- Grebenik, J.; Bingham, C.; Srivastava, S. Acoustic Diagnostics of Electrical Origin Fault Modes with Readily Available Consumer-Grade Sensors. IET Electr. Power Appl. 2019, 13, 1946–1953. [Google Scholar] [CrossRef] [Green Version]
- Shiri, H.; Wodecki, J.; Ziętek, B.; Zimroz, R. Inspection Robotic UGV Platform and the Proce-dure for an Acoustic Signal-Based Fault Detection in Belt Conveyor Idler. Energies 2021, 14, 7646. [Google Scholar] [CrossRef]
- Fang, S.; Li, S.-C.; Zhen, D.; Shi, Z.; Gu, F.; Ball, A.D. Acoustic Feature Extraction for Monitor-ing the Combustion Process of Diesel Engine Based on EMD and Wavelet Analysis. Int. J. COMADEM 2017, 20, 25–30. [Google Scholar]
- Zhen, D.; Wang, T.; Gu, F.; Tesfa, B.; Ball, A. Acoustic Measurements for the Combustion Di-agnosis of Diesel Engines Fuelled with Biodiesels. Meas. Sci. Technol. 2013, 24, 055005. [Google Scholar] [CrossRef]
- Anami, B.S.; Pagi, V.B. Acoustic Signal-Based Approach for Fault Detection in Motorcycles Using Chaincode of the Pseudospectrum and Dynamic Time Warping Classifier. IET Intell. Transp. Syst. 2014, 8, 21–27. [Google Scholar] [CrossRef]
- Amarnath, M.; Sugumaran, V.; Kumar, H. Exploiting Sound Signals for Fault Diagnosis of Bearings Using Decision Tree. Measurement 2013, 46, 1250–1256. [Google Scholar] [CrossRef]
- Pasha, S.; Ritz, C.; Stirling, D.; Zulli, P.; Pinson, D.; Chew, S. A Deep Learning Approach to the Acoustic Condition Monitoring of a Sintering Plant. In Proceedings of the APSIPA Annual Summit and Conference 2018, Hawaii, HI, USA, 12–15 November 2018. [Google Scholar]
- Giannakopoulos, T. PyAudioAnalysis: An Open-Source Python Library for Audio Signal Analysis. PLoS ONE 2015, 10, e0144610. [Google Scholar] [CrossRef]
- Russell, S.J.; Norvig, P. Artificial Intelligence: A Modern Approach, 3rd ed.; Pearson: Boston, MA, USA, 2010. [Google Scholar]
- Raschka, S. Python Machine Learning; Packt Publishing: Birmingham, UK, 2015. [Google Scholar]
- Mathew, S.K.; Zhang, Y. Acoustic-Based Engine Fault Diagnosis Using WPT, PCA and Bayesian Optimization. Appl. Sci. 2020, 10, 6890. [Google Scholar] [CrossRef]
- Altman, N.S. An Introduction to Kernel and Nearest-Neighbor Nonparametric Regression. Am. Stat. 1992, 46, 175–185. [Google Scholar]
- Yao, Z.; Ruzzo, W.L. A Regression-Based K Nearest Neighbor Algorithm for Gene Function Prediction from Heterogeneous Data. BMC Bioinform. 2006, 7, S11. [Google Scholar] [CrossRef] [Green Version]
- Kleyko, D.; Osipov, E.; Papakonstantinou, N.; Vyatkin, V.; Mousavi, A. Fault Detection in the Hy-perspace: Towards Intelligent Automation Systems. In Proceedings of the 2015 IEEE 13th International Conference on Industrial Informatics (INDIN), Cambridge, UK, 22–24 July 2015. [Google Scholar]
- Ghaderi, H.; Kabiri, P. Automobile Engine Condition Monitoring Using Sound Emission. Turk. J. Electr. Eng. Comput. Sci. 2017, 25, 1807–1826. [Google Scholar] [CrossRef]
- Lyon, R.F. Machine Hearing: An Emerging Field [Exploratory DSP]. IEEE Signal Process. Mag. 2010, 27, 131–139. [Google Scholar] [CrossRef]
- Lyon, R.F. Machine Hearing: Audio Analysis by Emulation of Human Hearing. In Proceedings of the 2011 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics (WASPAA), New Paltz, NY, USA, 13 May 2011; p. viii. [Google Scholar]
- Jombo, G.; Shriram, A. Evaluation of 2D Acoustic Signal Representations for Acoustic-Based Machine Condition Monitoring. In Proceedings of the PECS 2022 Physics, Engineering and Computer Science Research Conference, Kavala, Greece, 21–23 June 2022. [Google Scholar]
- Zhang, Y.; Dora, S.; Martinez-Garcia, M.; Bhattacharyaand, S. Machine Hearing for Industrial Acoustic Monitoring Using Cochleagram and Spiking Neural Network. In Proceedings of the 2022 IEEE/ASME International Conference on Advanced Intelligent Mechatronics (AIM), Sapporo, Japan, 11–15 July 2022; pp. 1047–1051. [Google Scholar]
- Tanveer, M.H.; Zhu, H.; Ahmed, W.; Thomas, A.; Imran, B.M.; Salman, M. Mel-Spectrogram and Deep CNN Based Representation Learning from Bio-Sonar Implementation on UAVs. In Proceedings of the 2021 Interna-tional Conference on Computer, Control and Robotics (ICCCR), Singapore, 23–26 April 2021; pp. 220–224. [Google Scholar]
- Li, J.; Zhang, X.; Huang, L.; Li, F.; Duan, S.; Sun, Y. Speech Emotion Recognition Using a Du-al-Channel Complementary Spectrogram and the CNN-SSAE Neutral Network. Appl. Sci. 2022, 12, 9518. [Google Scholar] [CrossRef]
- Ogundokun, R.O.; Maskeliunas, R.; Misra, S.; Damaševičius, R. Improved CNN Based on Batch Normalization and Adam Optimizer. In International Conference on Computational Science and Its Applications; Gervasi, O., Murgante, B., Misra, S., Rocha, A.M.A.C., Garau, C., Eds.; Springer International Publishing: Cham, Switzerland, 2022; pp. 593–604. [Google Scholar]
- Zhang, Y.; Martínez-García, M. Machine Hearing for Industrial Fault Diagnosis. In Proceedings of the 2020 IEEE 16th International Conference on Automation Science and Engineering (CASE), Hong Kong, China, 20–21 August 2020; pp. 849–854. [Google Scholar]
- Guo, W.; Fouda, M.E.; Eltawil, A.M.; Salama, K.N. Neural Coding in Spiking Neural Net-works: A Comparative Study for Robust Neuromorphic Systems. Front. Neurosci. 2021, 15, 638474. [Google Scholar] [CrossRef] [PubMed]
- Dora, S.; Kasabov, N. Spiking Neural Networks for Computational Intelligence: An Overview. Big Data Cogn. Comput. 2021, 5, 67. [Google Scholar] [CrossRef]
- Mittel, D.; Pröll, S.; Kerber, F.; Schöler, T. Mel Spectrogram Analysis for Punching Machine Operating State Classification with CNNs. In Proceedings of the 2021 26th IEEE International Conference on Emerging Technologies and Factory Automation (ETFA), Vasteras, Sweden, 7–10 September 2021; pp. 1–4. [Google Scholar]
- Tanabe, R.; Purohit, H.; Dohi, K.; Endo, T.; Nikaido, Y.; Nakamura, T.; Kawaguchi, Y. MIMII DUE: Sound Dataset for Malfunctioning Industrial Machine Investigation and Inspection with Domain Shifts Due to Changes in Operational and Environmental Conditions. arXiv 2021, arXiv:2006.05822. [Google Scholar]
- Harada, N.; Niizumi, D.; Takeuchi, D.; Ohishi, Y.; Yasuda, M.; Saito, S. ToyADMOS2: Another Dataset of Miniature-Machine Operating Sounds for Anomalous Sound Detection under Domain Shift Conditions. arXiv 2021, arXiv:2106.02369. [Google Scholar]
- Dohi, K.; Nishida, T.; Purohit, H.; Tanabe, R.; Endo, T.; Yamamoto, M.; Nikaido, Y.; Kawaguchi, Y. MIMII DG: Sound Dataset for Malfunctioning Industrial Machine Investigation and Inspection for Domain Generalization Task. arXiv 2022, arXiv:2205.13879. [Google Scholar]
- Ubhayaratne, I.; Xiang, Y.; Pereira, M.; Rolfe, B. An Audio Signal Based Model for Condition Monitoring of Sheet Metal Stamping Process. In Proceedings of the 2015 IEEE 10th Conference on Industrial Electronics and Ap-plications (ICIEA), Auckland, New Zealand, 15–17 June 2015; pp. 1267–1272. [Google Scholar]
- Wang, G.; Wang, Y.; Min, Y.; Lei, W. Blind Source Separation of Transformer Acoustic Signal Based on Sparse Component Analysis. Energies 2022, 15, 6017. [Google Scholar] [CrossRef]
- Michau, G.; Fink, O. Domain Adaptation for One-Class Classification: Monitoring the Health of Critical Systems Under Limited Information. arXiv 2019, arXiv:1907.09204. [Google Scholar]
- Wang, W.; Wang, H.; Ran, Z.-Y.; He, R. Learning Robust Feature Transformation for Domain Adaptation. Pattern Recognit. 2021, 114, 107870. [Google Scholar] [CrossRef]
- Schneider, J. Domain Transformer: Predicting Samples of Unseen, Future Domains. arXiv 2021, arXiv:2106.06057. [Google Scholar]
- Yang, Z.; Bozchalooi, I.S.; Darve, E. Anomaly Detection with Domain Adaptation. arXiv 2020, arXiv:2006.03689. [Google Scholar]
- Kumagai, A.; Iwata, T.; Fujiwara, Y. Transfer Anomaly Detection by Inferring Latent Domain Representations. In Proceedings of the 33rd Conference on Neural Information Processing Systems (NeurIPS 2019), Vancouver, BC, Canada, 8–14 December 2019; pp. 2471–2481. [Google Scholar]
- Wang, Q.; Michau, G.; Fink, O. Domain Adaptive Transfer Learning for Fault Diagnosis. In Proceedings of the 2019 Prognostics and System Health Management Conference (PHM-Paris), Paris, France, 2–5 May 2019; pp. 279–285. [Google Scholar]
- Yamaguchi, M.; Koizumi, Y.; Harada, N. AdaFlow: Domain-Adaptive Density Estimator with Application to Anomaly Detection and Unpaired Cross-Domain Translation. In Proceedings of the ICASSP 2019–2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 3647–3651. [Google Scholar]
- Motiian, S.; Jones, Q.; Iranmanesh, S.M.; Doretto, G. Few-Shot Adversarial Domain Adaptation. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; pp. 6673–6683. [Google Scholar]
- Zhang, W.; Shen, L.; Zhang, W.; Foo, C.-S. Few-Shot Adaptation of Pre-Trained Networks for Domain Shift. arXiv 2022, arXiv:2205.15234. [Google Scholar]
- Zhou, K.; Liu, Z.; Qiao, Y.; Xiang, T.; Loy, C.C. Domain Generalization: A Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2022. [Google Scholar] [CrossRef]
- Li, W.; Tsai, Y.P.; Chiu, C.L. The Experimental Study of the Expert System for Diagnosing Unbalances by ANN and Acoustic Signals. J. Sound Vib. 2004, 272, 69–83. [Google Scholar] [CrossRef]
- Siemens. Sound Fields: Free versus Diffuse Field, Near versus Far Field. 2020. Available online: https://community.sw.siemens.com/s/article/sound-fields-free-versus-diffuse-field-near-versus-far-field (accessed on 10 December 2022).
- ISO 3745:2012; Acoustics—Determination of Sound Power Levels and Sound Energy Levels of Noise Sources Using Sound Pressure—Precision Methods for Anechoic Rooms and Hemi-Anechoic Rooms. International Standards Organization: Geneva, Switzerland, 2012.
- Nave, C.R. Inverse Square Law, Sound. HyperPhysics. 2002. Available online: http://hyperphysics.phy-astr.gsu.edu/hbase/Acoustic/invsqs.html (accessed on 9 December 2022).
- Mesaros, A.; Heittola, T.; Virtanen, T. A Multi-Device Dataset for Urban Acoustic Scene Clas-sification. In Proceedings of the Detection and Classification of Acoustic Scenes and Events 2018 Workshop (DCASE2018), Surrey, UK, 19–20 November 2018; pp. 9–13. [Google Scholar]
- Erić, M.M. Some Research Challenges of Acoustic Camera. In Proceedings of the 2011 19thTelecommunications Forum (TELFOR) Proceedings of Papers, Belgrade, Serbia, 22–24 November 2011; pp. 1036–1039. [Google Scholar]
- Cariou, C.; Delverdier, O. Localizing Aircraft Noise Sources with Large Scale Acoustic An-tenna. In Proceedings of the 27th International Congress of the Aeronautical Sciences, Nice, France, 19–24 September 2010. [Google Scholar]
- Guidati, S. Advanced Beamforming Techniques in Vehicle Acoustics. In Proceedings of the 3rd Berlin Beamforming Conference, Berlin, Germany, 21–24 February 2010. [Google Scholar]
- Belsak, A.; Prezelj, J. Analysis of Vibrations and Noise to Determine the Condition of Gear Units. In Advances in Vibration Analysis Research; IntechOpen: London, UK, 2011; Available online: https://www.intechopen.com/chapters/14665 (accessed on 9 December 2022).
- Coutable, P.; Thomas, J.-H.; Pascal, J.-C.; Eveilleau, F. Bearing Fault Detection Based on Near-Field Acoustic Holography. In Proceedings of the International Conference Surveillance 6, Compiegne, France, 25–26 October 2011. [Google Scholar]
- Gfai Tech. The Acoustic Camera as an Innovative Tool for Fault Detection and Quality Control. 2022. Available online: https://www.gfaitech.com/applications/quality-control (accessed on 22 December 2022).
- Chiariotti, P.; Martarelli, M.; Castellini, P. Acoustic Beamforming for Noise Source Localization—Reviews, Methodology and Applications. Mech. Syst. Signal. Process. 2019, 120, 422–448. [Google Scholar] [CrossRef]
- Gfai Tech. How Does Delay-and-Sum Beamforming in the Time Domain Work? 2022. Available online: https://www.gfaitech.com/knowledge/faq/delay-and-sum-beamforming-in-the-time-domain (accessed on 22 December 2022).
- Gfai Tech. How Does Delay-and-Sum Beamforming in the Frequency Domain Work? 2022. Available online: https://www.gfaitech.com/knowledge/faq/delay-and-sum-beamforming-in-the-frequency-domain (accessed on 22 December 2022).
- Cabada, E.C.; Hamzaoui, N.; Leclere, Q.; Antoni, J. Acoustic Imaging Applied to Fault Detection in Rotating Machine. In Proceedings of the International Conference Surveillance 8, Roanne, France, 21–22 October 2015. [Google Scholar]
- Cardenas Cabada, E.; Leclere, Q.; Antoni, J.; Hamzaoui, N. Fault Detection in Rotating Ma-chines with Beamforming: Spatial Visualization of Diagnosis Features. Mech. Syst. Signal. Process. 2017, 97, 33–43. [Google Scholar] [CrossRef]
- Antoni, J.; Randall, R.B. The Spectral Kurtosis: Application to the Vibratory Surveillance and Diagnostics of Rotating Machines. Mech. Syst. Signal. Process. 2006, 20, 308–331. [Google Scholar] [CrossRef]
- Antoni, J. The Spectral Kurtosis: A Useful Tool for Characterising Non-Stationary Signals. Mech. Syst. Signal. Process. 2006, 20, 282–307. [Google Scholar] [CrossRef]
- Kosmider, M. Spectrum Correction: Acoustic Scene Classification with Mismatched Recording Devices. In Proceedings of the Interspeech 2020, Shanghai, China, 25–29 October 2020. [Google Scholar]
- Upadhyay, N.; Karmakar, A. Speech Enhancement Using Spectral Subtraction-Type Algorithms: A Comparison and Simulation Study. Procedia Comput. Sci. 2015, 54, 574–584. [Google Scholar] [CrossRef] [Green Version]
- Wildeboer, R.R.; Sammali, F.; Van Sloun, R.J.G.; Huang, Y.; Chen, P.; Bruce, M.; Rabotti, C.; Shulepov, S.; Salomon, G.; Schoot, B.C.; et al. Blind Source Separation for Clutter and Noise Suppression in Ultrasound Imaging: Review for Different Applications. IEEE Trans. Ultrason. Ferroelectr. Freq. Control 2020, 67, 1497–1512. [Google Scholar] [CrossRef] [PubMed]
- Farahani, A.; Voghoei, S.; Rasheed, K.; Arabnia, H.R. A Brief Review of Domain Adaptation. In Advances in Data Science and Information Engineering. Transactions on Computational Science and Computational Intelligence; Stahlbock, R., Weiss, G.M., Abou-Nasr, M., Yang, C.Y., Arabnia, H.R., Deligiannidis, L., Eds.; Springer: Cham, Switzerland, 2021. [Google Scholar] [CrossRef]
- Liu, X.; Yoo, C.; Xing, F.; Oh, H.; el Fakhri, G.; Kang, J.-W.; Woo, J. Deep Unsupervised Domain Adaptation: A Review of Recent Advances and Perspectives. APSIPA Trans. Signal. Inf. Process 2022, 11, e25. [Google Scholar] [CrossRef]
- Wang, J.; Lan, C.; Liu, C.; Ouyang, Y.; Qin, T. Generalizing to Unseen Domains: A Survey on Domain Generalization. In Proceedings of the 30th International Joint Conference on Artificial Intelligence (IJCAI-21), Montreal, QC, Canada, 19–26 August 2021. [Google Scholar]
- Li, Z.; Lv, Y.; Yuan, R.; Zhang, Q. Multi-Fault Diagnosis of Rotating Machinery via Iterative Multivariate Variational Mode Decomposition. Meas. Sci. Technol. 2022, 33, 125104. [Google Scholar] [CrossRef]
Input | Autoencoder Topology | Ref. |
---|---|---|
Frequency domain signal analysis: Log Mel spectrogram | Input layer
Dense layer (layers 1–4)
| [13] |
Frequency domain signal analysis: Log Mel spectrogram, MFCC, Spectrogram, Harmonic Percussive Source Separation (HPSS-h [harmonic], HPSS-p [percussive]) | Input layer
Dense layer (layers 1–4)
| [18] |
Frequency domain signal analysis: Log Mel spectrogram | Autoencoder architecture as [13] | [9,10] |
SN | Signal Analysis Domain | Acoustic Features | Ref. |
---|---|---|---|
1 | Time Domain | Zero Crossing Rate, | [19] |
Mean, Max, Min, Covariance of the waveform | |||
Short-time Energy, | |||
Frequency Domain | Entropy of Energy, | ||
Spectral Centroid, | |||
Spectral Spread, | |||
Spectral Entropy, | |||
Spectral Flux, | |||
Spectral Roll-off, | |||
MFCC, | |||
Chroma Vector, | |||
Chroma Deviation | |||
2 | Frequency Domain | Fisher Vectors | [20] |
3 | Frequency Domain | Log Mel Spectrogram | [21] |
Input | OE Topology | Ref. |
---|---|---|
Frequency domain signal analysis: Log Mel spectrogram | Input layer
| [11] |
Toy Car | Toy Conveyor | Toy Train | |||
---|---|---|---|---|---|
Parts | Condition | Parts | Condition | Parts | Condition |
Shaft | Bent | Tension pulley | Excessive tension | First carriage | Chipped wheel axle |
Gears | Deformed Melted | Tail pulley | Excessive tension Removed | Last carriage | Chipped wheel axle |
Tires | Coiled (plastic ribbon) Coiled (steel ribbon) | Belt | Attached metallic object 1 Attached metallic object 2 Attached metallic object 3 | Straight railway track | Broken Obstructing stone Disjointed |
Voltage | Over voltage Under voltage | Voltage | Over voltage Under voltage | Curved railway track | Broken Obstructing stone Disjointed |
Machine Type | Operations | Examples of Anomalous Conditions |
---|---|---|
Valve | Open/close repeat with different timing | More than two kinds of contamination |
Pump | Suction from discharge to a water pool | Leakage, contamination, clogging, etc. |
Fan | Normal work | Unbalanced, voltage change, clogging, etc. |
Slide rail | Slide repeat at different speeds | Rail damage, loose belt, no grease, etc. |
Datasets | Microphone Brand | Microphone Type | Sampling Frequency | Sample Duration | Ref. |
---|---|---|---|---|---|
ToyADMOS | Shure SM11-CN | Omni-directional Microphone | 16 kHz (Downsampled) | 10 s, and 10 min | [9] |
MIMII | TAMAGO-03 (Circular microphone array with 8 distinct microphones) | - | 16 kHz | 10 s. | [10] |
IDMT-ISA-ELECTRIC-ENGINE | - | - | 44.1 kHz | 3 s. | [14] |
MIMII DUE | TAMAGO-03 (Circular microphone array with 8 distinct microphones) | - | 16 kHz | 10 s. | [56] |
ToyADMOS2 | Shure SM11-CN TOMOCA EM-700 | Omni-directional Microphone Condenser Microphone | 48 kHz | 12 s. | [57] |
MIMII DG | TAMAGO-03 (Circular microphone array with 8 distinct microphones) | - | 16 kHz | 10 s. | [58] |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Jombo, G.; Zhang, Y. Acoustic-Based Machine Condition Monitoring—Methods and Challenges. Eng 2023, 4, 47-79. https://doi.org/10.3390/eng4010004
Jombo G, Zhang Y. Acoustic-Based Machine Condition Monitoring—Methods and Challenges. Eng. 2023; 4(1):47-79. https://doi.org/10.3390/eng4010004
Chicago/Turabian StyleJombo, Gbanaibolou, and Yu Zhang. 2023. "Acoustic-Based Machine Condition Monitoring—Methods and Challenges" Eng 4, no. 1: 47-79. https://doi.org/10.3390/eng4010004
APA StyleJombo, G., & Zhang, Y. (2023). Acoustic-Based Machine Condition Monitoring—Methods and Challenges. Eng, 4(1), 47-79. https://doi.org/10.3390/eng4010004