Next Article in Journal
A Revisiting Method Using a Covariance Traveling Salesman Problem Algorithm for Landmark-Based Simultaneous Localization and Mapping
Next Article in Special Issue
Classification of VLF/LF Lightning Signals Using Sensors and Deep Learning Methods
Previous Article in Journal
General Mental Health Is Associated with Gait Asymmetry
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Classification of Low Frequency Signals Emitted by Power Transformers Using Sensors and Machine Learning Methods

1
Department of Computer Science and Automatics, University of Bielsko-Biala, 43-309 Bielsko-Biala, Poland
2
Institute of Electric Power Engineering and Renewable Energy, Opole University of Technology, 45-758 Opole, Poland
*
Author to whom correspondence should be addressed.
Sensors 2019, 19(22), 4909; https://doi.org/10.3390/s19224909
Submission received: 14 October 2019 / Revised: 3 November 2019 / Accepted: 8 November 2019 / Published: 10 November 2019
(This article belongs to the Special Issue Artificial Intelligence and Sensors)

Abstract

:
This paper proposes a method of automatically detecting and classifying low frequency noise generated by power transformers using sensors and dedicated machine learning algorithms. The method applies the frequency spectra of sound pressure levels generated during operation by transformers in a real environment. The spectra frequency interval and its resolution are automatically optimized for the selected machine learning algorithm. Various machine learning algorithms, optimization techniques, and transformer types were researched: two indoor type transformers from Schneider Electric and two overhead type transformers manufactured by ABB. As a result, a method was proposed that provides a way in which inspections of working transformers (from background) and their type can be performed with an accuracy of over 97%, based on the generated low-frequency noise. The application of the proposed preprocessing stage increased the accuracy of this method by 10%. Additionally, machine learning algorithms were selected which offer robust solutions (with the highest accuracy) for noise classification.

1. Introduction

The operation of medium- and high-power transformers is associated with low frequency noise emission into the environment, the main source of which, among other things, include: cooler fans of the induced air circulation, insulation oil circulating pumps, and the magneto-strictive vibrations of the core. In very simple terms, magnetostriction means that if a piece of magnetic sheet steel is magnetized, it extends. When the source of magnetization is removed, it goes back to its original state. A transformer is magnetically excited by alternating voltage and current so that it extends and contracts twice during a full cycle of magnetization.
Identifying the problem of low frequency noise generated by new and worn out power transformers requires a wide range of tests. The reference measurement methodologies, methods of analysis, and the assessment of the nuisance caused by the generated noise should be clearly defined. Power transformers are considered as strong sources of low-frequency noise, with the most important spectrum components, in view of the noise level, being in the frequency range below 400 Hz. The initial results obtained in the experimental tests and measurements on real units can be referenced to the maximum allowable low frequency noise levels, which are defined in the standards relating to the working environment. Therefore, it is vital to be able to detect the low frequencies of power transformers and their levels (specificity).
The aim of this paper is to propose a method for the automatic detection of noise generated by a power transformer. The method makes it possible not only to detect the presence of noise from a transformer in the area, but also to determine the transformer type, its internal construction, and apparent power. Furthermore, the type of misclassification (i.e. detection of a transformer with higher apparent power) can be used as an initial diagnostic tool that detects the changing parameters of the device. It is worth noting that low-noise measurements can be made during the normal operation of the transformer (online).
The proposed method involves data acquisition using a dedicated low frequency sensor, a preprocessing stage (applying frequency analysis), and a classification stage using machine learning algorithms.
The scope of the analysis reported here includes the determination of waveforms showing changes in the sound pressure level as a function of frequency (amplitude spectra). For comparative purposes, the characteristics were determined separately for the examined transformers and background noise at the selected measurement points.
The paper is organized as follows. Section 2 describes previous works related to the low frequency sound analysis for transformers and machine learning methods. The proposed method for transformer detection based on sound pressure is presented in Section 3. Section 4 includes the experimental results and a comparison of the introduced method with state-of-the-art approaches. The conclusions and future research directions are given in Section 5.

2. Related Works

Infrasound noise is taken to be called noise, in which the spectrum is dominated by low frequencies, i.e., up to 20 Hz, in accordance with the PN-ISO 7196: 2002 standard [1]. There is no standardization of low frequency sounds; however, most scientists consider such frequencies to include the range between 10 and 200 Hz [2].
As part of the initial research, an attempt was made to determine the emission level of low-frequency signals generated by power transformers at rated conditions. The distribution transformers reducing the voltage from 15 kV to 0.4 kV of various types (indoor and overhead) and apparent powers (400 kVA and 2000 kVA) were tested [3].
The research conducted to date has demonstrated that power transformers are a source of low frequency signals. Our studies showed that they are characterized by similar waveforms of averaged amplitude spectra in terms of shape, as well as a similar character of time-frequency changes. The waveforms characteristic of the recorded sound pressure level has dynamically decreasing values, which occur within a frequency range of 10 Hz to 100 Hz [3,4]. In the conducted research, the Brüel & Kjær measuring equipment was applied, focusing only on the measurement of the sound pressure level without converting it to the sound power level. This approach can be effectively applied to determine the range of potential impact of low-frequency sound directly at the location of the measuring point.
Similar research was conducted for wind turbines, which are also covered by a wide range of research on infrasound and low-frequency noise [2]. Usually, the scope of the analysis includes the development of curves such as hysteresis that present variations in parameters over time, and the designation of the frequency spectra of the recorded low frequency signals for different meteorological conditions, often including wind speeds and direction. Usually, the sound pressure level of infrasonic noise for a given wind speed is determined as an arithmetic mean from all recorded sound pressure levels for the speed. A commonly used method for spectrogram analysis is the short time Fourier transform (STFT). The next step in the research study involved the analysis of the frequency spectra of sound pressure levels corrected with G frequency characteristics, which can be employed in demonstrating noise levels in conditions that are audible to the human ear [5,6]. Other studies present the effect of infrasound noise related to everyday human activities [7].
To date, measurements and signal analyses of power transformers were conducted by experts. This paper proposes the automation of a process involving the detection of operational transformers in the vicinity and their classification using Machine Learning algorithms (ML), which makes it possible to find complex relations and rules via data mining techniques. ML roots can be found in pattern recognition and computational learning theory. The method uses learning algorithms and example data (training set) to build the model, which can be adopted for classification or prediction purposes. There are three main categories of learning algorithms: supervised, unsupervised, and reinforced. In the case of supervised learning [8], a training set is given with the correct target values. In case of unsupervised learning [9], we tend to find the relation in some given data without knowing its original (correct) classification. Reinforcement learning [10] solves optimization problems, learning the optimal actions for a given situation. This paper is focused on supervised learning, due to the characteristics of the processed data. The classification algorithms commonly found in the literature includes: k-Nearest Neighbors [11], Naive Bayes Classification [12], Support Vector Machine [13,14], Random Forests [15], Bagging [16], and various types of Neural networks [17,18,19,20].
One of the major objectives in ML is to identify the useful internal representation of the input data by means of preprocessing methods, which transfer them into a new variable space. Preprocessing can simplify the model and improve its accuracy. However, in this case, a deeper understanding of the researched phenomenon is needed (i.e., a heuristics-based approach) to find the optimal feature representation [9] corresponding to the analyzed data. Commonly applied preprocessing methods include Principal Component Analysis [21,22,23,24] and Canonical Correlation Analysis [25,26].
It is worth noting that Neural Networks (especially deep ones), due to their specificity, find applications in both classification and pattern recognition. This approach overcomes the shortcomings of linear models by allowing a defined number of basic function parameters to learn within a defined network. There are many different types of neural networks (distinct by neurons construction and organization). They are easily adaptable to regression and classification problems [17,18]. Feed Forward Neural Networks (FFNN), also known as Multilayer Perceptron (MLP), are the most common type of neural networks. Based on the number of hidden layers, the network can solve more complex, non-linear problems. Unfortunately, they require a large amount of computation for their training (based on network complexity). Finding the optimum set of weights in cases of multiple hidden layer structures is a NP-complete problem [27]. Therefore, an alternative for more than one hidden layer MLP networks was proposed (deep neural network), in which layers have their functions, i.e., analyses of higher-level features based on the low-level features of previous layers [17,20]. Significant results using deep neural networks have led them becoming the most common classifiers in machine learning [27,28].
Previously presented ML algorithms have been proposed for sets of data that are independent. However, in the considered case, the measured time series contains sound pressure values of a defined period of time which is characterized by significant sequential correlations [9,29], and should be represented as a temporal feature. This task can be performed by sequence classification, in which the sequences of time series are taken into consideration [29]. There are a variety of machine learning models and methods that can perform these tasks. Examples of these models and methods appear under the names of Markov models [30], sliding-window methods [31], Kalman filters [32], conditional random fields [33], recurrent neural networks [34], graph transformer networks [17], the Welch method, and maximum entropy Markov models [35]. Further analysis can be found in [36]. The Welch method [37] makes it possible to determine the estimated spectral power density of the signal. As research shows [37], this method can be employed to minimize the effect of external noise, by averaging/smoothing the instantaneous spectrum. Additionally, it can be applied to identify the frequencies that could contain useful information for classification purposes. Therefore, in the proposed model, the Welsh method will be used in the preprocessing stage with the proposed feature discrimination method and various ML algorithms.

3. Methodology and Characteristics of the Analyzed Power Transformers

Based on previous research concerned with the transformer infrasound noise [1,3], it was observed that noise could be detected in all frequency spectra, but that some frequencies were found to be dominant. On the basis of this observation, a method was proposed to find the optimal frequency intervals and automatically detect noise using a selected machine learning algorithm. An algorithm providing the overview of this method is presented in Figure 1.
In a first step, the low frequency sound generated by the transformer or background is registered using a dedicated sensor device. The data are represented as time series X = [x1, x2, …, xn], where n is the number of samples. To reduce the volume of data in time series X, it is transformed into frequency domain vector F by the application of the Welch method. Then, using parameters P = [l, h, s, m1, m2, …, mk], the model is tuned. The parameters l = [2, 100], h = [2, 100], and l < h defines respectively lower and upper frequency bands, s N influences sample resolution, and mi, where i = 1, …, k, defines the ML algorithm parameters. In the first phase of the method (Figure 1a), initial parameters P’ are selected. The vector F size is reduced using l, h, and s parameters, and then it is used (together with class labels) to train the ML (tuned by mx parameters) using ten-fold cross-validation to increase the soundness of the result. The obtained classification accuracy (acc) is used to calculate fitness function ff (Equation (1)).
f f = a c c h l s × 10000 ,
where a c c   =   T P + T N T P + T N + F P + F N , TP is the true positive (correct classification), TN is the true negative (correct rejection), FP is false positive (type I error), and FN is false negative (type II error).
The proposed fitness function ensures that the result with highest accuracy will be selected, while results with fewer samples will be favored in cases when a comparable level of accuracy level is gained.
The number of iterations depends on the chosen optimization method. This research involved exhaustive search and heuristic methods. The parameters for the iteration with the highest ff function value and generated model M are used in the verification phase to estimate the model’s robustness.
Several optimization algorithms and ML methods were analyzed to find the optimal solution. In this section, all the used algorithms are described and discussed.

3.1. The Data Acquisition Method

The sensor device comprises a microphone type 4190, designed for accurate free field measurements, connected to a preamplifier type 2669L from Brüel & Kjær (Nærum, Denmark) and a digital signal meter with registration function LAN-XI type 3050-A-60, also from Brüel & Kjær. It was used to register low frequency signals, as shown in Figure 2. It is a professional tool used to measure sound pressure, intensity, and vibrations. Its implementation possibilities are wide ranging, i.e., from typical acoustic tests, such as noise measurements, the determination of sound power levels, noise mapping using beamforming techniques, testing the acoustic properties of materials, and determining the acoustic parameters of rooms, to specialist acoustic tests, such as machine diagnostics, modal analyses, and electroacoustic tests of acoustic transducers [2]. In the case of the used set, the range of the measured frequencies varied from 0.7 Hz to 20 kHz.
Before starting the measurements, the system was calibrated using the Brüel & Kjær type 4231 acoustic calibrator.
To operate the meter, the computer was used with dedicated software which was connected to a LAN cable measuring system. All operating parameters were defined using the PULSE LabShop application version 15.1.0, which forms an integral part of the setup shown in Figure 3. Apart from the option of the precise configuration of the device, this software provides tools with which to record measured signals and preprocess and visualize them in offline mode. This is a dedicated software package developed by Brüel & Kjær Sound & Vibration Measurement A/S.
The measurement was carried out through a continuous, multi-hour process using a sampling frequency of 51.2 kHz (full registration with listening capability). In addition, all analyzed power transformers were located away from major roads and motorways.

3.2. Feature Extraction and Optimization Methods

The feature extraction from time series X was performed using fast Fourier transform with a Hamming window. The procedure is called the Welch method [37]; it allowed us to determine the estimated spectral power density of the signal. The method aims to minimize the influence of external noise by averaging/smoothing the instantaneous spectrum. The parameters of the method were adapted to suit the characteristics of the test apparatus and the analyzed frequency range [2].
As a result, the vector F = [f1, f2, …, fn] was generated, which defines the values of sound pressure for a given frequency, in this case, a low frequency (2–100 Hz). The n value depends on the resolution of transformation. In this case, df = 0.125 Hz, yielding n = 784 values. To decrease quantity of input data, its resolution can be modified by the s, s N . parameter. The F’ vector was defined using the s parameter value (Equation (2)):
F = [ f 1 , f 2 , , f i , , f [ n s ] ] , f i = f ( i 1 ) * s + 1 , f i F .
For instance, for s = 8, the reduced vector stores only values representing data for a 1 Hz resolution (99 features). Finally, the vector F’ is further reduced, and it contains only values for a given frequency interval. The operation is performed using two filters tuned by the lower (l N ) and upper band (h N ) parameters. As a result, the final vector of features is generated (F″) (Equation (3)):
F = [ f 1 , f 2 , , f j , , f [ h l d f * s ] + 1 ] , f j = f [ l 2 d f * s + j ] , f F
which is used as an input for the ML algorithm. It is worth noting that l, h, and s are part of P parameter vector.
The search for optimal P vector values was performed using various optimization methods. Due to the small intervals of the analyzed frequency, it was possible to use an exhaustive method (EM) that made it possible to search through all the available parameter regions to find the optimal one. Nevertheless, several additional methods were also researched. The first is called Hill Climbing (HC), and starts with random parameter values. Then, the optimal solution was searched for among the surrounding values. If no neighbors improved the result, the optimization procedure was terminated. The second one, called random search (RS), also selects the initial parameters randomly, then the next position is selected randomly within the search space. The algorithm ends after the definition of 1000 steps or if the result does not improve over 20 steps. Finally, the Bayesian optimization (TPE) strategy was used which consists of two phases; the first is the warm-up, in which parameter combinations are randomly selected and evaluated. Based on the scores of the warm-up rounds, the second phase tries to find promising parameter combinations, which are then evaluated [38].

3.3. Machine Learning Model (ML)

The machine learning model M was designed and developed using reduced training data set F″ and parameters mx. The various ML algorithms were researched to check their applicability for the purposes of this task. They include k-Nearest Neighbors (KNN), Multilayer Perceptrons network (MLP), Classical Support Vector Machines (SVMs), and the Bayes approach. The implementation of these methods was based on the Weka library [39]. The KNN method classifies a new data vector by looking at the k given vectors that are closest to it in the feature space. In the proposed method, Euclidean distance, with k = m1 as a parameter, was selected. For the case of probabilistic classifiers, we can apply those that identify the naive Bayes family on the basis of Bayes’ theorem with the assumption of independence among the features; thus, no parameters are required. The SVMs non-probabilistic approach aims to identify hyperplanes that separate search classes. In this research, linear SVM was used that finds a hyperplane that is a linear function of the input features. Several parameters, like normal vector to the hyperplane by w and the parameter for controlling the offset b, as well as variable ξi, were preset based on research reported in [13,14]. For the random forests method, where instead of training a single tree, many trees are trained, the number of trees was defined as t = m1 parameter. Based on the initial research, the maximal number of trees was set to 10. Above this level, no improvement in accuracy was recorded. The MLP was selected to be representative of an artificial neural network. It proved adequate for the purposes of classification tasks. In this case, ≤ 2 hidden-layer (L = m1) and ≤ 20-neurons-per-layer (m2) structures were considered.

4. Results and Discussion

The proposed method was tested in a real-world testbed. The measurements of low frequency noise generated by the considered transformers were performed from a single reference point at a distance of 50 m from the sound source. The objects under study were four power transformers with various apparent powers and construction types. A detailed description of the transformers, with the sound pressure level, measured in the full frequency spectrum and narrowed to the low-frequency band, is provided in Table 1. Moreover, measurements were made in two series: firstly, when the transformer was nominally loaded, and secondly, when it was turned off. Figure 4 contains illustrative photos of the tested objects.
The scope of the study includes the analysis of the frequency spectra of sound pressure levels generated during operation by four representative transformers. It is worth noting that the A‑weighted sound level did not exceed 60 dBA in any of the researched transformers. Therefore, the transformers (with measured parameters) were not causing high levels of audible noise. During the research, it was noticed that the rms pressure value changes with the load of the power transformer (especially in low-frequency spectrum). Changes could be observed by relatively high rms pressure values for particular frequencies. The characteristics of the noise were presented as mean and rms values in Figure 5 and Figure 6, respectively.
Figure 5 and Figure 6 presents the frequency spectra of sound pressure levels generated by operating indoor- and overhead-type power transformers for a selected measurement point, i.e., 50 m. In the case of both type of transformers, the determined sound pressure levels of the generated noise were comparable; this applies to the whole range of the investigated frequencies. The largest values of the measured signals were recorded in the infrasound band from 1 Hz to 10 Hz. Several characteristic harmonic components could also be distinguished; the most significant was to be found for an indoor transformer in the range from 7 to 8 Hz, and for an overhead transformer in the range from 14 to 16 Hz. Also, the pick location depended on the transformer type, and was chosen to suit higher frequencies for the overhead transformer. For comparison purposes, the background noise was added in Figure 5 and Figure 6.
The tests were also carried out for distances of 75 m and 100 m. The distance did not influence the distribution amplitudes within the analyzed spectra (Figure 7). Additionally, there was a big problem finding objects around which there were no obstacles at a distance of more than 100 m (in the form of houses, trees, or busy roads). Therefore, a distance of 50 m was chosen as optimal for all the tested transformers.

4.1. Tuning Method for One Transformer

The tuning of the algorithm was performed using the method proposed in Section 3. In the first step, the nominally loaded transformer and its acoustic background were researched based on the resulting data (X). To simplify the tuning procedure and present robust results, the non-parametric, Naïve Bayes and exhaustive method was used. The results for resolution s = 4 (0.5 Hz), s = 8 (1 Hz), and s = 16 (2 Hz) for transformer 1 are presented in Figure 8, Figure 9 and Figure 10, respectively.
An analysis of Figure 8, Figure 9 and Figure 10 demonstrates that the resolution above 1 Hz (s = 8) decreases the stability of the results, which could be observed in Figure 10 as sinusoidal results above 80 Hz. A resolution range of up to 1 Hz gives similar results in terms of accuracy. Furthermore, four frequency ranges can be distinguished which offer a by which manner to classify transformers and their background with a high level of accuracy. The first frequency interval is 2–5 Hz, where the accuracy increases with range and obtains an optimal value for l = 2 Hz and h = 51 Hz. The second interval could be found at around 40–50 Hz. However, in this case, the accuracy did not increase so rapidly with the range (h-l). The third range could be found between 70 Hz and 85 Hz, but gave good results only for h-l = 1 Hz and h-l = 6 Hz. Finally, a high level of accuracy could be observed for frequencies close to 100 Hz for the smallest range (h-l = 1 Hz). A wide frequency range (over 30) makes it possible to obtain an accuracy of above 80% for any frequency. Transformers 2–4 share similar characteristics; however, some intervals were shifted slightly.
This observation was confirmed by heuristic (non-exhaustive) methods, which tended to find one of those frequency areas. Testing was performed using the Hill climbing method, random searches, and the Bayesian optimization method. The results of the best solution are shown in Table 2.

4.2. Tuning Model for Transformer Classification

The tuning of the algorithm for the classification of transformer type was performed in the same manner as for one transformer and its background. The results for resolution s = 4 (0.5 Hz), s = 8 (1 Hz), and s = 16 (2 Hz) are presented in Figure 11, Figure 12 and Figure 13, respectively.
A comparison of Figure 11, Figure 12 and Figure 13 shows that a resolution of up to s = 8 does not significantly decrease the accuracy. Three frequency ranges can be identified, which makes it possible to classify transformers with a high level of accuracy, i.e., the frequency for l = 2 to 5 Hz, where the accuracy increases with range (h-l parameter), yielding optimal values for l = 2 Hz and h = 51 Hz. A second area could be found at around 50 Hz, but for this frequency, the accuracy did not increase equally rapidly with range. Finally, the highest level of accuracy could be observed for frequencies close to 100 Hz. In this case, the small range (h-l = 1Hz) was sufficient to obtain a level of high accuracy. The automatically detected frequency intervals yielded comparable levels of accuracy (above 90%).
This observation was confirmed by heuristic (non-exhaustive) methods, which yielded the optimal values in those intervals. An example of the Hill Climbing method is presented in Figure 14. In the case of this method, the parameters l = 2, h = 44, and s = 5 were found. The black line illustrates the optimal solution.
Similar results were obtained using a random search: l = 4, h = 52, and s = 5. Finally, the Bayesan optimization method found the same interval as the exhausting method (l = 2, h = 32, and s = 5) parameters. The optimization procedure is presented in Figure 15.
Finally, using calculated parameter P, final verification was performed.

4.3. Verification of Tuned Model

Classification was performed for each transformer and its acoustic background. Various optimization methods and machine learning algorithms were tested for this purpose; the results are presented in Table 3 with and without (n/a) preprocessing procedures. In the case when a classifier had additional parameters, their values were tuned as part of the P vector.
The research firmly demonstrates that most up-to-date classifiers can obtain an excellent (nearly 100%) classification result and provide automatic identification of transformer noise from the background. The exhaustive search was based on a narrow range close to 1 Hz; thus, in the case of the SVM classifier, the hyperplane was not calculated correctly, and the results were close to 50%. In the case of wider ranges found by heuristic methods (HC and RS), the classifier provided an accuracy level of over 80%. The most stable results were obtained using the Random Trees method. This method also had the highest accuracy without using preprocessing method. The MLP-based solution offered similar results; however, the network tuning process was more complex due to the existence of two parameters to tune (m1-layer and m2-neuron number). The proposed preprocessing method made it possible to increase the classification accuracy by 15% on average.
Finally, the transformer type classification was performed for various optimization methods and using various machine learning algorithms. The results are presented in Table 4 using an additional measure, Cohens’ kappa [37].
Similarly, in the case of separate transformers, the SVM method proved to be inefficient (75%) for a small number of attributes. On the other hand, the MLP method, which proved to be a reliable solution for one classifier, provided a lower level of accuracy for the heuristic optimization method. Lower accuracy values were obtained due to fact that the accuracy varied significantly depending on the set of parameters used, and on random factors. Nevertheless, the optimization procedure makes it possible to increase the classification accuracy by 10% on average. The best classification results were obtained using k-NN for the exhaustive search. Nevertheless, most classifiers proved to be able to determine the transformers’ class. Misclassifications, in case of the best result, took place between transformers of the same structure but with different powers.

5. Conclusions

The proposed method, using a dedicated sensor, makes it possible to detect the state of transformers (on/off) based on the emitted low frequency with an accuracy of 99%. The best results were obtained using the Random Forest, KNN, and Naïve Bayes methods. The exhaustive search and Bayesian Optimization proved to be the best optimization methods for the transformer low frequency classification problem; however, the heuristic approaches gave better results in cases of separate transformers.
Further research may be applied to distinguish the type of transformer with a 97% level of accuracy using the KNN method; however, most of the classifiers made it possible to obtain accuracy levels of above 95%. The proposed preprocessing method makes it possible to significantly reduce the number of attributes and increase the detection accuracy by 10% on average. The method, with the proposed preprocessing module, could be adapted for edge computing in nodes due to its simplicity.
The present study has demonstrated that most low frequency noise (information) could be found near the following frequencies: 1Hz, 50Hz, and 100Hz. Therefore, future work will focus on the use multiple frequency ranges to train the classifiers. Furthermore, spectra analyses should be extended to over 100 Hz, due to the fact that vital information can be found near this frequency. This is a first step in future research; in the next stages, we also want to be able to distinguish the power of the transformer and its operating time. This set of parameters can be used for more precise diagnostics.
Finally, the method could be extended to detect noise form unknown transformers and to estimate their type and power level.

Author Contributions

Conceptualization, D.J.; Methodology, M.B.; Software, M.B., D.J., T.B.; Validation, T.B., D.J., M.B.; Formal Analysis, T.B.; Investigation, D.J., M.B., T.B.; Writing-Original Draft Preparation, M.B. and D.J.; Writing-Review & Editing, M.B., T.B. and D.J.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Jancarczyk, D.; Boczar, T.; Karpinski, M. Investigation of Low—Frequency Noise Emitted by Power Transformer. In Proceedings of the 19th International Multidisciplinary Scientific GeoConference SGEM 2019—Conference Proceedings Volume 19 Section Geoinformatics, Albena, Bulgaria, 28 June–7 July 2019. [Google Scholar]
  2. Malec, T.; Boczar, T.; Wotzka, D.; Fracz, P. Comparison of Low Frequency Signals Emitted by Wind Turbines of Two Different Generator Types. In Proceedings of the E3S Web of Conferences. International Conference Energy Environment and Material Systems, Polanica-zdro, Poland, 13–15 September 2017; Volume 19. [Google Scholar]
  3. Jancarczyk, D.; Bernaś, M.; Sidzina, M.; Janusz, J. Comparative Analysis of Infrasound Noise Emitted by Power Transformers; Scientific Papers Nr 59; Faculty of Electrical and Control Engineering, Gdansk University of Technology: Gdańsk, Poland, 2018. (In Polish) [Google Scholar]
  4. Jancarczyk, D. Research and Analysis of Infrasound Noise Emitted by Power Transformers. Electr. Eng. Pozn. Univ. Technol. Acad. J. 2018, 95, 153–161. (In Polish) [Google Scholar]
  5. Boczar, T.; Malec, T.; Wotzka, D. Studies on Infrasound Noise Emitted by Wind Turbine of Large Power. Acta Phys. Ppolonica 2012, 122, 850. [Google Scholar] [CrossRef]
  6. Pierzga, R.; Boczar, T.; Wotzka, D.; Zmarzly, D. Studies on Infrasound Noise Generated by Operation of Low-Power Wind Turbine. Acta Phys. Polonica A Opt. Acoust. Methods Sci. Technol. 2013, 124, 2013. [Google Scholar] [CrossRef]
  7. Zagubien, A.; Wolniewicz, K. Everyday Exposure to Occupational/Non-Occupational Infrasound Noise in Our Life. Arch. Acoust. 2016, 41, 659–668. [Google Scholar] [CrossRef] [Green Version]
  8. Barber Bayesian, D. Reasoning and Machine Learning; Cambridge University Press: Cambridge, UK, 2012. [Google Scholar]
  9. Mahdavinejad, M.S.; Rezvan, M.; Barekatain, M.; Adibi, P.; Barnaghi, P.; Sheth, A.P. Machine learning for internet of things data analysis: A survey. Digit. Commun. Netw. 2018, 4, 161–175. [Google Scholar] [CrossRef]
  10. Murphy, K.P. Machine Learning: A Probabilistic Perspective; MIT Press: Cambridge, MA, USA, 2012. [Google Scholar]
  11. Jagadish, H.V.; Ooi, B.C.; Tan, K.L.; Yu, C.; Zhang, R. Idistance: An adaptive b+-tree based indexing method for nearest neighbor search. ACM Trans. Database Syst. (TODS) 2005, 30, 364–397. [Google Scholar] [CrossRef]
  12. Zhang, H. The optimality of naive bayes. AA 2004, 1, 3. [Google Scholar]
  13. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  14. Scholkopf, B.; Smola, A.J. Learning with Kernels: Support Vector Machines, Regularization, Optimization, and Beyond; MIT Press: Cambridge, MA, USA, 2001. [Google Scholar]
  15. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  16. Breiman, L. Bagging predictors. Mach. Learn. 1996, 24, 123–140. [Google Scholar] [CrossRef] [Green Version]
  17. Glorot, X.; Bengio, Y. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the 13th International Conference Artificial Intelligence and Statistics (AISTATS), Sardinia, Italy, 13–15 May 2010; pp. 249–256. [Google Scholar]
  18. Eberhart, R.C. Neural Network PC Tools: A Practical Guide; Academic Press: Cambridge, MA, USA, 2014. [Google Scholar]
  19. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  20. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  21. Hotelling, H. Analysis of a complex of statistical variables into principal components. J. Educ. Psychol. 1933, 24, 417. [Google Scholar] [CrossRef]
  22. Jolliffe, I. Principal Component Analysis; Wiley Online Library: Hoboken, NJ, USA, 2002. [Google Scholar]
  23. Abdi, H.; Williams, L.J. Principal component analysis. Wiley Interdiscip. Rev. Comput. Stat. 2010, 2, 433–459. [Google Scholar] [CrossRef]
  24. Bro, R.; Smilde, A.K. Principal component analysis. Anal. Methods 2014, 6, 2812–2831. [Google Scholar] [CrossRef] [Green Version]
  25. Hotelling, H. Relations between two sets of variates. Biometrika 1936, 28, 321–377. [Google Scholar] [CrossRef]
  26. Bach, F.R.; Jordan, M.I. Kernel independent component analysis. J. Mach. Learn. Res. 2002, 3, 1–48. [Google Scholar]
  27. Blum, W.; Burghes, D.; Green, N.; Kaiser-Messmer, G. Teaching and learning of mathematics and its applications: First results from a comparative empirical study in england and Germany. Teach. Math. Appl. 1992, 11, 112–123. [Google Scholar] [CrossRef]
  28. Schmidhuber, J. Deep learning in neural networks: An overview. Neural Netw. 2015, 61, 85–117. [Google Scholar] [CrossRef] [Green Version]
  29. Bernas, M.; Placzek, B. Period-aware local modelling and data selection for time series prediction. Expert Syst. Appl. 2016, 59, 60–77. [Google Scholar] [CrossRef]
  30. Rabiner, L.R. A tutorial on hidden markov models and selected applications in speech recognition. Proc. IEEE 1989, 77, 257–286. [Google Scholar] [CrossRef]
  31. Sejnowski, T.J.; Rosenberg, C.R. Parallel networks that learn to pronounce english text. Complex Syst. 1987, 1, 145–168. [Google Scholar]
  32. Kalman, R.E.; Bucy, R.S. New results in linear filtering and prediction theory. J. Basic Eng. 1961, 83, 95–108. [Google Scholar] [CrossRef]
  33. Lafferty, J.; McCallum, A.; Pereira, F. Conditional random fields: Probabilistic models for segmenting and labeling sequence data. In Proceedings of the Eighteenth International Conference on Machine Learning, ICML, Williamstown, MA, USA, 28 June–1 July 2001; Volume 1, pp. 282–289. [Google Scholar]
  34. Williams, R.J.; Zipser, D. A learning algorithm for continually running fully recurrent neural networks. Neural Comput. 1989, 1, 270–280. [Google Scholar] [CrossRef]
  35. McCallum, A.; Freitag, D.; Pereira, F. Maximum Entropy Markov Models for Information Extraction and Segmentation. In Proceedings of the 17th International Conference on Machine Learning, ICML, Stanford, CA, USA, 29 June–2 July 2000; Volume 17, pp. 591–598. [Google Scholar]
  36. Baltrušaitis, T.; Ahuja, C.; Morency, L. Multimodal Machine Learning: A Survey and Taxonomy. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 41, 423–443. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. Welch, P. The use of fast Fourier transform for the estimation of power spectra: A method based on time averaging over short, modified periodograms. IEEE Trans. Audio Electr. 1967, 15, 70–73. [Google Scholar] [CrossRef] [Green Version]
  38. Bergstra, J.S.; Bardenet, R.; Bengio, Y.; Kégl, B. Algorithms for Hyper-Parameter Optimization. In Proceedings of the 25th Annual Conference on Neural Information Processing Systems, NIPS, Granada, Spain, 12 December 2011; pp. 2546–2554. [Google Scholar]
  39. Witten, I.H.; Frank, E.; Hall, M.A.; Pal, C.J. Data Mining: Practical Machine Learning Tools and Techniques; Morgan Kaufmann: Burlington, MA, USA, 2016. [Google Scholar]
Figure 1. Overview of proposed method.
Figure 1. Overview of proposed method.
Sensors 19 04909 g001
Figure 2. Microphone type 4190 digital and signal meter with LAN-XI type 3050-A-60.
Figure 2. Microphone type 4190 digital and signal meter with LAN-XI type 3050-A-60.
Sensors 19 04909 g002
Figure 3. PULSE LabShop application screenshot.
Figure 3. PULSE LabShop application screenshot.
Sensors 19 04909 g003
Figure 4. Photos of the tested power transformers: (a) first, (b) second, (c) third, and (d) fourth.
Figure 4. Photos of the tested power transformers: (a) first, (b) second, (c) third, and (d) fourth.
Sensors 19 04909 g004
Figure 5. The mean values of transformer noise in the 2–100 Hz frequency range.
Figure 5. The mean values of transformer noise in the 2–100 Hz frequency range.
Sensors 19 04909 g005
Figure 6. The rms pressure values of transformer noise in the 2–100 Hz frequency range.
Figure 6. The rms pressure values of transformer noise in the 2–100 Hz frequency range.
Sensors 19 04909 g006
Figure 7. Frequency spectra of sound pressure levels generated during the normal operation of an indoor type power transformer for the selected measurement points.
Figure 7. Frequency spectra of sound pressure levels generated during the normal operation of an indoor type power transformer for the selected measurement points.
Sensors 19 04909 g007
Figure 8. The accuracy value for Transformer 1, s = 4, and various l and h parameters.
Figure 8. The accuracy value for Transformer 1, s = 4, and various l and h parameters.
Sensors 19 04909 g008
Figure 9. The accuracy value for Transformer 1, s = 8, and various l and h parameters.
Figure 9. The accuracy value for Transformer 1, s = 8, and various l and h parameters.
Sensors 19 04909 g009
Figure 10. The accuracy value for Transformer 1, s = 16, and various l and h parameters.
Figure 10. The accuracy value for Transformer 1, s = 16, and various l and h parameters.
Sensors 19 04909 g010
Figure 11. Exhaustive tuning method for resolution 0.5 Hz.
Figure 11. Exhaustive tuning method for resolution 0.5 Hz.
Sensors 19 04909 g011
Figure 12. Exhaustive tuning method for resolution 1 Hz.
Figure 12. Exhaustive tuning method for resolution 1 Hz.
Sensors 19 04909 g012
Figure 13. Exhaustive tuning method for resolution 2 Hz.
Figure 13. Exhaustive tuning method for resolution 2 Hz.
Sensors 19 04909 g013
Figure 14. Hill climbing method optimization.
Figure 14. Hill climbing method optimization.
Sensors 19 04909 g014
Figure 15. Optimization using Bayesian approach.
Figure 15. Optimization using Bayesian approach.
Sensors 19 04909 g015
Table 1. The list of power transformers under study.
Table 1. The list of power transformers under study.
Transformer NameTransformer TypeApparent Power ProducerSPL
(1–20 kHz) [dB]
SPL
(1–100 Hz) [dB]
Transformer 1 indoor type 2000 kVASchneider Electric8981
Transformer 2indoor type 1250 kVASchneider Electric8971
Transformer 3overhead type400 kVAABB6125
Transformer 4overhead type250 kVAABB7949
SPL—sound pressure level.
Table 2. The best solutions for heuristic methods.
Table 2. The best solutions for heuristic methods.
Resolution [Hz]Heuristic MethodStart Frequency (l)Range (h-l)Accuracy
0.5Hill climbing (HC)382099%
Random search (RS)582899%
Bayesan optimization (TPE)23299%
1Hill climbing 671198%
Random search 891199%
Bayesan optimization99199%
Table 3. Comparison of various machine learning methods for each transformer and its acoustic background.
Table 3. Comparison of various machine learning methods for each transformer and its acoustic background.
Optimal. MethodML MethodML Params.Trans. 1 (Accuracy)Trans. 2 (Accuracy)Trans. 3 (Accuracy)Trans. 4 (Accuracy)
ESKNNm1 = 597%100%97%100%
naive Bayes-97%100%100%97%
MLPm1 = 1, m2 = 897%100%94%100%
RFm1 = 4100%100%94%100%
SVMm1 = 0.5, m2 = 367%50%61%50%
HCKNNm1 = 5100%100%100%100%
naive Bayes-100%100%100%100%
MLPm1 = 1, m2 = 892%100%92%100%
RFm1 = 4100%100%100%100%
SVMm1 = 0.5, m2 = 3100%100%83%100%
RSKNNm1 = 5100%100%100%100%
naive Bayes-100%100%100%100%
MLPm1 = 1, m2 = 8100%100%94%100%
RFm1 = 4100%100%100%100%
SVMm1 = 0.5, m2 = 3100%100%83%100%
TPEKNNm1 = 5100%100%100%97%
naive Bayes-100%100%100%100%
MLPm1 = 1, m2 = 897%97%100%92%
RFm1 = 4100%97%100%97%
SVMm1 = 0.5, m2 = 353%100%100%50%
n/aKNNm1 = 581%82%80%79%
naive Bayes-82%83%84%81%
MLPm1 = 1, m2 = 885%80%81%87%
RFm1 = 488%87%89%85%
SVMm1 = 0.5, m2 = 384%82%86%88%
EM: exhaustive method; HC: Hill climbing; RS: random search; TPE: Bayesian optimization strategy; KNN: k-Nearest Neighbors; naïve Bayes: Naive Bayes Classification; SVM: Support Vector Machine; RF: Random Forests; MLP: Multiple Layer Perceptron Network.
Table 4. Comparison of various machine learning methods.
Table 4. Comparison of various machine learning methods.
Optimal. MethodML MethodML Params.AccuracyCohens’ Kappa
ESKNNm1 = 797%96%
naive Bayes-96%94%
MLPm1 = 2, m2 = 1093%91%
RFm1 = 1096%94%
SVMm1 = 0.5, m2 = 375%67%
HCKNNm1 = 794%93%
naive Bayes-96%94%
MLPm1 = 2, m2 = 1086%81%
RFm1 = 1094%93%
SVMm1 = 0.5, m2 = 394%93%
RSKNNm1 = 794%93%
naive Bayes-96%94%
MLPm1 = 2, m2 = 1086%81%
RFm1 = 1094%93%
SVMm1 = 0.5, m2 = 394%93%
TPEKNNm1 = 794%93%
naive Bayes-94%93%
MLPm1 = 2, m2 = 1086%81%
RFm1 = 1096%94%
SVMm1 = 0.5, m2 = 396%94%
n/aKNNm1 = 783%78%
naive Bayes-86%81%
MLPm1 = 2, m2 = 1082%76%
RFm1 = 1085%80%
SVMm1 = 0.5, m2 = 388%83%
EM: exhaustive method; HC: Hill climbing; RS: random search; TPE: Bayesian optimization strategy; KNN: k-Nearest Neighbors; naïve Bayes: Naive Bayes Classification; SVM: Support Vector Machine; RF: Random Forests; MLP; Multiple Layer Perceptron Network.

Share and Cite

MDPI and ACS Style

Jancarczyk, D.; Bernaś, M.; Boczar, T. Classification of Low Frequency Signals Emitted by Power Transformers Using Sensors and Machine Learning Methods. Sensors 2019, 19, 4909. https://doi.org/10.3390/s19224909

AMA Style

Jancarczyk D, Bernaś M, Boczar T. Classification of Low Frequency Signals Emitted by Power Transformers Using Sensors and Machine Learning Methods. Sensors. 2019; 19(22):4909. https://doi.org/10.3390/s19224909

Chicago/Turabian Style

Jancarczyk, Daniel, Marcin Bernaś, and Tomasz Boczar. 2019. "Classification of Low Frequency Signals Emitted by Power Transformers Using Sensors and Machine Learning Methods" Sensors 19, no. 22: 4909. https://doi.org/10.3390/s19224909

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop