You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

1 September 2022

Unsupervised Outlier Detection in IOT Using Deep VAE

,
,
,
and
1
Department of Computer Engineering and Network, College of Computer and Information Sciences, Jouf University, Sakaka 72341, Al Jouf, Saudi Arabia
2
Electrical Engineering Department, Faculty of Engineering at Shoubra, Benha University, Cairo 13518, Egypt
3
University Institute of Information Technology, PMAS Arid Agricultural University, Rawalpindi 46000, Pakistan
4
Department of Computer Science, College of Computer and Information Sciences, Jouf University, Sakaka 72341, Al Jouf, Saudi Arabia
This article belongs to the Section Internet of Things

Abstract

The Internet of Things (IoT) refers to a system of interconnected, internet-connected devices and sensors that allows the collection and dissemination of data. The data provided by these sensors may include outliers or exhibit anomalous behavior as a result of attack activities or device failure, for example. However, the majority of existing outlier detection algorithms rely on labeled data, which is frequently hard to obtain in the IoT domain. More crucially, the IoT’s data volume is continually increasing, necessitating the requirement for predicting and identifying the classes of future data. In this study, we propose an unsupervised technique based on a deep Variational Auto-Encoder (VAE) to detect outliers in IoT data by leveraging the characteristic of the reconstruction ability and the low-dimensional representation of the input data’s latent variables of the VAE. First, the input data are standardized. Then, we employ the VAE to find a reconstructed output representation from the low-dimensional representation of the latent variables of the input data. Finally, the reconstruction error between the original observation and the reconstructed one is used as an outlier score. Our model was trained only using normal data with no labels in an unsupervised manner and evaluated using Statlog (Landsat Satellite) dataset. The unsupervised model achieved promising and comparable results with the state-of-the-art outlier detection schemes with a precision of ≈90% and an F1 score of 79%.

1. Introduction

IoTs are virtualized, integrated, and networked representations of objects in our environment. The integration of smart home devices, daily gadgets, health sector, and business are managed cloud infrastructure networks that are not bound by administrative, geographical, or cultural boundaries. Different sensors, instruments, actuators, and a variety of other devices and satellites enhance IoT-based applications. They are critical in gathering and comprehending the data supplied by smart city devices, agriculture, manufacturing, network security, and healthcare. This dataset collection is used to suggest and enhance services, security, and public utilities [1]. Wearable gadgets, distributed sensor-based smart devices, and smart devices, in general, may communicate information and services through a network, which is the basis of smart technology [2].
Outlier detection is another name for anomaly detection. Novelty discovery is the task of finding “odd” samples from a large dataset. Anomaly detection has grown in importance in enterprises where damaged goods or failures may be regarded as abnormal samples [3]. However, due to the rapid growth of big data technology and the Internet of Things (IoT), more research is being conducted in this area. Anomaly detection approaches have also received considerable interest in healthcare systems, smart city surveillance, and satellite-related information systems [4].
Despite their vast applications in various domains, there are several issues associated with IoTs and the data generated by IoT. First of all, different IoT-based solutions generate massive amounts of complex dimensional data, prompting the use of more advanced intelligence algorithms to evaluate and uncover hidden complex patterns in the data [5]. Second, conventional outlier detection techniques are frequently problem-specific and have supervised procedures [6]. The system based on these techniques needs domain knowledge and extensive data labels. As a result, in data-driven IoT systems, designing an unsupervised outlier detection approach is vital for smart and digital cities [7]. Among these issues outlier detection is a critical issue because of these challenges. Within that high-dimensional data, conventional anomaly detection techniques also fail to detect and classify anomalies [8].
Determined by the presence of labels in the dataset, outlier detection can be classified into three types, namely supervised, semi-supervised and unsupervised outlier detection techniques. Using properly labeled training data as well as test data, in training data, supervised outlier detection seeks to distinguish between normal and abnormal occurrences [9]. The model will then be implemented in both the training data and test data. A model using just normal examples for semi-supervised abnormality detection trains to create a standard “normal” model, allowing normal features to be learned equitably [10]. Outliers are defined as data that deviates significantly from the “normal” exemplary, as intended by scoring high. The unsupervised abnormality detection method is based on unlabeled data and does not differentiate between training and test data [11,12]. It makes a significant assumption, for example, that the detected data consist of a small number of abnormalities and a large number of typical cases. Models can be used to learn data’s intrinsic information. Outliers will be captured if there are any variances from the significant cases.
Unsupervised abnormality detection is considered the most challenging task among the three classes due to data variety and information scarcity. Furthermore, labeling is a costly and time-consuming needed task in real-world applications [13,14,15]. As a result, the lack of labels is common in many situations, making unsupervised abnormality detection techniques very useful and suitable.
A Variational Auto-Encoder (VAE) is a graphical model based on directed notes with a probabilistic approach, and its posterior is estimated by a neural network, resulting in an architecture similar to Auto-Encoder [16,17,18]. The encoder–decoder structure of the Variational Auto-Encoder allows for learning and training a mapping from highly complex and structured dimensional input and transforms to a latent representation having low dimensions. This is achieved while maintaining reconstruction accuracy at a higher level [19]. Furthermore, our strategy optimizes the performance of cooperative cooperation by combining feature extraction and model creation. The encoder–decoder structure uses a Fully Convolutional Network (FCN) without a completely connected layer. This layer retains relative spatial coordinates between the input dataset and the output features. This type of DNN is called a Fully Convolutional Variational Auto-Encoder (FC-VAE) [20]. The unsupervised abnormality detection challenges will be the focus of this research. Whether labels are provided or not, the anomaly detection problem may be viewed as a special unbalanced classification problem. It is well acknowledged that data quality has a significant impact on the outcomes of data-driven strategies.
This article proposes a deep learning system that is end-to-end and only training on unprocessed data using VAE. Our solution is based on the premise that normal data can be linked to a Variational Auto-Encoder. Then an outlier is defined as a test data instance that is not possible to link to any VAE component. Different varieties of features generally translate to better outcomes. However, most of the time, due to the presence of irrelevant and duplicated data, the quality of data cannot be certain. The worthless information not only reduces the accuracy of outlier detection but also lengthens processing complexity, and more time is consumed [3].
Our approach is based on the Variational Auto-Encoder, a probabilistic clustering model. The following are the research’s standout contributions.
  • We devised an unsupervised VAE that can detect abnormality effectively on satellite data.
  • We used the VAE from its latent representation to detect correlations between satellite data and identify anomalies.
  • We conducted an experimental evaluation to show that the proposed VAE outperforms cutting-edge approaches.
The remaining sections of the paper are organized as follows. Section 2 presents a summary of existing relevant research efforts. Section 3 describes the technique used to gather and analyze the satellite information, and Section 4 proposes a VAE-based system. Section 5 discusses the experiment and the outcomes, which are presented in detail. Section 6 concludes with a conclusion and possible research prospects.

3. Materials and Methods

In this section, we present the dataset description used in this research work. Once dataset is acquired, preprocessing is conducted on a tabular dataset of Satellite (Landsat Satellite). Then, the VAE is described in detail to show its working mechanism besides its importance in our work. After that, the proposed framework is presented that is based on VAE. In this study, we used a Satellite input database to train a VAE model.

3.1. Dataset Description

Data are central to deep learning models, and they serve as fuel for these learning models. We gathered satellite datasets in tabular form for this project from a publicly accessible data source. Statlog (Landsat Satellite) is a multi-class classification dataset from the UCI machine learning repository. The Statlog (Landsat Satellite) contains a total of 6435 samples, where the total normal samples = 4399 and the total outlier samples = 2036. These samples are stored in tabular form and have 36 features or dimensions.

3.2. Dataset Pre-Processing

The dataset of Satellite (Landsat Satellite) was available in tabular form. We standardize features by removing the mean and scaling to unit variance. This process is considered a pre-processing step in the proposed framework.

3.3. Variational Auto-Encoder Algorithm

A variational autoencoder (VAE) is a type of generative model in which the distribution of the encodings is regularized during training. In other words, VAE gives the latent space structure. The internal structure of the multidimensional latent space for a well-learned model defines its properties. The decoder component reconstructs the input using this information. The VAE architecture is made up of an encoder and a decoder with an intermediate compressed low-dimensional layer space, just as the AE architecture. The encoder maps the data to a posterior distribution, which is exclusive to VAE. The use of the univariate Gaussian distribution is a popular option in VAE. The Kullback–Leibler divergence measure between the approximation output and the target (input) characterizes the regularized term. The sum of the reconstruction loss and the Kullback–Leibler divergence is the overall loss in VAE, as shown in Algorithm 1.
A VAE has been shown to be successful in outlier detection as well as providing a wide range of complicated data. This assumes that normal samples’ latent representation is compatible with a Gaussian distribution. It means that all training datasets are grouped in feature space, and anomalies are located distant from the cluster [47]. Through a latent space, VAE can acquire the feature distribution for high-dimensional input data.
This study employs a fully connected neural network in both the encoder and decoder levels [48].
Algorithm 1 VAE base outlier detector
InputNormal dataset D
Dataset with abnormal d(i) = {1,2,3……N}
A thresh hold value α
OutputReconstruction probability class labels
StepsФѲ ← train using VAE with normal dataset D
   1.For loop from k = 1 to N
   2.Calculate
        Reconstruction   error   ( k ) = d k g θ ( f ( d k ) )
   3.   if Reconstruction error (k) > α then
                   outlier ← d(k)
          Else
                   Not outlier ← d(k)
          end if
          End for
Where gθ(f∅ (d(k)) is a recognition network also known as probabilistic encoder. f∅(d(k)) is a generative decoder as a generative network.
The VAE system provided here extracts useful characteristics for numerous standard unsupervised outlier detection applications. The reconstruction probability is computed using the stochastic latent variables that create the parameters from the distribution of the original input variable. The constraints of the input variable distribution, not the input variable itself, are reconstructed. This is the likelihood of data being generated from a particular latent variable pulled from the approximate posterior distribution.

3.4. Proposed Framework

This section explains an unsupervised-technique-based framework that relies on a deep Variational Auto-Encoder (VAE) to detect outliers in Satellite data by leveraging the characteristic of the reconstruction ability as well as VAE’s low-dimensional representation of the input data’s latent variables. Figure 2 demonstrates the DL approach that uses Landset Statlog (Landsat Satellite) to learn the candidate and favorable features to differentiate between normal and outlier samples. As in Algorithm 1, Standard Scaler is executed as a preprocessing phase. Then, in the next phase, A VAE is used to obtain the structure of the latent space. The VAE architecture, such as AE, is made up of two primary parts: a decoder and an encoder with an intermediate compressed low-dimensional layer space. The internal structure of a well-learned model defines the multidimensional latent space. The decoder section uses this information to rebuild the input. The sum of the reconstruction loss between the input and the reconstructed output and the Kullback–Leibler divergence is then used as the total loss in VAE, which in turn is used as an outlier score for our data.
Figure 2. Proposed Variational Auto-Encoder framework for outlier detection.
VAE is used to achieve these goals since it is capable of coping with data uncertainties and collinearity in the data set [49]. As previously explained, VAE is an unsupervised neural network whose purpose in its output layer is to generate/reconstruct the provided input samples [50]. Effective features for representing the high-dimensional, redundant, and noise-corrupted input data are generated by training a VAE model on input data. We may categorize the samples as normal or outlier based on the mean squared error between the original observation and the reconstructed one. During the optimization phase, a classifier’s threshold value is calculated.

4. Experimental Setup

Several tests were carried out using the Statlog (Landsat Satellite) dataset to illustrate the efficacy of the proposed VAE-based anomaly detection approach. These tests also yielded unique results, which were utilized to compare the suggested model to state-of-the-art techniques. The code for the proposed system was developed using TensorFlow -Keras libraries in Python Jupyter Notebook on Ubuntu 20.04.4 desktop pc with Core i7-11700F CPU with 16 GB RAM and NIVIDIA GeForce RTX 3060 GPU with 12 GB Memory. During the experiments, the training-to-testing ratio was 80:20%. According to the suggested training scheme, training was performed utilizing 80 percent random samples of Statlog (Landsat Satellite) Data Set for the proposed DL-based VAE system.
Only normal samples were used to train the model, so outlier samples from the training set were excluded and added to the test set. Hence, the model was tested using all 2036 outlier samples besides the test normal samples. The outlier score was generated by calculating the reconstruction error between the input and the reconstructed sample. The suggested framework is pre-trained on Statlog (Landsat Satellite) Data Set using the Adam with a learning rate technique that reduces the learning rate when learning gets stagnant for a period of time (i.e., validation patience). The Adam optimizer was trained with the following hyper-parameters: Total number of epochs = 200; batch size = 16, learning rate = 1 × 10−3, and loss function = “elbo”.

Assessment Methods

To assess performance, we used the performance indicators stated in Equations (1)–(6) to compare our proposed system to other systems. The proposed evaluation parameters were taken from existing research work [51,52].
T R P = T P ( T P + F N )
T R N = T N ( T N + F N )
P P V = T P ( T P + F P )
F N R = F N ( T P + F N )
F P R = F P ( F P + T N )
                      F 1 S c o r e = 2 ( P r e c i s i o n r e c a l l ) ( P r e c i s i o n + r e c a l l )
True positives (samples correctly recognized as having outlier), true negatives (samples correctly identified as not having outlier), false positives (samples with outliers other than anomaly), and false negatives (samples with anomaly identified as normal samples).

5. Discussion

The results of the suggested systems’ different trials utilizing the Statlog (Landsat Satellite) dataset are shown in this section. The outcomes defined the dataset into two categories: outliers and normal instances. Figure 3 depicts the frequency of normal instances of satellite images, as well as the frequency of outliers. From the given dataset, there are 4399 normal samples and 2036 outlier samples.
Figure 3. Visualization of the distribution of normal/outlier instances in the used dataset.
The effectiveness of the proposed technique is assessed using ROC and AUC assessments. The AUC measure is used for assessment since it is the benchmark in unsupervised outlier identification, as well as its interpretability. The ROC-based evaluation was used since it was insensitive to class distribution. It also indicates that one technique may outperform another at different thresholds. Furthermore, the Mean Precision–Recall (MPR) curve is used as a statistic to assess and compare the efficacy of outlier identification systems.
We also demonstrate reconstruction probabilities, when employed as the outlier score, may discriminate between the abnormal and the normal. Figure 4 displays the distribution of the obtained reconstruction error for both normal and anomalous samples. To properly reflect the distribution, we conducted operations on each sample by obtaining the Negative Logarithm of the Reconstruction Probability (NLRP). The likelihood of each sample being anomalous grows proportionately to the value of NLRP. According to Figure 4, the majority of normal samples have lower NLRP values than anomalous samples for the used dataset. The distributions of reconstruction probabilities from normal and anomalous samples are then easily differentiated with very minimal overlaps.
Figure 4. Different classes depicting reconstruction error.
The entire tradeoff between the true (genuine) positive rate and the false positive rate is generated using Receiver Operating Characteristic (ROC) curves. The area under the ROC curve can be used to summarize the ROC curve. This is referred to as the AUC. The AUC ROC curve is a graph that shows the classification performance of a model in terms of two variables: true positive (tp) and false positive rate (fpr). The area under the receiver operating characteristic (AUROC) and Mean Precision–Recall (MPR) are used as metrics throughout the assessment to evaluate and examine the effectiveness of outlier identification systems. An effective model must have a high AUROC, but AUPRC frequently exposes the gap between techniques when dealing with imbalanced datasets. The suggested FC-VAE framework’s efficiency was assessed using AUC. The AUC may be determined by incorporating the areas of small geometrical components under the ROC curve. Figure 5 displays the AUC assessments of the suggested architecture. Existing classification models are outperformed by the proposed model (AUC = 0.842).
Figure 5. Receiver operating characteristic curve (ROC) for Statlog (Landsat Satellite) dataset.
Figure 6 shows the confusion matrix of the outlier and normal sample from the Statlog (Landsat Satellite) dataset. According to the graph, 743 samples were correctly classified as normal samples, 164 normal were misclassified as abnormal samples, and 597 outlier samples were incorrectly classified as normal samples. The Satellite dataset contains 1439 outliers that are correctly classified as anomalies.
Figure 6. Confusion matrix with Statlog (Landsat Satellite) dataset.
This section shows the performance of the proposed Variational Auto-Encoder model using the MPR evaluation method and summarizes the results using the accuracy and recall curves. This curve focuses mostly on the performance in the abnormality class and is thus significantly more indicative when the abnormalities are only of interest to us. The Mean Precision–Recall (MPR) curve is used as a measure during the assessment to evaluate and compare the performance of abnormality detection systems. MPR summarizes the precision and recall curve, which focuses on performance in the abnormality class and is therefore much more revealing when the abnormalities are just of interest. MPR is calculated with the help of a confusion matrix using Equations (1)–(6). Precision or positive predictive value (PPV) = 0.897 is recorded, whereas recall is calculated as Recall (TPR) = 0.70; hence our model has achieved a higher PPV of ≈90%, as shown in Figure 7.
Figure 7. Precision and recall trend of Statlog (Landsat Satellite) dataset.
While inspecting our system results, the obtained F1 score is 0.79. This s crucial for a practical outlier detection system. Moreover, the proposed system shows that the FPR = 0.18 and specificity (TNR) appears to be 0.81, and the false negative ratio is FNR = 0.29 with AUC_ROC = 0.842. These results are then compared with state-of-the-art approaches used in the same domain. As shown in Table 2, FC-VAE gives remarkable results in terms of precision and accuracy like existing approaches. When compared with other state-of-the-art approaches that utilize satellite dataset (images or tabular form), different results are concluded in terms of precision, recall, F1 score, and accuracy. Most of the recall value lies in the range of 85–100%. The precision varies in the range of 78–92%. The accuracy of most of the techniques achieved during the evaluation reached 98%. F1 score, which is the harmonic mean of precision and recall, varied from 83% to 94%.
Table 2. Comparison of proposed FC-VAE with state-of-the-art techniques.
It is vital to note that comparing outcomes in this area of unsupervised learning is highly challenging owing to the wide range of accessible datasets and techniques. Furthermore, the clustering labels may seem different, and the performance indicators may vary. As certain dataset labels are never completely recognized, it may result in presenting findings in ambiguous test datasets. It is evident from this that the performance of the predicted FC-VAE may change if the prediction is based on a different set of training or test datasets.
The proposed FC-VAE model was tested on the KPI time series dataset used for VAE-GEN [54]. We made a public request to the researchers to reveal the dataset’s specifics. The authors were gracious enough to publish their model’s source code as well as the dataset on Jupiter Notebook. We executed FC-VAE model on their dataset and attained the acclaimed accuracy of 76% [54]. Table 3 compares the accuracy acquired as a consequence of this comparison experiment with the Statlog dataset used for FC-VAE. It is incredible to note that the satellite dataset provided considerably better findings in terms of precision and F1 score, indicating the generality of the proposed approach.
Table 3. Evaluation of proposed model with different datasets.

6. Conclusions

IoT-based applications are becoming prevalent in a variety of fields. IoT data may contain outliers or display aberrant behavior as a consequence of attack activity or device failure. The bulk of current outlier identification methods relies on labeled data, which is typically difficult to obtain in the IoT sector. In this paper, an unsupervised approach based on deep Variational Auto-Encoder (VAE) is presented. The purpose was to find outliers in IoT data by exploiting the VAE’s reconstruction capabilities and low-dimensional representation of the input data’s latent variables. The proposed model was trained in an unsupervised fashion using just normal data with no labels and assessed using the Statlog (Landsat Satellite) dataset. The unsupervised model produced promising and comparable results to state-of-the-art outlier identification approaches, with an accuracy of 90% and an F1 score of 79%. With such promising findings, we intend to expand our technique to utilize a better strategy employing AE architecture in order to manipulate the abnormality detector with datasets other than the ones it has been trained on, hence enhancing the performance and accuracy of unlabeled data.

Author Contributions

Conceptualization, W.G. and S.A.; methodology, W.G.; software, W.G.; validation, S.A., M.A. and G.A.; formal analysis, W.G.; investigation, S.T.; resources, G.A.; data curation, W.G.; writing—original draft preparation, S.T.; writing—review and editing, S.T.; visualization, W.G.; supervision, M.A.; project administration, M.A. and W.G.; funding acquisition, G.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Islam, M.; Rahaman, A. Development of smart healthcare monitoring system in IoT environment. SN Comput. Sci. 2020, 1, 185. [Google Scholar] [CrossRef] [PubMed]
  2. Moin, S.; Karim, A.; Safdar, Z.; Safdar, K.; Ahmed, E.; Imran, M. Securing IoTs in distributed blockchain: Analysis, requirements and open issues. Futur. Gener. Comput. Syst. 2019, 100, 325–343. [Google Scholar] [CrossRef]
  3. Smys, S.; Basar, A.; Wang, H. Hybrid intrusion detection system for internet of things (IoT). J. ISMAC 2020, 2, 190–199. [Google Scholar] [CrossRef]
  4. Cui, L.; Yang, S.; Chen, F.; Ming, Z.; Lu, N.; Qin, J. A survey on application of machine learning for Internet of Things. Int. J. Mach. Learn. Cybern. 2018, 9, 1399–1417. [Google Scholar] [CrossRef]
  5. Sobin, C.C. A survey on architecture, protocols and challenges in IoT. Wirel. Pers. Commun. 2020, 112, 1383–1429. [Google Scholar] [CrossRef]
  6. Yadav, E.P.; Mittal, E.A.; Yadav, H. IoT: Challenges and issues in indian perspective. In Proceedings of the 2018 3rd International Conference On Internet of Things: Smart Innovation and Usages (IoT-SIU), Bhimtal, India, 23–24 February 2018; pp. 1–5. [Google Scholar]
  7. Hamid, B.; Jhanjhi, N.; Humayun, M.; Khan, A.; Alsayat, A. Cyber Security Issues and Challenges for Smart Cities: A survey. In Proceedings of the 2019 13th International Conference on Mathematics, Actuarial Science, Computer Science and Statistics (MACS), Karachi, Pakistan, 14–15 December 2019; pp. 1–7. [Google Scholar]
  8. Sadeghi, A.-R.; Wachsmann, C.; Waidner, M. Security and privacy challenges in industrial internet of things. In Proceedings of the 2015 52nd ACM/EDAC/IEEE Design Automation Conference (DAC), San Francisco, CA, USA, 8–12 June 2015; pp. 1–6. [Google Scholar]
  9. Pang, G.; van den Hengel, A.; Shen, C.; Cao, L. Toward Deep Supervised Anomaly Detection: Reinforcement Learning from Partially Labeled Anomaly Data; Association for Computing Machinery: New York, NY, USA, 2021; Volume 1. [Google Scholar]
  10. Zavrak, S.; Iskefiyeli, M. Anomaly-Based Intrusion Detection from Network Flow Features Using Variational Autoencoder. IEEE Access 2020, 8, 108346–108358. [Google Scholar] [CrossRef]
  11. Zoppi, T.; Ceccarelli, A.; Capecchi, T.; Bondavalli, A. Unsupervised anomaly detectors to detect intrusions in the current threat landscape. ACM/IMS Trans. Data Sci. 2021, 2, 1–26. [Google Scholar] [CrossRef]
  12. Fan, C.; Xiao, F.; Zhao, Y.; Wang, J. Analytical investigation of autoencoder-based methods for unsupervised anomaly detection in building energy data. Appl. Energy 2018, 211, 1123–1135. [Google Scholar] [CrossRef]
  13. Yamanishi, K.; Takeuchi, J.-I.; Williams, G.; Milne, P. On-line unsupervised outlier detection using finite mixtures with discounting learning algorithms. Data Min. Knowl. Discov. 2004, 8, 275–300. [Google Scholar] [CrossRef]
  14. Yao, R.; Liu, C.; Zhang, L.; Peng, P. Unsupervised anomaly detection using variational auto-encoder based feature extraction. In Proceedings of the 2019 IEEE International Conference on Prognostics and Health Management (ICPHM), San Francisco, CA, USA, 17–20 June 2019. [Google Scholar]
  15. Nawaz, A.; Awan, A.A.; Ali, T.; Rana, M.R.R. Product’s behaviour recommendations using free text: An aspect based sentiment analysis approach. Cluster Comput. 2020, 23, 1267–1279. [Google Scholar] [CrossRef]
  16. Shen, B.; Ge, Z. Supervised nonlinear dynamic system for soft sensor application aided by variational auto-encoder. IEEE Trans. Instrum. Meas. 2020, 69, 6132–6142. [Google Scholar] [CrossRef]
  17. Fan, Y.; Wen, G.; Li, D.; Qiu, S.; Levine, M.D.; Xiao, F. Video anomaly detection and localization via Gaussian Mixture Fully Convolutional Variational Autoencoder. Comput. Vis. Image Underst. 2020, 195, 102920. [Google Scholar] [CrossRef]
  18. Tahir, S.; Hafeez, Y.; Abbas, M.A.; Nawaz, A.; Hamid, B. Smart Learning Objects Retrieval for E-Learning with Contextual Recommendation based on Collaborative Filtering. Educ. Inf. Technol. 2022, 27, 8631–8668. [Google Scholar] [CrossRef]
  19. Guo, Y.; Ji, T.; Wang, Q.; Yu, L.; Min, G.; Li, P. Unsupervised Anomaly Detection in IoT Systems for Smart Cities. IEEE Trans. Netw. Sci. Eng. 2020, 7, 2231–2242. [Google Scholar] [CrossRef]
  20. Nugroho, H.; Susanty, M.; Irawan, A.; Koyimatu, M.; Yunita, A. Fully convolutional variational autoencoder for feature extraction of fire detection system. J. Ilmu Komput. Dan Inf. 2020, 13, 9–15. [Google Scholar] [CrossRef]
  21. Liang, H.; Song, L.; Wang, J.; Guo, L.; Li, X.; Liang, J. Robust unsupervised anomaly detection via multi-time scale DCGANs with forgetting mechanism for industrial multivariate time series. Neurocomputing 2021, 423, 444–462. [Google Scholar] [CrossRef]
  22. Ni, F.; Zhang, J.; Noori, M.N. Deep learning for data anomaly detection and data compression of a long-span suspension bridge. Comput. Civ. Infrastruct. Eng. 2020, 35, 685–700. [Google Scholar] [CrossRef]
  23. Otoum, Y.; Liu, D.; Nayak, A. DL-IDS: A deep learning--based intrusion detection framework for securing IoT. Trans. Emerg. Telecommun. Technol. 2022, 33, e3803. [Google Scholar] [CrossRef]
  24. Song, H.M.; Woo, J.; Kim, H.K. In-vehicle network intrusion detection using deep convolutional neural network. Veh. Commun. 2020, 21, 100198. [Google Scholar] [CrossRef]
  25. Zahid, M.; Ahmed, F.; Javaid, N.; Abbasi, R.A.; Zainab Kazmi, H.S.; Javaid, A.; Bilal, M.; Akbar, M.; Ilahi, M. Electricity price and load forecasting using enhanced convolutional neural network and enhanced support vector regression in smart grids. Electronics 2019, 8, 122. [Google Scholar] [CrossRef] [Green Version]
  26. Caroppo, A.; Leone, A.; Siciliano, P. Comparison between deep learning models and traditional machine learning approaches for facial expression recognition in ageing adults. J. Comput. Sci. Technol. 2020, 35, 1127–1146. [Google Scholar] [CrossRef]
  27. Chalapathy, R.; Chawla, S. Deep Learning for Anomaly Detection: A Survey. arXiv 2019, arXiv:1901.03407. [Google Scholar]
  28. Hasan, M.; Choi, J.; Neumann, J.; Roy-Chowdhury, A.K.; Davis, L.S. Learning temporal regularity in video sequences. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 733–742. [Google Scholar]
  29. Li, Y.; Ma, R.; Jiao, R. A hybrid malicious code detection method based on deep learning. Int. J. Secur. Its Appl. 2015, 9, 205–216. [Google Scholar] [CrossRef]
  30. Ullah, Z.; Al-Turjman, F.; Mostarda, L.; Gagliardi, R. Applications of Artificial Intelligence and Machine learning in smart cities. Comput. Commun. 2020, 154, 313–323. [Google Scholar] [CrossRef]
  31. Chen, D.; Wawrzynski, P.; Lv, Z. Cyber security in smart cities: A review of deep learning-based applications and case studies. Sustain. Cities Soc. 2021, 66, 102655. [Google Scholar] [CrossRef]
  32. Garg, S.; Kaur, K.; Kumar, N.; Rodrigues, J.J.P.C. Hybrid deep-learning-based anomaly detection scheme for suspicious flow detection in SDN: A social multimedia perspective. IEEE Trans. Multimed. 2019, 21, 566–578. [Google Scholar] [CrossRef]
  33. Deng, X.; Huangfu, F. Collaborative variational deep learning for healthcare recommendation. IEEE Access 2019, 7, 55679–55688. [Google Scholar] [CrossRef]
  34. Gouda, W.; Almurafeh, M.; Humayun, M.; Jhanjhi, N.Z. Detection of COVID-19 Based on Chest X-rays Using Deep Learning. Healthcare 2022, 10, 343. [Google Scholar] [CrossRef]
  35. Mansour, R.F.; Escorcia-Gutierrez, J.; Gamarra, M.; Gupta, D.; Castillo, O.; Kumar, S. Unsupervised deep learning based variational autoencoder model for COVID-19 diagnosis and classification. Pattern Recognit. Lett. 2021, 151, 267–274. [Google Scholar] [CrossRef]
  36. Guo, F.; Xie, R.; Huang, B. A deep learning just-in-time modeling approach for soft sensor based on variational autoencoder. Chemom. Intell. Lab. Syst. 2020, 197, 103922. [Google Scholar] [CrossRef]
  37. Ionescu, R.T.; Khan, F.S.; Georgescu, M.I.; Shao, L. Object-centric auto-encoders and dummy anomalies for abnormal event detection in video. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 7834–7843. [Google Scholar]
  38. Chen, J.; Sathe, S.; Aggarwal, C.; Turaga, D. Outlier detection with autoencoder ensembles. In Proceedings of the 2017 SIAM International Conference on Data Mining, Houston, TX, USA, 27–29 April 2017; pp. 90–98. [Google Scholar]
  39. Lu, W.; Cheng, Y.; Xiao, C.; Chang, S.; Huang, S.; Liang, B.; Huang, T. Unsupervised sequential outlier detection with deep architectures. IEEE Trans. Image Process. 2017, 26, 4321–4330. [Google Scholar] [CrossRef] [PubMed]
  40. Golan, I.; El-Yaniv, R. Deep anomaly detection using geometric transformations. Adv. Neural Inf. Process. Syst. 2018, 31, 9758–9769. [Google Scholar]
  41. Wang, S.; Zeng, Y.; Liu, X.; Zhu, E.; Yin, J.; Xu, C.; Kloft, M. Effective end-to-end unsupervised outlier detection via inlier priority of discriminative network. Adv. Neural Inf. Process. Syst. 2019, 32, 1–14. [Google Scholar]
  42. Nguyen, M.-N.; Vien, N.A. Scalable and interpretable one-class svms with deep learning and random fourier features. In Joint European Conference on Machine Learning and Knowledge Discovery in Databases; Springer: Cham, Switzerland, 2018; pp. 157–172. [Google Scholar]
  43. Zong, B.; Song, Q.; Min, M.R.; Cheng, W.; Lumezanu, C.; Cho, D.; Chen, H. Deep autoencoding gaussian mixture model for unsupervised anomaly detection. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  44. Fan, S.; Shi, C.; Wang, X. Abnormal event detection via heterogeneous information network embedding. In Proceedings of the 27th ACM International Conference on Information and Knowledge Management, Torino Italy, 22–26 October 2018; pp. 1483–1486. [Google Scholar]
  45. Sabokrou, M.; Khalooei, M.; Fathy, M.; Adeli, E. Adversarially learned one-class classifier for novelty detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 3379–3388. [Google Scholar]
  46. Zheng, P.; Yuan, S.; Wu, X.; Li, J.; Lu, A. One-class adversarial nets for fraud detection. Proc. AAAI Conf. Artif. Intell. 2019, 33, 1286–1293. [Google Scholar] [CrossRef]
  47. Zhou, J.; Wang, X.; Yang, C.; Xiong, W. A novel soft sensor modeling approach based on difference-LSTM for complex industrial process. IEEE Trans. Ind. Inform. 2021, 18, 2955–2964. [Google Scholar] [CrossRef]
  48. Vincent, P.; Larochelle, H.; Lajoie, I.; Bengio, Y.; Manzagol, P.-A.; Bottou, L. Stacked denoising autoencoders: Learning useful representations in a deep network with a local denoising criterion. J. Mach. Learn. Res. 2010, 11, 3371–3408. [Google Scholar]
  49. Chen, J.; Pi, D.; Wu, Z.; Zhao, X.; Pan, Y.; Zhang, Q. Imbalanced satellite telemetry data anomaly detection model based on Bayesian LSTM. Acta Astronaut. 2021, 180, 232–242. [Google Scholar] [CrossRef]
  50. Xie, L.; Pi, D.; Zhang, X.; Chen, J.; Luo, Y.; Yu, W. Graph neural network approach for anomaly detection. Meas. J. Int. Meas. Confed. 2021, 180, 109546. [Google Scholar] [CrossRef]
  51. Shriram, S.; Sivasankar, E. Anomaly Detection on Shuttle data using Unsupervised Learning Techniques. In Proceedings of the 2019 International Conference on Computational Intelligence and Knowledge Economy (ICCIKE), Dubai, United Arab Emirates, 11–12 December 2019; pp. 221–225. [Google Scholar]
  52. Jiang, H.; Zhang, K.; Wang, J.; Wang, X.; Huang, P. Anomaly detection and identification in satellite telemetry data based on pseudo-period. Appl. Sci. 2020, 10, 103. [Google Scholar] [CrossRef]
  53. Bawono, A.H.; Bachtiar, F.A. Outlier Detection with Supervised Learning Method. In Proceedings of the 2019 International Conference on Sustainable Information Engineering and Technology (SIET), Lombok, Indonesia, 28–30 September 2019; pp. 306–309. [Google Scholar]
  54. Niu, Z.; Yu, K.; Wu, X. LSTM-Based VAE-GAN for Time-Series Anomaly Detection. Sensors 2020, 20, 3738. [Google Scholar] [CrossRef]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.