Next Article in Journal
Enhanced Learning Behaviors and Ability Knowledge Tracing
Next Article in Special Issue
Fault Detection Method Using Auto-Associative Shared Nearest Neighbor Kernel Regression for Industrial Processes
Previous Article in Journal
Detection of Manipulations in Digital Images: A Review of Passive and Active Methods Utilizing Deep Learning
Previous Article in Special Issue
An Integrated Stacking Ensemble Model for Natural Gas Purchase Prediction Incorporating Multiple Features
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dynamic Classifier Auditing by Unsupervised Anomaly Detection Methods: An Application in Packaging Industry Predictive Maintenance

by
Fernando Mateo
*,
Joan Vila-Francés
,
Emilio Soria-Olivas
,
Marcelino Martínez-Sober
,
Juan Gómez-Sanchis
and
Antonio José Serrano-López
Departamento de Ingeniería Electrónica, Escola Tècnica Superior d’Enginyería (ETSE), Avenida de la Universidad, 46100 Burjassot, Spain
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(2), 882; https://doi.org/10.3390/app15020882
Submission received: 16 December 2024 / Revised: 8 January 2025 / Accepted: 15 January 2025 / Published: 17 January 2025
(This article belongs to the Special Issue Artificial Intelligence Applications in Industry)

Abstract

:
Predictive maintenance in manufacturing industry applications is a challenging research field. Packaging machines are widely used in a large number of logistic companies’ warehouses and must be working uninterruptedly. Traditionally, preventive maintenance strategies have been carried out to improve the performance of these machines. However, these kinds of policies do not take into account the information provided by the sensors implemented in the machines. This paper presents an expert system for the automatic estimation of work orders to implement predictive maintenance policies for packaging machines. The central innovation lies in a two-stage process: a classifier generates a binary decision on whether a machine requires maintenance, and an unsupervised anomaly detection module subsequently audits the classifier’s probabilistic output to refine and interpret its predictions. By leveraging the classifier to condense sensor data and applying anomaly detection to its output, the system optimizes the decision reliability. Three anomaly detection methods were evaluated: One-Class Support Vector Machine (OCSVM), Minimum Covariance Determinant (MCD), and a majority (hard) voting ensemble of the two. All anomaly detection methods improved the baseline classifier’s performance, with the majority voting ensemble achieving the highest F1 score.

1. Introduction

Production lines in large companies rely on equipment to work properly. A failure in a device or component may cause a stop in the entire production line. Production stops are associated with huge costs, causing not only production loss due to downtime but also efforts to identify the cause of the failure and repair it [1]. In the new industrial era, predictive maintenance has been adopted by many companies to estimate when maintenance should be performed to avoid unnecessary losses [2].
In recent years, industry has developed towards the fourth stage of industrialization [3] (the so-called Industry 4.0). The connectivity of various devices, especially industrial ones, has experienced substantial growth, partly due to the advances and cost reduction of Ethernet-based buses. The current trend is that any industrial device, from a Programmable Logic Controller (PLC) to a complete manufacturing cell, is connected to a network. Usually, industrial devices send data related to the manufacturing process (performance, machine status, etc.). Thus, data analysis related to the improvement of industrial processes is a reality. A proof of this is the wide and prolific state of the art in this field [4,5]. The incorporation of advanced data-driven techniques and models enables analysis that goes beyond merely visualizing process parameters. These approaches provide valuable insights for optimizing maintenance and managing industrial manufacturing cells more effectively.
This paper addresses the problem of establishing predictive maintenance strategies in a group of packaging machines that operate in logistic centers. Packaging machines are devices used for applying sheets of plastic film on transport pallets. These machines are designed to wrap the load without having to turn around a platform in order to secure the load, protect it from dust or water, reduce its volume, strengthen the packaging, etc. An example of a fixed automatic packaging machine is shown in Figure 1. All these packaging machines are connected to an Ethernet network and periodically report a set of alarms associated with different types of anomalous behaviors. The status variables of each machine are received by a central server and are stored in a database. Despite the availability of the status variables of each machine in real time, several challenges arise when planning the development of an expert system to decide when the company must send a technician to take maintenance actions. These challenges are related to some main issues: (1) as in almost every predictive maintenance problem, there is a strong imbalance between the classes of the data set. This imbalance should be addressed before building a classifier to avoid the undesirable situation where the classifier consistently predicts the majority class, thereby neglecting all failure predictions [6]. The following steps are crucial: (1) extracting relevant alarms and constructing the dataset that will serve as input for the machine learning algorithms; (2) defining a strategy to assess the importance of past alarms in the classification system; (3) determining the frequency with which the models are updated; and (4) integrating the machine learning algorithms into the production lines of the manufacturing company.
The main contribution of this paper will be to introduce the framework and methodology to audit an expert system (a Random Forest-based classifier) for planning predictive maintenance actions (work orders) in the ERP (Enterprise Resource Planning) system of a packaging machine company by appending an unsupervised anomaly detection stage at its output that acts as a supervisor. The anomaly detection stage only uses the information given by the classifier during the first 30 days of run-time and is then applied sequentially to time windows of a determined duration to predict if the following day may be considered as an anomaly and would involve a maintenance action. We concluded that this auditing stage is useful to improve the precision and recall of the baseline classifier, which are measured globally as the F1 score.
The rest of the paper is organized as follows. Section 2 reviews the existing literature in the field. Section 3 formulates the objective of the work and provides a graphical description of the system. Materials and methods used are explained in Section 4, including the strategy to build the dataset and the anomaly detection methods and libraries used to fine-tune the expert system. Section 5 shows the obtained classification results and Section 6 discusses and analyzes those results.

2. Literature Review

Predictive maintenance has been gaining prominence in multidisciplinary research groups, proposing the creation and integration of lines of research related to data acquisition, infrastructure, storage, distribution, security, and intelligence [7]. The ability to predict the need for maintenance of assets at a specific future moment is one of the main challenges in Industry 4.0. New smart industrial facilities are focused on creating manufacturing intelligence from real-time data to support accurate and timely decision-making that can have a positive impact across the entire organization [8,9]. In particular, the possibility of performing predictive maintenance contributes to enhancing machine downtime, costs, control, and quality of production [1,2,7].
Emerging technologies such as Internet of Things (IoT) and Cyber Physical Systems (CPSs) are being embedded in physical processes to measure and monitor real-time data from across a factory, which will ultimately give rise to unprecedented levels of data production. This obliges resorting to Big Data technologies to manage these massive amounts of data [8,10,11].
Historically, preventive maintenance techniques have typically been used to minimize failures in manufacturing cells using maintenance strategies such as breakdown maintenance, preventive maintenance, and condition-based maintenance, including models and algorithms in manufacturing [12,13,14]. However, nowadays, other kinds of strategies are also considered. Examples of these techniques are cloud-based predictive maintenance [15] and mobile agent technologies [16]. According to [12], maintenance approaches capable of monitoring equipment conditions for diagnostic and prognostic purposes can be grouped into three main categories: statistical approaches, artificial intelligence approaches and model-based approaches. As model-based approaches need mechanistic knowledge and theory of the equipment to be monitored, and statistical approaches require mathematical background, artificial intelligence approaches, and machine learning techniques in particular, have been increasingly applied in predictive maintenance applications [17,18].
Examples of these techniques are Multiple Classifiers [19], Random Forests (RFs) [20,21] and Support Vector Machines (SVMs) [22,23,24,25], among many others [17]. These strategies take into account issues like real-time alarm monitoring in the manufacturing cell and the general status of the machine in order to determine the best timing to take maintenance actions. The acquired data collects and saves the machine state by means of communication interfaces connected between the machine and the computing servers that host the maintenance algorithms.
Anomaly detection refers to identifying data values that significantly deviate from typical behavior, which can be caused by various factors such as errors in the acquisition system or industrial equipment malfunction [26]. Anomaly detection can be approached through centralized or distributed solutions and using statistical or machine learning (ML) methods. Statistical approaches are based on the distribution of variables, while ML provides techniques for handling high-dimensional data and identifying hidden relationships in complex environments [27]. While previous works typically relied on fully labeled data, such scenarios are less common in practice, because labels are particularly difficult (or impossible) to obtain during the training stage. Furthermore, even when labeled data are available, there could be biases in the way samples are labeled, causing distribution differences. Such real-world data challenges limit the achievable accuracy of prior methods in detecting anomalies. In contrast, unsupervised methods perform their tasks without resorting to class labels. This framework aims at learning the steady state of a working machine and, from that knowledge, trying to predict anomalous behavior. One example of algorithm that follows this principle is the One-Class Classifier [28]. This framework makes it possible to work with very few training examples and to implement online anomaly detection solutions that are proven to complement a classifier in the task of detecting different types of anomalies [29].
In real-time operating systems, the ability to detect anomalies promptly is critical to maintaining system stability, safety, and performance. One of the main challenges in such systems is the dynamic nature of the data, which can vary continuously over time. Anomalies often manifest as deviations from typical patterns, but their detection requires a method that adapts to this changing environment. This is where the sliding window technique proves invaluable. By using a sliding window, the system can continuously monitor a fixed-size subset of the most recent data points, allowing for real-time anomaly detection without needing to process the entire dataset. This method helps in capturing local variations in data, making it especially useful in environments where system behavior fluctuates rapidly. The sliding window enables timely detection of outliers or failures, ensuring that the system can respond to anomalies as they emerge without the delay of waiting for the entire history to be processed. In the fields of industry and predictive maintenance, sliding-window-based anomaly detection has been successfully implemented to detect outliers in hydrological time series [30], to detect anomalous vibrations in robots [31], and for abnormal network traffic detection [32]. To the best of our knowledge, no prior research in the field of industrial predictive maintenance has explored the integration of machine learning model auditing to enhance performance through unsupervised anomaly detection techniques. Notably, this approach eliminates the need for access to the entire historical dataset, making it uniquely suited for deployment in real-time operational environments within the production chain. This represents a novel contribution, bridging the gap between model auditing and the practical constraints of industrial settings.

3. Expert System Description and Objective of the Work

Figure 2 shows the global structure of the proposed expert system. Packaging machine sensors periodically report information about the status of the machine to the company’s ERP system through a TCP/IP connection and an SQL database. This information is read by the expert system module and is preprocessed by the filtering module of the expert system in order to prepare the data to be used by the classification algorithms. Then, the classifier module writes the information related to the timing when a maintenance action has to be taken in the ERP SQL database. The objective of this work will be the development of an anomaly detection stage, following the existing pretrained classifier, to optimize its precision and recall. The main novelty and contribution of this work is the development of dynamic unsupervised anomaly detection techniques that (a) avoid the tedious labeling process and (b) allow for rapid deployment in a complex industrial maintenance and scheduling system for packaging machines, without the need for a long training process. To achieve this goal, we developed and tested an online framework based on sliding windows in combination with several anomaly detection techniques.

4. Methods

4.1. Data Preprocessing

The specification of the expert system imposed by the company was that it should provide a daily prediction to allow for programming of the daily schedule of work orders. Thus, the sampling period at which the patterns from each machine in the training set are acquired is one day. In order to build the dataset, daily information about all the events from a machine (alarm occurrence, counters, working orders, and tasks associated with these working orders) was collected, and then the information was organized in a vector containing all the information from that particular day. The alarm variables provided daily monitoring information that was used to train the classifier. An alarm event indicates some kind of machine malfunction. Table 1 shows the possible causes of an alarm event.
Historical events from the previous working order are of paramount importance to the company. Therefore, a first-order Infinite Impulse Response (IIR) filter was applied to the sensor signals from the machines to account for historical information. This filter weighs past information in a decreasing way, that is, the more recent the event, the more relevant it is, as represented in Equation (1):
y ( n ) = α y ( n 1 ) + x ( n ) ,
where x ( n ) is the current input, and y ( n 1 ) is the previous output. The α parameter ( 0 < α < 1 ) is a smoothing factor used to weigh the past information. This filtering provides a decreasing sum of the frequency of occurrence of each alarm from the date of the event to the last working order. The α parameter was empirically set to 0.63 after tests conducted by company experts, resulting in an IIR low-pass filter.
The target (binary) output defines whether there was a maintenance action for that day. The classes in the data set were extremely imbalanced (1.3% positive (maintenance action required) and 98.7% negative (maintenance not required)). Therefore, the SMOTE method [33] was employed to minimize the effect of class imbalance.

4.2. Classifier

The pretrained classifier used by the packaging company was also developed by the authors of this article and consists of a Random Forest (RF) used to classify the binary variable that encodes whether a maintenance action took place on a particular date for a specific machine. RFs are a popular machine learning algorithm that can be used for both classification and regression tasks. They work by creating multiple decision trees, each trained on a random subset of the data and a random subset of the features. The final prediction is then made by averaging the predictions of all individual trees [20]. Random forests are known for their high accuracy and robustness to overfitting. They have been successfully applied in various fields, such as finance, industry, and medicine.
In particular, a single RF was trained using the R caret package [34] to predict maintenance actions for all machines. The selected model is composed of 500 trees (ntree parameter) and uses 17 randomly sampled predictors (mtry parameter). The resulting Area Under the ROC was 0.848 on the test set (50% of the data). Theoretically, this result is satisfactory, but due to the significant imbalance between classes, a large number of false positives occurred, resulting in very low F1 scores (approximately 0.2 on average).

4.3. Anomaly Detection Methods

Anomaly detection (AD) is the identification of observations that do not conform to an expected pattern or other items in a dataset. It is an important data mining tool for discovering induced errors, unexpected patterns, and more general anomalies.
There are various ways of performing AD, and the approach to take will depend on the nature of the data and the types of anomalies to detect. When labels are scarce or not available, unsupervised learning schemes may be considered. Basically, this model identifies outliers during the fitting process, where it learns a steady state and can then infer abnormal variations (or novelties) from it. It can be used when outliers are defined as points that exist in low-density regions of the dataset. Thus, any new observation which does not belong to high-density areas will be labeled as such automatically by the algorithm.
In the anomaly detection stage of the developed framework, two of the most successful AD methods were compared, while also creating an ensemble of them. These are the One-Class Support Vector Machine (OCSVM) and Minimum Covariance Determinant (MCD).

4.3.1. OCSVM

The OCSVM is a variant of the traditional Support Vector Machine (SVM) that is trained on only one class of data, i.e., the normal data. The algorithm learns the boundaries of the normal data and identifies any data points that fall outside these boundaries as anomalous [35]. This algorithm can be used both in a supervised or unsupervised manner. The OCSVM has been widely used in various fields such as finance, cybersecurity, medical diagnosis, and fault detection in industrial systems.
The OCSVM model used in this work is a wrapper of Python’s scikit-learn one-class SVM class with more functionalities, included in the pyOD package [36]. In particular, we used an OCSVM with a radial basis function (RBF) kernel. Given a dataset { x i } i = 1 n , where x i R p , the objective of the OCSVM with the RBF kernel is to find the optimal hyperplane that separates the data from the origin, capturing the regions with the majority of the data points.
The RBF kernel function is used for mapping the input data into a higher-dimensional feature space and is defined by the following:
K ( x i , x j ) = e γ x i x j 2 ,
where γ > 0 is a parameter that controls the width of the Gaussian function.
The primal problem of the OCSVM with the RBF kernel can be formulated as follows:
min w , ξ i , ρ 1 2 w 2 + 1 ν n i = 1 n ξ i ρ ,
which is subject to the constraints
w , ϕ ( x i ) ρ ξ i , ξ i 0 , i ,
where we have the following definitions:
  • w is the weight vector in the feature space;
  • ξ i are the slack variables representing margin violations;
  • ρ is the offset term (the decision function’s threshold);
  • ϕ ( x i ) is the feature mapping function;
  • ν controls the fraction of outliers and the margin (a value between 0 and 1).
The goal is to find w , ξ i , and ρ such that the margin is maximized while minimizing the slack variables ξ i . It is more common to solve the dual problem for the OCSVM. The dual problem involves introducing Lagrange multipliers α i for the constraints. After applying the Lagrangian, the dual problem is defined as follows:
max α i = 1 n α i K ( x i , x i ) i , j = 1 n α i α j K ( x i , x j ) ,
which is subject to the constraints
i = 1 n α i = 1 , 0 α i 1 ν n , i ,
where α i are the Lagrange multipliers that correspond to each data point.
The dual problem involves maximizing the Lagrange multipliers α i subject to the given constraints. After solving the dual problem, the decision function is used to classify new data points x as normal or anomalous:
f ( x ) = s g n i = 1 n α i K ( x i , x ) ρ .
The classification of a point based on f ( x ) is as follows:
  • If f ( x ) = 1 , the point is classified as “normal”.
  • If f ( x ) = 1 , the point is classified as an “anomaly”.
Unlike linear or polynomial kernels, the RBF is more complex and efficient at the same time that it can combine multiple polynomial kernels multiple times and using different degrees to project the non-linearly separable data into a higher dimensional space so that they can be separable using a hyperplane.

4.3.2. MCD

The MCD estimator is another popular robust method for estimating multivariate location and scatter. It finds the best fitting elliptical distribution by minimizing the determinant of the covariance matrix subject to a constraint on the number of observations. First, it fits a minimum covariance determinant model and then computes the Mahalanobis distance as the outlier degree of the data.
Supposing the data are stored in an n × p matrix X = { x 1 x p } , the classical tolerance ellipse is defined as the set of p-dimensional points x whose Mahalanobis distance is defined in terms of the statistical distance (d) as follows:
M D ( x ) = d ( x , x ¯ , C o v ( X ) ) = ( x x ¯ ) C o v ( X ) 1 ( x x ¯ )
equals the cutoff value x p , 0.975 2 (We denote x p , α 2 as the α quantile of the x p 2 distribution). Here, x ¯ is the sample mean, and C o v ( X ) is the sample covariance matrix. The Mahalanobis distance M D ( x i ) should tell us how far away x i is from the center of the data cloud, relative to its size and shape, and consequently, it determines an outlier degree of the data point. On the other hand, the robust tolerance ellipse used by the MCD estimator is based on the robust distance:
R D ( x ) = d x , μ ^ MCD , Σ ^ MCD ,
where μ ^ MCD is the MCD estimate of location, and Σ ^ MCD is the MCD covariance estimate.
The MCD estimator is resistant to outliers and can handle high-dimensional data. It can also be applied both in supervised or unsupervised learning. A study by Rousseeuw and Driessen [37] showed that the MCD estimator outperformed other robust estimators in terms of efficiency and breakdown point. Another study by Croux and Haesbroeck [38] demonstrated the effectiveness of the MCD estimator in real-world applications.
The MCD model used in this work is a wrapper of Python’s scikit-learn MinCovDet class function with more functionalities that is included in the pyOD package [36].

4.3.3. AD Ensembles

Our initial findings when testing the aforementioned AD methods were that none of them obtained the best precision and recall in all scenarios. That is the reason why we proposed to combine their outputs using a model ensemble. Several types of model ensembles exist in the literature, like voting, averaging, weighted averaging, etc. In this work, a voting ensemble has been selected.
A voting ensemble works by combining the predictions from multiple models by returning the most voted result. It can be used for classification or regression. In the case of regression, this involves calculating the average of the predictions from the models. In the case of classification, the predictions for each label are summed, and the label with the majority vote is predicted. In the binary case (normal or abnormal observation), a majority voting with two models requires that both models must agree that the observation is abnormal in order to tag it as an anomaly.
Recent studies [39] have shown that ensemble methods outperform individual models in complex and imbalanced settings like anomaly detection. The ensemble approach works by reducing variance and bias, improving stability, and making the detection process more adaptable to different anomaly types and data distributions.

4.4. Streaming Framework

Real-time AD is crucial in industrial environments, as it helps in promptly recognizing and resolving errors before they cause major failure or even the destruction of industrial equipment. It can also provide insight on the causes that lead to those failures based on the ongoing activities of a system in near real time, enabling better decisions by detecting anomalies as they occur [40].
To enable a real-time implementation of the AD models in an industrial environment, we proposed a streaming framework based on sliding windows. For this purpose, this work profited from Python’s pySAD library [41]. This library is open-source and compatible with pyOD. The window scheme was based on the study by Manzoor et al. [42] that addressed the outlier detection problem for feature-evolving streams using an algorithm called xStream. In our work, we used a similar window scheme but with different detectors (the OCSVM, MCD, and voting ensemble).
Essentially, the AD models observe the probabilistic output from the classifier for a particular machine and for a determined period of time. This initial window makes it possible to train the AD models in an unsupervised way. In this stage, the models extract information about the normal or abnormal status of the machine. This initial window size was set to 30, which corresponds to 30 daily data samples. During this training period, no output is given by the AD system.
Once the initial window frame has passed, the probabilistic output from the classifier is fed sequentially to the AD models, which are fitted to the new observation and provide an anomaly score. This score is then normalized to the range [0, 1] to provide a new probability value. The probability value is then compared with the same decision threshold used by the classifier (that was optimized during its training stage) to convert the returned probability to a binary value (whether a work order should be taken or not). The streaming dynamics of the proposed approach are shown in Figure 3, and the final architecture proposed for the AD module is summarized in Figure 4.

4.5. Performance Metric

The performance metric used to compare the existing classifier (baseline) with the proposed methods was the F1 score, which is a robust metric that combines both precision and recall performance outcomes:
F 1 = 2 precision · recall precision + recall .
where precision and recall are defined as
precision = T P T P + F P ,
and
recall = T P T P + F N ,
where T P is the number of true positives, F P is the number of false positives, and F N is the number of false negatives.
All results refer to the test set, i.e., after the initial training window, to evaluate the generalization capabilities of the AD methods.

5. Results

We utilized machine records with the longest data history (more than 1000 days of run time) to test the proposed methodology. This approach ensured that the performance evaluation was realistic and reflective of long-term outcomes. This restriction reduced the number of machines to 23. The AD models were built and trained individually for each one of the machines. As an example, in Figure 5, we represent the classifier output probability and the anomaly likelihood obtained by the OCSVM for one of the machines. The true labels are also represented with a different line. The threshold applied was obtained during the classifier’s training to optimize the Area Under the Receiver Operating Characteristic Curve (AUROC) and determined the part of the signal where each method predicted a work order.
In this figure, we can observe that the proposed AD auditing stage started producing a non-null output after the first 30 samples used to train the model. After that period, it filtered the classifier output and made it more selective when predicting a maintenance action. This behavior was consistent with the rest of the machines. The effect of the AD stage was quantified by using a performance metric with respect to the true labels for each data point.
The data used to calculate the score were, in all cases, the inferred data after the initial training period of 30 days. The obtained F1 scores for the three AD methods are compared in Table 2. This table includes the baseline classifier F1 score to better assess the improvement offered by the AD methods.
In quantitative terms, the voting ensemble was the best method tested, with a test F1 score of 0.48 ± 0.17 for the 23 machines. The OCSVM and MCD obtained F1 scores of 0.44 ± 0.17 and 0.44 ± 0.19, respectively. All methods significantly improved the baseline model that produced a test F1 of 0.21 ± 0.10. The maximum absolute F1 score improvement was obtained for machine 25ARE2200:2AB-0118. Using the ensemble model, the F1 score was increased to 0.784 from a baseline of 0.277. In relative terms, the highest increase in F1 was achieved for the machine 25ARE22V2: 2BC-0248, where the ensemble method reached a +933.3% improvement compared to the baseline. The minimum absolute variation in the F1 score was obtained by the MCD for machine 25ARF22V2:1AA-0108 (−5.0%). However, this decrease in the F1 score with respect to the baseline only occurred with one machine, so it can be considered a very unlikely scenario.
A statistical analysis was conducted to determine the statistical significance of the differences between the three proposed methods and the baseline. Figure 6 shows the differences between the F1 distributions for each method. From the visual inspection of the boxplots, the distributions of the OCSVM, MCD, and the voting ensemble differ significantly from the baseline (non-audited classifier). Specifically, the interquartile ranges (Q1–Q3) do not overlap. The voting ensemble was the superior method in terms of the median score, followed by the OCSVM. A Shapiro–Wilk test was used to demonstrate the normality of each group (p > 0.1), and the one-way ANOVA test was used to indicate significant differences between the means (p < 0.001).
The isolated recall and precision scores are also shown in Table 3 and Table 4, respectively. The results in Table 3 show that, in terms of recall alone, the baseline method often performed better than the AD techniques proposed for the 23 packaging machines, followed by the OCSVM. The maximum average percentage improvement of the AD methods was in this case negative (−15.5%). Alternatively, Table 4 reflects a much better performance of the proposed AD methods in terms of precision. The leading method was in this case the voting ensemble, and the maximum percentage change in the F1 score using the best method for each machine reached 675.9% This suggests that while the standard classifier captured most anomalies, it also generated a large number of false positives, which were corrected by the AD methods. The classifier’s threshold used in this study was set to optimize the F1 score, which balances precision and recall. This is evident from the observed improvements in F1 scores across the machines when using the AD module and the fact that the precision was markedly improved while maintaining competitive recall.
One of the key challenges in predictive maintenance within industry settings is the accurate identification of anomalous events or failures to prevent costly downtime. In this context, the OCSVM and MCD methods have emerged as effective techniques to audit and correct the output of classifiers. The OCSVM algorithm is widely used for anomaly detection tasks and has shown promising results in various industrial applications, including fault detection and diagnosis. On the other hand, the MCD estimator is a robust estimator of multivariate location and scatter, which makes it suitable for detecting outliers in datasets with high-dimensional features. Numerous studies have demonstrated the effectiveness of the OCSVM and MCD methods in auditing and correcting classifier outputs for predictive maintenance tasks, improving the overall accuracy and reliability of anomaly detection. These methods provide a valuable means of refining the results of classifiers, enhancing the decision-making process, and facilitating more efficient maintenance strategies in industrial settings. However, according to our findings, none of the aforementioned methods has absolute superiority over the other, so they need to be evaluated separately, and their performance outcomes need to be compared.
We also proposed an ensemble method that tags a sample as an outlier only if both methods are in agreement (voting). Ensemble methods have gained significant popularity in various fields due to their ability to improve the robustness and overall performance of predictive models compared to individual methods. One key advantage of ensemble methods is their capacity to reduce the risk of overfitting by combining multiple models. By aggregating the predictions of these models, ensemble methods can effectively smooth out errors and biases that may be present in individual models, leading to more accurate and robust predictions. Moreover, ensemble methods can handle complex relationships and capture non-linear patterns by combining different modeling techniques or incorporating diverse feature representations. This versatility allows ensembles to adapt to a wide range of data distributions and increase the generalization capability of the model. In addition, ensemble methods can provide better exploration of the feature space and overcome the limitations of individual models, thus enhancing the model’s ability to capture important and relevant patterns.
On paper, the high AUC obtained on the test set by the original classifier previously developed by our group is satisfactory. However, when the client company deployed it to predict work orders in the ERP system, due to the class imbalance, no decision threshold on the signal captured from each machine was found to be optimal: it either prioritizes false positive detection or true negative detection, which results in a low F1 score. The main advantage of adding the proposed AD stage is the significant increase in the F1 score without modifying the input information to the model. Since the model is in production, this aspect is of paramount importance.
From the results of our experiments, we observe that all AD techniques contribute to improve the classifier performance. In particular, the ensemble method has obtained the best results in terms of its F1 score, followed by the MCD and OCSVM in that order. Out of the 23 machines, the ensemble method obtained the best performance in 20 cases, proving its consistency, although it tied with the MCD estimator in 14 of those cases. In four cases (machines 25ARE22V2:2BC-0248, 25ARF2200:101-0027, 25ARF22V2:1AA-0098, and 25ARF22V2:1AA-0113 (Aranco, Valencia, Spain)) it improved the performance of the MCD or OCSVM alone. With regard to the OCSVM, despite not being the best method in many cases, it achieved a similar average F1 score to that of the MCD and showed potential to improve beyond the ensemble performance in three of the cases.
The percentage of improvement in the F1 score with respect to the baseline model was led by the ensemble method, with an average improvement of 192.5%, followed by the MCD with an improvement of 146.7% and the OCSVM with 140.1%. Taking the best of the three methods, the maximum average percentage increase in the F1 score reached 198%, as stated in Table 2.
Although the heterogeneous results reflect differences in the time series of each machine, they also highlight the adaptability of the proposed methods to varying data characteristics.

6. Conclusions and Future Work

This paper proposes a framework to audit the classifier implemented in an expert system used for planning predictive maintenance policies in the ERP (Enterprise Resource Planning) system of a packaging machine company.
The study was conducted on 23 time series with more than 1000 data points, corresponding to daily predictions of the degree of likelihood of a work order. An unsupervised AD module was cascaded at the pretrained classifier’s output so that, for every new sample, it produced three outputs (based on the OCSVM, MCD, and an ensemble of them). These predictions were used to audit and correct the classifier’s output. This new score was normalized and thresholded to determine if the sample actually corresponded to an anomaly.
By examining the results, we observe that all AD techniques contributed to improve the classifier performance. More specifically, the ensemble method obtained the best results in terms of the F1 score with an average increase in the F1 score of 192.5% on the test set, although the individual AD methods also produced an average increase above 140%.
The main contribution of the proposed methods lies in their integration that streams unsupervised anomaly detection with a sliding window, enabling real-time auditing of classifiers as new samples arrive. The implemented methods demonstrate their ability to enhance classifier performance while maintaining simplicity and adaptability. This strategy significantly improves precision and recall in an already optimized classifier, yielding notable improvements for this specific system. Moreover, the approach has the potential to be easily transferred to other similar predictive maintenance settings.
Future work will focus on improving the applicability of methods by incorporating domain-specific features, refining ensemble weighting mechanisms, exploring adaptive approaches to improve performance consistency across diverse scenarios, and conducting comprehensive comparisons with state-of-the-art deep learning techniques to evaluate their relative effectiveness and potential integration.

Author Contributions

Conceptualization, A.J.S.-L., J.G.-S. and J.V.-F.; methodology, F.M., A.J.S.-L. and M.M.-S.; software, F.M. and J.V.-F.; validation, E.S.-O., J.G.-S. and F.M.; formal analysis, F.M., E.S.-O. and M.M.-S.; investigation, F.M. and A.J.S.-L.; resources, E.S.-O. and J.V.-F.; data curation, J.G.-S., A.J.S.-L. and M.M.-S.; writing—original draft preparation, F.M.; writing—review and editing, A.J.S.-L., E.S.-O. and J.V.-F.; visualization, F.M.; supervision, A.J.S.-L. and J.V.-F.; project administration, A.J.S.-L. and J.V.-F.; funding acquisition, A.J.S.-L. and J.V.-F. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by MCIN/AEI/10.13039/501100011033 “ERDF A way of making Europe” through grant number PID2021-127946OB-I00.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available upon request from the corresponding author due to privacy restrictions.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Ayvaz, S.; Alpay, K. Predictive Maintenance System for Production Lines in Manufacturing: A Machine Learning Approach Using IoT Data in Real-Time. Expert Syst. Appl. 2021, 173, 114598. [Google Scholar] [CrossRef]
  2. Froger, A.; Gendreau, M.; Mendoza, J.E.; Pinson, E.; Rousseau, L.M. Maintenance scheduling in the electricity industry: A literature review. Eur. J. Oper. Res. 2016, 251, 695–706. [Google Scholar] [CrossRef]
  3. Stock, T.; Seliger, G. Opportunities of sustainable manufacturing in industry 4.0. Procedia CIRP 2016, 40, 536–541. [Google Scholar] [CrossRef]
  4. Yin, S.; Li, X.; Gao, H.; Kaynak, O. Data-Based Techniques Focused on Modern Industry: An Overview. IEEE Trans. Ind. Electron. 2015, 62, 657–667. [Google Scholar] [CrossRef]
  5. Liu, Y.; Zhang, Y.; Yu, Z.; Zeng, M. Incremental supervised locally linear embedding for machinery fault diagnosis. Eng. Appl. Artif. Intell. 2016, 50, 60–70. [Google Scholar] [CrossRef]
  6. Lee, S.S. Noisy replication in skewed binary classification. Comput. Stat. Data Anal. 2000, 34, 165–191. [Google Scholar] [CrossRef]
  7. Zonta, T.; da Costa, C.A.; da Rosa Righi, R.; de Lima, M.J.; da Trindade, E.S.; Li, G.P. Predictive maintenance in the Industry 4.0: A systematic literature review. Comput. Ind. Eng. 2020, 150, 106889. [Google Scholar] [CrossRef]
  8. O’Donovan, P.; Leahy, K.; Bruton, K.; O’Sullivan, D.T. Big data in manufacturing: A systematic mapping study. J. Big Data 2015, 2, 20. [Google Scholar] [CrossRef]
  9. Muhuri, P.K.; Shukla, A.K.; Abraham, A. Industry 4.0: A bibliometric analysis and detailed overview. Eng. Appl. Artif. Intell. 2019, 78, 218–235. [Google Scholar] [CrossRef]
  10. Lee, J.; Bagheri, B.; Kao, H.A. Recent advances and trends of cyber-physical systems and big data analytics in industrial informatics. In Proceedings of the International Conference on Industrial Informatics (INDIN), Porto Alegre, Brazil, 27–30 July 2014; pp. 1–6. [Google Scholar]
  11. Rodríguez-Mazahua, L.; Rodríguez-Enríquez, C.A.; Sánchez-Cervantes, J.L.; Cervantes, J.; García-Alcaraz, J.L.; Alor-Hernández, G. A general perspective of Big Data: Applications, tools, challenges and trends. J. Supercomput. 2016, 72, 3073–3113. [Google Scholar] [CrossRef]
  12. Jardine, A.K.; Lin, D.; Banjevic, D. A review on machinery diagnostics and prognostics implementing condition-based maintenance. Mech. Syst. Signal Process. 2006, 20, 1483–1510. [Google Scholar] [CrossRef]
  13. Lu, B.; Durocher, D.B.; Stemper, P. Predictive maintenance techniques. IEEE Ind. Appl. Mag. 2009, 15, 52–60. [Google Scholar] [CrossRef]
  14. Chouikhi, H.; Khatab, A.; Rezg, N. A condition-based maintenance policy for a production system under excessive environmental degradation. J. Intell. Manuf. 2014, 25, 727–737. [Google Scholar] [CrossRef]
  15. Wang, J.; Zhang, L.; Duan, L.; Gao, R. A new paradigm of cloud-based predictive maintenance for intelligent manufacturing. J. Intell. Manuf. 2015, 28, 1125–1137. [Google Scholar] [CrossRef]
  16. Cucurull, J.; Martí, R.; Navarro-Arribas, G.; Robles, S.; Overeinder, B.; Borrell, J. Agent mobility architecture based on IEEE-FIPA standards. Comput. Commun. 2009, 32, 712–729. [Google Scholar] [CrossRef]
  17. Carvalho, T.P.; Soares, F.A.; Vita, R.; Francisco, R.d.P.; Basto, J.P.; Alcalá, S.G. A systematic literature review of machine learning methods applied to predictive maintenance. Comput. Ind. Eng. 2019, 137, 106024. [Google Scholar] [CrossRef]
  18. Zhang, W.; Yang, D.; Wang, H. Data-driven methods for predictive maintenance of industrial equipment: A survey. IEEE Syst. J. 2019, 13, 2213–2227. [Google Scholar] [CrossRef]
  19. Susto, G.A.; Schirru, A.; Pampuri, S.; McLoone, S.; Beghi, A. Machine learning for predictive maintenance: A multiple classifier approach. IEEE Trans. Ind. Inform. 2014, 11, 812–820. [Google Scholar] [CrossRef]
  20. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  21. Prytz, R.; Nowaczyk, S.; Rögnvaldsson, T.; Byttner, S. Predicting the need for vehicle compressor repairs using maintenance records and logged vehicle data. Eng. Appl. Artif. Intell. 2015, 41, 139–150. [Google Scholar]
  22. Vapnik, V.N. Support Vector Machines. In Encyclopedia of Biometrics; Li, S.Z., Jain, A.K., Eds.; Springer: Boston, MA, USA, 2009; pp. 1365–1368. [Google Scholar] [CrossRef]
  23. Gryllias, K.; Antoniadis, I. A Support Vector Machine approach based on physical model training for rolling element bearing fault detection in industrial environments. Eng. Appl. Artif. Intell. 2012, 25, 326–344. [Google Scholar] [CrossRef]
  24. Li, H.; Parikh, D.; He, Q.; Qian, B.; Li, Z.; Fang, D.; Hampapur, A. Improving rail network velocity: A machine learning approach to predictive maintenance. Transp. Res. Part C Emerg. Technol. 2014, 45, 17–26. [Google Scholar] [CrossRef]
  25. Langone, R.; Alzate, C.; Ketelaere, B.D.; Vlasselaer, J.; Meert, W.; Suykens, J.A. LS-SVM based spectral clustering and regression for predicting maintenance of industrial machines. Eng. Appl. Artif. Intell. 2015, 37, 268–278. [Google Scholar] [CrossRef]
  26. Erhan, L.; Ndubuaku, M.; Di Mauro, M.; Song, W.; Chen, M.; Fortino, G.; Bagdasar, O.; Liotta, A. Smart anomaly detection in sensor systems: A multi-perspective review. Inf. Fusion 2021, 67, 64–79. [Google Scholar] [CrossRef]
  27. Nunes, P.; Santos, J.; Rocha, E. Challenges in predictive maintenance—A review. CIRP J. Manuf. Sci. Technol. 2023, 40, 53–67. [Google Scholar] [CrossRef]
  28. Fernández, A.; García, S.; Galar, M.; Prati, R.C.; Krawczyk, B.; Herrera, F. Learning from Imbalanced Data Sets; Springer: Berlin/Heidelberg, Germany, 2018; Volume 10. [Google Scholar]
  29. Morselli, F.; Bedogni, L.; Mirani, U.; Fantoni, M.; Galasso, S. Anomaly Detection and Classification in Predictive Maintenance Tasks with Zero Initial Training. IoT 2021, 2, 590–609. [Google Scholar] [CrossRef]
  30. Kulanuwat, L.; Chantrapornchai, C.; Maleewong, M.; Wongchaisuwat, P.; Wimala, S.; Sarinnapakorn, K.; Boonya-aroonnet, S. Anomaly Detection Using a Sliding Window Technique and Data Imputation with Machine Learning for Hydrological Time Series. Water 2021, 13, 1862. [Google Scholar] [CrossRef]
  31. Zhong, Z.; Zhao, Y.; Yang, A.; Zhang, H.; Qiao, D.; Zhang, Z. Industrial Robot Vibration Anomaly Detection Based on Sliding Window One-Dimensional Convolution Autoencoder. Shock Vib. 2022, 2022, 1179192. [Google Scholar] [CrossRef]
  32. Li, Z.; Li, G.; Liu, Y.; Li, Z.; Li, G.; Liu, Y. Network traffic anomaly detection based on sliding window. In Proceedings of the 2011 International Conference on Electronics, Communications and Control (ICECC), Ningbo, China, 9–11 September 2011; pp. 1820–1823. [Google Scholar] [CrossRef]
  33. Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic Minority Over-sampling Technique. J. Artif. Int. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
  34. Kuhn, M. Building Predictive Models in R Using the caret Package. J. Stat. Softw. Artic. 2008, 28, 1–26. [Google Scholar] [CrossRef]
  35. Schölkopf, B.; Williamson, R.C.; Smola, A.; Shawe-Taylor, J.; Platt, J. Support vector method for novelty detection. Adv. Neural Inf. Process. Syst. 1999, 12, 582–588. [Google Scholar]
  36. Zhao, Y.; Nasrullah, Z.; Li, Z. PyOD: A Python Toolbox for Scalable Outlier Detection. J. Mach. Learn. Res. 2019, 20, 1–7. [Google Scholar]
  37. Rousseeuw, P.J.; Driessen, K.V. A fast algorithm for the minimum covariance determinant estimator. Technometrics 1999, 41, 212–223. [Google Scholar] [CrossRef]
  38. Croux, C.; Haesbroeck, G. Influence function and efficiency of the minimum covariance determinant scatter matrix estimator. J. Multivar. Anal. 1999, 71, 161–190. [Google Scholar] [CrossRef]
  39. Liu, Y.; Zhu, L.; Ding, L.; Huang, Z.; Sui, H.; Wang, S.; Song, Y. Selective ensemble method for anomaly detection based on parallel learning. Sci. Rep. 2024, 14, 1420. [Google Scholar] [CrossRef] [PubMed]
  40. Grunova, D.; Bakratsi, V.; Vrochidou, E.; Papakostas, G.A. Machine Learning for Anomaly Detection in Industrial Environments. Eng. Proc. 2024, 70, 25. [Google Scholar] [CrossRef]
  41. Yilmaz, S.F.; Kozat, S.S. PySAD: A Streaming Anomaly Detection Framework in Python. arXiv 2020, arXiv:2009.02572. [Google Scholar]
  42. Manzoor, E.; Lamba, H.; Akoglu, L. xstream: Outlier detection in feature-evolving data streams. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, London, UK, 19–23 August 2018; pp. 1963–1972. [Google Scholar]
Figure 1. Automatic fixed wrapping machine F-2200 (Aranco, Valencia, Spain). Source: https://www.aranco.com/en/services/sie-wrappers/fixed-wrapping-machine-f-2200 (accessed on 16 December 2024).
Figure 1. Automatic fixed wrapping machine F-2200 (Aranco, Valencia, Spain). Source: https://www.aranco.com/en/services/sie-wrappers/fixed-wrapping-machine-f-2200 (accessed on 16 December 2024).
Applsci 15 00882 g001
Figure 2. Predictive maintenance expert system. The main contribution of this work is the improvement of the existing classification module by means of unsupervised AD techniques (AD module).
Figure 2. Predictive maintenance expert system. The main contribution of this work is the improvement of the existing classification module by means of unsupervised AD techniques (AD module).
Applsci 15 00882 g002
Figure 3. Anomaly detection streaming procedure for online auditing of the classifier.
Figure 3. Anomaly detection streaming procedure for online auditing of the classifier.
Applsci 15 00882 g003
Figure 4. Proposed architecture for the AD module.
Figure 4. Proposed architecture for the AD module.
Applsci 15 00882 g004
Figure 5. Auditing example for one of the machines using OCSVM. The output from the classifier and the AD model output are represented as different lines, and the shaded part of each signal indicates if the threshold to call for a work order has been exceeded.
Figure 5. Auditing example for one of the machines using OCSVM. The output from the classifier and the AD model output are represented as different lines, and the shaded part of each signal indicates if the threshold to call for a work order has been exceeded.
Applsci 15 00882 g005
Figure 6. Box plots illustrating the differences between the F1 scores distributions obtained by the compared methods.
Figure 6. Box plots illustrating the differences between the F1 scores distributions obtained by the compared methods.
Applsci 15 00882 g006
Table 1. Set of alarms and their description.
Table 1. Set of alarms and their description.
AlarmDescription
A1Pulley failure
A2Arm engine failure
A3Maximum intensity in arm engine failure
A4Offset position failure
A5Communication failure
A6Minimum battery level failure
A7Maximum battery level failure
A8Emergency button
A9Pulley failure
A10Carriage failure
A11Carriage engine failure
A12Vertical bar failure
A13Horizontal bar failure
A14Maintenance failure 1
A15Maintenance failure 2
A16Latch failure
A17Brake communication system failure
A18Plastic film broken
A19Brake off
A20Excess strain on plastic film
A200Communication error with remote board
A201Extended time without communication failure
Table 2. Test set F1 score comparison between the three AD methods and the baseline classifier for 23 packaging machines. The best results for each machine have been highlighted in bold. The maximum relative percentage change is indicated in the rightmost column.
Table 2. Test set F1 score comparison between the three AD methods and the baseline classifier for 23 packaging machines. The best results for each machine have been highlighted in bold. The maximum relative percentage change is indicated in the rightmost column.
MachineBaselineOCSVMMCDEnsembleMax. % Change
25ARE2200:2AB-01180.2770.6000.7840.784+182.7%
25ARE2200:2AB-01400.2440.6150.6500.650+166.6%
25ARE22V2:2BC-02480.0220.0490.1330.222+933.3%
25ARE22V2:2BC-02640.2350.6200.5330.612+164.3%
25ARE22V2:2BC-02680.3210.5450.5960.596+85.9%
25ARF2200:101-00200.1670.4080.2080.408+144.9%
25ARF2200:101-00270.0470.2400.2610.273+481.8%
25ARF2200:101-00350.2880.7560.7560.756+162.6%
25ARF22V2:1AA-00420.1190.2670.2860.286+140.0%
25ARF22V2:1AA-00830.1290.2500.3330.333+ 158.3%
25ARF22V2:1AA-00940.2320.4470.5950.595+156.5%
25ARF22V2:1AA-00980.1250.5710.1710.600+380.0%
25ARF22V2:1AA-00990.2780.4250.4360.436+56.5%
25ARF22V2:1AA-01080.3920.4190.3720.381+6.9%
25ARF22V2:1AA-01090.2380.3080.3700.370+55.6%
25ARF22V2:1AA-01100.2120.3100.3330.333+56.9%
25ARF22V2:1AA-01130.1400.4210.3560.516+267.7%
25ARF22V2:1AA-01730.1950.4380.5240.524+167.9%
25ARF22V2:1AA-01760.1450.3670.2000.200+153.2%
25ARF22V3:1BB-01840.1180.4000.4000.400+240.0%
25ARF22V3:1BB-01880.1270.3640.3810.381+200.0%
25ARF22V3:1BB-01890.3410.6040.6720.672+97.0%
25ARF22V4:1CC-03080.3570.6760.7010.701+96.5%
Average ± std0.21 ± 0.010.44 ± 0.170.44 ± 0.190.48 ± 0.17+198.0%
Table 3. Recall performance comparison between baseline, OCSVM, MCD, and voting ensemble methods. The maximum relative percentage change is indicated in the rightmost column, with the values responsible for the maximum change highlighted in bold for each row.
Table 3. Recall performance comparison between baseline, OCSVM, MCD, and voting ensemble methods. The maximum relative percentage change is indicated in the rightmost column, with the values responsible for the maximum change highlighted in bold for each row.
MachineBaselineOCSVMMCDEnsembleMax. % Change
25ARE2200:2AB-01181.0000.8750.8330.833−12.5%
25ARE2200:2AB-01400.8910.6730.6530.653−24.4%
25ARE22V2:2BC-02480.1250.1250.1250.1250.0%
25ARE22V2:2BC-02640.8880.6130.5630.563−36.6%
25ARE22V2:2BC-02680.8890.7110.6560.656−26.3%
25ARF2200:101-00200.8950.5260.5260.526−41.2%
25ARF2200:101-00270.3750.3750.3750.3750.0%
25ARF2200:101-00350.8720.7230.7230.723−17.1%
25ARF22V2:1AA-00420.7140.5710.5710.571−20.0%
25ARF22V2:1AA-00830.2860.2860.2860.2860.0%
25ARF22V2:1AA-00940.8060.5810.5810.581−28.0%
25ARF22V2:1AA-00980.5710.5710.4290.429−25.0%
25ARF22V2:1AA-00990.9570.7390.7390.739−22.7%
25ARF22V2:1AA-01080.9050.4290.3810.381−57.9%
25ARF22V2:1AA-01091.0001.0001.0001.0000.0%
25ARF22V2:1AA-01100.7500.5630.5630.563−25.0%
25ARF22V2:1AA-01130.7270.7270.7270.7270.0%
25ARF22V2:1AA-01730.8390.7420.7100.710−15.4%
25ARF22V2:1AA-01760.8890.6110.1670.167−31.3%
25ARF22V3:1BB-01840.5000.5000.5000.5000.0%
25ARF22V3:1BB-01881.0001.0001.0001.0000.0%
25ARF22V3:1BB-01890.9150.8940.8510.851−7.0%
25ARF22V4:1CC-03081.0000.8410.8410.841−15.9%
Average ± std0.77 ± 0.240.64 ± 0.220.60 ± 0.240.60 ± 0.24−15.5%
Table 4. Precision comparison between the three AD methods and the baseline classifier for 23 packaging machines. The maximum relative percentage change is indicated in the rightmost column, with the values responsible for the maximum change highlighted in bold for each row.
Table 4. Precision comparison between the three AD methods and the baseline classifier for 23 packaging machines. The maximum relative percentage change is indicated in the rightmost column, with the values responsible for the maximum change highlighted in bold for each row.
MachineBaselineOCSVMMCDEnsembleMax. % Change
25ARE2200:2AB-01180.1610.4570.7410.741359.9%
25ARE2200:2AB-01400.1410.5670.6470.647358.0%
25ARE22V2:2BC-02480.0120.0300.1431.0008400.0%
25ARE22V2:2BC-02640.1350.6280.5060.672396.6%
25ARE22V2:2BC-02680.1960.4410.5460.546179.3%
25ARF2200:101-00200.0920.3330.1300.333262.7%
25ARF2200:101-00270.0250.1760.2000.214757.1%
25ARF2200:101-00350.1720.7910.7910.791359.0%
25ARF22V2:1AA-00420.0650.1740.1900.190193.3%
25ARF22V2:1AA-00830.0830.2220.4000.400380.0%
25ARF22V2:1AA-00940.1360.3640.6100.610350.3%
25ARF22V2:1AA-00980.0700.5710.1071.0001325.0%
25ARF22V2:1AA-00990.1630.2980.3090.30989.7%
25ARF22V2:1AA-01080.2500.4090.3640.38163.6%
25ARF22V2:1AA-01090.1350.1820.2270.22768.2%
25ARF22V2:1AA-01100.1240.2140.2370.23791.4%
25ARF22V2:1AA-01130.0780.2960.2350.400415.0%
25ARF22V2:1AA-01730.1110.3110.4150.415275.2%
25ARF22V2:1AA-01760.0790.2620.2500.250232.3%
25ARF22V3:1BB-01840.0670.3330.3330.333400.0%
25ARF22V3:1BB-01880.0680.2220.2350.235247.1%
25ARF22V3:1BB-01890.2100.4560.5130.513144.8%
25ARF22V4:1CC-03080.1960.6290.7040.704258.2%
Average ± std0.12 ± 0.060.36 ± 0.180.38 ± 0.200.48 ± 0.24675.9%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Mateo, F.; Vila-Francés, J.; Soria-Olivas, E.; Martínez-Sober, M.; Gómez-Sanchis, J.; Serrano-López, A.J. Dynamic Classifier Auditing by Unsupervised Anomaly Detection Methods: An Application in Packaging Industry Predictive Maintenance. Appl. Sci. 2025, 15, 882. https://doi.org/10.3390/app15020882

AMA Style

Mateo F, Vila-Francés J, Soria-Olivas E, Martínez-Sober M, Gómez-Sanchis J, Serrano-López AJ. Dynamic Classifier Auditing by Unsupervised Anomaly Detection Methods: An Application in Packaging Industry Predictive Maintenance. Applied Sciences. 2025; 15(2):882. https://doi.org/10.3390/app15020882

Chicago/Turabian Style

Mateo, Fernando, Joan Vila-Francés, Emilio Soria-Olivas, Marcelino Martínez-Sober, Juan Gómez-Sanchis, and Antonio José Serrano-López. 2025. "Dynamic Classifier Auditing by Unsupervised Anomaly Detection Methods: An Application in Packaging Industry Predictive Maintenance" Applied Sciences 15, no. 2: 882. https://doi.org/10.3390/app15020882

APA Style

Mateo, F., Vila-Francés, J., Soria-Olivas, E., Martínez-Sober, M., Gómez-Sanchis, J., & Serrano-López, A. J. (2025). Dynamic Classifier Auditing by Unsupervised Anomaly Detection Methods: An Application in Packaging Industry Predictive Maintenance. Applied Sciences, 15(2), 882. https://doi.org/10.3390/app15020882

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop