Next Article in Journal
Determination of Save Excavation Criteria in Velenje Coal Mine
Previous Article in Journal
Exploratory Review on Environmental Aspects of Enhanced Weathering as a Carbon Dioxide Removal Method
Previous Article in Special Issue
Acoustic Sensing of Fresh Feed Disturbances in a Locked-Cycle Laboratory AG/SAG Mill
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Monitoring of Mineral Processing Operations with Isolation Forests

1
Western Australian School of Mines: Minerals, Energy and Chemical Engineering, Curtin University, GPO Box U1987, Perth, WA 6845, Australia
2
Department of Process Engineering, Stellenbosch University, Private Bag X1, Matieland, Stellenbosch 7602, South Africa
*
Author to whom correspondence should be addressed.
Minerals 2024, 14(1), 76; https://doi.org/10.3390/min14010076
Submission received: 17 October 2023 / Revised: 9 December 2023 / Accepted: 25 December 2023 / Published: 8 January 2024

Abstract

:
Reliable monitoring of mineral process systems is key to more efficient plant operation. Multivariate statistical process control based on principal component analysis is well-established in industry but may not be effective when dealing with dynamic nonlinear or transient processes, where process behavior may change rapidly from time to time. Although a large variety of nonlinear models have been proposed to address these problems, the monitoring of complex dynamic process systems remains challenging. Isolation forests are unsupervised machine learning models that provide an interesting approach to process monitoring that has not been explored extensively yet. In this investigation, this approach is compared with traditional multivariate statistical process monitoring based on principal component models. Three real-world case studies are considered. The first case study is based on coal flotation, the second is based on features extracted from a platinum group metal flotation froth; and the third is based on data from an industrial semi-autogenous grinding circuit. In each case, the models were trained on data representing normal operating conditions and then tested on new process data that were generally different from the training data to test their ability to identify these data as out-of-control. The isolation forest models performed better than the principal component models when the data were nonlinear, but not when the data associated with normal operation and faulty conditions were linearly separable, as was the case with the flotation data.

Graphical Abstract

1. Introduction

Effective monitoring of process circuits is critical to efficient operation and key to predictive maintenance of equipment [1]. In particular, identification of process operations that may not obviously deviate from normal or optimal operating conditions can be difficult, especially when dealing with nonlinear systems. For example, grinding is influenced by a combination of factors, including mill design, operational parameters such as rotational speed, feed rate, and slurry density, and the characteristics of the mineral being processed. These factors interact in complex ways, leading to nonlinear responses that can be difficult to model and control. The same goes for froth flotation, where multiple variables have to be controlled simultaneously to optimize recovery and grade [2,3]. The interactions between these variables can create nonlinear behavior that complicates control strategies.
Process monitoring would typically be based on models capable of predicting the state of the process, whether it is in control or not, when presented with measurements of operational conditions. Although fundamental models can be used for this purpose, data-based models are more common, as these can be developed more rapidly and provide more reliable results than first-principle models. The disadvantage is that they may be more costly to maintain. Furthermore, monitoring of plant operations can either be supervised or unsupervised.
With supervised monitoring, the model is trained on labeled data from normal operating conditions (NOCs) as well as data from known fault conditions. Well-established models that are used for classification can be used for this purpose, such as random forests, boosted trees, or deep neural networks, as is the case in more recent studies [4,5,6]. The advantage of supervised monitoring is that known fault conditions can usually be rectified easily as their sources are well understood. The disadvantage of supervised process monitoring is that it may misclassify observations that do not belong to previously identified fault classes, as indicated schematically in the left-hand panel of Figure 1, where Fault 3 is an unknown fault that a machine learning model would be unable to identify or would misidentify, as it would not have been part of its training base. In addition, data from known fault conditions may be difficult to obtain.
In contrast, with unsupervised monitoring, the model is trained to simply predict whether the process is operating normally or in-control, or whether some unknown fault condition has arisen and the process is therefore out-of-control (e.g., [1,7,8,9]). This has the advantage that data regarding specific fault conditions, which may be scarce or unavailable, are not required to construct models. All that is required is data associated with normal operating conditions. The disadvantage of this approach is that it may be more challenging to rectify out-of-control operations as the source of the fault is generally unknown. In Figure 1, the right-hand panel indicates that a machine learning model constructs a representation (the dashed lines) of the NOC data. Any new data located on the outside of this decision surface would be classified as not NOC. The challenge is that these decision surfaces could have complex topologies in high-dimensional spaces that may not be easily learned from available process data.
In practice, multivariate approaches to unsupervised monitoring are mostly based on the use of latent variable methods, such as principal component analysis, that are capable of handling a wide variety of systems [10,11]. However, these methods may fail when process behavior is nonlinear.
Under these circumstances, nonlinear methods may be more suitable. Many such methods have been proposed in the past [12,13,14,15]. This includes, among other one-class classification methods [16], kernel principal component analysis [14,15,17,18], multiscale methods [19,20] and autoencoders [21,22]. These methods all address specific aspects of complex systems and may suffer from disadvantages related to robustness, training complexity that may become prohibitive when dealing with large-scale problems, computational cost, multimodal data, etc. As a consequence, the development of process monitoring systems remains an active area of research and development.
Isolation forests have the advantages of linear time complexity and a low memory requirement, which work well with high-volume data. Moreover, the algorithm relies upon the characteristics of anomalies, i.e., being few and different, in order to detect anomalies, which makes them suitable candidates for the development of unsupervised process monitoring systems. In this context, an anomaly is defined as a sample that can be isolated from the rest of the data with comparatively fewer partitions of the data space. This can be an effective approach to identifying samples in less dense regions of the data space, where traditional density-based methods might struggle when anomalies are globally rare. In addition, isolation forests are capable of capturing both linear and non-linear patterns. This flexibility makes them versatile for a wide range of anomaly detection tasks. Although they have been extensively applied in anomaly detection, their use in process monitoring frameworks has not been considered extensively as yet.
Therefore, in this investigation, the use of isolation forests that are trained on the NOC data set and then directly applied to new data that do not have the same statistical distribution as the training data are considered a novel framework for the monitoring of process systems.

2. Process Monitoring Methodology

From a process operator’s point of view, process data are available that have been recorded during normal process operation. The nature of these data would depend on the process or system under operation, but would typically consist of some set of variables or features that have been derived from equipment sensors. For example, this could include measurements of feed rates, feed densities, power consumption, in-line measurements of mineral grades, or features extracted from the vibrational spectra of equipment. Normally, these data would be collected under steady-state conditions. Typically, such data would have to be preprocessed to ensure a common time stamp (t), which may also require embedding with a time lag k and embedding dimension m in the form of X(t), X(t k), … X(t (m 1)k) prior to analysis. However, known direct indicators of process performance, such as recoveries or grades in flotation plants or the particle size of comminuted products in grinding circuits, would not be available.

2.1. Isolation Forests

Isolation forests (IFs) are ensembles of binary trees designed to recursively partition data [23], as indicated in Figure 2. A tree is grown by randomly splitting the feature space until each sample is contained in a terminal leaf node of the tree. Samples are assigned anomaly scores that are inversely related to the length of the branch between the root and the leaf node of the tree. In the final analysis, anomalies are identified based on their average scores across all the trees in the forest.
More formally, consider a data set X R N × M consisting of N samples and M process variables x j R M ,   j = 1,2 ,   M , representative of normal operating conditions on the plant. The construction of a tree in the isolation forest is accomplished by generating nodes in the tree t T by randomly sampling a subset of the NOC training data of size N S N . Subsequent to this, a variable j M , is randomly selected, together with a random split point s in the interval [ x j , m i n , x j , m a x ] . The subset of data x j s is assigned to the left branch of the tree and x j > s , with the right branch in the tree t .
The process is repeated for each of the descendent branches in the tree t and all the trees in the forest ( t = 1 ,   2 ,   T ), and terminated when the maximum tree depth ( L ) or the minimum terminal node size ( N m i n ) is reached.
Once an isolation forest has been constructed, the expected path length, h x i required to isolate the sample x i from the training data is computed from the mean of the path lengths in all the trees. The anomaly score a ( x i ) , for the i’th sample in the training data is defined in Equation (1).
a x i = 2 E [ h x i ] / c ( n )
In this equation, the expected path length is normalized by c ( n ) , which is the mean value of h x i for data sets of size n . This value can be computed using Equation (2), where H ( n ) is the n t h harmonic number, which can be approximated by ln n + γ = ln n + 0.577 , where γ is the Euler–Mascheroni constant [24].
c n = 2 H n 1 2 ( n 1 ) / n
This ensures that when E h x i = c ( n ) , the anomaly score of the sample is a x i = 0.5. Moreover, for large values of h x i , the anomaly score tends to 0, and for small values of h x i compared to c n , a x i tends to 1. This allows the specification of an anomaly threshold, a 0     [ 0 ,   1 ] that would flag a sample x i as an anomaly, that is, if a x i > a 0 .
Table 1 summarizes the hyperparameters that were used to construct the isolation forests used in this investigation. These hyperparameters were mostly determined through trial and error, but a grid search can also be conducted to optimize the models more systematically. More specifically, the anomaly threshold was set to ensure the designation of 5% of the NOC data as anomalous, in keeping with a 95% control limit commonly used in statistical process control systems.
Even so, it should be noted that the performance of the isolation forest was not particularly sensitive to most of these parameters. For example, smaller forests with 50 trees showed somewhat lower performance than isolation forests with 100 or 200 trees, but not by a large margin. The only critical parameter was the anomaly threshold, which was set to ensure a Type II error (false negatives) of less than 5%. This threshold was not set directly but was determined by a contamination factor, which controls the expected proportion of anomalies in the data set. This factor was set at 5%, unless otherwise indicated.
Since first proposed by Liu et al. [25], isolation forests have become established in a variety of applications, including fraud detection, network intrusion detection, and outlier detection in data mining. To date, they have mostly been used in mineral exploration [26,27,28], but some applications have more recently appeared in diagnostic systems in mineral processing [29,30,31].

2.2. Application of Isolation Forest in Process Monitoring

Although isolation forests are machine learning algorithms designed for anomaly detection, they can be applied to a wide range of data analysis tasks, including process monitoring, although few such studies have been reported to date [31,32,33,34].
In the context of process monitoring, anomalies can represent deviations from normal or expected process behavior. By applying isolation forests to process data, unusual patterns, events, or behaviors can be identified that may indicate process deviations or abnormalities.
The IF model accomplishes this through the implicit construction of a decision surface, as illustrated in Figure 3. This figure shows observations of variables X and Y arranged in a square for illustrative purposes. The samples on the corners of the square indicated in red are more easily separated from the other samples, as they would require two splits of the data space only, while the other samples would require three or more splits.
If the samples on the corners of the square are considered to be anomalies, then the decision surface can be indicated by the red and blue-shaded areas on the graph. Data in the red-shaded regions would be considered out-of-control (OOC), and data in the blue-shaded areas would be in-control (IC), even if they are very far away from the NOC data, which is essentially comprised of the five points within the sample that were not considered to be anomalies.

2.3. Multivariate Statistical Process Control Based on Principal Component Analysis

The principal component model is constructed by expressing the data matrix with N observations of M variables X R N × M , as the product of a score matrix ( T k R N × r ) and a loading matrix ( P k R M × r ) consisting of a small number of principal components ( r < M ), as well as a residual matrix ( E r ) [35].
X = T r P r + E r
For the i ’th observation, the process control diagnostics are the Hotelling’s T i 2 statistic and the squared prediction error, or Q i -statistic, as indicated in Equations (4) and (5), respectively. These statistics can be explained by considering that the principal component model is based on a plane that is fitted through the data if there are two variables, or more generally, a hyperplane when there are more than two variables. Hotelling’s T i 2 statistic is based on the projections of the data onto the principal component plane and captures the distance of each such projected point from the center of the plane.
Likewise, the Q -statistic accounts for the distance of the point from the plane. These two statistics are complementary in that points may be close to the plant but far from the center, or they may be projected close to the center of the plane but from a distance far away from the plane. In both of these cases, only one of the statistics will be able to identify the point as OOC. In Equations (4) and (5), x i , t i and e i are the i’th sample, principal component score, and residual.
T i 2 = t i λ r 1 t i = x i P r λ r 1 P r x i
Q i = e i e i = x i ( I r P r P r ) x i
Monitoring of the process is based on a control chart for each of the two statistics ( T i 2 and Q i ), each provided with an appropriate control limit, as indicated by Equations (6)–(9).
C L T 2 = r ( N + 1 ) ( N 1 ) N 2 N r F ( α , r , N r )
C L Q = θ 1 z α 2 θ 2 h 0 2 θ 1 + θ 2 h 0 ( h 0 1 ) θ 1 2 + 1 1 / h 0
h 0 = 1 2 θ 1 θ 3 3 θ 2 2
θ j = i = r + 1 M λ i j
Equation (6), F ( α , r , N r ) is the upper α ’th quantile of the F-distribution with r and N r degrees of freedom.
In Equations (7)–(9), z α is the upper α ’th quantile of the standard normal distribution and λ i is the I’th eigenvalue of the principal component model. The process is considered to be out-of-control (OOC), if either T i 2 > C L T 2 or Q i > C L Q , as illustrated in Figure 4.
When the data are not approximately normally distributed or the data distribution is unknown, the control limits can also be fitted to the training (NOC) data of the model, as was conducted in this investigation.
The following case studies illustrate and assess the performance of the IF and PCA-based methods on a comparative basis.

3. Case Study 1: Coal Flotation

In the first case study, the flotation of coal to separate fine coal from a coal slurry mixture is considered. Five variables were measured, namely the flow rate of the raw slurry into the mixing tank (x1), the density of the slurry (x2), the flow rate of water into the mixing tank (x3), the flow rate of frother into the tank (x4), as well as the flow rate of the kerosene collector into the tank (x5), as shown in Figure 5.
A training data set ( X N O C ) consisting of 200 NOC samples constituted normal operating conditions, while a test data set containing 65 samples with the same distribution as the training data and 200 samples different from the NOC data constituted the NEW data ( X N E W ).
Multiple runs were made to fit the models, each time with a different training and test data set. At each run, an isolation forest with 200 trees and a contamination factor of C = 0.05 was fitted to the training data consisting of 200 samples and tested on the test data set that consisted of 265 samples, 65 of which were NOC data and 200 of which were not.
The typical performance of the IF on the data set can be visualized in Figure 6. In this figure, training (NOC) data are represented by magenta-colored markers and test data by black markers. Samples correctly identified as IC or OOC are indicated by solid markers. In the training data set, 5% of the samples are shown as empty markers, as determined by the contamination factor setting of 5%. The test data indicated by black markers are further characterized by circular and square markers. The square markers are test data, which are IC, while the circular markers are OOC. Solid black square markers therefore indicate that the model correctly identified these markers as IC in the test data set. In contrast, empty black circular markers indicate samples that were incorrectly identified by the IF model as IC.
In this particular case, the model correctly identified 96.9% of the 65 samples in the test set that were IC as such and 41% of the 200 samples in the test set that were OOC as such. The high hit rate of the model on the IC data in the test set is not unexpected, as these data had the same distribution as the data on which the IF model had been trained. In Figure 6, it can be seen that these data (black squares) are located within the same area on the chart as the training data of the model (magenta markers).
In Table 2, the false positive rates of the optimal models on the test data (FP-test) are shown, together with the true positive rates on the test data (TP-test). The principal component model was based on two components and could on average correctly identify 28.0% of the out-of-control samples, although it had a marginally lower false positive rate on the test data than the IF model that could on average identify 32.0% of the OOC samples.
Figure 7 shows the effect of the contamination factor and isolation forest size on their performance in terms of their ability to correctly identify OOC samples in the test data. The performance of the principal component model is also included in this figure. In the boxplots shown in Figure 7, the central mark on each box indicates the median, and the bottom and top edges of the box indicate the 25th and 75th percentiles, respectively. The whiskers extend to the most extreme data points not considered outliers, and the outliers are plotted individually using the ‘+’ marker symbol.
In this figure, the first five boxes show the TP rates on the test data with IF models with various contamination factors (C) and trees (T). An increase in the contamination factor leads to an increase in the performance of the isolation forests, but at the expense of higher FP rates.
The box on the far right shows the performance of a principal component model with two components, which could explain 78% of the variance of the training data. Although the median TP rate of this model appears to be lower than that of the isolation forests with contamination factors of C = 5%, the overlapping notches suggest that these models may have performed statistically similarly. In addition, it is clear that the IF model with 1000 trees performed more consistently than the IF models with fewer trees (50 and 200), as well as the principal component model.

4. Case Study 2: Platinum Metal Group Flotation Froths

In the second case study, operational data from an industrial flotation plant are considered. A wealth of process information can be inferred from the appearance of the froth phase of flotation systems, which are routinely monitored in many industrial plants.

4.1. Operational Data

The small data set consisted of 297 samples of five gray-level co-occurrence matrix features that were extracted from images of platinum-group metal froths from a South African concentrator circuit. These studies were conducted in the early 1990s, and in this particular case, the original images and metadata associated with the images have not been retained. Nonetheless, the data serve as a practical real-world case study, given that the features were derived from froth structures that were classified by an experienced plant operator at the time.
Projections of the samples with a principal component and linear discriminant model are shown in Figure 8. The first two principal components could explain approximately 93.9% of the variance of the data. The froth conditions associated with normal operating conditions (NOCs) are shown as black circles, while the out-of-control conditions are shown as white circles. The linear discriminant projection was manipulated to project the data into three groups, namely the NOC data and the NEW data, which consisted of two natural clusters. As indicated by the linear discriminant score plot, which by design projects the data onto a plane to maximize the separability of the groups, the data are linearly separable to a high degree.
The data that were presented to the models consisted of a training data set consisting of a random selection of 89 samples of the NOC data shown in Figure 8. The test data consisted of a random selection of 10 samples from the NOC data (not present in the training data set), as well as all of the data indicated as NEW in Figure 8. This gave a total of 208 samples in the test data set. Multiple runs were used to construct models based on different training and test data sets in each run, given the random selection of NOC data.
Figure 9 shows typical results from each of the two models. In each figure, the markers with magenta outlines indicate training data, while markers with black outlines indicate test data. Empty magenta markers indicate data that were assigned as OOC data by the models during training. Likewise, empty black markers indicate data that were assigned as OOC data by the models during testing. As can be seen from these graphs, both models could identify almost all of the NEW data in the cluster, separate from the NOC data, as out-of-control. The statistical performance of the models is discussed in more detail in the following sections.

4.2. Isolation Forest

As before, Table 3 provides a summary of the performance of the two models. In this case, the PC model achieved a TP rate of 65.0% on average on the test data, with a FP rate of 2%, as compared with the 53.0% and 6.0% rates of the IF model.
Figure 10 shows the performance of the IF model with different contamination factor values. As these C values increase, the true positive rates of the IF models also improve. With a C-value of 0.05, it shows a true positive rate on the test set with a median value of 53.5%. This increases to more than 60% and more than 70% at C-values of 0.10 and 0.20, but also with a concomitant increase in the false positive rates on the test data, which increases from approximately 5% to 10% and close to 30% at respective contamination factors of 5%, 10%, and 20%.

4.3. Multivariate Statistical Process Monitoring Based on Principal Component Analysis

The Hotelling’s T2 and Q score plots of the principal component model are shown in Figure 11, where the NOC data are indicated by blue circles and the NEW data by red squares. The 95% and 99% control limits on the graphs are indicated by the coarse and finely dashed horizontal lines on the graphs. Data exceeding the 95% control limits were considered to be out-of-control. This includes approximately 5% of the NOC data (blue markers with indices from 1 to 89).
As can be seen from Figure 11, the cluster of new data separate from the NOC data in Figure 9 (samples 199 to 297) are easily detected by the Q-statistic of the model since these data are far from the principal component plane fitted to the NOC data.
Hotelling’s T2 and SPE statistics could collectively explain 68.7% of the new data as out-of-control, with a false positive rate of <5%. This is significantly better than what could be obtained with the isolation forest, which had similar false-positive rates of less than 5% on the test data.

5. Case Study 3: Monitoring of a SAG Circuit

5.1. Operational Data

The scaled operational data, shown in Figure 12, consisted of 6000 samples spanning 13 operational variables that characterized the steady-state operation of an industrial semi-autogenous grinding circuit, as discussed by Groenewald et al. [36]. In this case study, the NOC data (indices 1–3000 in Figure 12) and NEW data (indices 3001–6000 in Figure 12) were collected over different periods of circuit operation. Visualization of the data with a t-SNE and principal component score plot, as shown in Figure 13, shows that the data have different highly separable distributions. The principal component score plot could account for 64.8% of the variance of the data. The t-SNE plot was constructed with a perplexity parameter of 50.
As a further assessment of the separability of the designated NOC and NEW data sets, a classification tree was constructed on a training data set consisting of 2500 samples randomly selected from both classes. The tree was able to discriminate with 99.4% accuracy between the NOC and NEW data in an independent test set consisting of the remainder of the NOC and NEW data.
As an aside, it should be noted that in practice, data would likely be selected over a continuous period where the process is in-control. It is also assumed that the data would be suitably prepared for the construction of models by dealing with missing data, outliers, and other issues related to data quality separately. In addition, once the models used for process monitoring have been constructed, they would typically be used to monitor individual samples collected in real-time. Recalibration of the model would be required when operational conditions change as a result of changes in the ore feed composition or other operational procedures.
The data on which the models were trained consisted of 2900 observations sampled randomly from the NOC data set. The test data set consisted of all the samples in the NEW data as well as 100 randomly selected samples from the NOC data, i.e., the data in the NOC data not included in the training data set. As before, these data could be used to assess the generation of false positives by the models.
As before, models were run multiple times, which included the construction of new training and test sets for each run. Examples of the performance of the IF and PC models are shown in Figure 14 and discussed in more detail below.

5.2. Isolation Forest

Figure 15 shows the performance of the IF model based on ten runs. The box plot on the left in the figure shows the false positive rate generated by the model on the NOC data, which was 5% by design. The box in the middle shows the false positive rate on the test data, which averaged 4.4%. The box on the right shows the true positive rate, which averaged 76.3%, as indicated in Table 4.
As before, an increase in contamination factor leads to a significant increase in the TP rate of the model (from approximately 75% to 85% with an increase in C from 0.05 to 0.40). However, this comes at a commensurate increase in the FP rate of the model, closely in line with the contamination factor itself, i.e., increasing from approximately 5% to 40%.

5.3. Multivariate Statistical Process Monitoring Based on Principal Component Analysis

Unlike in Case Study 2, the IF model performed markedly better than the principal component model, as Hotelling’s T2 and Q-statistics could collectively identify only 18% of the new data as out-of-control. The performance of this model, based on five principal components capturing in excess of 80% of the variance in the data, is shown in Figure 16. The NOC and the new data were not linearly separable, in contrast to the data in the previous case study, which would explain the disparities in the performance of the PC and IF models.

6. Discussion

In this study, isolation forests were used as a basis for monitoring mineral process operations and compared with principal component models that are widely established in the process industries. Isolation forests are machine learning models designed to identify nonlinear relationships in data through recursive partitioning of the data space. Like most machine learning models, they can do so potentially very reliably in conventional applications, provided that the new data are statistically similar to the data that they had been trained on.
In this investigation, they were used to learn representations of NOC data instead, allowing them to extrapolate to test data markedly different from the data that they had been trained on. This is based on the assumption that at least some of the partitions are complementary to the partitions containing the actual training data and that, generally, these would be partitions associated with training data not considered representative of normal operating conditions.
All the models were designed to minimize type I errors or false positive errors during construction on the NOC data. The control limits of the principal component models were fitted to the NOC data, while the isolation forest threshold for anomaly scores was likewise set to identify 5% of the NOC data as anomalies or OOC data. The threshold at which these probabilities were set to indicate OOC data were tuned on the NOC data to yield an approximately 5% error, similar to what was established for the isolation forest and principal component models.
Future work will include more advanced variants of isolation forests as well as an investigation of the efficacy of isolation forests in process monitoring systems when dealing with larger and higher-dimensional data sets.

7. Conclusions

The following conclusions can be drawn from this investigation:
  • In all the case studies used in the analysis, the ability of the models to identify OOC data were assessed, subject to a 5% false-positive rate constraint. The IF models performed better than traditional MSPM models based on principal components in the SAG circuit case study, which showed complex nonlinear dynamics.
  • In the case of the coal flotation data set, the OOC clusters were near-linearly separable from the IC data, and the performance of the PCA and IF models was approximately similar.
  • In the case of the PGM froth features data set, where the OOC clusters were linearly separable from the IC data, the IF performed reasonably well, but not as well as the principal component models.
  • Overall, the IF models were robust, and control of the false positive rate on independent test data could be satisfactorily controlled by setting the contamination factor of the IF to the desired level on the training data.
The results of this study indicate the IFs can be used as a basis for unsupervised monitoring of complex process operations characterized by large sets of operational variables with nonlinear relationships.

Author Contributions

Conceptualization, C.A. and X.L.; methodology, C.A. and X.L.; software, C.A.; validation, C.A.; formal analysis, C.A.; investigation, C.A.; writing—original draft preparation, C.A. and X.L.; writing—review and editing, C.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Data are not available for confidentiality reasons.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

FPFalse positive
ICIn-control
IFIsolation forest
NEWNew operating conditions that may or may not be NOC
NOCNormal operating conditions
OOCOut-of-control
PCPrincipal component
PCAPrincipal component analysis
PGMPlatinum group metals
SPESquared prediction error
TPTrue positive

References

  1. Jemwa, G.T.; Aldrich, C. Kernel-based fault diagnosis on mineral processing plants. Miner. Eng. 2006, 19, 1149–1162. [Google Scholar] [CrossRef]
  2. Gomez-Flores, A.; Heyes, G.W.; Ilyas, S.; Kim, H. Prediction of grade and recovery in flotation from physicochemical and operational aspects using machine learning models. Miner. Eng. 2022, 183, 107627. [Google Scholar] [CrossRef]
  3. Cook, R.; Monyake, K.C.; Hayat, M.B.; Kumar, A.; Alagha, L. Prediction of flotation efficiency of metal sulfides using an original hybrid machine learning model. Eng. Rep. 2020, 2, e12167. [Google Scholar] [CrossRef]
  4. Al-Wahaibi, S.S.S.; Abiola, S.; Chowdhury, M.A.; Lu, Q. Improving convolutional neural networks for fault diagnosis in chemical processes by incorporating global correlations. Comput. Chem. Eng. 2023, 176, 108289. [Google Scholar] [CrossRef]
  5. He, Y.; Yang, Z.; Wang, D.; Gou, C.; Li, C.; Guo, Y. A fault diagnosis method for complex chemical process based on multi-model fusion. Chem. Eng. Res. Des. 2022, 184, 662–677. [Google Scholar] [CrossRef]
  6. Wu, D.; Zhao, J. Process topology convolutional network model for chemical process fault diagnosis. Process Saf. Environ. Prot. 2021, 150, 93–109. [Google Scholar] [CrossRef]
  7. Strydom, J.J.; Miskin, J.J.; McCoy, J.T.; Auret, L.; Dorfling, C. Fault diagnosis and economic performance evaluatio.n for a simulated base metal leaching operation. Miner. Eng. 2018, 123, 128–143. [Google Scholar] [CrossRef]
  8. Wakefield, B.J.; Lindner, B.S.; McCoy, J.T.; Auret, L. Monitoring of a simulated milling circuit: Fault diagnosis and economic impact. Miner. Eng. 2018, 120, 132–151. [Google Scholar] [CrossRef]
  9. Lindner, B.; Auret, L. Application of data-based process topology and feature extraction for fault diagnosis of an industrial platinum group metals concentrator plant. IFAC-PapersOnLine 2015, 48, 102–107. [Google Scholar] [CrossRef]
  10. Rantala, A.; Virtanen, H.; Saloheimo, K.; Jämsä-Jounela, S.-L. Using principal component analysis and self-organizing map to estimate the physical quality of cathode copper. IFAC Proc. Vol. 2000, 33, 357–362. [Google Scholar] [CrossRef]
  11. Le, T.M.K.; Mäkelä, M.; Schreithofer, N.; Dahl, O. A multivariate approach for evaluation and monitoring of water quality in mining and minerals processing industry. Miner. Eng. 2020, 157, 106582. [Google Scholar] [CrossRef]
  12. Aldrich, C.; Auret, L. Unsupervised Process Monitoring and Fault Diagnosis with Machine Learning Methods; Series: Advances in Pattern Recognition; Springer: London, UK, 2013; ISBN 978-1-4471-5184-5. Available online: http://www.springer.com/computer/ai/book/978-1-4471-5184-5 (accessed on 30 June 2013).
  13. Ge, Z. Review on data-driven modeling and monitoring for plant-wide industrial processes. Chemom. Intell. Lab. Syst. 2017, 171, 16–25. [Google Scholar] [CrossRef]
  14. Pilario, K.E.; Shafiee, M.; Cao, Y.; Lao, L.; Yang, S.-H. A review of kernel methods for feature extraction in nonlinear process monitoring. Processes 2020, 8, 24. [Google Scholar] [CrossRef]
  15. Pani, A.K. Non-linear process monitoring using kernel principal component analysis: A review of the basic and modified techniques with industrial applications. Braz. J. Chem. Eng. 2022, 39, 327–344. [Google Scholar] [CrossRef]
  16. Sukchotrat, T.; Kim, S.B.; Tsung, F. One-class classification-based control charts for multivariate process monitoring. IIE Trans. 2009, 42, 107–120. [Google Scholar] [CrossRef]
  17. Choi, W.S.; Lee, C.; Lee, J.-M.; Park, J.H.; Lee, I.-B. Fault detection and identification of nonlinear processes based on kernel PCA. Chemom. Intell. Lab. Syst. 2005, 75, 55–67. [Google Scholar] [CrossRef]
  18. Fezai, R.; Mansouri, M.; Taouali, O.; Harkat, M.F.; Bouguila, N. Online reduced kernel principal component analysis for process monitoring. J. Process Control. 2018, 61, 1–11. [Google Scholar] [CrossRef]
  19. Krishnannair, S.; Aldrich, C.; Jemwa, G.T. Detecting faults in process systems with singular spectrum analysis. Chem. Eng. Res. Des. 2016, 113, 151–168. [Google Scholar] [CrossRef]
  20. Nawaz, M.; Maulud, A.S.; Zabiri, H. Analysis of multiscale process monitoring in industrial processes from a bibliometric perspective. Comput. Chem. Eng. 2022, 168, 108055. [Google Scholar] [CrossRef]
  21. Yan, W.; Guo, P.; Gong, L.; Li, Z. Nonlinear and robust statistical process monitoring based on variant autoencoders. Chemom. Intell. Lab. Syst. 2016, 158, 31–40. [Google Scholar] [CrossRef]
  22. Yang, J.; Wang, L. Nonlocal, local and global preserving stacked autoencoder based fault detection method for nonlinear process monitoring. Chemom. Intell. Lab. Syst. 2023, 235, 104758. [Google Scholar] [CrossRef]
  23. Chabchoub, Y.; Togbe, M.U.; Boly, A.; Chiky, R. An in-depth study and improvement of Isolation Forest. IEEE Access 2022, 10, 10219–10237. [Google Scholar] [CrossRef]
  24. Lesouple, J.; Baudoin, C.; Spigai, M.; Tourneret, J.-Y. Generalized isolation forest for anomaly detection. Pattern Recognit. Lett. 2021, 149, 109–119. [Google Scholar] [CrossRef]
  25. Liu, F.T.; Ting, K.M.; Zhou, Z.-H. Isolation Forest. In Proceedings of the Eighth IEEE International Conference on Data Mining, Pisa, Italy, 15–19 December 2008; pp. 413–422. [Google Scholar] [CrossRef]
  26. Chen, Y.; Wu, W. Isolation forest as an alternative data-driven mineral prospectivity mapping method with a higher data-processing efficiency. Nat. Resour. Res. 2019, 28, 31–46. [Google Scholar] [CrossRef]
  27. Chen, Y.; Zhao, Q.; Lu, L. Combining the outputs of various k-nearest neighbor anomaly detectors to form a robust ensemble model for high-dimensional geochemical anomaly detection. J. Geochem. Explor. 2021, 231, 106875. [Google Scholar] [CrossRef]
  28. Mukherjee, J. A study on automated detection of surface and sub-surface coal seam fires using isolation forest from landsat 8 OLI/TIRS images. In Proceedings of the IGARSS 2022-2022 IEEE International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 17–22 July 2022; pp. 5512–5515. [Google Scholar]
  29. Zhang, J.; Tang, Z.; Xie, Y.; Ai, M.; Gui, W. Flotation fault diagnosis method using statistical approaches. In Proceedings of the 2019 7th International Conference on Advanced Cloud and Big Data, CBD 2019, Suzhou, China, 21–22 September 2019; Article number 8916511. pp. 266–271. [Google Scholar] [CrossRef]
  30. Zhang, J.; Tang, Z.; Xie, Y.; Ai, M.; Gui, W. Visual perception-based fault diagnosis in froth flotation using statistical approaches. Tsinghua Sci. Technol. 2021, 26, 172–184. [Google Scholar] [CrossRef]
  31. Liu, X.; Aldrich, C. Monitoring of a platinum group metal flotation plant with an isolation forest. In Proceedings of the Australian and New Zealand Control Conference (ANZCC 2022), Gold Coast, QLD, Australia, 24–25 November 2022; pp. 93–97. [Google Scholar] [CrossRef]
  32. Santis, R.; Costa, M. Extended isolation forests for fault detection in small hydroelectric plants. Sustainability 2020, 12, 6421. [Google Scholar] [CrossRef]
  33. Wang, R.; Chen, H.; Guan, C.; Gong, W.; Zhang, Z. Research on the fault monitoring method of marine diesel engines based on the manifold learning and isolation forest. Appl. Ocean. Res. 2021, 112, 102681. [Google Scholar] [CrossRef]
  34. He, S.; Chen, F.; Jiang, B. Physical intrusion monitoring via local-global network and deep isolation forest based on heterogeneous signals. Neurocomputing 2021, 441, 25–35. [Google Scholar] [CrossRef]
  35. MacGregor, J.F.; Kourti, T. Statistical process control of multivariate processes. Control. Eng. Pract. 1995, 3, 403–414. [Google Scholar] [CrossRef]
  36. Groenewald JD, V.; Coetzer, L.P.; Aldrich, C. Statistical monitoring of a grinding circuit: An industrial case study. Miner. Eng. 2006, 19, 1138–1148. [Google Scholar] [CrossRef]
Figure 1. Supervised monitoring (left) and unsupervised monitoring (right). In the case of supervised monitoring, the model would not be able to identify ‘Fault 3’, as it is not in its training base.
Figure 1. Supervised monitoring (left) and unsupervised monitoring (right). In the case of supervised monitoring, the model would not be able to identify ‘Fault 3’, as it is not in its training base.
Minerals 14 00076 g001
Figure 2. Anomaly detection with an isolation forest.
Figure 2. Anomaly detection with an isolation forest.
Minerals 14 00076 g002
Figure 3. Application of an isolation forest in process monitoring.
Figure 3. Application of an isolation forest in process monitoring.
Minerals 14 00076 g003
Figure 4. Process monitoring charts based on Hotelling’s T2 and Q statistics derived from principal component models. Data are considered out-of-control (purple markers), if any one of the control limits (dashed horizontal lines) on the charts are exceeded, otherwise the process is in-control (blue markers).
Figure 4. Process monitoring charts based on Hotelling’s T2 and Q statistics derived from principal component models. Data are considered out-of-control (purple markers), if any one of the control limits (dashed horizontal lines) on the charts are exceeded, otherwise the process is in-control (blue markers).
Minerals 14 00076 g004
Figure 5. Normalized measurements of the coal conditioning and flotation variables used in Case Study 1.
Figure 5. Normalized measurements of the coal conditioning and flotation variables used in Case Study 1.
Minerals 14 00076 g005
Figure 6. Typical performance of the isolation forest on the test data shown in Figure 5. Training data are indicated by magenta markers, and test data are indicated by black markers. Solid markers indicate samples correctly identified by the model as either IC or OOC.
Figure 6. Typical performance of the isolation forest on the test data shown in Figure 5. Training data are indicated by magenta markers, and test data are indicated by black markers. Solid markers indicate samples correctly identified by the model as either IC or OOC.
Minerals 14 00076 g006
Figure 7. Performance of IF and PCA models in Case Study 1.
Figure 7. Performance of IF and PCA models in Case Study 1.
Minerals 14 00076 g007
Figure 8. Principal component score plot of the flotation froth features in Case Study 2 (left) and a linear discriminant projection of the same data (right). Black circles and empty circles represent NOC and NEW data, respectively.
Figure 8. Principal component score plot of the flotation froth features in Case Study 2 (left) and a linear discriminant projection of the same data (right). Black circles and empty circles represent NOC and NEW data, respectively.
Minerals 14 00076 g008
Figure 9. Visualization of the performance of the isolation forest (left) and the principal component model (right). Magenta colored circles indicate training (NOC) data, black circles indicate new data to be tested, while open and solid circles indicate the model’s assignment of the data as OOC or IC respectively.
Figure 9. Visualization of the performance of the isolation forest (left) and the principal component model (right). Magenta colored circles indicate training (NOC) data, black circles indicate new data to be tested, while open and solid circles indicate the model’s assignment of the data as OOC or IC respectively.
Minerals 14 00076 g009
Figure 10. Performance of IF models with different contamination factors (C). The left, middle, and right boxes in each panel show the false positive rates on the training data, the false positive rates on the test data, and the true positive rates on the test data, respectively. On each box, median values are indicated by red lines, the edges of the box indicate the 25th and 75th percentiles, respectively. Whiskers extend to extreme data points not considered outliers, which are indicated by a ‘+’ marker symbol.
Figure 10. Performance of IF models with different contamination factors (C). The left, middle, and right boxes in each panel show the false positive rates on the training data, the false positive rates on the test data, and the true positive rates on the test data, respectively. On each box, median values are indicated by red lines, the edges of the box indicate the 25th and 75th percentiles, respectively. Whiskers extend to extreme data points not considered outliers, which are indicated by a ‘+’ marker symbol.
Minerals 14 00076 g010
Figure 11. Hotelling’s T2 and SPE (Q) plots of the MSPM model. The blue markers represent the NOC data in Case Study 2, while the red markers are the new data. The coarse and fine dashed lines are the 95% and 99% control limits of the statistics. Data above these limits are considered out-of-control.
Figure 11. Hotelling’s T2 and SPE (Q) plots of the MSPM model. The blue markers represent the NOC data in Case Study 2, while the red markers are the new data. The coarse and fine dashed lines are the 95% and 99% control limits of the statistics. Data above these limits are considered out-of-control.
Minerals 14 00076 g011
Figure 12. Normalized grinding circuit variables showing NOC data in blue (index 1–3000) and NEW data in red (index (3001–6000).
Figure 12. Normalized grinding circuit variables showing NOC data in blue (index 1–3000) and NEW data in red (index (3001–6000).
Minerals 14 00076 g012
Figure 13. Visualization of the NOC (black ‘o’ markers) and NEW data (white ‘o’ markers) representing the operation of a grinding circuit shown in Figure 12 with a principal component score plot (left) and a t-SNE score plot (right).
Figure 13. Visualization of the NOC (black ‘o’ markers) and NEW data (white ‘o’ markers) representing the operation of a grinding circuit shown in Figure 12 with a principal component score plot (left) and a t-SNE score plot (right).
Minerals 14 00076 g013
Figure 14. Visualization of the performance of the IF model (left) and PC model (right) in the principal component score space of the data. Magenta markers indicate NOC data on which the models were trained, and black markers indicate NEW data on which the models were tested. Solid and empty markers indicate model predictions of in-control (IC) and out-of-control (OOC) data.
Figure 14. Visualization of the performance of the IF model (left) and PC model (right) in the principal component score space of the data. Magenta markers indicate NOC data on which the models were trained, and black markers indicate NEW data on which the models were tested. Solid and empty markers indicate model predictions of in-control (IC) and out-of-control (OOC) data.
Minerals 14 00076 g014
Figure 15. Performance of isolation forest models trained on the SAG circuit data with different contamination factors (C) ranging from 0.05, 0.10, 0.20, and 0.40. Inside the panels, the boxes from left to right indicate false positives on the training data, false positives on the test data, and true positives on the test data.
Figure 15. Performance of isolation forest models trained on the SAG circuit data with different contamination factors (C) ranging from 0.05, 0.10, 0.20, and 0.40. Inside the panels, the boxes from left to right indicate false positives on the training data, false positives on the test data, and true positives on the test data.
Minerals 14 00076 g015
Figure 16. Hotelling’s T2 and SPE (Q) plots of the principal component model for the grinding circuit data in Case Study 3.
Figure 16. Hotelling’s T2 and SPE (Q) plots of the principal component model for the grinding circuit data in Case Study 3.
Minerals 14 00076 g016
Table 1. Hyperparameter settings of the isolation forests used in this investigation.
Table 1. Hyperparameter settings of the isolation forests used in this investigation.
ParameterDescriptionValue
TNumber of trees in the isolation forest200
NSSubsample size, no replacement min   ( N , 256 )
LMaximum tree depth, L = min ( l o g 2 ( N s ) 8
MSFeature sample size M
CContamination factor5%
Table 2. Performance of IF and PC models with the coal flotation data in Case Study 1.
Table 2. Performance of IF and PC models with the coal flotation data in Case Study 1.
Model/
Statistic
IF (C = 0.05, T = 1000)PCA (2 Components)
FP-TestTP-TestFP-TestTP-Test
Average6.032.01.928.0
Median6.231.81.530.3
StDev0.471.401.810.2
Minimum4.630.00.02.00
Maximum6.235.07.745.0
No of Runs20202020
Table 3. Performance of IF and PC models with the PGM data in Case Study 2.
Table 3. Performance of IF and PC models with the PGM data in Case Study 2.
Model/
Statistic
Isolation Forest (C = 0.05)PC Model (2 Components)
FP-TrainFP-TestTP-TestFP-TrainFP-TestTP-Test
Average4.56.053.052.065.0
Median4.5053.55064.4
StDev08.42.604.71.43
Minimum4.5049.05063.9
Maximum4.52057.151067.3
No of Runs10101051010
Table 4. Performance of the IF and PC models with the SAG circuit data in Case Study 3.
Table 4. Performance of the IF and PC models with the SAG circuit data in Case Study 3.
Model/
Statistic
Isolation Forest (C = 0.05)PC Model (5 Components)
FP-TrainFP-TestTP-TestFP-TrainFP-TestTP-Test
Average5.04.476.354.218.3
Median5.04.076.254.018.4
StDev02.70.300.840.18
Minimum4.92.076.153.018.1
Maximum5.09.076.755.018.5
No of Runs555555
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Aldrich, C.; Liu, X. Monitoring of Mineral Processing Operations with Isolation Forests. Minerals 2024, 14, 76. https://doi.org/10.3390/min14010076

AMA Style

Aldrich C, Liu X. Monitoring of Mineral Processing Operations with Isolation Forests. Minerals. 2024; 14(1):76. https://doi.org/10.3390/min14010076

Chicago/Turabian Style

Aldrich, Chris, and Xiu Liu. 2024. "Monitoring of Mineral Processing Operations with Isolation Forests" Minerals 14, no. 1: 76. https://doi.org/10.3390/min14010076

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop