Abstract
In the hot strip rolling process, many process parameters are related to the quality of the final products. Sometimes, the process parameters corresponding to different steel grades are close to, or even overlap, each other. In reality, locating overlap regions and detecting products with abnormal quality are crucial, yet challenging. To address this challenge, in this work, a novel method named kernel entropy component analysis (KECA)-weighted cosine distance is introduced for fault detection and overlap region locating. First, KECA is used to cluster the training samples of multiple steel grades, and the samples with incorrect classes are seen as the boundary of the sample distribution. Next, the concepts of recursive-based regional center and weighted cosine distance are introduced. For each steel grade, the regional center and the weight coefficients are determined. Finally, the weighted cosine distance between the testing sample and the regional center is chosen as the index to judge abnormal batches. The samples in the overlap region of multiple steel grades need to be focused on in the real production process, which is conducive to quality grade and combined production. The weighted cosine distances between the testing sample and different regional centers are used to locate the overlap region. A dataset from a hot steel rolling process is used to evaluate the performance of the proposed methods.
1. Introduction
With the constant increase of market demands for high-quality products, process monitoring and state evaluation technology become more and more attractive [1,2,3]. It is of great economic value to monitor the production process in real-time. In order to meet the different requirements in various applications, the demand for customized production has been augmented and the product types have been refined. The industrial production process gradually translates from mass production to high-mix, low-volume production.
Mechanical properties are important indicators of the final quality of hot rolled strips [4], which are affected by the chemical composition, temperature, and thickness of the hot strip, and many other important process parameters [5,6]. However, it is difficult to establish an accurate mechanical properties model since it is a highly automated and complex production process. To reduce the cost, people usually test a small number of products via sampling. Meanwhile, the inspection experiment is taken every several hours so that the result of product quality can be obtained in a timely manner. Thus, Wang established a novel neural network (NN) called PSO-BP NN, which combined the particle swarm optimization (PSO) and the back propagation (BP) rapid training algorithm to predict the mechanical properties [7]. Zhao introduces a semi-parametric single-index model to predict the mechanical properties of hot rolled strip [8]. It is a practical industrial problem that has to be solved urgently. Furthermore, with continuous maturity of production technology, the rate of unqualified products drops sharply. From a large amount of historical production data, it is concluded that there is sometimes an overlap in the process parameters of different steel grades, which is of great significance for quality control. In a real production process, the phenomenon of steel grade change and merger production is frequent, and it is particularly important to make rational use of this overlap. Therefore, how to judge the quality of products correctly and control the final quality according to the quality-related process parameters become important research topics.
In this paper, a clustering algorithm is employed to classify the datasets from different steel types, in which the data with incorrect classes are seen as the hypersurface since their process parameters deviate from the theoretical values. Kernel entropy component analysis (KECA) is first used to extract nonlinear characteristics via a kernel function [9,10,11]. Next, a distinct angular statistics method, named the Cauchy–Schwarz(CS) statistic, based on KECA is used to calculate the cosine value between batches [12]. In order to evaluate the mechanical properties of hot rolled steel offline more efficiently, KECA is used to extract features, and the kernel parameter and cluster number were selected adaptively to conduct the production state clustering [13]. KECA extracts features via Renyi entropy, and can achieve a better feature extraction performance than kernel principal component analysis (KPCA) [14]. However, the above work only focuses on the fault detection of one steel grade while the overlap region is not taken into consideration. Another entropy-based clustering approach is cross-entropy clustering [15]. The improvement of multi-objective optimization based on cross-entropy method is introduced in [16,17]. To enhance the cluster analysis, an information-theoretic-based method for cluster visualization using self-organizing maps (SOMs) is presented [18].
Support vector data description (SVDD) is derived from the support vector machine (SVM) and has been widely used in process monitoring. It focuses on determining a hypersphere with minimum volume to describe the characteristics of target classes [19]. In recent years, some researchers have applied SVDD in batch process monitoring [20,21,22]. Meanwhile, it is also widely applied in data classification [23,24,25] and outlier detection [26,27]. The SVDD method builds a hypersphere based on the training dataset to cover as many datasets as possible and minimize its volume. Similarly, a hypersurface is built in a high-dimensional space to distinguish normal and abnormal batches, where the batches inside the hypersurface are seen as normal. In this work, we consider not only the abnormal samples but also the overlap region of multiple steel grades. This paper introduces the recursive-based regional center, which is defined as a region that the weighted cosine distance from it to the boundary of the data distribution along different directions is a constant. Obviously, the constant is the radius of the hypersurface. The cosine distance is the index to distinguish normal and abnormal batches. First, the information of the data distribution is determined via the KECA algorithm. The training datasets are divided into two parts according to the data distribution, where the first part is on the boundary and the other part is inside the boundary. Then, the recursive-based regional center is determined. Finally, the test datasets are detected via the cosine distance. In order to control the quality of steel better, it is worth paying attention to the overlap of the process parameters of multiple steel grades. The weighted cosine distances between the test dataset and different regional centers are also used to judge which dataset is in the overlap.
The rest of this paper is organized as follows: Section 2 introduces the KECA clustering algorithm. In Section 3, we present a novel method named the KECA-weighted cosine distance for abnormal sample detection and overlap region locating. Section 4 validates the performance of our proposed method using numerical studies. Finally, Section 5 concludes the paper.
2. Kernel Entropy Component Analysis
2.1. Kernel Entropy Component Analysis
Consider a dataset with the probability density function . The Renyi quadratic entropy is defined as [28]:
and the Parzen window density estimator is described as:
where is a Mercer kernel function, and σ is the parameter of the kernel function, or Parzen window. Using the sample mean approximation of the expectation operator, we can express the Renyi quadratic entropy as follows:
which can be re-expressed as follows:
where K is the kernel matrix and I is an (n × 1) vector where each element equals one. The kernel matrix can be decomposed as using eigen decomposition, where is a diagonal matrix storing the eigenvalues and E is a matrix with the eigenvectors as columns. Therefore, the Renyi entropy estimator in Equation (3) can be expressed using the eigenvalues and eigenvectors:
Equation (5) demonstrates that each eigenvalue coupled with its corresponding eigenvector make a different contribution to the Renyi quadratic entropy. In this paper, the t eigenvalues that make the first t largest contributions to the Renyi quadratic entropy will be selected, which make the contributions greater than 85%. This yields a component matrix .
2.2. Cauchy–Schwarz (CS) Statistic
There is a distinct angular structure in the feature space of KECA, in which different clusters are usually located in different angular directions. Thus, it is necessary to calculate proper statistics on the feature space. To express the angular structure of a dataset, the CS divergence measurement between the probability density functions, which corresponds to the cosine of the angle between kernel feature space mean vectors, has been derived. Then, the clustering steps based on KECA are given as follows:
- (1)
- Obtain by KECA, and rank the eigenvectors ;
- (2)
- Initialize means , is the cluster number;
- (3)
- For all ;
- (4)
- Update mean vectors;
- (5)
- Repeat steps (3) and (4) until convergence.
In summary, we assign a data point to the cluster represented by the closest mean vector in terms of angular distance. In step 2, we initialize the mean vectors such that the two data points and , having the property that , are selected to represent and . Furthermore, is represented by , such that . Here, is the kernel feature space cluster mean vector corresponding to , and is the overall kernel space mean vector of the data. Based on our experience, this algorithm is robust to the initialization and usually converges within a few iterations. Convergence is measured by the change in a CS cost function from one iteration to the next [29].
The CS divergence measures the cosine of the angle between these mean vectors. An angle-based clustering cost function in terms of the kernel feature space dataset may be defined as:
3. Process Parameters Clustering via Weighted Cosine Distance
From the quality control perspective, we are interested in identifying the abnormal batches and overlap regions. To achieve this goal, we employ KECA for multiple steel grades clustering. KECA provides the information of the data distribution, which can be used for locating the overlap region. The weighted cosine distance is a novel distance measurement in the high-dimensional space and can be used as a metric for abnormal sample monitoring and locating the overlap region.
3.1. Weighted Cosine Distance
In this section, a weighted cosine distance is proposed to measure the distance in the high dimensional space between a sample and other samples.
Consider the datasets with samples and variables. The CS statistic matrix is a symmetric matrix, in which each element is the cosine value between the m-th sample and the n-th sample. Obviously, the cosine values range from –1 to 1. The closer the cosine values is to 1, the shorter the distance. In order to compare the distance conveniently, the cosine distance is defined as follow:
where, is the element of the cosine distance matrix and it represents the distance between the m-th sample and the n-th sample. The range of cosine distance is enlarged via the square operation. Obviously, the cosine distance ranges from 0 to 4.
Sometimes, many samples need to be considered as a whole. The new distance model named the One–Many cosine distance is introduced. Consider one sample and a dataset , where the dataset contains samples. For each samples in the dataset , we calculate its cosine value with sample . As a result, a vector is obtained. The One-Many cosine distance is defined as
In fact, the role of the samples in the datasets may be different. Thus, when calculating the One–Many cosine distance, the weights of different samples should be considered instead of the average. Assume a weight coefficient vector is given, where is the weight coefficient of the i-th sample, the weighted cosine distance is defined as follows:
3.2. Abnormal Samples Detection
The samples are projected into high-dimensional feature space via the kernel function. The hypersphere is built in the high dimensional feature space. If the sample is inside the hypersphere, it is seen as normal, otherwise it is seen as abnormal. As shown in Figure 1a, the blue points represent the abnormal samples and the black points represent the normal samples. Assume O(x0,y0) is the center of the hypersphere. If the hypersphere is selected with the radius , some samples are misclassified. For example, sample 1, which is abnormal, is inside the hypersphere, that is, it is misjudged as normal. Another example is sample 2, which is misclassified as abnormal.

Figure 1.
Comparison of two methods. (a) Schematic diagram of the traditional fault detection method; (b) schematic diagram of the recursive based regional center.
Assume that there is such a hypersurface in the high-dimensional feature space that contains all the normal samples, and all the abnormal samples are outside the hypersurface. No doubt, this hypersurface can replace the hypersphere and classify the process state more accurately. In order to overcome the problems caused by the hypersphere, this paper introduces a recursive-based regional center method. It defines as a region in high-dimensional feature space, which has the same distance from it to the hypersurface along different directions. As shown in Figure 1b, the regional center enlarged from the sample O to the regional center that is considered as a whole. The new radius R is defined as the mean of the cosine distances between the samples in the regional center and the hypersurface. Obviously, the choice of the regional center is the key problem. Thus, a recursive method is proposed.
However, it is difficult to determine a qualified region based on the irregular distribution in the high-dimensional feature space. Obviously, different samples in the regional center make different contributions to the distance calculation. Thus, the weight coefficients are selected to quantify the contribution of each sample in distance calculation. It is worth noting that both the radius and the weight coefficients are unknown. If the first one of them is set as a constant value, then we can obtain the other one via a recursive method and update the first one. The computation will loop continuously until convergence. To ensure that the regional center meets its definition, the radius must be a constant that is given firstly, the weight coefficients can be obtained via regression. Thus, the regression process can calculate as follows:
where is the weight coefficients, is the residual. is calculated by Equation (7) where is a vector when the regional center is selected, and all the samples in the center are seen as a whole.
Consider two datasets and with and samples, respectively, and variables, where dataset is the regional center and dataset is distributed on the hypersurface. According to Equation (11), it can be further obtained that:
where are the weight coefficients; is the radius vector whose elements are set as one value . This implies that the distance between the samples on the hypersurface and the regional center is a constant; is the cosine value between the m-th sample in dataset and the n-th in dataset ; is the residual vector. To guarantee the given radius stable, its residual must be stable. Thus, it is necessary to minimize the variance of residuals.
To achieve abnormal samples detection, the recursive-based regional center should be determined firstly. Consider two datasets and with and samples, respectively, and J variables, where dataset is inside the hypersurface and dataset is the boundary of the hypersurface. Calculate the cosine value matrix according to Equation (7) and take the sum of them as follows:
Then, sort the vectors in ascending order. Obviously, the smaller the sum of the cosine values, the closer it is to the regional center. The detailed procedure of the recursive-based regional center is described as Algorithm 1 below.
| Algorithm 1: Recursive-based regional center |
| Step 1. Calculate the sum of the cosine values according to Equation (13); Step 2. Reconstruct the datasets according to the ascending order of the vector ; Step 3. Let , the given radius is selected as ; Step 4. Select regional center , Calculate the residual vector according to Equation (12); Step 5. Let ovar = var(); Step 6. Do while k=k+1; =[;]; nvar = var(), where is the residual in Equation (12); if ovar ≥ nvar ovar = nvar; continue; else remove from ; end |
Last, the matrix is the regional center. The weighted coefficients is obtained according to Equation (12). Thus, the weighted cosine distance between each testing batch and the regional center is calculated. If the distance is larger than the given radius , it means the data sample is outside the hypersurface and is seen as abnormal. Otherwise, it is normal.
3.3. Locating the Overlap Region
As is known, many steel grades have similar mechanical properties due to their close process parameters. Sometimes their process parameters appear in an overlap region. Obviously, it is of great significance to make full use of the overlap region.
The schematic diagram of the data distribution from different steel types is shown in Figure 2, where black points are the datasets, and the lines of the three colors (black, blue, and yellow) represent the hypersurface composed of three steel grades in the high-dimensional space. If the mechanical properties of two steel grades are similar, the process parameters may still overlap in the high-dimensional space, for example, the area formed by the red dotted line. Assume three colors (black, blue, and yellow) represent three steel grades, steel 1, steel 2, and steel 3, respectively. Obviously, steel 1 is far away from the other steel grades, so there will be no overlap region. However, steel 2 is similar to steel 3. Despite the datasets in the overlap region being normal, the process parameters have deviated from their theoretical values. In fact, the datasets in the overlap region are also worth discussing since it can be used for steel grade change and production grade merger.
Figure 2.
Schematic diagram of data distribution for different steel types.
For each steel grade, the regional center is determined based on the recursive method. The weighted cosine distance between the datasets in the overlap region and the regional center of steels 2 and 3 is less than the given radius (the given radius can be the same for different steel grades). This characteristic is the necessary and sufficient condition for judging whether a sample is inside the overlap region.
3.4. Procedure of Process State Clustering
The procedure of the proposed method is shown in Figure 3. In the training process, the training datasets with the size of I×J are obtained, which contain C steel types. The KECA clustering is used to divide the training dataset into C clusters. The cluster number C is the number of steel grades. If the original number of steel grades is unknown, the optimal number of clusters can be determined by a novel cluster validity index (CVI) that is introduced for cluster validation [30]. Then, the dataset of each steel grade is divided into two parts according to the clustering result. The first part with correct clusters is seen as inside the hypersurface, and the second part with incorrect clusters is considered as the hypersurface. Then, we set the given radius as . The recursive-based region center and the weight coefficients are determined. The procedure of process state clustering via the KECA-weighted cosine distance is summarized below.
Figure 3.
Flowchart of the proposed method.
Off-line training process:
- (1)
- Obtain the training samples from historical normal operating conditions, where C represents the number of steel types;
- (2)
- Perform KECA clustering on datasets X, and then C clusters are obtained. For each product , it is divided into two parts, and that are seen as the hypersurface;
- (3)
- Set the given radius as ; the recursive-based regional center and the corresponding weight coefficients are determined;
- (4)
- Set the control limit as .
On-line fault detection:
- (1)
- Acquire the testing sample with the size of 1 × J;
- (2)
- Calculate the weighted cosine distance between the recursive-based regional center and the testing sample;
- (3)
- The weighted distance is compared with the control limit; if it exceeds the control limit, it would be judged as abnormal. Otherwise, it would be considered as normal;
- (4)
- For the normal sample, calculate the weighted cosine distances between it and the regional center of other steel types. If one or more weighted cosine distances are less than the control limit, it is inside the overlap region.
4. Experiments
In this section, a dataset from a hot strip rolling process is employed to demonstrate the effectiveness of the KECA-weighted cosine distance method. The proposed method is compared with the method ClusterSVDD [31] and the detailed results are described.
The hot strip rolling production is a complex industrial process with the purpose of turning reheated steel slabs into strips. The layout drawing of the hot rolling process is shown in Figure 4 [32]. The commonly hot strip rolling line consists of several major pieces of equipment, such as reheat furnaces, a roughing mill, a transfer table, a finishing mill, laminar cooling, and coilers. In the rough rolling process, the thickness of the reheat steel slabs is reduced quickly. To satisfy the requirements of the reduction, final qualities, and tolerances, steel plants execute the rolling process of tandem rolling with six or seven successive stands during the finishing rolling process in which the reheated steel slabs are narrowed into the final thickness via the repeated rolling process.
Figure 4.
Layout drawing of hot rolling process.
Mechanical performance is a significant quality of the final product. There are 11 process variables that are most related to the mechanical performance, and which are collected during the real process, as shown in Table 1.
Table 1.
Variables table of process parameters.
The real production data of three steel grades are used to illustrate the proposed method. A total of 65 samples per steel grades are collected as the training dataset. Thus, there are 195 samples with a size of 195 × 11. For the first steel grade, a total of 79 samples are selected as the testing dataset, which contains 14 fault samples and 65 normal samples. There are 87 samples consisting of 22 fault samples and 65 normal samples, seen as the testing datasets of the second steel grade. The third steel grade contains 81 samples where 16 are fault samples and 65 are normal samples.
Due to the imbalance of classes in the testing dataset, that is, there is a large difference in the proportion of normal and abnormal samples, the detection rate, ROC criteria, and AUC value are employed to evaluate different models. The detection rate is defined in Equation (14). The horizontal axis of the ROC is the false positive rate (FPR) and the vertical axis is the true positive rate (TPR). The area under the curve (AUC) is defined as the area under the ROC curve:
where TP represents the number of normal samples which are judged as normal; FP represents the abnormal samples which are judged as normal; similar, FP is the number of normal samples that are detected as abnormal; and TN is the number of abnormal samples that detected as abnormal.
The Gaussian kernel that is widely used in the field of process monitoring is employed to project the datasets into the high-dimensional feature space. Different parameters are selected to build the best models for each method, such as the given radius , the kernel parameter c ∈ {10–3, 10–2, …, 107}, and so on.
The clustering results via KECA are shown in Table 2. As we know, KPCA and KECA are very similar [12,33] dimension reduction methods. In order to verify the effectiveness of KECA, the clustering results via KPCA are shown in Table 3. Both the best kernel parameters of KECA and KPCA are selected as 102, which range from 103 to 107. Obviously, the results of KECA are better than the results of KPCA. In fact, how to select the features is the most important difference between KPCA and KECA. KPCA is subject to the assumption of a normal distribution of data, which selects the principal components according to the variances. However, industrial data has unpredictable complexity that is very difficult to obey the normal distribution. KECA has a clear theoretical foundation based on the concept of maximum entropy preservation. The novel strategy based on information entropy avoids the distribution assumption of data and chooses the information as the index directly. Therefore, it has better universality in industrial application.
Table 2.
The clustering results via KECA.
Table 3.
The clustering results via KPCA.
According to the clustering results via KECA, the third steel grade is recognized correctly while the first two steel grades have some samples with incorrect results. Obviously, the third steel grade is different from the first two steel grades. To confirm this, the mean of each process parameter for the three steel grades are shown in Table 4. The process parameters, such as the content of P, the entry temperature of the roughing milling and finish milling, the exit temperature of the finish milling, and the coiling temperature are different from other steel grades. Thus, the third steel grade overlap with the first two steel grades.
Table 4.
The mean of each process parameters for three steel grades.
The scatter plot between yield strength–elongation of the first two steel grades is shown in Figure 5. The samples in the blue curve are from steel grade 1; red hollow circles represent the samples with correct clusters, and yellow solid circles are these samples with incorrect clusters. The samples in the red curve are from steel grade 2; blue hollow circles represent the samples with correct clusters and green solid rectangles are those samples with incorrect clusters. In Figure 5, there are 130 batches. A large number of samples are shown in the same position because they have the same values. Although the KECA is performed on the process parameters, the distribution of the multivariable process parameters cannot be shown via visualization methods. Therefore, the clustering results can only be shown by their mechanical properties. For these samples with incorrect clusters, even though their mechanical properties do not appear in the overlap region, their process parameters deviate from their theoretical value, which needs focus in order to improve the quality of steel further. Thus, it is of great significance to focus on the overlap region between different steel grades.
Figure 5.
The yield strength–elongation rate diagram.
The kernel parameter is selected via a grid search and then verified using cross-validation. In the proposed method, the kernel parameter is selected as 102. The given radius is 160. The number of clusters is three, as there are three different steel grades. There are a total of 11 process parameters selected and the number of principal components is five. In ClusterSVDD, the number of principal components is still five. The number of cluster is three and the kernel parameter is selected as 50. The kernel parameter is selected from the range of 10−7–103. Similarly, the given radius is selected by the same way, which is in the range from 1 to 200.
The detection results of the proposed method and the ClusterSVDD are presented in Table 5. For the first steel grade, the detection rate of the KECA-weighted cosine distance is 75.95%, while the detection rate of ClusterSVDD is only 51.90%. Obviously, the computational results show that the weighted cosine distance gives better monitoring performance. The ROC criteria are shown in Figure 6a. The AUC value based on the weighted cosine distance is 0.7308, while the ClusterSVDD returns the AUC value of 0.5287, which demonstrates that the proposed method is more accurate.
Table 5.
Fault detection rates and AUC values of the weighted distance and ClusterSVDD for the hot strip rolling process

Figure 6.
The ROC criteria of two models. (a) The first steel grade; (b) the third steel grade; (c) the third steel grade.
For the second steel grade, the weighted cosine distance returns the highest detection rate of 86.21%, while the result of ClusterSVDD is 51.72%. The ROC criteria are shown in Figure 6b. The KECA-weighted cosine distance gives better performance than ClusterSVDD. Meanwhile, the AUC value of the weighted cosine distance is 0.7559, which is larger than the AUC value of ClusterSVDD.
Similarly, the result of the proposed method for the third steel grade is still better than the result of ClusterSVDD. The ROC criteria are shown in Figure 6c.
To verify the performance of the overlap region locating, a total of 33 samples are selected as the testing dataset. As mentioned above, the first two steel grades have an overlap region. All the weighted cosine distances from these samples to the two regional centers are less than the given radius . The detection rate is 72.73%, that is, there are 24 samples that are detected correctly.
In summary, the results for three steel grades demonstrate that the KECA-weighted cosine distance gives better monitoring performance than ClusterSVDD. Furthermore, the KECA-weighted cosine distance works in this case where different products have similar process parameters.
5. Conclusions
This paper introduces two new concepts of the weighted cosine distance and recursive-based regional center. In addition, a novel method, the KECA-weighted cosine distance, is derived for the abnormal samples detection and overlap region locating. Unlike traditional methods, the proposed method enlarges the center into a region and calculates the weighted cosine distance. The data from a hot strip rolling process is used to evaluate the process monitoring performance. The detection results of the KECA-weighted cosine distance are compared with the results of ClusterSVDD, and the results illustrate that the proposed method gives better monitoring performance. Meanwhile, it can also locate the overlap region of different steel grades well.
Some challenges are worth discussing in future works. For example, how to choose the given radius is worth considering and a meaningful topic. Furthermore, the dataset used to construct the hypersurface plays an important role, and an approach to determine the dataset is also an important topic to investigate in the future.
Author Contributions
Methodology, validation, writing-original draft preparation was performed by C.W.; Conceptualization, writing-review & editing, funding acquisition were provided by F.H.
Funding
This research is supported by the National Key Technology R and D Program of the 12th Five-year Plan of China (grant no. 2015BAF30B01), and the Open Foundation from the State Key Laboratory of Rolling and Automation, Northeastern University (grant no. 2018RALKFKT003), and the USTB-NTUT Joint Research Program (grant no. TW2019013).
Conflicts of Interest
The authors declare no conflict of interest.
References
- Wang, G.; Luo, H.; Peng, K.X. Quality-related fault detection using linear and nonlinear principal component regression. J. Frank. Inst. 2016, 353, 2159–2177. [Google Scholar] [CrossRef]
- Huang, K.K.; Wen, H.F.; Ji, H.Q.; Cen, L.H.; Chen, X.F.; Yang, C.H. Nonlinear process monitoring using kernel dictionary learning with application to aluminum electrolysis process. Control Eng. Pract. 2019, 89, 94–102. [Google Scholar] [CrossRef]
- Liu, M.G.; Li, X.S.; Lou, C.Y.; Jin, J. A Fault Detection Method Based on CPSO-Improved KICA. Entropy 2019, 21, 668. [Google Scholar] [CrossRef]
- Xu, Z.W.; Liu, X.M.; Zhang, K. Mechanical properties prediction for hot rolled alloy steel using convolutional neural network. IEEE Access 2019, 7, 47068–47078. [Google Scholar] [CrossRef]
- Zou, Q.; Chen, L.; Xiong, N.; Zou, S.; Wang, C. Prediction and key computer programming of mechanical properties of hot rolled plate based on BP neural network. In Proceedings of the 2009 International Conference on Computational Science and Engineering, Vancouver, BC, Canada, 29–31 August 2009; pp. 967–971. [Google Scholar]
- Beghini, M.; Bertini, L.; Monelli, B.D.; Santus, C.; Bandini, M. Experimental parameter sensitivity analysis of residual stresses induced by deep rolling on 7075-T6 aluminum alloy. Surf. Coat. Technol. 2014, 254, 175–186. [Google Scholar] [CrossRef]
- Wang, P.; Huang, Z.Y.; Zhang, M.Y.; Zhao, X.W. Mechanical Property Prediction of Strip Model Based on PSO-BP Neural Network. J. Iron Steel Res. Int. 2008, 15, 87–91. [Google Scholar] [CrossRef]
- Zhao, Y.H.; Weng, Y.; Peng, N.Q.; Tang, G.B.; Liu, Z.D. Prediction of Mechanical Properties of Hot Rolled Strip by Using Semi-Parametric Single-Index Model. J. Iron Steel Res. Int. 2013, 20, 9–15. [Google Scholar] [CrossRef]
- Zhou, H.D.; Shi, T.L.; Liao, G.L.; Xuan, J.P.; Duan, J.; Su, L.; He, Z.Z.; Lai, W.X. Weighted Kernel Entropy Component Analysis for Fault Diagnosis of Rolling Bearings. Sensors 2017, 17, 625. [Google Scholar] [CrossRef]
- Qi, Y.S.; Wang, Y.; Lu, C.X.; Wang, L. Improved batch process monitoring and diagnosis based on multiphase KECA. Int. Fed. Autom. Control 2018, 51, 827–832. [Google Scholar] [CrossRef]
- Qi, Y.S.; Meng, X.B.; Lu, C.X.; Gao, X.J.; Wang, L. KECA Similarity-Based Monitoring and Diagnosis of Faults in Multi-Phase Batch Processes. Entropy 2019, 21, 121. [Google Scholar] [CrossRef]
- Jenssen, R. Kernel entropy component analysis. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 847–860. [Google Scholar] [CrossRef] [PubMed]
- He, F.; Xu, J.W.; Liang, Z.G.; Wang, X.C. Hot rolled strip state clustering based on kernel entropy component analysis. J. Cent. South Univ. Sci. Technol. 2012, 43, 1732–1738. [Google Scholar]
- He, F.; Li, M.; Yang, J.H.; Xu, J.W. Adaptive clustering of production state based on kernel entropy component analysis. In Proceedings of the 2010 International Joint Conference on Neural Networks (IJCNN), Barcelona, Spain, 18–23 July 2010. [Google Scholar]
- Tabor, J.; Spurek, P. Cross-entropy clustering. Pattern Recognit. 2014, 47, 3046–3059. [Google Scholar] [CrossRef]
- Beruvides, G.; Quiza, R.; Haber, R.E. Multi-objective optimization based on an improved cross-entropy method. A case study of a micro-scale manufacturing process. Inf. Sci. 2016, 334, 161–173. [Google Scholar] [CrossRef]
- Haber, R.E.; Beruvides, G.; Quiza, R.; Hernandez, A. A simple multi-objective optimization based on the cross-entropy method. IEEE Access 2017, 5, 22272–22281. [Google Scholar] [CrossRef]
- Da Silva, L.E.B.; Wunsch, D.C. An Information-Theoretic-Cluster Visualization for Self-Organizing Maps. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 2595–2613. [Google Scholar] [CrossRef] [PubMed]
- Tax, D.; Duin, R. Support Vector Data Description. Mach. Learn. 2004, 54, 45–66. [Google Scholar] [CrossRef]
- Ge, Z.Q.; Gao, F.R.; Song, Z.H. Batch process monitoring based on support vector data description method. J. Process Control 2011, 21, 949–959. [Google Scholar] [CrossRef]
- Ge, Z.Q.; Song, Z.H. Bagging support vector data description model for batch process monitoring. J. Process Control 2013, 23, 1090–1096. [Google Scholar] [CrossRef]
- Lv, Z.M.; Yan, X.F. Hierarchical Support Vector Data Description for Batch Process Monitoring. Ind. Eng. Chem. Res. 2016, 55, 9205–9214. [Google Scholar] [CrossRef]
- Uslu, F.S.; Binol, H.; Ilarslan, M.; Bal, A. Improving SVDD classification performance on hyperspectral images via correlation based ensemble technique. Opt Lasers Eng. 2017, 89, 169–177. [Google Scholar] [CrossRef]
- Mygdalis, V.; Iosifidis, A.; Tefas, A.; Pitas, I. Semi-supervised subclass support vector data description for image and video classification. Neurocomputing 2018, 278, 51–61. [Google Scholar] [CrossRef]
- Luo, H.; Wang, Y.R.; Cui, J. A SVDD approach of fuzzy classification for analog circuit fault diagnosis with FWT as preprocessor. Expert Syst. Appl. 2011, 38, 10554–10561. [Google Scholar] [CrossRef]
- Azami, M.E.; Lartizien, C.; Canu, S. Converting SVDD scores into probability estimates: Application to outlier detection. Neurocomputing 2017, 268, 64–75. [Google Scholar] [CrossRef]
- Huan, Z.; Wei, C.; Li, G.H. Outlier Detection in Wireless Sensor Networks Using Model Selection-Based Support Vector Data Descriptions. Sensors 2018, 18, 4328. [Google Scholar] [CrossRef]
- Zhang, H.L.; Qi, Y.S.; Wang, L.; Gao, X.J.; Wang, X.C. Fault detection and diagnosis of chemical process using enhanced KECA. Chemom. Intell. Lab. Syst. 2017, 161, 61–69. [Google Scholar] [CrossRef]
- Jenssen, R.; Eltoft, T. A New Information Theoretic Analysis of Sum-of-Squared-Error Kernel Clustering. Neurocomputing 2008, 72, 23–31. [Google Scholar] [CrossRef]
- Guo, G.D.; Chen, L.F.; Ye, Y.F.; Jiang, Q.S. Cluster Validation Method for Determining the Number of Clusters in Categorical Sequences. IEEE Trans. Neural Netw. Learn. Syst. 2017, 28, 2936–2948. [Google Scholar] [CrossRef]
- Görnitz, N.; Lima, L.A.; Müller, K.-R.; Kloft, M.; Nakajima, S. Support Vector Data Descriptions and k-Means Clustering: One Class? IEEE Trans. Neural Netw. Learn. Syst. 2018, 20, 3994–4006. [Google Scholar] [CrossRef]
- Ma, L.; Dong, J.; Peng, K.X. Root cause diagnosis of quality-related faults in industrial multimode processes using robust Gaussian mixture model and transfer entropy. Neurocomputing 2018, 285, 60–73. [Google Scholar] [CrossRef]
- Deng, X.G.; Deng, J.W. Incipient Fault Detection for Chemical Processes Using Two-Dimensional Weighted SLKPCA. Ind. Eng. Chem. Res. 2019, 58, 2280–2295. [Google Scholar] [CrossRef]
© 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).