Next Article in Journal
Effect of Choice of Drilling Kinematic System on Cylindricity Deviation, Roundness Deviation, Diameter Error and Surface Roughness of Holes in Brass Alloy
Next Article in Special Issue
Integrating Improved Coati Optimization Algorithm and Bidirectional Long Short-Term Memory Network for Advanced Fault Warning in Industrial Systems
Previous Article in Journal
Effects of the PMMA Molecular Weight on the Thermal and Thermo-Oxidative Decomposition as the First Chemical Stage of Flaming Ignition
Previous Article in Special Issue
Multi-Objective Disassembly Sequence Planning in Uncertain Industrial Settings: An Enhanced Water Wave Optimization Algorithm
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhancing LightGBM for Industrial Fault Warning: An Innovative Hybrid Algorithm

1
Independent Researchers, No. 76 Chongmingdao East Road, Huangdao District, Qingdao 266000, China
2
Department of Agricultural Economics, Faculty of Agriculture, Ferdowsi University of Mashhad, Mashhad 9177948974, Iran
*
Author to whom correspondence should be addressed.
Processes 2024, 12(1), 221; https://doi.org/10.3390/pr12010221
Submission received: 18 December 2023 / Revised: 5 January 2024 / Accepted: 15 January 2024 / Published: 19 January 2024

Abstract

:
The reliable operation of industrial equipment is imperative for ensuring both safety and enhanced production efficiency. Machine learning technology, particularly the Light Gradient Boosting Machine (LightGBM), has emerged as a valuable tool for achieving effective fault warning in industrial settings. Despite its success, the practical application of LightGBM encounters challenges in diverse scenarios, primarily stemming from the multitude of parameters that are intricate and challenging to ascertain, thus constraining computational efficiency and accuracy. In response to these challenges, we propose a novel innovative hybrid algorithm that integrates an Arithmetic Optimization Algorithm (AOA), Simulated Annealing (SA), and new search strategies. This amalgamation is designed to optimize LightGBM hyperparameters more effectively. Subsequently, we seamlessly integrate this hybrid algorithm with LightGBM to formulate a sophisticated fault warning system. Validation through industrial case studies demonstrates that our proposed algorithm consistently outperforms advanced methods in both prediction accuracy and generalization ability. In a real-world water pump application, the algorithm we proposed achieved a fault warning accuracy rate of 90%. Compared to three advanced algorithms, namely, Improved Social Engineering Optimizer-Backpropagation Network (ISEO-BP), Long Short-Term Memory-Convolutional Neural Network (LSTM-CNN), and Grey Wolf Optimizer-Light Gradient Boosting Machine (GWO-LightGBM), its Root Mean Square Error (RMSE) decreased by 7.14%, 17.84%, and 13.16%, respectively. At the same time, its R-Squared value increased by 2.15%, 7.02%, and 3.73%, respectively. Lastly, the method we proposed also holds a leading position in the success rate of a water pump fault warning. This accomplishment provides robust support for the timely detection of issues, thereby mitigating the risk of production interruptions.

1. Introduction

In the current industrial landscape characterized by a fast-paced and highly automated environment, fault warning has become increasingly crucial [1]. As the complexity of production systems rises, the normal operation of industrial equipment becomes essential for maintaining production efficiency and product quality [2]. However, industrial equipment faces various risks such as aging, external environmental changes, and potential technical issues, increasing the likelihood of faults. Therefore, timely and accurate fault warning is a key factor in ensuring production continuity [3]. By detecting potential issues early and taking corresponding measures, businesses can reduce the risk of production interruptions, save on maintenance costs, and enhance production efficiency. The significance of fault warning systems lies in not only avoiding losses caused by unexpected downtime but also optimizing maintenance schedules for more targeted and efficient repairs.
Against this backdrop, the role of machine learning technology in fault warning practices becomes increasingly prominent [4]. It is noteworthy that among numerous machine learning algorithms, LightGBM stands out for its efficient training speed, capacity to handle large-scale data, excellent prediction accuracy, and customizable features, making it perform exceptionally well in fault warning tasks [5]. However, like many advanced technologies, LightGBM faces challenges in practical applications. Its numerous and challenging-to-determine parameters make adjusting and optimizing the model complex and challenging in different application scenarios [6]. To address this issue, innovative methods and strategies are needed to optimize the performance of the LightGBM model, making it better suited for specific industrial fault warning tasks.
With the rapid development of computer technology, metaheuristic algorithms have become powerful tools for tackling this complex challenge [7]. These algorithms are known for their flexibility and robustness in searching the solution space, effectively exploring the optimal solution by simulating natural systems or other heuristic principles. However, in the field of optimization, under the influence of the “no free lunch” theorem [8], continuous development and improvement of appropriate metaheuristic algorithms become particularly important to ensure good algorithm performance in specific problem domains. Therefore, this paper proposes a novel hybrid algorithm, i.e., simulated annealing-based Arithmetic Optimization Algorithm (SAOA), for optimizing the hyperparameters of LightGBM. The algorithm combines chaotic mapping and dynamic reverse learning strategies to generate excellent initial solutions. Additionally, it adopts a nonlinear inertia weight factor and an improved search strategy to enhance the robustness and accuracy of the AOA algorithm. Finally, the SA is applied for global search, further improving the algorithm’s performance.
To validate the effectiveness of the proposed method, we conduct experiments using real industrial system data. The experimental results demonstrate significant performance advantages of this method in fault warning. Compared to other state-of-the-art methods, this method not only exhibits higher accuracy and reliability but also provides new insights for fault warning research in the industrial domain.
In summary, this paper makes the following significant contributions to the field of fault warning:
  • Introducing a pioneering hybrid algorithm that combines SA and AOA. This marks the first application of such a hybrid algorithm to fault warning tasks, thereby expanding the applicative scope of AOA;
  • Proposing an improved LightGBM model specifically designed for fault warning tasks. In comparison to other advanced algorithms, this tailored model exhibits exceptional performance. Through effective optimization of LightGBM’s hyperparameter selection, the model achieves heightened accuracy and efficiency in predicting faults;
  • Conducting rigorous real-world industrial case validations to substantiate the effectiveness of the proposed methods. The validation results affirm the feasibility and efficacy of our approach in actual industrial settings, underscoring its potential and practical value in the realm of fault warning.
The remainder of the paper is structured as follows: Section 2 is an overview and summary of fault warning. Section 3 defines the problem we are trying to solve and the evaluation metrics used in this paper. Section 4 provides a detailed introduction and distributional elaboration of the proposed algorithm, Section 5 applies a real-world case study for the application of the proposed algorithm, Section 6 compares it with other state-of-the-art algorithms, and Section 7 concludes this paper with the conclusions, limitations, and directions that can be explored in the future.

2. Literature Review

In recent years, the field of fault warning has attracted increasing attention, leading to the emergence of numerous research methods and technological applications [9]. Prior to conducting the literature review, we initially select keywords pertinent to fault warning. These keywords encompass ‘fault warning’, ‘predictive maintenance’, ‘machine learning’, and so forth, which aid us in identifying the most relevant research. Subsequently, when we search for literature, we utilize Boolean operators to optimize the search results. For instance, we employ ‘AND’ to ensure all keywords appear in the search results, such as ‘fault warning AND predictive maintenance’. This enables us to locate research that encompasses both topics. We also use ‘OR’ to broaden the search scope, such as ‘fault warning OR machine learning’. This allows us to find research involving either topic. Upon obtaining the search results, we review the title and abstract of each piece of literature to ascertain their relevance to the research topic. We select the most relevant and recent research for reading and analysis. During the reading of the selected literature, we conduct a thorough evaluation. Subsequently, we incorporate this information into the research to form a comprehensive literature review. Finally, the literature we summarized is as follows: Li et al. [10] proposed a diagnostic method for mild-turn-to-turn short circuit faults in synchronous generator excitation windings. They utilized an improved particle swarm optimization algorithm to determine the structure parameters of the gated recurrent unit—convolutional neural network model and used the total bias distance as the fault detection criterion. The effectiveness of this method is verified through experiments. Additionally, Lu et al. [11] introduced a fault warning method for electric vehicle (EV) charging processes based on an adaptive deep belief network. They employed the Nesterov-accelerated adaptive moment estimation for optimization during training and constructed the adaptive deep belief network model for normal charging processes using historical EV charging data. The method achieved real-time fault prediction and demonstrated high accuracy. In the industrial equipment domain, Gao et al. [12] proposed a dynamic modeling approach, combining wavelet packet decomposition and graph theory for early fault detection and identification. Experimental results demonstrated the effectiveness of this method in engineering applications. Furthermore, Lyu et al. [13] presented a battery fault warning and localization method based on acoustic signals. They also proposed a wavelet transform-based false alarm prevention mechanism to ensure the reliability of fault warning and localization. Cai et al. [14] utilized historical observation data of power systems to build a predictive model for power distribution network outages using machine learning, particularly the XGBoost algorithm. The method enabled early warning and maintenance planning for distribution network faults. In the wind turbine field, Tan et al. [15] improved the social engineering optimizer method for optimizing BP networks in hydraulic turbine fault warning systems. Experimental results showed superior performance compared to other methods. Min et al. [16] proposed a fault prediction method for distribution networks using convolutional neural network and LightGBM, enhancing adaptability to imbalanced datasets. The experiments with fault data from a specific region in 2017–2018 demonstrated improved fault prediction performance. Yang et al. [17] focused on monitoring oil and gas pipeline safety using machine learning on distributed fiber optic sensor signals. Their model accurately detected and identified damage events in real time. Jing et al. [18] introduced a LightGBM-based microservices fault identification method, ensuring high availability by analyzing historical runtime information. Tao et al. [19] reduced operation and maintenance costs through an early fault warning system based on real operational data. A novel hybrid algorithm for fault detection and early warning in the manufacturing industry was proposed by Liu et al. [20]. The algorithm combines the stochastic wandering strategy, the balanced optimizer and the adaptive learning strategy of BP neural networks, aiming to overcome the limitations of traditional BP neural networks in practical applications. Zhang et al. [21] proposed a particle swarm optimization algorithm based on nonlinearly decreasing inertia weights and exponentially varying learning factors for optimizing the deep belief network, which is used for solving the early warning problem of wind turbine faults. Zhang et al. [22] established a BP neural network prediction model optimized based on dynamic cuckoo search algorithm for solving modern industrial equipment fault detection and early warning problems. Experiments show that the model has faster convergence speed and higher prediction accuracy. Similarly, Pi et al. [23] also optimized the parameters of the BP neural network. They used an improved sand cat group optimization algorithm and introduced a new search strategy. Wu et al. [24] proposed a three-step framework for multimodal industrial process monitoring. They used a deep local adaptive network, two-stage qualitative trend analysis, and a five-state Bayesian network to gradually implement fault detection, identification, and diagnosis. Huang et al. [25] proposed a gas concentration early warning model involving the spark streaming framework. This model combines the particle swarm optimization algorithm and the gated recurrent unit model in the spark streaming framework. Liu et al. [26] proposed a coupled thermal-hydraulic-mechanical nonlinear model for predict and warn of water inrush in mines. Finally, Kong et al. [27] proposed a fault diagnosis model using a particle swarm optimization algorithm to optimize the structure of a B-spline network, aiming to improve the prediction accuracy of nonlinear time series.
Through the literature analysis, several observations are made:
  • In the engineering domain, ensuring the reliable operation of equipment and systems is crucial. While machine learning techniques are widely employed for fault warning tasks, the utilization of efficient methods like LightGBM is relatively limited in practical fault warning applications;
  • LightGBM, despite its efficiency, faces challenges such as the complexity of parameter selection, potentially leading to prolonged diagnosis time, increased operational costs, and reduced warning accuracy;
  • The combination of metaheuristic algorithms and machine learning technologies proves to be an effective approach for enhancing fault warning performance.
Therefore, this paper proposes a novel fault warning model based on the LightGBM framework. Specifically, the proposed fault warning model, integrating LightGBM and a novel hybrid metaheuristic algorithm, offers a promising solution to the challenges identified in the literature. Through experimentation with real industrial cases, the method showcases enhanced fault warning performance, contributing to the assurance of reliable equipment and system operation while reducing diagnostic time and operational costs.

3. Problem Definition

3.1. Research Purpose

The performance of the LightGBM model is significantly influenced by its parameters. To enhance the model’s performance, it is crucial to identify an appropriate method for adjusting these parameters. Given the superior performance of meta-heuristic algorithms, our objective is to develop a higher-quality meta-heuristic algorithm to determine the parameters of LightGBM and improve the model’s performance. Due to the large number of LightGBM parameters, we optimize the four important hyperparameters to obtain a more accurate fault warning model. The specific information of hyperparameters is shown in Table 1, and we adopt the default values for the other parameters. In other words, our goal is to develop an efficient metaheuristic algorithm to find the optimal set of hyperparameters within the given space that yields the best performance for the LightGBM model.

3.2. Evaluation Metrics

Before we delve deeper into our analysis, it is first necessary to clarify the meaning and calculation methods of several evaluation metrics. In this paper, we have chosen three main evaluation metrics: RMSE, R-Squared, and Relative Percentage Deviation (RPD). We use RMSE and R-Squared to evaluate the performance of the model. We use the RPD to calibrate the parameters of SAOA.
RMSE is a metric used to measure the difference between predicted values and actual values. It calculates the square of the prediction errors for each data point, averages these squared errors, and then takes the square root. A smaller RMSE indicates more accurate predictions and less deviation from the actual values. Therefore, RMSE is an important metric for assessing the accuracy of the model [15].
RMSE = 1 m i = 1 m ( y i y ^ i ) 2
where m is the number of samples, y i is the actual value, and y ^ i is the predicted value.
R-Squared is a metric that measures the goodness of fit of the model, also known as the coefficient of determination. It represents the proportion of the target variable’s variance explained by the model. R-Squared ranges from 0 to 1, where values closer to 1 indicate a better fit and better explanation of the data variance. When R-Squared is close to 0, it means that the model cannot explain the variability of the target variable. R-Squared is a commonly used statistic for evaluating the model’s fit to the data [15].
R S q u a r e d = 1 S S residual S S total
where S S residual is the sum of squares of residuals and S S total is the total sum of squares.
We use RPD to measure the performance of the algorithm under various parameter combinations, that is, the deviation of each experiment from the best experiment’s RPD. RPD is calculated as shown in Equation (3).
R P D = A l g R M S E M i n R M S E M i n R M S E
where M i n R M S E represents the minimum RMSE from the algorithm across all experiments, and A l g R M S E represents the value from the algorithm in each individual experiment.

4. Proposed Methods

In this section, we present our hybrid model. We begin by outlining the LightGBM model (Section 4.1) and subsequently introduce the customized hybrid optimization algorithm using SA, AOA and new search strategies tailored for the LightGBM model (Section 4.2).

4.1. LightGBM Model

LightGBM is a fast, distributed, high-performance gradient-boosting framework based on decision tree algorithms, widely used in various machine learning problems such as regression, ranking, and classification [5]. It is a type of Boosting algorithm, which combines multiple weak machine learning models to generate a strong learning model. Boosting algorithms increase the weights of misclassified data and decrease the weights of correctly classified data, so that misclassified classifiers receive more attention in the next round of training. Ultimately, all machine learning models are linearly combined, and the weights of the combined model are adjusted based on the classifier’s error rate [5].
The core concept can be represented by the following Equation (4):
f x = q = 1 Q α q T x , θ q
where f x is the target value corresponding to the training sample; Q is the number of base learners; α q is the weight coefficient of the qth base learner; x is the training sample; θ q is the parameter for the learner’s classification; and T x , θ q is the qth base learner involved in training.
After determining the model’s loss function and training data, the training process of the Boosting algorithm is transformed into an optimization problem to minimize the loss function. The objective function is as follows:
a r g m i n h = 1 H L ( y h , f ( x h ) )
where H is the number of samples; h is the sample index; y h is the actual value of the data; f ( x h ) is the target value corresponding to the hth sample; and L y h , f x h is the loss function value for the hth sample.
In the GBM, which is a boosting tree model obtained based on the gradient descent algorithm, after each addition of a new sub-model, the selected loss function continuously decreases towards the gradient of the variable with the next highest information content, as show in Equation (6).
L ( F j ( x ) , Y ) < L ( F j 1 ( x ) , Y )
where L ( F j ( x ) , Y ) and L ( F j 1 ( x ) , Y ) are the loss function values for the jth and (j − 1)th iterations, respectively, F j ( x ) and F j 1 ( x ) are the target values corresponding to the jth and (j − 1)th samples, and Y is the true target value of the sample.
As a type of GBM, LightGBM effectively addresses the challenges encountered by GBM when dealing with massive data. This model has two main features [5]:
(a)
Leaf-wise tree growth method: It adopts a leaf-wise growth strategy as opposed to level-wise growth. This method has a smaller computational cost and successfully avoids overfitting issues by controlling the minimum data volume of leaf nodes and the depth of trees;
(b)
Histogram-based decision tree algorithm: LightGBM chooses a histogram-based decision tree algorithm. When selecting features, it only needs to traverse and find the optimal split point based on the discrete values of the histogram, thereby reducing computational and storage costs. This feature enables LightGBM to train and predict more efficiently when handling large-scale datasets.

4.2. Proposed Hybrid Algorithm

While LightGBM has shown remarkable performance across diverse domains, its numerous parameter configurations necessitate optimization for effective adaptation to specific problem contexts [5]. Conventional grid search algorithms for parameter tuning are plagued by extended computation times and high computational demands. Recognizing the complexity of this challenge and the absence of a universal solution, we introduce an innovative SAOA. This algorithm seamlessly integrates AOA, SA, and novel search strategies to streamline parameter adjustments, enhancing the efficiency of the LightGBM model. Our approach aims to tackle the challenges associated with parameter optimization, promoting adaptability in the LightGBM model, and simultaneously reducing computation time, thereby enhancing overall algorithm efficiency.
Given our objective of obtaining the optimal LightGBM parameters through the use of the SAOA, the input to the SAOA in our problem is the LightGBM parameter space that we have defined. This parameter space encompasses all possible combinations of LightGBM parameters that we intend for the SAOA algorithm to explore. We select the RMSE between the actual and predicted values of the LightGBM as the fitness value. Consequently, the output of the SAOA algorithm is the optimal combination of parameters that minimizes the RMSE. This combination of parameters can optimize the performance of the LightGBM model. Subsequently, we utilize the LightGBM model with this set of parameters to carry out training and fault warning tasks. In general, in our problem, the input to the SAOA algorithm is the LightGBM parameter space we have defined, and the output is the optimal combination of parameters. In the following section, we will provide a detailed description of the execution steps of the SAOA. This will include a step-by-step walkthrough of the algorithm, highlighting key operations and decision points, and explaining how these contribute to the overall objective of optimizing the LightGBM parameters.

4.2.1. Population Initialization

Traditional AOA typically begins the optimization process by randomly generating the initial population, which introduces a series of complexities and challenges for subsequent optimization stages [28]. The randomly generated population increases the difficulty for the algorithm to find the global optimum or converge to a stable state. In this case, more effective population initialization strategies and improvements to the optimization algorithm become particularly important. To address this challenge, we introduce chaotic mapping and a dynamic reverse learning strategy to enhance the formation process of the initial population, thereby improving the algorithm’s performance and robustness in optimization problems. Specifically, we use logistic mapping to update the initial population, and the chaotic sequence generation method of the logistic mapping is shown in Equation (7):
F i = x ( i + 1 ) = μ x ( i ) ( 1 x ( i ) )
where F i is the logistic mapping coefficient for the ith individual, and x ( i ) is the mapping variable, a random number between 0 and 1. The logistic mapping is in a completely chaotic state when 3.57 < μ ≤ 4, exhibiting chaotic behavior in this interval. When μ is set to 4, better results can be obtained; here, we take μ = 4.
Afterward, we utilize the generated chaotic sequences to form initial individuals that meet the specified conditions, as illustrated in Equation (8).
x i , j = L B j + F i ( U B j L B j )
where x i , j denotes the value of the jth dimension of the ith individual, U B j is is the lower limit of the value taken in the jth dimension, L B j is the upper limit of the value taken in the jth dimension, and F i is the chaotic mapping coefficient of the ith individual.
Finally, we apply the dynamic reverse learning strategy to further optimize the initially improved individuals and merge them with individuals improved by logistic mapping. The initial population is then composed of Npop individuals, selected based on their optimal fitness. The dynamic reverse learning strategy is depicted in Equation (9).
x i , j = x i , j + c 1 · ( c 2 · ( U B j + L B j x i , j ) x i , j )
where x i , j denotes the value of the jth dimension of the ith individual, L B j is the lower limit of the value taken in the jth dimension, U B j is the upper limit of the value taken in the jth dimension, and c 1 and c 2 are random numbers in the range (0, 1).

4.2.2. Improved Mathematical Optimizer Acceleration Function

The Arithmetic Optimization Algorithm is primarily divided into two stages: global exploration and local exploitation [28]. In the global exploration stage, operators with high dispersion, such as multiplication or division, are employed for the search. In the local exploitation stage, operators with high density, such as subtraction or addition, are used for the search. The transition between the exploration and exploitation stages is controlled by the function value calculated by the mathematical optimizer acceleration (MOA) function [29]. When r 1 > M O A , the algorithm performs global exploration, and when r 1 < M O A , the algorithm executes local exploitation. The mathematical expression for MOA is given by Equation (10) [2]:
M O A ( t ) = M i n + t ( M a x M i n M a x i t )
where t is the number of iterations, Min and Max are the minimum and maximum values of the acceleration function, respectively, and Maxit is the maximum number of iterations.
When using MOA to switch between “global exploration” and “local exploitation” in the standard AOA, the algorithm tends to have a lower probability of transitioning to local exploitation in the later stages of the search. This is inconsistent with the common search strategy employed by general intelligent optimization algorithms, which emphasizes global exploration in the early stages and shifts towards local exploitation in the later stages. This discrepancy weakens the algorithm’s ability for local development in the later stages, negatively impacting optimization speed and accuracy. Therefore, this paper proposes a new MOA(t), as shown in Equation (11):
M O A ( t ) = a × ( l e x p ( ( M a x i t t ) / M a x i t ) ) + b
where t is the current number of iterations, Maxit is the maximum number of iterations, and a and b are control parameters located in the range (0, 1).

4.2.3. AOA Global Exploration Phase

In the global exploration phase of AOA, mathematical calculations involving division and multiplication operators can yield values or decisions that align with the global exploration mechanism [28]. The position update formula during the global exploration phase is represented as Equation (12) [28].
x i , j t + 1 = x b j ÷ ( M O P + ε ) × ( ( U B j L B j ) × m μ + L B j ) , r 2 < 0.5 x b j × M O P × ( ( U B j L B j ) × m μ + L B j ) , r 2 0.5
where x i , j t + 1 is the updated value of the jth dimension of the ith individual, and x b j is the value of the jth dimension of the current optimal individual. L B j is the lower limit of the value of the jth dimension, and U B j is the upper limit of the value of the jth dimension. ε is a minimal value, and m μ is a control parameter, which is set to be 0. 499 in this paper.
MOP is the mathematical optimizer probability, and the mathematical expression of MOP is shown in Equation (13) [28].
M O P ( t ) = 1 t 1 a T 1 a
where t is the current iteration number, Maxit is the maximum iteration number, and α is a sensitivity parameter that defines the probing accuracy and takes the value 5.

4.2.4. AOA Local Exploration Phase

During the local exploitation phase of AOA, mathematical calculations involving subtraction and addition operators can yield high-density values, advantageous for the algorithm’s local exploitation [28]. The position update formula during the local exploitation phase is represented as Equation (14) [28].
x i , j t + 1 = x b j M O P × ( ( U B j L B j ) × m μ + L B j ) , r 3 < 0.5 x b j + M O P × ( ( U B j L B j ) × m μ + L B j ) , r 3 0.5
where x i , j t + 1 is the updated value of the jth dimension of the ith individual, and x b j is the value of the jth dimension of the current optimal individual. L B j is the lower limit of the value taken in the jth dimension, and U B j is the upper limit of the value taken in the jth dimension.

4.2.5. Nonlinear Inertia Weight Factor

To enhance the balance between the exploration and exploitation capabilities of the SAOA algorithm, this paper proposes an improvement to the SAOA formula by introducing a nonlinear inertia weight factor, denoted as w, as shown in Equation (15). Subsequently, the formulas for the global search and local search stages of SAOA are presented in Equations (16) and (17).
w ( t ) = w m a x · ( 1 s i n ( π t 2 M a x i t ) ) + w m i n s i n ( π t 2 M a x i t )
where w m a x and w m i n are the maximum and minimum values of weight change; we set w m i n   = 0.4 and w m a x     = 0.9. Maxit is the maximum number of iterations and t is the current number of iterations.
x i , j t + 1 = x b j M O P × ( ( U B j L B j ) × m μ + L B j + w x b , j t ) , r 3 < 0.5 x b j + M O P × ( ( U B j L B j ) × m μ + L B j + w x b , j t ) , r 3 0.5
x i , j t + 1 = x b j ÷ ( M O P + ε ) × ( U B j L B j × m μ + L B j + 1 w ( t ) ( x b , j t x i , j t ) ) , r 2 < 0.5 x b j × M O P × ( ( U B j L B j ) × m μ + L B j + 1 w ( t ) ( x b , j t x i , j t ) ) , r 2 0.5
where x i , j t + 1 is the updated value of the jth dimension of the ith individual, and x b j is the value of the jth dimension of the current optimal individual. L B j is the lower limit of the jth dimension of the individual, and U B j is the upper limit of the jth dimension of the individual. Ε is a minimal value; w ( t ) is the inertia weight of the tth iteration; and m μ is a control parameter, which is set to 0.499 in this paper.

4.2.6. SA Algorithm

Traditional AOA relies on updating based on the optimal solution obtained in each iteration [28]. However, this strategy may lead the algorithm to converge to a local optimum, hindering global optimization. To address this issue, the SA is introduced.
Key features of the SA include [29]:
  • Retention of inferior solutions: In certain probability scenarios, the SA chooses to retain inferior solutions. This practice helps avoid premature convergence to local optima, thereby enhancing the algorithm’s global search capability;
  • Increased population diversity: By preserving inferior solutions, the algorithm introduces a level of randomness, augmenting the diversity of the population. This is done to explore a broader range of potential solutions in the search space, rather than being confined to the vicinity of local optima;
  • Enhanced escape from local optima: The SA, through the strategy of accepting inferior solutions during the search process, increases the probability of escaping local optima. This makes the algorithm more robust and better adaptable to complex search spaces.
The calculation method is represented by Equation (18).
p = 1 ,   if   f new   < f old   e x p f new   f old   T ,   if   f new   f old  
where p is the likelihood of accepting the new solution, fnew denotes the function value for the new solution, and fold represents the same for the initial solution. Drawing inspiration from the annealing process, T serves as the existing temperature of the SA during this iteration. At the end of this subloop, we update this iteration as shown in Equation (19).
T u + 1 = T u · C
where T u is the temperature at the uth search and C is the rate of temperature reduction.

4.2.7. Overall Framework of the Algorithm

Combining the aforementioned components, the specific steps for SAOA’s hyperparameter optimization for LightGBM are as follows:
Step 1: Perform data preprocessing. Data standardization is performed;
Step 2: Set the basic parameters of SAOA algorithm and the range of hyperparameters of LightGBM;
Step 3: Perform SAOA individual initialization;
Step 4: Implement 10-fold cross-validation for LightGBM model training, using the RMRE as the fitness function. Calculate the fitness values of each initial individual.
The 10-fold cross-validation is a technique used in machine learning to assess model performance. The basic idea is to divide the dataset into ten equal parts (folds) and then perform ten rounds of model training and evaluation.
The process is as follows:
Step 1: Divide the dataset into ten parts, usually through random sampling or sequential splitting;
Step 2: Conduct ten rounds of iteration. In each round, use one part as the test set and the other nine parts as the training set. In each iteration, train the model using the training set and evaluate its performance on the test set;
Step 3: Record the model’s performance metric on the test set for each round;
Step 4: After completing ten rounds, average the results of all performance metrics to obtain a more robust performance evaluation;
Using 10-fold cross-validation provides a more robust evaluation and maximizes the use of data. By repeatedly using different data splits, 10-fold cross-validation reduces dependence on a single split, thereby improving the reliability of the evaluation;
Step 5: Select the individual with the optimal fitness value as the current best individual (i.e., the current best hyperparameter combination), and record the best fitness value and individual position in an external file;
Step 6: Use the SAOA operators to update individuals and calculate their fitness values;
Step 7: Find the updated best individual and compare it with the current best individual. According to the simulated annealing strategy, update the best individual’s position and best fitness value. Additionally, record the best fitness value and individual position in an external file;
Step 8: Check if the termination condition is met; if not, return to Step 6, otherwise, output the optimal hyperparameter combination from the external file, and the optimization process concludes.
Finally, Figure 1 shows the flowchart of our proposed algorithm.

5. Case Study

In this case study, we perform a comprehensive fault warning analysis utilizing a water pump within a power plant setting, as depicted in Figure 2. The investigation encompasses eleven critical measurement points linked to the pump’s operational condition: non-drive end bearing temperatures (Non-driving end bearing temperature 1 and Non-driving end bearing temperature 2), non-drive end bearing vibrations in the X and Y directions, stator winding temperatures of the condensate pump (A, B, C phase temperatures 1 and 2), and the pressure drop across the inlet filter. In addition, we use the entrance flow of the pump as an input measurement point because it can reflect many types of faults.
To conduct our analysis, we aggregate these measurement points into a cohesive sample. Our dataset comprises a total of 1500 samples sourced from the supervisory information system. These samples form the basis of our original dataset, which we subsequently partition into training and testing sets at a ratio of 7:3 to facilitate robust model evaluation.

5.1. Data Analysis

As previously mentioned, our data comprises eleven input measurement points and one output measurement point. To facilitate a clearer understanding of the data, we conduct an analysis. The upper and lower limits, standard deviation, median, and average of the data for each measurement point are displayed in Table 2 and Table 3. Additionally, we utilize box plots to illustrate the distribution of the data, which can be found in Appendix A. This comprehensive analysis allows for a more in-depth understanding of the data’s characteristics and underlying patterns, which is crucial for the subsequent modeling process.

5.2. SAOA Algorithm Parameter Calibration

Metaheuristic algorithms play a pivotal role in tackling intricate optimization problems, and their efficacy hinges on the meticulous tuning of parameters [30,31]. Drawing insights from a comprehensive literature analysis [15,20,21,22,29] and initial experiments, we established reference values for each parameter level. SAOA has six parameters, their meanings and predefined values are outlined in Table 4.
However, conducting a full experiment for parameter tuning would require significant computational resources. In light of this, the Taguchi method provides a robust experimental design framework for efficiently exploring the parameter space and determining optimal settings. Following the recommendations of the Taguchi method, we adopted the experimental structure outlined in Table 5, consisting of 18 experiments.
Next, we calculate the mean value of each parameter across all experiments to determine its optimal level.
We have 18 sets of experiments, each with a performance measure, namely RPD. Our goal is to find a set of parameter settings that optimize the performance of the algorithm. Specifically, we used the following steps to determine the optimal level of each parameter:
Step 1: For the Npop, we first consider its performance at level 1. We calculate the performance of Npop at level 1 (Mean RPD) by adding the RPD values of experiments 1–6 and then dividing by the number of experiments;
Step 2: For parameter a, we also need to calculate its performance at level 1. We add the RPD values of experiments 1, 2, 7, 8, 13, 14, and then divide by the number of experiments to get the performance of a at level 1;
Step 3: We repeat the above process, calculating for each level of each parameter, to get the performance of each parameter at each level;
Step 4: Finally, for each parameter, we choose the level that minimizes the average RPD as the optimal level for that parameter. In this way, we get the optimal setting for each parameter, which is the result of parameter calibration.
This systematic approach allows us to fine-tune the parameters of the algorithm, thereby optimizing its performance. The optimal parameter settings obtained through this process can significantly enhance the accuracy and reliability of the algorithm’s output.
Subsequently, the performance of each parameter at various levels is illustrated in Figure 3. Based on the insights from Figure 3, we establish the following parameter settings: Npop = 50, a = 0.8, b = 0.8, Maxit = 200, T1 = 1500, and C = 0.97.

5.3. Early Warning Strategy

After determining the parameters, we input the SAOA parameters, LightGBM parameter space, and the fault-measuring data points from Section 5 into the algorithm. We then start running the program to obtain the initial SAOA individuals (i.e., possible combinations under the LightGBM parameter space).
Next, we execute the SAOA optimization process with the RMSE of the LightGBM prediction test set serving as the value of the fitness function. This process is aimed at obtaining the optimal LightGBM parameters, which are shown in Table 6.
Finally, we assign the parameters in Table 6 to LightGBM for model training and to output the trained model. The test set outcomes are presented in Figure 4. Subsequently, we compute the residual E, capturing the variance between predicted and actual values, along with its mean ( μ ) and standard deviation ( σ ). For establishing the alarm limit, we set w l = μ 3 σ   a n d   w u = μ + 3 σ , where w l denotes the lower warning value, and w u signifies the upper warning value. To enhance precision, we incorporate a tolerance count (k = 3), indicating that a pump failure warning signal is activated only when the residuals from the predictions of three consecutive operational datasets surpass the alarm limit.

5.4. Early Warning Test

Here, we collected data for twenty sets of each of the four types of faults to verify the accuracy of the proposed method. These four types of faults are high bearing temperature (F1), excessive bearing vibration (F2), high winding temperature (F3), and high inlet filter pressure drop (F4). We input this faulty data into the trained model to assess its fault warning effectiveness. Finally, the results are presented in Figure 5.
The results reveal that the proposed method has effectively delivered precise fault warning outcomes, showcasing notable accuracy in the test data. This validation underscores the efficacy and reliability of the introduced approach. Additionally, we notice that F4 has the highest success rate. This is because our fault determination criterion is the inlet flow rate, and the high inlet filter pressure drop (F4) fault has a greater impact on the inlet flow rate compared to other types of faults, and is more sensitive; that is to say, its fault characteristics are more prominent in the data, the deviation from the data during normal operation is also greater, and it is more adaptable to the fault warning strategy we proposed.

5.5. Algorithm Application Discussion

This paper introduces a novel fault warning method with extensive practical applications, particularly in the realm of industrial equipment fault diagnosis. Our research offers a new optimization method by enhancing LightGBM with a hybrid meta-heuristic algorithm, which can bolster the accuracy and efficiency of fault warnings. This is of significant value for fault warning tasks of industrial equipment that necessitate handling large-scale data and making complex predictions.
Moreover, the SAOA optimization method we proposed can also be applied to other machine learning models that require extensive parameter tuning, thereby increasing its applicability, and broadening the warning effect of industrial equipment. In wider fields such as transportation, energy, and finance, our research results also hold substantial application potential. In the transportation sector, the fault warning system can be employed to monitor the operational status of transportation equipment (such as trains, airplanes, cars, etc.), and detect potential faults in advance, thereby averting accidents and ensuring the safety and efficiency of transportation. In the energy sector, the fault warning system can be utilized to monitor the operational status of energy equipment (such as wind turbines, solar panels, etc.), and detect potential faults in advance, thereby preventing energy waste and enhancing energy utilization efficiency. In the financial sector, the fault warning system can be used to monitor the dynamics of the financial market and detect potential risks in advance, thereby avoiding the occurrence of financial crises and safeguarding the interests of investors.
In conclusion, the novel fault warning method proposed in this paper, along with the SAOA optimization method, not only improves the performance of the LightGBM model but also has wide-ranging applications across various sectors, making it a valuable contribution to the field.

6. Comparison with Other Algorithms

In this section, we conduct a comprehensive comparison of the proposed SAOA algorithm with other methods, including ISEO-BP [15], LSTM-CNN [32], and GWO-LightGBM [33]. The reasons we chose these three algorithms as benchmarks are as follows:
  • ISEO-BP: The ISEO-BP is an enhanced version of the traditional BP network. BP networks are commonly used in fault warning systems, but they have some limitations, such as a tendency to fall into local optima. The ISEO method optimizes the BP network to mitigate this issue, thereby providing an efficient fault warning system. ISEO-BP improves the weight update mechanism, enabling the network to avoid local optima during the learning process, which enhances the accuracy and reliability of fault detection;
  • LSTM-CNN: The combination of LSTM and CNN is particularly effective for sequence prediction problems with spatial input. In fault warning, this method can help the model understand the time dependence in the data, thereby better predicting potential faults that may occur in the future. LSTM networks can capture long-term dependencies in time series data, while CNNs can extract useful features from spatial data. The combination of these two networks allows the model to handle the complexity of both time and space, thereby improving the accuracy of fault warnings;
  • GWO-LightGBM: GWO is a renowned optimization algorithm known for its high-quality solutions and efficiency. GWO simulates the hunting behavior of grey wolves, optimizing problems by searching for the optimal solution. When GWO is combined with LightGBM, it can further enhance the performance and accuracy of the model. By comparing GWO and our developed SAOA, we can demonstrate the superiority of SAOA in handling complex problems.
By comparing with these algorithms, we further validate the effectiveness of SAOA. All algorithm parameters are calibrated using the Taguchi method. To ensure fairness, each algorithm is run ten times, and the average values are taken. To comprehensively evaluate the performance of the algorithms, we use two metrics: RMSE and R-Squared.
Additionally, to ensure the accuracy of the statistical analysis, we perform a 95% confidence interval statistical analysis on the results of each algorithm. Table 7 demonstrates the average results of the performance metrics of each algorithm after ten runs, and Figure 6 shows its statistical results.
Based on the results in Table 5, the proposed hybrid algorithm achieves optimal values in both metrics compared to other state-of-the-art algorithms, which fully proves its superiority. In addition, the 95% confidence interval in Figure 6 demonstrates its stability and robustness.
Finally, we apply the models trained by each algorithm to the fault warning test for actual comparison. This step further validates the performance of SAOA in practical applications and provides a clear comparison of the relative performance of the algorithms in fault prediction. Table 8 demonstrates the results of their application testing, and Figure 7 provides a more visual graphical presentation.
The results above underscore the practical significance of SAOA-LightGBM, demonstrating its superior success rates in comparison to other algorithms across four fault warnings.
In summary, the algorithm introduced in this study not only outperforms alternative approaches but also showcases its potential for real-world applications by consistently delivering superior performance.

7. Conclusions and Future Works

In the industrial field, with the advancement of technology and the expansion of the industry, the demand for fault early warning has become increasingly important for the safe operation of equipment and the improvement of production efficiency. As a representative algorithm of machine learning, LightGBM has brought new breakthroughs to fault early warning. However, in practical applications, LightGBM still has a large number of parameters that are difficult to determine; effectively determining these parameters to optimize the performance of the model is a challenging problem.
To solve this problem, this study proposes a hybrid algorithm that combines AOA, SA, and innovative search strategies, namely SAOA. The proposed SAOA cleverly uses the advantages of AOA and SA, and through innovative search strategies, it optimizes the hyperparameters of LightGBM more effectively. Specifically, the combination of AOA and SA can find a good balance between global and local search, and the innovative search strategy can help the algorithm find the optimal solution faster. By combining SAOA with LightGBM, the parameters that make LightGBM perform best can be found to the greatest extent, making it more effective in practical applications.
Through a series of industrial case studies, we found that the proposed method is superior to other advanced methods in terms of fault prediction accuracy and generalization ability. This shows that our solution can detect problems in time, reduce the risk of production interruption, while also providing reliable support for industrial production. Therefore, we believe that this algorithm has broad application prospects in the industrial field and can help enterprises improve safety and production efficiency.
However, we also recognize that although this study has achieved significant results in the field of fault early warning, there are still some limitations, which provide possibilities for future research directions. First, the performance of the algorithm depends on data quality and feature selection, which requires further improvement in key research areas such as data preprocessing and feature engineering [34,35]. Secondly, there is room for improving the computational accuracy of the hybrid algorithm, which requires exploration of more effective optimization strategies and integration of more hybrid methods [36,37]. In addition, as the industrial field develops, future research can delve into combining deep learning and other machine learning methods with LightGBM to further improve the efficiency of fault early warning systems [38].

Author Contributions

Data curation, X.G.; Formal analysis, N.J. and A.D.; Methodology, M.Z.; Project administration, X.G.; Resources, A.D. and M.Z.; Software, S.L.; Supervision, N.J.; Validation, S.L.; Visualization, Y.Y.; Writing—original draft, S.L.; Writing—review & editing, A.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Figure A1. Statistical distribution of input data.
Figure A1. Statistical distribution of input data.
Processes 12 00221 g0a1aProcesses 12 00221 g0a1b
Figure A2. Statistical distribution of output data.
Figure A2. Statistical distribution of output data.
Processes 12 00221 g0a2

References

  1. Peng, J.; Kimmig, A.; Wang, D.; Niu, Z.; Zhi, F.; Wang, J.; Liu, X.; Ovtcharova, J. A systematic review of data-driven approaches to fault diagnosis and early warning. J. Intell. Manuf. 2023, 34, 3277–3304. [Google Scholar]
  2. Zhang, X.; Zhou, H.; Fu, C.; Mi, M.; Zhan, C.; Pham, D.T.; Fathollahi-Fard, A.M. Application and planning of an energy-oriented stochastic disassembly line balancing problem. Environ. Sci. Pollut. Res. Int. 2023, 30, 1–15. [Google Scholar] [CrossRef] [PubMed]
  3. Zhao, H.; Liu, H.; Hu, W.; Yan, X. Anomaly detection and fault analysis of wind turbine components based on deep learning network. Renew. Energy 2018, 127, 825–834. [Google Scholar] [CrossRef]
  4. Lei, Y.; Yang, B.; Jiang, X.; Jia, F.; Li, N.; Nandi, A.K. Applications of machine learning to machine fault diagnosis: A review and roadmap. Mech. Syst. Signal Process. 2020, 138, 106587. [Google Scholar] [CrossRef]
  5. Ke, G.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.; Liu, T.Y. Lightgbm: A highly efficient gradient boosting decision tree. In Advances in Neural Information Processing Systems; The MIT Press: Cambridge, MA, USA, 2017; Volume 30. [Google Scholar]
  6. Wang, D.; Zhang, Y.; Zhao, Y. LightGBM: An effective miRNA classification method in breast cancer patients. In Proceedings of the 2017 International Conference on Computational Biology and Bioinformatics, Newark, NJ, USA, 18–20 October 2017; pp. 7–11. [Google Scholar]
  7. Zhang, X.; Fu, A.; Zhan, C.; Pham, D.T.; Zhao, Q.; Qiang, T.; Aljuaid, M.; Fu, C. Selective disassembly sequence planning under uncertainty using trapezoidal fuzzy numbers: A novel hybrid metaheuristic algorithm. Eng. Appl. Artif. Intell. 2024, 128, 107459. [Google Scholar] [CrossRef]
  8. Wolpert, D.H.; Macready, W.G. No free lunch theorems for optimization. IEEE Trans. Evol. Comput. 1997, 1, 67–82. [Google Scholar] [CrossRef]
  9. Wang, H.; Chen, J.; Zhu, X.; Song, L.; Dong, F. Early warning of reciprocating compressor valve fault based on deep learning network and multi-source information fusion. Trans. Inst. Meas. Control 2023, 45, 777–789. [Google Scholar] [CrossRef]
  10. Li, J.; Liu, J.; Chen, Y. A fault warning for inter-turn short circuit of excitation winding of synchronous generator based on GRU-CNN. Glob. Energy Interconnect. 2022, 5, 236–248. [Google Scholar] [CrossRef]
  11. Lu, G.; Wen, X.; He, G.; Yi, X.; Yan, P. Early fault warning and identification in condition monitoring of bearing via wavelet packet decomposition coupled with graph. IEEE/ASME Trans. Mechatron. 2021, 27, 3155–3164. [Google Scholar] [CrossRef]
  12. Gao, D.; Wang, Y.; Zheng, X.; Yang, Q. A fault warning method for electric vehicle charging process based on adaptive deep belief network. World Electr. Veh. J. 2021, 12, 265. [Google Scholar] [CrossRef]
  13. Lyu, N.; Jin, Y.; Miao, S.; Xiong, R.; Xu, H.; Gao, J.; Liu, H.; Li, Y.; Han, X. Fault warning and location in battery energy storage systems via venting acoustic signal. IEEE J. Emerg. Sel. Top. Power Electron. 2021, 11, 100–108. [Google Scholar] [CrossRef]
  14. Cai, J.; Cai, Y.; Cai, H.; Shi, S.; Lin, Y.; Xie, M. Feeder fault warning of distribution network based on XGBoost. J. Phys. Conf. Ser. 2020, 1639, 012037. [Google Scholar] [CrossRef]
  15. Tan, Y.; Zhan, C.; Pi, Y.; Zhang, C.; Song, J.; Chen, Y.; Golmohammadi, A.M. A Hybrid Algorithm Based on Social Engineering and Artificial Neural Network for Fault Warning Detection in Hydraulic Turbines. Mathematics 2023, 11, 2274. [Google Scholar] [CrossRef]
  16. Min, F.; Yaling, L.; Xi, Z.; Huan, C.; Yaqian, H.; Libo, F.; Qing, Y. Fault prediction for distribution network based on CNN and LightGBM algorithm. In Proceedings of the 2019 14th IEEE International Conference on Electronic Measurement & Instruments (ICEMI), Changsha, China, 1–3 November 2019; pp. 1020–1026. [Google Scholar]
  17. Yang, Y.; Li, Y.; Zhang, H. Pipeline safety early warning method for distributed signal using bilinear CNN and LightGBM. In Proceedings of the ICASSP 2021—2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021; pp. 4110–4114. [Google Scholar]
  18. Jing, N.; Li, H.; Zhao, Z. A microservice fault identification method based on LightGBM. In Proceedings of the 2022 IEEE 8th International Conference on Cloud Computing and Intelligent Systems (CCIS), Chengdu, China, 26–28 November 2022; pp. 709–713. [Google Scholar]
  19. Tao, L.; Siqi, Q.; Zhaochao, M.; Feng, X.G. Early fault warning of wind turbine based on BRNN and large sliding window. J. Intell. Fuzzy Syst. 2020, 38, 3389–3401. [Google Scholar] [CrossRef]
  20. Liu, J.; Zhan, C.; Wang, H.; Zhang, X.; Liang, X.; Zheng, S.; Meng, Z.; Zhou, G. Developing a Hybrid Algorithm Based on an Equilibrium Optimizer and an Improved Backpropagation Neural Network for Fault Warning. Processes 2023, 11, 1813. [Google Scholar] [CrossRef]
  21. Zhang, Z.; Wang, S.; Wang, P.; Jiang, P.; Zhou, H. Research on Fault Early Warning of Wind Turbine Based on IPSO-DBN. Energies 2022, 15, 9072. [Google Scholar] [CrossRef]
  22. Zhang, W.; Han, G.; Wang, J.; Liu, Y. A BP neural network prediction model based on dynamic cuckoo search optimization algorithm for industrial equipment fault prediction. IEEE Access 2019, 7, 11736–11746. [Google Scholar] [CrossRef]
  23. Pi, Y.; Tan, Y.; Golmohammadi, A.M.; Guo, Y.; Xiao, Y.; Chen, Y. A Fault Warning Approach Using an Enhanced Sand Cat Swarm Optimization Algorithm and a Generalized Neural Network. Processes 2023, 11, 2543. [Google Scholar] [CrossRef]
  24. Wu, H.; Fu, W.; Ren, X.; Wang, H.; Wang, E. A Three-Step Framework for Multimodal Industrial Process Monitoring Based on DLAN, TSQTA, and FSBN. Processes 2023, 11, 318. [Google Scholar] [CrossRef]
  25. Huang, Y.; Li, S.; Fan, J.; Yan, Z.; Li, C. A Spark Streaming-Based Early Warning Model for Gas Concentration Prediction. Processes 2023, 11, 220. [Google Scholar] [CrossRef]
  26. Liu, W.; Zhao, J.; Nie, R.; Liu, Y.; Du, Y. A coupled thermal-hydraulic-mechanical nonlinear model for fault water inrush. Processes 2018, 6, 120. [Google Scholar] [CrossRef]
  27. Kong, L.; Gong, X.; Yuan, C.; Xiao, H.; Liu, J. Nonlinear time series prediction model based on particle swarm optimization B-spline network. IFAC-Pap. 2018, 51, 219–223. [Google Scholar] [CrossRef]
  28. Abualigah, L.; Diabat, A.; Mirjalili, S.; Abd Elaziz, M.; Gandomi, A.H. The arithmetic optimization algorithm. Comput. Methods Appl. Mech. Eng. 2021, 376, 113609. [Google Scholar] [CrossRef]
  29. Bertsimas, D.; Tsitsiklis, J. Simulated annealing. Stat. Sci. 1993, 8, 10–15. [Google Scholar] [CrossRef]
  30. Fathollahi-Fard, A.M.; Wong, K.Y.; Aljuaid, M. An efficient adaptive large neighborhood search algorithm based on heuristics and reformulations for the generalized quadratic assignment problem. Eng. Appl. Artif. Intell. 2023, 126, 106802. [Google Scholar] [CrossRef]
  31. Fathollahi-Fard, A.M.; Tian, G.; Ke, H.; Fu, Y.; Wong, K.Y. Efficient Multi-objective Metaheuristic Algorithm for Sustainable Harvest Planning Problem. Comput. Oper. Res. 2023, 158, 106304. [Google Scholar] [CrossRef]
  32. Li, J.; Chen, Y. Application of LSTM-CNN network in the early warning of inter-turn short-circuit faults of synchronous motor excitation windings. J. North China Electr. Power Univ. Nat. Sci. Ed. 2020, 47, 61–70. [Google Scholar]
  33. Mirjalili, S.; Mirjalili, S.M.; Lewis, A. Grey wolf optimizer. Adv. Eng. Softw. 2014, 69, 46–61. [Google Scholar] [CrossRef]
  34. Liu, Y.; Guo, W.; Feng, J.; Zhang, K. A Summary of Methods for Statistical Downscaling of Meteorological Data. Adv. Earth Sci. 2011, 26, 837. [Google Scholar]
  35. Pu, R.; Bonafoni, S. Thermal infrared remote sensing data downscaling investigations: An overview on current status and perspectives. Remote Sens. Appl. Soc. Environ. 2023, 29, 100921. [Google Scholar] [CrossRef]
  36. Pasha, J.; Nwodu, A.L.; Fathollahi-Fard, A.M.; Tian, G.; Li, Z.; Wang, H.; Dulebenets, M.A. Exact and metaheuristic algorithms for the vehicle routing problem with a factory-in-a-box in multi-objective settings. Adv. Eng. Inform. 2022, 52, 101623. [Google Scholar] [CrossRef]
  37. Tian, G.; Zhang, L.; Fathollahi-Fard, A.M.; Kang, Q.; Li, Z.; Wong, K.Y. Addressing a collaborative maintenance planning using multiple operators by a multi-objective Metaheuristic algorithm. IEEE Trans. Autom. Sci. Eng. 2023, 1–13. [Google Scholar] [CrossRef]
  38. Al Daoud, E. Comparison between XGBoost, LightGBM and CatBoost using a home credit dataset. Int. J. Comput. Inf. Eng. 2019, 13, 6–10. [Google Scholar]
Figure 1. SAOA flowchart.
Figure 1. SAOA flowchart.
Processes 12 00221 g001
Figure 2. Image of a water pump.
Figure 2. Image of a water pump.
Processes 12 00221 g002
Figure 3. Parameter calibration results.
Figure 3. Parameter calibration results.
Processes 12 00221 g003
Figure 4. Test set prediction results.
Figure 4. Test set prediction results.
Processes 12 00221 g004
Figure 5. Results of the model application.
Figure 5. Results of the model application.
Processes 12 00221 g005
Figure 6. Algorithm performance metrics statistics results.
Figure 6. Algorithm performance metrics statistics results.
Processes 12 00221 g006
Figure 7. Results of the application of each algorithm model.
Figure 7. Results of the application of each algorithm model.
Processes 12 00221 g007
Table 1. LightGBM hyperparameters selection range.
Table 1. LightGBM hyperparameters selection range.
HyperparametersUpper and Lower LimitsRoleMeaning
n_estimators[10,100]Controls training speedEstimates
Learning rate[0.15,0.35]Impact AccuracyModel training learning Rate
num_leaves[5,100]Impact AccuracyNumber of leaf nodes
max_depth[3,11]Avoid overfittingMaximum depth of tree
Table 2. Results of data analysis of measurement point input data.
Table 2. Results of data analysis of measurement point input data.
Measurement PointUnitUpper LimitLower LimitStandard DeviationMedianMean
Condensate pump non-driven end bearing temperature 1°C54.6053.410.6153.8753.93
Condensate pump non-driving end bearing temperature 2°C47.3045.900.7746.7346.41
Condensate pump non-driven bearing X-direction vibrationmm/s0.570.480.050.530.54
Condensate pump non-driven end bearing Y-direction vibrationmm/s0.740.640.050.680.70
Condensate pump stator winding phase A temperature 1°C65.2062.611.3964.8064.20
Condensate pump stator winding phase A temperature 2°C64.8961.821.5563.0063.24
Condensate pump stator winding phase B temperature 1°C65.3062.871.2264.3064.16
Condensate pump stator winding phase B temperature 2°C66.0063.201.5163.6064.27
Condensate pump stator winding phase C temperature 1°C64.8061.701.5862.8063.07
Condensate pump stator winding phase C temperature 2°C67.1062.502.4264.0364.37
Inlet screen differential pressurekPa5.715.280.225.395.48
Table 3. Results of data analysis of measurement point output data.
Table 3. Results of data analysis of measurement point output data.
Measurement PointUnitUpper LimitLower LimitStandard DeviationMedianMean
Entrance flowton/h522.15325.5243.45453.63440.81
Table 4. SAOA parameter reference levels.
Table 4. SAOA parameter reference levels.
ParametersMeaningLevel 1Level 2Level 3
NpopPopulation size304050
aMOA control parameters0.20.250.3
bMOA control parameter0.80.850.9
MaxitIteration number100150200
T1SA initial temperature100012001500
CCooling rate0.920.950.97
Table 5. Results of orthogonal test.
Table 5. Results of orthogonal test.
Number of Experiments NpopabMaxitT1CRPD
1111111 0.1572
2112233 0.1894
3121332 0.1942
4123123 0.2137
5132321 0.2018
6133212 0.1123
7211323 0.0189
8213132 0.2098
9222222 0.0745
10223311 0.2236
11231231 0.1567
12232113 0.1975
13312312 0.0304
14313221 0.1689
15321213 0.0801
16322131 0.0823
17331122 0.1527
18333333 0.0456
Table 6. LightGBM optimal parameters.
Table 6. LightGBM optimal parameters.
HyperparametersOptimal Values
n_estimators15
Learning rate0.18
num_leaves20
max_depth6
Table 7. Results of algorithm performance comparison (Optimal value is represented in bold).
Table 7. Results of algorithm performance comparison (Optimal value is represented in bold).
AlgorithmsRMSER-Square
ISEO-BP21.395.02
LSTM-CNN24.190.70
GWO-LightGBM22.893.58
SAOA-LightGBM19.897.07
Table 8. Results of the application of each algorithm model (Optimal value is represented in bold).
Table 8. Results of the application of each algorithm model (Optimal value is represented in bold).
AlgorithmsF1F2F3F4
ISEO-BP18/2016/2017/2016/20
LSTM-CNN15/2015/2015/2016/20
GWO-LightGBM16/2014/2016/2018/20
SAOA-LightGBM18/2017/2018/2019/20
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, S.; Jin, N.; Dogani, A.; Yang, Y.; Zhang, M.; Gu, X. Enhancing LightGBM for Industrial Fault Warning: An Innovative Hybrid Algorithm. Processes 2024, 12, 221. https://doi.org/10.3390/pr12010221

AMA Style

Li S, Jin N, Dogani A, Yang Y, Zhang M, Gu X. Enhancing LightGBM for Industrial Fault Warning: An Innovative Hybrid Algorithm. Processes. 2024; 12(1):221. https://doi.org/10.3390/pr12010221

Chicago/Turabian Style

Li, Shuai, Nan Jin, Azadeh Dogani, Yang Yang, Ming Zhang, and Xiangyun Gu. 2024. "Enhancing LightGBM for Industrial Fault Warning: An Innovative Hybrid Algorithm" Processes 12, no. 1: 221. https://doi.org/10.3390/pr12010221

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop