You are currently viewing a new version of our website. To view the old version click .
Mathematics
  • Article
  • Open Access

21 June 2022

Hybrid LSTM-ARMA Demand-Forecasting Model Based on Error Compensation for Integrated Circuit Tray Manufacturing

,
and
Department of Industrial Engineering and Management, Ming Chi University of Technology, New Taipei City 24303, Taiwan
*
Author to whom correspondence should be addressed.
This article belongs to the Topic Multi-Criteria Decision Making

Abstract

Demand forecasting plays a crucial role in a company’s operating costs. Excessive inventory can increase costs and unnecessary waste can be reduced if managers plan for uncertain future demand and determine the most favorable decisions. Managers are demanding increasing accuracy in forecasting as technology advances. Most of the literature discusses forecasting results’ inaccuracy by suspending the model and reloading the data for model retraining and correction, which is extensively employed but causes a bottleneck in practice since users do not have the sufficient ability to correct the model. This study proposes an error compensation mechanism and uses the individuals and moving-range (I-MR) control chart to evaluate the requirement for compensation to solve the current bottleneck using forecasting models. The approach is validated using the case companies’ historical data, and the model is developed using a rolling long short-term memory (LSTM) to output the predicted values; then, five indicators are proposed for screening to determine the prediction statistics to be subsequently employed. Root mean squared error (RMSE), mean absolute percentage error (MAPE), and mean absolute error (MAE) compare the LSTM, rolling LSTM combined index, and LSTM-autoregressive moving average (ARMA) models. The results demonstrate that the RMSE, MAPE, and MAE of LSTM-ARMA are smaller than those of the other two models, indicating that the error compensation mechanism that is proposed in this study can enhance the prediction’s accuracy.

1. Introduction

Companies are finding it difficult to accurately forecast customer demand using traditional demand-forecasting approaches, preferring to employ data science approaches for demand-forecasting modeling due to the explosion of competition and information technology in the market [1,2]. Companies can use predictive analytics to respond to changes in future market trends to enhance their competitive advantage. Predictive analytics is a critical tool for understanding future needs. Since customer demand is the basis for all activity planning, it is time-varying. Thus, accurate demand forecasting will prevent stock-outs and increase customer satisfaction [3].
In lean manufacturing, inventory is one of the significant wastes. Inventory is a major bottleneck in a company’s operating costs. Demand forecasting can reduce inventory, so establishing an accurate demand-forecasting system is a primary objective for enhancing competitiveness. Demand forecasting has been used in various fields, such as electricity demand forecasting [4], tourism demand forecasting [5,6], restaurant demand forecasting [7], oil production forecasting [8], and stock market forecasting [9]. Most of them can initially attain the expected accuracy using statistical approaches or machine learning models, but after some time, they fail to generate the expected answers due to over-reliance on time. The most crucial aspect of forecasting is to respond effectively to trends, and managers focus on continually reducing the error between forecast and actual values. To enhance forecast accuracy, the development of models overtime is necessary. Thus, rolling forecasts are crucial steps to facilitate model correction. The error in the rolling prediction increases with the number of prediction steps due to white noise in the time series, causing the prediction to deteriorate with the number of steps. The error compensation approach employs historical forecast error to correct the future forecast values based on multi-step forecasting to compensate for each forecasted error value [10]. Furthermore, the air quality index (AQI) indicator only provides current forecasts and cannot provide valid values for those who will be outdoors in the next few hours. Liang et al. proposed a rolling AQI prediction model for the next 6 h for seven air quality regions in Taiwan using characteristic curves [11]. The mean absolute normalized coarse-error margins for rolling forecasts are all below 9%. The results demonstrate that a rolling forecast model that is built using the monthly characteristic curve can produce highly accurate forecasts.
The COVID-19 pandemic has triggered a global public health crisis, forcing many to work from home [12]. In this globally connected world, every country will be affected by the disruption that has been caused by the pandemic. This has led to total lockdowns in many countries around the world. In this case, all business activities in all industries have come to a complete halt. With social restrictions and closing city gates to prevent the epidemic spread of the virus, companies have launched strategies such as remote office working and video conferencing [13]. Therefore, the demand for notebook computers and tablet computers has also increased.
Integrated circuit (IC) trays have a protective effect on electronic products’ packaging. During packaging and testing, the IC components are placed in trays and then placed into test machines for testing. IC trays are one of the indispensable materials for precision parts IC components packaging tools, as well as a necessity for packaging and testing. The demand for plastic IC trays has grown with the electronics industry’s development [14]. In addition to supply and demand factors, the price of plastic raw materials is closely related to oil. COVID-19, supply–demand imbalances, and volatility in international financial markets are all associated with global oil demand, as is the collapse in the industrial feedstock oil’s price. However, price fluctuations can offer some gains or risks to the economies in the relevant chain. The related relevant chain includes raw material for IC trays (modified polyphenylene oxide), as well as packaging and testing manufacturers. Therefore, the Organization of Petroleum Exporting Countries member countries and allies have met to negotiate production cuts to save oil prices. However, the breakdown of negotiations between Saudi Arabia and Russia increased oil volumes instead of decreasing them, causing an oil price war that brought oil prices to a new low, affecting the prices of plasticized products [15]. If we can predict IC tray demand, we can purchase raw materials at low prices to reduce costs and increase profits, which becomes an advantage of low-price inventory and can react rapidly to market changes.
The demand for IC trays has been increasing due to various strategies adopted by various countries to combat the epidemic. The price of raw materials for producing IC trays has changed because of the unstable international situation arising from different external factors. Thus, if we know the demand for IC trays in advance, we can purchase the number of raw materials that are required to produce them in advance, in order to rapidly respond to the changes in market demand and maintain the competitiveness of our company. Manufacturers’ requirements for the accuracy of forecast results are increasing as information technology improves. Companies can reduce costs, increase profits, and enhance customer satisfaction by improving demand-forecasting accuracy. Thus, this study proposes an error compensation mechanism that captures nonlinear data with a rolling long short-term memory (LSTM) and then compensates for the residuals with the rolling autoregressive moving average (ARMA) model, which can only reduce the forecasting model’s maintenance cost and reduce the time for model correction and training.

2. Materials and Methods

In this study, we propose a solution for an IC tray manufacturer’s data characteristics in Taiwan using industry–academia cooperation as an example, and verify the method’s feasibility. Recently, manufacturers of IC trays have been moving toward a less sample, more volume model, as customers’ requirements for IC shapes and sizes frequently change due to product diversification. The IC tray industry’s service characteristics emphasize order acceptance and speed of delivery, resulting in shorter production times. To reduce demand uncertainty’s impact and risk caused by market fluctuations, a systematic and highly accurate demand-forecasting model to assist in decision-making is urgently needed. The flowchart shown in Figure 1 was designed to include data gathering and pre-processing, the construction of a time series module, and rule-based explanations.
Figure 1. The analysis framework flowchart.

2.1. Data Collection and Preprocessing

Based on the top three products offered by Industry–Academy Cooperation from 1 January 2017 to 29 September 2019, the proposed approach was validated. Time series are continuous data, and the approach of adding zeros is employed to deal with missing values in this study. The main reason is that, from a practical perspective, missing values show that no products were sold on that day, and the system does not record what caused the missing values. Since the case company calculates sales data every day, IC trays calculate material purchases every week. This study converts data measurement frequency from daily to weekly to forecast raw material demand. Figure 2 shows the BGA 11.5X13 product converted from daily to weekly units.
Figure 2. Timeline conversion of the difference graph. (a) BGA 11.5X13 timeline is daily; (b) BGA 11.5X13 timeline is one week.

2.2. Long Short-Term Memory Rolling Algorithm

Rolling forecasts are a type of dynamic forecasting. The main idea of this method is to add new data and delete old data simultaneously in a rolling manner while maintaining the length of the data. Forecasting results become more accurate through continuous learning, correction, and adjustment. The forecasting capability of an enterprise is thus enhanced so that it can respond promptly to customer needs [16]. Figure 3 illustrates a rolling forecasting process.
Figure 3. Rolling forecast schematic.
In 1997, Hochreiter proposed LSTM, a neural network model that solves the vanishing gradient problem for time series data [17]. LSTM regulates information flow, passes relevant information, and forecasts outcomes based on the incoming information. The previous hidden state is passed on to the next unit through the tanh activation function during the operation. A neural network’s hidden state acts as its memory, storing data it has seen in the past. A gate (forgetting, input, and output gates) allows the unit to decide whether to retain or delete information before passing it on to the next unit. Figure 4 shows the LSTM structure. The LSTM’s core, including input, output, and forgetting gates, is described below [18].
Figure 4. LSTM operational architecture.
  • Forgetting gate: Based on the weights and the current input, the sigmoid function calculates a value f t between 0 and 1, determining which data will be retained and forgotten. Forgetting the data implies that they have been permanently removed from long-term memory. The forgetting gate’s equation is as follows:
    f t = σ ( W f · [ h t 1 , X t ] + b f )
    where σ is the sigmoid function, ∙ is the dot product, [ h t 1 , X t ] is a vector of input data, W f is the forgetting-gate weight, and b f is the forgetting-gate bias vector.
  • Input gate: Given the weight, hidden state, and current input, the sigmoid function calculates i t , a value between 0 and 1. This value is between 0 and 1, multiplied by the tanh activation function to calculate the cell status and determine which messages in long-term memory must be updated, including updated data and data that must be replaced to be employed in the future. The input gate’s equation is as follows
    i t = σ ( W i · [ h t 1 , X t ] + b i )  
    where σ is the sigmoid function, · is the dot product, [ h t 1 ,   X t ] is a vector of input data, W i is the input gate weight, and b i is the input gate bias vector.
    To update the long-term memory from C t 1 at t 1 to C t at t , tanh computes the following equation, where W c is the tanh weight and b c is the tanh bias vector.
    C ¯ t = t a n h ( W c · [ h t 1 , X t ] + b c )  
    If there are data that the forgetting gate and input gate wish to remove and add, respectively, the formula is as follows. First, to remove the forgetting gate’s data, f t   C t 1 is multiplied element by element, then the data that the input gate wishes to add are added element by element to update I t     C ¯ t to the long-term memory.
    C t = f t C t 1 + I t C ¯ t
  • Output gate: According to the weight, hidden state, and current output, the sigmoid function computes O t , which is between 0 and 1, and multiplies that value with the tanh activation function to determine which data should be removed from long-term memory C t to t + 1 for further use. The output gate equation is as follows, where σ is the sigmoid function, ∙ is the dot product operation, [ h t 1 ,   X t ] vector is the input data, W o is the output gate weight, and b O is the output gate bias vector.
    O t = σ ( W o · [ h t 1 , X t ] + b o )
    In Figure 4, at the top right, h t are the output data sent to the output layer, and at the bottom,   h t are the input data sent to t + 1 . The output data h t are obtained by multiplying the C t by the tanh activation function element by element after computing the output data O t .
    h t = O t tanh ( C t )

2.3. Individuals and Moving-Range (I-MR) Control Chart

Individual and moving-range charts comprise I and MR charts. I control charts are employed to demonstrate the means and variances of individual values, and MR control charts monitor the data’s variation. To effectively monitor all prediction processes, this study employs the I-MR control chart to determine whether error compensation is needed. If a point exceeds the control limits of the I-MR control chart, the error compensation mechanism is activated, and vice versa. The following are the steps to build an I-MR chart [19]:
Step 1.
Calculate the moving range (MR). The MR is calculated by calculating the distance between the 2 adjacent points.
M R i = | x i x i 1 |
Step 2.
Calculate average ( x ¯ ) and Moving Range average ( M R ¯ ) .
x ¯ = i = 1 k x i k   ,   M R ¯ = M R i k 1
where x i is the observed value of the i-th sample, k is the number of samples to establish the control chart.
Step 3.
Construct the control charts for the Individual and moving range.
Individual   control   chart = { U C L = x ¯ + 3 M R ¯ d 2 C L = x ¯ L C L = x ¯ 3 M R ¯ d 2
Moving   range   control   chart = { U C L = D 4 M R ¯ C L = M R ¯ L C L = D 3 M R ¯
where, d2, D3, and D4 are control chart constants [19].
It is important to remember that data that do not follow a normal distribution can affect the results when using I-MR control charts. Before using I-MR control charts, it is essential to check if the data are the normal distribution. It is recommended that the Box-Cox conversion method be used prior to plotting the I-MR chart if the data are not normally distributed.

2.4. Error Compensation Using ARMA Technique

It is difficult to develop an accurate model since the prediction error is uncertain [20]. The prediction results would be improved if an error compensation approach could be developed. More accurate prediction results can be obtained by compensating for the errors that are generated during training and combining them with the original prediction results [21].
Numerous scholars employ the ARMA model, which consists of autoregressive (AR) and moving average (MA), as an approach for error compensation, since it can determine the time series trend and predict future results. To study the evaporation process of alumina, Qian proposed particle-swarm optimization and ARMA error compensation [22]. Furthermore, microelectromechanical system gyroscopes are extensively employed in dynamic measurement devices. However, their measurement accuracy cannot be very accurate due to the external environment’s influence. Chen and Gao used the ARMA model for compensation to enhance the measurement accuracy of microelectromechanical system gyroscopes, which resulted in a 3.75% reduction in the measurement error [23].
In this study, the LSTM model was first developed. Then, the ARMA compensation model was developed by filtering the rolling prediction results by selected indicators (maximum, minimum, median, average, and the recent predicted results) and calculating the error from the actual values. This study employed the ARMA model since it does not need differencing, and the prediction results were corrected using the compensation values that were predicted by the ARMA model. The ARMA formula is as follows.
y t = + i = 1 p φ i y t i + ε t + i = 1 q θ i ε t i          
where y t is the time series observation at time t; α is the constant; φ i is the parameter of y t i when the lag time is i; ε t is the white noise at time t; p is the maximum lag-time step; ε t i is the white noise when the lag-time step is i; θ i is the constant of ε t i when the lag-time step is i; q is the maximum lag-time step; and ARMA (p,q) is the ARMA model with the maximum lag-time step of the self-regression term p and the maximum lag-time step of the MA term q.
In constructing an ARMA model, the first step is to visualize the data and determine whether there are missing values. If there are missing values, this study will employ the zero-completion approach. Thereafter, the ARMA model’s parameters (p,q) are estimated, and the combination of parameters is obtained from the autocorrelation and partial autocorrelation functions., Along with the model, this is validated to determine whether the fitted model’s residual series is white noise. In the residual sequence, white noise is checked and the Akaike Information Criterion is employed to compute different levels and select the (p,q) combination with the lowest akaike information criterion (AIC) value. Finally, (p,q) is substituted into the ARMA model for prediction. Based on the error compensation, the prediction steps are as follows:
Step 1.
Calculate the mean absolute percentage error (MAPE), mean absolute error (MAE), and root mean squared error (RMSE) of each indicator, and select the one with the lowest error rate from the LSTM model.
Step 2.
Determine whether compensation is required by calculating the indicators’ errors and plotting the I-MR control chart.
Step 3.
In rolling LSTM, the number of products are predicted and multiple predictions are obtained for the test set.
Step 4.
The prediction result is filtered to y i based on the index that is selected in Step 1.
Step 5.
If compensation is needed, develop a rolling ARMA compensation model and calculate the compensation value A i .
Step 6.
To obtain the compensation result y i , add the compensation value A i to the filtered prediction result y i :
y i = A i + y i .
Step 7.
Calculate the error E i between the actual value y and y i , i.e., E i = y i y , which are the data for the next compensation model.

2.5. Data Partitioning and Parameter Setting

Figure 5 shows the rolling prediction process. As a first rollup, the data were divided into 118 weeks of yellow data from 1 January 2017 to 7 April 2019, five weeks of purple data from 14 April 2019 to 12 May 2019, and five weeks of green data from 19 May 2019 to 16 June 2019 for LSTM model training, indicators establishment, and compensation model building, respectively. Forecasting was based on pink section data from 23 June 2019 to 21 July 2019. The second rollup is based on the first compensation model values (19 May 2019–16 June 2019), a compensation model that was built using the first prediction (23 June 2019–21 July 2019), and a prediction of the last five data (28 July 2019–25 August 2019). This third rollup consists of an indicator that was constructed using the values from the second compensation model 23 June 2019–21 July 2019) and an indicator to the compensation model’s construction using the results from the second prediction (23 June 2019–21 July 2019). The third rollup uses the second compensation model (23 June 2019–21 July 2019), the second forecast (28 July 2019–25 August 2019) and predicts the last five data (1 September 2019–29 September 2019), and so on. In this study, only three cuts and rolls were made to the data to test the proposed approach.
Figure 5. The rolling prediction process.
After the training samples are calculated, Table 1 shows the estimated LSTM rolling pattern parameters.
Table 1. LSTM parameter setting.

2.6. Model Performance Indicator

Different indicators are used to evaluate a model’s predictive power. Leonardo designed a forecasting model for sugarcane yield and compared the root mean squared error, mean absolute error, and mean absolute percentage error as indicators of the model’s predictive ability [24]. To predict the number of COVID-19 confirmations in Iran, Nasrin et al., developed a predictive model and evaluated the model’s performance using RMSE and MAE. The findings show that Iran was probably the most affected and needed more preventative measures [25]. Chang et al., developed a prediction model for air pollution and evaluated the model’s performance using the MAE, RMSE, and MAPE. The findings indicated that the model considerably enhanced the prediction’s accuracy [26]. MAE, RMSE, and MAPE, commonly used evaluation indicators in practice, are employed in this study as the criteria for model evaluation. This formula is defined as follows. When the index calculation result is small, predictability is better.
  • MAE: The mean absolute error between the predicted y t ^ and actual values y t of each datum is measured using the following formula:
      MAE = 1 k t = 1 k | y t y t ^ |
  • RMSE: A measure of prediction error ( y t y t ^ ) standard deviation. The formula is as follows:
    RMSE = 1 k t = 1 k ( y t y t ^ ) 2
  • MAPE: This model overcomes the limitations of both mean absolute deviation (MAD) and RMSE using the relative prediction error of each data, which has overly large calculation results due to the large value of each data.
    MAPE = 1 k t = 1 k | y t y t ^ | y t × 100 %    
    y t is the actual value, y t ^ is the predicted value, and k is the sample size. Mean square error (MSE) and MAPE are unaffected by unit and value sizes and are evaluated objectively; the larger the sample size, the higher the RMSE.

3. Analysis Results

A validation of the proposed methodology is provided in Section 3.1, Section 3.2 and Section 3.3 by analyzing and discussing the top three sales volumes of the product (BGA 11.5X13, BGA 8X13 and BGA 8X12.5).

3.1. Case Study of BGA 11.5X13

First, we employ rolling training with a fixed prediction window size, followed by initial training after every prediction. The initial training value is deleted upon each prediction and a new piece of data is added to the model. This is repeated until the specified period’s data have been predicted. The LSTM rolling model is employed for the 14 April 2019 to 12 May 2019 results, then the numerical results are computed based on maximum (Max), minimum (Min), median (Med), mean (Mean), and recent results (New), which are more commonly employed in statistics, and the values are filtered, as shown in Table 2.
Table 2. BGA 11.5X13 error calculation results of each indicator.
RMSE, MAPE, and MAE were computed for each indicator, and the lowest error indicator was selected. For inconsistent results, as shown in Table 3, the median was selected as the indicator after RMSE calculation, the minimum after MAPE calculation, and the indicator with the most evaluated indicators was selected as the final decision, thus, the minimum (Min) will be selected as the next screening value.
Table 3. BGA 11.5X13 1st indicator calculation.
In Figure 6, the five errors in the selected metrics’ minimum values are plotted on the I-MR control chart, and if there are phenomena that exceed the control limits, the subsequent prediction results will be compensated.
Figure 6. I-MR control chart for BGA 11.5X13 data for the first prediction roll.
As there is a condition of exceeding the control limits, an error compensation is required. Thereafter, the error is substituted into the ARMA model to obtain the compensation value A 1 . To obtain the compensation result y 1 , the compensation value A 1 is then summed with the prediction results after 23 June numerical screening. According to a previous rolling concept, the first 19 May error result is deleted, and a new 23 June error result is added. The new error is introduced in the compensation model from the result y 1 resulting from the 23 June compensation and the actual error E 1 . Then, the compensation value A 2 is compensated to the selected prediction of 30 June at the end of the week from 23 June to 21 July. Figure 7 shows a schematic representation of rolling error compensation.
Figure 7. Schematic representation of rolling error compensation.
A rolling LSTM was used for the prediction’s last five weeks, and the selected indicators’ values filtered the results, and Table 4 shows the calculated compensation values. The ARMA (p,q) value in Table 4 is (0,0) since this series is stationary, and ARMA is the most extensively employed model for fitting stationary series.
Table 4. The result of calculating compensation value at different time points.
Table 5 summarizes the three rolling prediction results, in which the errors are computed by subtracting the actual values from the predicted values. In Table 5, the rolling prediction’s error with indicator filtering (LSTM-Min) is smaller than the LSTM prediction. If the previous judgment requires error compensation (LSTM-ARMA), the error remains smaller than in the previous two approaches.
Table 5. BGA 11.5X13 rolling forecast results summary.

3.2. Case Study of BGA 8X13

Table 6 summarizes the results of the BGA 8X13’s three rolls, showing that the median was selected as the filtering indicator, and since the I-MR control chart did not exceed the control limits, no error correction was necessary. The mean was selected as the filtering indicator for the second roll, and no error compensation was required since the I-MR control chart did not exceed the control limits. The median was selected as the filtering indicator in the third roll, and since the I-MR control chart did not exceed the control limits, an error compensation was necessary. In the third rollover, the median was selected as the filter and the I-MR control chart exceeded the control limits, so error correction was necessary.
Table 6. BGA 8X13 three rolling results.
Table 7 summarizes the three rolling prediction results by subtracting the actual values from the predicted values and by calculating the errors. In Table 7, the rolling prediction’s error combined with the indicator filtering approach is smaller than the general LSTM prediction’s error. If the previous judgment requires error compensation, the error after compensation is smaller than the previous two approaches’ errors.
Table 7. BGA 8X13 rolling forecast results summary.

3.3. Case Study of BGA 8X12.5

Table 8 summarizes the three BGA 8X12.5 rollover results. It shows that the first rollover selects the maximum value as the numerical filter, and the I-MR control chart does not exceed the control limits; therefore, no error compensation is necessary. In the second rollover, the median is selected as the numerical filter, and the I-MR control chart does not exceed the control limits, thus, error compensation is necessary. In the third rollover, the nearest result is the numerical filter, and error compensation is not required if the I-MR control chart does not exceed the control limits.
Table 8. BGA 8X12.5 three rolling results.
Table 9 summarizes the results of the three rolling predictions, where the error is calculated by subtracting the actual values from the predicted values. In Table 9, the rolling prediction combined with the indicator filter is more accurate than the general LSTM prediction; if the previous judgment requires error compensation, the error after compensation is smaller than that of the previous two approaches.
Table 9. BGA 8X12.5 rolling forecast results summary.

3.4. Model Evaluation

These three cases study are compared for the traditional LSTM approach, the rolling LSTM prediction using the indicator filtering numerical approach, and the LSTM-ARMA compensation approach. Table 10 shows the evaluation results using the RMSE, the MAPE, and the MAE.
Table 10. Comparison of RMSE, MAPE, and MAE results for three different IC trays.
There is a decreasing trend in terms of the RMSE; however, the rolling LSTM prediction combined with pointer screening offers a better prediction result than LSTM. If error compensation is required, the LSTM-ARMA approach results indicate a significant decrease in RMSE compared with the aforementioned approaches. In terms of MAPE, there is a decreasing trend, however, in most cases, the rolling LSTM prediction combined with the screening of indicators presents a better prediction result than the LSTM, probably because the demand for the case company’s orders is more volatile and the model cannot be more accurate. The MAE trend is declining, but most of the rolling LSTM prediction combined with indicator filtering offers better predictions than the LSTM, perhaps because demands for the company’s orders are more volatile, and the model cannot predict more accurately. If the error compensation is required, the MAE of LSTM-ARMA tends to decrease compared with the aforementioned approaches.

4. Discussion and Conclusions

The prediction model results’ accuracy improved due to the demand of competitors in the market. The forecasting model that is developed in this study is based on actual data from the case companies and is compared with the way that it was previously used. This study had three primary findings. At one time, forecasts were performed for one point at a time, i.e., tomorrow’s results were predicted as they would be. We forecast five points at a time, i.e., tomorrow, the day after tomorrow, the day after tomorrow, the day after tomorrow, etc. In the case of a later forecast, there is not only one forecast, but also the findings from the current period, the previous period, and the previous period before that. Several values are filtered using indicators, instead of using the most recent result as the final forecast. Second, in the past, we would pause the model and reload the data for new model training if we encountered poor prediction results, but this process was tedious and complicated. Finally, in the past, error compensation was activated regardless of the prediction results. The proposed error compensation mechanism is based on the I-MR control chart that is plotted by the difference between the indicators’ filtered and actual values, and the need for error compensation is determined as opposed to simply activating compensation regardless of the prediction results.
LSTM, rolling LSTM combined with an index, and LSTM-ARMA are compared, and the LSTM-ARMA approach proves to be more accurate when the error compensation is required. Furthermore, the rolling LSTM combined with the index offers better predictions than the firm’s rule of thumb and LSTM when no residential error compensation is required. Table 11 shows the evaluation of the ARIMA model, demonstrating that the LSTM-ARMA method is more accurate when error compensation is required.
Table 11. Results of the comparison between traditional ARIMA and our proposed LSTM-ARMA for three IC trays.
As a contribution to academic work, we have proposed an approach to determine whether an error compensation is necessary using rolling data for forecasting. In cases of model inaccuracy, the original model can be used without model correction. Combining LSTM and ARMA rolling forecasting approaches can generate more accurate forecasting results and eliminate the bottleneck problems that have been experienced in the past. In the past, regarding industry contribution, models were trained and tested before going online, and once they were employed, they gradually became inaccurate and needed to be recorrected, which took a long time. This study develops a model that uses error compensation to reduce the time and cost of pausing the model correction when past forecasts are inaccurate and compares it with empirical rules that are employed by companies to offer more accurate forecasting results.

Author Contributions

Conceptualization, C.-C.W. and H.-T.C.; methodology, C.-C.W. and H.-T.C.; validation, C.-C.W. and H.-T.C.; formal analysis, C.-C.W. and H.-T.C.; data curation, H.-T.C. and C.-H.C.; writing—original draft preparation, C.-C.W., H.-T.C. and C.-H.C. writing—review and editing, C.-C.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was financially supported by the Ministry of Science and Technology (MOST) of Taiwan (110-2221-E-131-027-MY3).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ICIntegrated circuit
AQIAir quality index
ARMAAutoregressive moving average
ARAutoregressive
MAMoving average
LSTMLong short-term memory
I-MRIndividuals and moving range
UCLUpper control limits
CLCenterline limits
LCLLower control limits
MADMean absolute deviation
MAEMean absolute error
MAPEMean absolute percentage error

References

  1. Guo, D.Y.; Shi, H.Z.; Qian, Y.P.; Lv, M.; Li, P.G.; Su, Y.L.; Tang, W.H. Fabrication of β-Ga2O3/ZnO heterojunction for solar-blind deep ultraviolet photodetection. Semicond. Sci. Technol. 2017, 32, 03LT01. [Google Scholar] [CrossRef]
  2. Kumar, V.; Dwivedi, S.K. Hexavalent chromium stress response, reduction capability and bioremediation potential of Trichoderma sp. isolated from electroplating wastewater. Ecotoxicol. Environ. Saf. 2019, 185, 109734. [Google Scholar] [CrossRef]
  3. Wang, C.-C.; Chien, C.-H.; Trappey, A.J.C. On the Application of ARIMA and LSTM to Predict Order Demand Based on Short Lead Time and On-Time Delivery Requirements. Processes 2021, 9, 1157. [Google Scholar] [CrossRef]
  4. Aparna, S. Long Short-Term Memory and Rolling Window Technique for Modeling Power Demand Prediction. In Proceedings of the International Conference on Intelligent Computing and Control Systems, Madurai, India, 14–15 June 2018; pp. 1675–1678. [Google Scholar]
  5. Park, S.; Lehto, X.; Lehto, M. Self-service technology kiosk design for restaurants: An QFD application. Int. J. Hosp. Manag. 2020, 92, 102757. [Google Scholar] [CrossRef]
  6. Bi, J.-W.; Li, H.; Fan, Z.-P. Tourism demand forecasting with time series imaging: A deep learning model. Ann. Tour. Res. 2021, 90, 103255. [Google Scholar] [CrossRef]
  7. Tanizaki, T.; Hoshino, T.; Shimmura, T.; Takenaka, T. Demand forecasting in restaurants using machine learning and statistical analysis. Procedia CIRP 2019, 79, 679–683. [Google Scholar] [CrossRef]
  8. Liu, W.; Liu, W.D.; Gu, J. Forecasting oil production using ensemble empirical model decomposition based Long Short-Term Memory neural network. J. Pet. Sci. Eng. 2020, 189, 107013. [Google Scholar] [CrossRef]
  9. Yu, P.; Yan, X. Stock price prediction based on deep neural networks. Neural Comput. Appl. 2019, 32, 1609–1628. [Google Scholar] [CrossRef]
  10. Mayr, J.; Blaser, P.; Ryser, A.; Hernández-Becerro, P. An adaptive self-learning compensation approach for thermal errors on 5-axis machine tools handling an arbitrary set of sample rates. CIRP Ann. 2018, 67, 551–554. [Google Scholar] [CrossRef]
  11. Liang, X.; Zhang, S.; Wu, Y.; Xing, J.; He, X.; Zhang, K.M.; Wang, S.; Hao, J. Air quality and health benefits from fleet electrification in China. Nat. Sustain. 2019, 2, 962–971. [Google Scholar] [CrossRef]
  12. Ma, J. Telework Triggered by Epidemic: Effective Communication Improvement of Telecommuting in Workgroups during COVID-19. Am. J. Ind. Bus. Manag. 2021, 11, 202. [Google Scholar] [CrossRef]
  13. Vagal, A.; Reeder, S.B.; Sodickson, D.K.; Goh, V.; Bhujwalla, Z.M.; Krupinski, E.A. The impact of the COVID-19 pandemic on the radiology research enterprise: Radiology scientific expert panel. Radiology 2020, 296, E134–E140. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Jiang, J.; Marsh, T.L.; Tozer, P.R. Policy induced price volatility transmission: Linking the US crude oil, corn and plastics markets. Energy Econ. 2015, 52, 217–227. [Google Scholar] [CrossRef]
  15. Hu, C.; Liu, X.; Pan, B.; Chen, B.; Xia, X. Asymmetric impact of oil price shock on stock market in China: A combination analysis based on SVAR model and NARDL model. Emerg. Mark. Financ. Trade 2017, 54, 1693–1705. [Google Scholar] [CrossRef]
  16. O’Connor, M.; Remus, W.; Griggs, K. Does updating judgmental forecasts improve forecast accuracy? Int. J. Forecast. 2000, 16, 101–109. [Google Scholar] [CrossRef]
  17. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  18. Abbasimehr, H.; Shabani, M.; Yousefi, M. An optimized model using LSTM network for demand forecasting. Comput. Ind. Eng. 2020, 143, 106435. [Google Scholar] [CrossRef]
  19. Montgomery, D.C. Introduction to Statistical Quality Control; John Wiley & Sons: New York, NY, USA, 2009. [Google Scholar]
  20. He, X.; Deng, K.; Wang, X.; Li, Y.; Zhang, Y.; Wang, M. Lightgcn: Simplifying and powering graph convolution network for recommendation. In Proceedings of the 43rd International ACM SIGIR conference on research and development in Information Retrieval, Xi’an, China, 25–30 July 2020; pp. 639–648. [Google Scholar]
  21. Liang, C.-J.; Liang, J.-J.; Jheng, C.-W.; Tsai, M.-C. A rolling forecast approach for next six-hour air quality index track. Ecol. Inform. 2020, 60, 101153. [Google Scholar] [CrossRef]
  22. Qian, X. Parameter prediction based on Improved Process neural network and ARMA error compensation in Evaporation Process. In IOP Conference Series: Earth and Environmental Science; IOP Publishing: Bristol, UK, 2018; Volume 108, p. 022078. [Google Scholar]
  23. Chen, M.M.; Gao, G.W. Research on MEMS gyroscope random error compensation algorithm based on ARMA model. Appl. Mech. Mater. 2014, 602, 891–894. [Google Scholar] [CrossRef]
  24. Maldaner, L.F.; de Paula Corrêdo, L.; Canata, T.F.; Molin, J.P. Predicting the sugarcane yield in real-time by harvester engine parameters and machine learning approaches. Comput. Electron. Agric. 2021, 181, 105945. [Google Scholar] [CrossRef]
  25. Talkhi, N.; Fatemi, N.A.; Ataei, Z.; Nooghabi, M.J. Modeling and forecasting number of confirmed and death caused COVID-19 in IRAN: A comparison of time series forecasting methods. Biomed. Signal Process. Control 2021, 66, 102494. [Google Scholar] [CrossRef] [PubMed]
  26. Chang, Y.S.; Chiao, H.T.; Abimannan, S.; Huang, Y.P.; Tsai, Y.T.; Lin, K.M. An LSTM-based aggregated model for air pollution forecasting. Atmos. Pollut. Res. 2020, 11, 1451–1463. [Google Scholar] [CrossRef]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.