Next Article in Journal
Using Financial News Sentiment for Stock Price Direction Prediction
Previous Article in Journal
Local Lagrange Exponential Stability Analysis of Quaternion-Valued Neural Networks with Time Delays
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hybrid LSTM-ARMA Demand-Forecasting Model Based on Error Compensation for Integrated Circuit Tray Manufacturing

Department of Industrial Engineering and Management, Ming Chi University of Technology, New Taipei City 24303, Taiwan
*
Author to whom correspondence should be addressed.
Mathematics 2022, 10(13), 2158; https://doi.org/10.3390/math10132158
Submission received: 20 May 2022 / Revised: 12 June 2022 / Accepted: 20 June 2022 / Published: 21 June 2022
(This article belongs to the Topic Multi-Criteria Decision Making)

Abstract

:
Demand forecasting plays a crucial role in a company’s operating costs. Excessive inventory can increase costs and unnecessary waste can be reduced if managers plan for uncertain future demand and determine the most favorable decisions. Managers are demanding increasing accuracy in forecasting as technology advances. Most of the literature discusses forecasting results’ inaccuracy by suspending the model and reloading the data for model retraining and correction, which is extensively employed but causes a bottleneck in practice since users do not have the sufficient ability to correct the model. This study proposes an error compensation mechanism and uses the individuals and moving-range (I-MR) control chart to evaluate the requirement for compensation to solve the current bottleneck using forecasting models. The approach is validated using the case companies’ historical data, and the model is developed using a rolling long short-term memory (LSTM) to output the predicted values; then, five indicators are proposed for screening to determine the prediction statistics to be subsequently employed. Root mean squared error (RMSE), mean absolute percentage error (MAPE), and mean absolute error (MAE) compare the LSTM, rolling LSTM combined index, and LSTM-autoregressive moving average (ARMA) models. The results demonstrate that the RMSE, MAPE, and MAE of LSTM-ARMA are smaller than those of the other two models, indicating that the error compensation mechanism that is proposed in this study can enhance the prediction’s accuracy.

1. Introduction

Companies are finding it difficult to accurately forecast customer demand using traditional demand-forecasting approaches, preferring to employ data science approaches for demand-forecasting modeling due to the explosion of competition and information technology in the market [1,2]. Companies can use predictive analytics to respond to changes in future market trends to enhance their competitive advantage. Predictive analytics is a critical tool for understanding future needs. Since customer demand is the basis for all activity planning, it is time-varying. Thus, accurate demand forecasting will prevent stock-outs and increase customer satisfaction [3].
In lean manufacturing, inventory is one of the significant wastes. Inventory is a major bottleneck in a company’s operating costs. Demand forecasting can reduce inventory, so establishing an accurate demand-forecasting system is a primary objective for enhancing competitiveness. Demand forecasting has been used in various fields, such as electricity demand forecasting [4], tourism demand forecasting [5,6], restaurant demand forecasting [7], oil production forecasting [8], and stock market forecasting [9]. Most of them can initially attain the expected accuracy using statistical approaches or machine learning models, but after some time, they fail to generate the expected answers due to over-reliance on time. The most crucial aspect of forecasting is to respond effectively to trends, and managers focus on continually reducing the error between forecast and actual values. To enhance forecast accuracy, the development of models overtime is necessary. Thus, rolling forecasts are crucial steps to facilitate model correction. The error in the rolling prediction increases with the number of prediction steps due to white noise in the time series, causing the prediction to deteriorate with the number of steps. The error compensation approach employs historical forecast error to correct the future forecast values based on multi-step forecasting to compensate for each forecasted error value [10]. Furthermore, the air quality index (AQI) indicator only provides current forecasts and cannot provide valid values for those who will be outdoors in the next few hours. Liang et al. proposed a rolling AQI prediction model for the next 6 h for seven air quality regions in Taiwan using characteristic curves [11]. The mean absolute normalized coarse-error margins for rolling forecasts are all below 9%. The results demonstrate that a rolling forecast model that is built using the monthly characteristic curve can produce highly accurate forecasts.
The COVID-19 pandemic has triggered a global public health crisis, forcing many to work from home [12]. In this globally connected world, every country will be affected by the disruption that has been caused by the pandemic. This has led to total lockdowns in many countries around the world. In this case, all business activities in all industries have come to a complete halt. With social restrictions and closing city gates to prevent the epidemic spread of the virus, companies have launched strategies such as remote office working and video conferencing [13]. Therefore, the demand for notebook computers and tablet computers has also increased.
Integrated circuit (IC) trays have a protective effect on electronic products’ packaging. During packaging and testing, the IC components are placed in trays and then placed into test machines for testing. IC trays are one of the indispensable materials for precision parts IC components packaging tools, as well as a necessity for packaging and testing. The demand for plastic IC trays has grown with the electronics industry’s development [14]. In addition to supply and demand factors, the price of plastic raw materials is closely related to oil. COVID-19, supply–demand imbalances, and volatility in international financial markets are all associated with global oil demand, as is the collapse in the industrial feedstock oil’s price. However, price fluctuations can offer some gains or risks to the economies in the relevant chain. The related relevant chain includes raw material for IC trays (modified polyphenylene oxide), as well as packaging and testing manufacturers. Therefore, the Organization of Petroleum Exporting Countries member countries and allies have met to negotiate production cuts to save oil prices. However, the breakdown of negotiations between Saudi Arabia and Russia increased oil volumes instead of decreasing them, causing an oil price war that brought oil prices to a new low, affecting the prices of plasticized products [15]. If we can predict IC tray demand, we can purchase raw materials at low prices to reduce costs and increase profits, which becomes an advantage of low-price inventory and can react rapidly to market changes.
The demand for IC trays has been increasing due to various strategies adopted by various countries to combat the epidemic. The price of raw materials for producing IC trays has changed because of the unstable international situation arising from different external factors. Thus, if we know the demand for IC trays in advance, we can purchase the number of raw materials that are required to produce them in advance, in order to rapidly respond to the changes in market demand and maintain the competitiveness of our company. Manufacturers’ requirements for the accuracy of forecast results are increasing as information technology improves. Companies can reduce costs, increase profits, and enhance customer satisfaction by improving demand-forecasting accuracy. Thus, this study proposes an error compensation mechanism that captures nonlinear data with a rolling long short-term memory (LSTM) and then compensates for the residuals with the rolling autoregressive moving average (ARMA) model, which can only reduce the forecasting model’s maintenance cost and reduce the time for model correction and training.

2. Materials and Methods

In this study, we propose a solution for an IC tray manufacturer’s data characteristics in Taiwan using industry–academia cooperation as an example, and verify the method’s feasibility. Recently, manufacturers of IC trays have been moving toward a less sample, more volume model, as customers’ requirements for IC shapes and sizes frequently change due to product diversification. The IC tray industry’s service characteristics emphasize order acceptance and speed of delivery, resulting in shorter production times. To reduce demand uncertainty’s impact and risk caused by market fluctuations, a systematic and highly accurate demand-forecasting model to assist in decision-making is urgently needed. The flowchart shown in Figure 1 was designed to include data gathering and pre-processing, the construction of a time series module, and rule-based explanations.

2.1. Data Collection and Preprocessing

Based on the top three products offered by Industry–Academy Cooperation from 1 January 2017 to 29 September 2019, the proposed approach was validated. Time series are continuous data, and the approach of adding zeros is employed to deal with missing values in this study. The main reason is that, from a practical perspective, missing values show that no products were sold on that day, and the system does not record what caused the missing values. Since the case company calculates sales data every day, IC trays calculate material purchases every week. This study converts data measurement frequency from daily to weekly to forecast raw material demand. Figure 2 shows the BGA 11.5X13 product converted from daily to weekly units.

2.2. Long Short-Term Memory Rolling Algorithm

Rolling forecasts are a type of dynamic forecasting. The main idea of this method is to add new data and delete old data simultaneously in a rolling manner while maintaining the length of the data. Forecasting results become more accurate through continuous learning, correction, and adjustment. The forecasting capability of an enterprise is thus enhanced so that it can respond promptly to customer needs [16]. Figure 3 illustrates a rolling forecasting process.
In 1997, Hochreiter proposed LSTM, a neural network model that solves the vanishing gradient problem for time series data [17]. LSTM regulates information flow, passes relevant information, and forecasts outcomes based on the incoming information. The previous hidden state is passed on to the next unit through the tanh activation function during the operation. A neural network’s hidden state acts as its memory, storing data it has seen in the past. A gate (forgetting, input, and output gates) allows the unit to decide whether to retain or delete information before passing it on to the next unit. Figure 4 shows the LSTM structure. The LSTM’s core, including input, output, and forgetting gates, is described below [18].
  • Forgetting gate: Based on the weights and the current input, the sigmoid function calculates a value f t between 0 and 1, determining which data will be retained and forgotten. Forgetting the data implies that they have been permanently removed from long-term memory. The forgetting gate’s equation is as follows:
    f t = σ ( W f · [ h t 1 , X t ] + b f )
    where σ is the sigmoid function, ∙ is the dot product, [ h t 1 , X t ] is a vector of input data, W f is the forgetting-gate weight, and b f is the forgetting-gate bias vector.
  • Input gate: Given the weight, hidden state, and current input, the sigmoid function calculates i t , a value between 0 and 1. This value is between 0 and 1, multiplied by the tanh activation function to calculate the cell status and determine which messages in long-term memory must be updated, including updated data and data that must be replaced to be employed in the future. The input gate’s equation is as follows
    i t = σ ( W i · [ h t 1 , X t ] + b i )  
    where σ is the sigmoid function, · is the dot product, [ h t 1 ,   X t ] is a vector of input data, W i is the input gate weight, and b i is the input gate bias vector.
    To update the long-term memory from C t 1 at t 1 to C t at t , tanh computes the following equation, where W c is the tanh weight and b c is the tanh bias vector.
    C ¯ t = t a n h ( W c · [ h t 1 , X t ] + b c )  
    If there are data that the forgetting gate and input gate wish to remove and add, respectively, the formula is as follows. First, to remove the forgetting gate’s data, f t   C t 1 is multiplied element by element, then the data that the input gate wishes to add are added element by element to update I t     C ¯ t to the long-term memory.
    C t = f t C t 1 + I t C ¯ t
  • Output gate: According to the weight, hidden state, and current output, the sigmoid function computes O t , which is between 0 and 1, and multiplies that value with the tanh activation function to determine which data should be removed from long-term memory C t to t + 1 for further use. The output gate equation is as follows, where σ is the sigmoid function, ∙ is the dot product operation, [ h t 1 ,   X t ] vector is the input data, W o is the output gate weight, and b O is the output gate bias vector.
    O t = σ ( W o · [ h t 1 , X t ] + b o )
    In Figure 4, at the top right, h t are the output data sent to the output layer, and at the bottom,   h t are the input data sent to t + 1 . The output data h t are obtained by multiplying the C t by the tanh activation function element by element after computing the output data O t .
    h t = O t tanh ( C t )

2.3. Individuals and Moving-Range (I-MR) Control Chart

Individual and moving-range charts comprise I and MR charts. I control charts are employed to demonstrate the means and variances of individual values, and MR control charts monitor the data’s variation. To effectively monitor all prediction processes, this study employs the I-MR control chart to determine whether error compensation is needed. If a point exceeds the control limits of the I-MR control chart, the error compensation mechanism is activated, and vice versa. The following are the steps to build an I-MR chart [19]:
Step 1.
Calculate the moving range (MR). The MR is calculated by calculating the distance between the 2 adjacent points.
M R i = | x i x i 1 |
Step 2.
Calculate average ( x ¯ ) and Moving Range average ( M R ¯ ) .
x ¯ = i = 1 k x i k   ,   M R ¯ = M R i k 1
where x i is the observed value of the i-th sample, k is the number of samples to establish the control chart.
Step 3.
Construct the control charts for the Individual and moving range.
Individual   control   chart = { U C L = x ¯ + 3 M R ¯ d 2 C L = x ¯ L C L = x ¯ 3 M R ¯ d 2
Moving   range   control   chart = { U C L = D 4 M R ¯ C L = M R ¯ L C L = D 3 M R ¯
where, d2, D3, and D4 are control chart constants [19].
It is important to remember that data that do not follow a normal distribution can affect the results when using I-MR control charts. Before using I-MR control charts, it is essential to check if the data are the normal distribution. It is recommended that the Box-Cox conversion method be used prior to plotting the I-MR chart if the data are not normally distributed.

2.4. Error Compensation Using ARMA Technique

It is difficult to develop an accurate model since the prediction error is uncertain [20]. The prediction results would be improved if an error compensation approach could be developed. More accurate prediction results can be obtained by compensating for the errors that are generated during training and combining them with the original prediction results [21].
Numerous scholars employ the ARMA model, which consists of autoregressive (AR) and moving average (MA), as an approach for error compensation, since it can determine the time series trend and predict future results. To study the evaporation process of alumina, Qian proposed particle-swarm optimization and ARMA error compensation [22]. Furthermore, microelectromechanical system gyroscopes are extensively employed in dynamic measurement devices. However, their measurement accuracy cannot be very accurate due to the external environment’s influence. Chen and Gao used the ARMA model for compensation to enhance the measurement accuracy of microelectromechanical system gyroscopes, which resulted in a 3.75% reduction in the measurement error [23].
In this study, the LSTM model was first developed. Then, the ARMA compensation model was developed by filtering the rolling prediction results by selected indicators (maximum, minimum, median, average, and the recent predicted results) and calculating the error from the actual values. This study employed the ARMA model since it does not need differencing, and the prediction results were corrected using the compensation values that were predicted by the ARMA model. The ARMA formula is as follows.
y t = + i = 1 p φ i y t i + ε t + i = 1 q θ i ε t i          
where y t is the time series observation at time t; α is the constant; φ i is the parameter of y t i when the lag time is i; ε t is the white noise at time t; p is the maximum lag-time step; ε t i is the white noise when the lag-time step is i; θ i is the constant of ε t i when the lag-time step is i; q is the maximum lag-time step; and ARMA (p,q) is the ARMA model with the maximum lag-time step of the self-regression term p and the maximum lag-time step of the MA term q.
In constructing an ARMA model, the first step is to visualize the data and determine whether there are missing values. If there are missing values, this study will employ the zero-completion approach. Thereafter, the ARMA model’s parameters (p,q) are estimated, and the combination of parameters is obtained from the autocorrelation and partial autocorrelation functions., Along with the model, this is validated to determine whether the fitted model’s residual series is white noise. In the residual sequence, white noise is checked and the Akaike Information Criterion is employed to compute different levels and select the (p,q) combination with the lowest akaike information criterion (AIC) value. Finally, (p,q) is substituted into the ARMA model for prediction. Based on the error compensation, the prediction steps are as follows:
Step 1.
Calculate the mean absolute percentage error (MAPE), mean absolute error (MAE), and root mean squared error (RMSE) of each indicator, and select the one with the lowest error rate from the LSTM model.
Step 2.
Determine whether compensation is required by calculating the indicators’ errors and plotting the I-MR control chart.
Step 3.
In rolling LSTM, the number of products are predicted and multiple predictions are obtained for the test set.
Step 4.
The prediction result is filtered to y i based on the index that is selected in Step 1.
Step 5.
If compensation is needed, develop a rolling ARMA compensation model and calculate the compensation value A i .
Step 6.
To obtain the compensation result y i , add the compensation value A i to the filtered prediction result y i :
y i = A i + y i .
Step 7.
Calculate the error E i between the actual value y and y i , i.e., E i = y i y , which are the data for the next compensation model.

2.5. Data Partitioning and Parameter Setting

Figure 5 shows the rolling prediction process. As a first rollup, the data were divided into 118 weeks of yellow data from 1 January 2017 to 7 April 2019, five weeks of purple data from 14 April 2019 to 12 May 2019, and five weeks of green data from 19 May 2019 to 16 June 2019 for LSTM model training, indicators establishment, and compensation model building, respectively. Forecasting was based on pink section data from 23 June 2019 to 21 July 2019. The second rollup is based on the first compensation model values (19 May 2019–16 June 2019), a compensation model that was built using the first prediction (23 June 2019–21 July 2019), and a prediction of the last five data (28 July 2019–25 August 2019). This third rollup consists of an indicator that was constructed using the values from the second compensation model 23 June 2019–21 July 2019) and an indicator to the compensation model’s construction using the results from the second prediction (23 June 2019–21 July 2019). The third rollup uses the second compensation model (23 June 2019–21 July 2019), the second forecast (28 July 2019–25 August 2019) and predicts the last five data (1 September 2019–29 September 2019), and so on. In this study, only three cuts and rolls were made to the data to test the proposed approach.
After the training samples are calculated, Table 1 shows the estimated LSTM rolling pattern parameters.

2.6. Model Performance Indicator

Different indicators are used to evaluate a model’s predictive power. Leonardo designed a forecasting model for sugarcane yield and compared the root mean squared error, mean absolute error, and mean absolute percentage error as indicators of the model’s predictive ability [24]. To predict the number of COVID-19 confirmations in Iran, Nasrin et al., developed a predictive model and evaluated the model’s performance using RMSE and MAE. The findings show that Iran was probably the most affected and needed more preventative measures [25]. Chang et al., developed a prediction model for air pollution and evaluated the model’s performance using the MAE, RMSE, and MAPE. The findings indicated that the model considerably enhanced the prediction’s accuracy [26]. MAE, RMSE, and MAPE, commonly used evaluation indicators in practice, are employed in this study as the criteria for model evaluation. This formula is defined as follows. When the index calculation result is small, predictability is better.
  • MAE: The mean absolute error between the predicted y t ^ and actual values y t of each datum is measured using the following formula:
      MAE = 1 k t = 1 k | y t y t ^ |
  • RMSE: A measure of prediction error ( y t y t ^ ) standard deviation. The formula is as follows:
    RMSE = 1 k t = 1 k ( y t y t ^ ) 2
  • MAPE: This model overcomes the limitations of both mean absolute deviation (MAD) and RMSE using the relative prediction error of each data, which has overly large calculation results due to the large value of each data.
    MAPE = 1 k t = 1 k | y t y t ^ | y t × 100 %    
    y t is the actual value, y t ^ is the predicted value, and k is the sample size. Mean square error (MSE) and MAPE are unaffected by unit and value sizes and are evaluated objectively; the larger the sample size, the higher the RMSE.

3. Analysis Results

A validation of the proposed methodology is provided in Section 3.1, Section 3.2 and Section 3.3 by analyzing and discussing the top three sales volumes of the product (BGA 11.5X13, BGA 8X13 and BGA 8X12.5).

3.1. Case Study of BGA 11.5X13

First, we employ rolling training with a fixed prediction window size, followed by initial training after every prediction. The initial training value is deleted upon each prediction and a new piece of data is added to the model. This is repeated until the specified period’s data have been predicted. The LSTM rolling model is employed for the 14 April 2019 to 12 May 2019 results, then the numerical results are computed based on maximum (Max), minimum (Min), median (Med), mean (Mean), and recent results (New), which are more commonly employed in statistics, and the values are filtered, as shown in Table 2.
RMSE, MAPE, and MAE were computed for each indicator, and the lowest error indicator was selected. For inconsistent results, as shown in Table 3, the median was selected as the indicator after RMSE calculation, the minimum after MAPE calculation, and the indicator with the most evaluated indicators was selected as the final decision, thus, the minimum (Min) will be selected as the next screening value.
In Figure 6, the five errors in the selected metrics’ minimum values are plotted on the I-MR control chart, and if there are phenomena that exceed the control limits, the subsequent prediction results will be compensated.
As there is a condition of exceeding the control limits, an error compensation is required. Thereafter, the error is substituted into the ARMA model to obtain the compensation value A 1 . To obtain the compensation result y 1 , the compensation value A 1 is then summed with the prediction results after 23 June numerical screening. According to a previous rolling concept, the first 19 May error result is deleted, and a new 23 June error result is added. The new error is introduced in the compensation model from the result y 1 resulting from the 23 June compensation and the actual error E 1 . Then, the compensation value A 2 is compensated to the selected prediction of 30 June at the end of the week from 23 June to 21 July. Figure 7 shows a schematic representation of rolling error compensation.
A rolling LSTM was used for the prediction’s last five weeks, and the selected indicators’ values filtered the results, and Table 4 shows the calculated compensation values. The ARMA (p,q) value in Table 4 is (0,0) since this series is stationary, and ARMA is the most extensively employed model for fitting stationary series.
Table 5 summarizes the three rolling prediction results, in which the errors are computed by subtracting the actual values from the predicted values. In Table 5, the rolling prediction’s error with indicator filtering (LSTM-Min) is smaller than the LSTM prediction. If the previous judgment requires error compensation (LSTM-ARMA), the error remains smaller than in the previous two approaches.

3.2. Case Study of BGA 8X13

Table 6 summarizes the results of the BGA 8X13’s three rolls, showing that the median was selected as the filtering indicator, and since the I-MR control chart did not exceed the control limits, no error correction was necessary. The mean was selected as the filtering indicator for the second roll, and no error compensation was required since the I-MR control chart did not exceed the control limits. The median was selected as the filtering indicator in the third roll, and since the I-MR control chart did not exceed the control limits, an error compensation was necessary. In the third rollover, the median was selected as the filter and the I-MR control chart exceeded the control limits, so error correction was necessary.
Table 7 summarizes the three rolling prediction results by subtracting the actual values from the predicted values and by calculating the errors. In Table 7, the rolling prediction’s error combined with the indicator filtering approach is smaller than the general LSTM prediction’s error. If the previous judgment requires error compensation, the error after compensation is smaller than the previous two approaches’ errors.

3.3. Case Study of BGA 8X12.5

Table 8 summarizes the three BGA 8X12.5 rollover results. It shows that the first rollover selects the maximum value as the numerical filter, and the I-MR control chart does not exceed the control limits; therefore, no error compensation is necessary. In the second rollover, the median is selected as the numerical filter, and the I-MR control chart does not exceed the control limits, thus, error compensation is necessary. In the third rollover, the nearest result is the numerical filter, and error compensation is not required if the I-MR control chart does not exceed the control limits.
Table 9 summarizes the results of the three rolling predictions, where the error is calculated by subtracting the actual values from the predicted values. In Table 9, the rolling prediction combined with the indicator filter is more accurate than the general LSTM prediction; if the previous judgment requires error compensation, the error after compensation is smaller than that of the previous two approaches.

3.4. Model Evaluation

These three cases study are compared for the traditional LSTM approach, the rolling LSTM prediction using the indicator filtering numerical approach, and the LSTM-ARMA compensation approach. Table 10 shows the evaluation results using the RMSE, the MAPE, and the MAE.
There is a decreasing trend in terms of the RMSE; however, the rolling LSTM prediction combined with pointer screening offers a better prediction result than LSTM. If error compensation is required, the LSTM-ARMA approach results indicate a significant decrease in RMSE compared with the aforementioned approaches. In terms of MAPE, there is a decreasing trend, however, in most cases, the rolling LSTM prediction combined with the screening of indicators presents a better prediction result than the LSTM, probably because the demand for the case company’s orders is more volatile and the model cannot be more accurate. The MAE trend is declining, but most of the rolling LSTM prediction combined with indicator filtering offers better predictions than the LSTM, perhaps because demands for the company’s orders are more volatile, and the model cannot predict more accurately. If the error compensation is required, the MAE of LSTM-ARMA tends to decrease compared with the aforementioned approaches.

4. Discussion and Conclusions

The prediction model results’ accuracy improved due to the demand of competitors in the market. The forecasting model that is developed in this study is based on actual data from the case companies and is compared with the way that it was previously used. This study had three primary findings. At one time, forecasts were performed for one point at a time, i.e., tomorrow’s results were predicted as they would be. We forecast five points at a time, i.e., tomorrow, the day after tomorrow, the day after tomorrow, the day after tomorrow, etc. In the case of a later forecast, there is not only one forecast, but also the findings from the current period, the previous period, and the previous period before that. Several values are filtered using indicators, instead of using the most recent result as the final forecast. Second, in the past, we would pause the model and reload the data for new model training if we encountered poor prediction results, but this process was tedious and complicated. Finally, in the past, error compensation was activated regardless of the prediction results. The proposed error compensation mechanism is based on the I-MR control chart that is plotted by the difference between the indicators’ filtered and actual values, and the need for error compensation is determined as opposed to simply activating compensation regardless of the prediction results.
LSTM, rolling LSTM combined with an index, and LSTM-ARMA are compared, and the LSTM-ARMA approach proves to be more accurate when the error compensation is required. Furthermore, the rolling LSTM combined with the index offers better predictions than the firm’s rule of thumb and LSTM when no residential error compensation is required. Table 11 shows the evaluation of the ARIMA model, demonstrating that the LSTM-ARMA method is more accurate when error compensation is required.
As a contribution to academic work, we have proposed an approach to determine whether an error compensation is necessary using rolling data for forecasting. In cases of model inaccuracy, the original model can be used without model correction. Combining LSTM and ARMA rolling forecasting approaches can generate more accurate forecasting results and eliminate the bottleneck problems that have been experienced in the past. In the past, regarding industry contribution, models were trained and tested before going online, and once they were employed, they gradually became inaccurate and needed to be recorrected, which took a long time. This study develops a model that uses error compensation to reduce the time and cost of pausing the model correction when past forecasts are inaccurate and compares it with empirical rules that are employed by companies to offer more accurate forecasting results.

Author Contributions

Conceptualization, C.-C.W. and H.-T.C.; methodology, C.-C.W. and H.-T.C.; validation, C.-C.W. and H.-T.C.; formal analysis, C.-C.W. and H.-T.C.; data curation, H.-T.C. and C.-H.C.; writing—original draft preparation, C.-C.W., H.-T.C. and C.-H.C. writing—review and editing, C.-C.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was financially supported by the Ministry of Science and Technology (MOST) of Taiwan (110-2221-E-131-027-MY3).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ICIntegrated circuit
AQIAir quality index
ARMAAutoregressive moving average
ARAutoregressive
MAMoving average
LSTMLong short-term memory
I-MRIndividuals and moving range
UCLUpper control limits
CLCenterline limits
LCLLower control limits
MADMean absolute deviation
MAEMean absolute error
MAPEMean absolute percentage error

References

  1. Guo, D.Y.; Shi, H.Z.; Qian, Y.P.; Lv, M.; Li, P.G.; Su, Y.L.; Tang, W.H. Fabrication of β-Ga2O3/ZnO heterojunction for solar-blind deep ultraviolet photodetection. Semicond. Sci. Technol. 2017, 32, 03LT01. [Google Scholar] [CrossRef]
  2. Kumar, V.; Dwivedi, S.K. Hexavalent chromium stress response, reduction capability and bioremediation potential of Trichoderma sp. isolated from electroplating wastewater. Ecotoxicol. Environ. Saf. 2019, 185, 109734. [Google Scholar] [CrossRef]
  3. Wang, C.-C.; Chien, C.-H.; Trappey, A.J.C. On the Application of ARIMA and LSTM to Predict Order Demand Based on Short Lead Time and On-Time Delivery Requirements. Processes 2021, 9, 1157. [Google Scholar] [CrossRef]
  4. Aparna, S. Long Short-Term Memory and Rolling Window Technique for Modeling Power Demand Prediction. In Proceedings of the International Conference on Intelligent Computing and Control Systems, Madurai, India, 14–15 June 2018; pp. 1675–1678. [Google Scholar]
  5. Park, S.; Lehto, X.; Lehto, M. Self-service technology kiosk design for restaurants: An QFD application. Int. J. Hosp. Manag. 2020, 92, 102757. [Google Scholar] [CrossRef]
  6. Bi, J.-W.; Li, H.; Fan, Z.-P. Tourism demand forecasting with time series imaging: A deep learning model. Ann. Tour. Res. 2021, 90, 103255. [Google Scholar] [CrossRef]
  7. Tanizaki, T.; Hoshino, T.; Shimmura, T.; Takenaka, T. Demand forecasting in restaurants using machine learning and statistical analysis. Procedia CIRP 2019, 79, 679–683. [Google Scholar] [CrossRef]
  8. Liu, W.; Liu, W.D.; Gu, J. Forecasting oil production using ensemble empirical model decomposition based Long Short-Term Memory neural network. J. Pet. Sci. Eng. 2020, 189, 107013. [Google Scholar] [CrossRef]
  9. Yu, P.; Yan, X. Stock price prediction based on deep neural networks. Neural Comput. Appl. 2019, 32, 1609–1628. [Google Scholar] [CrossRef]
  10. Mayr, J.; Blaser, P.; Ryser, A.; Hernández-Becerro, P. An adaptive self-learning compensation approach for thermal errors on 5-axis machine tools handling an arbitrary set of sample rates. CIRP Ann. 2018, 67, 551–554. [Google Scholar] [CrossRef]
  11. Liang, X.; Zhang, S.; Wu, Y.; Xing, J.; He, X.; Zhang, K.M.; Wang, S.; Hao, J. Air quality and health benefits from fleet electrification in China. Nat. Sustain. 2019, 2, 962–971. [Google Scholar] [CrossRef]
  12. Ma, J. Telework Triggered by Epidemic: Effective Communication Improvement of Telecommuting in Workgroups during COVID-19. Am. J. Ind. Bus. Manag. 2021, 11, 202. [Google Scholar] [CrossRef]
  13. Vagal, A.; Reeder, S.B.; Sodickson, D.K.; Goh, V.; Bhujwalla, Z.M.; Krupinski, E.A. The impact of the COVID-19 pandemic on the radiology research enterprise: Radiology scientific expert panel. Radiology 2020, 296, E134–E140. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Jiang, J.; Marsh, T.L.; Tozer, P.R. Policy induced price volatility transmission: Linking the US crude oil, corn and plastics markets. Energy Econ. 2015, 52, 217–227. [Google Scholar] [CrossRef]
  15. Hu, C.; Liu, X.; Pan, B.; Chen, B.; Xia, X. Asymmetric impact of oil price shock on stock market in China: A combination analysis based on SVAR model and NARDL model. Emerg. Mark. Financ. Trade 2017, 54, 1693–1705. [Google Scholar] [CrossRef]
  16. O’Connor, M.; Remus, W.; Griggs, K. Does updating judgmental forecasts improve forecast accuracy? Int. J. Forecast. 2000, 16, 101–109. [Google Scholar] [CrossRef]
  17. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  18. Abbasimehr, H.; Shabani, M.; Yousefi, M. An optimized model using LSTM network for demand forecasting. Comput. Ind. Eng. 2020, 143, 106435. [Google Scholar] [CrossRef]
  19. Montgomery, D.C. Introduction to Statistical Quality Control; John Wiley & Sons: New York, NY, USA, 2009. [Google Scholar]
  20. He, X.; Deng, K.; Wang, X.; Li, Y.; Zhang, Y.; Wang, M. Lightgcn: Simplifying and powering graph convolution network for recommendation. In Proceedings of the 43rd International ACM SIGIR conference on research and development in Information Retrieval, Xi’an, China, 25–30 July 2020; pp. 639–648. [Google Scholar]
  21. Liang, C.-J.; Liang, J.-J.; Jheng, C.-W.; Tsai, M.-C. A rolling forecast approach for next six-hour air quality index track. Ecol. Inform. 2020, 60, 101153. [Google Scholar] [CrossRef]
  22. Qian, X. Parameter prediction based on Improved Process neural network and ARMA error compensation in Evaporation Process. In IOP Conference Series: Earth and Environmental Science; IOP Publishing: Bristol, UK, 2018; Volume 108, p. 022078. [Google Scholar]
  23. Chen, M.M.; Gao, G.W. Research on MEMS gyroscope random error compensation algorithm based on ARMA model. Appl. Mech. Mater. 2014, 602, 891–894. [Google Scholar] [CrossRef]
  24. Maldaner, L.F.; de Paula Corrêdo, L.; Canata, T.F.; Molin, J.P. Predicting the sugarcane yield in real-time by harvester engine parameters and machine learning approaches. Comput. Electron. Agric. 2021, 181, 105945. [Google Scholar] [CrossRef]
  25. Talkhi, N.; Fatemi, N.A.; Ataei, Z.; Nooghabi, M.J. Modeling and forecasting number of confirmed and death caused COVID-19 in IRAN: A comparison of time series forecasting methods. Biomed. Signal Process. Control 2021, 66, 102494. [Google Scholar] [CrossRef] [PubMed]
  26. Chang, Y.S.; Chiao, H.T.; Abimannan, S.; Huang, Y.P.; Tsai, Y.T.; Lin, K.M. An LSTM-based aggregated model for air pollution forecasting. Atmos. Pollut. Res. 2020, 11, 1451–1463. [Google Scholar] [CrossRef]
Figure 1. The analysis framework flowchart.
Figure 1. The analysis framework flowchart.
Mathematics 10 02158 g001
Figure 2. Timeline conversion of the difference graph. (a) BGA 11.5X13 timeline is daily; (b) BGA 11.5X13 timeline is one week.
Figure 2. Timeline conversion of the difference graph. (a) BGA 11.5X13 timeline is daily; (b) BGA 11.5X13 timeline is one week.
Mathematics 10 02158 g002
Figure 3. Rolling forecast schematic.
Figure 3. Rolling forecast schematic.
Mathematics 10 02158 g003
Figure 4. LSTM operational architecture.
Figure 4. LSTM operational architecture.
Mathematics 10 02158 g004
Figure 5. The rolling prediction process.
Figure 5. The rolling prediction process.
Mathematics 10 02158 g005
Figure 6. I-MR control chart for BGA 11.5X13 data for the first prediction roll.
Figure 6. I-MR control chart for BGA 11.5X13 data for the first prediction roll.
Mathematics 10 02158 g006
Figure 7. Schematic representation of rolling error compensation.
Figure 7. Schematic representation of rolling error compensation.
Mathematics 10 02158 g007
Table 1. LSTM parameter setting.
Table 1. LSTM parameter setting.
ParameterSetting Values
Time step1
Batch size 1
Loss functionMSE
Activation functionSigmoid
Neurons1
Hidden layers4
Epochs100
OptimizerAdam
Table 2. BGA 11.5X13 error calculation results of each indicator.
Table 2. BGA 11.5X13 error calculation results of each indicator.
Date14 April 201921 April 201928 April 20195 May 201912 May 2019
Actual value37,50027,20033,60034,20012,000
Forecasting values37,50827,29433,73534,32411,725
27,28933,68134,26411,829
33,63334,22611,758
34,39711,791
11,839
Max37,50827,29433,73534,39711,839
Min37,50827,28933,63334,22611,725
Med37,50827,29133,68134,29411,791
Mean37,50827,29133,68334,30311,789
New37,50827,28933,63334,39711,839
Table 3. BGA 11.5X13 1st indicator calculation.
Table 3. BGA 11.5X13 1st indicator calculation.
IndicatorRMSEMAPEMAE
Max135.28001.6083119
Min130.77160.978086
Med116.00701.218396
Mean118.70211.253999
New121.22531.285697
Table 4. The result of calculating compensation value at different time points.
Table 4. The result of calculating compensation value at different time points.
DateARMA (p,q)Compensation
23 June 2019ARMA (0,0)307
30 June 2019ARMA (0,0)203
7 July 2019ARMA (0,0)141
14 July 2019ARMA (0,0)79
21 July 2019ARMA (0,0)40
Table 5. BGA 11.5X13 rolling forecast results summary.
Table 5. BGA 11.5X13 rolling forecast results summary.
1DateActual ValueLSTMErrorLSTM-MinErrorLSTM-ARMAError
23 June 37,10036,894−20636,934−16637,14141
30 June 25,80025,437−36325,567−23325,791−9
7 July 20,20019,708−49219,999−20120,157−43
14 July 32,60032,476−12432,523−7732,61414
21 July 41,30041,136−16441,176−12441,227−73
2DateActual valueLSTMErrorLSTM-MinErrorLSTM-ARMAError
28 July 39,70039,94124139,491−20939,571−129
4 August15,40015,4929215,325−7515,399−1
11 August15,60015,76416415,572−2815,61616
18 August 19,70019,7606019,648−5219,674−26
25 August 250026611612452−482488−12
3DateActual valueLSTMErrorLSTM-MinErrorLSTM-ARMAError
1 September 16,40016,352−4816,378−2216,41212
8 September 16,00016,35235216,21921916,209209
15 September 37,60037,480−12037,70610637,65353
22 September30,20030,037−1633034614630,300100
29 September36,10036,30920936,24614636,19999
Table 6. BGA 8X13 three rolling results.
Table 6. BGA 8X13 three rolling results.
123
IndicatorMedMeanMin
I-MR control chart Mathematics 10 02158 i001 Mathematics 10 02158 i002 Mathematics 10 02158 i003
Error compensationNot requiredNot requiredRequired
Table 7. BGA 8X13 rolling forecast results summary.
Table 7. BGA 8X13 rolling forecast results summary.
1DateActual ValueLSTMErrorLSTM-MedErrorLSTM-ARMAError
23 June 36,90036,395−50536,595−305
30 June 40,70040,487−21340,576−124
7 July 50,10050,074−2650,11717
14 July 59,00058,989−1158,987−13
21 July 49,80049,770−3049,914114
2DateActual valueLSTMErrorLSTM-MeanErrorLSTM-ARMAError
28 July 16,00015,978−2215,988−12
4 August23,40023,321−7923,335−65
11 August69,20069,024−17669,302102
18 August 36,70036,7939336,75252
25 August 31,90031,9313131,863−37
3DateActual valueLSTMErrorLSTM-MinErrorLSTM-ARMAError
1 September 73,80073,638−16273,758−4274,011211
8 September 51,20051,2878751,020−18051,26969
15 September 45,80046,16936945,457−34345,602−198
22 September49,60049,95935949,371−22949,520−80
29 September29,50029,476−2429,269−23129,385−115
Table 8. BGA 8X12.5 three rolling results.
Table 8. BGA 8X12.5 three rolling results.
123
IndicatorMaxMedNew
I-MR control chart Mathematics 10 02158 i004 Mathematics 10 02158 i005 Mathematics 10 02158 i006
Error compensationNot requiredRequiredNot required
Table 9. BGA 8X12.5 rolling forecast results summary.
Table 9. BGA 8X12.5 rolling forecast results summary.
1DateActual ValueLSTMErrorLSTM-MaxErrorLSTM-ARMAError
23 June 20,90020,735−16520,93535
30 June 22,50022,334−16622,634134
7 July 1400143939144545
14 July 32,40032,405532,517117
21 July 13,60013,579−2113,704104
2DateActual valueLSTMErrorLSTM-MedErrorLSTM-ARMAError
28 July 10,90011,01411411,00410410,91616
4 August10,80010,99919910,98618610,86464
11 August21,40021,52912921,52912921,340−60
18 August 11,70011,663−3711,669−3111,71212
25 August 23,90023,835−6523,858−4223,877−23
3DateActual valueLSTMErrorLSTM-NewErrorLSTM-ARMAError
1 September 31,40031,4121231,42222
8 September 100366266268168
15 September 11,30011,212−8811,34141
22 September23,20022,924−27623,310110
Table 10. Comparison of RMSE, MAPE, and MAE results for three different IC trays.
Table 10. Comparison of RMSE, MAPE, and MAE results for three different IC trays.
IC Trays CodeRMSEMAPE MAE
LSTMLSTM-IndexLSTM-ARMALSTMLSTM-IndexLSTM-ARMALSTMLSTM-IndexLSTM-ARMA
BGA 11.5X131302.830169.33642.7461.0350.5760.116269.8160.236.0
2156.820105.09259.3851.8000.6760.206143.682.436.8
3205.357143.006115.3600.7860.5350.426178.4127.894.7
BGA 8X131245.801156.112NA0.4050.283NA157.0114.6NA
297.27461.298NA0.2160.151NA80.253.5NA
3244.717226.987147.0450.4000.4800.281200.2205.0134.6
BGA 8X12.51106.55395.615NA0.8971.021NA79.287.0NA
2122.337113.82341.5330.8160.7440.244108.898.435.0
3194.636110.687NA53.77233.907NA165.695.6NA
Notes: NA stated that it is not necessary to activate the compensation mechanism.
Table 11. Results of the comparison between traditional ARIMA and our proposed LSTM-ARMA for three IC trays.
Table 11. Results of the comparison between traditional ARIMA and our proposed LSTM-ARMA for three IC trays.
IC Trays CodeARIMALSTM-ARMA (In This Paper)
RMSEMAPEMAERMSEMAPEMAE
BGA 11.5X1313507.8830.341301.842.7460.11636.0
23707.60.21241.559.3850.20636.8
3424.530.53180.4115.3600.42694.7
BGA 8X1319108.20428.3160.2NANANA
29332.7426.250.2NANANA
39441.8319.8297.2147.0450.281134.6
BGA 8X12.511061.46931.21470.2NANANA
21204.7332.6368.841.5330.24435.0
31509.2133.97265.6NANANA
Notes: NA stated that it is not necessary to activate the compensation mechanism.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wang, C.-C.; Chang, H.-T.; Chien, C.-H. Hybrid LSTM-ARMA Demand-Forecasting Model Based on Error Compensation for Integrated Circuit Tray Manufacturing. Mathematics 2022, 10, 2158. https://doi.org/10.3390/math10132158

AMA Style

Wang C-C, Chang H-T, Chien C-H. Hybrid LSTM-ARMA Demand-Forecasting Model Based on Error Compensation for Integrated Circuit Tray Manufacturing. Mathematics. 2022; 10(13):2158. https://doi.org/10.3390/math10132158

Chicago/Turabian Style

Wang, Chien-Chih, Hsin-Tzu Chang, and Chun-Hua Chien. 2022. "Hybrid LSTM-ARMA Demand-Forecasting Model Based on Error Compensation for Integrated Circuit Tray Manufacturing" Mathematics 10, no. 13: 2158. https://doi.org/10.3390/math10132158

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop