You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

15 April 2024

An Intelligent Thermal Compensation System Using Edge Computing for Machine Tools

,
,
,
,
and
1
Department of Computer Science, Tunghai University, Taichung City 407224, Taiwan
2
Department of Informatics, Krida Wacana Christian University, Jakarta 11470, Indonesia
3
Industrial Technology Research Institute, Hsinchu 310401, Taiwan
4
Department of Mechanical Engineering, National Chin-Yi University of Technology, Taichung City 411030, Taiwan
This article belongs to the Special Issue Innovations in AI and ML-Based Techniques for Image and Video Analysis from Sensors

Abstract

This paper focuses on the use of smart manufacturing in lathe-cutting tool machines, which can experience thermal deformation during long-term processing, leading to displacement errors in the cutting head and damage to the final product. This study uses time-series thermal compensation to develop a predictive system for thermal displacement in machine tools, which is applicable in the industry using edge computing technology. Two experiments were carried out to optimize the temperature prediction models and predict the displacement of five axes at the temperature points. First, an examination is conducted to determine possible variances in time-series data. This analysis is based on the data obtained for the changes in time, speed, torque, and temperature at various locations of the machine tool. Using the viable machine-learning models determined, the study then examines various cutting settings, temperature points, and machine speeds to forecast the future five-axis displacement. Second, to verify the precision of the models created in the initial phase, other time-series models are examined and trained in the subsequent phase, and their effectiveness is compared to the models acquired in the first phase. This work also included training seven models of WNN, LSTNet, TPA-LSTM, XGBoost, BiLSTM, CNN, and GA-LSTM. The study found that the GA-LSTM model outperforms the other three best models of the LSTM, GRU, and XGBoost models with an average precision greater than 90%. Based on the analysis of training time and model precision, the study concluded that a system using LSTM, GRU, and XGBoost should be designed and applied for thermal compensation using edge devices such as the Raspberry Pi.

1. Introduction

With the popularization of Industry 4.0, the manufacturing industry is gradually transforming its manufacturing mindset from mass production to customized precision manufacturing, achieving the ultimate goal of smart manufacturing. The core of smart manufacturing is the integration and application of digital and physical systems, namely cyber-physical systems (CPS) and the Internet of Things (IoT). Industry 4.0 enables high flexibility in the development, diagnosis, maintenance, and operation of automation systems. When developing these systems, people can choose the best suppliers from many components, modules, and service providers. Some diagnostics can be performed by the user, and access to “big data” helps with automation. Information can be retrieved on demand, used intelligently, and correlated to achieve automatic diagnosis. Components can be ordered automatically from the cheapest manufacturer, solving the skill shortage problem [1]. Meanwhile, recent developments show that the concept of Industry 5.0 is beginning to emerge along with the development of artificial intelligence. Data-driven decision systems help humans and robots optimize production scheduling, running equipment, forecasting breakdowns, and evaluating industrial performance. Moreover, the collaboration between humans and machines allows for a faultless environment where the versatility of humans and the precision of machines may achieve production performance that is free from errors and optimized [2].
In the post-pandemic era, the manufacturing industry uses remote access to data and cloud-based processing of large amounts of data, but this method has some drawbacks, such as data transmission delays, data storage security problems, and large-scale use of network bandwidth [3]. Therefore, the combination of intelligent manufacturing and edge computing will greatly reduce the labor cost of the manufacturing industry and improve the security of its data, while the combination of edge computing with small devices will also greatly improve the portability of the system, helping factories move toward comprehensive intelligent manufacturing [4].
In the era of global trade, product quality and product precision have become increasingly important, so it is very important to control the tooling errors of manufactured products. Among machine tool errors, thermal error is also one of the most influential factors. Excess temperature will cause thermal deformation of the machine tool. There are two main sources of machine tool thermal deformation, namely internal heat sources and external heat sources. External heat sources come mainly from solar heat radiation or the influence of ambient temperature, while internal heat sources are mainly the heat generated by the machine tool and its components during operation. The traditional method of thermal compensation is to use coolant to cool the interior of the machine, but this method does not completely improve the thermal error of the machine tool [5]. Based on reviews by Konstantinidis et al. [6], this research could be carried out at level 4 that covers advanced features such as intelligent object identification capabilities and metadata transformation techniques to address interoperability issues. In particular, classifiers with less data are typically trained via (semi)supervised machine learning. On the other hand, distributed learning techniques enhance the effectiveness of models while ensuring privacy, as seen in Federated Learning. The metadata or decisions are shown on mobile interfaces or sent to Industrial IoT platforms for further utilization or interaction with other systems.
This paper aims to use deep learning models to predict the axial displacement of machine tools and integrate them into a lightweight edge computing device that can be directly connected to the machine tool to improve the precision of the machining. The study is divided into two parts. In the first part, data recorded during machine tool operation, such as spindle temperature, speed, and torque, are used. They are used to train and predict future temperature changes. In the second part, data of multiple temperature point changes, speed changes, and corresponding five-axis displacement changes are used to train and predict the five-axis displacement. Finally, a Raspberry Pi machine tool thermal compensation system is developed to predict the future five-axis displacement. This study also develops an intelligent machine tool thermal compensation system on the Raspberry Pi Edge device. The system will collect temperature data points from the machine tool and predict future five-axis displacement. Therefore, the contributions of this study are listed as follows:
  • This study designs two experiments to develop an appropriate prediction model for time-series data on lathe machine tools.
  • This study designs a combination of genetic algorithm (GA) and the LSTM model to improve the accuracy of predicting thermal displacement in turning machine tools.
  • This study compares multiple time-series models using the same set of lathe data and integrates their final prediction results.
  • This study develops an intelligent compensation system for thermal displacement in machine tools using Qt Creator and runs the system on the edge computing side of the Raspberry Pi and the cloud computing side of Windows operating systems.

3. Research Methodology and Framework

In this section, the research framework of the study will be introduced, which includes a detailed description of the GA-LSTM model used in Experiment 2.

3.1. Research Framework

Research processes have two main steps. First, feasible time-series variations are investigated based on the data collected of time, speed, torque, and temperature changes at different positions of the machine tool. Subsequently, using the identified feasible machine-learning models, the study analyzes different cutting conditions, temperature points, and speeds of the machine to predict the future 5-axis displacement. To ensure the accuracy of the models developed in the first step, additional time-series models are explored and trained in the second step, and their performance is compared with the models obtained in the first step.
Upon completion of the aforementioned research, a human-machine interface is built using Qt Creator in both Raspberry Pi and Windows environments, allowing users to predict real-time thermal displacements of the machine. Figure 1 illustrates the complete research framework.
Figure 1. Complete Research Framework.

3.2. Data Preprocessing

Due to the large number of temperature points in the data set, the prediction of the model may suffer from overfitting. Therefore, in this experiment, Pearson’s correlation analysis was performed to select the eight top temperature points that have a significant contribution to the displacement of the five axes. Pearson’s correlation analysis is used to explore the linear relationship between two continuous variables (x and y). If the absolute value of the correlation coefficient between the two variables is large, it indicates a high degree of mutual covariance. Generally, if the two variables are positively correlated, an increase in x will correspond to an increase in y. On the contrary, if the two variables are negatively correlated, an increase in x will result in a decrease in y. The formula for the Pearson correlation coefficient is as follows, where r represents the correlation coefficient and  C O V  is the covariance.
r ( x , y ) = C O V ( x , y ) S x S y = i = 1 n ( x i x ¯ ) ( y i y ¯ ) i = 1 n ( ( x i x ¯ ) 2 ) i = 1 n ( y i y ¯ ) 2

3.3. AI Model for Experiment 1

This experiment used the LSTM, GRU, TCN, and Ensemble Stacked models. Let us provide an overview and description of each model’s architecture.
The LSTM model requires the input data to be in a specific format. In this study, the data were reorganized to meet the input requirements of the LSTM model. The data were transformed into groups of 10 records, with each record consisting of 5 attribute values. The input data were then reshaped into the format [n, 10, 5], where n represents the total number of samples. The output data were shaped as [n, ].
The GRU model requires that the input data be in a specific format. In this study, the data were reorganized accordingly. The data were transformed into groups of ten samples, each sample consisting of five attributes. The input data were reshaped in the shape of [n, 10, 5], where n represents the total number of samples. The output data had the shape of [n, ].
The TCN (Temporal Convolutional Network) model also requires that the input data be in a specific format. For this study, the data were reorganized accordingly. The data were transformed into groups of ten samples, each sample consisting of five attributes. The input data were reshaped in the shape of [n, 10, 5], where n represents the total number of samples. The output data had the shape of [n, ].

3.4. AI Model for Experiment 2

In addition to the time-series models used in Experiment 1, this experiment also included the training of seven additional time-series models of different types. They are WNN, LSTNet, TPA-LSTM, XGBoost, BiLSTM, CNN, and GA-LSTM. The following describes its model content.
The usage condition for the BiLSTM model is that the input data should adhere to the input format of the BiLSTM model, restructuring the data set accordingly. For this experiment, the data were transformed into groups, each group consisting of one record that contains nine attribute values. The input data were then reshaped into the format [n, 1, 9], where n represents the total number of records. The output data format is [n, ].
The usage condition for the LSTNet model is that the input data should conform to the input format required by the LSTNet model. It necessitates reorganizing the dataset accordingly. For this experiment, the data were transformed into groups, each group consisting of 20 records that contain nine attribute values. The input data were then reshaped into the format [n, 20, 9], where n represents the total number of records. The output data format is [n, ].
The usage condition for the 1DCNN model is that the input data should conform to the input format required by the 1DCNN model. Reorganizing the data set accordingly is necessary. For this experiment, the data were transformed into groups, with each group consisting of one record containing nine attribute values. The input data were then reshaped in the format [n, 1, 9], where n represents the total number of records. The output data format is [n, ].
After multiple experiments, it was determined that for this experiment, XGBoost should be trained and used for predictions using the multivariate regression approach. Sklearn provides a wrapper class called MultiOutputRegressor that meets the requirements of this experiment. However, it is necessary to configure the parameters of MultiOutputRegressor accordingly.

3.5. GA-LSTM

In the second experiment, this study proposes a recurrent neural network GA-LSTM (Genetic Algorithm-Long-Short-Term Memory) to predict the displacement of the five axes based on the previous experimental results. Transforms the problem into genes using binary representation, combines multiple sets of genes into a population, and calculates the fitness of each chromosome in the population using the fitness function. Based on fitness, better chromosomes are selected for crossover and mutation to obtain solutions that are closer to the optimal solution. Iteration continues until the termination condition is satisfied.
This study combines GA (Genetic Algorithm) with LSTM (Long-Short-Term Memory) and uses GA to optimize the parameters of the LSTM training process. The main parameters optimized by GA include the data time steps (look_back, lb), the hidden layers of the LSTM model (lstm_nets, ls), the number of training epochs (epochs, ep), and the dropout rate (dp). After obtaining the optimal parameter set, the five-axis displacement dataset of the machine tool is used as input data, and the predicted values of the machine tool’s five-axis displacement are used as the output matrix. The model weights are adaptively adjusted, and the GA-LSTM model is constructed. The data set is trained using this model, and the predicted values are compared with the actual values.

4. Experimental Procedures

4.1. Experimental Environments

This study mainly involved training and prediction using a general computer and a Raspberry Pi. The hardware of the general computer utilized the VACS (Virtual AI Computing System), which includes the Nvidia Quadro RTX8000 GPU (NVIDIA, Santa Clara, CA, USA) for testing purposes. For the Raspberry Pi (Raspberry Pi Foundation, Cambridge, UK), two units of the 4B 8 GB RAM version were used. The operating system used was Raspberry Pi OS (64-bit), Debian version: 12 the official operating system provided by Raspberry Pi. This system is based on Debian Linux. One of the Raspberry Pi units served as the system monitoring host, while the other unit was installed with AI packages for the main testing and computation tasks. The system of this study is primarily implemented on Raspberry Pi 4 with 4 GB RAM hardware. The user interface is developed using QT Creator 4.14. Python is utilized on this platform to build AI models. Various functionalities are executed by calling Python functions through QT Creator. Figure 2 represents the architecture diagram of the system environment.
Figure 2. System Environment Architecture Diagram.

4.2. Experimental 1

4.2.1. Dataset Introduction

Based on our research on smart machine tools from international controller manufacturers such as Fanuc and Siemens and machine tool manufacturers such as DMG, Okuma, Mazak, and Mikron, we have found that there is limited publication and literature available on prediction studies specifically focused on rotary axes. Most of the major manufacturers mainly focus on thermal compensation for machine tools in their research and discussions. Additionally, we have observed that many manufacturers have developed new smart machine tools, but our search for the core technical literature on these types of machine tools did not reveal any related predictions regarding rotary axes. Therefore, in this study, we have selected models that demonstrate better training performance in the field of time-series prediction based on the collected time-series dataset.
The acquisition of the data set for this study mainly involved the collection of data on-site from the operating machine tools. We capture real-time speed and torque data from machine tools and save it. We used an infrared temperature sensor to measure and record temperatures at key points of the machine tool, as shown in Figure 3. The data set for this study mainly collected four temperature points, namely indoor temperature, condenser temperature, rotary axis temperature, and motor stator temperature. We recorded the variations of these four temperature points every minute during the experiment.
Figure 3. Temperature Point Display.
To ensure that the prediction results align closely with the actual working conditions of the machine tool, we observed temperature variations by introducing changes in speed. From the line graph, it is evident that there are slight changes in the temperature trend with variations in speed, as shown in Figure 4.
Figure 4. Temperature Line Chart.

4.2.2. Data Preprocessing

In this study, we selected a time difference of one minute for each data point. After normalizing the data, we transformed the time-series problem into a supervised learning problem using a sliding-window approach. According to the research proposal, the task requires predicting the rotary axis temperature for the next ten minutes based on the previous ten minutes of data. The input values consist of the motor stator temperature, speed, torque, indoor temperature, and condenser temperature at time t − 1, while the output value is the temperature on the inner side of the rotary axis at time t + 9.
Regarding data preprocessing, there are two main aspects of data cleansing. The first part involves checking the consistency of the data, specifically handling outliers or values that are too large or too small, to ensure that the data align with realistic values. The second part involves handling invalid or missing values in the data. In our study, data were collected by measuring on-site, and we performed simultaneous checks and processing for both aspects during the data collection process. Therefore, there is no need for further data cleansing in the data set used in this investigation.
During the data collection process, we collected six attribute values that can affect the temperature variation on the inner side of the rotary axis. To understand the correlation between attributes and temperature variation, we used Microsoft Azure to perform sensitivity analysis on the dataset used in this investigation. Microsoft Azure [36] is a Microsoft-provided public cloud service platform that offers AutoML capabilities for a complete automated analysis of datasets. The results of the sensitivity analysis for this research are shown in Figure 5. It can be observed that the attributes with the greatest impact on the temperature variation on the inner side of the rotary axis are the speed and the motor stator temperature.
Figure 5. Attribute Sensitivity Analysis Chart.

4.3. Experimental 2

Dataset Introduction

This dataset was provided by the Industrial Technology Research Institute and focuses mainly on the temperature variations and five-axis displacements of two different machine tools under different conditions. Five-axis machining is a machining mode of CNC machine tools that utilizes the linear interpolation motion of the X, Y, Z, A, B, and C axes. Machine tools used for five-axis machining are commonly referred to as five-axis machine tools or five-axis machining centers, as shown in the following Figure 6. The machine possesses the capacity to execute tool or workpiece displacement along five distinct axes concurrently. These axes are denoted by different arrows labeled X, Y, Z (for linear displacement) and A, B (for rotational displacement). The + and − symbols indicate the direction of movement along each axis. This capability enables the execution of complicated machining operations, such as drilling, milling, and tapping, at various angles without necessitating the repositioning of the workpiece.
Figure 6. Five-axis Machine Tool or Five-axis Machining Center Display.
As shown in Table 1, the first machine tool provides datasets for three different conditions, comprising a total of 41 columns. These columns include time, rotation speed, 34 temperature points, and displacement variations of the five axes. On the other hand, the second machine tool provides datasets for four different conditions, comprising a total of 60 columns. These columns include time, rotation speed, 53 temperature points, and displacement variations of the five axes.
Table 1. Dataset Introduction.
The experiment ultimately integrates the different operating conditions of the machine tool into a single dataset to increase the amount of data available for model training. Figure 7a displays the temperature variations at different points for Tool 1 in a line chart, while Figure 7b shows the corresponding displacement of the five axes. Similarly, Figure 7c represents the temperature variations at different points for Tool 2, and Figure 7d illustrates the corresponding displacement of the five axes for Tool 2.
Figure 7. The different operating conditions of the machine tool. (a) Tool 1 Temperature Line Chart. (b) 5-axis Displacement of Tool 1. (c) Tool 2 Temperature Line Chart. (d) 5-axis Displacement of Tool 2.

5. Research Results and Discussion

5.1. Experiment 1 Results

The dataset for Experiment 1 consists of a total of 300 records. In this study, the data set is divided into a training set and a test set in an 8:2 ratio for model training. To evaluate the prediction results of the aforementioned models, we mainly use RMSE (Root Mean Square Error) to check if they meet our requirements, as shown in Figure 8. It can be observed that both the GRU and LSTM models yield lower RMSE values, indicating that these two AI models provide more accurate predictions compared to actual values based on the collected data set.
Figure 8. Histogram of RMSE Results of AI model.
Although the TCN model also achieves an RMSE below 1 in most experiments, its performance is comparatively poorer than that of the LSTM and GRU models. In some cases, the TCN model does not converge during training, leading to increased instability in the predictions.
In addition, although the Stacking Ensemble Learning algorithm performs well in overall prediction, we have observed that its results can vary significantly when different rotational speeds are provided for prediction. Therefore, if the machine tool operates with a fixed set of rotational speeds, this AI model might be a good option due to its faster training speed and higher accuracy. However, if the machine tool requires a wide range of rotational speed changes, it is not recommended to use this model due to its inconsistency in predictions.
In addition, we compared the runtime of these four models on the Raspberry Pi, as shown in the above Figure 9. It can be observed that the inference time on the Raspberry Pi is significantly longer compared to a general computer (including the GPU). However, considering that the Raspberry Pi is a lightweight and relatively inexpensive product compared to a general computer, it can be considered a lightweight edge AI computing tool, especially when computational time is not a critical requirement.
Figure 9. Histogram of Computing Time between General Computer and Raspberry Pi.

5.2. Experiment 2 Results

For Experiment 2, Machine Tool 1 has a total of 6433 datasets, and Machine Tool 2 has a total of 7423 datasets. In this study, each data set is divided into a training set and a test set in an 8:2 ratio for model training. According to Pearson’s correlation analysis, Figure 10a shows the top eight temperature points and their corresponding importance values in relation to the five-axis displacement for Tool 1, and Figure 10b shows the top eight temperature points and their corresponding importance values in relation to the five-axis displacement for Tool 2. In the inspection by experts in the field, these results align well with their practical experience.
Figure 10. Histogram of the eight temperature points. (a) Tool 1. (b) Tool 2.
Based on the sensitivity analysis, we filtered the data set for the first eight temperature points related to the machine tool. We then trained the selected dataset using the prebuilt deep learning models. To evaluate the prediction results in this study, we adopted the determination coefficient ( R 2  score) as an assessment metric. In regression models, this coefficient mainly reflects the accuracy of model predictions compared to actual values. A higher coefficient indicates a higher accuracy of the model predictions, with values ranging from 0 to 1. The calculation formula for  R 2 _score is as follows:
R 2 ( y , y ^ ) = 1 i = 0 n s a m p l e 1 ( y i y ^ i ) 2 i = 0 n s a m p l e 1 ( y ^ i y i ) 2
As shown in Figure 11, it can be seen that, under the same dataset and hardware environment, the research results of this experiment indicate that the time-series models used in this study have achieved a  R 2  score of 0.8 or higher. The GA-optimized LSTM model developed in this investigation also demonstrates excellent performance with an accuracy of 0.99.
Figure 11. Histogram of Experimental Research Results.
In the early stages of system development, we evaluated the overall performance of AI models. Figure 12 shows the training times for each model on a general computer. It can be seen that although GA-LSTM achieved the highest prediction accuracy, it required a relatively longer training time. Compared to GA-LSTM, the LSTM, GRU, and XGBoost models, it provided overall more suitable results for practical applications.
Figure 12. Histogram of Model Training Time.

5.3. Discussion

Based on the entire experiment, the development of the system in this study consists mainly of three main functionalities: model training, model prediction, and model retraining, as shown in Figure 13.
Figure 13. Schematic Diagram of Platform Functions.
The first functionality of the system is the initial training of the model. The system performs sensitivity analysis on the user input data set to identify the top 8 temperature points that have the greatest impact on the 5-axis displacement. The data are then preprocessed, specifically for the 8 selected temperature points. The system calculates the temperature differences by subtracting the initial temperature values from the selected temperature values. The dataset is further split to suit the input format of the AI models.
Next, the system trains and validates the input data set using three different AI models. It outputs the predictions of the three models, along with the corresponding  R 2 _score and a line chart that compares the predicted values and the actual values. These results are provided to the user for selection. Based on the experience, the user can choose the most suitable model for the specific machine tool. Finally, the selected model is exported and made available to the user. The Raspberry Pi interface of the system is shown in Figure 14.
Figure 14. Raspberry Pi Interface 1.
The second functionality addresses the need to retrain the models as the dataset evolves over time. The platform provides the ability to retrain the models by preprocessing the new dataset and using the models obtained from the first part. The updated models are then outputted for further use. The Raspberry Pi interface for this functionality is shown in Figure 15.
Figure 15. Raspberry Pi Interface 2.

6. Conclusions

Based on the two experiments conducted, it was observed that applying different time-series models to the machine tool dataset yielded favorable prediction results. Furthermore, the GA-optimized LSTM model developed in the second experiment achieved the best performance. In future research, efforts will be made to optimize the training time of the GA-LSTM model, and additional algorithms will be applied to analyze multiple temperature points in the machine tool compensation data set. The goal is to identify the recommended features and integrate them into the developed Raspberry Pi smart compensation system. Based on the research results, three AI models were selected and developed for edge computing terminals, which demonstrated good performance. Additionally, the operating system was modified to enable its use in the machine cloud at the Industrial Technology Research Institute. Finally, the precision of the research models reached more than 0.96, which is in line with the current trend of using deep learning in the manufacturing industry to improve the thermal errors of the machining. Although this research has achieved a high accuracy of over 0.96 in predicting the five-axis displacement of machine tools, further improvements are needed to apply the developed system to real factory machines. For example, in this study, data were provided by selecting file paths, but in the future, our goal is to establish a direct connection with machine tools for real-time data prediction so that on-site personnel can make adjustments based on future five-axis displacement changes. The training time of the models in this research was time-consuming, and we hope to address this issue in future studies. Long-term time-series problems can lead to model drift, where the performance of the previously trained model deteriorates over time. Therefore, we also aim to develop a system that can effectively address this issue. In addition, this research can be applied to other machine tools.

Author Contributions

Conceptualization, C.-T.Y. and J.-C.L.; Methodology, C.-T.Y. and J.-C.L.; Software, L.-Y.W. and E.K.; Validation, J.-C.L. and C.-T.Y.; Formal Analysis, E.K. and L.-Y.W.; Investigation, C.-T.Y.; Resources, C.-K.H. and S.-J.W.; Data Curation, C.-K.H. and S.-J.W.; Writing—Original Draft Preparation, E.K. and L.-Y.W.; Writing—Review and Editing, C.-T.Y. and J.-C.L.; Visualization, C.-K.H. and S.-J.W.; Supervision, C.-T.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work was sponsored by the National Science and Technology Council (NSTC), Taiwan, under Grant No. 112-2622-E-029-003, 112-2621-M-029-004, 112-2221-E-126-004, 112-2811-E-029-003, and 110-2221-E-029-002-MY3.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author due to privacy concerns.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Jazdi, N. Cyber physical systems in the context of Industry 4.0. In Proceedings of the 2014 IEEE International Conference on Automation, Quality and Testing, Robotics, Cluj-Napoca, Romania, 22–24 May 2014; pp. 1–4. [Google Scholar]
  2. Konstantinidis, F.K.; Myrillas, N.; Mouroutsos, S.G.; Koulouriotis, D.; Gasteratos, A. Assessment of industry 4.0 for modern manufacturing ecosystem: A systematic survey of surveys. Machines 2022, 10, 746. [Google Scholar] [CrossRef]
  3. Moosavi, J.; Bakhshi, J.; Martek, I. The application of industry 4.0 technologies in pandemic management: Literature review and case study. Healthc. Anal. 2021, 1, 100008. [Google Scholar] [CrossRef] [PubMed]
  4. Chen, B.; Wan, J.; Celesti, A.; Li, D.; Abbas, H.; Zhang, Q. Edge computing in IoT-based manufacturing. IEEE Commun. Mag. 2018, 56, 103–109. [Google Scholar] [CrossRef]
  5. Ezugwu, E.O.; Bonney, J.; Da Silva, R.B.; Çakir, O. Surface integrity of finished turned Ti–6Al–4V alloy with PCD tools using conventional and high pressure coolant supplies. Int. J. Mach. Tools Manuf. 2007, 47, 884–891. [Google Scholar] [CrossRef]
  6. Konstantinidis, F.K.; Myrillas, N.; Tsintotas, K.A.; Mouroutsos, S.G.; Gasteratos, A. A technology maturity assessment framework for industry 5.0 machine vision systems based on systematic literature review in automotive manufacturing. Int. J. Prod. Res. 2023, 1–37. [Google Scholar] [CrossRef]
  7. Ren, S.; Kim, J.S.; Cho, W.S.; Soeng, S.; Kong, S.; Lee, K.H. Big data platform for intelligence industrial IoT sensor monitoring system based on edge computing and AI. In Proceedings of the 2021 International Conference on Artificial Intelligence in Information and Communication (ICAIIC), Jeju Island, Republic of Korea, 13–16 April 2021; pp. 480–482. [Google Scholar]
  8. Sun, W.; Liu, J.; Yue, Y. AI-enhanced offloading in edge computing: When machine learning meets industrial IoT. IEEE Netw. 2019, 33, 68–74. [Google Scholar] [CrossRef]
  9. Trinks, S.; Felden, C. Edge computing architecture to support real time analytic applications: A state-of-the-art within the application area of smart factory and industry 4.0. In Proceedings of the 2018 IEEE International Conference on Big Data (Big Data), Seattle, WA, USA, 10–13 December 2018; pp. 2930–2939. [Google Scholar]
  10. Azeem, M.; Haleem, A.; Javaid, M. Symbiotic relationship between machine learning and Industry 4.0: A review. J. Ind. Integr. Manag. 2022, 7, 401–433. [Google Scholar] [CrossRef]
  11. Totu, A.; Saito, K.; Tabacaru, M. Mazak-Intelligent Machine Tools. In Proceedings of the 16th International Conference on Manufacturing Systems—ICMaS, Bucharest, Romania, 2 November 2007; pp. 69–72. [Google Scholar]
  12. Mikron. 2023. Available online: https://www.mikron.com/ (accessed on 6 February 2023).
  13. Zhou, X.; Hu, Y.; Liang, W.; Ma, J.; Jin, Q. Variational LSTM enhanced anomaly detection for industrial big data. IEEE Trans. Ind. Inform. 2020, 17, 3469–3477. [Google Scholar] [CrossRef]
  14. Ren, L.; Wang, T.; Laili, Y.; Zhang, L. A data-driven self-supervised LSTM-DeepFM model for industrial soft sensor. IEEE Trans. Ind. Inform. 2021, 18, 5859–5869. [Google Scholar] [CrossRef]
  15. Mateus, B.C.; Mendes, M.; Farinha, J.T.; Cardoso, A.M. Anticipating future behavior of an industrial press using LSTM networks. Appl. Sci. 2021, 11, 6101. [Google Scholar] [CrossRef]
  16. Alazab, M.; Khan, S.; Krishnan, S.S.R.; Pham, Q.V.; Reddy, M.P.K.; Gadekallu, T.R. A multidirectional LSTM model for predicting the stability of a smart grid. IEEE Access 2020, 8, 85454–85463. [Google Scholar] [CrossRef]
  17. Liu, B.; Song, C.; Wang, Q.; Wang, Y. Forecasting of China’s solar PV industry installed capacity and analyzing of employment effect: Based on GRA-BiLSTM model. Environ. Sci. Pollut. Res. 2022, 29, 4557–4573. [Google Scholar] [CrossRef] [PubMed]
  18. Lan, M.; Luo, J.; Chai, S.; Chai, R.; Zhang, C.; Zhang, B. A novel industrial intrusion detection method based on threshold-optimized CNN-BiLSTM-Attention using ROC curve. In Proceedings of the 2020 39th Chinese Control Conference (CCC), Shenyang, China, 27–29 July 2020; pp. 7384–7389. [Google Scholar]
  19. Prihatno, A.T.; Nurcahyanto, H.; Ahmed, M.F.; Rahman, M.H.; Alam, M.M.; Jang, Y.M. Forecasting PM2.5 concentration using a single-dense layer bilstm method. Electronics 2021, 10, 1808. [Google Scholar] [CrossRef]
  20. Çavdar, T.; Ebrahimpour, N.; Kakız, M.T.; Günay, F.B. Decision-making for the anomalies in IIoTs based on 1D convolutional neural networks and Dempster–Shafer theory (DS-1DCNN). J. Supercomput. 2023, 79, 1683–1704. [Google Scholar] [CrossRef]
  21. Liang, Y.; Li, W.; Lou, P.; Hu, J. Thermal error prediction for heavy-duty CNC machines enabled by long short-term memory networks and fog-cloud architecture. J. Manuf. Syst. 2022, 62, 950–963. [Google Scholar] [CrossRef]
  22. Gui, H.; Liu, J.; Ma, C.; Li, M.; Wang, S. New mist-edge-fog-cloud system architecture for thermal error prediction and control enabled by deep-learning. Eng. Appl. Artif. Intell. 2022, 109, 104626. [Google Scholar] [CrossRef]
  23. Guo, J.; Xiong, Q.; Chen, J.; Miao, E.; Wu, C.; Zhu, Q.; Yang, Z.; Chen, J. Study of static thermal deformation modeling based on a hybrid CNN-LSTM model with spatiotemporal correlation. Int. J. Adv. Manuf. Technol. 2022, 119, 2601–2613. [Google Scholar] [CrossRef]
  24. Kuo, P.H.; Chen, Y.W.; Hsieh, T.H.; Jywe, W.Y.; Yau, H.T. A Thermal Displacement Prediction System with an Automatic LRGTVAC-PSO Optimized Branch Structured Bidirectional GRU Neural Network. IEEE Sens. J. 2023, 23, 12574–12586. [Google Scholar] [CrossRef]
  25. Kuo, P.H.; Chen, S.C.; Lee, C.H.; Luan, P.C.; Yau, H.T. Ensemble Model for Spindle Thermal Displacement Prediction of Machine Tools. Comput. Model. Eng. Sci. 2023, 137, 319–343. [Google Scholar] [CrossRef]
  26. Liu, Y.C.; Li, K.Y.; Tsai, Y.C. Spindle thermal error prediction based on LSTM deep learning for a CNC machine tool. Appl. Sci. 2021, 11, 5444. [Google Scholar] [CrossRef]
  27. Liu, J.; Ma, C.; Gui, H.; Wang, S. Thermally-induced error compensation of spindle system based on long short term memory neural networks. Appl. Soft Comput. 2021, 102, 107094. [Google Scholar] [CrossRef]
  28. Nguyen, D.K.; Huang, H.C.; Feng, T.C. Prediction of Thermal Deformation and Real-Time Error Compensation of a CNC Milling Machine in Cutting Processes. Machines 2023, 11, 248. [Google Scholar] [CrossRef]
  29. Zeng, S.; Ma, C.; Liu, J.; Li, M.; Gui, H. Sequence-to-sequence based LSTM network modeling and its application in thermal error control framework. Appl. Soft Comput. 2023, 138, 110221. [Google Scholar] [CrossRef]
  30. Ji, T.; Sivakumar, A.N.; Chowdhary, G.; Driggs-Campbell, K. Proactive anomaly detection for robot navigation with multi-sensor fusion. IEEE Robot. Autom. Lett. 2022, 7, 4975–4982. [Google Scholar] [CrossRef]
  31. Lee, M.C.; Lin, J.C.; Gan, E.G. ReRe: A lightweight real-time ready-to-go anomaly detection approach for time series. In Proceedings of the 2020 IEEE 44th Annual Computers, Software, and Applications Conference (COMPSAC), Madrid, Spain, 13–17 July 2020; pp. 322–327. [Google Scholar]
  32. Gupta, S.; Muthiyan, N.; Kumar, S.; Nigam, A.; Dinesh, D.A. A supervised deep learning framework for proactive anomaly detection in cloud workloads. In Proceedings of the 2017 14th IEEE India Council International Conference (INDICON), Roorkee, India, 15–17 December 2017; pp. 1–6. [Google Scholar]
  33. Spantideas, S.; Giannopoulos, A.; Cambeiro, M.A.; Trullols-Cruces, O.; Atxutegi, E.; Trakadas, P. Intelligent Mission Critical Services over Beyond 5G Networks: Control Loop and Proactive Overload Detection. In Proceedings of the 2023 International Conference on Smart Applications, Communications and Networking (SmartNets), Istanbul, Turkiye, 25–27 July 2023; pp. 1–6. [Google Scholar]
  34. Wang, Q.; Bu, S.; He, Z. Achieving predictive and proactive maintenance for high-speed railway power equipment with LSTM-RNN. IEEE Trans. Ind. Inform. 2020, 16, 6509–6517. [Google Scholar] [CrossRef]
  35. Psarommatis, F.; May, G.; Azamfirei, V.; Konstantinidis, F. Optimizing efficiency and zero-defect manufacturing with in-process inspection: Challenges, benefits, and aerospace application. Procedia Comput. Sci. 2024, 232, 2857–2866. [Google Scholar] [CrossRef]
  36. Barnes, J. Azure machine learning. In Microsoft Azure Essentials; Microsoft: Raymond, WA, USA, 2015; pp. 1–237. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.