Next Article in Journal
Evaluation of the Accessibility to Urban Mobility Services with High Spatial Resolution—Case Study: Valencia (Spain)
Previous Article in Journal
Multiscale Dynamic Attention and Hierarchical Spatial Aggregation for Few-Shot Object Detection
Previous Article in Special Issue
Explainable AI (XAI) Techniques for Convolutional Neural Network-Based Classification of Drilled Holes in Melamine Faced Chipboard
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Using Hybrid LSTM Neural Networks to Detect Anomalies in the Fiber Tube Manufacturing Process

1
College of Natural Sciences, University of Rzeszow, Rejtana St. 16C, 35-959 Rzeszow, Poland
2
FIBRAIN Sp. z o.o., Zaczernie 190F, 36-062 Zaczernie, Poland
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(3), 1383; https://doi.org/10.3390/app15031383
Submission received: 12 December 2024 / Revised: 6 January 2025 / Accepted: 13 January 2025 / Published: 29 January 2025
(This article belongs to the Special Issue Engineering Applications of Hybrid Artificial Intelligence Tools)

Abstract

:
The production process of tubes for fiber optic cables is a complex process, where proper execution is crucial to the quality of the final product. This process has a complex state vector whose structure and coordinates dynamically change during the tube extrusion process. Small fluctuations in process parameters, such as temperature, extrusion pressure, production speed, and optical fiber tension, affect the optical attenuation of the final product. Such defects necessitate the withdrawal of the product. Due to the high number of process coordinates and the technological inability to automatically label those segments of the production process that cause anomalies in the final product, the authors used data clustering methods to create a training set that enabled the use of neural tools for anomaly detection. The system proposed in the main part of the paper includes a hybrid Long short-term memory (LSTM) network model, which is fed with data streams recorded on the tube extrusion production line. The input module, which performs preprocessing of input data, conducts multiresolution analysis of recorded process parameters, and recommends the process state’s belonging to a set of classes describing individual production anomalies to appropriate LSTM network modules. The learning process of the three–channel network allowed effective recognition of five classes of the monitored tube production process. The fit level of the proposed network model reached R2 values of ≥0.85.

1. Introduction

Fiber optic cables in modern civilization are the primary medium used for efficient and reliable information transmission. The continuous expansion of computer network infrastructure drives a growing demand for such products. Solutions are expected to include both long segments of optical medium as well as multi–core and multi–mode solutions, which allow for parallelization and thus acceleration of information transmission. The basic raw material is usually optical fiber, which has specific parameters describing its attenuation, rigidity, and physical dimensions. It is obtained by process of drawing the fiber in special towers from prepared glass billets of high chemical purity, which ensures appropriately low optical attenuation. Typically, at this stage, the fiber is coated with a dedicated polymer that ensures light confinement within it. The polymer layers can be combined with a color that ensures proper separation of fibers already at the stage of digital network construction. The next stage of optical cable manufacturing in industrial conditions is the extrusion process of the tube, which involves continuous heating and extrusion of plastic or metal materials to form hollow tubes that serve as a hydrophobic coating surrounding bundles of optical fibers. The subsequent step is the twisting of tube assemblies, and the final product of this process is an optical cable.
As can be seen, the above process is complex, with many different state coordinates, and each stage of production can potentially be a source of defects in the final product. In this context, predicting process anomalies using data from production equipment and taking preventive action is crucial. This is a key leitmotif that motivates many research teams around the world [1,2,3] to seek solutions and tools for monitoring this process and preventing defects that degrade the quality of the final product. Typically, the anomaly detection process involves collecting and analyzing data from sensors that monitor temperature, pressure, vibrations, chemical parameters, diameter, ovality, and other indicators specific to a given production line’s state. In recent years, statistical algorithms (e.g., linear regression), machine learning algorithms (e.g., decision trees, Support Vector Machines (SVM), neural networks [2,4]), and deep learning algorithms (e.g., Recurrent Neural Networks (RNN), Autoencoders) have been used for this purpose. The primary task of these algorithms is to monitor the given process in real–time, maintain predictive maintenance [5], and ensure quality control. Currently, companies in the fiber optic cable manufacturing sector are developing various solutions to minimize the percentage of products manufactured out of specification. Among the challenges they face are the scalability and adaptability of algorithms, their ability to integrate with production systems, and their effectiveness in minimizing waste and downtime [6,7,8,9,10,11,12].
By reviewing the latest literature, State-of-the-Art Review on Anomaly Detection in Various Domains [13], several leading themes can be distinguished. The detection of anomalies has emerged as a critical focus across diverse domains, particularly within the Internet of Things (IoT), biomedical signal processing, industrial processes, and satellite telemetry data. This review synthesizes recent advances in methods and approaches, highlighting their unique characteristics, applications, and associated challenges. Anomaly detection in IoT systems often employs dual-view mechanisms, combining time reconstructors, frequency reconstructors, and dual-view adversarial learning [14,15]. These methods aim to analyze data streams comprehensively, addressing both temporal and spectral dimensions to enhance detection accuracy. Biomedical data, such as signals related to Parkinson’s disease, are processed using unsupervised deep networks. These methods focus on uncovering latent patterns within data without labeled training samples, providing significant insights into disease recognition while addressing specific anomaly types. The use of singular value decomposition (SVD) in deep neural network (DNN) models aids in encrypting hidden features of signals. Techniques incorporating Fourier transforms allow for robust analysis of anomalies while maintaining data integrity [16]. Recent developments include the implementation of self-adversarial variational autoencoders (VAEs) integrated with contrastive learning to enhance anomaly detection capabilities [16]. These architectures excel in capturing nuanced differences between normal and anomalous states. Multi-node knowledge graphs fused with expert knowledge and Bi-directional Long Short-Term Memory (Bi-LSTM) networks have been proposed [17] as powerful tools for local feature extraction in anomaly detection. This approach leverages both structural and temporal relationships within the data. The multi-scale sparse attention module effectively extracts global features across scales, preserving pattern information critical for detecting process disturbances [18]. This approach is particularly adept at handling data from industrial monitoring systems [19]. For multidimensional processes, local feature masking can amplify state differences to visually highlight anomalies [20]. Similarly, mechanisms for identifying highly divergent patterns based on data density help streamline the detection process [21]. A knowledge distillation approach [22], transferring insights from large models to smaller, more efficient models, has proven effective in speeding up anomaly detection. Stochastic approaches utilizing partial Markov chains are employed to mask irrelevant process features, such as in image data [23]. These methods improve reconstruction quality and focus anomaly detection efforts on significant features. An imbalance in training datasets is a persistent issue, especially in industrial processes where generating anomalies artificially is infeasible [24]. Few-shot anomaly detection using positive-unlabeled learning addresses this by preparing well-structured datasets, albeit with challenges in real-world implementation [25]. Preprocessing transformations, such as Kalman filtering, are used to simplify input data for anomaly detection in diverse systems, including ECG signals and cooling devices [26]. Additionally, federated learning enhances defect detection, achieving high accuracy, precision, recall, and F1 scores in certain industrial settings [24]. Frequency-domain techniques generate augmented pseudo-anomalous images, allowing for the discovery of key anomaly characteristics [27]. Analyzing short video sequences of industrial processes has achieved notable success, such as the MSTE environment yielding 90.7 % accuracy [28]. However, video-based methods face limitations due to measurement constraints and environmental variability. In satellite telemetry anomaly detection, deep learning algorithms contend with significant noise and data gaps [29]. To address data scarcity in anomaly states, mechanisms such as feature injection and heterogeneous spatiotemporal graphs have been developed [30]. These techniques enable artificial supplementation of training datasets, improving detection accuracy in processes with high anomaly deficits. In unsupervised scenarios, attributed networks leveraging graph-based models enhance the differentiation between normal states and anomalies by reducing interference [31].
In the study [6], anomaly detection in smart factories was explored using machine learning techniques, particularly addressing challenges related to diverse sensors and their transferability across different production lines. A combination of machine learning algorithms, including neural networks and transfer learning, was employed to facilitate knowledge transfer between production lines. The authors achieved high accuracy in anomaly detection using datasets from various sensors, and the knowledge transfer between sensors enhanced system effectiveness in new, unfamiliar production environments. The study [10] proposed the creation of a digital twin for a production line, enabling improved monitoring and real-time anomaly prediction. This approach integrated digital twin modeling with machine learning algorithms, such as Long Short-Term Memory (LSTM) networks and autoencoders. The system demonstrated high accuracy in anomaly prediction, while the digital twin allowed for dynamic production parameter adjustments, reducing defects by 40 % compared to standard methods. In [5], the authors introduced a system for predicting production costs in the context of Industry 4.0. This system integrated data from multiple sources, including industrial sensors and ERP systems, to forecast changes in production costs. Advanced ML models, such as LSTM networks and decision trees, were applied to analyze time-series data from various sources. The model achieved a 92 % accuracy rate in production cost predictions, significantly reducing financial losses from unplanned downtimes and production defects. The study [7] focused on anomaly detection in optical fiber networks using machine learning algorithms. The primary application involved monitoring the integrity of optical cables and detecting damages in communication networks. Support Vector Machines (SVM) and neural networks, including LSTM models, were used to process sensor data and detect cable irregularities. The LSTM model excelled in real-time anomaly detection, achieving 94 % accuracy, and effectively identified both minor and major damages. Finally, Ref. [12] investigated automatic anomaly detection in manufacturing machines based on sensor data, aiming to minimize machine failures and downtimes by predicting potential issues. Statistical algorithms such as linear regression and k-Nearest Neighbors (k-NN) were combined with advanced machine learning methods, including LSTM networks. These models successfully identified anomalies in 87 % of cases, leading to a 25 % reduction in production process downtimes.
Table 1 presents a summary of the technologies employed and the performance of the models in the context of research on anomaly detection and production process monitoring.
Regarding above, the industry practice shows, that often heuristic strategies are needed to overcome these industrial challenges, which frequently hinder effective detection [32,33]. Emerging trends such as federated learning, knowledge graphs, and advanced neural architectures offer promising directions for overcoming these challenges. By integrating domain expertise with sophisticated computational techniques, anomaly detection continues to evolve, bridging gaps across diverse applications. Although the recent reviews have attempted to unify diverse methods to facilitate objective comparisons, but unfortunately significant challenges persist due to process-specific measurement differences and the unique nature of anomalies across domains, which often hinder generalization [34,35,36,37,38,39,40].
In this study, the authors refer to the concept of anomalies as a set of various disturbances that may accompany the production process. These deviations from the correct process, which has its proper signature representing a normal and unthreatened production state, can become, in the absence of corrective actions, a potential source of defects in the final product. This may result in either a reduction in the product’s final quality or, in a worst-case scenario, render it unusable for further application.
This paper presents the results obtained from monitoring the tube extrusion process based on actual measurements recorded on the production line at the industrial plant FIBRAIN Sp. z o.o. This company specializes in the production and delivery of fiber optic technology solutions and is one of the leading manufacturers of fiber optic cables and telecommunication components in Central and Eastern Europe. The designed anomaly detection system allows for the reduction of disposal and repair costs for tube sections produced out of specification on the tube extrusion production line. The system described in the main section of the paper incorporates a hybrid Long Short-Term Memory (LSTM) network model, designed to process data streams collected from a tube extrusion production line. The input module, responsible for preprocessing, performs multiresolution analysis of the recorded process parameters. It subsequently assigns the current process state to one of the predefined classes, representing specific production anomalies, and directs it to the appropriate LSTM network modules. The three-channel network’s training process enabled the effective identification of five distinct classes within the monitored tube production process.

2. Materials and Methods

Given the current and continuously growing scale of production, the number of occurrences of various defects disqualifying a given fiber optic cable is a source of significant financial losses for the manufacturer in the form of wasted resources, the cost of disposing of defective cables, or their potential repair. The entire research work encompassed three different production lines, as shown in Figure 1: the RLR line for tube extrusion (a), the RLV line for coating extrusion on the cable core (b), and the RLM line for twisting cable cores from tubes (c).
These are stages where both the number and type of sensors used make it impossible to implement a uniform system that, in a hybrid form, would simultaneously record and detect anomalies from different processes. Therefore, at the current stage of work, the considerations presented by the research team focus on the production stage involving a single RLM production line.
In this paper, a hybrid LSTM network model is presented, which pertains to measurements from the RLM line. The other production lines are subjects of ongoing research and will be discussed in future works focusing on the engineering applications of AI tools. The RLM line is responsible for the production of micro–module cables with fiberglass rods (FRP) (MDC–FM, AERO–FM), aerial cables (ADSS, AERO–AS), road cables reinforced with fiberglass or metal rods in the cable sheath (AERO–DDF, Burry, VC–Tx), easy access cables (EAC), duct cables (BDC), micro cables (MK–LX), central tube cables (EXO), and steel tape armored cables (SST).

2.1. Factory RLM Line Data Acquisition

As part of the conducted work, the line was equipped with devices that allow for monitoring and recording the tension of individual modules or tight tubes, as well as the temperature of the process water located directly in the cooling baths. Based on the recommendations of tube manufacturing technology experts, 232 measurement points were designated, where the production process was recorded at a frequency of 1 s . This set included both the established process recipe parameters and process variables. The complete data set contained 180 randomly recorded production sessions lasting from 30 min to 150 min, corresponding to different working hours of the line and the teams of operators handling it. An example fragment of measurement data from a single session is presented in Table 2. This includes a small range of approximately a one–hour measurement session in which 20 km of optical cable was produced.
During the conducted work related to recording measurement data, the team observed random gaps in the process data. These gaps were caused either by improper operator handling of the production line or by malfunctions in individual sensors. Naturally, there are fundamentally different strategies to counteract such issues, ranging from simple methods of discarding damaged data sequences to advanced strategies for repairing them using sophisticated statistical methods, machine learning, or interpolation and extrapolation techniques. Each approach has a specific level of accuracy and computational complexity, which can be crucial when selecting the appropriate method for various industrial applications.
For the experiments involving the RLM line, the method of extrapolating missing data (“gaps”) was adopted, when the size of the gaps did not exceeded 10 s. In cases where the negative effect was more pronounced, the affected data packets were discarded. Considering the strict technological conditions, the occurrence of missing data or abrupt changes in the recorded values was extremely rare.
These data were subsequently analyzed using a sliding window with sizes of [64, 128, 256, 512, 1024]. Considering the dynamics of the manufacturing process, it was determined that the most optimal window size, from a technological standpoint and accounting for the possible reaction time of the line operator in the hypothetical event of an anomaly presence, would be approximately 4 min, corresponding to a 256-element window. Other time intervals might also highlight interesting properties of the process; however, for the purposes of this study and its limited scope, they were excluded from further consideration.

2.2. Preprocessing of Input Data for Training Set

A fundamental problem faced by research teams working on anomaly detection is the inability to precisely construct a training sequence, due to the fact that the vast majority of defects caused by anomalies are detected only after the completion of the production process during the final product quality checks. Usually, determining the precise temporal coincidence of a specific moment in the production process is either impossible or associated with a large spatial localization error. To address this, an initial assessment of the input data was performed by analyzing the significance of features in the recorded process state vector using the Discrete Wavelet Transform (DWT) with a Haar wavelet basis function, applying a 4–octave wavelet decomposition level. This resulted in sets of 32 features for successive windows, which illustrate different process states. For the purposes of further discussion, the following mathematical notation have been assumed. Let x i t represent the discrete input signal registered from the i-th sensor channel of the RLM production line, where: i = 1 , 2 , , S N , t = 0 , 1 , 2 , , W s , S N = 232 —denotes the number of mounted sensors in RLM line, W s = 256 —denotes the window size, (i.e., 256 time stamps in the moving window of length 256). The matrix representation of the complete signal data from all sensors of consecutive channels within the moving window can be represented as a matrix:
X t = x 1 t x 2 t x S N t
where X t R S N × W S . The moving window mechanism we assume respectively: t w denote the moving window size (here t w = W S ). For every sensors channels i, the windowed segment can be expressed to as: x i t = x i k k = t ,   t + 1 ,   ,   t + t w 1 . The initial research assumptions were that a pyramidal wavelet decomposition would be performed on the process coordinates so recorded, which, due to the absence of constraints on the periodicity and continuity of the signal characteristic of Fourier frequency decomposition, allows the extraction of selected components of the transformed signal into complementary components. For this purpose, we used the algorithm proposed by [41] which uses signal decomposition x i t into approximations and details at various levels j: j : D W T x i t = A j , D j j = 1 J , where A j is the approximation coefficients at level j, D j is the detail coefficients at level j and J is the maximum decomposition level, depending on the length of x i t . Taking into account the technological requirements determined by the size of the analyzed window, it was assumed for the purposes of the preliminary analysis that the J 2 , 3 , 4 , 5 , 6 . The multi-sensor processing with DWT is applied to each feed individually:
D W T X t = D W T x 1 t D W T x 2 t D W T x S N t
Presented above mathematical framework ensures clarity in processing, accommodating the 232 sensor channels, 256—sample moving window, and 1—second sampling interval. Below (Figure 2) are sample fragments of the signal recorded from the RLM production line for wavelet decomposition using the Haar mother wavelet with 2 and 4 octave sizes, respectively.
For all waveforms in Figure 2, the designations Applsci 15 01383 i001, Applsci 15 01383 i002 and Applsci 15 01383 i003 stand for: the waveform of the components A j , D j of the DWT transform, the waveforms of the leading 40 components of the production process, and the membership of the RLM line signal to the different classes extracted in the clustering process. The left column (Figure 2a,c,e,g,i,k) corresponds to the decomposition of the signal recorded during the measurement session from the RLM production line using a 4–octave Haar–based DWT, while the right column (Figure 2b,d,f,h,j,l) corresponds to the decomposition of the same signal using a 2–octave Haar–based DWT. The waveforms (Figure 2a–d) show example sections of RLM line acceleration, in which the time interval S B , , S E is marked with a rectangular area. The symbols S B and S E , denote the start of the fiber twisting process and the end of the twisting process of a given tube section, respectively. Graphs (Figure 2e–h) contain waveforms for the ongoing basic production process also for two different randomly selected sessions. The recorded process data from the RLM line X R S N × W S , were normalized a = 0 , b = 1 for each signal channel data independently:
x i , j , norm = a + b a · x i , j m i n X i , : m a x X i , : m i n X i , :
where: X i is the i-th row of X. The plots (Figure 2i–l) show sample stages of the finalization of the fiber twisting process for a given tube section, marked as C B , , C E , where the symbols C B and C E denote the start and end of the production completion process for the given tube section, respectively. All plots marked with the number Applsci 15 01383 i003 represent an attempt to visualize the potential membership of the production process state to a priori selected number of classes.
Subsequently, a univariate feature selection algorithm for classification using the Chi–square test was applied. This allowed for the visualization and assessment of the importance of individual leading features in the recorded data stream, represented as a cluster structure corresponding to 32 different process states, as shown in Figure 3. The Table 3 contains the coordinates and names of the individual leading components of the process.
In the course of observing the impact of individual process characteristics on cluster distribution, it was found that it is possible to use an automatic clustering mechanism to construct a training set for Deep Neural Networks (DNN). At the same time, it was noted that the DWT showed high sensitivity to momentary impacts of some parameters, such as the state of the cable length counter when changing the winding drum. Such a phenomenon caused strong interference, which negatively affected the segmentation process of individual process states. In addition, for analytical purposes, the authors attempted to graphically interpret the distribution of individual clusters and the distance between them, using the Hierarchical Clustering with Binary Clustering Tree method. Assuming that the given is a dataset with n observations, where each observation is represented as a d-dimensional vector x i = ( x i 1 , x i 2 , , x i d ) , the steps for agglomerative hierarchical clustering we obtained by iterative merging: by serching for the pair of clusters ( C i , C j ) with the smallest distance d ( C i , C j ) . Then merging clusters C i and C j into a new cluster C i j and eventually update the distance matrix to reflect the distance between consecutive new cluster C i j and all other. As the linkage criteria we assumed the average distance method to measure the distance between clusters:
d ( C i , C j ) = 1 | C i | | C j | x a C i x b C j d ( x a , x b )
By the repetition of merging loop we obtained a Binary Clustering Tree of the monitored process clusters. This preliminary preprocessing of RLM line data we performed for all sessions and the resulting dendrogram and 3D view of the most significant components are shown on the Figure 4 respectively.
As we can observe a binary clustering tree is as a kind of dendrogram, where each node represents a cluster, and each leaf node represents an individual observation. The height of the nodes in the tree corresponds to the distance (or dissimilarity) at which consecutive clusters are merged. A spatial visualization of the location of the leading features of the clusters is provided in Figure 4b. As can be seen, there is a classic “energy leakage effect” causing the extracted classes of process states to interpenetrate and individual clusters to differ to a very small degree. Therefore, a multi–resolution K—means Clustering mechanism was adopted in later part of this work to cluster the input data at different resolutions of the space of feature channels C h j 8 , 16 , 32 , for  j = 1 , 2 , 3 . So, having an RLM line dataset with n = 256 observations for each stepping window, where each observation is represented as a d-dimensional vector we have:
x i C h j = ( x i 1 C h j , x i 2 C h j , , x i d C h j )
Then we performed K—means clustering in each of the channels to partition the observations into channel independent clusters:
K C h j C 1 C h j , C 2 C h j , , C K C h j
such that sum of squared distances between observations and their corresponding cluster centroids is minimized. Given μ 1 , μ 2 , , μ k , as initial cluster centroids we assign each observation x i C h j to the cluster:
C k C h j = x i C h j arg min k x i C h j μ k 2 , k = 1 , , C h j
where x i C h j μ k is the Euclidean distance between observation x i C h j and centroid μ k . The update of the centroid for each cluster in a given channel C h j , is calculated as a mean of the observations assigned to that cluster:
μ k = 1 | C k C h j | x i C h j C k C h j x i
where | C k C h j | is the number of observations in cluster C k C h j . This way of preprocessing the input data was intended to maximize the disparity of the process states in order to facilitate the task of building a training set for the LSTM network. One of the key difficulties that arises in anomaly detection tasks is that technological limitations do not allow precise labeling of training set classes. Therefore, an analysis of individual measurement sessions was carried out in cooperation with RLM line technologists. As a result of it, it was indicated after which sessions the quality procedure checks showed any defects in the final product. In this way, it was possible to estimate the index of potential defect for each measurement session: p = 1 for normal production, p = 0 for the interval of the production process in which any defects were detected during the post manufacturing quality control in a given measurement session. A dendrogram and a spatial visualization of the most relevant features were constructed again for such a completed input data set, see Figure 5.
By incorporating this additional indicator for the rough spatial localization of defects, a sufficiently effective mechanism for distinguishing anomalies responsible for defects in the final product was achieved. The described data preprocessing stage was used to construct the training dataset.

2.3. Hybrid LSTM Network Model for Anomaly Detection

To process the data stream recorded during the manufacturing of fiber optic cables, which are subjected to K—means clustering in three channels simultaneously, a dedicated model of a hybrid LSTM network has been proposed. Its task is to analyze the states of the production line in each of the channels, see Figure 6. The integrating module of the network is an aggregating layer along with a set of auxiliary layers, whose task is to detect anomalies based on the recommendations of the three predictors operating in each of the channels.
Considering the three–channel architecture of the LSTM network, the following single cell model was used, see Figure 7.
Assuming that the input data of a single cell at time step t is the input vector x t , the mathematical model for a fixed time step enabling the update of the memory state c t is as follows:
f t C h = σ u f h t 1 + w f x t + b f
i t C h = σ u i h t 1 + w i x t + b i
c ˜ t C h = t a n h u c h t 1 + w c x t + b c
o t C h = σ u o h t 1 + w o x t + b o
h t C h = t a n h o t c t
c t C h = f t c t 1 + i t c t ˜
and the generalized matrix is written in the form:
W C h = w f , w i , w c , w o T
where: σ —is the sigmoid activation function, h t 1 —previous hidden state and t a n h is the hyperbolic tangent activation function and W C h denotes the given channel weights set. Figure 8 represents the structure of a single LSTM layer operating in each of the channels.
Assuming the clustering mechanism described by Equation (7), we obtain the relationship for the expected values of the signature at time t of the industrial process on the RLM line:
y C h j t = C C h j k X t
As in the supervised training model the goal of the LSTM model is to learn a function f : R S N × W S 1 , 2 , , C k C h j such that the output y ^ C h j t predicted by the model is as close as possible to the true label y C h j t . This we achieved by minimizing the loss function L W C h :
L W C h = 1 N i = 1 N k = 1 C k C h j y C h j l o g y ^ C h j W C h
where W C h represents the weight set collection sought and N denotes the number of samples in the learning set.
Individual channels of L S T M C h 8 , 16 , 32 have the following numerical details, see Table 4 and Table 5 .
For the chosen LSTM network model, a training set was constructed with parameters summarized in Table 6. They represent the time interval from which individual measurement sessions originate. In total, recorded coordinates of the production process and extracted classes corresponding to individual diagnostic states and the actual production state were used.

3. Results and Discussion

The training process of the hybrid LSTM network was conducted independently in each of the submodules L S T M C h 8 , 16 , 32 . Subsequently, for the frozen weight sets W C h 8 , 16 , 32 , a convergent learning process of the aggregating module was carried out using the Anomalies labeler block (see Figure 6). The course of this process and the achieved degree of model fit for the entire LSTM network are shown in Figure 9. The proportion of the training set split into training and testing sets was set at 70 % and 30 % , respectively. The obtained model fit coefficient R 2 > 0.85 demonstrates the high efficiency of the model on the tested data set.
RLM production line
Root Mean Squared Error: 0.475
F-statistic vs. constant model: 6.62 × 103, p-value = 0
R =
1.0000 0.9442
0.9442 1.0000
Obtained Pearson correlation coefficients: 0.9442
Obtained R2 determination coefficient: 0.892
Figure 9. Pearson coefficient (a) and Confusion matrix (b).
Figure 9. Pearson coefficient (a) and Confusion matrix (b).
Applsci 15 01383 g009
To evaluate the effectiveness of anomaly detection, a detailed experimental study was conducted using identical data sets, with models: Random Forest, SVM, and a simple RNN as benchmarks. The performance of each of these models was evaluated using standard metrics, including Precision, Recall, F-1 Score, and  R 2 , which allow an objective assessment of the quality of anomaly detection and prediction [42]. Detailed results are shown in Table 7.
The obtained results show that LSTM significantly outperforms other models, particularly in the context of detecting complex temporal and sequential dependencies. The Precision and Recall metrics for LSTM exceed 0.94 and 0.96, respectively, while the F1-score is >0.91, indicating a balance between the model’s accuracy and sensitivity. To thoroughly evaluate the effectiveness of the proposed model, a detailed performance analysis for individual classes was conducted using Precision, Recall, and F1-score metrics. The results, presented in Table 8, reveal that Proper achieved the highest Precision (0.9414) and Recall (0.9695), indicating an excellent fit of the model to this class. In contrast, the lowest Precision (0.6647) and Recall (0.7513) values were observed for Anomaly 4, highlighting some challenges for the model in detecting anomalies in this class.
The experimental results confirm that the proposed hybrid LSTM model outperforms other approaches in terms of anomaly detection effectiveness in production processes. The application of advanced deep learning methods represents a significant step toward improving anomaly detection quality in data with complex temporal structures, which is critically important in the context of industrial process monitoring.
The resulting network model was implemented in the FIBRAIN industrial plant in a technological supervision system with the task of providing real–time support for RLM line operators during the production process. Figure 10 and Figure 11 show examples of the monitoring view during: normal process operation and the occurrence of an anomaly threatening a defect in the final product, respectively. The bar graph represents the probability values of the current process signature belonging to one of the classes: Proper, Anomaly 1–4. For the operator’s convenience, the values on this graph correspond to the output signals of the penultimate Softmax layer of the aggregating module. The system operator’s task was to control and tune the line’s working parameters so that the process state signature recognized by the LSTM network corresponded to its proper operation. The number of distinguishable anomalies for the RLM line was set at 4, based on heuristic assumptions and consultations with experts supervising the production technology on this line.
During the period from 1 January 2024, to 28 June 2024, the constructed LSTM network model was test–integrated into the production process monitoring system. Based on the observations made, the following coincidences between detected anomalies and corrective actions taken were identified:
  • Anomaly 1: Excessive fluctuations in pressure and temperature alter the geometry and texture of the produced item. Significant pressure variations cause diameter changes within the range of 0.2 [mm] to 0.4 [mm], leading to the product being classified as non–compliant. In the cable coating extrusion process, large pressure changes result in discontinuities in the coating material, causing the final product to be divided into short segments, which are often unacceptable to customers.
  • Anomaly 2: Excessive deviations in the pressure and temperature of the hydrophobic gel. Pressure changes lead to variations in the external and internal diameters of the semi–finished product. Changes in the external diameter result in weakened strength at the constriction points of the semi–finished product.
  • Anomaly 3: Excessive production speed and the associated tension force of the production line and winding device. According to experts, this is a crucial anomaly that causes excess fiber in the tube, negatively affecting the transmission properties and strength of the finished fiber optic cable. Moreover, its frequent occurrence indicates wear and tear of the drive and consumable parts of the line.
  • Anomaly 4: Temperature fluctuations in the cooling bath water affect the surface condition of semi–finished and finished products, as well as the dynamics of secondary shrinkage, which negatively impacts semi–finished and finished cables many days after their production.
During the analyzed period of the test implementation, a reduction in the amount of defective product by 66.3% was achieved, which should be considered a significant reduction in unnecessary production costs per unit length of the final product. A notable drawback of the developed approach is the relatively high sensitivity of the monitoring module to sterile measurement conditions that must be maintained when using such a large number of sensors. Its susceptibility to transient spikes or temporary loss of measurement values may pose a significant challenge in adapting the proposed method to similar production processes.

4. Conclusions

Combining statistical methods, machine learning, and sensor-based monitoring provides a robust approach to detecting and addressing anomalies. The proposed approach effectively integrates the multi-resolution K-means clustering method with recursive LSTM network architectures. The analysis conducted in this study on the available data from the monitored RLM production line identified key parameters determining product quality during technological processes: pressure and temperature of the material and gel, and constant production speed, which is crucial for maintaining the correct dimensions of the extruded product. The achieved results indicate the validity of using deep learning with LSTM networks for the analysis and classification of technological data.
This work makes a significant contribution to the field of anomaly detection by applying an innovative hybrid architecture that combines clustering techniques with LSTM networks, enabling the automatic creation of training datasets even in the absence of manually labeled data. The multi-channel temporal data analysis system from production lines provided a new perspective on real-time monitoring of industrial processes, which can also be adapted to other industries. However, certain limitations of the developed model should be noted. The high sensitivity of the system to input data disturbances, particularly when using multiple sensors, necessitates precise measurement conditions. Additionally, the scalability of the model presents a challenge when deploying it across production lines with differing characteristics. Despite these challenges, the proposed method significantly reduces the number of defects and allows for better management of the production process.
The ultimate application of the described system is full integration with production lines in real industrial facilities, enabling further monitoring and optimization of the manufacturing processes for optical fiber tubes. The system facilitates a significant reduction in waste and improvement in the quality of final products, making an essential contribution both economically and environmentally. The application of anomaly detection methods using LSTM in other manufacturing processes, especially within the Industry 4.0 framework, can enhance quality control across a wide range of technologies. Ultimately, the findings of this study demonstrate that further advancements in optical cable production technology and advanced machine learning techniques will continue to improve the precision and efficiency of anomaly detection methods, thus supporting better quality control and the advancement of industrial process automation.

Author Contributions

Conceptualization, Z.G.; methodology, Z.G.; software, Z.G. and E.Z.; validation, Z.G. and E.Z.; formal analysis, Z.G.; investigation, Z.G., E.Z. and L.O.; resources, L.O.; data curation, Z.G.; writing—original draft preparation, Z.G. and E.Z.; visualization, Z.G. and E.Z.; supervision, Z.G.; project administration, Z.G.; funding acquisition, L.O., Z.G. and E.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This project received funding from the National Centre for Research and Development under grant agreement No. POIR.01.01.01-00-1425/20.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Acknowledgments

We gratefully acknowledge the support of the NCBR, as part of the competition: 6/1.1.1/2020 6/1.1.1/2020 SS Duze/MSP/JN 4, project number POIR.01.01.01-00-1425/20-00 and Fibrain Sp. z.o.o, which made this research possible.

Conflicts of Interest

Lukasz Olbrot was employed by the company FIBRAIN Sp. z o.o. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
DNNDeep Neural Network
DWTDiscrete Wavelet Transform
LSTMLong short-term memory
RNNRecurrent neural network
RLRline for tube extrusion
RLVline for coating extrusion on the cable core
RLMline for twisting cable cores from tubes
SVMSupport Vector Machines

References

  1. Gordan, M.; Sabbagh-Yazdi, S.R.; Ismail, Z.; Ghaedi, K.; Carroll, P.; McCrum, D.; Samali, B. State-of-the-art review on advancements of data mining in structural health monitoring. Measurement 2022, 193, 110939. [Google Scholar] [CrossRef]
  2. Kotsiopoulos, T.; Sarigiannidis, P.; Ioannidis, D.; Tzovaras, D. Machine Learning and Deep Learning in smart manufacturing: The Smart Grid paradigm. Comput. Sci. Rev. 2021, 40, 100341. [Google Scholar] [CrossRef]
  3. Ribeiro, R.; Pilastri, A.; Moura, C.; Morgado, J.; Cortez, P. A data-driven intelligent decision support system that combines predictive and prescriptive analytics for the design of new textile fabrics. Neural Comput. Appl. 2023, 35, 17375–17395. [Google Scholar] [CrossRef]
  4. Bock, F.E.; Aydin, R.C.; Cyron, C.J.; Huber, N.; Kalidindi, S.R.; Klusemann, B. A review of the application of machine learning and data mining approaches in continuum materials mechanics. Front. Mater. 2019, 6, 452701. [Google Scholar] [CrossRef]
  5. Soleimani, M.; Naderian, H.; Afshinfar, A.H.; Savari, Z.; Tizhari, M.; Agha Seyed Hosseini, S.R. A Method for Predicting Production Costs Based on Data Fusion from Multiple Sources for Industry 4.0: Trends and Applications of Machine Learning Methods. Comput. Intell. Neurosci. 2023, 2023, 6271241. [Google Scholar] [CrossRef] [PubMed]
  6. Abdallah, M.; Joung, B.G.; Lee, W.J.; Mousoulis, C.; Raghunathan, N.; Shakouri, A.; Sutherland, J.W.; Bagchi, S. Anomaly Detection and Inter-Sensor Transfer Learning on Smart Manufacturing Datasets. Sensors 2023, 23, 486. [Google Scholar] [CrossRef] [PubMed]
  7. Abdelli, K.; Cho, J.Y.; Azendorf, F.; Griesser, H.; Tropschug, C.; Pachnicke, S. Machine Learning-based Anomaly Detection in Optical Fiber Monitoring. J. Opt. Commun. Netw. 2022, 14, 365–375. [Google Scholar] [CrossRef]
  8. Abdula, S.P.; Llagas, M.J.; Fernandez, A.M.; Arboleda, E. Machine Learning Applications for Fault Tracing and Localization in Optical Fiber Communication Networks: A Review. Preprints 2024. [Google Scholar] [CrossRef]
  9. Glass, S.W.; Fifield, L.S.; Spencer, M.P. Transition to Online Cable Insulation Condition Monitoring. In Proceedings of the 2021 48th Annual Review of Progress in Quantitative Nondestructive Evaluation, QNDE 2021, Virtual, 28–30 July 2021. [Google Scholar]
  10. Kakavandi, F.; Gomes, C.; de Reus, R.; Badstue, J.; Jensen, J.L.; Larsen, P.G.; Iosifidis, A. Towards Developing a Digital Twin for a Manufacturing Pilot Line: An Industrial Case Study. In Digital Twin Driven Intelligent Systems and Emerging Metaverse; Springer: Singapore, 2023; pp. 39–64. [Google Scholar]
  11. Kane, A.P.; Kore, A.S.; Khandale, A.N.; Nigade, S.S.; Joshi, P.P. Predictive Maintenance Using Machine Learning; SPD Technology: London, UK, 2022. [Google Scholar]
  12. Pittino, F.; Puggl, M.; Moldaschl, T.; Hirschl, C. Automatic Anomaly Detection on In-Production Manufacturing Machines Using Statistical Learning Methods. Sensors 2020, 20, 2344. [Google Scholar] [CrossRef]
  13. Fan, J.; Wang, Z.; Wu, H.; Sun, D.; Wu, J.; Lu, X. An Adversarial Time–Frequency Reconstruction Network for Unsupervised Anomaly Detection. Neural Netw. 2023, 168, 44–56. [Google Scholar] [CrossRef] [PubMed]
  14. Zhu, W.; Li, W.; Dorsey, E.R.; Luo, J. Unsupervised anomaly detection by densely contrastive learning for time series data. Neural Netw. 2023, 168, 450–458. [Google Scholar] [CrossRef] [PubMed]
  15. Zhang, X.; Shi, S.; Sun, H.; Chen, D.; Wang, G.; Wu, K. ACVAE: A novel self-adversarial variational auto-encoder combined with contrast learning for time series anomaly detection. Neural Netw. 2024, 171, 383–395. [Google Scholar] [CrossRef] [PubMed]
  16. Yao, Y.; Ma, J.; Feng, S.; Ye, Y. SVD-AE: An asymmetric autoencoder with SVD regularization for multivariate time series anomaly detection. Neural Netw. 2024, 170, 535–547. [Google Scholar] [CrossRef] [PubMed]
  17. Li, Q.; Wang, Y.; Dong, J.; Zhang, C.; Peng, K. Multi-node knowledge graph assisted distributed fault detection for large-scale industrial processes based on graph attention network and bidirectional LSTMs. Neural Netw. 2024, 173, 106210. [Google Scholar] [CrossRef] [PubMed]
  18. Fan, J.; Ge, Y.; Zhang, X.; Wang, Z.; Wu, H.; Wu, J. Learning the feature distribution similarities for online time series anomaly detection. Neural Netw. 2024, 180, 106638. [Google Scholar] [CrossRef] [PubMed]
  19. Lyu, S.; Mo, D.; Wong, W.K. REB: Reducing biases in representation for industrial anomaly detection. Knowl.-Based Syst. 2024, 290, 111563. [Google Scholar] [CrossRef]
  20. Kang, B.; Zhong, Y.; Sun, Z.; Deng, L.; Wang, M.; Zhang, J. MSTAD: A masked subspace-like transformer for multi-class anomaly detection. Knowl.-Based Syst. 2024, 283, 111186. [Google Scholar] [CrossRef]
  21. Iqbal Basheer, M.Y.; Mohd Ali, A.; Abdul Hamid, N.H.; Mohd Ariffin, M.A.; Osman, R.; Nordin, S.; Gu, X. Autonomous anomaly detection for streaming data. Knowl.-Based Syst. 2024, 284, 111235. [Google Scholar] [CrossRef]
  22. Hong, J.; Kang, S. Score distillation for anomaly detection. Knowl.-Based Syst. 2024, 295, 111842. [Google Scholar] [CrossRef]
  23. Wang, X.; Li, W.; He, X. MTDiff: Visual anomaly detection with multi-scale diffusion models. Knowl.-Based Syst. 2024, 302, 112364. [Google Scholar] [CrossRef]
  24. Guo, W.; Jiang, P. Weakly Supervised anomaly detection with privacy preservation under a Bi-Level Federated learning framework. Expert Syst. Appl. 2024, 254, 124450. [Google Scholar] [CrossRef]
  25. An, S.; Kim, J.; Kim, S.; Chikontwe, P.; Jung, J.; Jeon, H.; Park, S.H. Few-shot anomaly detection using positive unlabeled learning with cycle consistency and co-occurrence features. Expert Syst. Appl. 2024, 256, 124890. [Google Scholar] [CrossRef]
  26. Shen, L.; Wei, Y.; Wang, Y.; Li, H. AFMF: Time series anomaly detection framework with modified forecasting. Knowl.-Based Syst. 2024, 296, 111912. [Google Scholar] [CrossRef]
  27. Bai, Y.; Zhang, J.; Chen, Z.; Dong, Y.; Cao, Y.; Tian, G. Dual-path Frequency Discriminators for few-shot anomaly detection. Knowl.-Based Syst. 2024, 302, 112397. [Google Scholar] [CrossRef]
  28. Liu, Y.; Ju, B.; Yang, D.; Peng, L.; Li, D.; Sun, P.; Li, C.; Yang, H.; Liu, J.; Song, L. Memory-enhanced spatial-temporal encoding framework for industrial anomaly detection system. Expert Syst. Appl. 2024, 250, 123718. [Google Scholar] [CrossRef]
  29. Di, Y.; Wang, F.; Zhao, Z.; Zhai, Z.; Chen, X. An interpretable graph neural network for real-world satellite power system anomaly detection based on graph filtering. Expert Syst. Appl. 2024, 254, 124348. [Google Scholar] [CrossRef]
  30. Chen, A.; Wu, J.; Zhang, H. FIAD: Graph anomaly detection framework based feature injection. Expert Syst. Appl. 2025, 259, 125216. [Google Scholar] [CrossRef]
  31. Lei, T.; Ou, M.; Gong, C.; Li, J.; Yang, K. An unsupervised deep global–local views model for anomaly detection in attributed networks. Knowl.-Based Syst. 2024, 300, 112185. [Google Scholar] [CrossRef]
  32. Mejri, N.; Lopez-Fuentes, L.; Roy, K.; Chernakov, P.; Ghorbel, E.; Aouada, D. Unsupervised anomaly detection in time-series: An extensive evaluation and analysis of state-of-the-art methods. Expert Syst. Appl. 2024, 256, 124922. [Google Scholar] [CrossRef]
  33. Chi, J.; Mao, Z. Deep domain-adversarial anomaly detection with robust one-class transfer learning. Knowl.-Based Syst. 2024, 300, 112225. [Google Scholar] [CrossRef]
  34. Zhu, T.; Liu, L.; Sun, Y.; Lu, Z.; Zhang, Y.; Xu, C.; Chen, J. Semi-supervised noise-resilient anomaly detection with feature autoencoder. Knowl.-Based Syst. 2024, 304, 112445. [Google Scholar] [CrossRef]
  35. Han, H.; Fan, H.; Huang, X.; Han, C. Self-supervised multi-transformation learning for time series anomaly detection. Expert Syst. Appl. 2024, 253, 124339. [Google Scholar] [CrossRef]
  36. Yu, L.-R.; Lu, Q.-H.; Xue, Y. DTAAD: Dual Tcn-attention networks for anomaly detection in multivariate time series data. Knowl.-Based Syst. 2024, 295, 111849. [Google Scholar] [CrossRef]
  37. Wei, S.; Wei, X.; Ma, Z.; Dong, S.; Zhang, S.; Gong, Y. Few-shot online anomaly detection and segmentation. Knowl.-Based Syst. 2024, 300, 112168. [Google Scholar] [CrossRef]
  38. Amini, A.; Kalantari, R. Gold price prediction by a CNN-Bi-LSTM model along with automatic parameter tuning. PLoS ONE 2024, 19, e0298426. [Google Scholar] [CrossRef] [PubMed]
  39. Huang, W.; Lin, Y.; Liu, M.; Min, H. Velocity-aware spatial-temporal attention LSTM model for inverse dynamic model learning of manipulators. Front. Neurorobotics 2024, 18, 1353879. [Google Scholar] [CrossRef] [PubMed]
  40. Liu, L.; Feng, J.; Li, J.; Chen, W.; Mao, Z.; Tan, X. Multi-layer CNN-LSTM network with self-attention mechanism for robust estimation of nonlinear uncertain systems. Front. Neurosci. 2024, 18, 1379495. [Google Scholar] [CrossRef] [PubMed]
  41. Misiti, M.; Misiti, Y.; Oppenheim, G.; Poggi, J.M. Clustering Signals Using Wavelets. In Computational and Ambient Intelligence; Sandoval, F., Prieto, A., Cabestany, J., Graña, M., Eds.; Springer: Berlin/Heidelberg, Germany, 2007; pp. 514–521. [Google Scholar]
  42. Campos, G.O.; Zimek, A.; Sander, J.; Campello, R.J.; Micenková, B.; Schubert, E.; Assent, I.; Houle, M.E. On the evaluation of unsupervised outlier detection: Measures, datasets, and an empirical study. Data Min. Knowl. Discov. 2016, 30, 891–927. [Google Scholar] [CrossRef]
Figure 1. The production lines used by FIBRAIN Sp. z o.o.: RLR (a), RLV (b) oraz RLM (c).
Figure 1. The production lines used by FIBRAIN Sp. z o.o.: RLR (a), RLV (b) oraz RLM (c).
Applsci 15 01383 g001
Figure 2. Preliminary analysis of the manufacturing process aimed at using the Haar-based DWT transform clustering algorithm.
Figure 2. Preliminary analysis of the manufacturing process aimed at using the Haar-based DWT transform clustering algorithm.
Applsci 15 01383 g002aApplsci 15 01383 g002b
Figure 3. Input data significance distribution.
Figure 3. Input data significance distribution.
Applsci 15 01383 g003
Figure 4. Production RLM line data feautures. The visualisation of the hierarchical, binary cluster tree (a), 3D view of clusters first three principal components (b).
Figure 4. Production RLM line data feautures. The visualisation of the hierarchical, binary cluster tree (a), 3D view of clusters first three principal components (b).
Applsci 15 01383 g004
Figure 5. Production RLM line data feautures with bad sessions indicator p component. The hierarchical, binary cluster tree (a) and 3D view of clusters first three principal components (b) respectively.
Figure 5. Production RLM line data feautures with bad sessions indicator p component. The hierarchical, binary cluster tree (a) and 3D view of clusters first three principal components (b) respectively.
Applsci 15 01383 g005
Figure 6. The diagram of the hybrid LSTM network designed for the classification of production process states.
Figure 6. The diagram of the hybrid LSTM network designed for the classification of production process states.
Applsci 15 01383 g006
Figure 7. LSTM cell from the given channel layer.
Figure 7. LSTM cell from the given channel layer.
Applsci 15 01383 g007
Figure 8. LSTM Layer Diagram.
Figure 8. LSTM Layer Diagram.
Applsci 15 01383 g008
Figure 10. Signalization of proper operation in the analysis of the technological line’s performance.
Figure 10. Signalization of proper operation in the analysis of the technological line’s performance.
Applsci 15 01383 g010
Figure 11. Signalization of proper operation in the analysis of the technological line’s performance.
Figure 11. Signalization of proper operation in the analysis of the technological line’s performance.
Applsci 15 01383 g011
Table 1. Summary of key elements, applied models, and results for selected works on anomaly detection in production processes.
Table 1. Summary of key elements, applied models, and results for selected works on anomaly detection in production processes.
Authors (Year)Research DescriptionApplied ModelsKey Results
Abdallah et al. (2023) [6]Anomaly detection in smart factories with sensor-to-sensor transferLSTM, Transfer learning, Neural networksHigh anomaly detection accuracy, effective knowledge transfer between sensors
Kakavandi et al. (2023) [10]Digital twin for real-time production line monitoringDigital twin, LSTM, Autoencoders40% defect reduction, dynamic adjustment of production parameters
Soleimani et al. (2023) [5]Production cost prediction in Industry 4.0LSTM, Decision trees92% prediction accuracy for costs, reduction of defect-related losses
Abdelli et al. (2022) [7]Anomaly detection in fiber optic monitoringLSTM, SVM94% anomaly detection accuracy, effective detection of optical cable damages
Pittino et al. (2020)  [12]Automatic anomaly detection in production machinesLSTM, Linear regression, k–NN87% anomaly detection accuracy, 25% downtime reduction
Table 2. Sample of process data recorded for RLM line.
Table 2. Sample of process data recorded for RLM line.
Time Stamps
29 November 2023
BAZ1
_iTens
BAZ2
_iLoad
BAZ2
_iMetLo
BAZ2
_iSpeed
EXT1
_iLoad
EXT1
_iSpeed
SPE2
_iLoad
SPE2
_iSpeed
22:17:140.2746586.4271153.984433.526630.95715.460255.0488
22:17:150.2197276.384282153.984433.78330.95717.21855.0879
22:17:160.1831066.494143053.994133.520530.996116.595554.7461
22:17:170.1464846.469733954.04333.557130.95715.594554.6094
22:17:180.1647956.512454853.994133.703630.947316.955654.4238
22:17:190.2014166.59185954.023433.734130.95716.180454.209
22:17:200.2014166.59186854.023433.502230.966816.668754.209
22:17:210.2746586.378177753.955133.544930.966817.462254.1406
22:17:220.2746586.500248653.984433.819630.91816.229254.3359
22:25:290.1281746.79321464454.960934.417731.66023.62549−0.2929
22:25:300.3662116.86035465454.970734.149231.61133.62549−0.2929
22:25:310.17.00073466354.990234.289631.61133.62549−0.2929
22:25:320.16.98242467354.970734.191931.60163.62549−0.2929
22:25:330.1464847.04346468254.970734.301831.57233.62549−0.2929
22:25:340.1464846.92749469154.980534.088131.62113.62549−0.2929
22:25:350.1464846.88477470155.029334.344531.63093.62549−0.2929
22:25:360.1281746.88477471154.980534.027131.65043.62549−0.2929
22:25:370.16.89697472155.009834.112531.57233.62549−0.2929
22:53:520.3479017.885742354479.970738.720746.89453.62549−0.2929
22:53:530.3479017.720952355880.019538.427746.89453.62549−0.2929
22:53:540.3479017.720952357180.019538.427746.95313.62549−0.2929
22:53:550.2197277.733152358580.058638.26946.93363.62549−0.2929
22:53:560.2197277.690432359979.980538.85546.99223.62549−0.2929
22:53:570.3662117.672122361279.990238.360646.96293.62549−0.2929
22:54:030.1464847.59277236968038.836747.0023.62549−0.2929
22:54:040.3479017.452392371080.039138.647546.93363.62549−0.2929
23:16:440.1098636.756594072853.974633.25231.34773.62549−0.2929
23:16:450.2380376.719974073753.994133.459531.30863.62549−0.2929
23:16:460.3845226.799324074654.013733.380131.25983.62549−0.2929
23:16:470.16.933594075554.003933.526631.25983.62549−0.2929
23:16:480.16.927494076653.994133.184831.2012−0.00610−0.1464
Table 3. Corresponding process coordinates names.
Table 3. Corresponding process coordinates names.
Column IndexesName
{[ 4]}{’BAZ1_iSpeed’}
{[30]}{’KAP1_iTens17_20’}
{[31]}{’KAP1_iTens21_24’}
{[32]}{’KAP1_iTens5_8’}
{[57]}{’POF1_iDancMode_N’}
{[58]}{’POF1_iFltCodeTrav’}
{[59]}{’POF1_iLoad’}
{[60]}{’POF1_iReelDiam’}
{[61]}{’POF1_iSpeed’}
{[63]}{’QSD2_ValDel’}
{[49]}{’MES_MeterCnt’}
{[56]}{’POF1_iDancerPos’}
{[50]}{’MES_MeterCnt2’}
{[52]}{’MES_OKaltRe’}
{[40]}{’MES_DWarmY’}
Table 4. Hybrid LSTM networks parameters Channels 1–3.
Table 4. Hybrid LSTM networks parameters Channels 1–3.
NameTypeActivationsLearnables
Seq 1Sequence input232
Seq 1LSTM512Input Weights2048 × 232
Recurrent Weights2048 × 512
Bias2048 × 1
Seq 3Full Connected8/16/32Weights8/16/32 × 512
Bias8/16/32 × 32
Seq 4Softmax8/16/32
Seq 5Classification Output8/16/32
Table 5. Hybrid LSTM networks parameters Aggregation Pipeline.
Table 5. Hybrid LSTM networks parameters Aggregation Pipeline.
NameTypeActivationsLearnables
Channel 16Feature input1
Channel 32Feature input1
Channel 8Feature input1
AggregationConcatenation3
Anomalies-organiserFull Connected50Weights50 × 3
Bias50 × 1
Anomalies-recognizerFull Connected5Weights5 × 50
Bias5 × 1
SoftmaxSoftmax5
AnommalyClassification Output5
Table 6. Training data set information.
Table 6. Training data set information.
NameData
Production lineRLM
Session measurement datesStart timeEnd time
1 June 20211 December 2023
Total measurement time688,896 [s], ∼191 [h]
Number of measurement points232
Size of the moving window256 [s]
Window shift step5
Number of samples in the learning set2691
Training set size70%
Test set size30%
Table 7. Summary of results for each forecasting model, taking into account Precision, Recall, F-1 Score and R 2 indicators (the higher the values, the better the performance).
Table 7. Summary of results for each forecasting model, taking into account Precision, Recall, F-1 Score and R 2 indicators (the higher the values, the better the performance).
MetricLSTMRandom ForestSVMRNN
Precision>0.94~0.85~0.88~0.87
Recall>0.96~0.83~0.85~0.86
F–1 Score>0.91~0.84~0.86~0.86
R 2 >0.86~0.85~0.85~0.85
Table 8. Performance indicator results for each class.
Table 8. Performance indicator results for each class.
NamePrecisionRecallF1-Score
Proper0.94140.96950.9553
Anomaly 10.91800.94380.9307
Anomaly 20.85540.75130.7997
Anomaly 30.91850.91850.9185
Anomaly 40.66470.82480.7368
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gomolka, Z.; Zeslawska, E.; Olbrot, L. Using Hybrid LSTM Neural Networks to Detect Anomalies in the Fiber Tube Manufacturing Process. Appl. Sci. 2025, 15, 1383. https://doi.org/10.3390/app15031383

AMA Style

Gomolka Z, Zeslawska E, Olbrot L. Using Hybrid LSTM Neural Networks to Detect Anomalies in the Fiber Tube Manufacturing Process. Applied Sciences. 2025; 15(3):1383. https://doi.org/10.3390/app15031383

Chicago/Turabian Style

Gomolka, Zbigniew, Ewa Zeslawska, and Lukasz Olbrot. 2025. "Using Hybrid LSTM Neural Networks to Detect Anomalies in the Fiber Tube Manufacturing Process" Applied Sciences 15, no. 3: 1383. https://doi.org/10.3390/app15031383

APA Style

Gomolka, Z., Zeslawska, E., & Olbrot, L. (2025). Using Hybrid LSTM Neural Networks to Detect Anomalies in the Fiber Tube Manufacturing Process. Applied Sciences, 15(3), 1383. https://doi.org/10.3390/app15031383

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop