Next Article in Journal
Response of the bGeigie Nano and CzechRad Monitors to Secondary Cosmic Radiation
Previous Article in Journal
Simultaneous Temperature and Relative Humidity Measurement Using Machine Learning in Rayleigh-Based Optical Frequency Domain Reflectometry
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Edge-Cloud Synergy for AI-Enhanced Sensor Network Data: A Real-Time Predictive Maintenance Framework

by
Kaushik Sathupadi
1,
Sandesh Achar
2,
Shinoy Vengaramkode Bhaskaran
3,
Nuruzzaman Faruqui
4,
M. Abdullah-Al-Wadud
5,* and
Jia Uddin
6,*
1
Google LLC, Sunnyvale, CA 94089, USA
2
Department of Software Engineering, Walmart Global Tech, Sunnyvale, CA 94086, USA
3
Zoom Video Communications, Sanjose, CA 95113, USA
4
Department of Software Engineering, Daffodil International University, Daffodil Smart City, Birulia, Savar, Dhaka 1216, Bangladesh
5
Department of Software Engineering, College of Computer and Information Sciences, King Saud University, Riyadh 11543, Saudi Arabia
6
AI and Big Data Department, Endicott College, Woosong University, Daejeon 34606, Republic of Korea
*
Authors to whom correspondence should be addressed.
Sensors 2024, 24(24), 7918; https://doi.org/10.3390/s24247918
Submission received: 13 November 2024 / Revised: 29 November 2024 / Accepted: 9 December 2024 / Published: 11 December 2024
(This article belongs to the Section Sensor Networks)

Abstract

:
Sensor networks generate vast amounts of data in real-time, which challenges existing predictive maintenance frameworks due to high latency, energy consumption, and bandwidth requirements. This research addresses these limitations by proposing an edge-cloud hybrid framework, leveraging edge devices for immediate anomaly detection and cloud servers for in-depth failure prediction. A K-Nearest Neighbors (KNNs) model is deployed on edge devices to detect anomalies in real-time, reducing the need for continuous data transfer to the cloud. Meanwhile, a Long Short-Term Memory (LSTM) model in the cloud analyzes time-series data for predictive failure analysis, enhancing maintenance scheduling and operational efficiency. The framework’s dynamic workload management algorithm optimizes task distribution between edge and cloud resources, balancing latency, bandwidth usage, and energy consumption. Experimental results show that the hybrid approach achieves a 35% reduction in latency, a 28% decrease in energy consumption, and a 60% reduction in bandwidth usage compared to cloud-only solutions. This framework offers a scalable, efficient solution for real-time predictive maintenance, making it highly applicable to resource-constrained, data-intensive environments.

1. Introduction

1.1. Background and Motivation

The rapid adaptation of sensor networks across various industries has made real-time monitoring, automation, and data-driven decision-making possible. Smart cities’ emergence and expeditious expansion are also the blessings of sensor networks [1]. However, real-time sensor network management is challenging because of the 3 Vs of Big data: volume, velocity, and variety [2]. Preprocessing the data with edge computers is a potential solution to overcome the challenge. It also reduces the latency and bandwidth requirements for data transmission. However, the computational limitations of the cost-effective edge devices curve the scope of utilizing Deep Learning (DL) models [3]. On the other hand, cloud computing offers powerful resources for large-scale data processing but suffers from network latency and potential bandwidth bottlenecks [4]. This dichotomy between cloud and edge computing has created the need for a synergistic framework where edge computers use lightweight models, and heavy computing is delegated to the heavier models hosted in the cloud servers. This hybrid approach set the foundation for predictive maintenance that enhances the operational efficiency of sensor networks, ensuring timely maintenance interventions and reducing downtime [5]. As a result, the predictive maintenance framework effectively optimizes the repair schedules, minimizes disruptions, and extends the lifespan of equipment.

1.2. Problem Statement

Significant advancements have been made in sensor network-based predictive maintenance systems. However, the existing field of research is dominated by reliance on either edge computing or cloud servers [6]. This bifurcation limits their efficiency in handling real-time sensor data streams effectively. On the one hand, edge computers are efficient in low-latency responses. However, their computational capability is unsuitable for complex Artificial Intelligence (AI) models [7]. On the other hand, cloud computing provides extensive computational power. However, transferring the big data introduces latency and bandwidth challenges. In addition, the existing predictive maintenance systems struggle with dynamic workload issues caused by sensor data volume fluctuations [8]. As a result, a rigid reliance on either edge or cloud resources can result in bottlenecks, leading to delayed responses and missed predictions. Ensuring optimal resource utilization is another critical challenge in cloud computing. Inefficient use of computational and network resources can lead to excessive energy consumption and increased operational costs [9]. As a result, it impacts the feasibility of the predictive maintenance frameworks in sensor networks in real-world, real-time applications. A seamless synergy between edge and cloud is required to balance these trade-offs while ensuring continuous sensor operations and accurate maintenance predictions. That is why a hybrid edge-cloud framework is needed to enable real-time decision-making for predictive maintenance systems. The hybrid approach solves the bottlenecks of either an edge or cloud-based approach. However, the problem is to ensure that the framework performs continuous data processing, minimizes energy consumption, and dynamically allocates workloads based on network conditions to prevent operational bottlenecks.

1.3. Research Objectives

Design, implementation, and experimentation with a hybrid edge-cloud framework for AI-enhanced sensor networks with the capability of real-time predictive maintenance that ensures optimal cloud resource consumption is the primary objective of this study. The aim of the framework is to utilize both edge and cloud computing to achieve a balanced trade-off among computing resource use, network latency, and energy consumption. The specific objectives of the research are as follows:
  • Hybrid AI Model Deployment: Developing a lightweight AI model suitable for edge devices for real-time anomaly detection complemented by deep learning models in the cloud for comprehensive analytics.
  • Dynamic Workload Management: Designing a novel algorithm to dynamically balance workloads between edge devices and cloud servers depending on network conditions, data volume, and sensor operations while ensuring uninterrupted performance.
  • Predictive Maintenance: To implement a full-fledged predictive maintenance system that analyzes time-series sensor data, predicts potential equipment failures and optimizes maintenance schedules to facilitate improved operational efficiency.
  • Resource Optimization: Evaluate strategies for efficient resource allocation between the edge and cloud to minimize energy consumption, network bandwidth usage, and latency.
Overall, this paper aims to utilize the advantages of both edge devices and cloud computing by coupling them through a novel real-time predictive maintenance framework that efficiently balances complementary characteristics of both systems while optimizing resource consumption.

1.4. Core Contributions

This paper contributes to the application domain by conducting rigorous studies on edge-cloud synergy. Applying existing AI solutions in under-explored domains introduces a novel solution to the edge-cloud bottleneck problem in sensor networks for real-time predictive maintenance systems. Coining a hybrid AI model deployment methodology equipped with an innovative, dynamic workload management algorithm for time-series data-dependent predictive maintenance to optimize resource utilization, ensuring energy and bandwidth, significantly contributes to sensor network research. The core contributions of this research are outlined as follows:
  • Hybrid AI Model Deployment: We propose the development of lightweight AI models deployed on edge devices to perform immediate anomaly detection. These models operate in conjunction with cloud-based deep learning algorithms, which provide comprehensive analytics and long-term insights.
  • Dynamic Workload Management Algorithm: A novel algorithm for real-time workload offloading is introduced, which dynamically shifts workloads between the edge and cloud-based on sensor activity, network conditions, and computational demands. This ensures seamless operation and prevents bottlenecks.
  • Predictive Maintenance Using Time-Series Data: The framework utilizes AI to analyze sensor data streams, predicting potential equipment failures. By doing so, it enables proactive scheduling of maintenance, reducing downtime and extending equipment lifespan.
  • Resource Optimization for Energy and Bandwidth Efficiency: We investigate resource allocation strategies to optimize the distribution of computational tasks between the edge and cloud. The goal is to minimize energy consumption and bandwidth usage while maintaining low-latency operations.
The remaining parts of this paper have been organized into six different sections. The relevant literature review is presented in Section 2, which supports the objectives of this paper. Section 3 presents the methodological development. The implementation of the proposed system is in Section 4. Section 5 presents the experimental results and evaluates the overall effectiveness of the system. The limitations of the proposed system are highlighted in Section 6. Finally, the paper has been concluded in Section 7.

2. Literature Review

2.1. Edge Computing in AI-Enhanced Sensor Networks

According to the study conducted by Sharma et al. [10], edge computing reduces latency and facilitates real-time decision-making. As a result, it is becoming increasingly relevant to sensor networks. A literature review conducted by Wu et al. [11] shows that the integration of AI algorithms to edge computers in sensor networks has gained widespread popularity for anomaly detection, data filtering, and initial processing. However, only lightweight AI models are suitable for edge devices because of the limited resource capability [12]. Techniques such as model pruning, quantization, and knowledge distillation have been applied to make AI models more suitable for edge deployment [13]. While these approaches enable edge devices to execute basic AI functionalities, they are not enough for complex analytics. The rigorous survey concluded that edge computers are effective for low-latency response. However, it struggles with constraints on computational power, memory, and energy efficiency [14]. This is the main barrier to implementing deep learning models in these devices for advanced analytics, as they require a significant amount of computing resources [15]. This paper aims to contribute to this research domain by presenting a hybrid approach that combines the strength of both edge and cloud computing to enhance the performance of sensor networks.

2.2. Cloud-Based AI Analytics

It is no exaggeration to say that cloud computing offers unlimited computing resources [16]. According to the methodology of [17,18,19], cloud-based platforms are feasible for hosting a complex DL model to develop real-time predictive analytics. Even if the model requires high-capacity storage and processing power, cloud servers are capable of allocating them. Concentrating all data in the cloud, training the DL model on those, and applying the model to develop advanced analytics from multiple sensors can provide insights beyond the capabilities of individual edge devices, helping to identify patterns and correlations across sensor data streams [20]. According to Ren et al. [21], a sensor network produces big data, and transmitting those to the cloud introduces significant network latency. As a result, real-time decision-making suffers from delays, often compromising the quality of services. The study conducted by [22] discusses the bandwidth cost and network resource consumption of big data transmission. According to this study, these costs grow rapidly to an unsustainable level unless appropriately managed. These issues highlight the need for a balanced approach, where the cloud’s analytical strengths are combined with the immediacy of edge processing to optimize predictive maintenance in sensor networks.

2.3. Hybrid Edge-Cloud Frameworks

Hybridization, which combines edge devices and cloud computing to leverage the complementary advantages of these two technologies, is a popular field of research. Multiple advancements have already been made in this area [23]. Grzesik et al. [24] conducted a rigorous study on frameworks that combine edge computing with color servers and summarized the opportunities and challenges. A similar study was conducted in [25] as well. These researches point toward the under-explored solutions to the challenges of managing dynamic workloads in sensor networks, which require intelligent decision-making to prevent bottlenecks and ensure resource efficiency. Moreover, the need for balanced energy consumption and bandwidth optimization are two marginally explored research areas. These gaps underscore the need for optimized workload management algorithms and resource allocation strategies to fully realize the potential of hybrid edge-cloud frameworks.

2.4. Predictive Maintenance in Sensor Networks

Predictive maintenance is emerging as a popular approach in industrial production, smart solutions, and automation because of its potential to minimize downtime and optimize maintenance schedules [26]. The underlying technology is the utilization of time-series data from the sensor networks in combination with AI models to forecast when maintenance is needed [27]. According to Maguluri et al. [28], AI-based predictive maintenance relies on Machine Learning (ML) and Deep Learning (DL) to identify trends, anomalies, and failure precursors within the data streams. Time-series data analysis plays a crucial role here where Recurrent Neural Networks (RNNs) and Long Short-Term Memory (LSTM) models are frequently used [29]. This is where the main challenge arises, as the complexities of these models are too high to run effectively on edge devices. Moreover, the frequency and variety of the data produced by the sensor networks overwhelm the edge and cloud resources. It leads to latency issues, and the systems start consuming excessive energy [30]. However, the state-of-the-art studies [31] and traditional predictive maintenance framework [32] exhibit a lack of mechanisms to balance real-time processing demands with resource efficiency. This calls for the development of hybrid approaches that can leverage edge processing for immediate anomaly detection and cloud analytics for in-depth failure prediction, thus maximizing both accuracy and efficiency.

2.5. Recent Advancements in Hybrid Edge-Cloud Frameworks

A rigorous review conducted by Boiko et al. [33] reveals the recent advancement in Hybrid Edge-Cloud frameworks. KRIOTA [34] is one of these frameworks, which is a situation-aware system for robotics. Although it performs well for predefined environments, unlike the proposed framework, it lacks adaptability to various environments. The ArchGPT developed by [35] comes with dynamic and adaptive algorithms to optimize resource allocation and task distribution. However, it does not explicitly incorporate the sensor network like the proposed framework does. An intelligent water quality monitoring system developed by Shahra et al. [36] is another recent application of edge-cloud hybridization. Bhoi et al. [37] applied the concept of edge-cloud hybridization in power monitoring in electronics systems. However, in these studies, the combination of dynamic workload management, predictive maintenance, and resource optimization is missing as the proposed methodology. It makes the approach presented in this paper the most comprehensive and practical approach, which is adaptable to various applications and capable of maintaining scalability within a wide range of data ingestion variations.

3. Methodology

3.1. Dataset Description and Preprocessing

This study uses a dataset consisting of historical time-series data collected from multiple sensors in an industrial setting. This dataset has been used to train the KNN model for anomaly detection. The LSTM learned to predict sensor failure from the same dataset. Key features include time, sensor readings, operational status, and failure labels.

3.1.1. Dataset Features

The dataset used in this paper contains mainly time-series features. These features are related to anomaly detection and failure prediction, which are recorded on the dataset. Each record contains a Timestamp t that specifies the precise time of the reading. This dataset contains the data of 20 sensors, which have been modeled as X t = { x t 1 , x t 2 , , x t 20 } . Each sensor x t i is the measurement from the i-th sensor at time t. A sample of the dataset is presented in Table 1 where up to six sensor readings are shown. In this dataset, y t is a binary label indicating the normal or anomalous state of operation, which is used to train the KNN model for real-time anomaly detection. The last column of the dataset contains the Failure States z t features. The LSTM network used in this paper has been trained considering z t as the target variable.

3.1.2. Data Preprocessing

The data from the sensors come in normalized form with a range from 0 to 1. As a result, normalization is not performed during preprocessing. However, there are multiple missing values in the dataset, which require cleaning. Sequential segmentation is also necessary for the LSTM network. Initially, the dataset has been analyzed manually using Microsoft Excel. It was observed in the initial analysis that there were missing values. Outliers have also been identified among the data. For consistent analysis, missing values were interpolated linearly, and outliers were removed. Outliers were identified by analyzing each sensor’s standard deviation and setting threshold limits to x t i , which has been measured using Equation (1).
x t i is an outlier if | x t i μ i | > α · σ i ,
In Equation (1), μ i and σ i denote the mean and standard deviation of sensor i, and α is an outlier threshold factor. To train the LSTM network, the data were segmented into sequences to capture temporal dependencies. Each input sequence X t spans n time steps, as expressed in Equation (2), where X t i represents the feature vector at time t i . The n is the sequence length. The target label for each sequence is the failure status z t + k after a prediction horizon k.
X t = { X t n + 1 , X t n + 2 , , X t } ,
There are a total of 123,013 instances in the dataset. It has been split into training, testing, and validation datasets with a ratio of 70:15:15. A temporal split was performed to maintain the chronological order of the data. The splitting supports the union operation, as defined in Equation (3) [38].
D = D train D val D test ,

3.2. Comparative Analysis for Hardware and AI Model Selection

The development of the predictive maintenance framework starts with selecting appropriate hardware and AI models. This has been achieved through comparative analysis. Computational capability, power consumption rate, cost, computational complexity, and accuracy are the selection criteria used in this section.

3.2.1. Edge Device Selection

The network sensors are connected to the edge devices, which perform the initial processing. The more sensors there are, the more devices will be required. Massive sensor networks require a lot of edge devices. That is why going for cost-effective devices is always preferred. However, making the right decision based on the trade-off between computational capability and cost is challenging. Table 2 has been prepared to address this challenge.
According to the comparative analysis in Table 2, the Raspberry Pi Zero 2W offers a balanced trade-off between processing power and power consumption. This trade-off makes it suitable for real-time anomaly detection in resource-constrained environments. Moreover, its low cost and moderate CPU performance support the deployment of lightweight AI models enough to detect the initial anomaly. That is why the Raspberry Pi Zero 2 W has been used as the edge device in this study.

3.2.2. AI and ML Model Selection

The AI model for the edge device must be lightweight and computationally efficient. As the edge devices are resource-constrained environments, complex models that require extensive memory are not feasible. From this perspective, a comparative study has been performed, and the outcomes are listed in Table 3. Based on the feature comparison of Table 3, the KNN has been selected as the AI model of the edge computer [39].
Although Table 3 shows that it has a moderate level of accuracy, it has been selected because of its low computational complexity and memory usage, making it suitable for deployment on the Raspberry Pi Zero 2 W. Although Decision Tree and SVM models provide higher accuracy, they are more resource-intensive, which would strain the edge device’s limited resources. Among the compared models, only KNN offers a reliable balance among computational complexity, memory usage, and accuracy. Considering the trade-off and practicality of the system, this research continued with KNN as the finalized AI model for edge devices.

3.2.3. Cloud Service and Model Selection

The experimental phase of this research requires the capability of handling massive volumes of data with numerous varieties at a very high velocity. That is why the primary selection criteria of the cloud service provider are high scalability and low latency. After the successful implementation of the proposed framework, the cloud server cost will be reduced. With this vision, the cost of provisioning the cloud server was less emphasized. The findings of the comparative analysis on different cloud service providers are listed in Table 4 [40].
Although the AWS Labda is costlier than other service providers, it has been selected for this research, considering its scalability and latency parameters. Furthermore, its ability to handle dynamic workloads is advantageous for deep learning applications that require significant computational resources.

3.2.4. Deep Learning Model Selection

The DL models run on the cloud, where the availability of computational resources is not an issue. That is why the suitability of time-series data analysis and accuracy have been given priority while making the selection. The findings of the exploratory analysis performed in this paper to identify the most suitable DL model are listed in Table 5.
According to multiple state-of-the-art studies, the LSTM model is most suitable for time-series data. Although the RNN is equally competitive in the race, it suffers from performance issues for long sequences. As we deal with big data, the data sequence length may vary significantly. In addition, GRU and CNN are not as promising as LSTM in terms of time-series data analysis [41,42,43,44]. Considering everything, the LSTM network has been selected to deploy in the cloud server to develop the predictive maintenance system.

3.2.5. Addressing Retraining Challenges in Machine Learning Models

Retraining machine learning models is critical for maintaining their performance over time, especially in dynamic environments such as sensor networks. The proposed hybrid framework incorporates strategies to address retraining challenges, ensuring that the KNN and LSTM models remain effective despite changes in data patterns or system conditions.
To minimize the need for complete retraining, the KNN model deployed at the edge is enhanced with incremental learning capabilities. Instead of retraining the model from scratch, the system updates the nearest neighbor data incrementally as new data points are received. This approach significantly reduces the computational overhead and ensures real-time responsiveness. The incremental update is governed by Equation (4), where X n e w represents the newly added data points.
X = X o l d X n e w
For the LSTM model in the cloud, periodic retraining is scheduled to incorporate new data and adapt to evolving patterns. To mitigate the computational burden of retraining, a subset of the most recent data, combined with a representative sample of historical data, is used for retraining. This subset selection ensures that the model captures both recent trends and long-term patterns. The retraining schedule is determined dynamically based on the rate of data drift, which is calculated using the Kullback–Leibler (KL) divergence between the historical and new data distributions, as shown in Equation (5).
D K L ( P | | Q ) = i P ( i ) log P ( i ) Q ( i )
The overall retraining workflow ensures that updates occur seamlessly without interrupting real-time operations. Edge-based incremental updates for KNN occur in parallel with ongoing anomaly detection, while LSTM retraining is conducted during periods of low system activity to minimize the impact on performance. This dual approach balances real-time performance with the adaptability required in dynamic environments.

3.2.6. Justification for the Choice of Metric in KNN

The choice of Euclidean distance as the metric for the KNN method in this study is based on its computational simplicity and suitability for the normalized dataset used. Since the data features are scaled to a range of 0 to 1, the Euclidean distance effectively captures the similarity between data points without being influenced by varying feature scales.
Alternative metrics, such as the Manhattan distance and Minkowski distance, were also considered. However, these metrics showed negligible improvement in classification accuracy during preliminary experiments while incurring additional computational overhead. The comparative analysis of metrics is presented in Table 6, which highlights the trade-offs in accuracy and computational efficiency.
Based on these results, Euclidean distance was chosen as it provides a balance between high classification accuracy and low computational cost, making it well-suited for deployment on resource-constrained edge devices.

3.2.7. Selection of Optimal Parameters for KNN

The performance of the KNN model is influenced significantly by two parameters: the number of nearest neighbors (k) and the window width for time-series segmentation. To ensure optimal parameter selection, this study conducted a grid search using cross-validation on the training dataset.
The number of nearest neighbors, k, was varied across a range of values ( k = 1 to k = 15 ) to identify the setting that maximized classification accuracy. The results of this analysis are presented in Figure 1, which shows the relationship between k and classification accuracy. Based on these results, k = 5 was selected, as it achieved the highest accuracy while maintaining computational efficiency. For time-series segmentation, the window width was varied from 10 to 50 time steps. The optimal window width was determined by evaluating the model’s performance using a combination of metrics, including classification accuracy and latency. The results of this analysis are presented in Table 7, which shows that a window width of 30 time steps provided the best trade-off between accuracy and processing time.

3.3. Data Transfer Protocols and Their Limitations

Effective data transfer is critical in hybrid edge-cloud frameworks, especially for ensuring real-time communication and minimizing latency. Table 8 presents a comparison of commonly used data transfer protocols, focusing on their characteristics, limitations, and suitability for the proposed system.

3.4. Conceptual Design of the Hybrid AI Model

The proposed hybrid AI model combines the KNN model and the LSTM network. The KNN runs from the edge device and is designed to detect anomalies. The LSTM network operates from the cloud server to predict the failure based on the time-series data analysis. This integration enables real-time responsiveness at the edge with deeper, long-term insights from the cloud.

3.4.1. Anomaly Detection at the Edge

The choice of data transfer protocols significantly impacts the performance of anomaly detection at the edge. Protocols such as HTTP introduce higher latency due to their inherent overhead, which can delay real-time detection. MQTT and WebSocket, with their low latency, are more suitable for timely anomaly alerts. However, MQTT’s reduced reliability in high-traffic scenarios may result in dropped messages. CoAP, being lightweight and efficient, is ideal for low-latency operations but struggles with scalability in larger deployments [45]. These trade-offs emphasize the importance of protocol selection in achieving consistent real-time performance.
In the conceptual design, the x t denotes the sensor reading received at time t and X = { x 1 , x 2 , , x n } represent a historical set of readings. The KNN model processes each new reading x t by computing its distance from the k nearest neighbors within the historical dataset X. After that, the anomaly score is calculated using the mathematical principle presented in Equation (6), where the anomaly score is expressed as S ( x t ) .
S ( x t ) = 1 k i = 1 k x t x ( i ) ,
The x ( i ) in Equation (6) represents the i-th nearest neighbor in X. The · represents the Euclidean distance. If S ( x t ) exceeds a predefined threshold S t h r e s h , an anomaly is flagged, prompting a data transfer to the cloud for further analysis. The KNN model used at the edge device is computationally less complex and requires low memory requirements. As a result, it smoothly runs on the Raspberry Pi Zero 2 W.
The Euclidean distance was used as the distance metric of the KNN. It was validated through comparative experiments with alternative metrics such as Manhattan and Minkowski distances. These alternatives showed similar classification accuracy but required higher computational resources, which is a significant constraint for edge devices. The Euclidean distance was, therefore, selected as it aligns with the resource constraints of the edge devices while maintaining high accuracy. The selection of the number of nearest neighbors (k) and the window width for time-series segmentation significantly impacts the performance of the KNN model. Through grid search and cross-validation, k = 5 and a window width of 30 time steps were determined to provide the best trade-off between classification accuracy and computational efficiency. These parameters ensured the model’s effectiveness while adhering to the constraints of the edge devices.

3.4.2. Threshold Calculation for Anomaly Detection

The anomaly detection threshold S t h r e s h is set based on the statistical distribution of anomaly scores S ( X ) = { S ( x 1 ) , S ( x 2 ) , , S ( x n ) } from historical data X. The mean μ S and standard deviation σ S of these scores are calculated using Equations (7) and (8), respectively.
μ S = 1 n i = 1 n S ( x i ) ,
σ S = 1 n i = 1 n ( S ( x i ) μ S ) 2 .
Finally, the threshold score, which is defined as S t h r e s h , is calculated using Equation (9). In this equation, the α is the sensitivity parameter. It is adjustable and used to control the system’s sensitivity to ensure optimal performance with a lower false positive rate.
S t h r e s h = μ S + α · σ S ,
This threshold allows the KNN model to detect meaningful anomalies, prompting cloud analysis only when necessary. That means the cloud server does not need to process each instance of the sensor. As long as the sensors are running properly, the data are not passed to the cloud server. Only when the anomaly score crosses the threshold does the sensor start sending data to the cloud server for further time-series prediction.

3.4.3. Failure Prediction in the Cloud

Once an anomaly is detected, the proposed framework considers the probability of a sensor failure high. From this point, the LSTM network hosted on the cloud server starts receiving data from the sensors directly. The sensor data are time-series data and organized as X t = { x t 1 , x t 2 , , x t n } . The LSTM network takes X t as input and predicts the probability of failure within a future horizon k. The prediction by the LSTM network is modeled as Equation (10).
y p r e d ( t + k ) = M L S T M ( X t ; θ ) ,
In Equation (10), M L S T M ( · ) represents the LSTM model where the θ denotes the parameters after training. The y p r e d ( t + k ) represents the predicted failure probability at time t + k .

3.4.4. Hybrid AI Model Operation Algorithm

The overall workflow of the proposed hybrid AI model has been presented as Algorithm 1. It demonstrates the procedure of integrating the KNN-based anomaly detection at the edge device. At the same time, it presents how the LSTM-based failure prediction in the cloud is integrated into the system. The algorithm establishes a logical connection between the components of the conceptual design, enabling efficient, real-time predictive maintenance.
Algorithm 1 Hybrid AI Model for Predictive Maintenance
1:
Input: Sensor reading x t at time t; historical dataset X = { x 1 , x 2 , , x n } ; predefined threshold S t h r e s h
2:
Output: Failure prediction y p r e d ( t + k ) for future time t + k
3:
Initialize: Load KNN model on Raspberry Pi Zero 2 W; load LSTM model on cloud server
4:
Compute mean μ S and standard deviation σ S of historical anomaly scores S ( X ) using Equations (7) and (8)
5:
Set threshold S t h r e s h = μ S + α · σ S (Equation (9))
6:
while True do                              ▹ Continuous monitoring loop
7:
     Receive new sensor reading x t from the sensor network
8:
     Compute anomaly score S ( x t ) using KNN (Equation (6))
9:
     if  S ( x t ) > S t h r e s h  then
10:
         Flag anomaly and send x t to cloud server
11:
         Collect recent time-series data X t = { x t 1 , x t 2 , , x t n }
12:
         Pass X t to LSTM model on the cloud server
13:
         Predict failure probability y p r e d ( t + k ) = M L S T M ( X t ; θ ) (Equation (10))
14:
         if  y p r e d ( t + k ) exceeds predefined failure probability threshold then
15:
             Trigger maintenance alert and log data for further analysis
16:
         else
17:
             Continue monitoring and anomaly detection at the edge
18:
         end if
19:
     else
20:
         No anomaly detected; continue monitoring with KNN at the edge
21:
     end if
22:
end while

3.5. Dynamic Workload Management Algorithm

The dynamic workload management algorithm designed and implemented in this paper is one of the core contributions of this research. It has been designed to allocate tasks between the edge and cloud computer in response to network conditions, data volume, and computational demands. The primary objective of the algorithm is to optimize workload distribution to minimize latency L and energy consumption E, ensuring uninterrupted sensor operations. The overall workload W ( t ) at time t consists of the workloads handled by the edge W e d g e ( t ) and cloud W c l o u d ( t ) expressed in Equation (11).
W ( t ) = W e d g e ( t ) + W c l o u d ( t ) .
If W e d g e ( t ) and W c l o u d ( t ) are automatically adjusted to complement each other, the latency will be minimized. The latency functions have been defined as expressed in Equation (12) where L e d g e ( W e d g e ( t ) ) is the latency function for the edge computer and L c l o u d ( W c l o u d ( t ) ) is the latency function for the cloud servers.
min W e d g e ( t ) , W c l o u d ( t ) L e d g e ( W e d g e ( t ) ) + L c l o u d ( W c l o u d ( t ) ) ,
However, Equation (12) is subject to constraints on bandwidth B and energy E. The workload is distributed based on bandwidth availability, which is represented in Equation (13).
B W c l o u d ( t ) T t r a n s ,
In the relation among B, W c l o u d ( t ) , and T t r a n s , T t r a n s represents data transmission time to the cloud. In low-bandwidth conditions, more tasks are processed at the edge to maintain responsiveness. The energy consumption constraint is modeled as Equation (14) where E is the energy constraint. The E e d g e and E c l o u d represent energy consumption at the edge and cloud, respectively, with E m a x as the allowable limit.
E = E e d g e ( W e d g e ( t ) ) + E c l o u d ( W c l o u d ( t ) ) E m a x ,
The overall dynamic workload management process follows the process flow presented as Algorithm 2. The incorporation of the dynamic workload management algorithm ensures that the proposed system maintains harmony between the edge computer and the cloud server.
Algorithm 2 Dynamic Workload Management Algorithm
1:
Input: Sensor data x t , bandwidth B, max energy E m a x
2:
Output: Optimal W e d g e ( t ) and W c l o u d ( t ) for each cycle
3:
while True do         ▹ Continuous monitoring and adjustment
4:
    Compute W ( t ) = W e d g e ( t ) + W c l o u d ( t )
5:
    Evaluate L e d g e ( W e d g e ( t ) ) and L c l o u d ( W c l o u d ( t ) )
6:
    if Bandwidth B is constrained then
7:
        Increase W e d g e ( t ) and reduce W c l o u d ( t ) to minimize transmission
8:
    else
9:
        Distribute W c l o u d ( t ) to leverage cloud resources
10:
    end if
11:
    if Energy E t o t a l > E m a x  then
12:
        Adjust W e d g e ( t ) and W c l o u d ( t ) to reduce power usage
13:
    end if
14:
end while

Reproducibility of the Algorithm

To ensure reproducibility, the dynamic workload management algorithm has been enhanced with clear references to previously defined mathematical formulations. The objective function minimizes the sum of latency L ( t ) and energy consumption E ( t ) , as detailed in Equation (11). Latency L ( t ) is defined as the sum of edge latency, cloud latency, and data transmission latency (Equation (12)). Similarly, energy consumption E ( t ) is modeled in Equation (15), capturing contributions from both edge and cloud resources.
The constraints governing bandwidth and energy are critical for managing workload distribution. The bandwidth constraint, expressed in Equation (13), ensures that data transmission to the cloud remains within the available bandwidth B ( t ) . The energy constraint, defined in Equation (14), ensures that the total energy consumption does not exceed the maximum allowable limit E max . Additionally, the workload balancing mechanism ensures that changes in workloads between edge and cloud components are dynamically adjusted to maintain system equilibrium.
By explicitly referencing these equations, the algorithm provides a reproducible framework, allowing researchers and practitioners to replicate and validate the proposed workload management strategy in diverse real-world scenarios.

3.6. Resource Optimization Mechanism

One of the contributions of this paper is an innovative resource optimization mechanism, presented as Algorithm 3, which focuses on minimizing both energy consumption E t o t a l and bandwidth usage B while adhering to latency constraints. The mechanism starts with the principle expressed in Equation (15), where E t o t a l is the total energy.
E t o t a l = E e d g e + E c l o u d + E t r a n s
In Equation (15), E e d g e = W e d g e · P e d g e , E c l o u d = W c l o u d · P c l o u d , and E t r a n s = T t r a n s · P t r a n s . Here, T t r a n s represents the time required for data transmission, and P t r a n s denotes the power consumption rate for transmission. Bandwidth B is determined by data rate R and transmission time T t r a n s , as defined by Equation (16).
B = R · T t r a n s .
The optimization objective is updated to include E t r a n s in Equation (17), which is subject to latency L L m a x , where L = L e d g e ( W e d g e ) + L c l o u d ( W c l o u d ) + L t r a n s ( T t r a n s ) .
min W e d g e , W c l o u d E t o t a l + λ B
Algorithm 3 Resource Optimization Mechanism
1:
Input: Workloads W e d g e , W c l o u d ; power rates P e d g e , P c l o u d , P t r a n s ; data rate R; latency L m a x
2:
while True do           ▹ Iterative loop for continuous optimization
3:
    Calculate E e d g e = W e d g e · P e d g e
4:
    Calculate E c l o u d = W c l o u d · P c l o u d
5:
    Calculate E t r a n s = T t r a n s · P t r a n s using T t r a n s = B R
6:
    Compute total energy E t o t a l = E e d g e + E c l o u d + E t r a n s
7:
    Compute latency L = L e d g e ( W e d g e ) + L c l o u d ( W c l o u d ) + L t r a n s ( T t r a n s )
8:
    if  L > L m a x or E t o t a l > E m a x  then
9:
        Adjust W e d g e and W c l o u d to minimize E t o t a l + λ B
10:
    end if
11:
end while

3.7. Security Considerations in Information Processing

The proposed edge-cloud hybrid framework processes and transmits critical sensor data, making it essential to address potential security threats. Security measures are required at every stage: transmission, storage, and processing. This subsection outlines the primary threats and the corresponding mitigation strategies implemented in the framework.

3.7.1. Transmission Security

During data transmission, the system is vulnerable to threats such as eavesdropping, man-in-the-middle attacks, and data tampering. To mitigate these risks, the proposed framework uses end-to-end encryption via Transport Layer Security (TLS). Additionally, secure protocols such as MQTT with SSL/TLS and CoAP with DTLS ensure data integrity and confidentiality during transmission.

3.7.2. Storage Security

Data stored on edge devices and in the cloud are susceptible to unauthorized access and tampering. To address these risks, all stored data are encrypted using Advanced Encryption Standard (AES) with a 256-bit key. Moreover, access controls and multi-factor authentication are implemented to restrict unauthorized access to storage resources.

3.7.3. Processing Security

During processing, security threats include unauthorized code execution and data leakage. To counter these threats, the cloud environment is configured with secure virtual machine isolation, ensuring that no malicious activity can compromise the integrity of the data processing pipeline. Additionally, intrusion detection systems (IDSs) are employed to monitor processing activities for anomalies.
The integration of these security measures ensures that the framework is robust against common security threats, maintaining data integrity, confidentiality, and availability throughout the information processing lifecycle.

4. Implementation

The sensor network used in this study is part of the packaging industry. There are more than 730 sensors in the entire industry. However, only 20 of them were accessible for research and development purposes. These 20 sensors, annotated in Figure 2, are connected to a roller-based winding machine, which is expressed in Figure 2a, an automated packaging machine presented in Figure 2b, a coating machine illustrated in Figure 2c, and conveyor system, which is in Figure 2d. The list of sensors, their specifications, types, and functionalities are listed in Table 9.

4.1. KNN Development and Implementation

The KNN applied in this paper to detect the anomaly calculates the Euclidean distance between a new data point and a set of historical data points provided in the dataset defined as X = { x 1 , x 2 , , x n } . Equation (18) has been used to calculate a distance where x i R d represents the feature vectors in a d-dimensional space.
D ( x t , x i ) = j = 1 d ( x t , j x i , j ) 2 .
The k-nearest neighbors are selected, and the anomaly score S ( x t ) is calculated as the average distance to these k neighbors, which is mathematically defined in Equation (19). In this equation, x ( i ) represents the i-th nearest neighbor in X. When S ( x t ) exceeds a predefined threshold S t h r e s h , it is considered an anomaly.
S ( x t ) = 1 k i = 1 k D ( x t , x ( i ) ) ,
The learning curve analysis presented in Figure 3 demonstrates the learning progress and, at the same time, validation of the learning progress, which justifies that the KNN is properly trained to detect anomalies. It was trained in Google Colab and the trained model was deployed on Raspberry Pi Zero 2 W. The algorithm to detect the anomaly and process the decision has been implemented using Python and the scikit-learn library.

4.2. LSTM Network for Failure Prediction in the Cloud

The LSTM network designed and developed in this study processed the sequential sensor data X t = { x t 1 , x t 2 , , x t n } , where n is the sequence length. In this network, each LSTM cell contains three ages, and they are an input gate i t , a forget gate f t , and an output gate o t . The working principles of these gates are governed by Equations (20), (21), and (22), respectively [46].
i t = σ ( W i x t + U i h t 1 + b i )
f t = σ ( W f x t + U f h t 1 + b f )
o t = σ ( W o x t + U o h t 1 + b o )
In these gates, W, U, and b are weight matrices and biases. The σ represents the sigmoid activation function. The cell and hidden states are denoted by c t and h t , respectively. The c t and h t have been implemented following their mathematical definitions expressed in Equations (23) and (24) [47].
c t = f t · c t 1 + i t · tanh ( W c x t + U c h t 1 + b c )
h t = o t · tanh ( c t )
The LSTM model’s objective is to predict the probability of failure over a future horizon k based on past readings. The LSTM network used in this paper has been developed by following the working principles explained earlier. The network architecture has been presented in Table 10. It has been implemented on AWS Lambda using Python and the TensorFlow library.
The LSTM network is trained using backpropagation through time (BPTT) to minimize the Mean Squared Error (MSE). Figure 4 illustrates the learning progress of the LSTM network. It shows a consistent decay of MSE of training and validation loss. That means the LSTM network has been trained properly.

4.3. Workflow

The workflow of the proposed system is illustrated in Figure 5. There are a total of 20 sensors in the experimental environment, which is presented in Figure 2. For ease of operation, five sensors were connected to one Raspberry Pi Zero 2 W. To cover all experimenting sensors, a total of four Raspberry Pi Zero 2 W were used. Each of the edge devices has the trained KNN running. The prediction from the KNN is generated in an array form consisting of zeros and ones where zero represents no anomaly, and one represents an anomaly. After that, these arrays are concatenated. If there are anomalies, only then will the concatenated array be transferred to the cloud. Otherwise, the sensor readings are directly sent to the database. On the cloud, the data are processed for the LSTM network. The sequences are formed as well. After processing, the data are passed to the LSTM network for prediction. The prediction from LSTM, sensor readings, and anomaly status stored in the database is used in the Dashboard, where the anomaly report, failure prediction, and device health are displayed.

5. Experimental Results and Performance Analysis

5.1. Anomaly Detection at the Edge

The KNN model operated from the edge device is responsible for real-time anomaly detection and classifying them into Normal or Anomalous classes. As a classification problem, confusion matrix analysis was used to evaluate performance, as illustrated in Figure 6. In this analysis, 18,452 instances have been used from the test dataset. True Positive (TP), True Negative (TN), False Positive (FP), and False Negative (FN) have been calculated from the confusion matrix. These values are used to calculate the accuracy, precision, recall, and F1-score. The definition of these metrics and their values are listed in Table 11.
To further validate the performance of the KNN applied in this paper, k-fold cross-validation has been performed at k = 10 . The performance in the cross-validation has been illustrated in Figure 7. The graphical demonstration shows that the performance of the anomaly detector at the edge devices is consistent for all classes for every fold. This consistent performance proves that the edge anomaly detector is effective in anomaly detection.

5.2. LSTM Failure Prediction at the Cloud

The LSTM network is operated from the AWS Lamda server and designed to predict the probability of failure based on the time-series sensor data. As a regression model, the Mean Absolute Error (MAE) and Root Mean Squared Error (RMSE) were used as evaluation metrics. The definition of MAE and RMSE and their corresponding values are listed in Table 12. The low MAE and RMSE values indicate that the LSTM model provides accurate predictions, enhancing the reliability of failure forecasts in the cloud.
A k-fold cross-validation has been performed at k = 10 to analyze the consistency of the performance of the failure predictor at the cloud. The performance variations are illustrated in Figure 8. The figure’s MAE and RMSE score range is from 0.02 to 0.16. On this marginal range, it is clearly visible that the LSTM maintains consistent performance in predicting sensor failure.

5.3. Latency Performance

The latency plays a crucial role in real-time communication. It has a significant impact on the proposed hybrid edge-cloud framework as well. As a matter of fact, one of the objectives of this paper is to minimize the latency. In the experimental analysis, the system’s latency is assessed by measuring the time taken for anomaly detection at the edge and failure prediction in the cloud. At the same time, we evaluated the end-to-end processing time. The findings from the experimental data are summarized in Table 13. According to the experimental analysis, the latency performance demonstrates the system’s capability for real-time operation. The average end-to-end latency is 500 ms, which is 35% less than the cloud-only solution. This comparison demonstrates that the proposed hybrid framework makes real-time anomaly detection faster.
The summary presented in Table 13 has been prepared from 20 discrete experiments conducted to evaluate the effect of the proposed hybrid framework on latency. Figure 9 illustrates the latency performance compared against a cloud-only baseline. The latency is reduced significantly because of the efficient workload distribution between the edge and the cloud components. The baseline latency of a cloud-only system is 800 ms, as shown in Figure 9. The green shaded area in the figure highlights the improvement in latency after applying the proposed system. This substantial reduction highlights the advantage of the hybrid model in providing real-time responsiveness, making the proposed system highly suitable for latency-sensitive applications in predictive maintenance.

5.4. Energy Consumption Performance

Another objective of the proposed framework is energy efficiency. However, tracking the energy consumption at different end-devices is challenging as it requires additional sensor setup. To avoid the complexities and keep the energy calculation relevant, the energy consumption is measured for anomaly detection at the edge, data transmission, and failure prediction in the cloud, which is not similar to traditional kilowatt/hour calculation. The energy consumption summary is listed in Table 14.
The summary of the energy consumption analysis experiment indicates that the hybrid edge-cloud approach provides significant energy savings. The edge devices are power-efficient computers. As the anomaly is detected at the edge, it saves the energy required to transfer the data and operate the intermediate devices. According to the experimental data, the proposed hybrid framework reduces 28% of energy compared to the cloud-only solution. Another surprising factor discovered during the experiment is that the proposed framework saves an additional 60% of energy savings achieved when operating with edge-only processing during low data flow conditions. The total energy per event remains low enough for scalable deployment in resource-constrained environments.
Figure 10 graphically represents the overall experimental data, the summary of which is in Table 14. It also presents the comparison of the energy savings against the cloud-only system baseline. According to the graphical demonstration, the total energy consumption of the hybrid system consistently falls below the cloud-only baseline, which is highlighted with a green shade. It proves that the proposed system saves substantial energy. The framework’s ability to significantly cut down energy demands demonstrates its suitability for resource-constrained environments, ensuring sustainable operation in real-time predictions.

5.5. Bandwidth Usage Performance

Bandwidth is intertwined with the operational cost and feasibility of large-scale sensor networks. Improper bandwidth management can flood the network with traffic, leading to significant transmission delays and impacting real-time communication. During the experiment, the performance of the proposed framework in bandwidth optimization has been evaluated by measuring data transmitted for anomaly reports, regular updates, and peak data flow. The summary of the findings in the experiment is presented in Table 15.
The bandwidth usage analysis shows that each anomaly detection results in 250 KB of data being sent to the cloud for further analysis, while regular updates consume only 50 KB. During peak events, up to 800 KB may be transferred to handle multiple anomalies. The hybrid approach reduces total data transmission by 70% compared to sending all raw data directly to the cloud, with average bandwidth usage per cycle at 150 KB. This represents a bandwidth saving of 60% compared to a cloud-only model, showcasing the benefits of edge processing.
Figure 11 illustrates the bandwidth usage performance of the edge-cloud framework across 20 experiments, comparing it to a cloud-only baseline. The green shaded area represents bandwidth savings achieved by limiting data transmission to significant anomaly events rather than continuously sending raw data to the cloud. This approach results in a 60% reduction in average bandwidth usage, demonstrating the efficiency of the framework in conserving network resources for predictive maintenance tasks in real-time.

5.6. Comparison with Fog Computing Approaches

Edge computing and fog computing are closely related. Both edge and fog computing are decentralized approaches and contribute to reducing latency [48]. Although these two technologies have some similarities, there are unique characteristics that separate them. Fog computing often involves intermediate nodes positioned closer to data sources than traditional cloud servers, enabling localized data aggregation and processing. It makes the proposed methodology flexible enough to incorporate in the fog computing environment as well. To keep this future scope of research open, a comparison between fog computing and the proposed approach is listed in Table 16. The comparison involves key metrics, including latency, scalability, energy efficiency, and bandwidth usage.

5.7. Comparison with State-of-the-Art Workload Management Algorithms

One of the novel contributions of this article is the dynamic workload management algorithm. It is adaptive and actively balances tasks between the edge and cloud resources. The decision is made based on real-time network conditions, data volume, and energy consumption constraints. However, the state-of-the-art workload management algorithms demonstrate promising performance [49]. That is why the performance of the proposed approach has been compared with the existing workload management algorithms. The comparison has been made using four metrics and they are latency, energy efficiency, and resource utilization. Table 17 summarizes the comparison.
The data presented in Table 17 demonstrate that the proposed method is better at workload balancing than other state-of-the-art methods. Because it is dynamic and adaptive, the proposed algorithm performs better and reduces the latency by 35%. At the same time, it improves energy efficiency by 40%. The proposed method shows better performance in scalability and resource utilization as well. This highlights the novelty of the proposed approach, especially in scenarios where computational resources are constrained.

5.8. Scalability Considerations for Larger Sensor Networks and Higher Data Velocities

Expansion and sudden increases in data volume are common for sensor networks. Any system developed for the sensor network must be scalable to efficiently perform during these events [55]. The proposed framework has been designed to keep these challenges in consideration. First of all, the system can distribute workloads across a network of devices. Each edge device processes data from a subset of sensors. This is how it ensures that computational demands are shared and no single device is overwhelmed. Secondly, the workload management algorithm dynamically redistributes tasks between edge devices and the cloud based on network conditions and data rates. This adaptability ensures that the system remains efficient even as the scale of the sensor network increases. Lastly, the cloud component of the framework utilizes auto-scaling features to allocate additional computational resources when data velocities are high.
This ensures that the LSTM model continues to provide timely predictions without bottlenecks. An experiment was conducted to analyze the effects of these features. The purpose of this experiment is to find out if these features make the system scalable. The performance-related data obtained from this experiment are listed in Table 18. The results listed in this table demonstrate the framework’s ability to maintain low latency and high efficiency under varying conditions, confirming its applicability in large-scale, real-world deployments.

5.9. Comparison with Existing Hybrid Approaches

To further validate the effectiveness of the proposed edge-cloud hybrid framework, a comparative analysis was conducted with existing hybrid approaches available in the literature. The comparison focused on key metrics, including latency, energy efficiency, and bandwidth usage, as summarized in Table 19.
The results indicate that the proposed framework achieves the highest latency reduction (35%), energy efficiency (40%), and bandwidth savings (60%), outperforming other hybrid approaches. Additionally, its scalability is categorized as “Very High” due to the integration of dynamic workload management and cloud auto-scaling mechanisms. These comparisons underscore the novelty and robustness of the proposed system in addressing key challenges in real-time predictive maintenance.

6. Limitation and Future Direction

Despite the remarkable achievement of the proposed edge-cloud hybrid framework in latency, energy consumption, and bandwidth optimization, it has several limitations. These limitations are the scope of improvements that provide the future direction of this study. This section highlights these limitations and future directions.

6.1. System Downtime

The current framework has no expectation handling mechanism if the cloud server fails to respond. It is not uncommon to experience server downtime in cloud computing [59]. However, this paper considers the ideal case only where the cloud systems ensure uninterrupted service. A backup server during the outage of the main server to support the operation of the hybrid framework is a potential solution to this limitation.

6.2. Extendability Issue

The current framework uses a Raspberry Pi Zero 2 W as the edge device. It is a resource-constrained environment with limited capability [60]. Although it is possible to introduce additional features to the existing framework on the cloud server, the scope at the edge node is very limited. From this context, the system suffers from an extendability issue. However, replacing it with a more powerful edge device is an easier solution to this problem. There is a cost-benefit trade-off here, which will be explored in the future scope of this study.

6.3. Constraint in Workload Distribution

Although the dynamic workload management algorithm developed and presented in this paper performs properly, because of the limited computational capability at the edge device, the distribution capability suffers from resource constraints. Enhancing the capability of the edge devices makes extending balancing functionality possible. Subsequent versions of this study will explore the potential of this approach.

6.4. Sensor Lifespan

One critical issue is the limited lifespan of sensors, especially in harsh environmental conditions, where exposure to dust, humidity, or extreme temperatures can degrade sensor performance over time. Additionally, sensor readings can be influenced by environmental noise, leading to inaccuracies in data collection and processing.

6.5. Maintenance Cost

Another significant limitation is the high maintenance cost associated with large-scale sensor networks, as frequent calibration, repair, and replacement are required to ensure consistent performance. For sensors deployed in remote areas, energy efficiency becomes a critical concern, as power supply and data transmission capabilities are often constrained.
Limitations are natural to any system, and the proposed hybrid framework is no exception. However, the way the limitations are handled makes the difference. In this paper, the limitations are considered as the scope of improvements. Further studies are being conducted where these improvements will be presented.

7. Keys Aspects and Discussion

The experimental data prove the effectiveness of the proposed edge-cloud hybrid model. This approach is excellent in anomaly detection and sensor failure prediction and significantly reduces operational costs by optimizing resources. This section highlights key aspects of the proposed method.

7.1. Industrial Automation and Smart Manufacturing

The edge-cloud hybrid model developed in this paper has the potential to be integrated into industrial automation and smart manufacturing systems. As these systems depend on sensor data to make intelligent decisions, the proposed system is an ideal candidate for ensuring optimal resource usage [61]. Future applications of this model could involve integrating additional data sources like machine telemetry, environmental conditions, and usage history to refine predictive accuracy further and improve maintenance schedules.

7.2. Energy Sector and Smart Grids

Another potential field of application of this innovative hybrid model is the energy sector, especially in the smart grids. The modern economy is dependent on the power sector. Even the slightest impact of infrastructural or technical damage on the power sector causing power supply disruption can seriously impact numerous businesses [62]. That is why predictive maintenance is crucial in this sector. The successful integration of the proposed edge cloud can significantly enhance the reliability of power-generating systems, including wind turbines, solar panels, hydro turbines, and other sources managed by the smart grids.

7.3. Environmental Monitoring and Agriculture

The proposed edge-cloud hybrid model has the potential to ensure sustainability in smart environmental monitoring and agriculture. These two sectors are heavily dependent on sensors, which are often damaged because of the harsh environmental effects [63]. This is one of the reasons that significantly impacts the sustainability of this field. However, applying the proposed edge-cloud hybrid model can change and enhance sustainability by predicting sensor failure with minimal computational resources.
The innovative edge-cloud hybrid model supported by the KNN at the edge and LSTM network at the cloud is a technology with the potential of being integrated with a sensor network wherever predictive maintenance is required. It has unlocked a wide range of opportunities, which will be explored in the future scope of this study.

8. Conclusions

This paper presented a novel edge-cloud hybrid framework designed to address key challenges in predictive maintenance for sensor networks. By strategically combining the low-latency capabilities of edge computing with the extensive computational power of cloud servers, this framework achieves significant improvements in latency, energy consumption, and bandwidth usage. The KNN-based anomaly detection model, deployed at the edge, enables real-time anomaly detection, reducing the frequency of data transmissions to the cloud and minimizing latency. The cloud-hosted LSTM network, in turn, provides accurate long-term failure predictions, allowing for timely maintenance scheduling and operational optimization.
The experimental results demonstrate that this hybrid approach provides a 35% reduction in latency and 28% energy savings compared to a cloud-only model. Additionally, the bandwidth usage is reduced by 60%, underscoring the efficiency of processing data locally at the edge before transmission. These results highlight the framework’s applicability in real-time predictive maintenance, particularly in resource-constrained environments. The proposed edge-cloud hybrid framework was also compared with existing hybrid approaches from the literature. As shown in Table 19, it outperformed other methods in latency reduction, energy efficiency, and bandwidth savings while demonstrating superior scalability. These findings further establish the framework’s competitive advantage and its potential for practical applications in resource-constrained environments.
In conclusion, this hybrid edge-cloud approach provides a promising solution for real-time predictive maintenance in sensor networks. The balanced use of edge and cloud resources demonstrated in this study serves as a foundation for further research into scalable, efficient, and cost-effective frameworks for the next generation of smart industrial applications.

Author Contributions

Conceptualization, K.S., S.A., S.V.B., N.F., M.A.-A.-W. and J.U.; methodology, K.S., S.A., S.V.B. and N.F.; software, K.S., S.A. and S.V.B.; validation, K.S., S.A., S.V.B. and N.F.; formal analysis, S.V.B. and N.F.; investigation, K.S., S.A. and N.F.; resources, M.A.-A.-W. and J.U.; data curation, N.F. and S.A.; writing—original draft preparation, K.S., S.A. and N.F.; writing—review and editing, S.V.B., M.A.-A.-W. and J.U.; visualization, K.S., S.A. and S.V.B.; supervision, M.A.-A.-W. and J.U.; project administration, S.A. and N.F.; funding acquisition, M.A.-A.-W. and J.U. All authors have read and agreed to the published version of the manuscript.

Funding

The authors extend their appreciation to the Researchers Supporting Project, King Saud University, Saudi Arabia, for Funding this research work through Project No. RSPD2024R951.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data utilized in this experiment are derived from real-world industrial sources and contain confidential information. These data are available for research purposes upon reasonable request, provided that the recipient agrees not to use them for commercial purposes and not to share them in any form. Upon acceptance of these terms, access to the data will be granted by the corresponding authors.

Conflicts of Interest

Kaushik Sathupadi was employed by Google LLC. Sandesh Achar was employed by Walmart Global Tech. Shinoy Vengaramkode Bhaskaran was employed by Zoom Video Communications. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AIArtificial Intelligence
DLDeep Learning
MLMachine Learning
KNNK-Nearest Neighbors
LSTMLong Short-Term Memory
MAEMean Absolute Error
MSEMean Squared Error
RMSERoot Mean Squared Error
IoTInternet of Things
3VsVolume, Velocity, Variety
AWSAmazon Web Services
TPTrue Positive
TNTrue Negative
FPFalse Positive
FNFalse Negative

References

  1. Karthick Raghunath, K.; Koti, M.S.; Sivakami, R.; Vinoth Kumar, V.; NagaJyothi, G.; Muthukumaran, V. Utilization of IoT-assisted computational strategies in wireless sensor networks for smart infrastructure management. Int. J. Syst. Assur. Eng. Manag. 2024, 15, 28–34. [Google Scholar] [CrossRef]
  2. Aggarwal, A. Identification of quality parameters associated with 3V’s of Big Data. In Proceedings of the 2016 3rd International Conference on Computing for Sustainable Global Development (INDIACom), New Delhi, India, 16–18 March 2016; pp. 1135–1140. [Google Scholar]
  3. Choudhury, A.; Ghose, M.; Islam, A.; Yogita. Machine learning-based computation offloading in multi-access edge computing: A survey. J. Syst. Archit. 2024, 148, 103090. [Google Scholar] [CrossRef]
  4. Ali, S.; Wadho, S.A.; Yichiet, A.; Gan, M.L.; Lee, C.K. Advancing cloud security: Unveiling the protective potential of homomorphic secret sharing in secure cloud computing. Egypt. Inform. J. 2024, 27, 100519. [Google Scholar] [CrossRef]
  5. Gong, L.; Chen, Y. Machine learning-enhanced loT and wireless sensor networks for predictive analysis and maintenance in wind turbine systems. Int. J. Intell. Netw. 2024, 5, 133–144. [Google Scholar] [CrossRef]
  6. Pech, M.; Vrchota, J.; Bednář, J. Predictive maintenance and intelligent sensors in smart factory. Sensors 2021, 21, 1470. [Google Scholar] [CrossRef]
  7. Mukherjee, M.; Guo, M.; Lloret, J.; Zhang, Q. Leveraging intelligent computation offloading with fog/edge computing for tactile internet: Advantages and limitations. IEEE Netw. 2020, 34, 322–329. [Google Scholar] [CrossRef]
  8. Abid, A.; Jemili, F.; Korbaa, O. Real-time data fusion for intrusion detection in industrial control systems based on cloud computing and big data techniques. Clust. Comput. 2024, 27, 2217–2238. [Google Scholar] [CrossRef]
  9. Alhosban, A.; Pesingu, S.; Kalyanam, K. CVL: A Cloud Vendor Lock-In Prediction Framework. Mathematics 2024, 12, 387. [Google Scholar] [CrossRef]
  10. Sharma, M.; Tomar, A.; Hazra, A. Edge computing for industry 5.0: Fundamental, applications and research challenges. IEEE Internet Things J. 2024, 11, 19070–19093. [Google Scholar] [CrossRef]
  11. Wu, N.; Jiang, R.; Wang, X.; Yang, L.; Zhang, K.; Yi, W.; Nallanathan, A. AI-Enhanced Integrated Sensing and Communications: Advancements, Challenges, and Prospects. IEEE Commun. Mag. 2024, 62, 144–150. [Google Scholar] [CrossRef]
  12. Hong, C.H.; Varghese, B. Resource management in fog/edge computing: A survey on architectures, infrastructure, and algorithms. ACM Comput. Surv. (CSUR) 2019, 52, 1–37. [Google Scholar] [CrossRef]
  13. Jiang, Y.; Wang, S.; Valls, V.; Ko, B.J.; Lee, W.H.; Leung, K.K.; Tassiulas, L. Model pruning enables efficient federated learning on edge devices. IEEE Trans. Neural Netw. Learn. Syst. 2022, 34, 10374–10386. [Google Scholar] [CrossRef] [PubMed]
  14. Capra, M.; Peloso, R.; Masera, G.; Ruo Roch, M.; Martina, M. Edge computing: A survey on the hardware requirements in the internet of things world. Future Internet 2019, 11, 100. [Google Scholar] [CrossRef]
  15. Djigal, H.; Xu, J.; Liu, L.; Zhang, Y. Machine and deep learning for resource allocation in multi-access edge computing: A survey. IEEE Commun. Surv. Tutor. 2022, 24, 2449–2494. [Google Scholar] [CrossRef]
  16. Zhao, L.; Sakr, S.; Liu, A.; Bouguettaya, A.; Zhao, L.; Sakr, S.; Liu, A.; Bouguettaya, A. Cloud computing. In Cloud Data Management; Springer: Cham, Switzerland, 2014; pp. 9–20. [Google Scholar]
  17. Schmitt, J.; Bönig, J.; Borggräfe, T.; Beitinger, G.; Deuse, J. Predictive model-based quality inspection using Machine Learning and Edge Cloud Computing. Adv. Eng. Inform. 2020, 45, 101101. [Google Scholar] [CrossRef]
  18. Simaiya, S.; Lilhore, U.K.; Sharma, Y.K.; Rao, K.B.V.B.; Maheswara Rao, V.V.R.; Baliyan, A.; Bijalwan, A.; Alroobaea, R. A Hybrid Cloud Load Balancing and Host Utilization Prediction Method Using Deep Learning and Optimization Techniques. Sci. Rep. 2024, 14, 1337. [Google Scholar] [CrossRef]
  19. Tengku Asmawi, T.N.; Ismail, A.; Shen, J. Cloud failure prediction based on traditional machine learning and deep learning. J. Cloud Comput. 2022, 11, 47. [Google Scholar] [CrossRef]
  20. Raviprasad, B.; Mohan, C.R.; Devi, G.N.R.; Pugalenthi, R.; Manikandan, L.; Ponnusamy, S. Accuracy determination using deep learning technique in cloud-based IoT sensor environment. Meas. Sens. 2022, 24, 100459. [Google Scholar] [CrossRef]
  21. Ren, Y.; Wang, T.; Zhang, S.; Zhang, J. An intelligent big data collection technology based on micro mobile data centers for crowdsensing vehicular sensor network. Pers. Ubiquitous Comput. 2023, 27, 563–579. [Google Scholar] [CrossRef]
  22. Bergui, M.; Najah, S.; Nikolov, N.S. A survey on bandwidth-aware geo-distributed frameworks for big-data analytics. J. Big Data 2021, 8, 40. [Google Scholar] [CrossRef]
  23. Tang, B.; Luo, J.; Zhang, J. Container-based data-intensive application scheduling in hybrid cloud-edge collaborative environment. Softw. Pract. Exp. 2024, 54, 1217–1240. [Google Scholar] [CrossRef]
  24. Grzesik, P.; Mrozek, D. Combining Machine Learning and Edge Computing: Opportunities, Challenges, Platforms, Frameworks, and Use Cases. Electronics 2024, 13, 640. [Google Scholar] [CrossRef]
  25. Xu, X.; Zang, S.; Bilal, M.; Xu, X.; Dou, W. Intelligent architecture and platforms for private edge cloud systems: A review. Future Gener. Comput. Syst. 2024, 160, 457–471. [Google Scholar] [CrossRef]
  26. Ucar, A.; Karakose, M.; Kırımça, N. Artificial intelligence for predictive maintenance applications: Key components, trustworthiness, and future trends. Appl. Sci. 2024, 14, 898. [Google Scholar] [CrossRef]
  27. Prabha, S.; Upadhyay, R.; Goswami, P.K. The Smart Predictive Network Maintenance Model using AI/ML Time Series Analysis in Cloud Networks. In Proceedings of the 2024 International Conference on Optimization Computing and Wireless Communication (ICOCWC), Debre Tabor, Ethiopia, 29–30 January 2024; pp. 1–7. [Google Scholar]
  28. Maguluri, L.P.; Suganthi, D.; Dhote, G.M.; Kapila, D.; Jadhav, M.M.; Neelima, S. AI-enhanced predictive maintenance in hybrid roll-to-roll manufacturing integrating multi-sensor data and self-supervised learning. Int. J. Adv. Manuf. Technol. 2024, 1–10. [Google Scholar] [CrossRef]
  29. Chiang, S.; Zito, J.; Rao, V.R.; Vannucci, M. Time-series analysis. In Statistical Methods in Epilepsy; Chapman and Hall/CRC: London, UK, 2024; pp. 166–200. [Google Scholar]
  30. Leng, J.; Guo, J.; Wang, D.; Zhong, Y.; Xu, K.; Huang, S.; Liu, J.; Yu, C.; Ye, Z.; Liu, Q. Blockchain-of-Things-Based Edge Learning Contracts for Federated Predictive Maintenance Toward Resilient Manufacturing. IEEE Trans. Comput. Soc. Syst. 2024, 11, 7990–8004. [Google Scholar] [CrossRef]
  31. Souza, P.; Ferreto, T.; Calheiros, R. Maintenance Operations on Cloud, Edge, and IoT Environments: Taxonomy, Survey, and Research Challenges. ACM Comput. Surv. 2024, 56, 1–38. [Google Scholar] [CrossRef]
  32. Ren, L.; Liu, Y.; Wang, X.; Lü, J.; Deen, M.J. Cloud–edge-based lightweight temporal convolutional networks for remaining useful life prediction in iiot. IEEE Internet Things J. 2020, 8, 12578–12587. [Google Scholar] [CrossRef]
  33. Boiko, O.; Komin, A.; Malekian, R.; Davidsson, P. Edge-Cloud Architectures for Hybrid Energy Management Systems: A Comprehensive Review. IEEE Sens. J. 2024, 24, 15748–15772. [Google Scholar] [CrossRef]
  34. Chauhan, M.A.; Babar, M.A.; Shen, H. KRIOTA: A framework for Knowledge-management of dynamic Reference Information and Optimal Task Assignment in hybrid edge-cloud environments to support situation-aware robot-assisted operations. Future Gener. Comput. Syst. 2024, 160, 489–504. [Google Scholar] [CrossRef]
  35. Zhang, J.; Xiang, R.; Kuang, Z.; Wang, B.; Li, Y. ArchGPT: Harnessing large language models for supporting renovation and conservation of traditional architectural heritage. Herit. Sci. 2024, 12, 220. [Google Scholar] [CrossRef]
  36. Shahra, E.Q.; Wu, W.; Basurra, S.; Aneiba, A. Intelligent Edge-Cloud Framework for Water Quality Monitoring in Water Distribution System. Water 2024, 16, 196. [Google Scholar] [CrossRef]
  37. Bhoi, S.K.; Chakraborty, S.; Verbrugge, B.; Helsen, S.; Robyns, S.; El Baghdadi, M.; Hegazy, O. Intelligent data-driven condition monitoring of power electronics systems using smart edge–cloud framework. Internet Things 2024, 26, 101158. [Google Scholar] [CrossRef]
  38. Faruqui, N.; Yousuf, M.A.; Chakraborty, P.; Hossain, M.S. Innovative automation algorithm in micro-multinational data-entry industry. In Proceedings of the Cyber Security and Computer Science: Second EAI International Conference, ICONCS 2020, Dhaka, Bangladesh, 15–16 February 2020; Proceedings 2. Springer: Berlin/Heidelberg, Germany, 2020; pp. 680–692. [Google Scholar]
  39. Zhang, C.; Liu, Y.; Tie, N. Forest Land Resource Information Acquisition with Sentinel-2 Image Utilizing Support Vector Machine, K-Nearest Neighbor, Random Forest, Decision Trees and Multi-Layer Perceptron. Forests 2023, 14, 254. [Google Scholar] [CrossRef]
  40. Arpaci, I.; Masrek, M.N.; Al-Sharafi, M.A.; Al-Emran, M. Evaluating the actual use of cloud computing in higher education through information management factors: A cross-cultural comparison. Educ. Inf. Technol. 2023, 28, 12089–12109. [Google Scholar] [CrossRef]
  41. Khan, S.; Kumar, V. A novel hybrid GRU-CNN and residual bias (RB) based RB-GRU-CNN models for prediction of PTB Diagnostic ECG time series data. Biomed. Signal Process. Control. 2024, 94, 106262. [Google Scholar] [CrossRef]
  42. ArunKumar, K.; Kalaga, D.V.; Kumar, C.M.S.; Kawaji, M.; Brenza, T.M. Comparative analysis of Gated Recurrent Units (GRU), long Short-Term memory (LSTM) cells, autoregressive Integrated moving average (ARIMA), seasonal autoregressive Integrated moving average (SARIMA) for forecasting COVID-19 trends. Alex. Eng. J. 2022, 61, 7585–7603. [Google Scholar] [CrossRef]
  43. Siraj, M.S.; Ahad, M. A hybrid deep learning framework using CNN and GRU-based RNN for recognition of pairwise similar activities. In Proceedings of the 2020 Joint 9th International Conference on Informatics, Electronics & Vision (ICIEV) and 2020 4th International Conference on Imaging, Vision & Pattern Recognition (icIVPR), Kitakyushu, Japan, 26–29 August 2020; pp. 1–7. [Google Scholar]
  44. Amalou, I.; Mouhni, N.; Abdali, A. Multivariate time series prediction by RNN architectures for energy consumption forecasting. Energy Rep. 2022, 8, 1084–1091. [Google Scholar] [CrossRef]
  45. Ghaleb, B.; Al-Dubai, A.Y.; Ekonomou, E.; Alsarhan, A.; Nasser, Y.; Mackenzie, L.M.; Boukerche, A. A survey of limitations and enhancements of the ipv6 routing protocol for low-power and lossy networks: A focus on core operations. IEEE Commun. Surv. Tutor. 2018, 21, 1607–1635. [Google Scholar] [CrossRef]
  46. Faruqui, N.; Yousuf, M.A.; Whaiduzzaman, M.; Azad, A.; Alyami, S.A.; Liò, P.; Kabir, M.A.; Moni, M.A. SafetyMed: A novel IoMT intrusion detection system using CNN-LSTM hybridization. Electronics 2023, 12, 3541. [Google Scholar] [CrossRef]
  47. Paula, L.P.O.; Faruqui, N.; Mahmud, I.; Whaiduzzaman, M.; Hawkinson, E.C.; Trivedi, S. A novel front door security (FDS) algorithm using GoogleNet-BiLSTM hybridization. IEEE Access 2023, 11, 19122–19134. [Google Scholar] [CrossRef]
  48. Yousefpour, A.; Fung, C.; Nguyen, T.; Kadiyala, K.; Jalali, F.; Niakanlahiji, A.; Kong, J.; Jue, J.P. All one needs to know about fog computing and related edge computing paradigms: A complete survey. J. Syst. Archit. 2019, 98, 289–330. [Google Scholar] [CrossRef]
  49. Alsadie, D. A Comprehensive Review of AI Techniques for Resource Management in Fog Computing: Trends, Challenges and Future Directions. IEEE Access 2024, 12, 118007–118059. [Google Scholar] [CrossRef]
  50. Ramsauer, R.; Huber, S.; Schwarz, K.; Kiszka, J.; Mauerer, W. Static Hardware Partitioning on RISC-V: Shortcomings, Limitations, and Prospects. In Proceedings of the 2022 IEEE 8th World Forum on Internet of Things (WF-IoT), Yokohama, Japan, 26 October–11 November 2022; pp. 1–6. [Google Scholar] [CrossRef]
  51. Yekta, M.; Shahhoseini, H. A Review on Machine Learning Methods for Workload Prediction in Cloud Computing. In Proceedings of the 2023 13th International Conference on Computer and Knowledge Engineering (ICCKE), Mashhad, Iran, 1–2 November 2023; pp. 306–311. [Google Scholar] [CrossRef]
  52. Cai, L.; Yao, K.; Xian, W.; Gong, C. Edge Computing Resource Management Based on Genetic Algorithm. In Proceedings of the 2023 IEEE 13th International Conference on Electronics Information and Emergency Communication (ICEIEC), Beijing, China, 14–16 July 2023; pp. 153–157. [Google Scholar] [CrossRef]
  53. Abyaneh, A.H.A.; Liao, M.; Zahedi, S. Malcolm: Multi-agent Learning for Cooperative Load Management at Rack Scale. ACM SIGMETRICS Perform. Eval. Rev. 2023, 51, 39–40. [Google Scholar] [CrossRef]
  54. Avinash, S.; Srikar, A.S.; Naik, V.P.; Bhaskaran, S. SDN-based Hybrid Load Balancing Algorithm. In Proceedings of the 2023 3rd International Conference on Intelligent Technologies (CONIT), Hubli, India, 23–25 June 2023; pp. 1–5. [Google Scholar] [CrossRef]
  55. Kavitha, S.; Rao, D.D.; Jain, A.; Sharma, S.; Pandey, S.V.P.R. Effectual Energy Optimization Stratagems for Wireless Sensor Network Collections Through Fuzzy-Based Inadequate Clustering. SN Comput. Sci. 2024, 5, 1–10. [Google Scholar]
  56. Symvoulidis, C.; Kiourtis, A.; Mavrogiorgou, A.; Tom-Ata, J.D.T.; Manias, G.; Kyriazis, D. Dynamic deployment prediction and configuration in hybrid cloud/edge computing environments using influence-based learning. In Proceedings of the 2023 10th International Conference on Electrical Engineering, Computer Science and Informatics (EECSI), Palembang, Indonesia, 20–21 September 2023; pp. 315–320. [Google Scholar] [CrossRef]
  57. Silva, J.; Marques, E.R.B.; Lopes, L.M.B.; Silva, F.M.A. Jay: A software framework for prototyping and evaluating offloading applications in hybrid edge clouds. Softw. Pract. Exp. 2023, 53, 2007–2025. [Google Scholar] [CrossRef]
  58. Chen, Q.; You, Z.; Wen, D.; Zhang, Z. Enhanced Hybrid Hierarchical Federated Edge Learning over Heterogeneous Networks. IEEE Trans. Veh. Technol. 2023, 72, 14601–14614. [Google Scholar] [CrossRef]
  59. Faruqui, N.; Yousuf, M.A.; Kateb, F.A.; Hamid, M.A.; Monowar, M.M. Healthcare as a Service (HAAS): CNN-based cloud computing model for ubiquitous access to lung cancer diagnosis. Heliyon 2023, 9, e21520. [Google Scholar] [CrossRef]
  60. Janek, M. Investigating the Wavelength and Polarization of Electromagnetic Wave using Raspberry Pi Zero W. In Proceedings of the 2022 International Conference on Communications, Information, Electronic and Energy Systems (CIEES), Veliko Tarnovo, Bulgaria, 20–22 November 2022; pp. 1–5. [Google Scholar]
  61. Peruzzini, M.; Prati, E.; Pellicciari, M. A framework to design smart manufacturing systems for Industry 5.0 based on the human-automation symbiosis. Int. J. Comput. Integr. Manuf. 2024, 37, 1426–1443. [Google Scholar] [CrossRef]
  62. Saleem, M.U.; Shakir, M.; Usman, M.R.; Bajwa, M.H.T.; Shabbir, N.; Shams Ghahfarokhi, P.; Daniel, K. Integrating smart energy management system with internet of things and cloud computing for efficient demand side management in smart grids. Energies 2023, 16, 4835. [Google Scholar] [CrossRef]
  63. Shaikh, F.K.; Karim, S.; Zeadally, S.; Nebhen, J. Recent trends in internet-of-things-enabled sensor technologies for smart agriculture. IEEE Internet Things J. 2022, 9, 23583–23598. [Google Scholar] [CrossRef]
Figure 1. Impact of k on KNN classification accuracy.
Figure 1. Impact of k on KNN classification accuracy.
Sensors 24 07918 g001
Figure 2. The sensor locations on four different types of machines.
Figure 2. The sensor locations on four different types of machines.
Sensors 24 07918 g002
Figure 3. The learning curve analysis of the KNN anomaly detector. The graph is divided into three distinct regions: (a) rapid improvement region, where the accuracy improves significantly with increasing training data; (b) stable region, where the rate of improvement slows down; and (c) convergence region, where accuracy plateaus, indicating diminishing returns with additional data. This division highlights the efficiency of the KNN model in achieving high performance with a moderately sized dataset.
Figure 3. The learning curve analysis of the KNN anomaly detector. The graph is divided into three distinct regions: (a) rapid improvement region, where the accuracy improves significantly with increasing training data; (b) stable region, where the rate of improvement slows down; and (c) convergence region, where accuracy plateaus, indicating diminishing returns with additional data. This division highlights the efficiency of the KNN model in achieving high performance with a moderately sized dataset.
Sensors 24 07918 g003
Figure 4. The learning curve analysis of the LSTM failure prediction model. The graph shows three key regions: (a) rapid improvement region, where both training and validation losses decrease substantially; (b) stable region, where the rate of decrease slows, showing consistency; and (c) convergence region, where losses reach minimal levels, indicating a well-trained model with low overfitting risk. These regions emphasize the LSTM’s effective learning and generalization capabilities.
Figure 4. The learning curve analysis of the LSTM failure prediction model. The graph shows three key regions: (a) rapid improvement region, where both training and validation losses decrease substantially; (b) stable region, where the rate of decrease slows, showing consistency; and (c) convergence region, where losses reach minimal levels, indicating a well-trained model with low overfitting risk. These regions emphasize the LSTM’s effective learning and generalization capabilities.
Sensors 24 07918 g004
Figure 5. The workflow of the implemented system.
Figure 5. The workflow of the implemented system.
Sensors 24 07918 g005
Figure 6. Confusion matrix analysis of the anomaly prediction by KNN.
Figure 6. Confusion matrix analysis of the anomaly prediction by KNN.
Sensors 24 07918 g006
Figure 7. The consistency of performance in all evaluation metrics in k-fold cross-validation.
Figure 7. The consistency of performance in all evaluation metrics in k-fold cross-validation.
Sensors 24 07918 g007
Figure 8. The consistency of MAE and RMSE in the LSTM network in k-fold cross-validation.
Figure 8. The consistency of MAE and RMSE in the LSTM network in k-fold cross-validation.
Sensors 24 07918 g008
Figure 9. Latency performance of the proposed edge-cloud framework across 20 experiments, showing reductions in end-to-end latency compared to a cloud-only baseline along with other parameters.
Figure 9. Latency performance of the proposed edge-cloud framework across 20 experiments, showing reductions in end-to-end latency compared to a cloud-only baseline along with other parameters.
Sensors 24 07918 g009
Figure 10. Energy consumption performance of the edge-cloud framework across 20 experiments, showing energy savings compared to a cloud-only baseline.
Figure 10. Energy consumption performance of the edge-cloud framework across 20 experiments, showing energy savings compared to a cloud-only baseline.
Sensors 24 07918 g010
Figure 11. Bandwidth usage performance of the edge-cloud framework across 20 experiments, showing substantial reductions in bandwidth consumption compared to a cloud-only baseline.
Figure 11. Bandwidth usage performance of the edge-cloud framework across 20 experiments, showing substantial reductions in bandwidth consumption compared to a cloud-only baseline.
Sensors 24 07918 g011
Table 1. Sample of dataset with key features.
Table 1. Sample of dataset with key features.
Timestamp x t 1 x t 2 x t 3 x t 4 x t 5 x t 6 y t z t
2024-11-08 00:00:000.750.400.650.800.330.55Normal0
2024-11-08 00:01:000.770.380.680.820.320.56Normal0
2024-11-08 00:02:000.720.410.630.790.350.57Anomalous1
2024-11-08 00:03:000.760.390.660.810.340.58Normal0
Table 2. Comparison of edge devices for real-time anomaly detection.
Table 2. Comparison of edge devices for real-time anomaly detection.
DeviceManufacturerCPURAMPower ConsumptionCost
Raspberry Pi Zero 2 WRaspberry Pi Foundation, Cambridge, UKQuad-core 1.0 GHz512 MBLowLow
NVIDIA Jetson NanoNVIDIA Corporation, Santa Clara, CA, USAQuad-core 1.43 GHz4 GBModerateHigh
Coral Edge TPUGoogle, Mountain View, CA, USADual-core Cortex-A531 GBLowModerate
Arduino Nano 33 IoTArduino AG, Somerville, MA, USAARM Cortex-M0+32 KBVery LowVery Low
Table 3. Comparison of AI models for edge-based anomaly detection.
Table 3. Comparison of AI models for edge-based anomaly detection.
ModelComputational ComplexityMemory UsageAccuracy
K-Nearest Neighbors (KNNs)LowLowModerate
Decision TreeModerateModerateHigh
Support Vector Machine (SVM)HighHighHigh
Naive BayesLowVery LowLow
Table 4. Comparison of cloud computing services for deep learning.
Table 4. Comparison of cloud computing services for deep learning.
Cloud ServiceScalabilityLatencyCost
AWS LambdaHighLowHigh
Google Cloud FunctionsHighModerateModerate
Microsoft Azure FunctionsModerateLowModerate
IBM Cloud FunctionsModerateModerateLow
Table 5. Comparison of deep learning models for cloud-based failure prediction.
Table 5. Comparison of deep learning models for cloud-based failure prediction.
ModelSuitability for Time-SeriesComputational ComplexityAccuracy
LSTMHighHighHigh
GRUModerateModerateModerate
CNNLowHighModerate
RNNHighModerateHigh
Table 6. Comparison of metrics for the KNN method.
Table 6. Comparison of metrics for the KNN method.
MetricClassification Accuracy (%)Computational Complexity
Euclidean Distance95.71Low
Manhattan Distance95.72Moderate
Minkowski Distance (p = 3)95.70High
Table 7. Impact of window width on KNN Performance.
Table 7. Impact of window width on KNN Performance.
Window Width (Time Steps)Classification Accuracy (%)Processing Time (ms)
1093.41100
2094.86120
3095.71140
4095.30160
5094.90190
Table 8. Comparison of data transfer protocols and their limitations [45].
Table 8. Comparison of data transfer protocols and their limitations [45].
ProtocolLatencyBandwidth
Efficiency
Security
Features
Limitations
MQTTLowHighBasic encryption
(TLS)
Limited Quality of Service (QoS)
in high-traffic networks
HTTPModerateModerateRobust (HTTPS)High overhead for frequent small
messages
WebSocketLowModerateBasic encryption
(TLS)
Susceptible to connection
interruptions
CoAPVery LowHighBasic encryption
(DTLS)
Limited scalability for large
deployments
Table 9. Specifications and functionalities of sensors used in the study.
Table 9. Specifications and functionalities of sensors used in the study.
SensorSpecificationsTypeFunctionality
Vibration
Sensor
Range: ±25 g,
Frequency: 10 Hz to 1 kHz,
Sensitivity: 100 mV/g
AccelerometerMonitors vibration levels
to detect mechanical faults
Temperature
Sensor
Range: −50 °C to +200 °C,
Accuracy: ±0.5 °C
Thermocouple,
RTD, Thermistor
Measures temperature to
identify overheating and
lubrication issues
Rotational
Speed Sensor
Range: 0 to 10,000 RPM,
Accuracy: ±1%
Optical,
Magnetic
Monitors rotational speed
to ensure safe operation limits
Proximity
Sensor
Detection Range:
Up to 30 mm
Inductive,
Capacitive,
Ultrasonic
Detects object presence/
absence for positioning
and safety
Optical
Sensor
Range: 10 mm to 2 m,
Resolution: 0.01 mm
Photodetector,
Infrared
Detects objects, measures
distances, verifies positioning
Pressure
Sensor
Range: 0–500 psi,
Accuracy: ±0.25% FS
Piezoresistive,
Capacitive
Measures pressure for
hydraulic/ pneumatic systems
Tension
Sensor
Range: 0–1000 N,
Accuracy: ±0.5% FS
Strain GaugeMeasures material tension to
maintain quality
Thickness
Sensor
Range: 0–100 mm,
Resolution: 0.01 mm
Laser-based,
Ultrasonic
Measures thickness in processes
like rolling and coating
Speed
Sensor
Same as Rotational
Speed Sensor
Optical,
Magnetic
Measures speed of parts in
assembly lines for synchronization
Load
Cell
Range: 0–500 kg,
Accuracy: ±0.1% FS
Strain Gauge,
Piezoelectric
Measures force/load, commonly
for weight measurements
Current
Sensor
Range: 0–100 A,
Accuracy: ±1%
Hall Effect,
Shunt Resistor
Monitors electrical current to
detect overloads or malfunctions
Voltage
Sensor
Range: 0–1000 V,
Accuracy: ±0.5%
Potential
Transformer,
Hall Effect
Measures voltage levels for
safe operation
Table 10. Characteristics and features of the LSTM network for failure prediction.
Table 10. Characteristics and features of the LSTM network for failure prediction.
FeatureDescription
Input Sequence Length30 time steps (sensor data sequences)
Input Dimensions20 (number of sensor features per time step)
LSTM Layers2 layers
Units per LSTM Layer50 units
Activation FunctionSigmoid for gates, Tanh for cell states
Output LayerFully connected layer for regression output
Loss FunctionMean Squared Error (MSE)
OptimizerADAM
Learning Rate0.001 (initial learning rate)
Batch Size32
Epochs100
Dropout Rate0.2 (for regularization to prevent overfitting)
Training DataHistorical time-series sensor data
Deployment EnvironmentAWS Lambda
ScalabilityAutomatically scales with AWS Lambda for large datasets
Prediction TargetFailure probability within future horizon
Table 11. Performance metrics for KNN anomaly detection at the edge.
Table 11. Performance metrics for KNN anomaly detection at the edge.
MetricFormulaValue
Accuracy T P + T N T P + T N + F P + F N 95.71%
Precision T P T P + F P 96.79%
Recall T P T P + F N 94.71%
F1-Score 2 · Precision × Recall Precision + Recall 95.74%
Table 12. Performance metrics for LSTM failure prediction in the cloud.
Table 12. Performance metrics for LSTM failure prediction in the cloud.
MetricFormulaValue
Mean Absolute Error (MAE) 1 N i = 1 N | y a c t u a l ( i ) y p r e d ( i ) | 0.12
Root Mean Squared Error (RMSE) 1 N i = 1 N ( y a c t u a l ( i ) y p r e d ( i ) ) 2 0.15
Table 13. Latency performance of the edge-cloud framework.
Table 13. Latency performance of the edge-cloud framework.
MetricDescriptionValue (ms)
Edge Latency (Detection)Average time for KNN anomaly detection120
Cloud Latency (Prediction)Average time for LSTM failure prediction300
Data Transmission LatencyAverage time to transmit data from edge to cloud80
End-to-End LatencyTotal time for edge detection and cloud prediction500
Peak Latency (Under Load)Max observed latency during high data flow600
Latency Reduction (vs. Cloud-only)Improvement over a cloud-only solution35%
Table 14. Energy consumption performance of the edge-cloud framework.
Table 14. Energy consumption performance of the edge-cloud framework.
MetricDescriptionValue (J)
Edge Energy Consumption (Idle)Baseline energy consumption in idle state0.30
Edge Energy Consumption (Detection)Energy per anomaly detection event (KNN)0.45
Cloud Energy Consumption (Prediction)Energy per failure prediction event (LSTM)2.30
Data Transmission EnergyEnergy for transmitting data to cloud per event0.12
Total Energy Consumption (Edge + Cloud)Combined energy for detection, transmission, and prediction2.87
Energy Reduction (vs. Cloud-only)Energy savings compared to cloud-only processing28%
Energy Efficiency (Edge-only Processing)Energy saved when processing anomalies at edge only60%
Table 15. Bandwidth usage performance of the edge-cloud framework.
Table 15. Bandwidth usage performance of the edge-cloud framework.
MetricDescriptionValue (KB)
Data per AnomalyAverage data sent to cloud per anomaly detected250
Regular Update DataData sent periodically to update cloud without anomalies50
Peak Data TransferData transferred during peak anomaly events800
Data Reduction (vs. Raw Transmission)Reduction from transmitting only anomalies70%
Average Bandwidth UsageAverage bandwidth used per operation cycle150
Bandwidth Savings (vs. Cloud-only)Savings from processing data at the edge60%
Table 16. Comparison of proposed hybrid framework with fog computing approaches.
Table 16. Comparison of proposed hybrid framework with fog computing approaches.
MetricFog ComputingProposed Hybrid Framework
LatencyVery Low (localized processing)Low (edge anomaly detection,
cloud prediction)
ScalabilityLimited by fog node resourcesHigh (scalable cloud resources)
Energy EfficiencyModerate (fog nodes consume
power)
High (edge devices are energy-
efficient)
Bandwidth UsageLow (localized processing
reduces data transfer)
Low (edge filtering reduces data
transfer)
Computational
Complexity
High (fog nodes handle
extensive tasks)
Balanced (edge for lightweight tasks,
cloud for heavy analytics)
Table 17. Comparison of workload management algorithms.
Table 17. Comparison of workload management algorithms.
AlgorithmLatency
Reduction (%)
Energy
Efficiency (%)
Resource
Utilization (%)
Scalability
Static Partitioning [50]152065Moderate
Heuristic-Based
Approach [51]
202570High
AI-Driven Dynamic [52]303585High
Multi-Agent Learning
(Malcolm) [53]
283480High
SDN-Based Hybrid
Load Balancing [54]
253075High
Proposed Algorithm354090Very High
Table 18. Scalability analysis of the proposed framework.
Table 18. Scalability analysis of the proposed framework.
ScenarioSensor CountData Velocity (MB/s)Latency (ms)Energy Efficiency (%)
Small Network20150090
Medium Network1001060085
Large Network5005070080
Very Large Network100010080075
Table 19. Comparison of hybrid frameworks.
Table 19. Comparison of hybrid frameworks.
ApproachLatency Reduction (%)Energy Efficiency (%)Bandwidth Savings (%)Scalability
Dynamic Deployment Framework [56]202550Moderate
Jay: Offloading Framework [57]303555High
Hybrid Federated Edge Learning [58]253060High
Proposed Edge-Cloud Framework354060Very High
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sathupadi, K.; Achar, S.; Bhaskaran, S.V.; Faruqui, N.; Abdullah-Al-Wadud, M.; Uddin, J. Edge-Cloud Synergy for AI-Enhanced Sensor Network Data: A Real-Time Predictive Maintenance Framework. Sensors 2024, 24, 7918. https://doi.org/10.3390/s24247918

AMA Style

Sathupadi K, Achar S, Bhaskaran SV, Faruqui N, Abdullah-Al-Wadud M, Uddin J. Edge-Cloud Synergy for AI-Enhanced Sensor Network Data: A Real-Time Predictive Maintenance Framework. Sensors. 2024; 24(24):7918. https://doi.org/10.3390/s24247918

Chicago/Turabian Style

Sathupadi, Kaushik, Sandesh Achar, Shinoy Vengaramkode Bhaskaran, Nuruzzaman Faruqui, M. Abdullah-Al-Wadud, and Jia Uddin. 2024. "Edge-Cloud Synergy for AI-Enhanced Sensor Network Data: A Real-Time Predictive Maintenance Framework" Sensors 24, no. 24: 7918. https://doi.org/10.3390/s24247918

APA Style

Sathupadi, K., Achar, S., Bhaskaran, S. V., Faruqui, N., Abdullah-Al-Wadud, M., & Uddin, J. (2024). Edge-Cloud Synergy for AI-Enhanced Sensor Network Data: A Real-Time Predictive Maintenance Framework. Sensors, 24(24), 7918. https://doi.org/10.3390/s24247918

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop