You are currently viewing a new version of our website. To view the old version click .
Mathematics
  • Article
  • Open Access

31 July 2024

Organized Optimization Integration Validation Model for Internet of Things (IoT)-Based Real-Time Applications

,
,
and
1
Department of Industrial Engineering, Faculty of Engineering, University of Tabuk, Tabuk 47512, Saudi Arabia
2
Department of Civil Engineering, Faculty of Engineering, University of Tabuk, Tabuk 47512, Saudi Arabia
3
Industrial Engineering Department, College of Engineering, University of Bisha, Bisha 61922, Saudi Arabia
*
Authors to whom correspondence should be addressed.
This article belongs to the Special Issue Internet of Things Security: Mathematical Perspective

Abstract

Emerging technology like the Internet of Things (IoT) has great potential for use in real time in many areas, including healthcare, agriculture, logistics, manufacturing, and environmental surveillance. Many obstacles exist alongside the most popular IoT applications and services. The quality of representation, modeling, and resource projection is enhanced through interactive devices/interfaces when IoT is integrated with real-time applications. The architecture has become the most significant obstacle due to the absence of standards for IoT technology. Essential considerations while building IoT architecture include safety, capacity, privacy, data processing, variation, and resource management. High levels of complexity minimization necessitate active application pursuits with variable execution times and resource management demands. This article introduces the Organized Optimization Integration Validation Model (O2IVM) to address these issues. This model exploits k-means clustering to identify complexities over different IoT application integrations. The harmonized service levels are grouped as a single entity to prevent additional complexity demands. In this clustering, the centroids avoid lags of validation due to non-optimized classifications. Organized integration cases are managed using centroid deviation knowledge to reduce complexity lags. This clustering balances integration levels, non-complex processing, and time-lagging integrations from different real-time levels. Therefore, the cluster is dissolved and reformed for further integration-level improvements. The volatile (non-clustered/grouped) integrations are utilized in the consecutive centroid changes for learning. The proposed model’s performance is validated using the metrics of execution time, complexity, and time lag.

1. Introduction

Navigating the intricate landscape of IoT integration within real-time applications is crucial for ensuring optimal functionality [1]. Challenges arise from variable execution times and resource management intricacies, necessitating innovative strategies for effective resolution [2]. Data integration processes with optimized energy utilization need to consider an IoT communication system [3]. IoT integrates and organizes sensors and devices for optimal analysis and administration. IoT can manage real-time data from numerous sources for decision-making and analysis. IoT devices monitor and communicate data to improve real-time operations and procedures. Adaptability emerges as a cornerstone in addressing the evolving demands of integration, facilitating seamless operation. Drawing insights from past integration experiences is an asset, informing future strategies and enhancing overall efficiency [4]. Rigorous performance evaluations play a pivotal role in validating the effectiveness of integration approaches and ensuring optimal outcomes [5]. Accurate complexity estimation is fundamental for successful deployment, which guides decision-making processes and resource allocation. Meticulous attention to detail and a commitment to continuous improvement are indispensable in navigating the complexities of IoT integration and ensuring the efficacy of real-time applications [6].
Efficiently managing IoT integration complexities within real-time applications is critical for smooth operation. These complexities, ranging from fluctuating execution times to intricate resource management, present significant challenges that require innovative solutions [7]. Harmonizing integration processes and ensuring validation accuracy are crucial components, providing seamless functionality [8]. Flexibility is essential in adapting to the dynamic nature of integration requirements, facilitating agility and responsiveness [9]. Leveraging insights from past experiences enables the refinement of integration strategies, enhancing overall effectiveness and performance. Rigorous performance evaluations validate the success of integration efforts, guiding future optimizations [10]. Accurate complexity estimation guides decision-making processes and resource allocation, ensuring efficient deployment. Continuous refinement and improvement are vital for achieving seamless integration and optimizing the performance of real-time IoT applications [11]. The associated cloud framework integrates an efficient k-means clustering technique with an accurate diabetes prediction that manages enormous dynamic data [12].
There are many reasons for implementing IoT for real-time applications. Some of these include enhancing the quality of everyday life, utilizing resources more efficiently, fostering a stronger connection between humans and the environment, enhancing transportation for everyone, ensuring global internet connectivity, simplifying use, and enhancing accessibility. These many benefits have propelled IoT to the status of a new industrial revolution. Some of the most lauded uses of IoT include “smart” houses, farms, transportation, healthcare (through the Internet of Medical Things, or IoMT), businesses, and supply chains. Healthcare is one actual use of IoT, which offers many benefits over older medical treatment methods. Optimal hospital administration, cost and risk reduction in medical procedures, faster diagnosis, better use of healthcare resources, telemedicine, self-service Medicare, early illness detection, and support for these applications are all possible thanks to their development and implementation. IoT presents many opportunities as well as numerous obstacles. The architecture is the biggest concern when deploying IoT in real-time. Companies like Cisco, Intel, and IBM have built their reference architectures around how they approach creating IoT technologies; thus, there is no universally accepted IoT architecture. Interchangeability, security, reliability, safety, variation, data management, and connectivity are just a few of the obstacles the IoT architectural standards must overcome. With the proliferation of smart devices comes an explosion in data, which in turn causes problems with data management. Devices from different manufacturers, use of various protocols, and adherence to different architecture standards complicate communication in a heterogeneous environment. Because IoT devices communicate via multiple methods, interoperability is a significant concern. Protecting users’ data from disclosure in the IoT ecosystem is crucial for ensuring privacy and building consumers’ trust. Ensuring IoT’s scalability is critical for devices to be able to adjust to future needs.
Several companies use one service technique to maximize the efficiency of their resources. Unfortunately, few consider the risks of ignoring asset deterioration early in defect creation. The three main types of maintenance are corrective, preventive, and predictive. Predictive and corrective approaches prioritize asset age and maintenance periodicity. These assets often break down between checks, increasing maintenance costs and safety issues. Many companies believe that you can predict when an asset will fail by examining its service life and performing breakdown maintenance. IoT connects physical objects to the internet so they may exchange data. This capacity could simplify massive data collection, boosting organizations’ bottom lines and forecasts. Embedded hardware, including sensors and other IoT-managed intelligent equipment, is transforming production and industrial maintenance. When used for predictive maintenance, IoT can be practical.
For real-time applications to run smoothly, it is essential to efficiently manage IoT integration complexity. These intricacies, which include complex resource management and execution durations that might vary greatly, pose big problems that need creative answers. Investigating learning strategies that might reduce integration complexity within IoT paradigms is vital for more efficient operation. These methods provide helpful information for improving the efficacy and efficiency of integration procedures in real-time applications. To achieve this goal of constant improvement, it is essential to embrace fresh approaches and use previous experiences. Organizations can improve their efficiency and adaptability by adjusting and perfecting their integration methods in response to what they learn. Evaluation of integration performance is crucial for driving optimization efforts and ensuring continual improvement and refinement. Users can use the intelligent environment’s many interoperable capabilities to communicate with real-time IoT applications. The organized optimization validation model proposed in this paper aims to identify and reduce complexity in IoT integration for large applications. By separating unstructured classes based on integration degree, time latency, and complexity using modified k-means clustering, the performance of the proposed model is validated through evaluation and verification using simulation metrics.
Exploring learning techniques to mitigate integration complexity within IoT paradigms is essential for streamlined operations. These techniques offer valuable insights into optimizing integration processes for real-time applications, fostering efficiency and effectiveness [13,14]. Embracing innovative methodologies and leveraging past experiences are key strategies in this pursuit, facilitating continuous improvement [15]. By adapting and refining integration strategies based on learned insights, organizations can increase efficiency and adaptability. A rigorous evaluation of integration performance ensures constant improvement and refinement, driving optimization efforts [16,17]. Attention to the multiple interoperable features that allow users to communicate with real-time IoT applications exploits the advantages of an intelligent environment [18]. By transferring data processing responsibilities from overloaded edge devices to unused cloud servers, an industrial IoT system and smart agriculture can dynamically reallocate computing power according to the OOIVM paradigm. This helps to reduce processing latency and prevent system bottlenecks.
The proposed model’s contribution is presented as follows:
  • The organized optimization validation model identifies and reduces complexity in IoT integration for large-scale applications.
  • Modified k-means clustering is used to segregate non-organized classifications based on complexity, time lag, and integration levels.
  • The verification and assessment of the proposed model are conducted using simulation-related metrics to validate the proposed model’s performance.

3. Organized Optimization Integration Validation Model (O2IVM)

Based on these contributions, a diagrammatic illustration of the proposed model’s flow is presented in Figure 1. The process flow of the abstract is as follows: The integration of the IoT application under different service levels is clustered based on complexity and lags. The grouped possibilities are used to verify and validate integration improvements. The changes in grouping deviations are used to train the clustering process. The proposed model, OOIVM, must process dynamic, constant information streams coming from a considerable number of IoT devices. This makes it hard to keep processing in real time, which means the system will be slow to respond to new data patterns. The model takes into account the difficulties of grouping and processing overhead. This enhances the chances of integration, reducing complexities (Figure 1).
Figure 1. (a) Diagrammatic illustration of the proposed model’s flow. (b) Flow graph representation of the proposed model.
By streamlining application processes, the IoT paradigm solves the complexity occurrence identification challenge. In this case, data from two sources are used: complexity addressing a problem using machine learning and organized optimization modeling of its outputs. IoT in real-time applications usually involves continuous data transmission to the infrastructure. At all times, numerous sensors and gadgets gather data and send it to a centralized location for processing via the web or regional networks. The internet implementation of the optimization challenge is also standard. To respond quickly and efficiently to shifting situations, the system optimizes and adjusts settings in real time using data. This online tuning is essential to keeping IoT applications running well and reliably in ever-changing contexts.
The need for complexity analysis and grouping is determined by the nature of the real-time applications for improving their resource projection quality, modeling, and representation based on interactive devices/interfaces. In the proposed O2IVM, for precise integration evaluation, non-optimized classification and prioritization are devised using the centroid deviation knowledge of the organized integration levels. Moreover, in most organized integration cases, the output of the problem of sequential centroid changes by one criterion and is reduced to prevent validation lags with accurate measures.
This flowchart is connected to the primary objective of enhancing efficiency, scalability, and immediate processing in IoT systems. The data collection phase involves IoT devices collecting data, which is subsequently analyzed using k-means clustering for improved management. Optimization and validation guarantee the ongoing efficiency and accuracy of the system—the driving force behind the desire to decrease intricacy and enhance data administration. Using k-means clustering, OOIVM organizes comparable data points into groups, resulting in a more controllable and efficient system. OOIVM optimizes resource allocation by adjusting resources according to the current system load. This approach successfully uses computing and network resources, reducing costs and improving performance. The OOIVM undergoes continuous validation and refinement to ensure its adaptability to new data and circumstances, maintaining its effectiveness and reliability.
For using centroids for a single entity, harmonized service levels are required to minimize additional complexity demands and thereby confine complexity lag problems based on centroid deviation.
= 1 , , n , , N m a x , n = 1 , N ¯
where = 1 , , n , , N is the number of real-time applications over the IoT platform. The complexity occurs at the time of device integration based on the service level changes. The model utilizes the complexity analysis, ω = ω 1 , , ω n , , ω N , and identification of currently processing applications, = 1 , , n , , N , for precisely addressing the problems above. The integration is evaluated as the number of service levels observed at different N instances. In this organized integration case, the validation lags, , are due to the non-optimized classification issues of n . The variable ε t signifies the execution time of currently processing applications based on their requirements. This analysis differentiates the same properties of IoT applications and determines the identification of complexities, a = a 1 , , a s , , a S , which contains the same ratio of non-optimized classification, , and centroid deviation, ∆; for the condition, a and a represent identified complexity levels, and a ,   a ε t   a a occurs if and only if
ε t = 1 n i = 1 n a + a w h e r e , a = 1 π a ε t ε t d . ε t a n d a = 1 π a ε t ε t d . ε t
such that
ω a a
In Equation (2), IoT application integrations are estimated for the applications defined in Equation (1). The specific service level is represented as a s , s = 1 , α ¯ ; it either has a single entity or a harmonized service level is considered. The mechanism for minimizing the integration complexity can be prevented with the following procedures.
Non-optimized integration classification is performed based on service level changes in real-time application representation; it also contains optimal outputs of a s , s = 1 , α ¯ required from the output of integrated optimization for each of the specific service level, n , n = 1 , N ¯ . The introduction of a single entity between the harmonized service levels is n , n = 1 , N ¯ , which allows the organized integration cases to be addressed and segregated to augment resource projection quality through appropriate optimized integration, , as a function of the harmonized service level, = 1 , N . The condition of a and a ( a a ) occurs if and only if
= N ω N a s s u c h   t h a t a r g   min t ω N ε t
where
a a ε t
Thus, the organized optimization integration of a is introduced based on grouping the service levels according to condition (2). The identification of accurate integration complexity, a 0 ω , in the initial non-optimized classification problem is reduced to minimize the problem of additional complexity demands as follows:
n = n . a ε t ω = 0
and
N = ω n a ω ε t ω 0
such that
m i n a ω a
In this case, the service level is defined by a single entity with weight measurements; the complexity in the organized integration = 1 , , n is minimized and prioritized. This is achieved using k-means clustering with weight measurements, implementing the intense complexity minimization levels for detecting precise complexity levels:
a s = N × a × ε t     n ε t   ,     ω = 0 n × N a 0 × ε t     n , ω ε t ,     ω 0
such that
a = a
Complexity minimization (5) can be performed to require the varying execution time, ε t , only if the harmonized service level, n , n = 1 , N ¯ , satisfies the following complexity demands. Specific service levels, n , n = 1 , N ¯ , are suitable for minimizing and prioritizing complexity in non-optimized integrations. Each level can be assigned for certain non-optimized classifications, differentiating the non-optimized integrations in grouping to other service levels. Specific service levels, n , n = 1 , N ¯ , are used to prevent validation lags or have a single entity.
Figure 2 above presents the analysis for different devices under lags and levels. The lags are computed using ω a 1 = ω a and n = n . a conditions. For the different levels of integration, a = ω n is the computable complexity that halts N. In both cases, the lag detection is defined using halted and N = N 1 instances such that the complexity-causing factors are identified. The proportion of lag, level, or both in the identified complexity is used to group and independently identify its cause.
Figure 2. Time-based complexity analysis.
Vital to improving the precision and dependability of current information validation to guarantee the quality of choices and actions supported by IoT data is verification precision. Optimizing the efficiency of algorithmic optimization means making them more responsive to new data patterns and shifting system requirements. With these key performance indicators in mind, the OOIVM sets out to build a solid foundation for handling the intricacies of immediate-time IoT application management, aiming to enhance reliability, scalability, and performance.
The OOIVM expands upon the principles established by prior studies in IoT technology’s cooperation, grouping, and resource management. It enhances clustering efficiency, resource consumption, flexibility, computational complexity, implementation speeds, and reliability of validation in a targeted manner. These improvements cumulatively lead to a more robust, more effective, and easily expandable architecture for real-time IoT applications.

3.1. K-Means Clustering Model

It refers to the grouping of similar kinds of complexities in the form of clusters. The real-time applications are responsible for integrating organized optimization from the IoT paradigms. The input data include service levels, complexities, and time lag. In the clustering process, initial data received A from the IoT paradigm is evaluated as follows.
A = S V L i ( C o m p l m a x C o m p l m i n ) T L a g + ω a a + α
and
α = 1 2 π C o m p l m i n C o m p l m a x a a 2 a a
where S V L i is the number of service levels, and C o m p l m i n and C o m p l m a x are the minimum and maximum complexities detected in different instances. The variable α denotes the segregation of organized integration cases from the centroid for each cluster. The centroids estimate the number of non-optimized classifications (validation lags) observed over the different N service levels. The clustering initialization for grouping is illustrated in Figure 3.
Figure 3. Clustering initialization for grouping.
In Figure 3 above, the A -based representation for grouping initialization is presented. By defining a and a , the condition defines the centroid changes. Both minimization, m i n a ω α , and maximization, m a x a ω a , represent the number of α required. As α increases, ω = 0 is achieved by classifying a and a . Thus, the grouping consolidates levels and ties to enable integration feasibility. The deviation is observed for a to a switchover or ω = 1 detection. Therefore, the grouping requires multiple considerations to improve integrations. In this organized case, complexity occurs during integration due to non-optimized classification problems. Therefore, the validation lags impact IoT integration at any level of the service for which the organized integration cases are segregated, S I N T G , using the centroid deviation knowledge. This segregation is represented as follows.
S I N T G = d e v t m a x C o m p l m i n C o m p l m a x d e v t σ 2 + a a α
and
d e v t σ = 1 N 1 n 1 i = 1 N C d v P d v 2
where the variables C d v and P d v signify the current centroid deviation and knowledge of previous deviations. Based on the centroid deviation, the segregation is pursued to follow the maximum centroid deviation d e v t m a x and standard deviation d e v t σ for accurately balancing the integration levels. Here, the number of sharing levels and the center point is used to identify the number of clusters K for which appropriate non-complex processing, integration level balancing, and time-lagging integrations from different levels are required for minimizing integration complexity.
Based on the current and previous deviations, the sequential row of clustering α C d v , P d v is estimated as follows:
α C d v , P d v = d e v t σ d e v t m a x 1 2 + d e v t σ d e v t m a x 2 2 + + 1 C d v P d v n 2 , n N
In Equation (10), the clustering for a sequence is performed until the changes occur in the centroid. The centroid deviation knowledge confines complexity lags in the service levels. The above clustering sequence is pursued using the k-means clustering process and machine learning to precisely identify complexities. In this analysis and grouping process, k-means clustering is exploited for detecting complexity lags across various IoT application integrations. The harmonized service levels are grouped in appropriate and accurate time instances to reduce validation and complexity lags. In addition, the cluster is dissolved and reformed to be instantaneous to meet complex demands for integration. Therefore, k-means clustering and machine learning are used to prevent validation lags. The sequential clustering for complexity and time lag is analyzed in Figure 4.
Figure 4. Sequential clustering analysis for complexity and time lag.
The complexity and T L a g -based K requirements analysis is presented in Figure 4. Considering the d e v t σ and d e v t m a x , the C d v , P d v for the sequential process is defined. Further, ω = 0 and ω 0 classifications are performed under d e v t σ d e v t m a x computation to ensure maximum α is identified. The integration complexity is thus estimated for four levels N = 4 , i.e., field level, resource connectivity, response, and storage. Based on the increasing level of A , the above K is analyzed. In particular, the similar grouping rate is analyzed for four levels, as shown in Figure 5.
Figure 5. S(INTG) analysis.
The analysis in Figure 5 presents the integration segregation for complexity + lag, lag + levels, and all three (joint). As the K increases, α C d v , P d v increases to identify the maximum complexity causing integration. Therefore, the identified grouping is segregated towards multiple instances of organized integration. The clusters are disintegrated and reformed under maximum N and α detected under reduced d e v t σ . It also ensures d e v t m a x is not reached until a new integration request or complexity arises due to N = N + 1 . The machine learning output aims to identify and segregate the organized integration cases with standard deviation and previous deviation knowledge updates based on training. In this learning process, the first step is to address the issues above using the clustering process. The number of clusters, K , in service levels is achieved using the condition 1 C d v P d v for updating the previous deviation knowledge update. O2IVM is implemented to improve the integration level of complexity analysis based on centroid deviation. The centroid deviation is detected from different service levels, i.e., organized integration is observed in various time intervals for further clustering. This proposed model aims to reduce the validation lags in analyzing complexity. The challenge is to update the centroid deviation and grouping sequence with previous deviation knowledge. The centroid deviations are stored as previously identified centroid deviation knowledge.
This proposed approach effectively meets the need for high availability by utilizing machine learning and k-means clustering to minimize time lags and response time. The centroids are identified from the clustering output to rectify suboptimal classifications and validate delays across various N service levels. Machine learning is utilized to discern the intricacies of fluctuating execution time and resource allocation. The structured integration scenarios are classified based on the centroid deviation to determine and distinguish the maximum centroid deviation (devt_max) and standard deviation (devt_σ) for balance integration levels. The current gadget reduces complexity by employing machine training and clustering techniques, reducing time lag and complexity and improving integration evaluation. This article employs machine learning to mitigate the intricacies in random situations characterized by high availability parameters, in contrast to other aspects.

3.2. Learning Process for Integration Validation

In the service level grouping process, machine learning is used to find out the complexity and centroid deviation estimation. As this machine learning process relies on previous knowledge of centroid deviations, a more accurate complexity lag is achievable. The number of clusters may vary based on service levels, though the previously stored centroid deviation knowledge helps to classify non-optimized integration. Machine learning helps to pursue two types of segments: sequence grouping and centroid deviation identification. Service levels, complexity, and time lags are identified in the sequence grouping process to update the previously stored centroid deviation knowledge. In the centroid deviation identification, the clustering process balances integration levels to augment the same ratio of non-optimized classification , thereby reducing centroid deviation along with better grouping. In the grouping process G , the levels, complexity, and time lags are classified independently using k-means clustering. The integration improvement is achieved based on dissolved and reformed clusters, after which volatile integration is used to update the previous centroid deviation knowledge. The machine learning output sequence from G 1 to G T is computed as follows:
G 1 = C d v 1 G 2 = 2 C d v 2 2 S V L 1 1 G 3 = 3 C d v 3 3 S V L 2 3 G N = n C d v N n S V L i N 1
U 1 = C 1 U 2 = C 2 + 1 C d v P d v 1 U 3 = C 3 + 1 C d v P d v 2 ω a a 1 U N = C K + 1 C d v P d v n 1 ω a a n 2
The grouping process generates two outputs, namely, levels, complexity, and time lag from G 1 to G N and update of the sequence from U 1 to U N . Based on the number of clusters, the grouping is pursued using k-means clustering based on the complexity of the lag occurring in the sequence. Each service level is analyzed to identify and confirm the validation lags. The number of clusters dissolved must not be equal to the number of clusters reformed, which is the optimal grouping condition, to prevent additional complexity demands. If the occurrence of validation lags is identified in the first sequence, then previous centroid deviation knowledge is updated using C K . This means the time lag is classified according to the occurrence of non-clustered integrations, and C K + 1 C d v P d v n 1 ω a a n 2 is the previous knowledge updating condition. In the grouping process, from the initial organized integration cases, U N , C K is segregated with the appropriate centroid deviation knowledge. The grouping and updating output matching are checked and independently grouped in this clustering process. Previous deviation knowledge occurs in the first level from which that level is grouped for integration improvement. After the grouping process, the harmonized service levels of a single entity are compared with multiple entities based on complexity grouping. Here, each cluster’s serving inputs, sharing levels, and midpoint are used to update the previous knowledge.
In the above cluster balancing, the condition of A > α is said to be 1 , and the condition of A < α is said to be 0 based on the centroid deviation. If 0 is observed at any level, then the complexity lags are identified; hence, non-optimized classification is performed for further integration-level improvement. Now, the volatile integrations must be verified by machine learning, and the previous centroid deviation knowledge must be updated. In the centroid deviation knowledge data retrieval, the cluster and update are fetched for complexity analysis to identify if any validation lags are taking place. Here, the k-means clustering, C K , and update is different from the grouping process, represented using (13) and (14):
C 1 = S V L 1 C 2 = 2 S V L 2 1 + G 1 C 3 = 3 S V L 3 2 + G 2 C K = n S V L N N + G N
U 1 = 1 t d 1 + t s 1 t d 1 U 2 = ( C o m p l m a x C o m p l m i n ) T L a g 1 C 1 U 3 = ( C o m p l m a x C o m p l m i n ) T L a g 3 C 2 U N = ( C o m p l m a x C o m p l m i n ) T L a g N 1 C K
The centroid deviation knowledge update is performed at the end of all service levels or before the start of a new service. The update is pursued along with consecutive centroid changes that are updated using machine learning for integration improvement. It is to be noted that the volatile integrations are utilized and grouped under centroid deviation. Using machine leaning and k-means, clustering complexity and time lags are reduced, and integration is improved.
The analysis of device availability under c o m p l m a x and c o m p l m i n is presented in Figure 6. This analysis shows the elasticity of the proposed method for different U N instances.
Figure 6. Availability and time lag analysis.
The K rate improvement increases the S ( I N T G ) for the observed new integration. The and of the centroid converges the combination of complexity, lag, and levels to identify α . If α demands are high, then c o m p l m a x and c o m p l m i n are segregated at ease using availability. If the availability is high, U N , for grouping, it is estimated that the time lag is confined. Thus, the feasibility of integration is increased under fewer complexities (Figure 6).
Conventional IoT integration and optimization research has helped us understand this field’s issues and solutions. Preliminary data handling and analysis research has revealed the importance of actual time data processing and management in IoT. Researchers have examined dispersed data processing, fog computing, and edge computing to manage the fast influx of IoT data. K-means and its variations have excelled at clustering and analyzing IoT data. These studies illustrate that clustering simplifies data and improves insights. Research has shown that IoT systems need efficient resource management to sustain and efficiently operate IoT, optimize assets, network bandwidth, and energy use. Real-time system research has focused on latencies and data input responsiveness. The unique timing needs of IoT applications have led to priority-based computing and real-time management.
Several practical considerations must be addressed to feasibly and effectively apply OOIVM in real-world IoT applications. To improve the quality of its resource projections, simulations, and depictions based on interactive devices/interfaces, real-world applications are characterized by a need for complexity evaluation and grouping. The proposed O2IVM for precise incorporation evaluation ensures non-optimized categorization and prioritization by utilizing the centroid deviation knowledge of the ordered integration levels. Important factors include harmonized service levels that are needed to decrease complexity demands and limit centroid deviation-based complexity lags for a single entity. OOIVM can streamline and scale real-time IoT applications with proper planning and execution, making it a viable solution for modern IoT.
This proposed model simultaneously categorizes structured and non-organized integrations, utilizing time lag and levels to enhance the response rate. The unmanaged clusters with the highest complexity level are dissolved to decrease the computational complexity. This model successfully decreases complexity by 9.04% and enhances device availability by 12.46% to meet maximum device limits. While this model effectively solves the complexity issue, it is limited by the energy-based difficulties of the devices. Hence, energy harvesting optimization and multi-constraint clustering need to be developed to reduce integration complexity.
Segmentation in IoT typically does not use neural networks or other sophisticated ML models due to the high computing cost, extended training durations, and massive volumes of labeled information that these models require. Regarding the massive, ever-changing characteristics of IoT settings, where flexibility and swift adaptation are of utmost importance, k-means clustering is the way to go because it is more effective, quicker, and simpler to deploy for real-time applications.

4. Results and Discussion

Discussions on execution time, complexity, time lag, response time, and availability are described in this section. The above parameters are obtained from OPNET simulations containing 2600 devices serving 11 applications. The applications are heterogeneous and served from 24 storage units assembled in IoT. A device’s maximum request processing capacity is 10–15/h. Integration of multiple devices with IoT paradigms occurs up to a maximum ratio of 60. The simulation is validated for service responses from 300 s to 1800 s devices. Therefore, the devices and the integration demands are the varying parameters set in the discussion for the evaluation metrics. The proposed model is accompanied by EEDTO [27], RERM [29], and ChameLIoT [35] methods from the Related Works section.
IoT relies on k-means clustering because it is scalable, efficient, and simple—three qualities crucial for dealing with the massive, ever-changing datasets in IoT settings. It is perfect for use in real-time situations, primarily due to its rapid data adaptation and simplicity-enhancing centroid-based optimization. The computationally costly or less scalable characteristics of other models, such as hierarchical clustering, DBSCAN, and Gaussian mixture models, make them unsuitable for IoT systems’ rapid and expanding characteristics.

4.1. Execution Time

Figure 7 displays the execution time of the proposed model for application integration and service level from the IoT paradigm using k-means clustering to meet integration demand. For accurate issue identification, this approach avoids service level-based adjustments. The non-optimized categories for minimizing complexity address resource management and execution times in protracted device integrations. IoT application integration uses device/interface complexity to improve non-optimized categorization. Thus, the initial O2IVM validation output defines the difficulty of analyzing and grouping for evaluating resource projection quality, modeling, and representation. The conditions ω = (ω_1, …, ω_n, …, ω_N) and ∃ = (∃_1, …, ∃_n, …, ∃_N) are investigated to reduce execution time at various service levels for resource management and execution. To discover IoT application integration challenges, single entities are grouped. Therefore, the ideal output for 1400 devices involves IoT application device integrations that segregate organized integration scenarios in 1.74 s.
Figure 7. Execution time.

4.2. Integration Complexity

The coordinated optimization integration validation model addresses complexity. K-means clustering reduces complexity, as shown in Figure 8. The resource forecast quality was increased with suitable service level grouping based on execution time and complexity. For exact integration evaluation, we identified service levels, delays, and complications between protracted centroids device integrations. A simpler model took 1.25 s to execute than the other. Clustering to find issues balances integration levels and improves complexity minimization and prioritizing. Minimization helps evaluate complexity gaps between service tiers to prevent time lags. The clustering technique matches centroid changes with service level adjustments for each device integrated to find midpoints and avoid complexity. Better integration improves resource management by minimizing complexity and prioritizing. Therefore, the probability is simpler than the other factors.
Figure 8. Integration complexity.

4.3. Time Lag

Figure 9 demonstrates the clustering method output service levels, complexity, and latency to fix all device integration issues at different times. Centroid deviation knowledge balances time-lagging integrations, integration levels, and non-complex processing in IoT applications. Matching current and previous centroid deviation knowledge with clustering output prevents time lag and improves resource projection quality. Grouping harmonized service levels with correct centroid deviation knowledge reduces complexity associated with non-optimized classification. Cluster dissolving improves complexity minimization and prioritizing, and reformation reduces time lag. The proposed model addresses midpoints using RL machine learning and organized optimization model output. The centroid reduces complexity and time lags with k-means clustering. The real-time application-based model is considered for combining single entities to reduce prioritization time.
Figure 9. Time lag.

4.4. Response Time

For device integrations with high availability characteristics, the proposed O2IVM is used to achieve less response time. As shown in Figure 10, this is accomplished in this model by simplifying the response time and, as a result, decreasing the time lag. Based on the different service levels, k-means clustering is used to find the complexity using the centroid. The first step in aligning service levels is to group entities into initial single entities, where n = (1, N). This grouping finds and reduces the complexity to improve resource management using centroid deviation knowledge, which is necessary to meet the criteria mentioned earlier. Therefore, complexity lags are suppressed by the integration evaluation. A reduction in complexity based on centroids is achieved by obtaining the integration requirement. The incorporation of organized optimization for correctly grouping the service levels according to Equation (2) achieves a high level of complexity minimization. Less response time can be achieved through integration evaluation using the levels, lags, and complexities that result from minimization of complexity.
Figure 10. Response time.

4.5. Availability Parameter

Based on minimum and maximum complexity of real-time application at distinct instances, centroid deviation knowledge is analyzed. As seen in Figure 11, this assessment has great availability. With minimal time lags and reaction time, this model meets the high availability parameter using machine learning and k-means clustering. Centroids are recognized from the clustering output to overcome non-optimized classifications and validation lags across N service levels. Machine learning is used to identify execution time and resource management complexities. To determine the maximum centroid deviation, d e v t m a x , and standard deviation, d e v t σ , for balanced integration levels, the organized integration cases are separated based on centroid deviation. The current model reduces complexity and time latency with machine learning and clustering to improve integration evaluation. This method simplifies random instances with high availability characteristics using machine learning.
Figure 11. Availability parameters.
Many hurdles prevent widespread deployment of IoT applications and services. Due to IoT technology standard gaps, architecture is the biggest challenge. IoT architecture design prioritizes safety, capacity, confidentiality, variability, data management, and networking. To quickly make IoT system decisions and maneuvers, the OOIVM addresses criteria such as streamlining data input reception and response time. Better use of computing and network infrastructure reduces waste and improves operations, lowering costs. Scalability lets the network handle more data and IoT devices without slowing down. In terms of computing complexity, the system simplifies IoT data integration and processing and makes systems easier to manage. Verification precision improves information validation to ensure IoT data-supported choices and actions are accurate. Making algorithmic optimization more adaptable to new data patterns and system needs improves efficiency. Reliability, scalability, and performance are OOIVM’s goals, as it builds a solid platform for immediate-time IoT application administration.
Substantial energy consumption from the ongoing operation of IoT devices, computational units, and data transportation adds to the functioning, preservation, cloud service, and hardware expenses associated with OOIVM model implementation. Optimization of resource allocation, use of energy-efficient hardware, utilization of edge computing to decrease reliance on the cloud, and adoption of adaptive energy control methods are strategies to mitigate these impacts. Possible obstacles to adoption include safeguarding personal information, establishing interoperability, keeping data consistent and high-quality, controlling latency and bandwidth, and making sure the system can scale. Robust security mechanisms, consistent protocols, effective data management, and scalable system architecture are necessary to overcome these obstacles.

5. Conclusions

The integration of IoT to meet maximum real-time application demands has become of primary interest. This article proposes and discusses an organized optimization integration model for achieving this feature. This model groups the complexity-causing factors, such as response time lag, and the highest integration levels for identifying unisons. A modified k-means clustering method is used for this purpose, performing either disintegration or unison-based grouping with the update. The complexity factors converge towards a single entity through grouping knowledge, thereby reforming the clustering process. The proposed model thus classifies organized and non-organized integrations jointly or independently using time lag and levels to improve the response rate. The unattended clusters with the maximum number of instances are dissolved to reduce the computation complexity. Therefore, this model reduces the complexity by 9.04% and improves device availability by 12.46% for maximum device limits. This model successfully addresses the complexity problem, but its scope is limited by the energy-based complexities of the devices. Therefore, energy harvesting optimization and multi-constraint clustering will need to be improved to sustain less integration complexity.

Author Contributions

Conceptualization, A.A. and M.K.A.; methodology, T.S.A.; validation, K.A.A.; formal analysis, A.A.; investigation, T.S.A.; resources; writing—original draft preparation, K.A.A.; writing—review and editing, K.A.A.; visualization, M.K.A.; funding acquisition, A.A. and M.K.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data will be made available by the authors on request.

Acknowledgments

The authors extend their appreciation to the Deanship of Research and Graduate Studies, at the University of Tabuk for financial and technical support.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Algarni, F.; Khan, M.A.; Alawad, W.; Halima, N.B. P3S: Pertinent Privacy-Preserving Scheme for Remotely Sensed Environmental Data in Smart Cities. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 5905–5918. [Google Scholar] [CrossRef]
  2. Liu, L.; Essam, D.; Lynar, T. Complexity measures for IoT network traffic. IEEE Internet Things J. 2022, 9, 25715–25735. [Google Scholar] [CrossRef]
  3. Alqarni, M.A.; Alharthi, A.; Alqarni, A.; Khan, M.A. A transfer-learning-based energy-conservation model for adaptive guided routes in autonomous vehicles. Alex. Eng. J. 2023, 76, 491–503. [Google Scholar] [CrossRef]
  4. Shibu, N.S.; Devidas, A.R.; Balamurugan, S.; Ponnekanti, S.; Ramesh, M.V. Optimising Microgrid Resilience: Integrating IoT, Blockchain, and Smart Contracts for Power Outage Management. IEEE Access 2024, 12, 18782–18803. [Google Scholar] [CrossRef]
  5. Mukherjee, A.; Goswami, P.; Khan, M.A.; Manman, L.; Yang, L.; Pillai, P. Energy-Efficient Resource Allocation Strategy in Massive IoT for Industrial 6G Applications. IEEE Internet Things J. 2021, 8, 5194–5201. [Google Scholar] [CrossRef]
  6. Liu, H.; Guan, X.; Bai, R.; Qin, T.; Chen, Y.; Liu, T. Designing a medical information diagnosis platform with IoT integration. Heliyon 2024, 10, e25390. [Google Scholar] [CrossRef] [PubMed]
  7. Hu, X.; Assaad, R.H. BIM-enabled digital twin framework for real-time indoor environment monitoring and visualization based on autonomous LIDAR-based robotic mobile mapping, IoT sensing, and indoor positioning technology. J. Build. Eng. 2024, 86, 108901. [Google Scholar] [CrossRef]
  8. Varriale, V.; Cammarano, A.; Michelino, F.; Caputo, M. Integrating blockchain, RFID and IoT within a cheese supply chain: A cost analysis. J. Ind. Inf. Integr. 2023, 34, 100486. [Google Scholar] [CrossRef]
  9. Najim, A.H.; Kurnaz, S. Study of integration of wireless sensor network and Internet of Things (IoT). Wirel. Pers. Commun. 2023, 1–14. [Google Scholar] [CrossRef]
  10. Zhou, Z.; Shojafar, M.; Alazab, M.; Abawajy, J.; Li, F. AFED-EF: An energy-efficient VM allocation algorithm for IoT applications in a cloud data center. IEEE Trans. Green Commun. Netw. 2021, 5, 658–669. [Google Scholar] [CrossRef]
  11. Abuhasel, K.A.; Khan, M.A. A Secure Industrial Internet of Things (IIoT) Framework for Resource Management in Smart Manufacturing. IEEE Access 2020, 8, 117354–117364. [Google Scholar] [CrossRef]
  12. Shakeel, P.M.; Baskar, S.; Dhulipala, V.S.; Jaber, M.M. Cloud based framework for diagnosis of diabetes mellitus using K-means clustering. Health Inf. Sci. Syst. 2018, 6, 16. [Google Scholar] [CrossRef] [PubMed]
  13. Presciuttini, A.; Cantini, A.; Costa, F.; Portioli-Staudacher, A. Machine learning applications on IoT data in manufacturing operations and their interpretability implications: A systematic literature review. J. Manuf. Syst. 2024, 74, 477–486. [Google Scholar] [CrossRef]
  14. Alghamdi, N.S.; Khan, M.A. Energy-Efficient and Blockchain-Enabled Model for Internet of Things (IoT) in Smart Cities. Computers, Mater. Contin. 2021, 66, 2509–2524. [Google Scholar] [CrossRef]
  15. Rizvi, S.; Zwerling, T.; Thompson, B.; Faiola, S.; Campbell, S.; Fisanick, S.; Hutnick, C. A modular framework for auditing IoT devices and networks. Comput. Secur. 2023, 132, 103327. [Google Scholar] [CrossRef]
  16. Jazaeri, S.S.; Asghari, P.; Jabbehdari, S.; Javadi, H.H.S. Composition of caching and classification in edge computing based on quality optimization for SDN-based IoT healthcare solutions. J. Supercomput. 2023, 79, 17619–17669. [Google Scholar] [CrossRef] [PubMed]
  17. Abohamama, A.S.; El-Ghamry, A.; Hamouda, E. Real-time task scheduling algorithm for IoT-based applications in the cloud–fog environment. J. Netw. Syst. Manag. 2022, 30, 54. [Google Scholar] [CrossRef]
  18. Sheron, P.F.; Sridhar, K.P.; Baskar, S.; Shakeel, P.M. A decentralized scalable security framework for end-to-end authentication of future IoT communication. Trans. Emerg. Telecommun. Technol. 2020, 31, e3815. [Google Scholar] [CrossRef]
  19. Wang, H.; Li, W.; Sun, J.; Zhao, L.; Wang, X.; Lv, H.; Feng, G. Low-complexity and efficient dependent subtask offloading strategy in IoT integrated with multi-access edge computing. IEEE Trans. Netw. Serv. Manag. 2023, 21, 621–636. [Google Scholar] [CrossRef]
  20. Si-Mohammed, S.; Bardou, A.; Begin, T.; Lassous, I.G.; Vicat-Blanc, P. NS+ NDT: Smart Integration of Network Simulation in Network Digital Twin, Application to IoT Networks. Future Gener. Comput. Syst. 2024, 157, 124–144. [Google Scholar] [CrossRef]
  21. Brandín, R.; Abrishami, S. IoT-BIM and blockchain integration for enhanced data traceability in offsite manufacturing. Autom. Constr. 2024, 159, 105266. [Google Scholar] [CrossRef]
  22. Jin, S.; Karki, B. Integrating IoT and Blockchain for Intelligent Inventory Management in Supply Chains: A Multi-Objective Optimization Approach for the Insurance Industry. J. Eng. Res. 2024. [Google Scholar] [CrossRef]
  23. Alulema, D.; Criado, J.; Iribarne, L.; Fernández-García, A.J.; Ayala, R. SI4IoT: A methodology based on models and services for the integration of IoT systems. Future Gener. Comput. Syst. 2023, 143, 132–151. [Google Scholar] [CrossRef]
  24. Irwanto, F.; Hasan, U.; Lays, E.S.; De La Croix, N.J.; Mukanyiligira, D.; Sibomana, L.; Ahmad, T. IoT and Fuzzy Logic Integration for Improved Substrate Environment Management in Mushroom Cultivation. Smart Agric. Technol. 2024, 7, 100427. [Google Scholar] [CrossRef]
  25. Chakour, I.; Mhammedi, S.; Daoui, C.; Baslam, M. Unlocking QoS Potential: Integrating IoT services and Monte Carlo Control for heterogeneous IoT device management in gateways. Comput. Netw. 2024, 238, 110134. [Google Scholar] [CrossRef]
  26. Narang, D.; Madaan, J.; Chan, F.T.; Chungcharoen, E. Managing open loop water resource value chain through IoT focused decision and information integration (DII) modelling using fuzzy MCDM approach. J. Environ. Manag. 2024, 350, 119609. [Google Scholar] [CrossRef]
  27. Wu, H.; Wolter, K.; Jiao, P.; Deng, Y.; Zhao, Y.; Xu, M. EEDTO: An energy-efficient dynamic task offloading algorithm for blockchain-enabled IoT-edge-cloud orchestrated computing. IEEE Internet Things J. 2020, 8, 2163–2176. [Google Scholar] [CrossRef]
  28. Yang, W.; Qin, Y.; Yi, Z.; Wang, X.; Liu, Y. Providing cache consistency guarantee for ICN-based IoT based on push mechanism. IEEE Commun. Lett. 2021, 25, 3858–3862. [Google Scholar] [CrossRef]
  29. Eroshkin, I.; Vojtech, L.; Neruda, M. Resource Efficient Real-Time Reliability Model for Multi-Agent IoT Systems. IEEE Access 2021, 10, 2578–2590. [Google Scholar] [CrossRef]
  30. Chahed, H.; Usman, M.; Chatterjee, A.; Bayram, F.; Chaudhary, R.; Brunstrom, A.; Kassler, A. AIDA—A holistic AI-driven networking and processing framework for industrial IoT applications. Internet Things 2023, 22, 100805. [Google Scholar] [CrossRef]
  31. Zhang, K.; Zhou, Y.; Wang, C.; Hong, H.; Chen, J.; Gao, Q.; Ghobaei-Arani, M. Towards an automatic deployment model of IoT services in Fog computing using an adaptive differential evolution algorithm. Internet Things 2023, 24, 100918. [Google Scholar] [CrossRef]
  32. Wang, H.; Li, L.; Cui, Y.; Wang, N.; Shen, F.; Wei, T. MBSNN: A multi-branch scalable neural network for resource-constrained IoT devices. J. Syst. Archit. 2023, 142, 102931. [Google Scholar] [CrossRef]
  33. Guan, Z.; Wang, Z.; Cai, Y.; Wang, X. Deep reinforcement learning based efficient access scheduling algorithm with an adaptive number of devices for federated learning IoT systems. Internet Things 2023, 24, 100980. [Google Scholar] [CrossRef]
  34. Kim, E.; Son, T.; Ha, S. A novel hierarchical edge-based architecture for service oriented IoT. Internet Things 2023, 24, 100939. [Google Scholar] [CrossRef]
  35. Silva, M.; Gomes, T.; Ekpanyapong, M.; Tavares, A.; Pinto, S. ChamelIoT: A tightly-and loosely-coupled hardware-assisted OS framework for low-end IoT devices. Real-Time Syst. 2023, 60, 1–47. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.