Next Article in Journal
Skin Cancer Detection and Classification Through Medical Image Analysis Using EfficientNet
Previous Article in Journal
Real-Time Industrial Water Pollution Evaluation Using Edge–Cloud IoT Architecture and Multi-Parameter Sensing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

AI-Driven Energy-Efficient Data Aggregation and Routing Protocol Modeling to Maximize Network Lifetime in Wireless Sensor Networks

1
Electronics and Communication Engineering, KGiSL Institute of Technology, Coimbatore 641035, India
2
Information Technology, KGiSL Institute of Technology, Coimbatore 641035, India
3
Mechatronics Engineering, Sri Krishna College of Engineering and Technology, Coimbatore 641008, India
*
Author to whom correspondence should be addressed.
Submission received: 5 July 2025 / Revised: 18 September 2025 / Accepted: 21 September 2025 / Published: 25 September 2025

Abstract

The research work presents an artificial intelligence-driven, energy-aware data aggregation and routing protocol for wireless sensor networks (WSNs) with the primary objective of extending overall network lifetime. The proposed scheme leverages reinforcement learning in conjunction with deep Q-networks (DQNs) to adaptively optimize both Cluster Head (CH) selection and routing decisions. An adaptive clustering mechanism is introduced wherein factors such as residual node energy, spatial proximity, and traffic load are jointly considered to elect suitable CHs. This approach mitigates premature energy depletion at individual nodes and promotes balanced energy consumption across the network, thereby enhancing node sustainability. For data forwarding, the routing component employs a DQN-based strategy to dynamically identify energy-efficient transmission paths, ensuring reduced communication overhead and reliable sink connectivity. Performance evaluation, conducted through extensive simulations, utilizes key metrics including network lifetime, total energy consumption, packet delivery ratio (PDR), latency, and load distribution. Comparative analysis with baseline protocols such as LEACH, PEGASIS, and HEED demonstrates that the proposed protocol achieves superior energy efficiency, higher packet delivery reliability, and lower packet losses, while adapting effectively to varying network dynamics. The experimental outcomes highlight the scalability and robustness of the protocol, underscoring its suitability for diverse WSN applications including environmental monitoring, surveillance, and Internet of Things (IoT)-oriented deployments.

1. Introduction

Wireless Sensor Networks (WSNs) have emerged as a key enabling technology in modern digital infrastructures, supporting diverse applications such as precision agriculture, environmental monitoring, disaster management, military surveillance, healthcare systems, and smart city development [1,2,3]. A typical WSN consists of a large number of spatially distributed sensor nodes equipped with limited sensing, computation, and wireless communication capabilities. These nodes are generally powered by non-rechargeable batteries and are deployed to measure ambient parameters including temperature, humidity, vibration, and pressure, with the sensed data transmitted to a base station (sink) for further processing [4,5].
Despite their versatility, WSNs face critical operational challenges, with energy consumption being the most significant. Since sensor nodes are usually deployed in inaccessible or hostile environments, battery replacement is impractical. Node energy depletion leads to sensing coverage gaps, routing failures, and eventually, the collapse of network connectivity [6]. Consequently, the design of energy-efficient clustering and routing protocols remains central to prolonging network lifetime while ensuring reliable data transmission.
Several clustering-based protocols have been introduced to address these concerns. Early approaches such as Low-Energy Adaptive Clustering Hierarchy (LEACH) [7], Power-Efficient Gathering in Sensor Information Systems (PEGASIS) [8], and Hybrid Energy-Efficient Distributed Clustering (HEED) [9] significantly reduced redundant transmissions and minimized direct communication with the sink. However, their decision-making strategies relied on static or probabilistic criteria, offering limited adaptability under dynamic conditions such as node failures, heterogeneous energy states, and varying traffic demands. For instance, LEACH randomly selects cluster heads, often causing uneven energy distribution. PEGASIS organizes nodes into linear chains, but its rigid structure cannot efficiently adapt to dynamic topologies. HEED partially addresses these issues by considering residual energy in cluster head selection, yet its reliance on fixed parameters restricts scalability and flexibility.
Recent advancements in Artificial Intelligence (AI), particularly Reinforcement Learning (RL), provide a promising alternative for addressing the limitations of traditional protocols [10,11]. RL enables distributed decision-making by allowing nodes to learn from their interactions with the network environment and adapt strategies that optimize long-term energy efficiency and reliability. Deep Reinforcement Learning (DRL), and specifically Deep Q-Networks (DQNs), further enhance this capability by approximating Q-values through deep neural networks, thereby enabling efficient learning in large-scale state-action spaces inherent to WSNs [12].
Motivated by these insights, this paper proposes an AI-driven, energy-efficient data aggregation and routing protocol for WSNs. The contributions of this work can be summarized as follows:
  • Adaptive Clustering Strategy: Residual energy, node proximity, and local traffic load are jointly considered in cluster-head selection, ensuring balanced energy consumption and preventing premature node failures.
  • Intelligent Routing Mechanism: A DQN-based routing scheme is employed to dynamically determine energy-efficient forwarding paths, considering both reliability and latency requirements.
  • Performance Validation: The proposed protocol is evaluated against conventional approaches such as LEACH, PEGASIS, and HEED using metrics including network lifetime, energy consumption, packet delivery ratio (PDR), latency, and load distribution. Simulation results demonstrate superior performance, scalability, and adaptability under dynamic network conditions.
The research introduces an AI-based energy-aware protocol for WSNs that combines adaptive clustering with a DQN-driven routing strategy. Experimental results show improved lifetime, efficiency, reliability, and scalability over traditional methods.
The integration of Artificial Intelligence (AI) into WSN protocols introduces significant distinctions compared to conventional approaches. Unlike traditional schemes, AI-enabled protocols exhibit adaptive behavior, allowing the network to respond dynamically to variations in environmental conditions. This adaptability is particularly critical in WSNs, where node failures, fluctuating traffic patterns, and progressive energy depletion are common. By leveraging learning mechanisms, AI-driven protocols are capable of making time-optimal decisions through experience-based adaptation, thereby continuously refining energy management strategies. This enables efficient distribution of energy-intensive tasks, such as data transmission and aggregation, with respect to the overall lifetime of the network. Furthermore, deep reinforcement learning methods, such as Deep Q-Networks (DQNs), inherently support scalability, making them suitable for large-scale deployments involving hundreds or thousands of nodes.
The effectiveness of the proposed AI-assisted data aggregation and routing protocol, as shown in Figure 1, is evaluated using extensive simulations that replicate real-world WSN scenarios. These scenarios consider large-scale random node deployment for monitoring environmental conditions. Performance assessment is carried out through key metrics, including network lifetime, energy consumption, packet delivery ratio (PDR), latency, and load distribution. Comparative analysis with established protocols such as LEACH, PEGASIS, and HEED demonstrates substantial improvements in terms of energy efficiency, packet delivery reliability, and overall network sustainability. The results confirm that the AI-driven protocol ensures balanced energy utilization across nodes, reduces packet losses through intelligent routing, and significantly extends operational lifetime.
A key outcome of this study is the reinforcement learning-based cluster head (CH) selection strategy, which dynamically rotates the CH role among nodes based on residual energy levels and spatial distribution. This real-time adaptability prevents overburdening of individual nodes with data aggregation responsibilities, thereby mitigating premature failures and avoiding network partitioning. In addition, the DQN-based routing mechanism identifies energy-efficient forwarding paths that minimize communication overhead and enhance throughput. Collectively, these mechanisms yield a robust, scalable, and energy-aware solution for WSN communication challenges.
The proposed AI-driven energy-efficient data aggregation and routing protocol provides a comprehensive solution to the energy and reliability limitations of traditional WSN protocols. The integration of adaptive clustering with DQN-based routing enhances energy conservation, improves load balancing, and extends network longevity, while maintaining high-quality data delivery. These contributions establish a solid foundation for future advancements in AI-assisted WSN protocols, with potential applications in environmental monitoring, smart city infrastructure, healthcare systems, and other IoT-driven domains. As WSNs continue to play a critical role in next-generation IoT ecosystems, the demand for intelligent, energy-efficient communication strategies will become increasingly vital.

2. Related Works

Recent advances in Internet of Things (IoT) and Wireless Sensor Networks (WSNs) research emphasize optimization techniques that enhance energy efficiency, security, and adaptability through the integration of Artificial Intelligence (AI). Approaches under investigation include fuzzy clustering models, hybrid machine-learning-based intrusion detection frameworks, blockchain-assisted communication protocols, and AI-driven routing mechanisms. Collectively, these studies converge on several key requirements: minimization of energy expenditure, robust data protection, scalability for large-scale deployments, and adaptability to dynamic operating conditions. By integrating state-of-the-art technologies such as blockchain, machine learning, and deep reinforcement learning, researchers have developed solutions aimed at improving the lifetime, reliability, and resilience of IoT-enabled WSNs, particularly in domains such as healthcare, industrial automation, and smart city infrastructures.
Javadpour et al. [1] introduced an optimization-based fuzzy clustering method to enhance energy management in IoT networks. Their work focused on refining cluster head selection using fuzzy logic combined with hybrid optimization algorithms, leading to extended network lifetime and more balanced power usage across sensor nodes. They further suggested that predictive AI-based models could be incorporated to adapt clustering strategies in real time. Gebremariam et al. [2] proposed a hybrid machine learning-based intrusion detection system (IDS) tailored for hierarchical WSNs. By combining decision trees and support vector machines within a multilayer IDS, they achieved improved accuracy while minimizing false positives. They also noted that deep learning extensions could further strengthen adaptability against evolving cyber threats.
Energy-efficient clustering solutions were also studied by Labib et al. [3], who developed an enhanced threshold-sensitive distributed protocol designed for IoT-based WSNs. By dynamically adjusting cluster head thresholds according to network conditions, their approach achieved balanced energy consumption and extended network longevity. Similarly, Priyadarshi et al. [4] surveyed a wide spectrum of energy-aware routing algorithms, including genetic algorithms, particle swarm optimization, and reinforcement learning. Their findings highlighted the potential of AI-based models for achieving real-time optimization under dynamic network conditions and recommended hybrid frameworks that combine meta-heuristics with AI for greater efficiency.
In the domain of secure routing, Haseeb et al. [5] proposed an AI-assisted sustainable model for Mobile WSNs (MWSNs). Their protocol predicts and mitigates security threats while optimizing transmission energy costs, with adaptability to time-varying conditions. They further suggested blockchain integration to guarantee data integrity. Kumari and Tyagi [6] reviewed the design and applications of WSNs, stressing challenges related to scalability, energy conservation, and secure communication, and emphasized the potential of digital twins and blockchain for future deployments in urban environments.
Application-specific AI frameworks have also been investigated. Basingab et al. [7] introduced an AI-based decision support system for optimizing WSN resource allocation in consumer electronics. By predicting traffic patterns with machine learning, their system reduced energy usage while improving quality of service in e-commerce applications. Hu et al. [8] proposed a deep reinforcement learning (DRL)-based security mechanism for WSNs, integrating DRL with traditional security measures to achieve both secure and energy-efficient communication. They recommended federated learning as a future enhancement to improve scalability. Bairagi et al. [9] presented a recursive geographic forwarding protocol that reduced forwarding costs by exploiting spatial information, with potential extensions through machine learning for predictive routing.
Additional works have focused on integrating blockchain with AI. Satheeskumar et al. [10] developed a hybrid framework combining neural networks and blockchain to improve routing and ensure secure communication in WSNs. Ntabeni et al. [11] analyzed device-level energy-saving strategies for machine-type communications (MTC), including power management and RF optimization, and suggested AI-based approaches for real-time adaptation in resource-constrained IoT devices. Chinnasamy et al. [12] designed a blockchain-6G-enabled framework that incorporates machine learning for security management in IoT applications, recommending further advancements in predictive AI algorithms for real-time optimization.
Research into domain-specific IoT applications has also gained momentum. Venkata Prasad et al. [13] reviewed lightweight secure routing techniques for the Internet of Medical Things (IoMT), emphasizing hybrid models that combine AI-driven optimization with classical routing protocols. Kaur et al. [14] proposed an NSGA-III-based fog-assisted WSN architecture for emergency evacuation in large buildings, demonstrating its ability to optimize multiple objectives such as latency and energy efficiency. Ibrahim et al. [15] presented a 6G-IoT resource allocation framework leveraging Bayesian game theory and packet scheduling, with AI integration for adaptive reconfiguration in real time.
Overall, these studies reveal a growing trend toward AI- and blockchain-assisted WSN optimization, where clustering, routing, and security are designed to handle large-scale, heterogeneous, and dynamic environments. The collective findings demonstrate that energy efficiency, reliability, and scalability remain central challenges, while AI-driven adaptive solutions present the most promising direction for future IoT and WSN deployments.

3. Methodology

In wireless sensor networks (WSNs), energy conservation largely depends on efficient clustering mechanisms. Clustering involves designating specific nodes as cluster heads, which are responsible for aggregating data from their associated member nodes and transmitting it to the base station or sink. To optimize this process, AI-driven clustering algorithms have been introduced, where the selection of cluster heads is performed dynamically. The decision is influenced by factors such as the residual energy of candidate nodes, their relative proximity to neighboring nodes, and the prevailing traffic conditions. Reinforcement learning techniques are further employed to enhance this selection strategy, thereby extending network lifetime and ensuring balanced energy consumption across the nodes.

3.1. Energy-Aware Cluster Head Selection via Reinforcement Learning

Efficient energy utilization is a fundamental challenge in wireless sensor networks (WSNs), where limited battery capacity constrains the operational lifetime of sensor nodes. One effective strategy to mitigate excessive energy consumption is clustering, in which a set of nodes is designated as cluster heads (CHs) to aggregate and forward data to the sink. The selection of suitable CHs is critical, as uneven or static allocation can result in rapid energy depletion of specific nodes, leading to premature network failures. To address this issue, reinforcement learning (RL) offers an adaptive solution for CH selection by dynamically optimizing the decision-making process based on real-time network conditions. In this approach, nodes are evaluated using multiple parameters, including residual energy, distance to neighboring nodes, and current traffic load. The RL agent interacts with the network environment and progressively learns the optimal CH selection policy through trial-and-error updates, ensuring energy-efficient operation. The integration of RL not only balances energy consumption across the network but also enhances scalability and adaptability in heterogeneous environments. By preventing early node exhaustion and maintaining uniform energy distribution, RL-driven CH selection significantly improves network longevity, reduces packet loss, and ensures reliable data transmission. This makes it a promising approach for resource-constrained applications such as environmental monitoring, smart cities, and industrial IoT systems.

Adaptive Reward Function for Energy-Efficient Clustering

Efficient clustering is a critical mechanism in Wireless Sensor Networks (WSNs) to reduce redundant transmissions, minimize energy consumption, and extend the overall network lifetime. Traditional clustering protocols, such as LEACH and HEED, employ static or probabilistic mechanisms for cluster head (CH) selection, which often results in unbalanced energy usage among sensor nodes. To overcome this limitation, reinforcement learning (RL)-based clustering schemes introduce adaptive decision-making capabilities by dynamically adjusting CH selection strategies based on network state information.
In this context, the design of a reward function plays a pivotal role, as it determines how the agent (a sensor node or the system) evaluates its actions, learns optimal clustering policies, and achieves energy efficiency. An adaptive reward function incorporates multiple factors such as residual energy, distance to the base station, intra-cluster communication cost, and load distribution, thereby ensuring balanced energy consumption across nodes.
A.
System Model
Consider a WSN consisting of NNN sensor nodes uniformly deployed in a two-dimensional region. Each node iii has an initial energy E0E_{0}E0 and can act either as a member node or as a cluster head. The base station (BS) is assumed to be located either at the center or outside the sensing field.
The energy consumption for transmission is modeled using the first-order radio energy model:
ETX(k, d) = Eelec · k + ϵamp · k · dα
where
  • k = packet size (bits),
  • d = transmission distance,
  • Eelec = energy consumed per bit for processing,
  • ϵamp = amplifier energy,
  • α = path loss exponent (typically 2 ≤ α ≤ 4).
The energy consumed for receiving is:
ERX(k) = Eelec · k
Cluster heads consume additional energy for aggregation:
Eagg(k) = EDA · k
where EDA is the energy per bit required for data aggregation.
B.
Reinforcement Learning Framework
In the RL model, the clustering problem is formulated as a Markov Decision Process (MDP), defined by:
State (S): The state of a node is represented by a vector including residual energy (Eres), distance to BS (dB), average intra-cluster distance (dintra), and node degree (Ndeg).
Action (A): Each node chooses between acting as a cluster head (CH) or a member node (MN).
Reward (R): The adaptive reward function evaluates the energy efficiency and communication cost of the selected action.
C.
Adaptive Reward Function
The adaptive reward function is defined to balance energy preservation, load distribution, and network connectivity. The reward for node iii at time step t is given as:
Ri(t) = β1 · Eres(i, t)/E0 − β2 · di,BS/dmax − β3 · dintra(i)/davg − β4 · L(i, t)
where
  • Eres(i, t): Residual energy of node i at time t,
  • di,BS: Distance from node i to the BS
  • dmax: Maximum distance to BS in the network,
  • dintra(i): Average distance between node i and its cluster members,
  • davg: Average intra-cluster distance of the network,
  • L(i, t): Current load factor on node i, defined as the number of member nodes associated with it when acting as a CH,
  • β1, β2, β3, β4: Weighting factors that determine the importance of each parameter.
This reward formulation ensures that nodes with higher residual energy and shorter distances to BS are more likely to be selected as CHs, while penalizing nodes with high load or excessive intra-cluster communication cost.
D.
Optimization Objective
The goal is to maximize the cumulative reward:
J π = E [ t = 0 T γ ^ t   R i t ]
where
  • J(π) = expected return under policy π;
  • E = expectation operator;
  • γ = discount factor;
  • Rᵢ(t) = reward at time step t for agent i;
  • T = time horizon.
E.
Simulation Considerations
To validate the performance, the proposed adaptive reward-based clustering is compared against benchmark protocols such as LEACH, PEGASIS, and HEED under different node densities and network sizes. Key performance metrics include:
  • Network Lifetime (FND, HND, LND);
  • Energy Consumption per Round;
  • Packet Delivery Ratio (PDR);
  • Average Latency;
  • Load Distribution among CHs.
The adaptive reward function effectively integrates residual energy, communication distance, and traffic load to guide cluster head selection. Unlike static clustering, the reinforcement learning approach dynamically evolves based on network state transitions. As a result, the system achieves:
  • Balanced Energy Utilization: Prevents overburdening of specific nodes.
  • Scalability: Adapts well to large-scale WSNs.
  • Extended Lifetime: Minimizes early node failures through fair CH rotation.
  • Improved Reliability: Ensures higher data delivery success.

3.2. Deep Reinforcement Learning Model for Routing Efficiency

Routing optimization in wireless sensor networks (WSNs) and Internet of Things (IoT) systems is a complex problem due to dynamic network topologies, constrained resources, and varying traffic demands. Traditional routing protocols often rely on static heuristics or probabilistic approaches, which may fail to adapt under changing network conditions such as node failures, heterogeneous energy states, and fluctuating traffic loads. To overcome these limitations, Deep Reinforcement Learning (DRL), particularly Deep Q-Networks (DQNs), has emerged as a powerful tool for intelligent and adaptive routing.
In the context of routing, the environment is modeled as a Markov Decision Process (MDP), where each sensor node represents an agent that learns to make forwarding decisions based on local state observations. The state at time t typically includes parameters such as residual energy, node distance to sink, queue length, and link quality. The action at corresponds to the selection of the next-hop node for data forwarding. The agent receives a reward Rt after each action, which is designed to capture routing objectives such as minimizing energy consumption, reducing latency, and maximizing Packet Delivery Ratio (PDR).

Q-Learning Update Rule

The Q-learning update rule is given as:
Q(st, at) ← Q(st, at) + α[Rt + γa′maxQ(st + 1,a′) − Q(st,at)]
where α is the learning rate, and γ is the discount factor that balances immediate and long-term rewards.
However, in large-scale WSNs, the state–action space becomes prohibitively large, making traditional Q-learning inefficient. To address this, Deep Q-Networks approximate the Q-value function using a deep neural network parameterized by θ:
Q(s,a;θ) ≈ E[Rt + γmax a′Q(s′, a′; θ′)]
Here, the neural network generalizes across unseen states, enabling effective decision-making in complex environments. Techniques such as experience replay and target networks are employed to stabilize training, mitigate correlation between samples, and avoid oscillations during learning.
By employing DQNs, routing decisions adapt dynamically to network variations. Nodes learn to balance between exploiting energy-efficient routes and exploring alternative paths to avoid congestion or node depletion. Simulation results reported in the literature demonstrate that DQN-based routing significantly improves network lifetime, reduces packet loss, and enhances overall data reliability compared to conventional protocols such as LEACH, PEGASIS, and HEED.
Thus, integrating DQNs into WSN routing provides a scalable, robust, and intelligent solution, well-suited for applications in environmental monitoring, disaster response, smart cities, and industrial IoT.

4. Experimental Analysis

The evaluation of the proposed protocol is performed through simulations in a WSN environment composed of randomly deployed sensor nodes with limited energy resources across a defined geographical region. Performance is assessed using key metrics, including network lifetime, overall energy consumption, packet delivery ratio (PDR), and latency. Comparative analysis with conventional schemes such as LEACH, PEGASIS, and HEED illustrates that the proposed protocol achieves improved energy efficiency and enhanced network sustainability under varying operating conditions.
Figure 2 represents the number of alive nodes across several rounds for the AI-driven RL protocol and traditional protocols, such as LEACH, PEGASIS, and HEED. With an increase in rounds, it is quite apparent that the AI-driven RL protocol had maintained a higher value of alive nodes compared to other protocols, thereby depicting its efficiency in distributing energy consumption and preventing the premature death of nodes. The above improvement by the AI is due to adaptive decision-making regarding cluster head selection and routing to optimize the network lifetime. The steep decline in network lifetime is noted for LEACH, PEGASIS, and HEED networks because of static decision-making approaches.
In Round 1, the network topology illustrates the initial cluster formation in a wireless sensor network (WSN), where several nodes are randomly distributed across the field. Cluster heads (CHs) are denoted by red crosses, while ordinary sensor nodes are represented with blue crosses. The black dotted lines indicate associations between sensor nodes and their corresponding CHs, thereby forming clusters. Furthermore, the green solid lines highlight the communication paths from CHs to the sink (base station). This clustering mechanism enables energy-efficient communication, as individual nodes transmit data to nearby CHs rather than directly to the sink, thereby reducing overall energy consumption.
In Round 2, the selection of CHs is influenced by both the residual energy of nodes and their relative proximity. The newly elected CHs, again marked with red crosses, differ from those in the previous round, while the blue crosses continue to represent ordinary nodes. The changes in node–CH connections reflect a dynamic restructuring of clusters, which balances energy dissipation and prolongs the operational lifetime of the WSN. Such reorganization prevents excessive energy burden on specific nodes, thus mitigating uneven energy distribution and avoiding premature node failures.
In Round 3, further re-clustering occurs as new CHs are elected, represented once again by red crosses, with updated associations to sensor nodes. This continuous adaptive process ensures that energy expenditure remains distributed across the network. The CH selection considers real-time parameters such as residual energy levels and inter-node distances, demonstrating the adaptive capability of AI-assisted clustering protocols. By optimizing the utilization of available resources, this approach enhances the network lifetime while preventing rapid energy depletion of critical nodes, as illustrated in Figure 3.
Figure 4 illustrates the residual energy distribution among sensor nodes during the 10th round of the energy consumption balancing simulation. A color gradient is employed for visualization, where nodes highlighted in green correspond to higher residual energy, while those shaded in red represent nodes with relatively low remaining energy. Cluster heads (CHs) are denoted by larger red crosses, distinguishing them from ordinary sensor nodes. The dashed lines indicate the established communication links between sensor nodes and their designated CHs.
Figure 4, demonstrates the adaptive nature of the applied clustering algorithm. By periodically rotating the role of CHs and reconfiguring routing paths, the protocol effectively prevents excessive energy depletion of specific nodes. Such adaptive reorganization ensures that no single node is overburdened, thereby maintaining a balanced distribution of energy consumption across the network. Consequently, the network lifetime is extended, while simultaneously enhancing the efficiency of data aggregation and reducing redundant transmissions. Furthermore, this energy-aware mechanism guarantees reliable connectivity and equitable load sharing, which are critical factors in sustaining large-scale wireless sensor network (WSN) deployments.
Figure 5, presents the network coverage obtained in the fifth simulation round over a field of 100 × 100 units. The active sensor nodes are represented by green crosses, while the sink node is indicated by a red diamond. Each sensor node has a predefined sensing radius.
From the figure, it can be observed that the sensor nodes are randomly scattered, providing significant coverage of the deployment area. Several regions exhibit overlapping coverage zones, which increases fault tolerance by enabling multiple nodes to monitor the same area. This redundancy improves the reliability of data acquisition even in the presence of node failures. The placement of the sink node near the center facilitates direct communication with nearby nodes, while distant nodes rely on multi-hop transmission.
Overall, the results indicate that the network maintains effective spatial coverage with limited uncovered regions. Such coverage characteristics are crucial in wireless sensor networks, as they directly influence network lifetime, connectivity, and quality of service.
Figure 6 illustrates the time-evolution of cluster stability during the fifth simulation round. The network is organized into multiple clusters, each governed by a cluster head, while the remaining nodes act as members associated with their respective heads. The figure highlights the placement of cluster heads and the grouping of nodes, which collectively demonstrate the stable structure maintained across rounds of the clustering algorithm.
The formation of clusters demonstrates that sensor nodes consistently associate with their respective cluster heads, thereby minimizing the need for frequent reconfiguration. Such stability reduces the overhead associated with cluster re-formation and supports efficient energy utilization. The distribution of cluster heads across the sensing area further enables load balancing by avoiding excessive communication demand in localized regions.
By ensuring relatively stable cluster membership, the algorithm effectively prolongs the operational lifetime of the network. Nodes communicate with nearby cluster heads at lower transmission costs, while the balanced placement of heads across the field prevents premature depletion of energy in critical zones. These characteristics indicate that the clustering scheme contributes to energy efficiency, scalability, and enhanced reliability of the wireless sensor network.
A simulation was conducted to evaluate packet delivery performance in a wireless sensor network (WSN) with randomized node deployment. A total of 36 sensor nodes were distributed within a 100 × 100-unit area, each attempting to transmit a data packet to a centrally located sink node positioned at (55, 50). The results recorded 15 successful deliveries and 21 failed transmissions, as shown in Figure 7. The delivery outcomes were randomly assigned to reflect variable communication conditions such as signal attenuation, interference, or node energy levels. Analysis of the spatial distribution revealed that nodes located closer to the sink generally experienced higher success rates, whereas those positioned near the boundaries of the simulation field were more prone to delivery failures. This observation aligns with known limitations in WSNs, where greater transmission distances and limited relay options can adversely affect network reliability. The study underscores the importance of considering spatial placement and routing strategies in WSN design to enhance data delivery efficiency and reduce transmission loss. These findings provide a foundational understanding for future work involving dynamic routing protocols and energy-aware communication models in sensor networks.
The load distribution comparison, as shown in Figure 8, contains four routing protocols—LEACH, PEGASIS, HEED, and the AI-driven reinforcement learning approach—was evaluated over multiple simulation rounds. The horizontal bar chart indicates the number of packets handled by each protocol at different intervals.
In the initial rounds, all protocols exhibited comparable load distribution, with slight variations in packet handling. LEACH and PEGASIS showed moderate packet handling capabilities, whereas HEED maintained relatively balanced load allocation. However, the AI-driven reinforcement learning protocol consistently achieved higher load distribution, particularly in later rounds, demonstrating its adaptability and robustness.
At mid-level rounds (30–60), PEGASIS and HEED displayed improved performance compared to LEACH, suggesting better energy utilization and network balance. Despite these improvements, the AI-driven reinforcement learning scheme outperformed the traditional protocols by sustaining higher packet throughput.
In the final rounds (80–100), the superiority of the AI-driven reinforcement learning approach became more evident. While traditional protocols exhibited performance fluctuations, the AI-based protocol maintained steady and higher load distribution, indicating better scalability and prolonged network efficiency.
Overall, the results confirm that incorporating reinforcement learning mechanisms enhances load balancing, reduces network congestion, and ensures effective packet delivery when compared with conventional clustering and chain-based protocols.
The results in Table 1, clearly demonstrate that the AI-driven method consistently outperforms the baseline approaches across most metrics. In terms of network lifetime, the AI-driven protocol sustains operation for 90 rounds, compared to 58–80 rounds for the other schemes. This indicates better energy management and node participation. Similarly, throughput is substantially improved, reaching 185 packets, while LEACH, PEGASIS, and HEED achieve values in the range of 95–115 packets.
The latency performance also shows that the AI-driven protocol delivers lower delay (150 ms) relative to PEGASIS (195 ms) and the other two methods (165 ms each), ensuring faster communication. Furthermore, residual energy reaches 300 Joules in the AI-driven scheme, substantially higher than the 210–235 Joules observed in conventional protocols, highlighting the effectiveness of energy utilization.
In terms of energy efficiency, the AI-driven approach attains 95%, exceeding the 73–84% range of the other methods. The stability period is also prolonged to 75 rounds, demonstrating resilience and balanced cluster formation. Finally, packet loss is minimized to 35%, which is significantly lower than the 40–52% observed in existing schemes.
Overall, the comparative analysis indicates that the AI-driven protocol not only enhances energy utilization and packet delivery but also ensures higher stability and reduced communication overhead. These improvements collectively contribute to extending the network’s operational lifetime and reliability, making the AI-driven scheme more suitable for large-scale WSN deployments.

5. Conclusions

The proposed AI-driven energy-efficient Data Aggregation and Routing Protocol has demonstrated significant improvements in the performance of Wireless Sensor Networks (WSNs). By integrating reinforcement learning (RL) with deep Q-networks (DQNs), the protocol effectively addresses the fundamental challenge of energy efficiency while maintaining network stability and scalability. Comparative analysis with classical approaches, such as LEACH, PEGASIS, and HEED, reveals that the AI-based method consistently achieves longer network lifetime, higher throughput, reduced latency, and improved energy efficiency. Furthermore, packet loss is considerably minimized, ensuring reliable data delivery across the network.
The simulation outcomes emphasize that the incorporation of adaptive cluster head selection and intelligent routing strategies allows for balanced energy consumption, thereby reducing reconfiguration overheads and prolonging the overall network operation. These advantages make the protocol highly suitable for large-scale applications, including environmental monitoring, smart city deployments, and Internet of Things (IoT) ecosystems.
In addition, this work highlights the potential of artificial intelligence, particularly RL and DQNs, to address both energy and scalability issues that are intrinsic to WSNs. The results establish a pathway for more robust, resilient, and sustainable sensor networks. Future research directions include the adoption of multi-agent reinforcement learning (MARL) techniques to enable collaborative learning among nodes, the consideration of node or sink mobility to adapt to dynamic real-world environments, and the incorporation of security mechanisms to safeguard against malicious activities such as jamming or data manipulation. Real-world large-scale deployment will further validate the scalability and practical viability of the proposed protocol.
In conclusion, the findings of this study underscore the role of AI-driven methods as a transformative approach in advancing WSN protocols, setting the foundation for next-generation IoT and smart sensing applications.

Author Contributions

Conceptualization, R.A.C. and C.S.; methodology, R.A.C.; software, M.A.; validation, R.A.C., C.S. and M.B.; formal analysis, M.A.; investigation, C.S.; resources, R.A.C.; data curation, M.B.; writing—original draft preparation, R.A.C.; writing—review and editing, C.S. and M.A.; visualization, M.B.; supervision, R.A.C.; project administration, R.A.C.; funding acquisition, R.A.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Javadpour, A.; Sangaiah, A.K.; Zaviyeh, H.; Ja’fari, F. Enhancing energy efficiency in IoT networks through fuzzy clustering and optimization. Mob. Netw. Appl. 2023, 29, 1594–1617. [Google Scholar] [CrossRef]
  2. Gebremariam, G.G.; Panda, J.; Indu, S. Design of advanced intrusion detection systems based on hybrid machine learning techniques in hierarchically wireless sensor networks. Connect. Sci. 2023, 35, 2246703. [Google Scholar] [CrossRef]
  3. Labib, M.N.; Omran, N.F.; Mohamed, E.M.; Abdelnapi, N.M. Design of an enhanced threshold sensitive distributed energy efficient clustering routing protocol for WSN-based IoT. Int. J. Electron. 2023, 110, 1373–1392. [Google Scholar] [CrossRef]
  4. Priyadarshi, R. Energy-efficient routing in wireless sensor networks: A meta-heuristic and artificial intelligence-based approach: A comprehensive review. Arch. Comput. Methods Eng. 2024, 31, 2109–2137. [Google Scholar] [CrossRef]
  5. Haseeb, K.; Alruwaili, F.F.; Khan, A.; Alam, T.; Wafa, A.; Khan, A.R. AI assisted energy optimized sustainable model for secured routing in mobile wireless sensor network. Mob. Netw. Appl. 2024, 29, 867–875. [Google Scholar] [CrossRef]
  6. Kumari, S.; Tyagi, A.K. Wireless sensor networks: An introduction. In Digital Twin and Blockchain for Smart Cities; John Wiley & Sons, Inc.: Hoboken, NJ, USA, 2024; pp. 495–528. [Google Scholar]
  7. Basingab, M.S.; Bukhari, H.; Serbaya, S.H.; Fotis, G.; Vita, V.; Pappas, S.; Rizwan, A. AI-based decision support system optimizing wireless sensor networks for consumer electronics in e-commerce. Appl. Sci. 2024, 14, 4960. [Google Scholar] [CrossRef]
  8. Hu, L.; Han, C.; Wang, X.; Zhu, H.; Ouyang, J. Security enhancement for deep reinforcement learning-based strategy in energy-efficient wireless sensor networks. Sensors 2024, 24, 1993. [Google Scholar] [CrossRef] [PubMed]
  9. Bairagi, P.P.; Dutta, M.; Babulal, K.S. An energy-efficient protocol based on recursive geographic forwarding mechanisms for improving routing performance in WSN. IETE J. Res. 2024, 70, 2212–2224. [Google Scholar] [CrossRef]
  10. Satheeskumar, R.; Prakash, B.; Velliangiri, S.; Shajin, F.H. Evolutionary gravitational neocognitron neural network based blockchain technology for a secured dynamic optimal routing in wireless sensor networks. J. Exp. Theor. Artif. Intell. 2024, 36, 435–451. [Google Scholar] [CrossRef]
  11. Ntabeni, U.; Basutli, B.; Alves, H.; Chuma, J. Device-level energy efficient strategies in machine type communications: Power, processing, sensing, and RF perspectives. IEEE Open J. Commun. Soc. 2024, 5, 5054–5087. [Google Scholar] [CrossRef]
  12. Chinnasamy, P.; Babu, G.C.; Ayyasamy, R.K.; Amutha, S.; Sinha, K.; Balaram, A. Blockchain 6G-based wireless network security management with optimization using machine learning techniques. Sensors 2024, 24, 6143. [Google Scholar] [CrossRef] [PubMed]
  13. Prasad, K.V. Revolutionary of secure lightweight energy efficient routing protocol for internet of medical things: A review. Multimedia Tools Appl. 2024, 83, 37247–37274. [Google Scholar]
  14. Kaur, L.; Kaur, R. Energy-efficient smart architecture for fog-based WSN using NSGA-III and improved layer-wise clustering for enhanced building evacuation safety. Ann. Oper. Res. 2024, 1–24. [Google Scholar] [CrossRef]
  15. Ibrahim, A.M.; Chen, Z.; Eljailany, H.A.; Yu, G.; Ipaye, A.A.; Abouda, K.A.; Idress, W.M. Advancing 6G-IoT networks: Willow catkin packet transmission scheduling with AI and Bayesian game-theoretic approach-based resource allocation. Internet Things 2024, 25, 101119. [Google Scholar] [CrossRef]
Figure 1. Data Aggregation and Routing Protocol.
Figure 1. Data Aggregation and Routing Protocol.
Ndt 03 00022 g001
Figure 2. Network Lifespan Comparison.
Figure 2. Network Lifespan Comparison.
Ndt 03 00022 g002
Figure 3. Cluster Formation.
Figure 3. Cluster Formation.
Ndt 03 00022 g003
Figure 4. Energy Consumption Balancing Network Simulation.
Figure 4. Energy Consumption Balancing Network Simulation.
Ndt 03 00022 g004
Figure 5. Network Coverage.
Figure 5. Network Coverage.
Ndt 03 00022 g005
Figure 6. Cluster Node Detection and Stability.
Figure 6. Cluster Node Detection and Stability.
Ndt 03 00022 g006
Figure 7. Packet Delivery.
Figure 7. Packet Delivery.
Ndt 03 00022 g007
Figure 8. Load Distribution Comparison.
Figure 8. Load Distribution Comparison.
Ndt 03 00022 g008
Table 1. WSN Protocol Comparison—Extended Performance.
Table 1. WSN Protocol Comparison—Extended Performance.
MetricLEACHPEGASISHEEDAI-Driven Method
Network Lifetime (Rounds)80585890
Throughput (Packets)9596115185
Latency (ms)165195165150
Residual Energy (Joules)210222235300
Energy Efficiency (%)73768495
Network Stability (Rounds)64626475
Packet Loss (%)40455235
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chakravarthy, R.A.; Sureshkumar, C.; Arun, M.; Bhuvaneswari, M. AI-Driven Energy-Efficient Data Aggregation and Routing Protocol Modeling to Maximize Network Lifetime in Wireless Sensor Networks. NDT 2025, 3, 22. https://doi.org/10.3390/ndt3040022

AMA Style

Chakravarthy RA, Sureshkumar C, Arun M, Bhuvaneswari M. AI-Driven Energy-Efficient Data Aggregation and Routing Protocol Modeling to Maximize Network Lifetime in Wireless Sensor Networks. NDT. 2025; 3(4):22. https://doi.org/10.3390/ndt3040022

Chicago/Turabian Style

Chakravarthy, R. Arun, C. Sureshkumar, M. Arun, and M. Bhuvaneswari. 2025. "AI-Driven Energy-Efficient Data Aggregation and Routing Protocol Modeling to Maximize Network Lifetime in Wireless Sensor Networks" NDT 3, no. 4: 22. https://doi.org/10.3390/ndt3040022

APA Style

Chakravarthy, R. A., Sureshkumar, C., Arun, M., & Bhuvaneswari, M. (2025). AI-Driven Energy-Efficient Data Aggregation and Routing Protocol Modeling to Maximize Network Lifetime in Wireless Sensor Networks. NDT, 3(4), 22. https://doi.org/10.3390/ndt3040022

Article Metrics

Back to TopTop