Energy Harvesting Techniques for Wireless Sensor Networks/Radio-Frequency Identification: A Review

In the near future, symmetry technologies for the Internet of Things (IoT), along with symmetry and asymmetry applications for IoT security and privacy, will re-design the socio-ecological human terrain morphology. The IoT ecosystem deploys diverse sensor platforms connecting billions of heterogeneous objects through the Internet. Most sensors are low-energy consuming devices which are designed to transmit sporadically or continuously. However, when we consider the billions/trillions of connected sensors powering various user applications, their energy efficiency (EE) becomes a critical issue. Therefore, the importance of EE in IoT technology cannot be overemphasised, specifically the development of EE solutions for sustainable IoT technology. Propelled by this need, EE proposals are expected to address IoT’s EE issues. Consequently, many developments have been displayed, and highlighting them to provide clear insights into eco-sustainable and green IoT technologies is becoming a crucial task. To pursue a clear vision of green IoT, this article aims to describe the current state-of-the art insights into energy-saving practices and strategies on green IoT. The major contribution of this study is the review and discussion of the substantial issues enabling hardware green IoT to focus on green wireless sensor networks and green radio-frequency identification. This review paper will contribute significantly to the future implementation of green and eco-sustainable IoT.


Introduction
The Internet of Things (IoT) is a paradigm which aims to advance telecommunications in all spheres of human life. In the near future, symmetry technologies for the Internet of Things (IoT), along with symmetry and asymmetry applications for IoT security and privacy, will re-design the socio-ecological human terrain morphology. This advancement leads to substantial improvement in quality of human life and the world's economic growth at large. IoT is considered the backbone of emerging applications (Figure 1) [1], as innovation plays a key role in the massive evolution of machine communications. Machine-to-machine (M2M) traffic is estimated to account for approximately 45% of the total Internet traffic by 2022 [2]. IoT creates a platform in which physical objects can mimic certain human sensory capabilities, such as perception, vision, hearing, smell and thinking. Buoyed by these human sensory capabilities and the emerging tactile Internet, machines can communicate with one another, share relevant information and make real-time decisions with less human input, especially now that we are migrating to the 5G era, in which the expected wireless network delay is 1 ms. In this scenario, divergent sensors, such as radio-frequency identification (RFID) and sensing technology, can undertake the collaborative function of sensing, collecting and transmitting sensor information (RFID) and sensing technology, can undertake the collaborative function of sensing, collecting and transmitting sensor information through the Internet [2]. Experts report that IoT holds incredible potential for smart homes, smart cities and healthcare applications [3]. Healthcare sectors can generate an annual revenue of $1.1-2.5 trillion by 2025 [4]. Moreover, the worldwide economic effect of IoT is approximated at between $2.7 and $6.2 trillion by 2025 [2]. With the rise in ubiquitous IoT innovations, it has dawned on the telecom operators that innovative, and creative solutions are the essential tools needed to tackle the challenges and the potential of IoT. Today, the number of objects with Internet connectivity capability is greater than that of the Earth's population. As the capabilities of the IoT expand, it is expected that close to 50 billion devices will have Internet connectivity by 2020 [5]. Generally, the IoT will be capable of interconnecting billions/trillions of heterogeneous objects through the Internet by using different types of sensors like RFID and sensing technology. Both technologies will undertake collaborative functions of sensing, collecting and transmitting sensor information via the Internet.
Though active RFID tags are low energy consuming wireless devices [6], when these RFID tags are connected in an industrial scale involving billions of them, they generate millions of payload transmitted to the data centres for processing. To process these enormous sensor payload, huge processing and analytics capabilities are needed, which, in turn, consumes huge energy resources [7,8]. Herein, it is necessary to consider the issue of large-scale consumption of energy resources by IoTs. Motivated by this, a relatively novel research interest has evolved known as "green IoTs". This initiative not only tend to improve EE but also reduce the emission of CO2 from ICT products [9]. It is generally accepted that IoT will have great economic and ecological significance in years to come. Driven by these realities, it has become imperative to analyse the state-of-the art techniques and approaches capable of tackling the energy needs of these billions of energy hungry devices. Thus, the green IoT has become an important topic for researchers and vendors now more than ever as the conventional energy resources are dwindling and the energy consumption increasing exponentially.
Green IoT initiatives concentrate on the IoT EE. Thus, green IoT is defined as the process of achieving EE in IoT technology. This entails making every design process in IoT green until the implementation phase [10]. This has motivated researchers in both academia and industry to develop different techniques to improve IoT EE. Accordingly, research on the green IoTs encompasses wide spectrum of topics, research issues and challenges. Green IoT techniques can be classified into two main categories [11]: (i) Software, which focuses on energy wastage as a result of inefficient resource utilization, i.e., algorithms; and (ii) Hardware, which focuses on improving the EE in IoT components, and it is our target in this study. With the rise in ubiquitous IoT innovations, it has dawned on the telecom operators that innovative, and creative solutions are the essential tools needed to tackle the challenges and the potential of IoT. Today, the number of objects with Internet connectivity capability is greater than that of the Earth's population. As the capabilities of the IoT expand, it is expected that close to 50 billion devices will have Internet connectivity by 2020 [5]. Generally, the IoT will be capable of interconnecting billions/trillions of heterogeneous objects through the Internet by using different types of sensors like RFID and sensing technology. Both technologies will undertake collaborative functions of sensing, collecting and transmitting sensor information via the Internet.
Though active RFID tags are low energy consuming wireless devices [6], when these RFID tags are connected in an industrial scale involving billions of them, they generate millions of payload transmitted to the data centres for processing. To process these enormous sensor payload, huge processing and analytics capabilities are needed, which, in turn, consumes huge energy resources [7,8]. Herein, it is necessary to consider the issue of large-scale consumption of energy resources by IoTs. Motivated by this, a relatively novel research interest has evolved known as "green IoTs". This initiative not only tend to improve EE but also reduce the emission of CO 2 from ICT products [9]. It is generally accepted that IoT will have great economic and ecological significance in years to come. Driven by these realities, it has become imperative to analyse the state-of-the art techniques and approaches capable of tackling the energy needs of these billions of energy hungry devices. Thus, the green IoT has become an important topic for researchers and vendors now more than ever as the conventional energy resources are dwindling and the energy consumption increasing exponentially.
Green IoT initiatives concentrate on the IoT EE. Thus, green IoT is defined as the process of achieving EE in IoT technology. This entails making every design process in IoT green until the implementation phase [10]. This has motivated researchers in both academia and industry to develop different techniques to improve IoT EE. Accordingly, research on the green IoTs encompasses wide spectrum of topics, research issues and challenges. Green IoT techniques can be classified into two main categories [11]: (i) Software, which focuses on energy wastage as a result of inefficient resource utilization, i.e., algorithms; and (ii) Hardware, which focuses on improving the EE in IoT components, and it is our target in this study.
In the recent times, green IoT survey papers have attracted wide attention [1,10,[12][13][14][15]. The authors in [12] analysed various strategies for achieving green IoT; however, explicit green IoT models were never considered. Baliga et al. [13] discussed various cloud energy consumption scenarios. However, their models never discussed Quality of Service (QoS) metrics capable of increasing the energy consumption further in some scenarios. Reference [14] provided an in-depth discussions on energy harvesting in wireless sensor network (WSN) by exploiting various environmental resources. However, storing the harvested energy in a different medium other than the device will result in energy loss and hence, more work is needed in this regard. Reference [15] proposed that by implementing EE in heating, ventilating and air conditioning, could result in high energy saving design. Though extensive work has been done in the area of green IoT, energy conservation models are yet to be analysed. Reference [1] provided an extensive analysis of green IoT strategies, leading to the proposition of five green IoT principles. Moreover, the authors considered the case study approach to be a vital tool of IoT (in smart phones). However, the above studies lack depth in their explanation.
Since then, both sensor networking and RFID have been considered the two main pillars of IoT applications. In a hot research topic in ICT like green WSN and green RFID, there are many developments that quickly come into the spotlight and need to be highlighted in order to provide clear insights for researchers to choose the best solutions that provide green IoT and eco-sustainability. Thus, this review paper is different from the other review papers related to green IoT, in that this study gives full consideration to discussing green WSN and green RFID, which is considered one of the two main pillars of IoT applications, and this study aims to provide an overview of current state-of-the-art energy saving practices and strategies for green WSN and green RFID. To this end, this study tries to incorporate as many directions into this article as possible. Restricted by size constraints, the hot research topics have been investigated deeply based on their respective sub-domains to achieve a precise, concrete and concise conclusion at the end of this article. The key contributions of this study are summarized as follows: • Presents an overview of popular research topics on green WSNs and RFID ecosystem, covering the recent industry development in the main areas of application, challenges and key players; • Addresses several substantial design choices and features for WSNs and RFID, both of which are considered the top priorities of green IoT technologies. These features are deeply investigated on the basis of their respective sub-domains to achieve a precise, concrete and concise conclusion; • Provides new references to other researchers who need insights into enabling hardware green IoT which provides eco-sustainability.
The organization of this article is as thus; Section 2 presents an overview of the green WSNs. Section 3 provides a detailed discussion of the RFID EE solutions. Finally, Section 4 concludes the work.

Green Wireless Sensor Network
Wireless sensors are an integral component of smart applications based on IoT technology. They are miniaturized, inexpensive devices equipped with the capability to detect parameter of interest and periodically transmit the results to the collection point. Mostly, they are powered by battery. A set of interconnected wireless sensors known as WSN is deployed. A WSN architecture comprises of wireless sensor nodes linked to a base station (BS) acting as the sink node. In formulating WSN standards, nodes source energy and computing resources are of great importance. Two topologies exist for WSN, which can either be infrastructure-based or infrastructure-less (ad-hoc)-based. In infrastructure topology, the BS performs the primary network core functions of; scheduling, resource allocation, interference management, routing and serving as the gateway to the local area network. Conversely, in infrastructure-less system architecture, there is no centralized BS, as each sensor can undertake the role of system coordination [16]. A description of low-power wireless sensor communication standards including: Long Range (LoRa), Bluetooth, LR-WPAN, Mobile communication, WiMAX, and WiFi is summarized in [17][18][19]. Table 1 provides a summarised comparison of wireless systems on the basis of notable attributes, such as standard, energy consumption, frequency band, data rate, transmission range and cost. Table 2 shows the suitability of these wireless technologies for IoT applications. Wireless sensors are designed to continuously and autonomously transmit data for a long period. WSNs that continuously send data are sensitive applications of weather forecast, traffic report, water quality, healthcare and embedded system. Wireless sensors are powered by batteries. However, the replacement or maintenance cost of exhausted batteries is prohibitive, and these batteries are difficult to replace, especially in remote locations due to geographical limitations (e.g., challenging terrain) [20]. Accordingly, developing energy-aware solutions to increase battery life and reduce replacement costs has become indispensable for WSN sustainability.  Therefore, EE solutions have become crucial. The summary of the proposed major existing energy-saving mechanisms is as follows: (i) radio optimization techniques (transmission power control [21,22], cooperative communication [23,24], and modulation optimization [25,26]); (ii) sleep/wakeup schemes (topology control [27,28], and duty cycling schemes [29][30][31]); (iii) energy harvesting and wireless charging (utilizing energy harvesting [32][33][34], and wireless charging [35][36][37]); (iv) energy-efficient routing and WSN's architecture (cluster architectures [38,39], multipath routing [40,41], and relay node placement [42][43][44]); and (v) data reduction mechanisms (aggregation [45], adaptive sampling [29], compression [46], and network coding [47]). The summarised taxonomy of green WSNs is given in Figure 2. In the following sections, a detailed discussion on the existing energy-saving mechanisms is presented. Symmetry 2019, 11, x FOR PEER REVIEW 5 of 23

Radio Optimisation Techniques
The radio unit is the most pronounced energy consumption unit in the WSN. The energy depletion by the radio unit is caused by two parts: (i) powering the circuit, and (ii) powering of the transmitted signal. Short distances utilize more energy in powering the circuit, while powering the transmitted signal in long range communication consumes more power. Several references [22,48] have investigated the strategies to enhance EE by adjusting transmission power level dynamically. In addition, the authors in [21] proposed advance saving energy cooperative topology, in which sensor nodes with higher remaining energy is at liberty to increase transmitting power leading to other nodes to decrease their own transmitting power. Moreover, the proposed topology reduces interference and improves connectivity due to the decrease in transmission power. However, an increase in delay potentials is expected, because more hops will be required for packet forwarding. Nevertheless, the problem of delay can be overcome through cooperative communications among the neighbouring sensor nodes, which creates a virtual multiple-antenna environment (spatial

Radio Optimisation Techniques
The radio unit is the most pronounced energy consumption unit in the WSN. The energy depletion by the radio unit is caused by two parts: (i) powering the circuit, and (ii) powering of the transmitted signal. Short distances utilize more energy in powering the circuit, while powering the transmitted signal in long range communication consumes more power. Several references [22,48] have investigated the strategies to enhance EE by adjusting transmission power level dynamically. In addition, the authors in [21] proposed advance saving energy cooperative topology, in which sensor nodes with higher remaining energy is at liberty to increase transmitting power leading to other nodes to decrease their own transmitting power. Moreover, the proposed topology reduces interference and improves connectivity due to the decrease in transmission power. However, an increase in delay potentials is expected, because more hops will be required for packet forwarding. Nevertheless, the problem of delay can be overcome through cooperative communications among the neighbouring sensor nodes, which creates a virtual multiple-antenna environment (spatial diversity). Virtual multiple-antenna reduces data retransmission, effectively improving the quality of the received signal by overcoming multi-path fading and shadowing phenomenon. References [49] and [50] extended the communication range among the sensor nodes as well as higher energy conservation and lower end-to-end (E-2-E) delays over certain broadcasting coverage as reported by Cui et al. [23]. Jayaweera [24], compared the energy consumption of both Single Input Single Output and virtual multiple-antenna (Multiple Input and Multiple Output) systems and showed that virtual multiple-antenna systems can provide higher energy savings and minimize E-2-E delays over certain propagation range distances. On the other hand, Cui et al. [25] examined the relationship among the energy consumption, transmission time, and bit error rate. The results showed that optimising the transmission time could have minimised the energy utilization needed to attain a stated bit error rate as well as delay requirement. Moreover, the authors in [26] presented a comparative study on the EE of three modulation techniques to select the optimal modulation scheme that yields the lowest energy utilization with various distances between nodes. To address the problem of joint transmission power and rate allocation in the uplink of a cellular wireless network, several works have considered transmission power and rate allocation. The two basic approaches that have been proposed in the recent literature are explained below.
• Joint rate and power control are modelled as two distinct games (i.e., an uplink transmission rate allocation and an uplink transmission power allocation problem), which are based on the game theoretic perspective. Users determine first their uplink transmission rate and then given their uplink transmission rate, they apply power control to allocate their uplink transmission powers [51,52]. The main drawback of this approach is that the optimisation problem is solved asynchronously and separately considering the two systems' resources. Thus, the combined outcome of the two distinct optimisation problems is less efficient than jointly solving the problem [53].

•
The joint rate and power control problem is amended in a single-variable problem of the ratio of uplink transmission rate to the uplink transmission power [54]. However, this approach is limited in realistic cases and can only be applied in specific studies where simplified forms of utility functions are assumed (i.e., where the ratio of uplink transmission rate to power appears). As a result, the use of this approach strongly depends on problem formulation. The single variable problem is solved with respect to the substituted ratio. To determine users' optimal pair of uplink transmission rate and power, the maximum value of one resource is assumed and the other one is determined, so the ratio is equal to the optimal one. Although users update their uplink transmission rate and power in the same step, the obtained solution remains inferior compared with the corresponding solution of the actual joint two-variable optimisation problem discussed in [53].
Tsiropoulou et al. [53] proposed a novel utility-based game theoretic framework to address the problem of joint transmission power and rate allocation in the uplink of a cellular wireless network. Initially, each user is associated with a generic utility function that can properly express and represent mobile users' degree of satisfaction in relation to the allocated system's resources for heterogeneous services with various transmission rates. Then, a joint utility-based uplink power and rate allocation (JUPRA) game is formulated, where each user selfishly aims to maximise his/her utility-based performance under the imposed physical limitations, and its unique Nash equilibrium is determined with respect to both variables (i.e., uplink transmission power and rate). The JUPRA game's convergence to its unique Nash equilibrium is proven, and a distributed, iterative and low-complex algorithm for computing JUPRA game's equilibrium is introduced. The results showed the superiority of the proposed framework over other various state-of-the-art approaches.

Sleep/Wake Up Techniques
Switching off (sleep mode) the non-active transceivers have become the ultimate approach towards the realization of EE in Information and communications technology, due to the fact that it can save large amount of energy. The philosophy behind the proposed approach is to exploit dense and redundant deployment of sensor nodes, leading to a small coverage area. The sensor nodes' off/on switching approach is more desirable for improving WSN EE and prolonging the battery lifetime of the wireless sensors. However, the coverage issue should be considered, and it should be guaranteed by the remaining active nodes. Misra et al. [27] proposed a subset solution in which nodes with minimum overlap areas are activated and must be capable of reducing network energy. Meanwhile, Karasabun et al. [28] modelled the EE issue as a subset selection problem of active connected sensors for correlated data payload gathering. Using spatial correlation, the sensor information of non-active sensor nodes can be obtained from those of active nodes, which makes it a good strategy. Equation (1) gives the average power consumption as the sleep power multiplied by the percentage of duration the system is in sleep mode plus the active power multiplied by the percentage of duration the system is in active mode all divided by 100. In a situation in which the system is designed to have bigger sleep energy comparable to the active energy, it is then feasible to engage power reduction strategy by tuning the sensor node to its lowest power mode. There are two scenarios in which the active power term can be larger than the sleep power term either (i) the power ratio per event is large or (ii) active power events have higher frequency.
On the other hand, one can exploit the duty cycling schemes to make a sensor node switched on/off based on network activity (traffic conditions). Duty cycling schemes can be classified into three categories: on-demand, asynchronous, and scheduled rendezvous. Meanwhile, duty cycle-based protocols are certainly the most EE [29,30]. However, it should be taken into account that the low duty cycle has the capability to conserve a large big volume of energy but can lead to high communication delays. To reduce the delay, the protocol parameters can be tuned before deployment for ease, although it may result in inflexibility, or dynamical settings can be deployed to reflect the instantaneous traffic conditions. Moreover, the active period of nodes in order to optimise power consumption is a function of the traffic load, buffer overflows, delay requirements or harvested energy are discussed in [31].

Energy Harvesting and Wireless Charging Techniques
Key features of wireless sensors energy source such as sustainability and reliability, as well as reduction of greenhouse gas emissions can be met through advances in renewable energy technology [55]. Moreover, the renewable energy technology is one of the promising ways to address the EE issue of WSN located in rural and remote areas. In this terrain profile, it is difficult to replace batteries due to geographical limitations (challenging terrain), which makes access to these sites difficult [32]. The solar cells have low maintenance needs and high reliability, with an expected life span of 20-30 years. Additionally, new sensor technologies have emerged that harness power from their immediate environments, such as wind and kinetic energy [33]. The harvested energy is then converted to electrical signals, which are either consumed directly or stored for later usage. For example, using solar panels to charge a rechargeable battery during daytime. At night, nodes switch to conservative mode drawing energy from the stored power. During the protocol design stage, consideration must be given to uneven residual energy distribution, which is the difference in the quantity of energy collected [34]. In a case that there is no power to harness, battery life cycle capacity is formulated to calculate whether its total storing capability referenced to the magnitude of charge/discharge cycles, stated as depth-of discharge (DoD), is enough for the job. The life cycle capacity is estimated as: Life cycle capacity = Rated battery capacity × Rated charge _ discharge cycle life × DoD (2) For example, for the ML1220 rechargeable coin cell, the rated capacity is 17 mAh, charge_discharge cycles is 1000 cycles, and DoD is 10%; thus life cycle capacity = 17 mAh × 1000 cycles × 10% per cycle = 1.7 Ah.
Renewable energy technology is associated with energy estimation schemes for astute energy management. Thus, there is the need to undertake inept energy-saving mechanisms in addition to renewable energy technology in order to attain a high reliability status. The sensors may incorporate dynamic behaviour tendencies in the face of the estimated energy not been able to sustain them in the next recharge cycle. Hence, they can optimise decisive parameters such as sampling rate, transmit power and duty cycling to adapt their power consumption according to the periodicity and magnitude of the harvestable source. On the other hand, it is justifiable to allocate sensor nodes with large residual power with bigger sleep duration and shorter RF range, whereas, those with bigger residual power are selected as the preferable routing route [56]. However, efforts have not been made to develop protocols assuming battery degradation over time (leakage, storage loss), which will impact WSN performance.
The attributes and operations of the renewable energy sources available in outdoor environmental conditions are very different from those found in indoor industrial and commercial environments. Table 3 shows a summary of the indoor and outdoor energy sources and their characteristics. Given the wide spectrum of IoT device formats, applications, and use cases to choose from, it is nearly impossible to authenticate with assurance that a given device will make good a IoT energy harvesting device without prior knowledge of application specifics and system operations. This notwithstanding, some indices are available to grade some devices as likely energy harvesting devices or otherwise. Inspiration can be drawn from the viewpoint of cost-effectiveness and technical standpoint. Table 4 is a sample IoT devices/applications and their suitability for use with energy harvesting sources in South Korea.
The evolution of wireless power charging technology has made it possible for energy constraint devices to maintain functionality in a more controllable manner, thus increasing the sustainability and reliability of WSNs [17]. Today, we can see that the wireless power charging concept has already been applied in numerous applications, such as power medical sensors and implantable devices [57], to restock sensors embedded in concrete wall [58], and to power a ground sensor from an unmanned aerial vehicle [59]. Generally, energy transfer techniques can be classified into (i) non-radiative coupling-based charging, which can be classified into three techniques: magnetic inductive coupling [60], magnetic resonance coupling [61], and capacitive coupling [62]; and (ii) radiative RF-based charging, which can be classified into two techniques: directive RF power beamforming and non-directive RF power transfer [63]. However, in capacitive coupling, the achievable amount of coupling capacitance is dependent on the available area of the device [64]. Nevertheless, for a typical-sized portable electronic device, it is hard to generate sufficient power density for charging, which imposes a challenging design limitation. As for directive RF power beamforming, the limitation lies in fact that the charger needs to know the exact location of the energy receiver [65]. Due to the obvious limitation of above two techniques, wireless charging is usually realized through three other techniques: magnetic inductive coupling, magnetic resonance coupling, and non-directive RF radiation [16]. In non-directive RF radiation schemes, electric energy is sent as an electromagnetic radiation within the RF spectrum of 300 GHz and 3 kHz [66]. RF energy transfer is suitable for far-field communications. Experiences have indicated that RF power transfer has poor RF-to-DC energy conversion efficiency when confronted with RF harvested power. Detailed information on this subject is presented in [67,68]. Using Maxwell's equation, electric current is initially generated from a magnetic coupling that is tuned to resonate at the centre frequency [69]. Electric energy is then transported through the magnetic field. Lastly, magnetic resonance coupling is generated by an evanescent field, which generates and sends electrical energy between two resonators [61]. To attain this type of resonator, a capacitance is inserted between an induction coil. Inductive and magnetic resonance coupling techniques are classified as short-range communications, also known as near-field wireless communications (NFC) [70]. Near-field wireless transmission is characterised by high power conversion efficiency, which is dependent on the coupling coefficient and the distance between two coils/resonators. The operating range (distance between transmitter and receiver) is a major challenge for near-field wireless transmissions because power rapidly decreases as the distance between devices increases (i.e., inverse square law) [71]. The advantages, disadvantages, and effective charging distance of these three techniques are summarised in Table 5.
Wireless energy delivery to deployed sensor nodes have been investigated in several studies [35][36][37]. Wireless energy delivery is a new frontier that must be explored by wireless charging technologies, because it creates the environment in which sensor nodes are able to share energy between neighbours. Therefore, in the nearest future, wireless networks nodes are envisioned to incorporate energy sharing scheme by harvesting energy from the environment and transfer some of these energy to other sensor nodes, making a self-sustaining network [72]. To realise this paradigm, multi-hop energy harvesting techniques have been studied [68], thus opening a new dimension in the development of wireless charging protocols and energy cooperative systems, as well as energy-efficient routing. Simultaneous Wireless Information and Power Transfer (SWIPT) is a technique that can reduce the total recharging cost to enhance the lifespan of WSNs. The concept of SWIPT was first introduced in [73]. SWIPT has recently attracted attention in the field of wireless communication networks [66,74,75]. In the era of 5G communication, SWIPT technology could be fundamentally important for energy and information transmissions within numerous types of modern communications networks.
In WSNs, power usage is usually divided into three parts [76]: sensing, data processing and communication. Among the three, communication is the most costly function in a typical sensor node [77]. Hence, local data processing is important to minimise the power utilisation of WSNs because the failure of any node can re-route the structure. In addition, robust communication protocols are critical to efficiently exploit energy sources. Peer et al. [78] proposed a hybrid TS and PS spectrum sharing protocol for EH wireless sensor nodes. However, communication protocols must be further investigated to offer an optimal self-sustainable solution. Moreover, clustered WSNs with SWIPT that re-charge low energy relay nodes in cooperative clustered WSNs have been considered in [79], in which the authors considered a WSN consisting of multiple clusters of sensors and a sink node that collects data from the sensors in the clusters. To enable SWIPT, the head node of each cluster works as an information and power transferred. Relay node receiver harvests the RF energy from the head node of the cluster and recharges the batteries. Guo et al. [79] proposed a distributed iteration algorithm with closed-form transmission power, PS ratio and relay selection by utilising decomposition. Tong et al. [80] focused on the development of network deployment and its routing strategy. The idea is to reduce the total recharging cost to enhance the lifespan of WSNs. An optimisation problem is formulated using joint network deployment and routing in [81] by assuming that the sensors can be continuously recharged before their power source is exhausted and by obtaining complete knowledge of CSI. A few prototype implementations of sensor nodes are investigated in [81] and [82]. By reaping the benefits of SWIPT, low power WSNs deployed in the health care sector can recharge power sources while in operation, thereby improving the QoE [83]. Certain works have suggested using a wireless energy charger application for WSNs because it supports mobility [84,85]. A practical, real-time wireless recharging protocol for dynamic wireless recharging in sensor networks is proposed in [86]. The real-time recharging framework supports single or multiple mobile vehicles. Similarly, a collaborative wireless charging scheme has been presented in [87]. Recharging selected nodes in remote areas by using unmanned aerial vehicles can address the issue of recharging a single node. Johnson et al. [88] simulated a limited recharging system and verified the effectiveness of recharging a single node. However, their work does not describe a robust system model and is limited to simulations. A prototype-based implementation for similar EH architectures is provided in [89]. Further research must be conducted to select a proper energy storage mechanism, which is directly related to efficiency and long operational lifetime of sensors.

Energy-Efficient Routing and WSN Architecture
Generally, designing of single-path routing protocol is easier than a multipath routing protocol. The drawback of a single-path protocol is that it swiftly drenches the energy when selected as the path. Additionally, in scenarios where a single-path protocol node is out of energy, a fresh route must be recomputed. Meanwhile, multipath routing creates a platform to equally re-distribute the energy among the sensor nodes by rotating the forwarding nodes. These have the capacity to increase network reliability by provisioning multiple routes, speeding up network recovering rate from a failure. For readers interested in the multipath routing protocols for WSN, a comprehensive survey is given in [40]. In terms of the energy efficiency of the multipath routing protocols for WSN, the Energy-Efficient Multipath Routing Protocol (EEMRP), discussed in [41], focuses on discovering multiple node-disjointed paths based on a cost function driven by the energy levels and hop distances of the nodes, and subsequently allocates the traffic rate to each selected path. Moreover, Energy-Efficient and Collision Aware (EECA), discussed in [90], is proposed as a dual node disjointed and collision-free route considering source and sink. The results showed that the efficiency of the multipath routing protocols in terms of the energy are better than single-path routing protocols. Moreover, it is capable of further improving both the EE as well as lifetime of the WSN, if the routing algorithms are not only the function of the shortest paths, but consider the residual energy before selecting the next hop, as reported in [91]. Liu et al. [91] proposed dual novel energy-aware cost functions to improve the energy-balancing performance of the routing protocol by considering nodes in hotspots consumes more energy: (i) Exponential and Sine Cost Function-based Route (ESCFR) function, maps a miniscule variation in remaining nodal energy to a big variation in the cost function value. The idea of the ESCFR, operates by giving higher preference to sensor nodes having bigger remaining energy during route selection, thus creating energy equilibrium. (ii) The Double Cost Function-based Route (DCFR) protocol makes decisions by taking into consideration the energy consumption rate of nodes, as well as the residual energy, which enhances the energy-balancing performance of the routing protocol, even in networks facing obstructions. Unfortunately, the location of the sensor nodes may deplete energy in a given region or create energy holes. However, optimal sensor node placement via uniform distribution or by including a few sensor relay nodes with enhanced capabilities can be deployed to address the issue. Generally, this leads to energy balance improvement among the sensor nodes, avoiding hot-spots sensor nodes and guarantee RF coverage and link connectivity [92]. Plethora of research have focussed on locating the least number of sensor relay nodes or optimal sensor relay placement that will extend the network lifetime [42][43][44]. Meanwhile, other studies have proposed a cluster architecture approach, which organizes the sensor nodes into clusters. The motive of this approach is dependent on the cooperation among sensor nodes in the same cluster. Meanwile, each cluster is managed by a selected node known as the cluster head, which is responsible for coordinating the members' activities and communicating with other cluster heads or the base station [38,39]. Cluster architectures is one of the most desirable approaches to improving the EE of the WSN. Cluster architecture comes with many benefits, such as: improvement in WSN energy-efficiency and network scalability by maintaining a hierarchy in the network. To fully derive these benefits, these strategies must be considered: i.
Reduction of transmitting distance of cluster members requiring lower transmission power. ii.
Cluster heads limiting the transmissions frequency as a result of fusion. iii.
Mandating the cluster head to perform all the energy-sapping functions, such as coordination and aggregation. iv.
Permit to power-off some cluster members while the cluster head assumes the forwarding roles. v.
Alternate the choice of cluster head among the nodes so as energy consumption in the network.

Aggregation and Reduction of the Data
Obviously, data transmission and processing are not cheap with reference to energy consumption. Therefore, efficiency in handling data delivered to the sink nodes leads to energy saving. Reducing and aggregating the data quantity being delivered to the sink nodes are considered efficient solutions to increase EE during the transmission process [17]. In data aggregation schemes [45], nodes are permitted to only re-transmit the average or the lesser of the received information. However, information aggregation may lead to latency reduction since traffic is reduced, thereby reducing network delays. The drawback of this approach is that it may impact negatively on the accuracy of the data collected. If an optimal aggregation function is not deployed, it may become difficult to recover the original data sent to the sink [46]. Therefore, it is not recommended to use this technique with applications that need high accuracy; but adaptive sampling techniques are used when the criteria are formed in terms of coverage or information precision. In adaptive sampling approaches, the sampling rate are adjusted at each sensor and at the same time, making sure that the application requirements are achieved referenced to range or data precision. For instance, in a supervision task, low-power acoustic sensors can be deployed to notice an imposition. In a scenario where an event is stated, power-hungry cameras can be instructed to gather better-grained information [29]. Three-dimensional correlation is a good candidate for reducing the sampling rate in areas experiencing low variation in sensed data. In human activity recognition applications, Yan et al. [93] suggested that that sampling acquisition be based on core user activity, rather than taking samples in all unnecessary instances such as sitting, jumping, biking or running. Conversely, network coding can be deployed to reduce the overall data traffic in broadcast environment by transmitting a linear aggregation of several packets rather than a copy of each packet. Between computation and communications, communications utilize lesser energy because computations are generally regarded as power hungry application as a result, network coding exploits this gain. Wang et al. [47] fuse network coding and connected dominating sets to additionally decrease energy utilization in broadcast events. AdapCode [94] is an information broadcasting protocol designed by allowing a node to broadcast N messages received at several other nodes, resulting in energy conservation. The resultant energy savings from the bandwidth is (N − 1)/N compared to naive flooding. The receiver node can recover the original packets by Gaussian elimination after receiving N coded packets successfully. Moreover, AdapCode enhances reliability by adjusting N to the sensor node numbers, because when N rises and the number declines, the packets recovery rate for data decoding decreases. Reliability can be further improved by permitting more sensor nodes to acquire less than N packets and send a negative acknowledgement to recover loss information.

Green Radio-Frequency Identification
The RFID system consists of RFID tags with a unique identifier electronic product code (EPC), an RFID reader, and middleware [95]. For an object to be tracked by an RFID enabled system, RFID tags must be appended to the target object. The design nomenclature of RFID tags consists of a small microchip linked to an antenna. As with other wireless devices, the antenna is responsible for transmitting and receiving of the radio signals. It is not necessary compulsory that RFID tags must be direct line of sight with the RFID reader. RFID tags can be read utilizing non-line of sight technology. Boasted by the memory capacity, the EPC and other valuable data can be read and traced using RFID readers effortlessly. An RFID reader can be considered as the base station (BS) or access point of the system and it is responsible for energizing, sending data and commands to connect RFID tags attached on a fixed or mobile object [96]. Meanwhile, RFID systems is characterized as having very low (i.e., a few meters) RF coverage. Some of the notable spectrum bandwidths of interest are very low frequencies (VLF) at 124-135 kHz up to ultrahigh frequencies (UHF) at 860-960 MHz [16].
RFID tags can be classified into two types in terms of power source: (i) Active RFID tags have a local power source (battery) and operate hundreds of meters from the RFID reader. (ii) Passive RFID tags are without batteries and collect energy from a nearby RFID reader's interrogating radio waves by means of the principle of induction. Thus, the EE is a key requirement for the wider acceptance of the active RFID systems that utilize battery-constrained tags. Additionally, active RFID is less advantageous than passive RFID in terms of its tag cost and size, but more advantageous in terms of sensing distance, sensing rate, and stability [16,97]. Semi-passive tags are a third variant of RFID, and they are equipped with batteries; however, these batteries are not an integral part of RF transmission. Hence, they are not considered active RFID tags.

Passive RFID Systems
Passive RFID tags lack an inbuilt energy source; they are powered by the electromagnetic energy emitted from an RFID reader instead. Passive RFID tags are popular in several applications, such as access control, file tracking, race timing, supply chain management and smart labels. The cheap price of RFID tags allows their wide deployment across a wide spectrum of the industry. RFIDs have already been promoted as the most cost-effective solution with satisfactory achievements in smart parking systems [98]. RFID in smart parking systems have mainly been used for check-ins and check-outs of passing vehicles in parking areas; such vehicles are equipped with e-pass cards with active RFID tags [99]. Each vehicle that enters a parking area has a unique identification number that can be paired with its license plate. The barriers only open if the vehicle is recognised as being registered. Moreover, the authors in [100] combined RFID technology with wireless sensors to collect information on the occupancy state of parking spaces and direct drivers to the nearest vacant parking lot. Other for vehicle validation at the entrance and exit of a parking area, RFID technology has been utilised for parked vehicle localisation by exploiting the time difference of arrival and the received signal strength measurements [101]. Furthermore, the authors in [102] presented an analysis and evaluation of a passive RFID-based framework in the context of smart parking systems. The main novelty of the proposed framework is the adoption of RFID tag-to-tag communication, which can support a more energy-efficient collection of information than the conventional direct type of communication. The overall framework is considered and examined within a smart parking system use case scenario, in which RFID technology is adopted for: (a) vehicle validation when entering and exiting the parking area; (b) identification and localisation of parked vehicles; and (c) smart parking space management. The superiority of the proposed framework over conventional direct RFID reader tag communication is demonstrated in terms of: (i) the reduction of RFID readers' transmission power to the required minimum to guarantee connectivity, and (ii) the expansion of RFID readers' coverage areas to communicate with many distant tags. A summary of the passive RFID system is given in Figure 3.
Electromagnetic energy transfer techniques can be classified into three main techniques: (i) RF energy harvesting, (ii) inductive coupling, and (iii) magnetic resonance coupling (Table 5). Passive RFID system energy transfer techniques are NFC, either inductive coupling or backscatter. The RFID reader emits a sinusoidal signal. The tag antenna is tuned to receive the signal from the reader. The internal IC of the passive tag contains a rectifier circuit that converts the power into DC, enabling the tag circuitry to work. The circuitry modulates the signal to an extent and then returns it to the reader. During this process, the tag does not create a separate signal; it merely modulates the signal received from the reader. An explanation of the backscattering principle is given in detail in [96].
The Passive RFID tag feature is low power consumption, making it a suitable candidate in wireless sensing applications. However, operating coverage distance is a notable challenge. In best case scenario, the maximum RF coverage distance of passive tags are up to 7-15 m. The reason for this is that the system is powered using electromagnetic induction, which is itself relatively weak. Additionally, path loss, which is considered to be one of the most important parameters in any wireless communication, must be included. Accordingly, path loss is a crucial design parameter in RFID transmission. Path loss is due to many effects, such as free-space loss, refraction, scattering and diffraction, reflection, the height of antenna, the surrounding environment and weather (dry or moist air), the distance between the transmitter and the receiver, the height and location of antennas [95]. Thus, choice of calculation path loss model is a vital decision. In [103], the author discusses and compares more accurate path loss predicting models extensively, such as Friis's model, CCIR model, Hata model, etc., for different environments. However, these models have application-specific device operating frequencies. For more of clarity, we discuss, in the following, the relationship among distance, operating frequency, and path loss based on the most rudimentary mathematical model to calculate path loss, Free Space Path Loss (FSPL) model that can be written as follows: where f and c denote the operating frequency (in Hz) and speed of light (in metres/second), respectively.
Also, f c is called the signal wavelength (λ). The term d refers to the distance between the RFID reader and RFID tag antennas. λ and d are in the same unit of length (in metres). If d >> λ, it is assumed that both antennas are in the far field of each other. Equation (3) shows that the operating frequency is directly proportional to the square root of the path loss considering all other variables constant. In addition, the distance between the antennas is also directly proportional to the square root of the path loss considering all other variables constant including the operating frequency, which means that the received signal power at RFID tag rapidly decreases with increasing transmit-receive distance, which also results in an increase in path losses.
The wake-up signals deployed in duty cycle leads to energy inefficiency as lots of energy are utilized. Low energy powered radios can be deployed to wake up a sensor node only when need arises, such as sending or receiving packets. Meanwhile, power-consuming transceivers are deployed for information transmission. Ba et al. [104] suggests a network consisting of non-active RFID wake-up transceiver called WISP-Motes and RFID readers. A non-active RFID wake-up transceiver utilizes the energy derived from the reader transmitter to initialize an interruption that awakens the node. In reality, it is impossible to equip all the sensor nodes with RFID readers because of its power-hungry capability. The aforementioned issue and the limited operational range have restricted passive RFID from being utilized as only as a single-hop scenario. Software simulations have indicated that WISP-Motes can conserve a greater quantity of energy at the cost of more hardware and magnified latency in information delivery. The authors showcase their advantages in the scenario of lightly delay-tolerant system mobile elements accompanied with RFID readers. Moreover, the authors in [100] combined RFID technology with wireless sensors to collect information on the occupancy state of parking spaces and direct drivers to the nearest vacant parking lot. Other for vehicle validation at the entrance and exit of a parking area, RFID technology has been utilised for parked vehicle localisation by exploiting the time difference of arrival and the received signal strength measurements [101]. Furthermore, the authors in [102] presented an analysis and evaluation of a passive RFID-based framework in the context of smart parking systems. The main novelty of the proposed framework is the adoption of RFID tag-to-tag communication, which can support a more energy-efficient collection of information than the conventional direct type of communication. The overall framework is considered and examined within a smart parking system use case scenario, in which RFID technology is adopted for: (a) vehicle validation when entering and exiting the parking area; (b) identification and localisation of parked vehicles; and (c) smart parking space management. The superiority of the proposed framework over conventional direct RFID reader tag communication is demonstrated in terms of: (i) the reduction of RFID readers' transmission power to the required minimum to guarantee connectivity, and (ii) the expansion of RFID readers' coverage areas to communicate with many distant tags. A summary of the passive RFID system is given in Figure 3. Electromagnetic energy transfer techniques can be classified into three main techniques: (i) RF energy harvesting, (ii) inductive coupling, and (iii) magnetic resonance coupling (Table 5). Passive RFID system energy transfer techniques are NFC, either inductive coupling or backscatter. The RFID reader emits a sinusoidal signal. The tag antenna is tuned to receive the signal from the reader. The internal IC of the passive tag contains a rectifier circuit that converts the power into DC, enabling the tag circuitry to work. The circuitry modulates the signal to an extent and then returns it to the reader.

Active RFID System
A battery-assisted RFID tag that uninterruptedly broadcasts its signal is known as Active RFID. Active RFID tags are usually deployed as "beacons", which precisely trail the real-time position of targets or in high-speed settings such as tolling. Active tags which are battery powered have a greater reading range when compared to passive tags; however, they are also much costlier. A summary of the active RFID system is given in Figure 4. The most important issue when considering an active RFID system is the energy storage device. The batteries and capacitors are the two most common storage devices. A summary comparison for these two commonly storage devices are provided in Table 6. Table 6. Comparison of the two most common storage devices.

Issues
Battery Capacitor Advantages • High power density.
• High energy density.
• Easily replaceable in most cases.
• High power density.
• Does not loose ability to retain power with time.

Challenges
• Poor to "form factor".
• Rechargeable form loses ability to retain power with time. • Have to be replaced periodically, and difficult to replace in larger numbers and some applications in remote areas.
• Low energy density.
• Susceptible to damage with current fluctuations. • Capacity is highly dependent on size.
• Harder to replace compared to batteries. Improvements • Introduction of thin film technology has removed the issues related to form factor. • There are a range of new technologies which could have been tested, from lithium-ion varieties to redox-flow batteries.
• Introduction of ultra-capacitors have enables the best of both components with high power and energy densities. • Size of ultra-caps are also significantly smaller than regular capacitors.
Batteries have been the most notable energy sources for countless devices since their invention, and are still the most used portable power source in the world [105]. However, the problem with batteries is that they have a limited lifespan. After this certain period, they have to be replaced. Even with rechargeable batteries, there is a certain period after which the energy retaining capabilities of the battery diminish considerably. Additionally, form factor morphology is another key parameter that is considered when choosing the right storage device for a particular use-case. However, the The most important issue when considering an active RFID system is the energy storage device. The batteries and capacitors are the two most common storage devices. A summary comparison for these two commonly storage devices are provided in Table 6. Batteries have been the most notable energy sources for countless devices since their invention, and are still the most used portable power source in the world [105]. However, the problem with batteries is that they have a limited lifespan. After this certain period, they have to be replaced. Even with rechargeable batteries, there is a certain period after which the energy retaining capabilities of the battery diminish considerably. Additionally, form factor morphology is another key parameter that is considered when choosing the right storage device for a particular use-case. However, the relationship between these two characteristics (capacity and form factor) is most often conflicting. As the size of the storage device rises, normally, the device size will become larger, and vice versa [106]. Taking the capacity and form factor issue into consideration, batteries are slowly becoming a less viable option for RFID environments. Consequently, a compromise is needed to maintain desirable battery design in terms of both practicality and size, in order to retain batteries as options as storage devices for use in RFID environments. Extensive research has been carried out by academic institutes and the industrial sector on how to make energy storage devices as small and flexible as possible while having ample capacity and lasting usability in an RFID tag environment. With the introduction of thin film technology, batteries are starting to adapt to form factors appropriate for usage in wireless sensors. The end-goal of this process is a system which will enable electronic devices fabricated in a paper-thin width range. Carmo et al. [107] describes a thin film battery scheme in conjunction with suitable option for thermo-electric micro-systems. They recommend a strategy for fabricating thin film solid-state rechargeable batteries. This affords a deeper analysis and evidence in support of the notion that it is ideal for applications involving settings where a thermal difference is evidently accessible, e.g., human body. The popularity of the thin film approach is again supported in [108]. The aforementioned paper provides a design of rechargeable battery in an RFID tag environment in great detail.
Capacitors are very effective components for storing energy. They are equipped with an inherent ability to hold energy instantaneously, thus making them the preferred device for energy storage purposes. Meanwhile, there is a direct proportionality between capacitor size and energy storing capacity; the capacity to hold energy greatly depends on the size of the capacitor. Evidently, bigger sized capacitors store more current than smaller sized devices. Capacitors are also prone to damage in the face of high current and voltage fluctuations, making them dependent on the deployed environment. There is an evolution towards the development of super-capacitors and ultra-capacitors that have the capacity to store greater amounts of current. When ultra-capacitors are integrated into rechargeable batteries, they provide both longevity and solidity to the rechargeable batteries. Ultra-capacitors can store current in many times the amount of conventional capacitors, and as a result have become the design choice of many researchers and industry players. These ultra-capacitors are optimally appropriate in power conversion electronics circuits, as shown in [109]. Hybrid models have also emerged in which two models are utilized rather than a single model. Such models have found application and deployment in solar-powered WSN nodes [110]. In [111], the authors proposed a model, named "Prometheus", involving the deployment of a hybrid method. In this configuration, a dual phase storage scheme is discussed, consisting of a super-capacitor in one phase and a lithium rechargeable battery in the second phase. This scheme is deployed as a safeguard for running a Berkley's Telos Mote from a PV solar panel system. The experimental results are encouraging, suggesting that this system can run for 43 years in an application having 1% load, and can last for 1 year with a 100% load value.
Powering the electronics embedded into a smart garment is essential for e-textiles. Regarding energy storage devices, batteries are accepted as one of the most important and efficient ways to establish electricity networks. However, other technologies can be used, such as super-capacitors, which can now achieve energy density levels that are comparable to lead-acid batteries. In addition, super-capacitors are environmentally friendly and offer high power density, fast charging/discharging speed, and a long lifecycle [112]. The requirements of batteries change significantly depending on the wearable energy consumption. Three different types of devices can be distinguished, depending on their power needs [35,113,114]: • Low-end devices. These devices are similar to traditional watches or pacemakers that can be powered for a long time (years) by button type batteries, given their low energy consumption (usually under 100 µW).
• Mid-range devices. These devices consume an average of 500 mW, considering their use of wireless communications transceivers. They usually last less than a day (often just several hours) when transmitting continuously, although certain technologies make use of sleep modes or periodic transmissions (e.g., BLE beacons) to last long. This kind of device requires bulkier batteries than low-end devices (e.g., AA or AAA batteries), making them less appropriate to be embedded into smart clothing.

•
High-end devices. These are devices similar to smartphones or laptops that consume up to 50 W. They usually utilise Li-ion batteries, which can be bulky and add weight to a garment. Table 7 presents certain examples of wearable devices with their power consumption and the basic characteristics of the battery that they usually carry. The operating period of a battery is determined by its power requirements and energy density. Energy density is the amount of electrical energy stored in a battery in a given weight (Wh/kg). Moreover, the selection of energy storage devices depends on the type of applications. Recent research has revealed that hybrid models that use both components can achieve good results, considering their own capabilities. On the other hand, references [6,10,115,116] considered the transmission power issue. The sleep mode technique, when the reader is not interrogating, to achieve energy saving was proposed by [10]. The authors in [6] proposed an algorithm and protocol to avoid overheating when the reader is not interrogating, in order to achieve energy saving. Meanwhile, the authors in [115,116] proposed an algorithm and protocol to avoid tag collision during the transmission. Meanwhile, reference [22,48] proposed adjusting the transmission power level dynamically, which could achieve energy saving, but less than a sleep mode technique. Recall that the major driver of green technology is ecological and energy fears. Reduction in RFID tags size should be exploited, as there is a direct correlation between quantity of non-degradable substance deployed in their engineering (e.g., biodegradable RFID tags, printable RFID tags, paper-based RFID tags), since the RFID tags themselves are generally hard to reuse [97].
Security and privacy are major concerns for IoT deployment. Implementing security algorithms requires a substantial device processing. Along with security, we must investigate appropriate mechanisms that consider energy consumption and the required QoS. The potentials of energy-efficient and secure mechanisms remain in their infancy, thus encouraging extensive research and development in this area. IoT involves resource-constrained devices, such as RFID, sensor nodes and high-end data servers. Therefore, finding and exploiting trade-offs is important to provide security among heterogeneous devices in the green IoT paradigm. Tags are typically energy-constrained, thus limiting the implementation of complex security algorithms. Energy-depleted RFID tags are a source of potential attacks from intruders with malicious intent. Moreover, modifying the data in RFID tags is also a security threat. Therefore, a natural solution is to password-protect the memory in RFID tags. Doing so is a challenge from an energy consumption perspective, because cryptographic algorithms require a large amount of energy and battery (active tag). In this regard, several methods with low energy requirements have been recently proposed for these cryptographic solutions [117,118]. The authors in [119] developed a communication strategy where a tag periodically communicates its security status to the reader and is authenticated. The information exchange between tags and the reader is secured using AES encryption. In addition, they used a PUF and a power sensing circuit to identify malicious readers and tags. The concept of semi-active tags, in which an energy-depleted RFID tag can use neighbouring tags, is also introduced to communicate their security status and be authenticated by the reader. The authors in [120,121] presented a technical survey research on privacy and security problems for RFID with the consideration of the energy case.

Conclusions
Green technologies play an important role in enabling energy-efficient IoT. This study has provided an overview of popular research topics on green WSNs and RFID, which are considered to be the two main pillars of IoT applications and cover recent industry developments. We initially provided a taxonomy of the energy-saving techniques for green WSNs and RFID. A detailed discussion of each technique was subsequently presented to pursue a vision of green WSNs and RFID. By considering energy consumption and the required QoS, we suggest conducting further research to select a proper energy storage mechanism that is directly related to efficiency and long operational lifetime.