Next Article in Journal
Selected Chemical Parameters of Cereal Grain Influencing the Development of Rhyzopertha dominica F.
Previous Article in Journal
A Tale of Two Cities: How Can We Use Social Media to Engage Young People with an Environmentally Friendly Lifestyle in Hong Kong and Guam?
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimal Electric Vehicle Battery Management Using Q-learning for Sustainability

by
Pannee Suanpang
1,* and
Pitchaya Jamjuntr
2
1
Department of Information Technology, Faculty of Science & Technology, Suan Dusit University, Bangkok 10300, Thailand
2
Department of Electronic and Telecommunication, Faculty of Engineering, King Mongkut’s University of Technology Thonburi, Bangkok 10140, Thailand
*
Author to whom correspondence should be addressed.
Sustainability 2024, 16(16), 7180; https://doi.org/10.3390/su16167180
Submission received: 10 July 2024 / Revised: 10 August 2024 / Accepted: 19 August 2024 / Published: 21 August 2024

Abstract

:
This paper presents a comprehensive study on the optimization of electric vehicle (EV) battery management using Q-learning, a powerful reinforcement learning technique. As the demand for electric vehicles continues to grow, there is an increasing need for efficient battery-management strategies to extend battery life, enhance performance, and minimize operating costs. The primary objective of this research is to develop and assess a Q-learning-based approach to address the intricate challenges associated with EV battery management. This paper starts by elucidating the key challenges inherent in EV battery management and discusses the potential advantages of incorporating Q-learning into the optimization process. Leveraging Q-learning’s capacity to make dynamic decisions based on past experiences, we introduce a framework that considers state-of-charge, state-of-health, charging infrastructure, and driving patterns as critical state variables. The methodology is detailed, encompassing the selection of state, action, reward, and policy, with the training process informed by real-world data. Our experimental results underscore the efficacy of the Q-learning approach in optimizing battery management. Through the utilization of Q-learning, we achieve substantial enhancements in battery performance, energy efficiency, and overall EV sustainability. A comparative analysis with traditional battery-management strategies is presented to highlight the superior performance of our approach. A comparative analysis with traditional battery-management strategies is presented to highlight the superior performance of our approach, demonstrating compelling results. Our Q-learning-based method achieves a significant 15% improvement in energy efficiency compared to conventional methods, translating into substantial savings in operational costs and reduced environmental impact. Moreover, we observe a remarkable 20% increase in battery lifespan, showcasing the effectiveness of our approach in enhancing long-term sustainability and user satisfaction. This paper significantly enriches the body of knowledge on EV battery management by introducing an innovative, data-driven approach. It provides a comprehensive comparative analysis and applies novel methodologies for practical implementation. The implications of this research extend beyond the academic sphere to practical applications, fostering the broader adoption of electric vehicles and contributing to a reduction in environmental impact while enhancing user satisfaction.

1. Introduction

In the epoch of digital transformation, profound technological advancements, particularly in the realm of artificial intelligence (AI), have engendered disruptions across diverse industries on a global scale. The pervasive adoption of AI has precipitated substantial ramifications across all domains, propelled by the utilization of sophisticated information technologies. Concurrently, there has been a notable surge of interest in electric vehicles (EVs), spurred by numerous nations as a sustainable mode of transportation [1,2]. EVs are lauded for their environmental compatibility, which is attributed to their emission-free operation. Moreover, they present economic advantages over conventional gasoline-powered engines, coupled with user-friendly interfaces and seamless controls [1,3,4,5]. Due to their zero emissions, electric vehicles are considered environmentally benign. They are also less expensive to operate than conventional gasoline engines and have smooth operation controls [6,7]. Furthermore, it is estimated that over 951.9 billion EVs will be in use worldwide by 2030 [8]. Nevertheless, a high infiltration by EVs is a significant issue that affects the electricity-distribution system, leading to problems such as power quality degradation, increased line damming, distribution transformer failure, increased distortion, and a higher fault current [7,8].
The integration of EVs into smart cities offers a promising avenue for sustainable transportation and reduced environmental impact [1,2]. As the demand for EVs within smart tourism destinations increases, the need for efficient and intelligent battery-management strategies becomes paramount. Ineffective battery management can lead to shortened battery lifespan, diminished user experience, and strain on the power grid [1,2].
The performance, lifespan, and overall sustainability of EVs are intricately linked to the optimal management of their energy-storage systems, primarily lithium-ion batteries. However, traditional battery-management strategies face significant challenges in adapting to the dynamic nature of driving conditions, charging infrastructure, and the diverse patterns of energy consumption [9,10]. As battery technology continues to progress, the significance of thermal-management systems intensifies in maintaining peak performance and safety [9,10]. A proficient thermal-management system not only enhances battery efficiency but also contributes to overall energy conservation, maximizing range and prolonging battery longevity, thus fostering sustainability within the realm of transportation [11,12,13].

1.1. Problem Statement

The widespread integration of EVs hinges significantly on the optimization of their functionality, effectiveness, and resilience. Within the integral components of an EV, the battery pack assumes a critical role in ensuring the vehicle’s reliable operation and prolonged lifespan. The following key problems in EV battery management motivate the research, including:
  • Dynamic adaptation to driving patterns: EVs encounter a wide range of driving conditions, from city traffic to highway speeds. Traditional battery-management systems may struggle to dynamically adjust to these diverse patterns, resulting in suboptimal battery performance and efficiency [11].
  • Optimal state-of-charge management: maintaining an optimal state-of-charge is pivotal for extending battery life and ensuring consistent performance. Existing approaches may not effectively balance the energy demands of driving with charging and discharging cycles, leading to challenges in achieving and maintaining an ideal state of charge [1,2].
  • Integration of charging infrastructure: The availability and utilization of charging infrastructure significantly influence the charging patterns and overall health of EV batteries. Current battery-management systems may not fully exploit the potential benefits offered by existing infrastructure, limiting the optimization of charging and discharging cycles [1,2].
  • Real-time decision making: The advent of machine learning and reinforcement learning opens avenues for real-time decision making based on past experiences. Integrating these advanced techniques into EV battery-management systems has the potential to address challenges posed by dynamic driving conditions and diverse charging scenarios [1,2,13].
The research problem at the core of this study revolves around the need to develop advanced battery-management strategies that maximize the benefits of EVs for both manufacturers and consumers. This encompasses the challenge of optimizing battery performance, extending battery life, and minimizing operational costs. Achieving these objectives while considering the dynamic nature of EV usage patterns, charging infrastructure, and the degradation of battery health is a complex and multifaceted problem [1,14].
Several studies have explored the use of reinforcement learning (RL) techniques, including Q-learning, in optimizing battery management for electric vehicles [13,14]. These approaches leverage the ability of reinforcement learning to adapt to changing conditions and make dynamic decisions.
  • State-of-Charge and State-of-Health Management: the literature in this domain often emphasizes the importance of managing both the state of charge (SoC) and state of health (SoH) of EV batteries. Q-learning frameworks are discussed as a potential solution for balancing the trade-off between maximizing energy usage and minimizing degradation [15,16].
  • Dynamic Decision Making in Charging Infrastructure: charging infrastructure plays a crucial role in EV battery management [17]. Some studies investigate how Q-learning algorithms can optimize charging decisions, considering factors such as charging station availability, charging rates, and their impact on battery performance.
  • Real-world Data Integration: successful implementation of Q-learning in battery management often involves the use of real-world data [18]. Researchers discuss the challenges and benefits of incorporating actual driving patterns, environmental conditions, and battery characteristics into the learning process to enhance the model’s accuracy.
  • Comparative Analysis with Traditional Approaches: researchers have conducted comparative analyses between Q-learning-based battery management and traditional methods [19]. These studies typically highlight the potential improvements in terms of energy efficiency, battery life extension, and overall performance achieved through Q-learning.
  • Machine Learning and Battery Management Optimization: beyond Q-learning, there is a broader literature on the application of various machine learning techniques for optimizing EV battery management. Researchers explore different algorithms, including neural networks and other reinforcement learning methods, to address the complexities of battery systems [19].
  • Implications for Sustainability and Environmental Impact: some works discuss the broader implications of Q-learning-based battery management for the sustainability of EVs [18,19]. This includes the considerations of a reduced environmental impact through enhanced battery efficiency, potentially contributing to the overall sustainability of electric transportation.
The significance of using Q-learning in this context is rooted in its ability to make intelligent, data-driven decisions, improving battery performance, energy efficiency, and user satisfaction [1]. By allowing EVs to learn and adapt their battery-management strategies, we can unlock the potential for prolonged battery life, reduced environmental impact, and greater overall value for EV owners. Furthermore, the integration of Q-learning can lead to more cost-effective and sustainable EV operation, which is crucial in the competitive automotive market.
This paper highlights the novelty of the Q-learning approach and its practical relevance in improving vehicle efficiency and battery longevity and suggests future research into more complex driving scenarios and real-world testing to validate the findings [20]. This paper introduces a method to integrate electric vehicles and stationary battery storage into optimization problems using exclusively linear relationships and validates its effectiveness with real-world data. It examines multiple aspects of four different charging strategies and conducts sensitivity analyses [21].

1.2. Research Gap

Despite the advancements in battery management and Q-learning, several limitations and gaps in existing research persist. Many studies have primarily focused on theoretical models and simulations, with limited real-world application [22]. Practical implementation, scalability, and validation of Q-learning strategies in large-scale EV fleets are areas where further research is required [23,24]. Additionally, existing studies often lack comprehensive comparisons with traditional battery-management methods, hindering the assessment of Q-learning’s effectiveness.
The significant problem of battery management in EV cars in Thailand primarily revolves around infrastructure challenges for charging stations and the sustainable disposal of spent batteries. As the adoption of EVs grows in Thailand, the demand for reliable and accessible charging infrastructure has become crucial [1,2]. The lack of an extensive charging network across the country hampers the widespread use of EVs and creates range anxiety among potential buyers. Moreover, the proper disposal and recycling of lithium-ion batteries used in EVs poses environmental concerns. Thailand, like many other countries, faces the challenge of managing end-of-life batteries effectively to minimize their impact on the environment and human health [1,2,25].
These issues are highlighted in various reports and articles. For instance, a study by the International Energy Agency (IEA) in 2021 emphasized the necessity for Thailand to invest in charging infrastructure and develop policies to manage the lifecycle of EV batteries sustainably [26]. Addressing these challenges demands collaborative efforts between the government, automakers, and other stakeholders to establish an efficient charging network and implement robust policies for battery recycling and disposal.

1.3. Objective of This Paper

This study aims to address this problem by leveraging the power of reinforcement learning, specifically Q-learning, to design a novel approach to EV battery management. Q-learning, a well-established reinforcement learning algorithm, has shown significant promise in optimizing dynamic decision-making processes by learning from past experiences. By applying Q-learning, we can develop a battery-management strategy that adapts to the ever-changing conditions and demands of EV operation.

1.4. Contribution of This Paper

In presenting the ‘contribution’ of this study, we highlight the integration and advancement of Q-learning within EV battery management by building upon previous research in this area. Our work is distinguished by its novel application and critical comparison with existing literature, which we elucidate as follows:
(1)
Advancement of knowledge: This research bridges the gap between theoretical frameworks and practical implementation by deploying a Q-learning-based battery-management strategy in real-world EVs [2,27].
(2)
Comparative analysis: this analysis offers a broader perspective on both the benefits and potential drawbacks of applying Q-learning within the EV domain, contributing to a deeper understanding of its viability and effectiveness [2].
(3)
The practical implications of this research for EV battery management and sustainable transportation are significant and multifaceted. By employing Q-learning for battery management, we introduce an advanced method that not only optimizes the performance and lifespan of EV batteries but also contributes to broader sustainable transportation goals [1,2,3,4]. This research on EV battery management using Q-learning presents significant practical implications for sustainable transportation. In summary, this research offers practical benefits that extend from individual EV users to industry stakeholders and policymakers, fostering a more sustainable and efficient transportation system.
By distinguishing our research through these facets, we contribute substantively to the literature on battery management for electric vehicles, providing both theoretical insights and practical solutions that could facilitate the broader adoption and optimization of EV technologies. By doing so, this study provides a practical and realistic perspective on the application of Q-learning to EV battery management. It sheds light on the potential benefits and challenges, ultimately contributing to the broader adoption of advanced battery-management strategies in the growing electric vehicle market.
In the sections that follow, we will delve into the specifics of our methodology, data collection, experimental results, and discussions, all of which will showcase the effectiveness and potential of Q-learning for EV battery management. This research contributes to the ongoing discourse on sustainable transportation solutions and highlights the role of cutting-edge machine learning techniques in enhancing the performance of electric vehicles.

2. Literature Review

Electric vehicle (EV) battery management and the application of reinforcement learning, particularly Q-learning, have been subjects of increasing interest in recent years. This section provides an overview of relevant studies and research in EV battery management and Q-learning, highlights the limitations and gaps in existing research, and explains how our paper contributes to the existing body of knowledge.

2.1. Electric Vehicle Trends

The global uptake of EVs has exhibited a consistent rise in recent years, propelled by the urgent need to mitigate greenhouse gas (GHG) emissions and lessen dependence on fossil fuels. This review of post-2018 literature on EVs encompasses an analysis of their environmental impact, policy framework, and consumer preferences. They accentuate the potential of EVs to markedly curtail GHG emissions, particularly when powered by renewable energy sources [1,28,29]. The trends in the development of EVs include the growing fleet of EVs, with many car manufacturers reporting that electric vehicles will account for half of the models produced after 2030 [30,31]. The demand for EVs is increasing due to their better road handling capacity, stability control, and torque top-up capability compared to fuel-based vehicles [32,33]. EVs also offer clean, silent, and pollution-free operation [34,35]. The market analysis of EVs shows trends in the traction motor type and drive train controller type. EVs have the potential to significantly reduce transportation-related emissions of greenhouse gases. The development of electric transport in European countries depends on factors such as price, infrastructure development, and government incentive programs. The interaction between EVs and the power system is based on the concept of a smart grid, which allows for efficient network infrastructure use and peak load shifting.
Electric vehicle trends in Thailand are influenced by various factors, such as government support, charging infrastructure, EV knowledge, vehicle prices, and battery capacities [36,37]. The adoption of EVs in Thailand requires extensive public policy frameworks and the development of the EV manufacturing industry [38]. Factors influencing EV manufacturing include government support, taxes, and subsidies, as well as battery costs [39]. On the other hand, factors influencing consumer purchasing decisions include individual judgment, EV performance, vehicle price, battery issues, driving range, and charging time [40]. The demand for EV charging stations is increasing in Thailand due to the growing number of electric vehicles and the need for clean air and reduced emissions. Video marketing plays a significant role in influencing millennials’ intention to purchase EVs in Thailand. Overall, the EV market in Thailand is driven by environmental concerns, government support, technological advancements, and consumer preferences.

2.2. Electric Vehicle Component

2.2.1. Electric Vehicle Car Component

Figure 1 provides a comprehensive overview of the dynamics and modeling equations for key components of an EVs, outlined as follows [2,41]:
  • Battery pack: this is the main energy-storage system in an EV, characterized by its state of charge (SOC) and state of health (SOH). The SOC indicates the current energy level, while the SOH shows the battery’s degradation over time. Models such as the Thevenin equivalent circuit and electrochemical models like the Doyle–Fuller–Newman are used to describe the battery’s voltage dynamics.
  • Electric motor/generator: this component converts electrical energy from the battery into mechanical energy to propel the vehicle. It is modeled through equations that depict its torque–speed characteristics, linking motor torque, speed, and electrical power input.
  • Final power transmission typically involves a single-speed transmission or direct drive in EVs; this system transmits power from the motor to the wheels. Its dynamics are simpler than those of internal combustion engine vehicles, as it lacks complex gear mechanisms.
  • Internal combustion engine (IC engine): absent in pure EVs but present in hybrid variants like HEVs or PHEVs, the IC engine serves as a range extender or additional power source. Its dynamics are described by equations covering fuel consumption, combustion efficiency, and emissions.
  • Fuel tank: this is nonexistent in pure EVs but included in hybrids to store gasoline. Its dynamics are governed by equations related to the fuel level, consumption rate, and refueling behaviors.
The dynamics of an electric vehicle (EV) involve the complex interplay of various components that convert electrical energy into mechanical propulsion. Key components include the following [2,41,42]:
  • Battery dynamics: The battery pack serves as the main energy source, managing energy storage and release. It powers the electric motor during acceleration and recharges during regenerative braking, with critical factors including the state of charge (SOC), state of health (SOH), and internal resistance.
  • Electric motor dynamics: The motor converts electrical to mechanical energy, with dynamics governed by torque–speed characteristics, efficiency, and response times. These are influenced by motor design and control algorithms.
  • Power electronics dynamics: Inverters and converters control the flow of electricity between the battery and motor, adjusting the voltage and current based on driving conditions. This involves managing switching behavior and thermal management to ensure efficiency.
  • Transmission dynamics: EVs range from simple single-speed systems to complex multi-speed or dual-motor setups, with dynamics involving gear shifts and torque distribution.
  • Vehicle dynamics: These encompass overall component interaction affecting acceleration, braking, handling, and stability. Influencing factors include weight distribution, tire characteristics, and suspension setup. Control systems like traction and stability control, along with regenerative braking, optimize these dynamics for enhanced safety and efficiency.

2.2.2. Electric Vehicle Car Modeling Equation

The modeling equations for the main components of an EV are shown in Figure 1 [2] as follows:
1. Battery Pack: Thevenin Equivalent Circuit Model:
V = V o c I R i n t
This equation represents the voltage output of the battery pack, which is modeled using Thevenin’s theorem. ( V ) is the terminal voltage, ( V o c ) is the open-circuit voltage, ( I ) is the current, and ( R i n t )   is the internal resistance.
State of charge (SOC) dynamics:
d S O C d t = 1 Q I
State of charge (SOC) dynamics: This equation describes how the state of charge of the battery pack changes over time. ( d S O C d t ) represents the rate of change of SOC, ( Q ) is the battery capacity, and ( I ) is the current. State of health (SOH) dynamics (simplified):
d S O H d t = k I 2
State of charge (SOC) dynamics: This equation describes how the state of charge of the battery pack changes over time. ( d S O H d t )   represents the rate of change of SOC, ( Q ) is the battery capacity, and ( I ) is the current.
2. Electric Motor: Motor Torque Equation:
T = k T I
Motor Torque Equation: This equation relates the motor torque ( T ) to the motor current ( k T I ) is the torque constant. Motor Speed Equation: This equation describes how the motor speed ( ω ) is related to the torque ( T ) and the inertia ( J ) of the system.
ω = T J
Motor Efficiency Equation: This equation calculates the efficiency ( η ) of the motor based on the input power ( P i n ) , torque ( T ) , and speed ( ω ) .
η = T ω P i n
3. Power Electronics: Inverter Efficiency: This equation computes the efficiency ( η i n v ) of the inverter based on the output power ( P o u t ) and the input power ( P i n ) .
η i n v = P o u t P i n
DC-DC Converter Efficiency: This equation determines the efficiency ( η d c d c ) of the DC-DC converter based on the output power ( P o u t   ) and the input power ( P i n ).
η d c d c = P o u t P i n
4. Transmission: This equation describes how the output angular velocity ( ω o u t p u t ˙ ) of the transmission is related to the input angular velocity ( ω i n p u t ) ˙ and the gear ratio ( I ).
ω o u t p u t ˙ = 1 I ω i n p u t ˙
Losses in Transmission (simplified): This equation calculates the power loss ( P l o s s ) in the transmission system based on the input torque ( T i n p u t ), output torque ( T o u t p u t ) , angular velocities, and transmission efficiency ( η t r a n s ).
P l o s s = 1 η t r a n s T i n p u t ω i n p u t T o u t p u t ω o u t p u t
These equations represent simplified models of the dynamics of each component. Actual modeling may involve more complex equations that account for factors such as temperature effects, nonlinear behavior, and control strategies. Additionally, the specific parameters   V o c ,     R i n t ,   Q ,     k ,     k T ,   J ,     P i n ,   P o u t p u t , I , ω ,   T i n p u t   ,   T o u t p u t , η i n v , η d c d c , η t r a n s , etc., would need to be determined based on the characteristics of the particular components used in the EV.

2.2.3. Electric Vehicle Batteries

One invention involves dividing the battery shell into multiple unit cells, with positive and negative plates stacked in a staggered manner and connected with busbars [41]. Another invention focuses on a low-voltage battery with a housing formed from two parts, containing rechargeable electrochemical cells and equipped with a desiccant and a two-way pressure valve for moisture prevention and pressure regulation [42]. A different approach utilizes an aluminum shell with soft bag lithium iron phosphate batteries, providing high efficiency, good performance, long service life, and low cost [1]. A power battery for electric vehicles integrates a battery pack and a battery-management system compactly within a box body, ensuring data acquisition precision and facilitating management and maintenance [1,43]. Lastly, an electric vehicle battery system includes a concentrated power module, a battery module, and an oil constant-temperature module, achieving accurate control, constant temperature, optimal battery performance, and stability [43].
Lithium-ion (Li-ion) batteries have become the preferred power source for electric vehicles (EVs) due to their high energy density, low self-discharge rate, and long cycle life [44]. Design trends in Li-ion batteries include increasing cell dimensions, with the longest cells reaching 500 mm (pouch) and almost 1000 mm (prismatic) in 2021, increasing differentiation between high-energy or low-cost cathode and anode materials, and increasing cell energy, equivalent to gaining about 100% (energy density) and 70% (specific energy) compared to the 2010 and 2021 averages [2]. Vehicles equipped with 800 V battery packs are being released to reduce wiring thickness, power and heat generation loss, and improve mileage and performance [45]. China has a complete lithium battery industry chain and faces challenges from domestic and foreign competitors in the market [46]. Battery diagnostic and prognostic technologies are being developed to inform EV owners about battery condition and performance over its lifetime [47].
Electric vehicle batteries face several challenges. One major challenge is the limited driving range and high initial costs, which hinder the broader market penetration of EVs [48]. Another challenge is the need for more efficient energy-storage solutions to power the driving motor of EVs [49]. The battery power density, longevity, adaptable electrochemical behavior, and temperature tolerance are important factors that need to be understood and improved [50]. Additionally, the development of battery technology, including advancements in battery capacity, performance, and cost reduction, is crucial for the widespread adoption of EVs [51]. Battery-management systems (BMS) are also essential in EVs, requiring voltage and current monitoring, charge and discharge estimation, protection, equalization, and thermal management [52]. Furthermore, the installation of sufficient charging facilities and addressing concerns related to interoperability and cybersecurity are important challenges in the EV battery sector. Overall, addressing these challenges will contribute to the affordability, convenience, and sustainability of EV batteries.

2.3. EV Battery-Management Studies

Numerous studies have addressed the importance of effective battery management in EVs. Battery-management systems (BMS) for electric vehicles have been the focus of several studies. Researchers are interested in EVs due to their potential to minimize greenhouse impacts, reduce pollution, and provide freedom from fossil fuels [53]. The aim of BMS is to improve the efficiency of EVs, and Very Large-Scale Integration (VLSI) plays a role in achieving this [54]. The integration of simulation-based design optimization and artificial intelligence/machine learning (AI/ML) in BMS has improved battery performance prediction and design efficiency [55]. Online BMS is critical for the safe and reliable operation of EVs, and algorithm development is a challenging area of research [56]. Model-based and non-model-based data-driven methods are suitable for developing algorithms and control for online BMS [57]. Overall, these studies highlight the importance of BMS in improving the performance, safety, and efficiency of EV batteries.

2.4. Q-learning in Battery Management

Q-learning is a reinforcement learning algorithm that has been applied to battery-management systems in various contexts. In the context of modular multilevel converters (MMC) for battery cells, Karnehm et al. implemented and evaluated a Q-learning algorithm for state-of-charge (SoC) balancing [58]. Ali et al. [59] proposed a dual-layer Q-learning strategy for real-time energy management of battery storage in microgrids, where the upper layer generates directive commands offline, and the lower layer refines these commands in real-time [59]. Ye et al. [60] introduced a digital twin-enhanced Q-learning energy management system for battery and ultracapacitor electric vehicles, which improved energy efficiency and reduced battery degradation [60]. Additionally, a study on residential houses equipped with solar panels and battery energy-storage systems developed an intelligent and real-time battery energy-storage control based on a Q-learning model, resulting in more efficient control and cost savings [61,62].
Figure 2 illustrates how machine learning, specifically reinforcement learning, can be applied to EV control systems. It shows the flow of information from real-world driving conditions (Environment) through a decision-making process (Agents) informed by both empirical data (Expert Knowledge) and computational objectives (Objective Function), leading to the execution of vehicle actions that are constantly refined based on feedback (Reward).

2.5. Fundamentals of Q-learning and Its Suitability for Battery Management

Q-learning is a reinforcement learning algorithm that has been used for battery management in various applications. It has been applied to the real-time energy management of battery storage in grid-connected microgrids [63]. Q-learning has also been used for state-of-charge (SoC) balancing of battery cells in a modular multilevel converter (MMC) system [64]. Additionally, Q-learning has been employed in the energy-management system for battery and ultracapacitor electric vehicles, where it has been enhanced using the digital twin methodology [60,65]. Furthermore, Q-learning has been utilized for intelligent and real-time battery energy storage control in residential houses equipped with solar photovoltaic panels and a battery energy-storage system [66]. These studies demonstrate the suitability of Q-learning for battery management in various domains, including microgrids, MMC systems, electric vehicles, and residential energy systems.
Q-learning, a reinforcement learning algorithm, has gained prominence across diverse fields due to its capacity for making optimal decisions in dynamic and uncertain environments. The fundamental principles of Q-learning render it highly suitable for battery management in electric vehicles [67]:
-
Learning from Experience: the algorithm adopts a trial-and-error approach, learning from past experiences to make informed decisions. In EV battery management, this means adapting and optimizing strategies based on historical battery performance [60,68].
-
Adaptability: Q-learning can dynamically adjust decisions in response to changing conditions, such as driving patterns, environmental factors, and charging infrastructure. This adaptability is pivotal for maximizing the benefits of EV batteries [1,68].
-
Policy Optimization: Q-learning seeks the optimal policy, a set of rules or actions maximizing cumulative rewards. In battery management, this translates to enhancing battery efficiency, prolonging its lifespan, and improving overall vehicle performance [1,60].
-
State–Action–Reward Framework: Q-learning operates on the state–action–reward framework, associating states (e.g., battery SoC and SoH) with actions (e.g., charging, discharging) that yield the highest rewards (e.g., improved energy efficiency). This framework aligns seamlessly with the dynamic nature of EV battery management [60,68].
In conclusion, Q-learning’s adaptability and capacity to learn from experience make it an invaluable tool for addressing the intricate challenges of EV battery management. Its application can maximize battery performance, extend battery life, and minimize operational costs, contributing significantly to the sustainable and efficient operation of electric vehicles [69,70].

2.6. Sustainability in Smart Tourism City Implementation

Sustainability in smart tourism cities is a growing area of research. The relationship between sustainability and smart tourism has not been clearly outlined in the literature [1,71]. However, several studies have highlighted the potential benefits of integrating sustainability and smart city initiatives in tourism destinations. One study found that increasing tourism in cities can drive the city to advance a green transition, contributing to the advancement of smart tourism and smart city debate [72]. Another study focused on successful smart initiatives in Amsterdam, Barcelona, Seoul, and Stockholm, and emphasized the importance of local authorities and community involvement in shaping sustainable and technologically advanced smart features [73]. Additionally, the application of smart tourism data-mining technology has been shown to improve foreign exchange income, increase employment in tourism, and drive the development of tourism-related industries [74]. Overall, the integration of sustainability and smart technologies in tourism cities has the potential to enhance the tourist experience, promote economic growth, and contribute to the overall sustainability of the destination [75,76,77,78].

3. Methodology

3.1. Research Design

The research methodology adopted a Q-table approach for EV battery management because of its simplicity and interpretability, essential for decision-making transparency in this sector. This method allows stakeholders to easily understand the mapping from states to actions, making model behaviors clear and verifiable. Although neural networks could capture more complex data patterns and potentially enhance performance, they add significant complexity and computational demands, which can be counterproductive in real-time applications like EV battery management. The Q-table approach provides a balance between performance and interpretability, ensuring the reinforcement learning framework remains comprehensible and practical for real-world deployment, supporting our goal of developing an accessible and effective solution for managing EV batteries [70,71,72,73,74,75].
(1)
Reward Function for EV Battery Management
We employ a weighted sum of different objectives, each contributing to the overall desirability of the agent’s actions in EV battery management. While the specific formulation may vary based on the particular objectives and constraints of the system, a generalized representation can be provided as follows:
[ R = w 1 E + w 2 H + w 3 P ]
where:
( R ) represents the total reward accumulated by the agent over a given time period.
( E ) denotes energy efficiency, quantified as the ratio of distance traveled to energy consumed.
(H) represents battery health, encompassing factors such as state of charge (SOC), state of health (SOH), and temperature.
( P ) signifies vehicle performance, including metrics like acceleration, speed, and smoothness of operation.
w 1 , w 2 , and w 3 denote the weights assigned to each objective, reflecting their relative importance in the overall reward computation.
The reward function is designed to incentivize actions that lead to optimal energy utilization, minimal battery degradation, and efficient vehicle operation under varying driving conditions. Each objective term is weighed based on its significance in achieving the desired outcomes of EV battery management. Additionally, penalty terms may be incorporated to discourage undesirable behaviors, such as excessive battery discharge or overheating. Overall, the reward function provides a guiding signal for the reinforcement learning agent, facilitating the learning of optimal decision-making strategies that align with the objectives of EV battery management.
(2)
The use of Q-learning
Q-learning presents several key advantages in the context of EV battery management:
1. Model-Free Learning: Q-learning is a model-free reinforcement learning algorithm, meaning it does not require explicit knowledge of the system dynamics. This is particularly beneficial for EV battery management, where the underlying physics can be intricate and nonlinear. Other machine learning techniques, such as model-based approaches, might struggle to accurately capture these dynamics, especially in dynamic real-world environments. Q-learning, on the other hand, can adapt well to these complexities without needing an explicit model [1,2,75,76,77,78].
2. Flexibility and Adaptability: Q-learning offers flexibility in handling nonlinear relationships and complex decision-making processes. In EV battery management, optimal control strategies can vary based on driving conditions, user preferences, and battery health. Q-learning’s flexibility allows it to adapt to changing circumstances without requiring extensive redesign or recalibration, which might be necessary for other techniques like supervised learning models [1,2].
3. Data-Driven Optimization: Q-learning facilitates data-driven optimization by learning optimal strategies directly from data. In EV battery management, historical data, such as EV telemetry data, are abundant. Q-learning can leverage these data to identify patterns and trends that lead to efficient battery-management strategies. This continuous-learning process allows Q-learning to optimize performance over time, making it well-suited for the evolving nature of EV battery management [76].
4. Interpretability and Simplicity: Choosing a Q-table approach for Q-learning offers distinct advantages, particularly for its simplicity and transparency. Q-tables provide an easily interpretable representation of Q-values for each state–action pair, which is essential in safety-critical systems like EV battery management. They also offer lower computational complexity due to their fixed size, enhancing training and inference efficiency. This makes Q-learning more practical for real-time applications compared to more complex models like neural networks, which require significant computational resources and are less interpretable.
5. Guaranteed Convergence: Q-learning, under certain conditions, is guaranteed to converge to the optimal policy. This provides a level of reliability and predictability in the learning process, which is crucial for applications like EV battery management, where safety and efficiency are paramount. While neural networks and other function approximation methods can handle larger and continuous state spaces, they often lack the same level of theoretical guarantees for convergence and stability.
Overall, Q-learning provides a balance between performance and interpretability, ensuring the reinforcement learning framework remains comprehensible and practical for real-world deployment. This balance is crucial for developing an accessible and effective solution for managing EV batteries, supporting our goal of improving energy efficiency, prolonging battery life, and ensuring vehicle reliability. By using Q-learning, we can create a robust and adaptive EV battery-management system that continuously improves based on real-world data and evolving conditions.
Q-learning is chosen for EV battery management for several reasons, but it is essential to consider alternative methods:
Q-learning presents several key advantages in the context of EV battery management. Firstly, as a model-free reinforcement learning algorithm, Q-learning operates without explicit knowledge of system dynamics. This is particularly beneficial in EV battery management, where the underlying physics of battery behavior can be intricate and nonlinear. Unlike model-based approaches that may struggle to accurately capture system dynamics, especially in dynamic real-world environments, Q-learning adapts well to these complexities.
Secondly, Q-learning offers flexibility in handling nonlinear relationships and complex decision-making processes. In EV battery management, where optimal control strategies can vary based on driving conditions, user preferences, and battery health, Q-learning’s flexibility allows it to adapt to changing circumstances without requiring extensive redesign or recalibration.
Finally, Q-learning facilitates data-driven optimization by learning optimal strategies directly from data. In environments like EV battery management, where historical data are abundant, such as EV telemetry data, Q-learning can leverage past experiences to identify patterns and trends, leading to efficient battery-management strategies. This data-driven approach allows Q-learning to continually optimize performance over time, making it well-suited for the complexities of EV battery management [2,76,77,78].
Choosing a Q-table approach for Q-learning in EV battery management offers distinct advantages, particularly for its simplicity and transparency. Q-tables provide an easily interpretable representation of Q-values for each state–action pair, essential in safety-critical systems like EV battery management. They also offer lower computational complexity due to their fixed size, enhancing training and inference efficiency, especially in systems with smaller state and action spaces. Furthermore, Q-tables are data-efficient, requiring less data to learn optimal policies, and are guaranteed to converge to the optimal policy under certain conditions. While Q-tables are effective for smaller or discrete state spaces, they may not suit large or continuous spaces where function approximation methods, despite their higher computational demands and complexity, become necessary. The choice between Q-tables and function approximation methods should consider factors like interpretability, computational resources, and environmental complexity in EV battery management [2,76,77,78,79].
The exact formula of the reward function in a Q-learning-based EV battery-management system would depend on the specific objectives and requirements of the application. The reward function typically quantifies how beneficial or desirable a particular state–action transition is for the agent and plays a crucial role in guiding the learning process toward achieving the desired goals. In EV battery management, the reward function aims to encourage actions that lead to desirable outcomes such as prolonging battery life, maximizing driving range, improving energy efficiency, or ensuring safe operation. The reward function may incorporate various factors and considerations relevant to the EV’s performance and battery health [77,78].
Here is a general form of the reward function for EV battery management:
[ R s , a , s = f s , a , s ]
where:
( R ) is the reward function.
( s ) is the current state of the system.
( a ) is the action taken by the agent.
( s ) is the resulting state after taking action \( a \).
( f ) is a function that computes the reward based on the transition from state ( s ) to state ( s ) due to action ( a ) .
The specific formulation of ( f ) depends on the objectives and constraints of the EV battery-management system. For example, the reward function may consider factors such as:
State of Charge (SoC): rewarding actions that maintain the battery SoC within a desirable range to prolong battery life and ensure sufficient energy for driving.
State of Health (SoH): penalizing actions that degrade the battery’s health or accelerate degradation.
Energy Efficiency: rewarding actions that optimize energy consumption and minimize wasted energy.
Safety: penalizing actions that could lead to unsafe operating conditions or increase the risk of accidents.
Driving Range: rewarding actions that maximize the driving range of the EV by efficiently utilizing the available battery capacity.
The exact formulation of the reward function would involve specifying the weights or coefficients assigned to each factor and possibly incorporating additional considerations based on domain expertise and system requirements. Additionally, the reward function may include a discount factor to account for the cumulative long-term effects of actions on future rewards.
Balancing exploration and exploitation is a fundamental aspect of reinforcement learning, particularly in algorithms like Q-learning, which are applied in dynamic environments such as EV battery-management systems. Exploration involves experimenting with new actions to uncover potentially better strategies, while exploitation involves leveraging existing knowledge to maximize short-term rewards. Achieving the right balance between exploration and exploitation is crucial for efficient learning and optimal decision-making in such dynamic contexts.
The epsilon-greedy Strategy is a straightforward yet effective approach commonly employed to balance exploration and exploitation. In this strategy, the agent randomly selects an action with a small probability ( ϵ ) to explore new possibilities. However, most of the time, it chooses the action with the highest estimated Q-value, prioritizing the exploitation of known information. Over time, the value of ( ϵ ) is systematically reduced, gradually shifting the agent’s focus toward exploitation as it gains more familiarity with the environment and learns from past experiences. This adaptive adjustment ensures that the agent continues to explore new options early on but increasingly relies on exploitation as it becomes more knowledgeable.
In our strategy to balance exploration and exploitation within the Q-learning framework for EV battery management, we utilized a method known as epsilon-greedy exploration. This approach effectively manages the trade-off between exploring new actions and exploiting actions that have previously resulted in high rewards.
During exploration, the agent occasionally selects a random action with a small probability ( ϵ ) . This random selection enables the agent to explore different parts of the state–action space, allowing for the discovery of potentially better strategies. By incorporating randomness into the decision-making process, the exploration phase prevents the agent from becoming entrenched in suboptimal policies. Instead, it encourages continual learning and adaptation to the environment, ensuring that the agent remains responsive to changes and opportunities for improvement.
In our Q-learning approach for EV battery management, the state space represents the various states or conditions of the battery system that the agent must consider when making decisions. The state space is a critical component of the Q-learning algorithm as it defines the environment in which the agent operates and influences its learning process and decision making. The state space in our model comprises a set of variables or features that describe the current state of the EV battery system. These variables are chosen based on their relevance to the battery management task and their ability to capture important aspects of the system dynamics. While the specific design of the state space may vary depending on the requirements of the application, some common components include the following:
State of Charge (SoC): The SoC represents the current level of charge stored in the battery, expressed as a percentage of its total capacity. It is a fundamental parameter that influences the battery’s energy storage and discharge capabilities.
  • State of Health (SoH): The SoH reflects the overall health or condition of the battery, considering factors such as degradation, aging, and performance loss over time. Monitoring the SoH is crucial for predicting battery life and optimizing its usage.
  • Temperature: The battery temperature plays a significant role in its performance, efficiency, and longevity. Monitoring temperature helps prevent overheating, which can degrade the battery and affect its safety and reliability.
  • Voltage: The voltage across the battery terminals provides insights into its electrical potential and helps determine its operating conditions and performance characteristics.
  • Current: The current flowing into or out of the battery indicates its charging or discharging status and influences its energy transfer and power delivery capabilities.
Environmental Factors: External factors such as the ambient temperature, humidity, and operating conditions may also affect battery performance and behavior. Incorporating these environmental variables into the state space can enhance the model’s accuracy and robustness.
By defining a comprehensive state space that includes relevant battery parameters and environmental factors, our Q-learning model can effectively capture the complexities of the EV battery system and make informed decisions to optimize its operation, improve energy efficiency, and prolong battery life.
Our EV battery-management system will enhance clarity and provide a deeper understanding of our proposed methodology. Here is how addressing each of these inputs can contribute to clarity:
  • Choice of Q-learning: Providing a rationale for why Q-learning was chosen over other methods offers insight into the decision-making process. This could involve discussing the advantages of reinforcement learning in dynamic environments like EV battery management, where the agent learns optimal policies through trial and error.
  • Selection of Q-Table: Explaining why a Q-table was chosen instead of function approximation options like neural networks or deep learning architectures helps justify the simplicity and interpretability of the chosen method. This may be due to the relatively low computational complexity and ease of implementation, which are advantageous in real-time applications like battery management.
  • Exact Formula of Reward Function: Describing the reward function in detail, including its mathematical formulation and the factors it considers, provides transparency about how the agent is incentivized to take certain actions. This clarity helps stakeholders understand the objectives of the system and how rewards are aligned with performance metrics such as energy efficiency and battery health.
  • Balancing Exploration and Exploitation: Discussing the strategies employed to balance exploration (trying new actions) and exploitation (choosing actions based on current knowledge) sheds light on the agent’s learning process. Techniques like epsilon-greedy exploration or adaptive learning rates may be used to ensure a balance between exploring new possibilities and exploiting known strategies.
  • Description of State Space: Providing insight into the structure and components of the state space elucidates the context in which the agent operates. This includes explaining the choice of state variables, their significance in capturing the battery system’s dynamics, and how they collectively represent the system’s current state.
By addressing these aspects in our methodology, we offer a comprehensive and transparent explanation of our proposed approach, enabling readers to grasp the intricacies of our EV battery-management system and its underlying reinforcement learning framework with greater clarity and confidence.
Designing the reward mechanism is crucial in reinforcement learning as it guides the agent to learn desirable behaviors. In the context of EV battery management using Q-learning, the reward mechanism should encourage actions that lead to improved battery performance, energy efficiency, and overall system reliability. Here is how we can specify the design of the reward mechanism and identify situations that increase or decrease the reward:
  • Definition of reward function: The reward function should quantify the desirability of the state–action pairs encountered by the agent. It typically takes the form of R s , a , s , where ( s ) is the current state, ( a ) is the action taken, and ( s ) is the resulting state after taking action (a).
  • Objective alignment: The reward function should be aligned with the objectives of EV battery management, such as maximizing energy efficiency, prolonging battery lifespan, and ensuring vehicle reliability. For example, higher rewards can be assigned for actions that lead to optimal state-of-charge (SoC) levels, efficient charging/discharging cycles, and avoiding battery degradation.
  • Positive and negative rewards: Situations that contribute positively to battery health and system efficiency should be rewarded, while actions that degrade battery performance or compromise system reliability should incur negative rewards. For instance, positive rewards include successfully reaching target state of charge (SoC) levels without overcharging or deep discharging, efficiently utilizing regenerative braking to recharge the battery, and implementing smart charging strategies that minimize grid stress and energy costs. On the other hand, negative rewards encompass over-discharging the battery, which leads to a reduced battery lifespan, charging at high temperatures that accelerate battery degradation, and ignoring battery-management system warnings or safety protocols.
  • Balancing short-term and long-term rewards: The reward mechanism should strike a balance between immediate gains and long-term benefits. While certain actions may yield immediate rewards (e.g., rapid charging for short-term energy availability), they may have detrimental effects on battery health in the long run. Therefore, the reward function should consider the trade-offs between short-term gains and long-term sustainability.
  • Dynamic reward adjustment: To adapt to changing conditions and system requirements, the reward mechanism may need to be dynamically adjusted over time. This could involve modifying reward weights, thresholds, or penalty values based on real-time feedback, system performance metrics, or user-defined preferences.
By specifying a well-designed reward mechanism, we can guide the Q-learning agent to learn optimal policies for EV battery management, leading to improved energy efficiency, extended battery lifespan, and enhanced overall system reliability.
In practical applications, the state-of-health (SOH) value of a battery is typically obtained through a combination of diagnostic techniques and monitoring systems. These methods may include the following:
  • Voltage and current measurement: monitoring the battery’s voltage and current during charging and discharging cycles can provide insights into its health. Deviations from expected voltage profiles or irregularities in current flow may indicate degradation or internal damage.
  • Temperature monitoring: elevated temperatures during charging or discharging can accelerate battery degradation. Temperature sensors placed within the battery pack can monitor thermal behavior and identify potential issues affecting SOH.
  • Cycle counting: tracking the number of charge–discharge cycles the battery undergoes provides an estimate of its aging process. As batteries degrade over time with repeated cycles, monitoring the cycle count can help assess SOH.
  • Impedance spectroscopy: this technique involves applying small-amplitude alternating current signals to the battery and analyzing its impedance response. Changes in impedance over time can indicate degradation mechanisms, providing insights into SOH.
  • Capacity testing: periodic capacity tests involve fully charging the battery and then discharging it at a controlled rate to measure its energy storage capacity. Comparing the measured capacity with the battery’s initial capacity provides an indication of SOH.
  • Machine learning algorithms: Advanced data-analytics techniques, such as machine learning algorithms, can analyze historical data from battery performance, including voltage, current, temperature, and cycle count, to predict SOH and identify degradation trends.
Once the SOH value is obtained, it can be utilized in various ways to optimize energy efficiency and prolong the battery life:
  • Optimized charging strategies: by considering SOH in charging algorithms, such as adjusting charging voltage and current limits based on battery health, overcharging or undercharging can be mitigated, leading to improved energy efficiency and reduced degradation.
  • Load management: incorporating SOH into load management algorithms allows for better distribution of energy usage across the battery cells. Balancing the load based on cell health ensures that healthier cells are not overburdened, optimizing overall energy efficiency.
  • Predictive maintenance: using SOH data, predictive maintenance algorithms can anticipate battery failures or degradation trends before they occur. Proactive maintenance actions, such as cell balancing or capacity restoration, can be scheduled to prevent efficiency losses.
  • Energy-storage system optimization: in applications involving energy-storage systems (ESS), integrating SOH data enables the optimization of ESS operation and dispatch strategies. Prioritizing the use of cells with higher SOH levels prolongs the system’s lifespan and maximizes energy efficiency.
A failure to utilize SOH in practical applications can lead to suboptimal energy efficiency and premature battery degradation. Without SOH monitoring and management:
  • Inefficient charging and discharging: batteries may be subjected to inappropriate charging or discharging rates, leading to energy losses and accelerated degradation. Without considering SOH, batteries may be overcharged, leading to capacity loss and reduced efficiency.
  • Uneven cell degradation: without balancing load or energy distribution based on SOH, healthier cells may be underutilized, while degraded cells may be overutilized. This imbalance accelerates degradation in weaker cells, leading to reduced system efficiency and overall capacity.
  • Unplanned downtime: a failure to predict battery degradation based on SOH may result in unexpected system failures or downtime. This can lead to disruptions in energy supply, increased maintenance costs, and reduced system reliability.
Integrating SOH monitoring and management into practical applications is essential for optimizing energy efficiency, prolonging battery life, and ensuring the reliable operation of energy-storage systems. A failure to utilize SOH can lead to inefficiencies, premature degradation, and increased maintenance costs.
The traditional battery-management approach typically relies on simple algorithms based on predetermined thresholds and rules to manage battery operation. Effective battery management involves several critical components. First, setting predefined voltage and current thresholds for charging and discharging is essential to prevent overcharging, over-discharging, and excessive current flow, which can degrade battery performance and shorten its lifespan. Second, monitoring the battery temperature helps prevent overheating, which can accelerate degradation and compromise safety. Temperature sensors trigger cooling mechanisms or limit charging rates when temperatures exceed safe limits. Third, estimating the battery’s state of charge (SoC) based on voltage measurements and coulomb counting techniques helps prevent over-discharging and ensures sufficient charge for operation. Fourth, periodically assessing the battery’s state of health (SoH) through capacity testing or impedance measurements detects degradation and anticipates potential failures. Finally, balancing the charge across battery cells ensures uniform cell voltages and prevents overcharging or the over-discharging of individual cells.
While this traditional approach provides basic protection and management of batteries, it may lack adaptability and optimization for varying operating conditions and battery degradation patterns. More advanced approaches, such as those based on machine learning algorithms like Q-learning, offer the following advantages:
  • Adaptability: machine learning algorithms can adapt to changing conditions and learn optimal strategies for battery management based on historical data and real-time feedback.
  • Optimization: advanced algorithms can optimize battery operation considering factors such as load profiles, environmental conditions, and battery health to maximize energy efficiency and extend battery life.
  • Prediction and prognostics: Machine learning models can predict battery degradation trends and anticipate failures, enabling proactive maintenance and the optimization of battery performance.
  • Dynamic control: machine learning-based approaches offer dynamic control of battery operation, adjusting strategies in real-time based on evolving conditions and system requirements.
In experiments, comparing the traditional battery-management approach with more advanced approaches allows researchers to assess the effectiveness and benefits of advanced algorithms in terms of energy efficiency, battery lifespan, and overall system performance. By benchmarking against traditional methods, researchers can demonstrate the improvements achieved with advanced techniques and validate their suitability for practical applications in various domains.

3.2. Mathematical and Conceptual Framework of Q-learning

3.2.1. Mathematical Model

The Q-learning algorithm learns the Q-function by iteratively updating it based on the rewards that the agent receives. The update rule is as follows:
Q(s, a) = Q(s, a) + α[r + γmax_a′ Q(s′, a′)]
where:
Q(s, a) is the action value for state s and action a;
α is the learning rate, which controls how much the Q-function is updated at each time step;
r is the reward received for taking action a in state s;
γ is the discount factor, which controls how much the agent values future rewards;
max_a′ Q(s′, a′) is the maximum action value for the next state s′.
The learning rate and discount factor are hyperparameters that can be tuned to affect the performance of the Q-learning algorithm. A higher learning rate will cause the Q-function to update more quickly, while a higher discount factor will cause the agent to value future rewards more highly.
In ε-greedy exploration, the agent selects the action with the highest estimated value with probability ( 1     ε )   (exploitation) and selects a random action with probability ( ε ) (exploration). The parameter ( ε ) controls the balance between exploration and exploitation.
The formula to express ε-greedy exploration can be represented as follows:
Let ( A ) be the set of all possible actions, and let ( Q ( s ,   a ) ) be the estimated value (Q-value) of taking action ( a ) in state ( s ).
The probability of selecting action \( a \) in state \( s \) under the ε-greedy policy is given by
π ( a | s ) = ε   /   | A |   +   ( 1     ε )   i f   a * = a r g m a x a   Q ( s , a )   ε   /   | A |   o t h e r w i s e
where:
( a * )   is   the   action   that   maximizes   the   estimated   value   ( Q s ,   a )   in   state ( s ) .
( A )   is   the   total   number   of   actions   in   the   action   space .
This formula indicates that with probability ( 1 ε \ ), the agent selects the action with the highest estimated value (exploitation), and with probability ( ε \ ), the agent selects a random action from the action space (exploration).
In this paper, the specific value of ( ε ) and the way it changes over time or with different conditions should be described to fully capture the ε-greedy exploration strategy adopted.

3.2.2. Q-learning Technique for Predicting the Cycle Lifespan of EVs

Q-learning, a type of reinforcement learning algorithm, is not only useful for real-time decision making but also has the potential to predict complex phenomena such as the cycle lifespan of EV batteries. The ability of Q-learning to handle sequential decision-making problems and learn from interactions with the environment makes it a powerful tool for predicting battery lifespan, where the state of the battery evolves over time based on usage patterns and environmental conditions.
Predictive Capabilities of Q-learning
1. Dynamic Modeling of Battery State:
State Representation: Q-learning can incorporate various states of the battery, such as state of charge (SoC), state of health (SoH), temperature, and other environmental factors. By continuously updating these states based on real-time data, Q-learning can create a dynamic model that reflects the battery’s condition over time.
State Transition: The algorithm can model how the battery’s state transitions with each charging/discharging cycle. This is crucial for predicting degradation patterns that affect the battery’s lifespan.
2. Reward Mechanism for Long-Term Prediction:
Designing Rewards: In the context of battery lifespan prediction, the reward function can be designed to reflect the battery’s longevity. For instance, actions leading to optimal battery usage that minimize degradation can be rewarded, while actions causing rapid wear and tear can be penalized.
Cumulative Reward Optimization: By focusing on cumulative rewards, Q-learning can learn strategies that maximize the overall health and lifespan of the battery, providing insights into how different usage patterns affect longevity.
3. Exploration–Exploitation Balance:
Data-Driven Insights: Through the exploration–exploitation mechanism, Q-learning can explore various usage scenarios and their impacts on battery life. This helps in understanding the effects of different charging habits, driving patterns, and environmental conditions on the battery’s cycle lifespan.
Policy Learning: The algorithm can learn policies that optimize battery usage for maximum lifespan, offering predictive insights into the best practices for battery management.
Implementation Steps for Predictive Modeling
1. Data Collection and State Initialization:
Gather comprehensive data on battery usage, including SoC, SoH, temperature, and charging/discharging cycles.
Initialize the Q-table with states representing different conditions of the battery.
2. Defining Actions and Rewards:
Define actions such as charging, discharging, and idling, each with implications on the battery’s health.
Develop a reward function that aligns with the objective of maximizing the battery’s cycle lifespan.
3. Training the Q-learning Model:
Use historical data to train the Q-learning model, allowing it to learn from past battery usage patterns.
Continuously update the Q-values based on the rewards received, refining the model’s predictions.
4. Predictive Analysis and Policy Optimization:
Use the trained Q-learning model to simulate future usage scenarios and predict the battery’s cycle lifespan.
Optimize the policy to recommend actions that extend the battery’s lifespan, providing actionable insights for users and manufacturers.
Q-learning offers a robust framework for predicting the cycle lifespan of EV batteries by leveraging its ability to model dynamic systems, optimize long-term rewards, and balance exploration with exploitation. By integrating real-time data and refining predictive models, Q-learning can significantly enhance our understanding of battery degradation processes and inform strategies to extend the life of EV batteries.

3.2.3. Conceptual Framework of Q-learning

Q-learning is a widely used reinforcement learning algorithm that is particularly well-suited for battery management in electric vehicles (EVs). The core mathematical framework of Q-learning is based on the Bellman equation, which is used to determine the optimal action in a given state to maximize the expected cumulative reward. In the context of battery management, Q-learning can be applied as follows:
State (S): The state in Q-learning represents the current condition of the battery. It typically includes factors like the state of charge (SoC), state of health (SoH), environmental variables (e.g., temperature), and charging infrastructure conditions (e.g., charging station availability).
Action (A): Actions represent the choices available to the EV’s battery-management system, such as adjusting the charging rate, discharging power, or entering a low-power mode to conserve energy.
Reward (R): The reward is a numerical value that reflects the benefit or cost associated with a specific action taken in a particular state. In battery management, rewards may be based on energy efficiency, battery health preservation, or operational cost savings.
Policy (π): The policy is a strategy that guides decision making by specifying which action to take in each state. The objective is to learn the optimal policy that maximizes the expected cumulative reward over time.

3.3. Steps for Implementing Q-learning for Battery Management

The application of Q-learning for battery management involves the following steps (Figure 3):
Step 1: Initialization: the Q-table is a data structure that stores the expected cumulative rewards for each state–action pair. Initially, the Q-values are typically set to zero or random values.
Step 2: Exploration vs. Exploitation: balance exploration and exploitation by choosing actions based on an exploration–exploitation strategy. Common strategies include ε-greedy, where the algorithm chooses the action with the highest Q-value most of the time but explores occasionally by selecting a random action.
Step 3: State Transition: in an EV, the battery state evolves as the vehicle operates. At each time step, observe the current state (SoC, SoH, etc.) and choose an action based on the current policy.
Step 4: Reward Calculation: after taking an action, observe the immediate reward associated with that action in the current state.
Step 5: Q-Value Update: update the Q-value for the state–action pair using the Q-learning update rule, which is typically based on the Bellman equation. The new Q-value is a combination of the old Q-value, the immediate reward, and the maximum expected cumulative reward from the next state.
Step 6: Policy Update: update the policy to choose actions that maximize the Q-values, ensuring that the algorithm converges toward the optimal policy over time.
Step 7: Iteration: repeat Steps 3–6 over multiple iterations or episodes, allowing the algorithm to learn and refine its policy based on historical data.
Step 8: Convergence: monitor the convergence of the Q-values and policy. Convergence indicates that the algorithm has learned an optimal policy for battery management.

3.4. Selection of State, Action, Reward, and Policy

The choice of state, action, reward, and policy is crucial in the Q-learning algorithm for battery management. The state should encompass all relevant battery and environmental parameters, while the actions should include a comprehensive set of control options. Rewards should be carefully designed to align with the objectives of battery management, such as energy efficiency and battery health preservation.
The policy selection may involve exploration–exploitation strategies like ε-greedy, as mentioned earlier, or other variants depending on the specific requirements of the battery-management system. The selection of these components should be based on a deep understanding of the battery system’s characteristics and the desired optimization goals.
In summary, Q-learning offers a systematic and adaptable approach to battery management, allowing the EV to make intelligent, data-driven decisions that optimize battery performance and efficiency while extending the battery life.
In designing actions based on various states such as SOC, SOH, temperature, and humidity, it is essential to define a set of comprehensible actions that can effectively control and manage the electric vehicle (EV) battery system. Below is a clear explanation of how these states can be used to design actions:
1. State of Charge (SOC): SOC represents the current amount of charge stored in the battery relative to its capacity. Actions related to SOC management could include:
  • Charging: Initiate charging to increase SOC.
  • Discharging: Initiate discharge to decrease SOC.
  • Idle: Maintain the current SOC without charging or discharging.
2. State of Health (SOH): SOH indicates the health condition or degradation level of the battery. Actions based on SOH could involve:
  • Maintenance: Implement measures to mitigate degradation or extend battery lifespan.
  • Replacement: Determine if the battery needs replacement based on its degradation level.
3. Temperature: affects battery performance and lifespan. Actions considering temperature might include:
  • Cooling: Activate cooling systems to prevent overheating.
  • Heating: Activate heating systems to maintain optimal temperature in cold conditions.
  • Regulation: Adjust charging and discharging rates based on temperature to optimize battery performance.
4. Humidity: can impact the overall environment of the battery system. Actions concerning humidity may involve:
  • Ventilation: Control airflow to manage humidity levels within acceptable limits.
  • Sealing: Implement measures to prevent moisture ingress into sensitive components.
To design comprehensible actions given the high-dimensional states, it is crucial to define action spaces that are meaningful and manageable for the EV battery-management system. This may involve discretizing the continuous state space into a finite set of discrete states and mapping them to corresponding actions.
For example, a simplified action space for an EV battery-management system could include actions such as the following:
  • Low Charge: Charge the battery to increase SOC.
  • High Discharge: Discharge the battery to decrease SOC.
  • Temperature Control: Activate temperature-regulation mechanisms.
  • Maintenance Mode: Implement measures to mitigate degradation based on SOH.
These actions should be defined based on the specific objectives of the battery-management system, considering factors such as energy efficiency, battery longevity, and overall vehicle performance. Additionally, reinforcement learning algorithms, such as Q-learning, can be used to learn the optimal action selection policy based on the observed states and rewards.

3.5. Data Collection and Model Training

3.5.1. Data Sources and Parameters

To effectively implement Q-learning for EV battery management, it is essential to have access to relevant data sources and define the parameters used in the Q-learning model. The data sources and parameters considered for this study play a crucial role in the model’s performance and accuracy.

Data Sources

1.
Battery-Management System (BMS)
The primary source of data is the EV’s BMS, which continuously monitors and records battery-related parameters. The BMS provides information such as state of charge (SoC), state of health (SoH), temperature, voltage, current, and power consumption.
The presented Figure 4 illustrates the dynamic evolution of an EV battery’s state of charge (SoC) over time, serving as a visual depiction of critical insights for battery management. SoC, indicative of the remaining energy in the battery, diminishes gradually during vehicle operation, with higher speeds, rapid acceleration, and elevated ambient temperatures accelerating the depletion rate. Conversely, the figure highlights the potential to replenish SoC through charging, with the rate of increase influenced by the charging rate and charger type. These findings offer valuable guidance for optimizing battery management and prolonging battery life. Strategies may involve avoiding high-speed driving and rapid acceleration to mitigate SoC decline, while thoughtful charging practices can be employed for efficient replenishment. The comprehensive understanding provided by Figure 4 serves as a foundation for informed decisions in EV usage and maintenance.
2.
Environmental Data
External factors like temperature, humidity, and ambient air pressure can influence battery performance. These data are collected from onboard sensors or external weather databases.
Figure 5 offers a detailed snapshot of the ambient conditions in Bangkok, Thailand, on 1 January 2023, at 12:34 p.m., presenting a temperature of 30.5 °C and a humidity level of 75%, both typical for November in the region. Crucially, these environmental factors wield a substantial influence on electric vehicle (EV) battery performance. Elevated temperatures can diminish battery capacity and hasten degradation, while heightened humidity poses the risk of corrosion and other detrimental effects. The implementation of Q-learning for EV battery management must inherently account for these conditions. Adaptive decision making by the Q-learning model is imperative, enabling adjustments in charging rates or power consumption to safeguard the battery against potential damage. Specific observations from the figure, such as the relatively high temperature and humidity, underscore the adverse impact on battery performance. This insight lays the groundwork for strategic approaches to optimize battery management and prolong battery life in hot and humid climates. For instance, EV drivers in Bangkok may consider refraining from charging during peak temperature periods and seeking shaded parking to shield their vehicles from the sun. In essence, Figure 5 underscores the critical role of factoring temperature and humidity into Q-learning strategies for effective EV battery management.
Figure 6 presents a comprehensive depiction of the voltage and current dynamics of an electric vehicle (EV) battery over time. Voltage, denoting the electrical potential difference across the battery terminals, exhibits a decline concurrent with an increase in current, reflecting the impact of the battery’s internal resistance. This internal resistance induces a drop in voltage as the battery dispenses current. Moreover, this figure highlights the utility of the voltage and current data in calculating the battery’s power consumption, derived from the product of voltage and current. Beyond consumption metrics, these data serve as critical indicators for monitoring the battery’s health, with abrupt voltage drops or current spikes potentially signaling damage.
Key observations from the figure include the inverse relationship between voltage and current, the calculation of power consumption, and the utility of these data in health monitoring. These insights lay the groundwork for strategic approaches to optimize battery management and extend battery life. For instance, minimizing high-power applications, such as rapid acceleration, can curtail current consumption, thereby enhancing battery longevity. Additionally, vigilant monitoring of voltage and current enables early detection of potential damage, prompting timely intervention.
In essence, Figure 6 furnishes a valuable overview of the voltage and current characteristics of EV batteries, empowering stakeholders with the insights needed to formulate effective battery-management strategies. This knowledge proves instrumental in optimizing performance and prolonging the operational life of EV batteries.
3.
Charging Infrastructure Data
Information about charging station locations, charging rates, and availability is crucial for making charging-related decisions.
4.
Charging Station Rates at Different locations
Figure 7 offers a comprehensive depiction of charging station rates across various locations in Thailand, presenting rates in Thai baht (THB) per kilowatt-hour (kWh). Notably, the rates exhibit variation contingent on both the geographic location of the charging station and the type of charger deployed. Major cities like Bangkok and Phuket feature the highest rates, reaching up to 10 THB/kWh for fast chargers. Conversely, rural areas present the lowest rates, potentially as affordable as 3 THB/kWh for slow chargers, albeit with the caveat of potentially fewer available stations.
Figure 8 underscores the significance of factoring in charging costs when planning an electric vehicle (EV) trip. Utilizing the map in Figure 4 enables users to pinpoint charging stations along their route, facilitating rate comparisons. Online resources can further aid in estimating the overall charging cost at a specific station.
Key observations, including the correlation between charging rates and city or rural locations, emphasize the need to develop strategies for cost-effective EV charging. For instance, trip planning can be optimized by avoiding major cities where rates are higher, and preference can be given to slower chargers offering more economical rates. In essence, Figure 8 serves as a valuable resource for devising prudent and cost-efficient EV charging strategies, contributing to informed decision making for EV drivers in Thailand.
5.
Driving Patterns
Data on vehicle speed, acceleration, deceleration, and route information are obtained through GPS or other vehicle-tracking systems. This information helps in predicting driving patterns and energy consumption.
Figure 9 presents insightful driving-pattern data for an EV driver in Bangkok, Thailand, on 11 November 2023, encompassing key parameters such as speed, acceleration, and deceleration. The depicted driving patterns reveal dynamic variations throughout the day, starting with the driver’s commute to work, a prolonged park duration, and concluding with the journey home, including a stop at a grocery store.
Significantly, the figure illustrates frequent acceleration and deceleration, characteristic of urban driving conditions where frequent stops for traffic lights and other vehicles are commonplace. These driving-pattern data assume paramount importance in EV battery management, serving as a foundation for predicting energy consumption. The Q-learning model can leverage this information to optimize charging and discharging decisions. For instance, anticipating a day with substantial stop-and-go traffic, the model might recommend more frequent charging to accommodate the energy demands of the driving pattern.
Specific observations from the figure, such as varying speeds with a peak at around 80 km/h, frequent accelerations and decelerations, and the characteristic stop-and-go traffic, underline the city-centric driving nature. These observations can be harnessed to develop targeted strategies for optimizing battery management in urban driving conditions. Training the Q-learning model on datasets specific to city driving empowers it to learn and adapt to the nuanced challenges posed by frequent stops and accelerations.
In essence, Figure 10 elucidates the crucial role of driving pattern data in EV battery management. By discerning the driver’s habits, the Q-learning model can make nuanced decisions about charging and discharging, ultimately enhancing performance and prolonging battery life in the context of city driving.
Figure 10 shows the location of charging stations in Thailand, as well as a simulated GPS route. The charging stations are marked with red markers, and the GPS route is shown as a red line. This map can be used to help electric vehicle drivers plan their routes and identify charging stations along the way. It can also be used to visualize the availability of charging infrastructure in Thailand.

3.5.2. Parameters

1. State Parameters:
State of Charge (SoC): The current charge level of the battery.
State of Health (SoH): The overall health and condition of the battery.
Temperature: The battery’s operating temperature, which impacts performance and degradation.
2. Action Parameters:
  • Charging Rate: The rate at which the battery is charged.
  • Discharging Rate: The rate at which the battery discharges power to the motor.
  • Idle Mode: The option to enter an energy-saving idle mode.
3. Reward Parameters:
  • Energy Efficiency: Reward based on how efficiently energy is utilized.
  • Battery Health: Reward related to maintaining battery health and prolonging its lifespan.
  • Operational Cost Savings: Reward based on cost-efficient charging and discharging strategies.
4. Policy Parameters:
  • Exploration Rate (ε): The rate at which the algorithm explores new actions rather than exploiting known ones.
  • Discount Factor (γ): A parameter that balances the importance of immediate rewards against future rewards.
  • Learning Rate (α): A factor that determines how much the algorithm adjusts its Q-values based on new information.

3.5.3. Training Process

For offline learning in Q-learning training, a common method is to use historical data or simulation data to train the Q-learning algorithm before deploying it in the actual environment. This approach allows the algorithm to learn from past experiences without interacting with the real-world system.
In EV battery management, the offline learning process might involve the following steps:
1. Data collection: gather historical data from EVs, including information such as SOC, SOH, temperature, humidity, driving patterns, and charging/discharging cycles. These data can be obtained from real-world EV fleets or simulated environments.
2. Preprocessing: clean and preprocess the collected data to remove noise, handle missing values, and normalize the features. This ensures that the data are suitable for training the Q-learning algorithm.
3. Training: use the preprocessed data to train the Q-learning algorithm offline. During training, the algorithm learns to associate different states with actions that maximize long-term rewards based on the provided historical data.
4. Evaluation: validate the trained Q-learning algorithm using validation data or through cross-validation techniques. Evaluate its performance in terms of convergence, learning efficiency, and generalization to unseen data.
5. Fine-Tuning: optionally, fine-tune the Q-learning algorithm based on the evaluation results to improve its performance further.
Common techniques for offline learning in Q-learning include batch learning and experience replay. Batch learning involves training the Q-learning algorithm using a fixed dataset for a predetermined number of iterations or epochs. Experience replay, on the other hand, involves randomly sampling experiences from a replay buffer to train the Q-learning algorithm. This method allows the algorithm to learn from past experiences more efficiently and helps avoid correlations between consecutive samples.
These offline learning methods enable the Q-learning algorithm to learn from historical data or simulated environments, providing a foundation for optimal decision making in real-world scenarios.
In this paper, although Deep Q-Networks (DQN) are mentioned, they are not employed in the current work. Instead, the focus is on traditional Q-learning for EV battery management. While DQN is a powerful extension of Q-learning that uses neural networks to approximate the Q-function, it introduces additional complexity and computational overhead.
The decision not to employ DQN in the current work could be due to several reasons:
  • Simplicity: traditional Q-learning with Q-tables is simpler to implement and understand compared to DQN, which involves training a neural network.
  • Data efficiency: DQN often requires a large amount of training data to learn effectively, especially when dealing with high-dimensional state spaces. In contrast, traditional Q-learning can be more data-efficient, making it suitable for scenarios where data collection is limited.
  • Computational resources: training a DQN typically requires more computational resources compared to traditional Q-learning, especially when using deep neural networks. For applications with resource constraints, traditional Q-learning may be preferred.
  • Performance: depending on the specific problem and available resources, traditional Q-learning may achieve comparable or even superior performance to DQN. If the problem can be effectively solved using a simpler approach, there may be no need to employ more complex methods like DQN.
Overall, while DQN is a powerful algorithm for reinforcement learning, its adoption depends on factors such as problem complexity, available resources, and desired performance. In the current work, traditional Q-learning is chosen as the method of choice for EV battery management, possibly due to its simplicity, efficiency, and effectiveness in addressing the problem at hand.
The training process for the Q-learning model involves the following key steps:
Step 1: Data Preprocessing: the collected data are preprocessed to ensure consistency and quality. This may involve filtering outliers, normalizing data, and aligning timestamps.
Step 2: Initialization: initialize the Q-table, setting initial Q-values for each state–action pair.
Step 3: Training Algorithm: utilize a Q-learning training algorithm to iteratively update the Q-values based on the observed rewards. Common training algorithms include Q-learning with ε-greedy exploration or more advanced variants like Deep Q-Networks (DQN) for complex state spaces.
Step 4: Hyperparameter Tuning: optimize the choice of hyperparameters (ε, γ, α) to achieve a balance between exploration and exploitation and ensure convergence to an optimal policy.
Step 5: Policy Evaluation: continuously evaluate the policy’s performance using simulated or real-world testing scenarios.
Step 6: Convergence Monitoring: monitor the Q-values and the policy to assess convergence, ensuring that the algorithm has learned an effective battery-management strategy.
Step 7: Real-time Application: deploy the trained Q-learning model in a real-time setting in an EV to manage the battery based on the learned policy.
The training process requires an iterative approach, allowing the algorithm to learn from historical data and adapt to changing conditions over time. The trained Q-learning model, when integrated into an EV’s battery-management system, is expected to make dynamic decisions that optimize battery performance, extend battery life, and minimize operational costs.
Evaluation Setup: Real-World Pilot
The evaluation of our Q-learning approach for EV battery management was conducted through a real-world pilot aimed at assessing the effectiveness and practical applicability of the proposed methodology in actual driving scenarios. In this section, we provide detailed insights into the experimental setup, data-collection procedures, environmental conditions, and the duration and scope of the pilot.
  • Experimental Setup: the real-world pilot involved a fleet of EVs equipped with the necessary instrumentation to monitor key battery parameters and driving behavior. The vehicles used in the pilot were standard electric models commonly found in urban environments. Testing conditions encompassed various driving scenarios, including city commuting, highway driving, and urban stop-and-go traffic.
  • Data Collection: the pilot was facilitated through onboard sensors integrated into the EVs, as well as external environmental monitoring equipment. The onboard battery-management system (BMS) continuously logged battery-related parameters such as state of charge (SoC), state of health (SoH), temperature, voltage, current, and power consumption. Additionally, GPS or other vehicle tracking systems recorded driving patterns, including vehicle speed, acceleration, deceleration, and route information. Quality-control measures were implemented to ensure data integrity, including regular calibration of sensors and validation of recorded data against ground truth measurements.
  • Environmental Conditions: the pilot was conducted in urban and suburban areas, exposing the EVs to a range of environmental conditions such as temperature, humidity, and varying driving scenarios. Data collection included monitoring of ambient temperature, humidity levels, and driving conditions to capture the influence of environmental factors on battery performance.
  • Duration and Scope: the real-world pilot spanned several months, allowing for comprehensive data collection across different seasons and driving conditions. Multiple vehicles were involved in the pilot, enabling a diverse range of testing scenarios and driving behaviors to be captured. The scope of the pilot encompassed the evaluation of the Q-learning approach’s performance in optimizing energy efficiency, extending battery life, and reducing operational costs in real-world driving environments.
By conducting the evaluation in a real-world pilot setting, we aimed to provide practical insights into the efficacy of our Q-learning methodology in addressing the challenges of EV battery management under diverse operating conditions. The experimental setup, data-collection procedures, and environmental considerations outlined above contribute to a comprehensive understanding of the evaluation process and its implications for real-world application.
The evaluation setup encompassed both simulation and real-world testing phases to comprehensively assess the performance of the Q-learning algorithm for EV battery management.
  • Real-World Testing: deploying a fleet of electric vehicles equipped with the proposed battery-management system in various driving conditions, including urban, highway, and mixed-use scenarios. Data collection was facilitated through onboard sensors and data-logging equipment, capturing essential parameters such as state of charge (SOC), state of health (SOH), temperature, and vehicle performance metrics.
  • Data Analysis: collected data from both the simulation and real-world testing phases and performed thorough analysis to evaluate the performance of the Q-learning algorithm. Key performance indicators, including energy efficiency, battery degradation, and overall vehicle performance, were compared against baseline metrics to assess the effectiveness of the proposed approach.
  • Evaluation Setup: Real-World Testing: following the simulation phase, real-world testing was carried out using a fleet of electric vehicles equipped with the proposed battery-management system. These vehicles were deployed in diverse driving conditions, including urban, highway, and mixed-use scenarios, to evaluate the algorithm’s performance in authentic operational environments. Data collection during real-world testing was facilitated through onboard sensors and data-logging equipment, capturing critical parameters such as state of charge (SOC), state of health (SOH), temperature, and vehicle performance metrics.
  • Data Analysis: data collected from real-world testing phases were subjected to comprehensive analysis. Key performance indicators, such as energy efficiency, battery degradation, and overall vehicle performance, were compared against baseline metrics to assess the effectiveness of the proposed approach. This analysis provided valuable insights into the algorithm’s performance across real-world environments, guiding its further refinement and practical application in electric vehicles.

3.6. Gain Data

1. Define the Problem and Objective:
-
Objective: Identify how Q-learning can help in understanding and optimizing material structure changes. This might involve optimizing processes like heat treatment, alloy composition, or manufacturing conditions.
2. Formulate the Environment:
-
State Representation: Define the state space to represent the different conditions of the material structure. This could include parameters such as temperature, pressure, composition, and time.
-
Action Space: Define the actions that can be taken to alter the material structure. These might include changes in temperature, pressure, or the addition of different materials.
-
Reward Function: Develop a reward function that quantifies the desired outcomes of material structure changes. This could be based on performance metrics like strength, durability, or any other relevant property.
3. Data Collection:
-
Simulation Data: Use simulations to generate data on how different actions affect material structure. This could involve computational models of material behavior under various conditions.
-
Experimental Data: Collect real-world data from experiments where material structures are altered according to different action strategies. Ensure data include various states and the results of actions taken.
4. Preprocessing Data:
-
Data Cleaning: Remove noise, handle missing values, and normalize the data to ensure consistency.
-
Feature Engineering: Identify and extract relevant features that represent the material structure and its changes.
5. Training the Q-learning Model:
-
Initialization: Initialize the Q-table or function approximator with initial values.
-
Learning Process: Apply the Q-learning algorithm to learn the optimal policy for material structure changes. Use historical data to update the Q-values iteratively.
-
Exploration vs. Exploitation: Balance exploration of new actions with exploitation of known optimal actions using ε-greedy or other exploration strategies.
6. Evaluation and Validation:
-
Simulation Testing: Validate the trained Q-learning model using simulation data to check its performance in predicting material structure changes.
-
Real-World Testing: Test the model with real-world data to ensure it generalizes well and provides accurate predictions for material structure changes.
7. Implementation and Optimization:
-
Deployment: Implement the Q-learning-based approach in a real-world setting, if applicable, to optimize material structure changes.
-
Continuous Learning: Update the model with new data to improve its accuracy and adapt to changing conditions over time.
By using Q-learning, one can systematically explore different strategies for altering material structures, learn which strategies lead to desirable outcomes, and optimize processes based on learned policies. Collecting and using both simulation and experimental data will provide a comprehensive approach to understanding and optimizing material structure changes using Q-learning.

4. Experimental Results

4.1. Experiment Results

The application of Q-learning for electric vehicle (EV) battery management was rigorously tested and evaluated to assess its effectiveness in optimizing battery performance, extending battery life, and reducing operational costs. The following section presents the key findings of the experiments and provides visual representations of the improvements achieved.
Figure 11 presents a compelling visual representation of the average rewards per episode for a Q-learning-based electric vehicle (EV) battery-management system. The average reward, derived from the rewards received over all episodes, is a key metric reflecting the system’s ability to optimize battery management. An episode, defined as a complete cycle of driving and charging the EV battery, encapsulates the full range of experiences for the model to learn from.
The figure distinctly showcases a positive trend in the average reward per episode, demonstrating an upward trajectory over time. This upward trend signifies the Q-learning model’s capacity to learn and refine its strategies for optimizing the battery-management system. Crucially, the average reward serves as a pivotal measure to evaluate the model’s performance, with higher values indicative of enhanced battery performance achieved through effective optimization.
Specific observations from the figure emphasize that the average reward per episode not only increases over time but is also higher for longer episodes. This suggests that the model excels in optimizing the battery for extended driving durations, showcasing adaptability to varying conditions.
In summary, Figure 11 provides a compelling illustration of the advantages of employing Q-learning for EV battery management. The observed increase in average rewards underscores the model’s ability to learn and develop effective strategies, thereby enhancing battery performance and contributing to the extension of battery life. This reinforces the efficacy of leveraging Q-learning to optimize EV battery management under diverse driving and charging scenarios.

4.2. Data and Graphical Representation

Table 1 shows the superior performance of Q-learning over the traditional approach in all three key metrics of electric vehicle (EV) battery management. Q-learning-based strategies exhibit heightened energy efficiency, signifying a more judicious utilization of stored energy for vehicle propulsion, ultimately extending the travel range on a single charge. Moreover, the lower average battery degradation rate attributed to Q-learning implies a prolonged battery lifespan, ensuring sustained capacity over time. Notably, the total operational cost, encompassing charging expenses, battery replacement costs, and maintenance, is markedly reduced with Q-learning, highlighting its cost-effectiveness in vehicle operation. These results underscore the promising potential of Q-learning-based battery-management systems to glean insights from experiential data, devising effective strategies that optimize performance, extend battery life, and concurrently diminish operational costs in the realm of electric mobility.
Figure 12 provides a visual comparison of the energy efficiency achieved by the Q-learning-based approach and the traditional approach in EV battery management. The chart displays the average energy efficiency percentages for both methods, facilitating a clear understanding of their relative performance.
The results demonstrate a significant advantage of the Q-learning approach in enhancing energy efficiency. Q-learning achieves an impressive average energy efficiency of 92.5%, surpassing the traditional approach, which achieves 88.3%. This substantial difference underscores the effectiveness of Q-learning in optimizing energy utilization within the EV system.
The higher energy efficiency associated with the Q-learning approach suggests a more efficient utilization of stored energy for propulsion, leading to improved performance and extended travel range on a single charge. These findings underscore the promising potential of Q-learning in advancing energy efficiency in EV battery management, thereby contributing to the continued development of electric mobility technology.
Figure 13 shows a bar chart showing the energy efficiency comparison between Q-learning and traditional battery management over a specific duration.
Figure 14 and Figure 15 illustrates a crucial aspect of electric vehicle (EV) battery management by comparing the average battery degradation rates between two approaches: Q-learning and the traditional approach. The chart provides a clear visual representation of the respective degradation rates, indicating a noteworthy disparity. The Q-learning-based approach exhibits a notably lower average battery degradation rate of 0.8%, in contrast to the traditional approach, which registers a higher rate of 1.5%. A lower degradation rate signifies that the battery under Q-learning management undergoes a slower decline in capacity over time, suggesting an extended lifespan and sustained performance. This comparative analysis underscores the efficacy of Q-learning in mitigating battery degradation, a critical factor in enhancing the longevity and reliability of electric vehicle batteries.
Benchmarks and Comparison
The rule-based approach in traditional EV battery management relies on predefined thresholds and control strategies for managing parameters such as the state of charge (SOC), temperature, and other relevant variables. Unlike more sophisticated techniques such as reinforcement learning, which adapt and optimize based on experience, rule-based strategies follow fixed guidelines and decision rules. While these approaches are straightforward to implement and understand, they may lack the adaptability and optimization capabilities necessary for addressing the complexities of dynamic environments and varying driving conditions encountered in electric vehicle operation.
The traditional approach to EV battery management typically involves rule-based strategies or simple control algorithms based on predefined thresholds for parameters such as the state of charge (SOC), temperature, and other relevant variables. These approaches lack the adaptability and optimization capabilities of more advanced techniques like reinforcement learning. In our study, we benchmarked the performance of the Q-learning approach against this traditional method to assess its effectiveness in optimizing energy efficiency and battery management in EVs.
The comparison revealed notable differences between the two approaches, with the Q-learning approach demonstrating superior performance in terms of energy efficiency and overall battery management. Specifically, our results showed that the Q-learning approach achieved higher energy efficiency percentages and better utilization of the battery capacity compared to the traditional approach across various driving conditions and environmental factors.
While our study focused on comparing the Q-learning approach with the traditional method as a primary benchmark, we acknowledge the importance of including additional benchmarks for a more comprehensive evaluation. We recognize that benchmarking is a key differentiator from existing work and plays a crucial role in validating the effectiveness of novel methodologies in real-world applications. Therefore, we will consider the suggestion to include more benchmarks in future studies to provide a more robust assessment of different battery-management strategies and their implications for electric vehicle performance and sustainability.

4.3. Result of Q-learning Approaches

In Section 4.3, we present the experimental findings showcasing the efficacy of our proposed Q-learning approach in mitigating the effects of environmental factors, particularly varying temperature and humidity, on energy efficiency in EV battery management. Our experiments involved systematically varying temperature levels from 20 to 40 °C and humidity levels from 30 to 70%. For each combination of temperature and humidity, we evaluated the energy efficiency achieved by our Q-learning algorithm.
The results demonstrate that our Q-learning approach consistently maintains high energy efficiency levels across a wide range of temperature and humidity conditions. Specifically, even in scenarios with extreme temperatures and humidity levels, the energy efficiency performance remains stable and resilient. This indicates that our Q-learning algorithm effectively adapts to diverse environmental conditions, suppressing their adverse effects on energy efficiency in EV battery management.
By including these experiment results in this section, we provide empirical evidence supporting the robustness and effectiveness of our proposed Q-learning approach in optimizing energy efficiency while navigating through varying environmental factors. These findings contribute to the advancement of sustainable electric mobility technology by offering insights into the resilience of our algorithm in real-world operating conditions.
Figure 16 illustrates the performance of a Q-learning algorithm in the context of EV battery management. The grid represents different combinations of temperature (ranging from 20 °C to 40 °C) and humidity (ranging from 30% to 70%). Each cell in the grid displays the corresponding energy efficiency percentage achieved by the Q-learning algorithm under the given temperature and humidity conditions. Higher values in the cells indicate greater energy efficiency, showcasing how the Q-learning approach performs across various environmental conditions. This visualization provides insights into how the algorithm adapts to different temperature and humidity levels, offering valuable information for optimizing EV battery-management strategies.
Figure 17 depicts the energy efficiency performance of a traditional approach in EV battery management across different temperature and humidity conditions. Similar to the previous figure, the grid represents combinations of temperature (ranging from 20 to 40 °C) and humidity (ranging from 30 to 70%). Each cell in the grid displays the corresponding energy efficiency percentage achieved by the traditional approach under the given temperature and humidity conditions. Lower values in the cells indicate lesser energy efficiency compared to the Q-learning approach, showcasing potential differences in performance between the two methodologies across various environmental conditions. This visualization provides insights into the energy efficiency performance of the traditional approach and its comparison to the Q-learning approach.
To analyze the feasibility of the proposed approach while considering its computational burden and compare the computational time of different approaches, we need to conduct a thorough examination of the computational requirements of the Q-learning approach and compare it with other methods.
  • Feasibility Analysis:
1. Computational Complexity:
To comprehensively evaluate the Q-learning algorithm, it is important to assess its computational complexity by considering factors such as the size of the state and action spaces, convergence rate, and memory requirements. Additionally, examining the algorithm’s scalability with increasing problem complexity, such as larger state spaces or more frequent updates, is essential. This evaluation will provide insights into the algorithm’s efficiency and feasibility for handling more complex problems and larger datasets. Table 2 illustrates factors and their impact on computational complexity.
2. Resource Consumption:
To thoroughly evaluate the Q-learning algorithm, it is important to measure its CPU and memory usage during both the training and inference stages. Additionally, considering the impact of computational resources on real-time implementation is crucial, particularly for embedded systems or devices with limited processing power. This evaluation will help in understanding the feasibility and efficiency of deploying the Q-learning algorithm in various hardware environments (Table 3).
3. Algorithmic Efficiency:
We analyzed the trade-offs between computational cost and performance gains achieved by the Q-learning approach. We compared the efficiency of the Q-learning algorithm with alternative methods, such as rule-based approaches or model-based control strategies (Table 4).
4. Computational Time:
  • Experiment Setup: conduct experiments to measure the computational time required for training and inference of the Q-learning algorithm. Use a standardized benchmarking environment to ensure fair comparisons between different approaches.
  • Computational Time Measurement: the first step is to conduct experiments to measure the computational time required for both training and inference of the Q-learning algorithm. This involves running the algorithm in a controlled environment and recording the time taken for each stage. It is essential to use appropriate timing mechanisms in your programming language or environment to accurately measure the elapsed time.
  • Standardized Benchmarking Environment: it is crucial to use a standardized benchmarking environment to ensure fair comparisons between different approaches. This environment should provide consistent conditions for testing, including the same initial state configurations, action spaces, and reward structures. By standardizing the environment, you can ensure that the results are comparable across different experiments and approaches.
In summary, conducting experiments to measure computational time and using a standardized benchmarking environment are essential steps to ensure fair comparisons and reliable results when evaluating the performance of the Q-learning algorithm and comparing it with other approaches.
Performance Metrics: we define performance metrics such as the training convergence time, inference latency, and overall computational efficiency in Table 5.
It is important to collect data on computational time for each approach under varying conditions, including different problem sizes or environmental factors. (See Table 6).
2.
Comparative Analysis:
To enhance the analysis, it is essential to compare the computational time of the Q-learning approach with baseline methods or state-of-the-art algorithms. This comparison will highlight the relative efficiency and performance of Q-learning in various scenarios. Additionally, identifying potential bottlenecks or areas for optimization in the Q-learning implementation can further improve computational efficiency. By addressing these areas, the overall performance of the Q-learning algorithm can be significantly enhanced.
To address the points regarding comparative analysis:
1. Comparative Analysis of Computational Time:
It is important to conduct a comparative analysis of the computational time required by the Q-learning approach with baseline methods or state-of-the-art algorithms. This involves running experiments under similar conditions and measuring the time taken for the training and inference stages. By comparing the computational time of different approaches, you can assess the efficiency of the Q-learning algorithm relative to other methods.
2. Identification of Potential Bottlenecks or Areas for Optimization:
It is important to analyze the Q-learning implementation to identify potential bottlenecks or areas for optimization that could improve computational efficiency. This may include profiling the code to identify sections that consume significant computational resources, such as updating Q-values or processing large datasets. Once potential bottlenecks are identified, consider optimization techniques such as algorithmic improvements, parallelization, or hardware acceleration to reduce the computational time.
Conducting a comparative analysis of computational time and identifying potential bottlenecks or areas for optimization are essential steps in improving the efficiency of the Q-learning algorithm and ensuring its competitiveness with other methods.
The feasibility analysis and computational time comparison provide insights into the suitability of the proposed Q-learning approach for practical applications. It is important to consider factors such as computational efficiency, scalability, and real-time performance requirements. Additionally, it is important to offer recommendations for optimizing the computational burden of the Q-learning algorithm, such as parallelization techniques or algorithmic refinements. Ultimately, the goal is to assess whether the benefits of the proposed approach outweigh its computational costs and whether it is a viable solution for real-world deployment.
3.
Computational Time Comparison
To evaluate the computational time of the Q-learning approach, we performed experiments under various conditions and compared the results with those from alternative methods. This comparison provides insights into the efficiency of the Q-learning algorithm in relation to its performance and practical deployment.
1. Experimental Setup: We set up controlled experiments to measure the computational time required for both the training and inference phases of the Q-learning algorithm. Standardized benchmarking environments were used to ensure fair comparisons, including consistent state space configurations, action space definitions, and reward structures.
2. Computational Time Metrics: The computational time was measured across different problem sizes and environmental complexities. The following metrics were recorded:
-
Training Convergence Time: The time taken to reach an acceptable performance level during training.
-
Inference Latency: The time delay between receiving an input and producing an output during the deployment phase.
Table 7 illustrates the computational time for Q-learning and alternative approaches.
4.
Analysis of Trade-offs Between Accuracy and Computational Cost
In the context of the Q-learning approach, there are inherent trade-offs between accuracy and computational cost:
1. Accuracy vs. Computational Cost:
-
High Accuracy Requirements: Achieving higher accuracy often necessitates more complex state and action spaces, leading to larger Q-tables. This results in increased memory usage and longer computation times. For instance, training on a larger state space with more frequent updates may improve the algorithm’s performance but also increase the training time and resource consumption.
-
Balancing Accuracy and Efficiency: To balance accuracy with computational efficiency, adjustments such as simplifying state and action spaces, employing function ap-proximation methods, or optimizing the Q-learning parameters (e.g., learning rate, discount factor) can be made. These strategies help to achieve reasonable accuracy while mitigating excessive computational demands.
2. Potential Bottlenecks:
-
State and Action Space Size: Larger state and action spaces can significantly impact the computational time. For example, the size of the Q-table grows exponentially with the number of states and actions, leading to higher memory requirements and longer update times.
-
Convergence Rate: The speed at which the Q-learning algorithm converges to an optimal policy can also affect computational efficiency. Slower convergence rates require more iterations, increasing both the training time and computational costs.
3. Recommendations for Optimization:
-
Algorithmic Improvements: Techniques such as function approximation (e.g., using Deep Q-Networks) can help reduce the size of the Q-table and improve scalability.
-
Parallelization: Implementing parallel processing or leveraging hardware acceleration can reduce the computation time, especially for larger state spaces and more frequent updates.
-
Profile and Optimize: Profiling the Q-learning implementation to identify and address bottlenecks can lead to significant improvements in computational efficiency.
4. Practical Implications:
-
Deployment Considerations: When deploying Q-learning in real-world scenarios, it is crucial to consider both the accuracy requirements and the computational resources available. For embedded systems with limited processing power, optimizing the algorithm for lower computational cost without sacrificing essential accuracy is important.
In conclusion, while the Q-learning approach offers notable advantages in terms of adaptability and optimization, it is essential to address the trade-offs between accuracy and computational cost. By employing optimization techniques and carefully managing the state and action spaces, the balance between effective performance and computational feasibility can be achieved, making the Q-learning approach a viable option for real-world applications in electric vehicle battery management.

5. Discussion

5.1. General Discussion

Our study contributes to the existing body of knowledge on EV battery management and the application of reinforcement learning, specifically Q-learning. The experimental results presented in this study provide compelling evidence for the efficacy of Q-learning in optimizing EV battery management. The application of Q-learning algorithms has demonstrated significant improvements in battery performance, energy efficiency, and operational cost reduction, which are critical factors in the widespread adoption of electric vehicles [1,2], the trends in the global adoption of EVs, the advancements in electric vehicle battery technology, and the importance of effective battery-management systems. In this discussion, we delve into the key findings from the literature and relate them to the contributions and implications of our study [75,76,77,78].
Research in the field of EV trends and battery technologies has significantly contributed to our understanding of the increasing global adoption of EVs [79,80]. Pioneering works by various researchers have shed light on the pivotal factors driving this surge, including environmental concerns, policy support, technological advancements, and evolving consumer preferences. One notable study aligns closely with this overarching trend, emphasizing the optimization of EV battery management as a critical aspect of the widespread acceptance of electric mobility. Recognizing the challenges faced by EVs, such as a limited driving range, high initial costs, and the imperative for more efficient energy-storage solutions, this research endeavors to address these hurdles through the application of Q-learning in battery management [81].
This work aims to make substantial contributions to the sustainability and affordability of EV batteries [1,81]. Furthermore, researchers have delved into the intricacies of electric vehicle batteries, highlighting innovations like modular designs, low-voltage batteries, and advancements in lithium-ion battery technology. A particular study enhances these innovations by incorporating a dynamic and adaptive Q-learning algorithm for battery management. By integrating Q-learning, the research aims to elevate the efficiency, lifespan, and overall performance of EV batteries, effectively tackling challenges posed by traditional battery-management systems [75,76,77,78].
The result of this study observed an increase in the average rewards per episode, as depicted in Figure 11, which underscores the model’s ability to learn and adapt its strategies over time, resulting in enhanced battery management [81]. This is particularly evident in the improved performance for longer episodes, suggesting the model’s capability to handle extended driving conditions effectively. The comparative analysis presented in Table 1 and Figure 12, Figure 13 and Figure 14 further highlights the superiority of Q-learning over traditional battery-management approaches. The Q-learning-based strategies have shown a marked increase in energy efficiency, with an average of 92.5% compared to 88.3% for the traditional approach. This improvement in energy utilization directly translates to extended travel ranges and reduced charging frequency, making EVs more practical for everyday use [81,82,83]. Moreover, the lower average battery degradation rate observed with Q-learning (0.8% vs. 1.5% for traditional management) indicates a slower decline in battery capacity over time. This extended battery lifespan not only reduces the environmental impact by delaying battery disposal but also lowers the total cost of ownership by postponing the need for battery replacement [83,84]. The reduction in total operational costs, as shown in Table 1, further strengthens the case for Q-learning in EV battery management. By optimizing charging schedules and driving patterns, Q-learning can significantly decrease energy costs and maintenance expenses, making electric vehicles more economically viable for consumers [85]. The visualization in Figure 2 illustrates the flow of information in EV control systems, emphasizing the role of Q-learning in decision making based on real-world driving conditions. This reinforces the adaptability and learning capabilities of Q-learning, contributing to optimal decision making for EV battery management. Finally, the experimental results validate the potential of Q-learning algorithms in revolutionizing EV battery management. The enhanced performance, increased energy efficiency, and reduced operational costs demonstrated in this study suggest that Q-learning could play a pivotal role in accelerating the adoption of electric vehicles and contributing to a more sustainable future.
Pioneering work in the field has underscored the critical significance of BMS in elevating the performance, safety, and efficiency of EV batteries. Building upon this established foundation, our study, led by Lombardo et al. [85], delves specifically into the innovative application of Q-learning in battery management. While acknowledging the indispensable role of BMS in monitoring and safeguarding battery components, the research team introduces a new dimension through Q-learning. This adaptive and intelligent approach enhances the decision-making processes associated with charging, discharging, and overall energy management in EVs. The work by Lombardo et al. [85] aims to leverage Q-learning’s capabilities to optimize these critical processes, contributing significantly to the advancement of battery performance in electric vehicles.
Our study aligns with previous research on Q-learning in battery management, showcasing its applicability in various domains, including microgrids, modular multilevel converters (MMC) systems, electric vehicles, and residential energy systems. The fundamental principles of Q-learning, such as learning from experience, adaptability, policy optimization, and the state–action–reward framework, are discussed in our study. We emphasize how these principles make Q-learning a suitable and valuable tool for addressing the intricate challenges of EV battery management [85].

5.2. Theoretical Implications

The application of Q-learning in EV battery management, as demonstrated in our study, holds significant theoretical implications for the field of reinforcement learning and its integration into energy systems. The observed increase in the average rewards per episode, depicted in Figure 11, serves as a testament to the model’s ability to learn and adapt its strategies over time, resulting in enhanced battery management [86]. This finding aligns with the core principles of reinforcement learning, where an agent learns to make optimal decisions through interactions with its environment [82]. The comparative analysis underscores the superiority of Q-learning over traditional battery-management approaches. The marked increase in energy efficiency and the reduction in the battery degradation rate highlight the potential of Q-learning to address key challenges in EV battery management, such as extending the battery’s lifespan and improving its performance [82,83]. These results contribute to the growing body of literature that advocates for the integration of intelligent algorithms in energy systems to enhance their efficiency and sustainability [84]. Moreover, the reduction in the total operational costs illustrates the economic viability of employing Q-learning in battery management. This finding supports the theoretical notion that reinforcement learning algorithms can provide cost-effective solutions in complex decision-making scenarios, a concept that is increasingly relevant in the context of electric mobility [85,86].
The integration of Q-learning into EV battery management also has theoretical implications for the broader field of energy systems optimization. The adaptability and learning capabilities of Q-learning, as evidenced by the improved performance for longer episodes, suggest that reinforcement learning algorithms can effectively handle the dynamic and complex nature of energy management in electric vehicles [84]. This aligns with the ongoing research efforts to develop intelligent and adaptive energy systems that can respond to changing conditions and optimize their operations accordingly [83,84].
Our study contributes to the theoretical understanding of the application of reinforcement learning in energy systems, specifically in the context of EV battery management. The demonstrated effectiveness of Q-learning in optimizing battery performance, extending its lifespan, and reducing operational costs highlights the potential of reinforcement learning algorithms to address the challenges faced by electric mobility and contribute to the development of more efficient and sustainable energy systems.

5.3. Managerial Implications

The findings from our study on the application of Q-learning for EV battery management have several important managerial implications. These implications are particularly relevant for stakeholders in the EV industry, including manufacturers, policymakers, and energy service providers.
Firstly, the superior performance of Q-learning in optimizing battery performance, energy efficiency, and reducing operational costs suggests that adopting this technology could be a strategic advantage for EV manufacturers. By integrating Q-learning algorithms into their battery-management systems, manufacturers can enhance the appeal of their vehicles by offering improved range, longevity, and lower total cost of ownership [86].
Secondly, the evidence supporting the cost-effectiveness of Q-learning-based battery-management systems highlights the potential for reducing the overall cost of EV ownership. This could play a significant role in addressing one of the major barriers to EV adoption, which is the perception of high initial costs. Policymakers could leverage these findings to design incentives and subsidies that promote the adoption of intelligent battery-management systems in EVs, thereby accelerating the transition to electric mobility [1,83]. Thirdly, the adaptability and learning capabilities of Q-learning algorithms underscore the importance of investing in data analytics and machine learning expertise within the EV industry. Companies that develop competencies in these areas will be better positioned to innovate and lead in the development of next-generation battery-management systems. This could also spur collaborations between the automotive industry and tech companies, fostering a cross-sectoral ecosystem of innovation in electric mobility [1,83,84].
Thirdly, the adaptability and learning capabilities of Q-learning algorithms underscore the importance of investing in data analytics and machine learning expertise within the EV industry. Companies that develop competencies in these areas will be better positioned to innovate and lead in developing next-generation battery-management systems. This could also spur collaborations between the automotive industry and tech companies, fostering a cross-sectoral ecosystem of innovation in electric mobility [1,83,84].
Finally, the findings of this study have implications for energy service providers, particularly those involved in charging infrastructure. The ability of Q-learning-based systems to optimize charging schedules and patterns can contribute to more efficient use of the grid and reduce peak demand pressures. Energy providers could explore partnerships with EV manufacturers to develop smart-charging solutions that leverage the capabilities of Q-learning, enhancing the integration of EVs into the energy system [1,83,84,85].

5.4. Limitations and Future Research Directions

While the findings of our study are promising, several limitations warrant consideration:
1. Geographic and Scenario Specificity: This study’s focus on a particular geographic area or driving scenario may limit the generalizability of the findings. The results observed in this specific context might not fully translate to different regions or diverse driving conditions. Future research should aim to encompass a broader range of geographical areas and driving environments to validate the applicability of Q-learning-based battery-management systems across various contexts.
2. Scope of Parameters: The research primarily centers on a limited set of parameters related to battery performance, potentially overlooking other influential factors. Key aspects such as variations in battery chemistry, ambient temperature effects, and diverse driving behaviors were not extensively considered. Expanding the scope to include these additional factors could provide a more comprehensive understanding of Q-learning’s impact on battery management.
3. Dependence on Historical Data: The reliance on historical data for training the Q-learning model introduces potential biases and assumptions that may affect the accuracy and robustness of the predictions. Historical data might not fully capture the variability and dynamic nature of real-world conditions, which could limit the model’s performance under novel or changing scenarios. Incorporating real-time data and continuous model updates could address this limitation and enhance prediction accuracy.
4. Model Scalability: This study’s focus on a specific set of conditions and parameters raises questions about the scalability of the Q-learning model. While the model demonstrates promising results in the studied context, its performance in larger-scale or more complex scenarios remains uncertain. Further research is needed to evaluate the scalability and adaptability of Q-learning-based approaches to different scales and operational contexts.
5. Integration with Real-World Systems: The practical integration of Q-learning with existing EV battery-management systems and real-world applications was not extensively explored. Challenges related to seamless integration, system compatibility, and practical deployment need to be addressed to ensure the effective implementation of Q-learning-based strategies in commercial EV battery-management systems.
Acknowledging these limitations is crucial for guiding future research and improving the robustness of Q-learning applications in EV battery management. Addressing these challenges through expanded research and practical testing will contribute to a more comprehensive understanding of the potential and limitations of reinforcement learning in optimizing EV battery performance [1,83,84,85].
Furthermore, the integration of Q-learning with other intelligent systems, such as predictive maintenance or smart grid technologies, represents a promising avenue for future research. Investigating the synergy between Q-learning and these technologies could lead to more holistic and efficient energy management solutions for electric vehicles [86]. Finally, the ethical and societal implications of deploying reinforcement learning algorithms in EV battery management should be carefully considered. Ensuring transparency, accountability, and fairness in decision-making processes is essential to gaining public trust and acceptance of these technologies [86].
In conclusion, our study highlights the potential of Q-learning to revolutionize EV battery management by enhancing performance, extending the battery lifespan, and reducing operational costs. These findings underscore the need for further research to validate and refine these approaches in real-world settings, explore other reinforcement learning techniques, and integrate Q-learning with other intelligent systems. Addressing these limitations and future research directions will contribute to developing more efficient, sustainable, and cost-effective energy solutions for electric mobility.

6. Conclusions

The findings of this study underscore the transformative potential of Q-learning in EV battery management, marking a significant advancement in sustainable transportation. The application of Q-learning has demonstrated notable improvements in battery performance, including enhanced energy efficiency and reduced degradation rates. This adaptability to dynamically changing operating conditions, coupled with its ability to achieve substantial operational cost savings, positions Q-learning as a cornerstone in optimizing EV battery-management strategies [86]. However, this study also acknowledges the challenges ahead, such as data quality, model scalability, and seamless integration with real-world systems, which are crucial for widespread adoption [87].
The broader implications of this research extend beyond technological advancements, touching upon environmental, user-centric, and economic domains. Q-learning’s contribution to optimized energy use and reduced greenhouse gas emissions aligns with global efforts to combat climate change, promising a more sustainable future. From a user-centric perspective, the enhanced battery lifespan, lower operational costs, and improved driving range are likely to increase satisfaction among EV owners. Economically, the cost savings from reduced battery degradation enhance the viability of electric vehicles, potentially reshaping the competitive landscape against traditional internal combustion engine vehicles [88].
Data-driven approaches for EV battery management are crucial for sustainability. One approach is Q-learning, which uses reinforcement learning to optimize battery performance. By collecting data on driving patterns and behaviors, a Q-learning algorithm can be trained to make decisions that minimize fuel consumption, battery aging, and state of charge (SoC) sustainability penalty [73]. Another approach is sequential learning, which allows for the adaptation of previously learned knowledge to new tasks. This flexibility enables the accurate prediction of battery condition and state of charge, leading to effective battery management [74]. Additionally, deep learning techniques can be used for fault detection and classification of battery sensors and transmission data, enhancing the safety and dependability of battery-management systems [75]. These data-driven approaches, combined with real-time monitoring systems and IoT technology, enable efficient battery swapping and management, contributing to the sustainability of electric vehicles [76,77].
The implications of this research transcend the boundaries of technology, echoing across environmental, user-centric, and economic domains. From a global perspective, this study holds promise for a more sustainable future, as Q-learning contributes to reduced environmental impact through optimized energy use and decreased greenhouse gas emissions, aligning seamlessly with international efforts to combat climate change. On a user-centric level, enhanced battery lifespan, lower operational costs, and improved driving range promise increased satisfaction among EV owners. Furthermore, the economic viability of electric vehicles is bolstered by the cost savings stemming from reduced battery degradation, potentially reshaping the competitive landscape vis-à-vis traditional internal combustion engine vehicles [87,88].
Yet, it is essential to acknowledge the limitations inherent in this study. One notable limitation is the focus on a specific geographic area or driving scenario, which may restrict the generalizability of the findings to broader contexts. Additionally, the scope primarily revolves around a specific set of parameters, potentially overlooking other factors that could influence EV battery performance. Furthermore, the study’s reliance on historical data might introduce biases or assumptions that could affect the accuracy of the Q-learning model’s predictions. These limitations highlight the need for future research to encompass diverse environmental conditions and driving patterns for a more comprehensive understanding [89].
Looking forward, the recommendations for future research underscore the commitment to continuous improvement and innovation in EV battery management. The exploration of advanced machine learning techniques, such as Deep Q-Networks (DQN), hints at the untapped potential for even more adaptive and efficient strategies. The call for real-world fleet testing on a larger scale speaks to the commitment to practicality and the need to evaluate the long-term impacts and scalability of Q-learning-based approaches. Moreover, the integration of Q-learning with renewable energy sources, like solar panels, signifies a broader commitment to clean energy solutions within the EV ecosystem. Finally, delving into the realm of human–machine interaction opens exciting possibilities for customization, enabling drivers to tailor preferences and learn from their unique driving patterns [90].
This study on the application of Q-learning for EV battery management has significant implications for the EV industry, particularly for manufacturers, policymakers, and energy service providers. Q-learning’s ability to optimize battery performance and reduce costs provides a strategic advantage for manufacturers by enhancing vehicle appeal through improved range and longevity. This technology’s cost-effectiveness also supports policy initiatives promoting electric mobility. Additionally, the adaptability of Q-learning underscores the importance of investment in data analytics and machine learning, fostering innovation and collaboration within the industry [91,92]. Furthermore, its potential to optimize charging schedules presents opportunities for energy providers to partner with EV manufacturers in developing smart charging solutions [91,92,93].
Finally, this research goes beyond being a technical breakthrough—it represents a paradigm shift in the evolution of electric vehicle technology. By offering a compelling and promising avenue for sustainable transportation, it lays the groundwork for a future where environmental responsibility and cutting-edge technology converge seamlessly in the automotive industry for sustainability [94,95].

Author Contributions

Conceptualization, P.S.; research design, P.S.; literature review, P.S. and P.J.; methodology, P.S. and P.J.; algorithms, P.S. and P.J.; software, P.S. and P.J.; validation, P.S. and P.J.; formal analysis, P.S. and P.J.; investigation, P.S. and P.J.; resources, P.S.; data curation, P.J.; writing—original draft preparation, P.S. and P.J.; writing—review and editing, P.S. and P.J.; visualization, P.S.; supervision, P.S.; project administration, P.S.; funding acquisition, P.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Suan Dusit University under the Ministry of Higher Education, Science, Research and Innovation, Thailand, grant number FF67—innovative process for inspiring chefs to become chef innovators for supporting the tourism and hospitality industry to Michelin standards.

Institutional Review Board Statement

This study was conducted in accordance with ethical guidelines and approved by the Ethics Committee of Suan Dusit University (SDU-RDI-SHS 2023-043, 1 June 2023) for studies involving humans.

Informed Consent Statement

This article does not contain any studies involving human participants performed by any of the authors.

Data Availability Statement

The data presented in this study are available upon request from the corresponding author.

Acknowledgments

The authors wish to express their gratitude to the Hub of Talent in Gastronomy Tourism Project (N34E670102), funded by the National Research Council of Thailand (NRCT), for facilitating the research collaboration that contributed to this study. We also extend our thanks to Suan Dusit University and King Mongkut’s University of Technology Thonburi for their research support and the network of researchers in the region where this research was conducted. Additionally, we are grateful to the Tourism Authority of Thailand (TAT) and the Rayong Smart City Project for providing essential data in the study areas.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Suanpang, P.; Jamjuntr, P. Optimizing Electric Vehicle Charging Recommendation in Smart Cities: A Multi-Agent Reinforcement Learning Approach. World Electr. Veh. J. 2024, 15, 67. [Google Scholar] [CrossRef]
  2. Suanpang, P.; Jamjuntr, P.; Kaewyong, P.; Niamsorn, C.; Jermsittiparsert, K. An Intelligent Recommendation for Intelligently Accessible Charging Stations: Electronic Vehicle Charging to Support a Sustainable Smart Tourism City. Sustainability 2023, 15, 455. [Google Scholar] [CrossRef]
  3. Li, Y.; Tao, J.; Xie, L.; Zhang, R.; Ma, L.; Qiao, Z. Enhanced Q-learning for real-time hybrid electric vehicle energy management with deterministic rule. Meas. Control. 2020, 53, 1493–1503. [Google Scholar] [CrossRef]
  4. Khamis, M.A.H.; Hassanien, A.E.; Salem, A.E.K. Electric Vehicle Charging Infrastructure Optimization: A Comprehensive Review. IEEE Access 2020, 8, 23676–23692. [Google Scholar]
  5. Kim, N.; Kim, J.C.D.; Lee, B. Adaptive Loss Reduction Charging Strategy Considering Variation of Internal Impedance of Lithium-Ion Polymer Batteries in Electric Vehicle Charging Systems. In Proceedings of the 2016 IEEE Applied Power Electronics Conference and Exposition (APEC), Long Beach, CA, USA, 20–24 March 2016; pp. 1273–1279. [Google Scholar] [CrossRef]
  6. Sedano, J.; Chira, C.; Villar, J.R.; Ambel, E.M. An Intelligent Route Management System for Electric Vehicle Charging. Integr. Comput. Aided Eng. 2013, 20, 321–333. [Google Scholar] [CrossRef]
  7. Market & Market. Electric Vehicle Market. Available online: https://www.marketsandmarkets.com/Market-Reports/electric-vehicle-market-209371461.html (accessed on 9 July 2024).
  8. Brenna, M.; Foiadelli, F.; Leone, C.; Longo, M. Electric Vehicles Charging Technology Review and Optimal Size Estimation. J. Electr. Eng. Technol. 2020, 15, 2539–2552. [Google Scholar] [CrossRef]
  9. Rizvi, S.A.A.; Xin, A.; Masood, A.; Iqbal, S.; Jan, M.U.; Rehman, H. Electric vehicles and their impacts on integration into power grid: A review. In Proceedings of the 2nd IEEE Conference on Energy Internet and Energy System Integration (EI2), Beijing, China, 20–22 October 2018. [Google Scholar] [CrossRef]
  10. Papadopoulos, P.; Cipcigan, L.M.; Jenkins, N. Distribution networks with electric vehicles. In Proceedings of the 44th International Universities Power Engineering Conference (UPEC), Glasgow, UK, 1–4 September 2009. [Google Scholar]
  11. Cao, X.; Li, Y.; Zhang, C.; Wang, J. Reinforcement learning-based battery management system for electric vehicles: A comparative study. J. Energy Storage 2021, 42, 102383. [Google Scholar] [CrossRef]
  12. Xiong, R.; Sun, F.; Chen, Z.; He, H. A data-driven multi-scale extended Kalman filtering based parameter and state estimation approach of lithium-ion polymer battery in electric vehicles. Appl. Energ. 2014, 113, 463–476. [Google Scholar] [CrossRef]
  13. Al-Alawi, B.M.; Bradley, T.H. Total cost of ownership, payback, and consumer preference modeling of plug-in hybrid electric vehicles. Appl. Energ. 2013, 103, 488–506. [Google Scholar] [CrossRef]
  14. Yang, K.; Zhang, L.; Zhang, Z.; Yu, H.; Wang, W.; Ouyang, M.; Zhang, C.; Sun, Q.; Yan, X.; Yang, S.; et al. Battery State of Health Estimate Strategies: From Data Analysis to End-Cloud Collaborative Framework. Batteries 2023, 9, 351. [Google Scholar] [CrossRef]
  15. Mishra, S.; Swain, S.C. Utilizing the Unscented Kalman Filter for Estimating State of Charge in Lithium-Ion Batteries of Electric Vehicles. In Proceedings of the Fifteenth Annual IEEE Green Technologies (GreenTech) Conference, Denver, CO, USA, 19–21 April 2023; pp. 1–6. [Google Scholar]
  16. Topan, P.A.; Ramadan, M.N.; Fathoni, G.; Cahyadi, A.I.; Wahyunggoro, O. State of Charge (SOC) and State of Health (SOH) estimation on lithium polymer battery via Kalman filter. In Proceedings of the 2016 2nd International Conference on Science and Technology-Computer (ICST), Yogyakarta, Indonesia, 27–28 October 2016; pp. 93–96. [Google Scholar] [CrossRef]
  17. Nath, A.; Rather, Z.; Mitra, I.; Srinivasan, L. Multi-Criteria Approach for Identification and Ranking of Key Interventions for Seamless Adoption of Electric Vehicle Charging Infrastructure. IEEE Trans. Veh. Technol. 2023, 72, 8697–8708. [Google Scholar] [CrossRef]
  18. Karnehm, D.; Pohlmann, S.; Neve, A. State-of-Charge (SoC) Balancing of Battery Modular Multilevel Management (BM3) Converter using Q-Learning. In Proceedings of the 2023 IEEE Green Technologies Conference (GreenTech), Denver, CO, USA, 19–21 April 2023; pp. 107–111. [Google Scholar] [CrossRef]
  19. Shateri, M. Privacy-Cost Management in Smart Meters: Classical vs Deep Q-Learning with Mutual Information. In Proceedings of the 2023 IEEE 11th International Conference on Smart Energy Grid Engineering (SEGE), Oshawa, ON, Canada, 19–21 April 2023; pp. 109–113. [Google Scholar] [CrossRef]
  20. Ahmadian, S.; Tahmasbi, M.; Abedi, R. Q-learning based control for energy management of series-parallel hybrid vehicles with balanced fuel consumption and battery life. Energy AI 2023, 11, 100217. [Google Scholar] [CrossRef]
  21. Corinaldesi, C.; Lettner, G.; Schwabeneder, D.; Ajanovic, A.; Auer, H. Impact of Different Charging Strategies for Electric Vehicles in an Austrian Office Site. Energies 2020, 13, 5858. [Google Scholar] [CrossRef]
  22. Kene, R.O.; Olwal, T.O. Energy Management and Optimization of Large-Scale Electric Vehicle Charging on the Grid. World Electr. Veh. J. 2023, 14, 95. [Google Scholar] [CrossRef]
  23. Li, Y.; Wang, J.; Zhang, C.; Sun, C. Q-learning-based battery management system with real-time state-of-charge estimation for electric vehicles. IEEE Trans. Veh. Technol. 2021, 70, 6931–6944. [Google Scholar]
  24. Zhang, W.; Hu, X.; Sun, C.; Zou, C. A Q-learning-based battery management system with adaptive state-of-charge estimation for electric vehicles. IEEE Trans. Ind. Electron. 2022, 69, 11718–11728. [Google Scholar]
  25. Suanpang, P.; Jamjuntr, P.; Jermsittiparsert, K.; Kaewyong, P. Tourism Service Scheduling in Smart City Based on Hybrid Genetic Algorithm Simulated Annealing Algorithm. Sustainability 2022, 14, 16293. [Google Scholar] [CrossRef]
  26. International Energy Agency (IEA). EV Charging in Thailand Market Overview 2023–2027. Available online: https://www.reportlinker.com/market-report/Electric-Vehicle/726537/Electric_Vehicle_Charging (accessed on 9 July 2024).
  27. Zhang, W.; Wang, J.; Sun, C.; Ecker, M. A review of reinforcement learning for battery management in electric vehicles. J. Power Sources 2020, 459, 228025. [Google Scholar] [CrossRef]
  28. Alanazi, F. Electric Vehicles: Benefits, Challenges, and Potential Solutions for Widespread Adaptation. Appl. Sci. 2023, 13, 6016. [Google Scholar] [CrossRef]
  29. Turrentine, T.; Kurani, K. Consumer Considerations in the Transition to Electric Vehicles: A Review of the Research Literature. Energy Policy 2019, 127, 14–27. [Google Scholar] [CrossRef]
  30. Ahmed, A.; El Baset, A.; El Halim, A.; Ehab, E.H.; Bayoumi, W.; El-Khattam, A.; Ibrahim, A. Electric vehicles: A review of their components and technologies. Int. J. Power Electron. Drive Syst. 2022, 13, 2041–2061. [Google Scholar] [CrossRef]
  31. Szumska, E.; Jurecki, R. Technological Developments in Vehicles with Electric Drive. Combust. Engines 2023, 194, 38–47. [Google Scholar] [CrossRef]
  32. Murali, N.; Mini, V.P.; Ushakumari, S. Electric Vehicle Market Analysis and Trends. In Proceedings of the 2022 IEEE 19th India Council International Conference (INDICON), Kochi, India, 24–26 November 2022. [Google Scholar] [CrossRef]
  33. Ouramdane, O.; Elbouchikhi, E.; Amirat, Y.; Le Gall, F.; Sedgh Gooya, E. Home Energy Management Considering Renewable Resources, Energy Storage, and an Electric Vehicle as a Backup. Energies 2022, 15, 2830. [Google Scholar] [CrossRef]
  34. Kostenko, A. Overview of European Trends in Electric Vehicle Implementation and the Influence on the Power System. Syst. Res. Energy 2022, 2022, 62–71. [Google Scholar] [CrossRef]
  35. Boonchunone, S.; Nami, M.; Krommuang, A.; Suwunnamek, O. Exploring the Effects of Perceived Values on Consumer Usage Intention for Electric Vehicle in Thailand: The Mediating Effect of Satisfaction. Acta Logist. 2023, 10, 151–164. [Google Scholar] [CrossRef]
  36. Butler, D.; Mehnen, J. Challenges for the Adoption of Electric Vehicles in Thailand: Potential Impacts, Barriers, and Public Policy Recommendations. Sustainability 2023, 15, 9470. [Google Scholar] [CrossRef]
  37. Chinda, T. Manufacturer and Consumer’s Perceptions towards Electric Vehicle Market in Thailand. J. Ind. Integr. Manag. 2023. [Google Scholar] [CrossRef]
  38. Dokrak, I.; Rakwichian, W.; Rachapradit, P.; Thanarak, P. The Business Analysis of Electric Vehicle Charging Stations to Power Environmentally Friendly Tourism: A Case Study of the Khao Kho Route in Thailand. Int. J. Energy Econ. Policy 2022, 12, 102–111. [Google Scholar] [CrossRef]
  39. Bhovichitra, P.; Shrestha, A. The Impact of Video Marketing on Social Media Platforms on the Millennial’s Purchasing Intention toward Electric Vehicles. J. Econ. Finance Manag. Stud. 2022, 5, 3726–3730. [Google Scholar] [CrossRef]
  40. Qi, S.; Cheng, Y.; Li, Z.; Wang, J.; Li, H.; Zhang, C. Advanced Deep Learning Techniques for Battery Thermal Management in New Energy Vehicles. Energies 2024, 17, 4132. [Google Scholar] [CrossRef]
  41. Li, C.; Hu, X.; Zhang, S.; Wang, Y.; Duan, Q. Electric Vehicle Battery. U.S. Patent US20160202640A1, 8 January 2016. Available online: https://patents.google.com/patent/US20160202640A1 (accessed on 9 July 2024).
  42. Xu, G.; Tang, M.; Cai, W.; Tan, Y. Electric Vehicle Battery Pack. U.S. Patent US20150093510A1, 2 April 2015. Available online: https://patents.google.com/patent/US20150093510A1 (accessed on 9 July 2024).
  43. Dhameja, S. Electric Vehicle Batteries, 1st ed.; Newnes: Oxford, UK, 2002. [Google Scholar] [CrossRef]
  44. Link, S.; Neef, C. Trends in Automotive Battery Cell Design: A Statistical Analysis of Empirical Data. Batteries 2023, 9, 261. [Google Scholar] [CrossRef]
  45. Liu, C.; Bi, C. Current Situation and Trend of Electric Vehicle Battery Business-Take CATL as an example. Tech. Soc. Sci. J. 2022, 38, 7975. [Google Scholar] [CrossRef]
  46. Dimitriadou, K.; Hatzivasilis, G.; Ioannidis, S. Current Trends in Electric Vehicle Charging Infrastructure; Opportunities and Challenges in Wireless Charging Integration. Energies 2023, 16, 2057. [Google Scholar] [CrossRef]
  47. Zhao, J.; Burke, A. Electric Vehicle Batteries: Status and Perspectives of Data-Driven Diagnosis and Prognosis. Batteries 2022, 8, 142. [Google Scholar] [CrossRef]
  48. Garg, V.K.; Kumar, D. A review of electric vehicle technology: Architectures, battery technology and its management system, relevant standards, application of artificial intelligence, cyber security, and interoperability challenges. IET Electr. Syst. Transp. 2023, 3, 2083. [Google Scholar] [CrossRef]
  49. Mehar, D.; Varshney, P.; Saini, D. A Review on Battery Technologies and Its Challenges in Electrical Vehicle. In Proceedings of the 2023 3rd International Conference on Energy, Systems, and Applications (ICESA), Pune, India, 4–6 December 2023; p. 63067. [Google Scholar] [CrossRef]
  50. Ahasan, H.A.K.M.; Masrur, H.; Islam, M.S.; Sarker, S.K.; Hossain, M.S. Lithium-Ion Battery Management System for Electric Vehicles: Constraints, Challenges, and Recommendations. Batteries 2023, 9, 152. [Google Scholar] [CrossRef]
  51. Roy, H.; Gupta, N.; Patel, S.; Mistry, B. Global Advancements and Current Challenges of Electric Vehicle Batteries and Their Prospects: A Comprehensive Review. Sustainability 2022, 14, 16684. [Google Scholar] [CrossRef]
  52. Santhira Sekeran, M.; Živadinović, M.; Spiliopoulou, M. Transferability of a Battery Cell End-of-Life Prediction Model Using Survival Analysis. Energies 2022, 15, 2930. [Google Scholar] [CrossRef]
  53. Ranjan, R.; Yadav, R. An Overview and Future Reflection of Battery Management Systems in Electric Vehicles. J. Energy Storage 2023, 35, 567. [Google Scholar] [CrossRef]
  54. Thilak, K.; Kumari, M.S.; Manogaran, G.; Khuman, R.; Janardhana, M.; Hemanth, V.K. An Investigation on Battery Management System for Autonomous Electric Vehicles. IEEE Trans. Ind. Appl. 2023, 56, 5341. [Google Scholar]
  55. Bhushan, N.; Desai, K.; Suthar, R. Overview of Model- and Non-Model-Based Online Battery Management Systems for Electric Vehicle Applications: A Comprehensive Review of Experimental and Simulation Studies. Sustainability 2022, 14, 15912. [Google Scholar] [CrossRef]
  56. Hossain, M.S.; Islam, M.T.; Ali, A.B.; Abdullah-Al-Mamun, M.; Hassan, M.R.; Nizam, M.T.; Jayed, M.A. Smart Battery Management Technology in Electric Vehicle Applications: Analytical and Technical Assessment toward Emerging Future Directions. Batteries 2022, 8, 219. [Google Scholar] [CrossRef]
  57. Kumar, R.S.; Kamal, A.H.; Selvam, S.P.; Rajesh, M. Battery Management System for Renewable E-Vehicle. IEEE Trans. Energy Convers. 2023, 24, 1092. [Google Scholar] [CrossRef]
  58. Suanpang, P.; Jamjuntr, P. Machine Learning Models for Solar Power Generation Forecasting in Microgrid Application Implications for Smart Cities. Sustainability 2024, 16, 6087. [Google Scholar] [CrossRef]
  59. Ali, H.; Tsegaye, D.; Abera, M. Dual-Layer Q-Learning Strategy for Energy Management of Battery Storage in Grid-Connected Microgrids. Energies 2023, 16, 1334. [Google Scholar] [CrossRef]
  60. Ye, Y.; Zhao, J.; Li, M.; Zhu, Y. Reinforcement Learning-Based Energy Management System Enhancement Using Digital Twin for Electric Vehicles. In Proceedings of the 2022 IEEE Vehicle Power and Propulsion Conference (VPPC), Gijón, Spain, 26–29 October 2022; pp. 343–347. [Google Scholar] [CrossRef]
  61. Xu, B.; Wang, L.; Li, X.; He, H.; Sun, Y. Hierarchical Q-learning network for online simultaneous optimization of energy efficiency and battery life of the battery/ultracapacitor electric vehicle. J. Energy Storage 2022, 52, 103925. [Google Scholar] [CrossRef]
  62. Muriithi, G.; Chowdhury, S.P. Deep Q-network application for optimal energy management in a grid-tied solar PV-Battery microgrid. J. Eng. 2022, 21, 12128. [Google Scholar] [CrossRef]
  63. Sousa, T.J.C.; Pedrosa, D.; Monteiro, V.; Afonso, J.L. A Review on Integrated Battery Chargers for Electric Vehicles. Energies 2022, 15, 2756. [Google Scholar] [CrossRef]
  64. Suanpang, P.; Jamjuntr, P.; Jermsittiparsert, K.; Kaewyong, P. Autonomous Energy Management by Applying Deep Q-Learning to Enhance Sustainability in Smart Tourism Cities. Energies 2022, 15, 1906. [Google Scholar] [CrossRef]
  65. Qi, S.; Zhao, Y.; Li, Z.; Zhao, Y. Control Strategy of Energy Storage Device in Distribution Network Based on Reinforcement Learning Q-learning Algorithm. In Proceedings of the 2022 International Conference on Control, Automation and Systems Integration Technology (ICCASIT), Tokyo, Japan, 27 November–1 December 2022; p. 6984. [Google Scholar] [CrossRef]
  66. Rokh, S.B.; Soltani, M.; Eftekhari, S.; Rakhshan, F.; Hosseini, S. Real-Time Optimization of Microgrid Energy Management Using Double Deep Q-Network. In Proceedings of the 2023 IEEE International Conference on Smart Grid and Clean Energy Technologies (ICSGCET), Shanghai, China, 13–15 October 2023; p. 66355. [Google Scholar] [CrossRef]
  67. Xiao, G.; Chen, Q.; Xiao, P.; Zhang, L.; Rong, Q. Multiobjective Optimization for a Li-Ion Battery and Supercapacitor Hybrid Energy Storage Electric Vehicle. Energies 2022, 15, 2821. [Google Scholar] [CrossRef]
  68. Kong, H.; Wu, Z.; Luo, Y.; Xu, J.; Huang, X. Energy management strategy for electric vehicles based on deep Q-learning using Bayesian optimization. Neural Comput. Appl. 2020, 32, 4556–4574. [Google Scholar] [CrossRef]
  69. Bignold, A.; Parisini, T.; Mangini, M. A conceptual framework for externally-influenced agents: An assisted reinforcement learning review. J. Ambient. Intell. Humaniz. Comput. 2023, 21, 3489. [Google Scholar] [CrossRef]
  70. Hu, X.; Liang, C.; Xie, J.; Li, Z. A survey of reinforcement learning applications in battery management systems. IEEE Trans. Intell. Transp. Syst. 2021, 22, 11119. [Google Scholar]
  71. Vardopoulos, I.; Papoui-Evangelou, M.; Nosova, B.; Salvati, E. Smart ‘Tourist Cities’ Revisited: Culture-Led Urban Sustainability and the Global Real Estate Market. Sustainability 2023, 15, 4313. [Google Scholar] [CrossRef]
  72. Xu, R. Framework for Building Smart Tourism Big Data Mining Model for Sustainable Development. Sustainability 2023, 15, 5162. [Google Scholar] [CrossRef]
  73. Madeira, C.; Rodrigues, P.; Gómez-Suárez, M. A Bibliometric and Content Analysis of Sustainability and Smart Tourism. Urban Sci. 2023, 7, 33. [Google Scholar] [CrossRef]
  74. Lata, S.; Jasrotia, A.; Sharma, S. Sustainable development in tourism destinations through smart cities: A case of urban planning in jammu city. Enlightening Tour. Pathmaking J. 2022, 12, 6911. [Google Scholar] [CrossRef]
  75. Zearban, M.; Abdelaziz, M.; Abdelwahab, M. The Temperature Effect on Electric Vehicle’s Lithium-Ion Battery Aging Using Machine Learning Algorithm. Eng. Proc. 2024, 70, 53. [Google Scholar] [CrossRef]
  76. Suanpang, P.; Pothipassa, P.; Jermsittiparsert, K.; Netwong, T. Integration of Kouprey-Inspired Optimization Algorithms with Smart Energy Nodes for Sustainable Energy Management of Agricultural Orchards. Energies 2022, 15, 2890. [Google Scholar] [CrossRef]
  77. Suanpang, P.; Pothipassa, P.; Jittithavorn, C. Blockchain of Things (BoT) Innovation for Smart Tourism. Int. J. Tour. Res. 2024, 26, e2606. [Google Scholar] [CrossRef]
  78. Zhao, J.; Liu, S.; Yin, Z.; Wang, S. Deep reinforcement learning for battery management system of electric vehicles: A review. CSEE J. Power Energy Syst. 2022, 8, 1334. [Google Scholar]
  79. Chan, C.K.; Wong, Y.; Sun, Q.; Ng, B. Optimizing Thermal Management System in Electric Vehicle Battery Packs for Sustainable Transportation. Sustainability 2023, 15, 11822. [Google Scholar] [CrossRef]
  80. Kosuru, V.S.R.; Balakrishna, K. A Smart Battery Management System for Electric Vehicles Using Deep Learning-Based Sensor Fault Detection. World Electr. Veh. J. 2023, 14, 101. [Google Scholar] [CrossRef]
  81. Liu, H.; Wang, Z.; Liu, L. A comprehensive review of energy management strategies for electric vehicles. Renew. Sustain. Energy Rev. 2020, 119, 109595. [Google Scholar] [CrossRef]
  82. Sutton, R.S.; Barto, A.G. Reinforcement Learning: An Introduction; MIT Press: Cambridge, MA, USA, 2018. [Google Scholar] [CrossRef]
  83. Wang, Y.; Zhang, X.; Chen, Z. A review of battery management technologies and algorithms for electric vehicle applications. J. Clean. Prod. 2019, 215, 1279–1291. [Google Scholar] [CrossRef]
  84. Zhang, L.; Wang, H.; You, S. A review of machine learning for energy management in electric vehicles: Applications, challenges and opportunities. Appl. Energy 2021, 281, 115966. [Google Scholar] [CrossRef]
  85. Lombardo, T.; Duquesnoy, M.; El-Bouysidy, H.; Årén, F.; Gallo-Bueno, A.; Jørgensen, P.B.; Bhowmik, A.; Demortière, A.; Ayerbe, E.; Alcaide, F.; et al. Artificial Intelligence Applied to Battery Research: Hype or Reality? Chem. Rev. 2022, 122, 10899–10969. [Google Scholar] [CrossRef]
  86. Lombardo, A.; Nguyen, P.; Strbac, G. Q-Learning for Battery Management in Electric Vehicles: Theoretical and Experimental Analysis. Energies 2022, 15, 1234. [Google Scholar]
  87. Arunadevi, R.; Saranya, S. Battery Management System for Analysing Accurate Real Time Battery Condition using Machine Learning. Int. J. Comput. Sci. Mob. Comput. 2023, 12, 1–10. [Google Scholar] [CrossRef]
  88. Ezzouhri, A.; Charouh, Z.; Ghogho, M.; Guennoun, Z. A Data-Driven-based Framework for Battery Remaining Useful Life Prediction. IEEE Access 2023, 11, 76142–76155. [Google Scholar] [CrossRef]
  89. Tang, X.; Zhang, J.; Pi, D.; Lin, X.; Grzesiak, L.M.; Hu, X. Battery Health-Aware and Deep Reinforcement Learning-Based Energy Management for Naturalistic Data-Driven Driving Scenarios. IEEE Trans. Transp. Electrif. 2021, 6, 3417. [Google Scholar] [CrossRef]
  90. Swarnkar, R.; Harikrishnan, R.; Thakur, P.; Singh, G. Electric Vehicle Lithium-ion Battery Ageing Analysis under Dynamic Condition: A Machine Learning Approach. Africa Res. J. 2023, 24, 2788. [Google Scholar] [CrossRef]
  91. Baberwal, K.; Yadav, A.K.; Saini, V.K.; Lamba, R.; Kumar, R. Data Driven Energy Management of Residential PV-Battery System Using Q-Learning. In Proceedings of the 2023 IEEE International Conference on Recent Advances in Systems Science and Engineering (RASSE), Kerala, India, 8–11 November 2023; pp. 1–6. [Google Scholar] [CrossRef]
  92. Matare, T.N.; Folly, K.A. Energy Management System of an Electric Vehicle Charging Station Using Q-Learning and Artificial Intelligence. In Proceedings of the 2023 IEEE PES GTD International Conference and Exposition (GTD), Istanbul, Turkey, 22–25 May 2023; pp. 246–250. [Google Scholar] [CrossRef]
  93. Bennehalli, B.; Singh, L.; Stephen, S.D.; Venkata Prasad, P.; Mallala, B. Machine Learning Approach to Battery Management and Balancing Techniques for Enhancing Electric Vehicle Battery Performance. J. Electr. Syst. 2024, 20, 885–892. [Google Scholar]
  94. Huang, J.; Wang, P.; Liu, Y. Emerging Trends and Market Dynamics in the Electric Vehicle Industry. Energy Rep. 2023, 8, 11504–11529. [Google Scholar]
  95. Lee, J.; Park, S. Reducing Grid Stress with Adaptive MARL for EV Charging. IEEE Trans. Ind. Electron. 2023, 69, 4931–4940. [Google Scholar]
Figure 1. EV car components.
Figure 1. EV car components.
Sustainability 16 07180 g001
Figure 2. Q-learning in battery management.
Figure 2. Q-learning in battery management.
Sustainability 16 07180 g002
Figure 3. System flowchart.
Figure 3. System flowchart.
Sustainability 16 07180 g003
Figure 4. The stage of charge over time.
Figure 4. The stage of charge over time.
Sustainability 16 07180 g004
Figure 5. Temperature and humidity.
Figure 5. Temperature and humidity.
Sustainability 16 07180 g005
Figure 6. Voltage and current data.
Figure 6. Voltage and current data.
Sustainability 16 07180 g006
Figure 7. Charting station at difference location.
Figure 7. Charting station at difference location.
Sustainability 16 07180 g007
Figure 8. EV charging station locations in Bangkok, Thailand.
Figure 8. EV charging station locations in Bangkok, Thailand.
Sustainability 16 07180 g008
Figure 9. Driving-pattern data.
Figure 9. Driving-pattern data.
Sustainability 16 07180 g009
Figure 10. Charging station map with GPS route.
Figure 10. Charging station map with GPS route.
Sustainability 16 07180 g010
Figure 11. Average rewards per episode.
Figure 11. Average rewards per episode.
Sustainability 16 07180 g011
Figure 12. Energy efficiency comparison.
Figure 12. Energy efficiency comparison.
Sustainability 16 07180 g012
Figure 13. Energy efficiency comparison of Q-learning and traditional battery management.
Figure 13. Energy efficiency comparison of Q-learning and traditional battery management.
Sustainability 16 07180 g013
Figure 14. Average battery degradation rate comparison.
Figure 14. Average battery degradation rate comparison.
Sustainability 16 07180 g014
Figure 15. Battery degradation rate comparison.
Figure 15. Battery degradation rate comparison.
Sustainability 16 07180 g015
Figure 16. Performance heatmap of Q-learning algorithm for EV battery management across temperature and humidity conditions.
Figure 16. Performance heatmap of Q-learning algorithm for EV battery management across temperature and humidity conditions.
Sustainability 16 07180 g016
Figure 17. Performance heatmap of traditional approach for EV battery management across temperature and humidity conditions.
Figure 17. Performance heatmap of traditional approach for EV battery management across temperature and humidity conditions.
Sustainability 16 07180 g017
Table 1. Summary of Q-learning vs. traditional battery management.
Table 1. Summary of Q-learning vs. traditional battery management.
Q-learningTraditional Approach
Average Energy Efficiency (%)92.588.3
Average Battery Degradation Rate (%)0.81.5
Total Operational Cost (THB)45205850
Table 2. Factors and impact on computational complexity.
Table 2. Factors and impact on computational complexity.
FactorDescriptionData Impact on Computational Complexity and Scalability
State and Action Space SizeThe size of the state and action spaces directly influence the complexity of the Q-learning algorithm. As the number of states and actions increases, the Q-table grows.100–10,000 states,
4–100 actions
Larger spaces lead to a larger Q-table, requiring more memory and computation for updates. This can significantly impact scalability for very complex environments.
Convergence RateThe speed at which the algorithm learns the optimal policy.100–10,000 iterations Slower convergence necessitates more iterations, increasing computation time. This becomes more critical in complex environments where learning takes longer.
Memory RequirementsThe amount of memory needed to store the Q-values for all state–action pairs.10 MB–1 GB memory Larger state and action spaces lead to increased memory demands. This might become a bottleneck for problems with extensive state representations.
Larger State SpacesAs the state space grows, more computational resources and time are needed to explore and update Q-values.10,000–1,000,000 states Higher complexity often corresponds to larger state spaces, potentially pushing the limits of Q-learning’s scalability due to the sheer number of Q-values to manage.
More Frequent UpdatesEnvironments with frequent updates to Q-values increase computational complexity due to continuous exploration and updates.Updates every episode–every step May increase computational complexity, particularly in dynamic or rapidly changing environments.
Table 3. Resource consumption.
Table 3. Resource consumption.
StageResourceMeasurementData Impact
TrainingCPUAverage CPU usage70–80%High CPU usage during training can affect other processes running on the system.
MemoryPeak memory usage2–4 GBHigh memory usage can lead to memory exhaustion and system slowdowns.
InferenceCPUAverage CPU usage20–30%CPU usage during inference should be monitored to ensure real-time performance.
MemoryAverage memory usage100–200 MBMemory usage during inference is lower, making it suitable for deployment on devices with limited processing power.
Table 4. Algorithmic efficiency.
Table 4. Algorithmic efficiency.
FactorDescriptionDataImpact on Computational Complexity and Scalability
State and Action Space SizeThe size of the state and action spaces directly influences the complexity of the Q-learning algorithm. As the number of states and actions increases, the Q-table grows.100–10,000 states, 4–100 actions Larger spaces lead to a larger Q-table, requiring more memory and computation for updates. This can significantly impact scalability for very complex environments.
Convergence RateThe speed at which the algorithm learns the optimal policy.100–10,000 iterations Slower convergence necessitates more iterations, increasing computation time. This becomes more critical in complex environments where learning takes longer.
Memory RequirementsThe amount of memory needed to store the Q-values for all state–action pairs.10 MB–1 GB memory Larger state and action spaces lead to increased memory demands. This might become a bottleneck for problems with extensive state representations.
Larger State SpacesAs the state space grows, more computational resources and time are needed to explore and update Q-values.10,000–1,000,000 states Higher complexity often corresponds to larger state spaces, potentially pushing the limits of Q-learning’s scalability due to the sheer number of Q-values to manage.
More Frequent UpdatesEnvironments with frequent updates to Q-values increase computational complexity due to continuous exploration and updates.Updates every episode—every step May increase computational complexity, particularly in dynamic or rapidly changing environments.
Table 5. Performance metrics used for Q-learning.
Table 5. Performance metrics used for Q-learning.
Performance MetricDescription
Training Convergence TimeThe time taken for the Q-learning algorithm to achieve an acceptable level of performance during the training phase. This could be measured by observing the average reward achieved by the agent over time or by tracking the change in Q-values.
Inference LatencyThe time delay between receiving input from the environment (e.g., state of the game) and producing an output (e.g., action to be taken) during the deployment stage. This is crucial for real-time applications where quick decisions are necessary.
Overall Computational EfficiencyA combined measure of the algorithm’s efficiency considering both training and inference stages. This could involve metrics like total training time, memory usage during training and inference, and energy consumption (if relevant).
Table 6. Comparison of computational time for Q-learning and alternative approaches across different problem sizes and environmental complexities.
Table 6. Comparison of computational time for Q-learning and alternative approaches across different problem sizes and environmental complexities.
ConditionProblem SizeEnvironmental ComplexityQ-learning Computational Time (Estimated Range)Time
Small problem sizeSmallLow1–10 s0.1–1 s
Medium problem sizeMediumModerate10–100 s1–10 s
Large problem sizeLargeHigh100–1000 s (or more)10–100 s (or more)
Table 7. Computational time for Q-learning and alternative approaches.
Table 7. Computational time for Q-learning and alternative approaches.
ConditionProblem SizeEnvironmental
Complexity
Q-learning Time (Estimated Range)Alternative Approach Time (Estimated Range)
ConditionSmallLow1–10 s0.1–1 s
Medium problem sizeMediumModerate10–100 s1–10 s
Large problem sizeLargeHigh100–1000 s (or more)10–100 s (or more)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Suanpang, P.; Jamjuntr, P. Optimal Electric Vehicle Battery Management Using Q-learning for Sustainability. Sustainability 2024, 16, 7180. https://doi.org/10.3390/su16167180

AMA Style

Suanpang P, Jamjuntr P. Optimal Electric Vehicle Battery Management Using Q-learning for Sustainability. Sustainability. 2024; 16(16):7180. https://doi.org/10.3390/su16167180

Chicago/Turabian Style

Suanpang, Pannee, and Pitchaya Jamjuntr. 2024. "Optimal Electric Vehicle Battery Management Using Q-learning for Sustainability" Sustainability 16, no. 16: 7180. https://doi.org/10.3390/su16167180

APA Style

Suanpang, P., & Jamjuntr, P. (2024). Optimal Electric Vehicle Battery Management Using Q-learning for Sustainability. Sustainability, 16(16), 7180. https://doi.org/10.3390/su16167180

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop