Next Article in Journal
Sector Coupling and Migration towards Carbon-Neutral Power Systems
Previous Article in Journal
Influence of Thermochemical Conversion Technologies on Biochar Characteristics from Extensive Grassland for Safe Soil Application
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Regenerative Braking Algorithm for Parallel Hydraulic Hybrid Vehicles Based on Fuzzy Q-Learning

1
College of Mechanical Engineering, Zhejiang University of Technology, Hangzhou 310014, China
2
School of Intelligent Manufacturing, Taizhou University, Taizhou 318000, China
*
Author to whom correspondence should be addressed.
Energies 2023, 16(4), 1895; https://doi.org/10.3390/en16041895
Submission received: 3 January 2023 / Revised: 31 January 2023 / Accepted: 3 February 2023 / Published: 14 February 2023
(This article belongs to the Section E: Electric Vehicles)

Abstract

:
The use of regenerative braking systems is an important approach for improving the travel mileage of electric vehicles, and the use of an auxiliary hydraulic braking energy recovery system can improve the efficiency of the braking energy recovery process. In this paper, we present an algorithm for optimizing the energy recovery efficiency of a hydraulic regenerative braking system (HRBS) based on fuzzy Q-Learning (FQL). First, we built a test bench, which was used to verify the accuracy of the hydraulic regenerative braking simulation model. Second, we combined the HRBS with the electric vehicle in ADVISOR. Third, we modified the regenerative braking control strategy by introducing the FQL algorithm and comparing it with a fuzzy-control-based energy recovery strategy. The simulation results showed that the power savings of the vehicle optimized by the FQL algorithm were improved by about 9.62% and 8.91% after 1015 cycles and under urban dynamometer driving schedule (UDDS) cycle conditions compared with a vehicle based on fuzzy control and the dynamic programming (DP) algorithm. The regenerative braking control strategy optimized by the fuzzy reinforcement learning method is more efficient in terms of energy recovery than the fuzzy control strategy.

1. Introduction

In urban conditions of frequent starting and stopping, about 30–50% of the kinetic energy is consumed by the frictional braking process. Due to the large mass and more frequent start–stop events of city buses, a large amount of braking energy is wasted. Therefore, recovering this energy is a way to increase the travel mileage of the vehicle [1,2]. At present, it is more common to charge the braking current to the battery or supercapacitor. However, this approach to energy recovery efficiency is not ideal, due to battery life, safety, and other issues [3]. Hydraulic accumulators have higher power densities and better charging–discharging energy ratios than batteries, supercapacitors, and other energy storage components [4].
The regenerative braking control strategy determines the distribution coefficient of the braking force and the intervention time during regenerative braking, which has an important impact on the safety of the vehicle braking, fuel economy, and power of the vehicle [5,6]. Taghavipour et al. found that the use of radial-basis-based neural networks improved the fuel economy of the vehicle while obtaining better maneuvering and stability [7]. Rule-based control strategies are widely accepted because they are simple to implement, but they take a long time to optimize because they usually require a large amount of experimental data and expert experience [8,9]. At the same time, it is difficult to explore the full potential of the regenerative braking system and to adapt flexibly to different driving conditions.
Based on the complex nonlinear time-varying characteristics of the regenerative braking of hybrid vehicles, many scholars have transformed the regenerative braking control problem of hybrid vehicles into an optimal solution problem [10,11,12]. Shangguan globally optimized the main parameters of a parallel hydraulic hybrid vehicle based on the dynamic programming algorithm, and the simulation results showed that the energy recovery efficiency was improved [13]. Larsson et al. simplified the dynamic planning algorithm by reducing the number of grid points generated by discretization, thereby reducing the computational effort [14]. Tate et al. proposed a control strategy based on stochastic dynamic planning, which is similar to the Markov decision process involved in artificial intelligence theory [15]. Global optimization algorithms such as dynamic programming techniques can calculate the optimal solution of each parameter of the system under known cycle conditions. However, due to the complexity of the hybrid vehicle system and the many constraints, the calculations are large and difficult to apply.
Along with the rapid development of learning-based artificial intelligence technology, many researchers have started to apply machine learning, deep learning, and other algorithms to automotive control [16,17]. Tian et al. separately collected data recorded during past vehicle driving experiences and used machine learning methods to learn automotive control strategies from them [18]. Reinforcement learning algorithms can learn through continuous repetitive experiments using the model, do not require the accumulation of preliminary data, and have a strong self-adjustment capability [19,20,21]. Qi et al. developed a control strategy for a vehicle by continuously rewarding and penalizing their model through the Q algorithm [22]. The learning algorithm based on neuron dynamic programming does not depend on the known information of cyclic working conditions and can adjust the energy management strategy parameters by itself, having good adaptability to different working conditions [23].
However, most of the research on hydraulic regenerative braking systems involves theoretical analyses and simulation experiments, and there is a lack of experiments further verifying the influence of each component on the braking effect in the process of hydraulic regenerative braking. Meanwhile, most of the existing regenerative braking force distribution schemes are rule based. If the setting cannot be changed, it cannot be applied to different driving conditions, and the braking energy of the vehicle cannot be fully recovered. This paper proposes a hydraulic regenerative braking energy recovery efficiency optimization algorithm based on fuzzy Q-learning (FQL) and a reward function based on the hydraulic regenerative braking energy recovery system, which can solve the problems of Q-learning and speed up the computation by introducing the method of fuzzy logic. Xian et al. and Zhou et al. made improvements on the basis of fuzzy logic and obtained a predictive model that is superior to other methods [24,25]. Furthermore, reinforcement learning is a knowledge-free online learning process that can adapt the regenerative braking control strategy to different driving conditions over time, which is more advantageous than fuzzy control, which requires expert experience [26,27].
In Section 2, this paper establishes the vehicle dynamics model and the mathematical model based on parallel hydraulic hybrid power systems (PHHPS). In Section 3, an optimized algorithm based on FQL for braking energy recovery efficiency is proposed. In Section 4, we describe our experimental bench. By comparing the experimental and simulated data, we obtain an accurate simulation model and combine it with the ADVISOR. Finally, the simulation simulates the braking energy recovery of the fuzzy control, DP algorithm and the fuzzy reinforcement learning control strategy under the 1015 cycle and UDDS cycle. The conclusions are presented in Section 5.

2. Model of Hydraulic Regenerative Braking System (HRBS)

Electric vehicle batteries are energy storage devices that have high energy density and low power density. However, such batteries cannot withstand high-power inrush currents, resulting in electric regenerative braking being less effective than expected, impacting the battery life.
The considered hydraulic braking energy recovery system is a parallel hydraulic hybrid system (PHHS) consisting of a battery, an electric motor, a secondary component, a hydraulic accumulator, a clutch, and a torque coupling. By using the torque coupler to match the torque and speed, the torque is transmitted to the gearbox, the final reducer, and finally the wheels (Figure 1). Secondary components and hydraulic accumulators are used to recover the braking energy. Since the two power systems, namely the electric drive and hydraulic drive, are connected in parallel, the parallel hybrid vehicle has multiple operating modes, including a pure electric drive mode, pure hydraulic drive mode, and hybrid drive mode.
The motor output torque and the hydraulic system output torque of the two-axis parallel structure are connected through a torque coupler. Since the motor and the secondary components are on two separate drive shafts, the electric system and the hydraulic system can output torque independently, ignore the influence of speed and operate independently in their respective high-efficiency zones, and make it easy to match the two power sources. At this time, since the double-shaft parallel structure is relatively simple, there are fewer energy transmission levels, which reduces the energy loss in the mechanical transmission process.
During the vehicle braking process, the secondary component works as a hydraulic pump that converts the kinetic energy into hydraulic energy, and then this energy is stored in the accumulator. During the vehicle starting and hill climbing processes, the secondary component acts as a motor, releasing the energy stored in the accumulator to drive the vehicle. When the hydraulic accumulator is unable to provide the torque in the drive phase, the electric motor will participate instead. During daily driving, the hydraulic system also participates in providing torque when the motor output torque is insufficient.
In the HRBS, the secondary component has both energy storage and driving functions. The torque provided by the secondary component is calculated as follows N m [28]:
T p / m = Δ p q p / m η p / m i 0 i g i 2 π
where Δ p is the amount of accumulator pressure change Mpa , q p / m is the displacement of the hydraulic pump motor mL / r , η p / m is the mechanical efficiency of the secondary component, i 0 is the torque coupling ratio, i g is the speed ratio of the gearbox, and i is the speed ratio of the final drive.
The braking force required to brake the vehicle is expressed as:
T r e q = T r e q + = T m o t o r + T p / m + ( t r a c t i o n ) T r e q = T m e c h + T r e g = T m e c h + T p / m ( b r a k i n g )
where T m o t o r is the driving force provided for the motor, T p / m + is the driving force provided by the secondary component, T m e c h is the mechanical friction braking force, and T p / m is the braking force provided by the secondary component.
The main parameters of the hydraulic accumulator include the initial pressure, volume, displacement, and maximum working pressure. In the hydraulic system, when the initial pressure of the accumulator is high, the braking torque and energy recovery efficiency provided by the HRBS will also be high, but the maximum energy absorbed will be reduced. The smaller the volume of the accumulator, the faster the accumulator pressure rises during braking, the greater the regenerative braking torque, and the higher the regenerative braking efficiency, but this will also cause the same energy storage capacity problem. In consideration of the airtightness and safety of the accumulator, the maximum pressure of the accumulator is usually 31.5 Mpa. According to Boyle’s law:
P 0 V 0 n = P 1 V 1 n = P 2 V 2 n = c o n s t
During braking, the energy recovered by the hydraulic accumulator is calculated as:
E r e g = P 1 V 1 n 1 P 1 p 1 n n 1
where P 0 is the initial inflation pressure, V 0 is the initial volume of the accumulator, P 1 is the minimum working pressure, V 1 is the minimum working volume, P 2 is the maximum working pressure of the accumulator, and E r e g denotes the maximum energy that can be stored in the accumulator (when p = P 2 ).
The kinetic energy of the vehicle is selected as the main objective in braking energy recovery, which is simplified as follows:
E a c c = 1 2 m u 0 2 1 2 m u 1 2
where m is the mass of vehicle, u 0 is the speed of the vehicle before braking, and u 1 is the speed of the vehicle at the end of the braking phase.
The braking energy recovery efficiency can be expressed as follows:
η = E r e g E a c c
Combined with the operating characteristics of the PHHS, the secondary component needs to meet the following requirements: (1) in the starting phase of the vehicle, the hydraulic system can drive the vehicle independently, i.e., the hydraulic system’s output torque is greater than the vehicle’s demand torque; (2) in the braking phase, the hydraulic pump needs to provide as much regenerative braking force as possible to improve the efficiency of the braking energy recovery.
To meet the requirements of the starting phase, the torque provided by the hydraulic motor needs to be greater than the combined force of the rolling resistance, aerodynamics resistance, and slope resistance. Therefore, the motor displacement needs to satisfy the following equation:
q p / m 2 π r F Δ p η p / m i 0 i g η M
F = m g f cos α + m g sin α + ρ C D A u 2 2
where η M is the total mechanical efficiency, r is the wheel radius, g is the gravitational constant, F is the sum of the vehicle driving resistance, f is the rolling resistance coefficient, α is the ground inclination angle, ρ is the density of air, C D is the aerodynamics resistance coefficient, and A is the front area of the vehicle.
In pursuit of maximum energy recovery efficiency, the HRBS needs to meet the light braking conditions for all braking forces provided by the hydraulic pump; when the braking strength is 0.1 (braking strength is defined as the ratio of deceleration to gravitational acceleration), the hydraulic pump displacement needs to meet:
q p / m 2 π r m g z Δ p η p / m i 0 i g η M
where z is the braking strength.
When the stored energy of the energy storage system reaches its limit, it is unable to receive the recovered energy from the outside world and cannot effectively deliver the required braking torque of the vehicle. In some braking situations, the increased braking force of the HRBS cannot fully provide the required braking torque for the vehicle, so friction braking is required in the achieving braking process. The total braking force of the vehicle consists of the front and rear braking forces:
F car = F f r o n t + F r e a r = f m g
where F f r o n t consists of mechanical braking and regenerative braking, and F r e a r consists of mechanical braking only. During vehicle braking, the braking effect is related to the utilization rate of the road attachment conditions.
When the adhesion conditions are not fully utilized, the vehicle is likely to slide sideways or show braking instability. When the front and rear axles of the car are clutched, the utilization rate of the road adhesion conditions is the highest, and the stability of the vehicle when braking is also the best. Therefore, when both the front and rear axles are locked:
F f r o n t = f m g b + z h g L
F r e a r = f m g a z h g L
where a is the horizontal distance from the center of mass of the vehicle to the front axle, b is the horizontal distance from the center of mass of the vehicle to the rear axle, L is the wheelbase of the vehicle, and h g is the height of the center of mass of the vehicle.
For each possible coefficient of adhesion, the following method can be used to obtain the front wheel braking force with the front and rear wheels locked simultaneously:
F f r o n t = 1 2 m g h g a 2 4 h g L m g F r e a r 2 F r e a r m g a h g
The regenerative braking factor K proposed in the HRBS is defined as the braking force provided by the hydraulic pump divided by the demanded braking torque of the front wheel:
K = T p / m / r F f
The regenerative braking distribution strategy is based on safety, first through the distribution of the front and rear axle braking torque so that the vehicle brake safely, and then in the form of a combined regenerative and mechanical braking distribution in the front axle.

3. FQL Algorithm and Models

A large number of researchers have introduced fuzzy control theory to the vehicle control process and achieved some results, but fuzzy control still relies on expert experience, and for some nonlinear systems, fuzzy control cannot achieve optimal control results [29]. In addition, once a fuzzy controller has been designed and produced, the vehicle equipped with the controller will follow this control rule in all operating conditions, leading to neglect of some of the recoverable energy.
Therefore, to solve the problem whereby fuzzy control cannot handle complex nonlinear systems, we introduce reinforcement learning algorithms and adjust the fuzzy rules of the fuzzy controller using the Q-Learning algorithm. In our proposed controller, the FQL algorithm acts as a decision engine that learns approaches to map the input states to the desired output decisions. The controller can maintain the original expert experience, while the FQL has an exploration function that improves the braking energy recovery performance.
The Q-learning algorithm is a table-driven learning algorithm. On the one hand, the use of fuzzy logic can solve the computational capacity limit and storage problems of Q-learning in the face of large-scale continuous state action problems, and fuzzy logic can improve the generalization ability of the reinforcement learning state action space. On the other hand, the FQL can optimize the control effect of the controller. Therefore, the FQL algorithm makes up for the shortcomings of fuzzy control and Q-learning algorithms.

3.1. FQL Model

The Markov decision process is the basic model of reinforcement learning. To solve a specific learning task, an agent is placed in an unknown environment, the agent takes an action based on the state in the environment, and the action can change its state in the environment and return a delayed numerical reward to the agent. The goal of the agent is to learn a strategy with which to take an action that allows the agent to maximize the accumulated reward in the task. Q- learning is a popular reinforcement learning algorithm that learns knowledge by updating a Q-table through a reward mechanism. The Q value is the expected cumulative reward that can be obtained by following the optimal policy after taking a certain action in each state. After learning, the agent is able to construct an optimal policy by simply selecting the action with the highest Q value in each state. The RL problem is modeled as follows:
  • X = x 1 , , x n , , x N is the set of all environment states, x t X indicates the state of the agent at the moment t ;
  • O = o 1 , , o n , , o N is the set of all actions that the controlled object can perform, and u t U indicates the actions performed by the agent at the moment t ;
  • r t is a scalar quantity indicating that the object of the controlled pair is under the state x t , where at this time action u t is taken and the environmental state shifts to x t + 1 , at which time the controlled object gets an immediate reward;
  • The Q value corresponding to each state action is updated by the temporal difference (TD) method with the following rules:
    Q t + 1 x t , a t = Q t x t , a t + β r t + 1 + γ max a Q t x t , a Q t x t , a t
Here, r t + 1 is the observed reward, β is the learning rate, and 0 ≤ β ≤ 1. High values of β result in rapid learning and adaptation, while low values of β slow down the learning and prevent the impact on the q-table from possible outliers. The max a Q t x t , a denotes the state x t of the estimated optimal value of Q under the discount factor γ , 0 ≤ γ ≤ 1. The value of γ determines whether the optimization process should consider the long-term reward or not.
Q-learning is not good at dealing with situations where the state space is relatively large, since a large amount of memory is needed to store the q-table. Even if a large amount of memory can be provided, agent learning requires a lot of trials and time to learn the required behavior, and vehicles driving urban routes require different environments to satisfy q-table updates in all cases, thereby causing a lot of learning costs.
FQL is a fuzzy extension of Q-learning that can overcome this problem [30]. We can encapsulate expert knowledge into a learning table to speed up the learning process. In FQL, the decision part is represented by a fuzzy inference system (FIS), which takes continuous and large discrete states as the inputs. The idea of the FQL algorithm is to use the FIS to integrate continuous and large discrete state inputs into a so-called q-table, which is different from the original q-table; the new q-table is based on evaluating the fuzzy rules in the FIS as the basis for updating the q-table. Compared to the infinite action- state space, fuzzy rules are finite and are used as inputs to the q-table so that the q-table can learn knowledge without relying on large amounts of memory.
In FQL, the FIS is represented by a set of fuzzy rules J . Rule j J is defined as:
I F x 1 i s L j 1 A N D x n i s L j n A N D x N i s L j N T H E N a = o j w i t h q L j , o j
where L j n is the label of the input variable x n base on rule j t h , o j is the action of the rule j t h , L j = L j 1 , , L j n , , L j N is the modal vector under rule j , and q L j , o j k is the modal vector L j and the action o j k of rule j t h in the Q value.
In the fuzzification layer of the FIS (Figure 2), the input states are fuzzified into the affiliation degree of the corresponding label by the affiliation function, and each input state is fuzzified to obtain a fuzzy set μ L n = μ 1 x n , , μ n x n , , μ N x n . Let J x denote the set of all rules in the rule evaluation layer, and each rule is a scalar obtained by multiplying the fuzzy sets corresponding to the input states, then the set of rules is the Cartesian product of the fuzzy sets corresponding to the input states. For the rules j t h ( j t h J ), α denotes the activation degree of this rule, which can be expressed as follows:
α j x = n = 1 N u L j n x n
In the FQL algorithm, we take a two-stage action selection approach. In the first stage, we select the local action according to the ε g r e e d y strategy to select local actions o j l . The agent makes it possible to explore untried actions instead of the action with the largest Q value, in order to ensure a higher long-term payoff. At each time step, the agent selects random actions with a fixed probability of 1 ε . At the beginning of the exploration, we let the agent take more random actions instead of greedily choosing the optimal action in the q-table.
ε                 j J x : o j l = arg max x k K q L j , o j k 1 ε       j J x : o j l = random   x k K ( o j k )
where o j k is the local action of the rule j t h . In the second stage of the action selection, the activation degree of the rule is obtained according to the input state, and the activation degree multiplied by the local action can be used to obtain the final action; here, the agent chooses the final action with the highest activation degree of the rule; then the final action output can be expressed as:
a = max j J x α j x o j l
By adding the time index to the equation, the Q value can be approximated as:
Q x t , a = j J x α j x t × q L j , o j l
After taking action, the agent obtains the status x t + 1 and reward r t + 1 . Then, the state x t + 1 of the Q value is calculated as follows:
V x t + 1 = j J x + 1 α j x t × max k L j , o j k
Based on the above equation, the error is calculated by the TD algorithm as follows [31]:
Δ Q = r t + 1 + γ V x t + 1 Q x t , a
Ultimately, we can update the Q value of each activation rule by using the above equation as follows:
q t + 1 L j , o j l = q t L j , o j l + β α j x t Δ Q
where γ and β are the same as in Equation (15).

3.2. Regenerative Braking Based on FQL

In this section, the above fuzzy reinforcement learning model will be simulated jointly with the regenerative braking system for electric vehicles. Here, the HRBS for the electric vehicles (HRBEV) is built in MATLAB and Simulink, which will be introduced later. The FQL model is built in Python and simulated jointly with Python and MATLAB. At each time step, the HBREV outputs state x t as the input of the FQL, which is fuzzed in the FIS. Then, the optimal action is selected according to the fuzzy rules and q-table, and the agent outputs action a to the HRBEV. In the next time step, the state x t + 1 is output to the FQL model from the HRBEV. The critic gives r t + 1 from x t + 1 . Finally, the q-table is updated using the TD method (Figure 3).

3.2.1. FQL Components

The RBFQL consists of the HRBEV and FQL, and the FQL consists of the following components.
  • State: The input state consists of the vehicle speed, braking intensity, accumulator pressure, demand braking torque, hydraulic pump providing torque, and other components. Here, the vehicle speed, braking intensity, and pressure of the accumulator are used as inputs of the FIS, while other components are used as inputs to the FQL observation reward value:
x t = u , z , p
where u denotes the vehicle speed, z denotes the braking intensity, and p is the accumulator pressure.
  • Action: The output action is the regenerative braking coefficient. The brake distribution factor needs to meet the vehicle braking safety and reduce the output when the accumulator is full:
a = α j o j 1 , α j o j 2 , , α j o j n
  • Reward function: The design of the reward function is the key to building a reinforcement learning system. The agent gets a reward based on the observed state and uses the reward to update the q-table. The goal of reinforcement learning is to maximize the cumulative reward over time, and the agent seeks to produce the action with the maximum Q value. In our system, the reinforcement learning goal is to seek to maximize the regenerative braking energy. Therefore, our reward function can be expressed as:
r t + 1 = 1 u t + 1 u r e q + Δ p P max P 0
where u r e q is the required speed at time t + 1 and u t + 1 is the real vehicle speed at time t + 1 , Δ p is the accumulator pressure change under braking conditions, P max is the maximum pressure of the accumulator (31.5 Mpa here), and P 0 is the minimum working pressure of the accumulator. Obviously, to obtain a larger reward, it is necessary to make the real vehicle speed u t + 1 as small as possible and the accumulator pressure variation as large as possible. If the reward is close to 0, this means that the action is invalid, and a negative reward means a penalty for the action. For example, if the agent takes an aggressive action in a certain state, i.e., takes a larger regenerative braking factor, the accumulator pressure change becomes large, and a large reward is obtained. However, due to the torque limitation of the pump, the pump does not provide enough braking torque, which will result in the u t + 1 u r e q being large, and the penalty will increase at this time. Therefore, the method for dealing with the relationship between the reward and punishment is the key to reinforcement learning.

3.2.2. Fuzzy Inference System

In this paper, we simulate the vehicle dynamics model in urban cyclic conditions (1015 cycle and UDDS cycle). Due to the limitation of the vehicle′s operating environment, it is not possible for all action states to be traversed, but we need the agent to output a more appropriate action even when facing an inexperienced environment. Therefore, we choose the fuzzy control approach to add knowledge to the action selection process in the form of rules, which can improve the speed of the agent in dealing with complex continuous state spaces or multidimensional discrete spaces [32].
The FQL system inputs three input variables: (1) vehicle speed u , (2) braking intensity z , and (3) accumulator pressure p ; and one output—(4) regenerative braking distribution coefficient k . The fuzzy division of these four variables is as follows:
  • The vehicle speed u . T u = U 1 , , U 5 , and the range of the theoretical domain is (0–80) ( km / h );
  • The braking strength z . T z = Z 1 , , Z 5 , and the theoretical domain range is (0–1) ( m / s 2 g );
  • The hydraulic accumulator pressure p . T p = P 1 , P 2 , P 3 , and the theoretical domain range is (10–31.5) ( Mpa );
  • The regenerative braking coefficient k . T k K 1 , , K 5 , and the theoretical domain range is (0–1).
The Sugeno method has the advantage of allowing simple calculations and facilitating data analyses, so we chose to set up the fuzzy control system using the first-order Sugeno method. The fuzzy control system performs the fuzzy inference process using fuzzy rules to obtain the output under different operating conditions. The knowledge base of the fuzzy system is described by a series of fuzzy logic rules in the form of IF–THEN, and the fuzzy rules are used as follows:
I F u i s U i A N D z i s Z l A N D p i s P m T H E N k i s K n
where U i T i , i = 1 , 2 , 3 , 4 , 5 , Z l T z , l = 1 , 2 , 3 , 4 , 5 , and P m T p , m = 1 , 2 , 3 . The regenerative braking coefficient T k = K 1 , K 2 , K 3 , K 4 , K 5 = 0 , 0.25 , 0.5 , 0.75 , 1 (Figure 4). For example, when the vehicle speed is very low, the braking intensity is medium, and the accumulator pressure is very low, the regenerative braking distribution coefficient will be very high. When the vehicle speed is high, the braking intensity will be high, and the accumulator pressure is medium, the regenerative braking distribution coefficient will be very low. The idea of the rule set is to try to recover more energy at low speeds and ensure safety at high speeds, and regenerative braking is not involved in the work. The total number of rules in the rule set is 75 5 × 5 × 3 .
The FIS will be improved using the QL method to obtain better performance.

3.2.3. FQL Setup

In the FQL setup, we will set the discount factor through the phasing γ , learning rate β , and random action selection probability 1 ε . In the early stage of learning and the exploration stage, we set the discount factor γ to 0.2, because in the early stage of training, Q s and V s do not converge, meaning the agent cannot correctly measure the expected future benefits. A higher discount factor will cause the agent to incorrectly estimate the current action value, leading to instability when the algorithm is updated, even making it difficult to converge. The discount factor will gradually increase as the training process proceeds. Meanwhile, we set the learning rate β value to 0.9, as a larger learning rate at the early stage of learning can reduce the training time, and the subsequent reduction in the learning rate is beneficial to maintaining the system stability. The probability of random action selection in the exploration phase 1 ε is set to 0.5; the agent needs to balance the maximum value action and develop potential higher value actions, while the random action selection probability will decrease with the learning process. Here, we compare the regenerative braking system with the fuzzy controller without adding the QL algorithm, so that we can judge the learning process and modify the above parameters. In addition, we initialize the q-table to zero to start the knowledge learning process from zero.
So far, this paper has completed the modeling of the hydraulic parallel hybrid braking system and vehicle dynamics, combined with the FQL to model the vehicle regenerative braking problem. The next section will verify the above system and the simulation of vehicle braking efficiency through experimental and simulation comparison methods.

4. Experiment and Simulation Results

In this section, we will apply the method proposed above for vehicle regenerative braking energy recovery in three steps. First, we build a test bench and the above mathematical model in Simulink to compare the test and simulation results. Then, we integrate the obtained Simulink model into the ADVISOR environment to model the regenerative braking torque distribution through fuzzy control. Finally, we use MATLAB API engine to simulate ADVISOR in parallel with the fuzzy reinforcement learning algorithm in Python to validate the proposed method.

4.1. Experiment Test

The experimental platform is built based on the PHHS established in the previous paper. The difference from Figure 1 is that we use the flywheel moment of inertia to simulate the kinetic energy of the vehicle to study the energy recovery efficiency of the hydraulic system. The kinetic energy simulation part consists of the flywheel, magnetic powder brake, electromagnetic clutch, three-phase synchronous motor, inverter, and console. Among them, the magnetic powder brake is used to simulate the mechanical friction brake, the electromagnetic clutch is used to open and close the direct connection between the motor and the flywheel, the motor provides the initial kinetic energy of the flywheel, the inverter is used to control the motor and adjust its speed, and the console controls the engagement of the electromagnetic clutch and the switch of the inverter (Figure 5). The hydraulic brake energy recovery part mainly consists of the secondary component, a hydraulic accumulator, and an oil tank. The secondary component converts the kinetic energy of flywheel rotation into hydraulic potential energy and fills the hydraulic accumulator with liquid. The hydraulic accumulator is the energy storage element, and the hydraulic oil is supplied from the oil tank. In addition to the above components, a speed sensor, voltage sensor, current sensor, pressure sensor, electrical parameter tester, and acquisition card are required to collect the experimental data. Table 1 shows the parameters of each component of the experimental stand.
In the bench experiment, the flywheel starts braking when the hydraulic pump starts working. The kinetic energy possessed by the flywheel is expressed as:
E f = 1 2 J ω 2
We performed three different sets of experiments based on the above experimental bench. By comparing the experimental data with the simulation data, we verified the accuracy of our simulation model. The simulation model will be used as the basis of the proposed algorithm.
First of all, we performed experiments and simulations with different initial flywheel speeds (Figure 6a). Here, the initial pressure of the accumulator is set to 5.6 MPa; the volume of the accumulator is 25 L; the displacement of the hydraulic pump is 50 mL/r; and the flywheel speeds are 360 r/min, 300 r/min, and 240 r/min when the electromagnetic clutch starts to close, and the flywheel drives the oil pump to rotate. It can be found that the rising trend of experimental and simulated accumulator pressures are almost the same, and the error is approximately 10%. In the experimental results, because the energy consumed by the friction will increase with the increase in speed of the flywheel, the energy recovery efficiency will decrease. Additionally, higher flywheel speeds can achieve higher energy recovery rates in the hydraulic regeneration system. The difference between the experimental and simulated response times is caused by the hydraulic system needing a response time of about 0.7 s, but the results of experiments and simulations will be basically unchanged.
To find the energy recovery efficiency of the braking process, the braking energy recovery efficiencies at different initial speeds can be calculated according to Equations (5) and (28), as shown in Table 2.
Secondly, we performed experiments and simulations with different initial accumulator pressures (Figure 6b). Here, the initial speed of the flywheel is 300 r/min; the displacement of the hydraulic pump is 50 mL/r; the initial volume of the accumulator is 10 L; and the initial pressures of the accumulator are 4.5 MPa, 5 Mpa, and 5.6 MPa. It can be found that an accumulator with a high initial pressure will also have a high final pressure after the same braking process. The experimental results will be slightly lower than the simulation results.
Thirdly, we performed experiments and simulations with respect to flywheel speeds with different initial accumulator pressures (Figure 6c). It can be observed from the results that the larger the initial accumulator pressure, the shorter the braking process time and the larger the power of the hydraulic regeneration system to provide recovery energy.
For the trend of the three sets of experiments, it can be found that the greater the accumulator instantaneous pressure, the greater the power of the hydraulic regeneration system to provide recovery energy, and more braking energy can be recovered. The experimental flywheel deceleration time is smaller than the simulated one, which is caused by the existence of larger friction in the system.
Regarding the effect of the initial pressure of the accumulator on the braking energy recovery efficiency, the simulation and experimental results are shown in Table 3.
The displacement of the accumulator affects the volume change in the accumulator oil and the output torque per unit of time. Here, we set the flywheel speed to 360 rpm; the accumulator volume is 16 L; the initial accumulator pressure is 5.7 MPa; and the hydraulic pump displacement rates are 60 mL/r, 50 mL/r, and 38 mL/r for the experiment (Figure 7).
From the experimental and simulation results, it can be seen that with larger displacement rates of the hydraulic pump, the braking torque output by the hydraulic pump increases rapidly, and the accumulator pressure rises rapidly at the early stage of braking. During the later stage, the braking torque output by the hydraulic pump tends to level off and the accumulator pressure rises slowly, then the pressure of the accumulator is consistent finally. Therefore, the simulation and experimental results generally tend to be consistent, and the simulation model has good accuracy.

4.2. Simulation Results

The vehicle model is based on the ADVISOR pure electric vehicle model for secondary development. A pure electric vehicle is selected in the ADVISOR module library, and its model parameters are modified according to a pure electric bus developed by a bus company. Here, we add the above hydraulic regenerative braking simulation model, modify the original vehicle control strategy, and train the FQL algorithm. Figure 8 shows the torque transfer diagram of the HRBS. Figure 9 shows how we combined the electric vehicle HRBS model in the ADVISOR environment. In addition, according to the characteristics of the frequent starting and stopping of urban buses, the simulation environments for the 1015 cycle and UDDS cycle are selected, including the acceleration, deceleration, uniform speed, uphill and downhill conditions, and other working conditions, as shown in Figure 10.
The hydraulic pump displacement and the transmission ratio of the torque coupler in the HRBS are selected according to the following rules. The hydraulic pump displacement is 0.025 L/r for the 1015 operating conditions and 0.03 L/r for the UDDS operating conditions. The transmission ratio for both torque couplers is 2.5. The state of charge (SOC) of the vehicle battery is set to 1 at the beginning of the simulation, and the vehicle SOC decreases as the simulation proceeds, serving as an evaluation index of the energy recovery effect.
Compared with the two driving cycles, the 1015 cycle has a low braking frequency, low intensity, and high percentage of regenerative braking forces. The UDDS cycle has a high braking frequency and high intensity, so the UDDS regenerative braking torque is higher than for the 1015 cycle (Figure 11 and Figure 12). For each braking process, the HRBS with reinforcement learning can obtain a greater final pressure rate from the accumulator, as well as greater braking energy recovery efficiency and regenerative braking torque than the HRBS with fuzzy control based on expert experience and the HRBS with the DP algorithm.
It can be seen that the HRBS with FQL saves 9.62% more power than the hydraulic braking system with fuzzy control and the DP algorithm under the 1015 conditions and 8.91% more under the UDDS conditions (Table 4). Therefore, the FQL-algorithm-based optimization strategy is effective and can be adapted to more operating conditions for adjustment.

5. Conclusions

This paper presents an algorithm for optimizing the energy recovery efficiency of the HRBS based on FQL for an urban electro-hydraulic hybrid bus. The experimental bench of the HRBS was built to verify the simulation model and investigate the influence of the HRBS parameters on the energy recovery characteristics. The whole hydraulic regenerative braking electric vehicle model was established by integrating the Simulink model into the ADVISOR. The vehicle regenerative braking control strategy was optimized using the FQL algorithm and simulated under the 1015 cycle and UDDS cycle conditions. The results showed that the control strategy optimized by the FQL algorithm recovered more braking energy than the fuzzy control strategy based on expert experience and the DP algorithm. In addition, the control strategy based on the FQL algorithm can output actions that are more consistent with the training environment, which is more flexible than the fuzzy control method. Therefore, we verified that this method can reduce the power consumption of electric vehicles and improve the efficiency of the hydraulic regenerative braking process.

Author Contributions

Conceptualization, X.N. and J.W.; methodology, N.L. and Y.Y.; software, J.W. and N.B.; validation, J.S. and J.W.; formal analysis, X.N. and J.W.; investigation, J.W.; resources, X.N.; data curation, J.W. and N.B.; writing—original draft preparation, J.W.; writing—review and editing, X.N., Y.Y. and J.S.; visualization, J.W. and N.B.; supervision, X.N.; project administration, X.N.; funding acquisition, X.N., Y.Y. and N.L.; All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Basic Public Welfare Research Program of Zhejiang Province, grant number LGG22E050019; and the National Natural Science Foundation of China (NSFC), grant numbers 51375452 and 51905483.

Data Availability Statement

The data that support the findings of this study are available from the corresponding author, Xiaobin Ning, upon reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sabri, M.F.M.; Danapalasingam, K.A.; Rahmat, M.F. A review on hybrid electric vehicles architecture and energy management strategies. Renew. Sustain. Energy Rev. 2016, 53, 1433–1442. [Google Scholar] [CrossRef]
  2. Li, N.; Jiang, J.P.; Sun, F.L.; Ye, M.R.; Ning, X.B.; Chen, P.Z. A cooperative control strategy for a hydraulic regenerative braking system based on chassis domain control. Electronics 2022, 11, 4212. [Google Scholar] [CrossRef]
  3. Xu, W.; Chen, H.; Zhao, H.Y. Braking energy optimization control for four in-wheel motors electric vehicles considering battery life. Control Theory Appl. 2019, 36, 1942–1951. [Google Scholar]
  4. Ramakrishnan, R.; Hiremath, S.S.; Singaperumal, M. Design strategy for improving the energy efficiency in series hydraulic/electric synergy system. Energy 2014, 67, 422–434. [Google Scholar] [CrossRef]
  5. Kim, D.H.; Kim, J.M.; Hwang, S.H.; Kim, H.S. Optimal brake torque distribution for a four-wheel-drive hybrid electric vehicle stability enhancement. Proc. Inst. Mech. Eng. Part D—J. Automob. Eng. 2007, 221, 1357–1366. [Google Scholar] [CrossRef]
  6. Zhai, L.; Sun, T.M.; Wang, J. Electronic stability control based on motor driving and braking torque distribution for a four in-wheel motor drive electric vehicle. IEEE Trans. Veh. Technol. 2016, 65, 4726–4739. [Google Scholar] [CrossRef]
  7. Taghavipour, A.; Foumani, M.S.; Boroushaki, M. Implementation of an optimal control strategy for a hydraulic hybrid vehicle using CMAC and RBF networks. Sci. Iran. 2012, 19, 327–334. [Google Scholar] [CrossRef]
  8. Chindamo, D.; Economou, J.T.; Gadola, M.; Knowles, K. A neurofuzzy-controlled power management strategy for a series hybrid electric vehicle. Proc. Inst. Mech. Eng. Part D—J. Automob. Eng. 2014, 228, 1034–1050. [Google Scholar] [CrossRef]
  9. Hassan, S.; Khanesar, M.A.; Kayacan, E.; Jaafar, J.; Khosravi, A. Optimal design of adaptive type-2 neuro-fuzzy systems: A review. Appl. Soft Comput. 2016, 44, 134–143. [Google Scholar] [CrossRef]
  10. Filipi, Z.; Kim, Y.J. Hydraulic hybrid propulsion for heavy vehicles: Combining the simulation and engine-in-the-loop techniques to maximize the fuel economy and emission benefits. Oil Gas Sci. Technol.-Rev. IFP Energ. N. 2010, 65, 155–178. [Google Scholar] [CrossRef]
  11. Li, N.; Ning, X.B.; Wang, Q.C.; Li, J.L. Hydraulic regenerative braking system studies based on a nonlinear dynamic model of a full vehicle. J. Mech. Sci. Technol. 2017, 31, 2691–2699. [Google Scholar] [CrossRef]
  12. Moura, S.J.; Fathy, H.K.; Callaway, D.S.; Stein, J.L. A stochastic optimal control approach for power management in plug-in hybrid electric vehicles. IEEE Trans. Control Syst. Technol. 2011, 19, 545–555. [Google Scholar] [CrossRef]
  13. Ning, X.; Shangguan, J.; Xiao, Y.; Fu, Z.; Xu, G.; He, A.; Li, B. Optimization of energy recovery efficiency for parallel hydraulic hybrid power systems based on dynamic programming. Math. Probl. Eng. 2019, 2019, 9691507. [Google Scholar] [CrossRef]
  14. Larsson, V.; Mardh, L.J.; Egardt, B.; Karlsson, S. Commuter route optimized energy management of hybrid electric vehicles. IEEE Trans. Intell. Transp. Syst. 2014, 15, 1145–1154. [Google Scholar] [CrossRef]
  15. Tate, E.D.; Grizzle, J.W.; Peng, H.E. SP-SDP for fuel consumption and tailpipe emissions minimization in an EVT hybrid. IEEE Trans. Control Syst. Technol. 2010, 18, 673–685. [Google Scholar] [CrossRef]
  16. Zhu, M.X.; Wang, X.S.; Wang, Y.H. Human-like autonomous car-following model with deep reinforcement learning. Transp. Res/Part C-Emerg. Technol. 2018, 97, 348–368. [Google Scholar] [CrossRef]
  17. Grigorescu, S.; Trasnea, B.; Cocias, T.; Macesanu, G. A survey of deep learning techniques for autonomous driving. J. Field Robot. 2020, 37, 362–386. [Google Scholar] [CrossRef]
  18. Tian, H.; Li, S.E.; Wang, X.; Huang, Y.; Tian, G. Data-driven hierarchical control for online energy management of plug-in hybrid electric city bus. Energy 2018, 142, 55–67. [Google Scholar] [CrossRef]
  19. Chen, L.; Hu, X.M.; Tang, B.; Cheng, Y. Conditional DQN-based motion planning with fuzzy logic for autonomous driving. IEEE Trans. Intell. Transp. Syst. 2022, 23, 2966–2977. [Google Scholar] [CrossRef]
  20. Kuutti, S.; Bowden, R.; Jin, Y.C.; Barber, P.; Fallah, S. A survey of deep learning applications to autonomous vehicle control. IEEE Trans. Intell. Transp. Syst. 2021, 22, 712–733. [Google Scholar] [CrossRef]
  21. Liu, T.; Zou, Y.; Liu, D.X.; Sun, F.C. Reinforcement learning–based energy management strategy for a hybrid electric tracked vehicle. Energies 2015, 8, 7243–7260. [Google Scholar] [CrossRef] [Green Version]
  22. Qi, X.; Wu, G.; Boriboonsomsin, K.; Barth, M.J.; Gonder, J. Data-driven reinforcement learning-based real-time energy management system for plug-in hybrid electric vehicles. Transp. Res. Rec. 2016, 142, 1–8. [Google Scholar] [CrossRef]
  23. Chen, Z.Y.; Liu, W.; Yang, Y.; Chen, W.Q. Online energy management of plug-in hybrid electric vehicles for prolongation of all-electric range based on dynamic programming. Math. Probl. Eng. 2015, 2015, 368769. [Google Scholar] [CrossRef]
  24. Xian, S.D.; Chen, K.Y.; Cheng, Y. Improved seagull optimization algorithm of partition and XGBoost of prediction for fuzzy time series forecasting of COVID-19 daily confirmed. Adv. Eng. Softw. 2022, 173, 103212. [Google Scholar] [CrossRef] [PubMed]
  25. Zhou, J.D.; Li, X.; Wang, X.; Chai, Y.P.; Zhang, Q.P. Locally weighted factorization machine with fuzzy partition for elderly readmission prediction. Knowl -Based Syst. 2022, 242, 108326. [Google Scholar] [CrossRef]
  26. Maia, R.; Mendes, J.; Araujo, R.; Silva, M.; Nunes, U. Regenerative braking system modeling by fuzzy Q-Learning. Eng. Appl. Artif. Intell. 2020, 93, 103712. [Google Scholar] [CrossRef]
  27. Jin, L.Q.; Tian, D.Y.; Zhang, Q.X.; Wang, J.J. Optimal torque distribution control of multi-axle electric vehicles with in-wheel motors based on DDPG algorithm. Energies 2020, 13, 1331. [Google Scholar] [CrossRef]
  28. Sun, H.; Jiang, J.H.; Wang, X. Parameters matching and control method of hydraulic hybrid vehicles with secondary regulation technology. Chin. J. Mech. Eng. 2009, 22, 57–63. [Google Scholar] [CrossRef]
  29. Maia, R.; Silva, M.; Araujo, R.; Nunes, U. Electrical vehicle modeling: A fuzzy logic model for regenerative braking. Expert Syst. Appl. 2015, 42, 8504–8519. [Google Scholar] [CrossRef]
  30. Shi, H.B.; Li, X.S.; Hwang, K.S.; Pan, W.; Xu, G.J. Decoupled visual servoing with fuzzy Q-learning. IEEE Trans. Ind. Inf. 2018, 14, 241–252. [Google Scholar] [CrossRef]
  31. Elfwing, S.; Uchibe, E.; Doya, K. Sigmoid-weighted linear units for neural network function approximation in reinforcement learning. Neural Netw. 2018, 107, 3–11. [Google Scholar] [CrossRef] [PubMed]
  32. Pu, Z.Y.; Cui, Z.Y.; Tang, J.J.; Wang, S.; Wang, Y.H. Multimodal traffic speed monitoring: A real-time system based on passive Wi-Fi and Bluetooth sensing technology. IEEE Internet Things J. 2021, 9, 12413–12424. [Google Scholar] [CrossRef]
Figure 1. Parallel hydraulic hybrid power structure: (1) accumulator; (2) relief valve; (3) pressure sensor; (4) electromagnetic digital valve; (5) variable hydraulic pump/motor; (6) torque coupler; (7) gearbox; (8) battery; (9) motor; (10) electromagnetic clutch; (11) tank.
Figure 1. Parallel hydraulic hybrid power structure: (1) accumulator; (2) relief valve; (3) pressure sensor; (4) electromagnetic digital valve; (5) variable hydraulic pump/motor; (6) torque coupler; (7) gearbox; (8) battery; (9) motor; (10) electromagnetic clutch; (11) tank.
Energies 16 01895 g001
Figure 2. The FQL structure.
Figure 2. The FQL structure.
Energies 16 01895 g002
Figure 3. The RBFQL structure.
Figure 3. The RBFQL structure.
Energies 16 01895 g003
Figure 4. The fuzzy partition: (a) speed; (b) braking strength; (c) accumulator state of charge.
Figure 4. The fuzzy partition: (a) speed; (b) braking strength; (c) accumulator state of charge.
Energies 16 01895 g004
Figure 5. The HRBS test bench: 1—motor; 2—electromagnetic clutch; 3—magnetic powder brake; 4—flywheel; 5—revolution speed transducer; 6—commutator; 7—electromagnetic clutch No. 2; 8—hydraulic pump and motor; 9—tank; 10—relief valve; 11—hydraulic pressure sensor; 12—hydraulic accumulator; 13—hub motor; 14—proportional amplifier; 15—oil discharge valve.
Figure 5. The HRBS test bench: 1—motor; 2—electromagnetic clutch; 3—magnetic powder brake; 4—flywheel; 5—revolution speed transducer; 6—commutator; 7—electromagnetic clutch No. 2; 8—hydraulic pump and motor; 9—tank; 10—relief valve; 11—hydraulic pressure sensor; 12—hydraulic accumulator; 13—hub motor; 14—proportional amplifier; 15—oil discharge valve.
Energies 16 01895 g005
Figure 6. Comparison between bench experiment and simulation results.
Figure 6. Comparison between bench experiment and simulation results.
Energies 16 01895 g006
Figure 7. Pressure change diagram of the accumulator under different displacement rates.
Figure 7. Pressure change diagram of the accumulator under different displacement rates.
Energies 16 01895 g007
Figure 8. Torque flow of the parallel hydraulic regenerative braking system.
Figure 8. Torque flow of the parallel hydraulic regenerative braking system.
Energies 16 01895 g008
Figure 9. The electric vehicle model combined with the proposed HRBS model: (a) vehicle control system; (b) driving condition system; (c) vehicle transmission system; (d) hydraulic hybrid system; (e) electric vehicle system; (f) flywheel subsystem; (g) accumulator subsystem.
Figure 9. The electric vehicle model combined with the proposed HRBS model: (a) vehicle control system; (b) driving condition system; (c) vehicle transmission system; (d) hydraulic hybrid system; (e) electric vehicle system; (f) flywheel subsystem; (g) accumulator subsystem.
Energies 16 01895 g009
Figure 10. Duty cycles: (a) 1015; (b) UDDS.
Figure 10. Duty cycles: (a) 1015; (b) UDDS.
Energies 16 01895 g010
Figure 11. Simulation results under the 1015 cycle.
Figure 11. Simulation results under the 1015 cycle.
Energies 16 01895 g011
Figure 12. Simulation results under the UDDS cycle.
Figure 12. Simulation results under the UDDS cycle.
Energies 16 01895 g012
Table 1. Main component parameters of the kinetic energy simulation part.
Table 1. Main component parameters of the kinetic energy simulation part.
NameModelKey Parameters
MotorY132S-4Rated speed: 1440 r/min
Rated voltage: 380 V
Maximum power: 5.5 KW
Flywheel Rotational inertia: 45.3 kg m 2
Secondary ComponentSecondary ComponentMaximum Displacement: 107 mL/r
Minimum Displacement: 38 mL/r
Hydraulic AccumulatorNXQNominal volume: 25 L
Nominal pressure: 31.5 MPa
Voltage SensorMIK-DZU-100VMeasuring range: 0~100 V
Output: 0~10 V
Current SensorMIK-DZI-150AMeasuring range: 0~150 A
Output: 0~10 V
Pressure sensorsMCY-XMeasuring range: 0~35 MPa
Sensitivity: 2 MPa/V
Photoelectric speed sensorDK890Operating voltage: 10~36 V
Measurement range: 0~10 kHz
Acquisition CardsUSB3103ASignal channels: 16
Sampling frequency: 500 kHz
Table 2. Final pressure and energy recovery efficiency of the accumulator at various speeds.
Table 2. Final pressure and energy recovery efficiency of the accumulator at various speeds.
ComponentDescription240 r/min300 r/min360 r/min
Experimental resultspressure6.45 MPa6.79 MPa7.23 MPa
efficiency84.96%85.7%85.97%
Experimental resultspressure6.25 MPa6.62 MPa7.05 MPa
efficiency78.05%76.65%74.01%
Table 3. Final pressure and energy recovery efficiency rates of the accumulator at various initial pressures.
Table 3. Final pressure and energy recovery efficiency rates of the accumulator at various initial pressures.
ComponentDescription4.5 MPa5 MPa5.6 MPa
Experimental resultspressure6.78 MPa7.23 MPa7.81 MPa
efficiency80.3%85.7%81.2%
Experimental resultspressure5.91 MPa6.51 MPa7.56 MPa
efficiency70.55%73.65%72.01%
Table 4. The amount of electricity consumed.
Table 4. The amount of electricity consumed.
CycleHRBEV with FQLHRBEV with DPHRBEV with Fuzzy ControlAverage Growth Rate (%)
10150.02340.02590.02549.62%
UDDS0.06510.07030.07158.91%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ning, X.; Wang, J.; Yin, Y.; Shangguan, J.; Bao, N.; Li, N. Regenerative Braking Algorithm for Parallel Hydraulic Hybrid Vehicles Based on Fuzzy Q-Learning. Energies 2023, 16, 1895. https://doi.org/10.3390/en16041895

AMA Style

Ning X, Wang J, Yin Y, Shangguan J, Bao N, Li N. Regenerative Braking Algorithm for Parallel Hydraulic Hybrid Vehicles Based on Fuzzy Q-Learning. Energies. 2023; 16(4):1895. https://doi.org/10.3390/en16041895

Chicago/Turabian Style

Ning, Xiaobin, Jiazheng Wang, Yuming Yin, Jiarong Shangguan, Nanxin Bao, and Ning Li. 2023. "Regenerative Braking Algorithm for Parallel Hydraulic Hybrid Vehicles Based on Fuzzy Q-Learning" Energies 16, no. 4: 1895. https://doi.org/10.3390/en16041895

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop