Figure 1.
Asymmetrical Half-Bridge (ASHB) converter for a 4ϕ Switched Reluctance Motor (SRM).
Figure 1.
Asymmetrical Half-Bridge (ASHB) converter for a 4ϕ Switched Reluctance Motor (SRM).
Figure 2.
Magnetization state of the ASHB converter.
Figure 2.
Magnetization state of the ASHB converter.
Figure 3.
(a,b). Free-wheeling state of the ASHB converter.
Figure 3.
(a,b). Free-wheeling state of the ASHB converter.
Figure 4.
Demagnetization state of the ASHB converter.
Figure 4.
Demagnetization state of the ASHB converter.
Figure 5.
Phase voltage and current waveforms characteristics associated with an ASHB.
Figure 5.
Phase voltage and current waveforms characteristics associated with an ASHB.
Figure 6.
Overall system architecture of the proposed DDPG-based reinforcement learning framework for a torque ripple minimization scheme for an 8/6 SRM drive.
Figure 6.
Overall system architecture of the proposed DDPG-based reinforcement learning framework for a torque ripple minimization scheme for an 8/6 SRM drive.
Figure 7.
(a) Overall experimental setup for inductance profile 8/6 Switched Reluctance Motor. (b) LCR meter display showing the measured phase inductance and winding resistance of the 8/6 SRM at a fixed rotor position during experimental characterization. (c) Experimental setup for encoder response measurement of the 8/6 Switched Reluctance Motor.
Figure 7.
(a) Overall experimental setup for inductance profile 8/6 Switched Reluctance Motor. (b) LCR meter display showing the measured phase inductance and winding resistance of the 8/6 SRM at a fixed rotor position during experimental characterization. (c) Experimental setup for encoder response measurement of the 8/6 Switched Reluctance Motor.
Figure 8.
(a) Experimentally obtained encoder voltage responses as a function of rotor position. (b) Measured inductance variation in SRM phases over a full electrical cycle obtained through discrete rotor position stepping.
Figure 8.
(a) Experimentally obtained encoder voltage responses as a function of rotor position. (b) Measured inductance variation in SRM phases over a full electrical cycle obtained through discrete rotor position stepping.
Figure 9.
Reinforcement learning framework for the proposed SRM control scheme.
Figure 9.
Reinforcement learning framework for the proposed SRM control scheme.
Figure 10.
Structural representation of the Deep Deterministic Policy Gradient (DDPG) algorithm used in the proposed SRM control system.
Figure 10.
Structural representation of the Deep Deterministic Policy Gradient (DDPG) algorithm used in the proposed SRM control system.
Figure 11.
Critic neural network architecture, employed in the DDPG-based SRM torque ripple minimization controller.
Figure 11.
Critic neural network architecture, employed in the DDPG-based SRM torque ripple minimization controller.
Figure 12.
Actor neural network architecture, employed in the DDPG-based SRM torque ripple minimization controller.
Figure 12.
Actor neural network architecture, employed in the DDPG-based SRM torque ripple minimization controller.
Figure 13.
Learning policy flow of the Deep Deterministic Policy Gradient (DDPG) algorithm.
Figure 13.
Learning policy flow of the Deep Deterministic Policy Gradient (DDPG) algorithm.
Figure 14.
Simulink model of the proposed DDPG-based reinforcement learning control architecture for torque ripple minimization in a Switched Reluctance Motor.
Figure 14.
Simulink model of the proposed DDPG-based reinforcement learning control architecture for torque ripple minimization in a Switched Reluctance Motor.
Figure 15.
Flowchart of the proposed reinforcement learning-based DDPG training strategy for torque ripple minimization in the 8/6 Switched Reluctance Motor drive.
Figure 15.
Flowchart of the proposed reinforcement learning-based DDPG training strategy for torque ripple minimization in the 8/6 Switched Reluctance Motor drive.
Figure 16.
Episode reward and value function convergence of the DDPG-based SRM control agent.
Figure 16.
Episode reward and value function convergence of the DDPG-based SRM control agent.
Figure 17.
Phase current waveforms of the 8/6 SRM under the proposed DDPG-based actor–critic control.
Figure 17.
Phase current waveforms of the 8/6 SRM under the proposed DDPG-based actor–critic control.
Figure 18.
Four-phase flux linkage characteristics of the 8/6 SRM under the proposed DDPG-based actor–critic control.
Figure 18.
Four-phase flux linkage characteristics of the 8/6 SRM under the proposed DDPG-based actor–critic control.
Figure 19.
Electromagnetic torque and rotor speed response of the SRM at 1000 RPM using the proposed DDPG actor–critic reinforcement learning controller.
Figure 19.
Electromagnetic torque and rotor speed response of the SRM at 1000 RPM using the proposed DDPG actor–critic reinforcement learning controller.
Figure 20.
Electromagnetic torque and rotor speed response of the SRM at 1500 RPM using the proposed DDPG actor–critic reinforcement learning controller.
Figure 20.
Electromagnetic torque and rotor speed response of the SRM at 1500 RPM using the proposed DDPG actor–critic reinforcement learning controller.
Figure 21.
Electromagnetic torque and rotor speed response of the SRM at 2000 RPM using the proposed DDPG actor–critic reinforcement learning controller.
Figure 21.
Electromagnetic torque and rotor speed response of the SRM at 2000 RPM using the proposed DDPG actor–critic reinforcement learning controller.
Figure 22.
Electromagnetic torque and rotor speed response of the SRM at 2500 RPM using the proposed DDPG actor–critic reinforcement learning controller.
Figure 22.
Electromagnetic torque and rotor speed response of the SRM at 2500 RPM using the proposed DDPG actor–critic reinforcement learning controller.
Figure 23.
Electromagnetic torque and rotor speed response of the SRM at 3000 RPM using the proposed DDPG actor–critic reinforcement learning controller.
Figure 23.
Electromagnetic torque and rotor speed response of the SRM at 3000 RPM using the proposed DDPG actor–critic reinforcement learning controller.
Figure 24.
Reward convergence and peak-to-peak torque ripple minimization using the DDPG-based learning controller for SRM drive. The reward curve (yellow) indicates the progressive improvement in learning performance, while the torque ripple curve (blue) shows a significant reduction and stabilization after 0.15 s, confirming optimal policy convergence.
Figure 24.
Reward convergence and peak-to-peak torque ripple minimization using the DDPG-based learning controller for SRM drive. The reward curve (yellow) indicates the progressive improvement in learning performance, while the torque ripple curve (blue) shows a significant reduction and stabilization after 0.15 s, confirming optimal policy convergence.
Figure 25.
Torque ripple comparison of a Switched Reluctance Motor (SRM) operating at 1500 RPM using (a) the proposed DDPG reinforcement learning (DDPG-RL) control, (b) Torque Sharing Function (TSF) control, and (c) Direct Instantaneous Torque Control (DITC). The steady-state electromagnetic torque responses highlight the superior torque ripple reduction achieved by the proposed DDPG-RL approach compared with the conventional control strategies.
Figure 25.
Torque ripple comparison of a Switched Reluctance Motor (SRM) operating at 1500 RPM using (a) the proposed DDPG reinforcement learning (DDPG-RL) control, (b) Torque Sharing Function (TSF) control, and (c) Direct Instantaneous Torque Control (DITC). The steady-state electromagnetic torque responses highlight the superior torque ripple reduction achieved by the proposed DDPG-RL approach compared with the conventional control strategies.
Figure 26.
Torque ripple comparison of a Switched Reluctance Motor (SRM) operating at 3000 RPM using (a) the proposed DDPG–reinforcement learning (DDPG-RL) control, (b) Torque Sharing Function (TSF) control, and (c) Direct Instantaneous Torque Control (DITC). The steady-state electromagnetic torque responses highlight the effectiveness of the proposed DDPG-RL approach in reducing torque ripple compared with the conventional control strategies.
Figure 26.
Torque ripple comparison of a Switched Reluctance Motor (SRM) operating at 3000 RPM using (a) the proposed DDPG–reinforcement learning (DDPG-RL) control, (b) Torque Sharing Function (TSF) control, and (c) Direct Instantaneous Torque Control (DITC). The steady-state electromagnetic torque responses highlight the effectiveness of the proposed DDPG-RL approach in reducing torque ripple compared with the conventional control strategies.
Figure 27.
Torque ripple comparison of three SRM control strategies (DDPG RL, TSF, and DITC) at 1500 RPM and 3000 RPM.
Figure 27.
Torque ripple comparison of three SRM control strategies (DDPG RL, TSF, and DITC) at 1500 RPM and 3000 RPM.
Figure 28.
Hardware experimental setup of the proposed DDPG-based SRM drive system.
Figure 28.
Hardware experimental setup of the proposed DDPG-based SRM drive system.
Figure 29.
DC generator coupled to the SRM.
Figure 29.
DC generator coupled to the SRM.
Figure 30.
(a–e) Experimental torque–speed meter readings (a) 1000 RPM, (b) 1500 RPM, (c) 2000 RPM, (d) 2500 RPM, and (e) 3000 RPM.
Figure 30.
(a–e) Experimental torque–speed meter readings (a) 1000 RPM, (b) 1500 RPM, (c) 2000 RPM, (d) 2500 RPM, and (e) 3000 RPM.
Table 1.
Comparative analysis of propulsion motors for EV applications.
Table 1.
Comparative analysis of propulsion motors for EV applications.
| Motor Parameters | SRM | PMSM | SCIM |
|---|
| Size | Compact | Moderate | Moderate |
| Weight | Low | Moderate | Moderate |
| Cost | Low | High | Low |
| Ruggedness | High | Low | High |
| Constant Torque/Speed Range | Wide | Wide | Moderate |
| Torque Ripple | High | Low | Low |
| Noise & Vibration | High | Low | Low |
| Power Converters | Specific | Modular | Modular |
| Permanent Magnets | No | Yes | No |
| Efficiency | Moderate | High | Low |
Table 2.
Heuristic learning-behavior interpretations used for calibration of the DDPG-based SRM control framework.
Table 2.
Heuristic learning-behavior interpretations used for calibration of the DDPG-based SRM control framework.
| Property | Heuristic Interpretation/Calibration Rationale | Practical Meaning in SRM Control |
|---|
| Critic Learning Behavior | A high discount factor (γ = 0.99) and a moderate critic learning rate (2 × 10−4) are selected to promote numerically stable Bellman updates and bounded Q-function estimation. A large replay buffer (106 samples) and mini-batch training (batch size = 128) reduce temporal correlation in torque ripple data. | Prevents divergence in value estimation and stabilizes the learning of the torque ripple cost landscape during training. |
| Actor Learning Smoothness | A smaller actor learning rate (1 × 10−4) relative to the critic, together with a physically bounded action space for firing angle offsets, promotes gradual policy updates. | Ensures smooth adaptation of phase firing angle offsets, thereby avoiding excessive torque ripple caused by abrupt or aggressive commutation. |
| Actor–critic Time-Scale Separation | A two-time-scale learning strategy is achieved by faster critic updates relative to actor updates, reinforced through a soft target network update rate (τ = 0.005). | Maintains stable interaction between value estimation and policy improvement throughout the learning process. |
| Torque Ripple Energy Reduction (Empirical Indicator) | Reward shaping penalizes peak-to-peak torque ripple, while bounded exploration using Ornstein–Uhlenbeck noise (μ = 0, σ = 0.25) encourages local exploration. | Produces progressively smoother electromagnetic torque and reduced torque ripple in closed-loop SRM operation. |
Table 3.
Simulation parameters of the 8/6 Switched Reluctance Motor drive system in MATLAB SIMULINK.
Table 3.
Simulation parameters of the 8/6 Switched Reluctance Motor drive system in MATLAB SIMULINK.
| Parameter | Value/Range | Unit | Description |
|---|
| SRM Type | 8/6 (4-Phase) | - | Switched Reluctance Motor Configuration |
| Rated Power | 75 | kW | Nominal Power Output |
| DC Link Voltage | 220 | V | Input DC Bus Voltage to ASHB Converter |
| Simulated Speed Range | 1000–3000 | RPM | Speed Range simulated in Simulink |
| Load Torque | 20 | Nm | Applied Load Torque |
| Simulated Phase Current | 5–10 | A | Current range used in simulation |
| Aligned Inductance | 55 | mH | Inductance at Aligned Position |
| Unaligned Inductance | 12 | mH | Inductance at Unaligned Position |
| Maximum Flux Linkage | 1.5 | Wb | Per-Phase Maximum Flux Linkage |
Table 4.
Training hyperparameters adopted in the DDPG-based actor–critic learning framework for torque ripple minimization.
Table 4.
Training hyperparameters adopted in the DDPG-based actor–critic learning framework for torque ripple minimization.
| Parameter | Value/Setting |
|---|
| RL Algorithm | DDPG (Actor–Critic) |
| Learning Rate (Actor Network) | 1 × 10−4 |
| Learning Rate (Critic Network) | 2 × 10−4 |
| Discount Factor (γ) | 0.99 |
| Target Network Update Rate (τ) | 0.005 |
| Replay Buffer Size | 1 × 106 samples |
| Mini-Batch Size | 128 |
| Exploration Noise Model | Ornstein–Uhlenbeck Noise (μ = 0, σ = 0.25) |
| Maximum Training Episodes | 500 Episodes |
| Episode Length | 1 s (Simulation Time) |
Table 5.
Actor–Critic Neural Network configuration used for continuous control learning.
Table 5.
Actor–Critic Neural Network configuration used for continuous control learning.
| Network | Layer Structure | Activation Function | Output Dimension |
|---|
| Actor Network | Input Layer → Dense (256) → Dense (128) → Dense (64) → Output Layer | ReLU (hidden), Tanh (output) | 4 |
| Critic State Path | Input Layer → Dense (256) → Dense (128) | ReLU | - |
| Critic Action Path | Input Layer → Dense (64) | ReLU | - |
| Merged Critic Output Path | Dense (128) → Dense (64) → Scalar Q-Value Output | ReLU (hidden), Linear (output) | 1 |
Table 6.
Observation-state representation and continuous action command format used in the RL control loop.
Table 6.
Observation-state representation and continuous action command format used in the RL control loop.
| Category | Dimension | Variables | Role in Learning |
|---|
| Observation (State) Vector | 13 | Phase Currents, Flux Linkage, Electromagnetic Torque, Rotor Speed, Rotor Position (two scalars), Previous Action | Captures the motor dynamic and magnetic states for adaptive control. |
| Action Vector | 4 | Firing Angle Offset (4 phases) | Directly modulates the torque production profile. |
| Action Range | −1 to +1 (normalized) | Scaled to Realistic α Firing Angle Bounds | Ensures safe actuator operation and smooth switching. |
Table 7.
Torque ripple reduction results using the proposed DDPG-RL framework under various speeds.
Table 7.
Torque ripple reduction results using the proposed DDPG-RL framework under various speeds.
| Speed Setpoint (RPM) | Steady Speed (RPM) | Load Torque (Nm) | Mean Torque (Nm) | Tmax (Nm) | Tmin (Nm) | Settling Time (s) | Torque Ripple (p-p %) |
|---|
| 1000 | 1012.4 | 20 | 20.53 | 20.74 | 20.31 | 0.3 | 2.08 |
| 1500 | 1513.3 | 20 | 20.79 | 20.96 | 20.65 | 0.3 | 1.48 |
| 2000 | 2008.6 | 20 | 21.06 | 21.21 | 20.94 | 0.3 | 1.29 |
| 2500 | 2506.9 | 20 | 21.32 | 21.51 | 21.22 | 0.3 | 1.38 |
| 3000 | 3038.6 | 20 | 21.61 | 21.89 | 21.51 | 0.3 | 1.75 |
Table 8.
Common operating and motor parameters used for control strategies.
Table 8.
Common operating and motor parameters used for control strategies.
| Parameter | Value | Remark |
|---|
| SRM configuration | 8/6, 4-phase | Same motor model |
| Converter topology | 4-phase ASHB | Identical power stage |
| DC link voltage | 220 V | Variable |
| Load torque | 20 Nm | Constant load |
| Operated speeds | 1500, 3000 RPM | Identical test conditions |
| Phase current limit | 5–10 A | Same constraints |
| Torque reference | 20 Nm | Same command |
Table 9.
Controller-specific parameters and tuning settings.
Table 9.
Controller-specific parameters and tuning settings.
| Control Strategy | Parameter | Value/Description |
|---|
| DDPG-RL | Actor learning rate | 1 × 10−4 |
| Critic learning rate | 2 × 10−4 |
| Discount factor (γ) | 0.99 |
| Target update rate (τ) | 0.005 |
| Replay buffer size | 1 × 106 |
| Mini-batch size | 128 |
| Exploration noise | Ornstein–Uhlenbeck (μ = 0, σ = 0.25) |
| TSF | Torque Sharing Function | Cubic TSF lookup table |
| Current reference shaping | Fixed, non-adaptive |
| Overlap/conduction angle | Fixed |
| DITC | Torque hysteresis band | Fixed |
| Switching logic | Hysteresis-based |
| Commutation angles | Fixed |
Table 10.
Specifications of Switched Reluctance Motor (SRM).
Table 10.
Specifications of Switched Reluctance Motor (SRM).
| Specifications | Values/Range |
|---|
| Shaft Power | 3.7 kW/5 HP |
| Supply Voltage | 220 V |
| Rated Current | 5–10 A |
| SR Motor Configuration | 8/6 (8 Stator Poles, 6 Rotor Poles) |
| Rated Speed | 3000 RPM |
| Rated Torque | 8–25 Nm |
| Motor Winding Resistance | 0.22 Ω per phase |
| Moment of Inertia | 0.0055 Kgm2 |
| Encoder Resolution | 24 PPR |
| Efficiency | 90% |
Table 11.
Experimental performance of SRM under constant load torque.
Table 11.
Experimental performance of SRM under constant load torque.
| Speed (RPM) | Output Torque (kgm) | Output Torque (Nm) | Output Power (kW) |
|---|
| 1002 | 2.09 | 20.51 | 2.16 |
| 1503 | 2.11 | 20.71 | 3.26 |
| 2001 | 2.13 | 20.90 | 4.38 |
| 2504 | 2.15 | 21.09 | 5.55 |
| 3005 | 2.17 | 21.28 | 6.67 |