Next Article in Journal
Adaptive Transient Synchronization Support Strategy for Grid-Forming Energy Storage Facing Inverter Faults
Previous Article in Journal
GhostBlock-Augmented Lightweight Gaze Tracking via Depthwise Separable Convolution
Previous Article in Special Issue
Evolutionary Framework with Binary Decision Diagram for Multi-Classification: A Human-Inspired Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Article

SVRG-AALR: Stochastic Variance-Reduced Gradient Method with Adaptive Alternating Learning Rate for Training Deep Neural Networks

School of Computer, Electronics and Information, Guangxi University, Nanning 530004, China
*
Author to whom correspondence should be addressed.
Electronics 2025, 14(15), 2979; https://doi.org/10.3390/electronics14152979
Submission received: 7 June 2025 / Revised: 19 July 2025 / Accepted: 24 July 2025 / Published: 25 July 2025
(This article belongs to the Special Issue Advances in Machine Learning for Image Classification)

Abstract

The stochastic variance-reduced gradient (SVRG) theory is particularly well-suited for addressing gradient variance in deep neural network (DNN) training; however, its direct application to DNN training is hindered by adaptation challenges. To tackle this issue, the present paper proposes a series of strategies focused on adaptive alternating learning rates to effectively adapt SVRG for DNN training. Firstly, within the outer loop of SVRG, both the full gradient and the learning rate specific to DNN training are computed. For two distinct formulas used for calculating the learning rate, an alternating strategy is introduced that employs them alternately across iterations. This approach allows for simultaneous provision of diverse guidance information regarding parameter change rates and gradient change rates during DNN weight updates. Additionally, a threshold method is utilized to correct the learning rate into an appropriate range, thereby accelerating convergence. Secondly, in the inner loop of SVRG, DNN weights are updated using mini-batch average gradient along with the proposed learning rate. Concurrently, mini-batch average gradients from each iteration within the inner loop are refined and aggregated into a single gradient exhibiting reduced variance through an inertia strategy. This refined gradient is then relayed back to the outer loop to recalculate the new learning rate. The efficacy of the proposed algorithm has been validated on models including LeNet, VGG11, ResNet34, and DenseNet121 while being compared against several classic and advanced optimizers. Experimental results demonstrate that the proposed algorithm exhibits remarkable training robustness across DNN models with diverse characteristics. In terms of training convergence, the proposed algorithm demonstrates competitiveness with state-of-the-art algorithms, such as Lion, developed by the Google Brain team.
Keywords: SVRG; DNN training; adaptive alternating learning rate; inertial correction strategy; aggregation of the mini-batch average gradients SVRG; DNN training; adaptive alternating learning rate; inertial correction strategy; aggregation of the mini-batch average gradients

Share and Cite

MDPI and ACS Style

Zou, S.; Qin, H.; Yang, G.; Wang, P. SVRG-AALR: Stochastic Variance-Reduced Gradient Method with Adaptive Alternating Learning Rate for Training Deep Neural Networks. Electronics 2025, 14, 2979. https://doi.org/10.3390/electronics14152979

AMA Style

Zou S, Qin H, Yang G, Wang P. SVRG-AALR: Stochastic Variance-Reduced Gradient Method with Adaptive Alternating Learning Rate for Training Deep Neural Networks. Electronics. 2025; 14(15):2979. https://doi.org/10.3390/electronics14152979

Chicago/Turabian Style

Zou, Shiyun, Hua Qin, Guolin Yang, and Pengfei Wang. 2025. "SVRG-AALR: Stochastic Variance-Reduced Gradient Method with Adaptive Alternating Learning Rate for Training Deep Neural Networks" Electronics 14, no. 15: 2979. https://doi.org/10.3390/electronics14152979

APA Style

Zou, S., Qin, H., Yang, G., & Wang, P. (2025). SVRG-AALR: Stochastic Variance-Reduced Gradient Method with Adaptive Alternating Learning Rate for Training Deep Neural Networks. Electronics, 14(15), 2979. https://doi.org/10.3390/electronics14152979

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop