You are currently on the new version of our website. Access the old version .

144 Results Found

  • Article
  • Open Access
3 Citations
2,061 Views
12 Pages

Pipelined Stochastic Gradient Descent with Taylor Expansion

  • Bongwon Jang,
  • Inchul Yoo and
  • Dongsuk Yook

26 October 2023

Stochastic gradient descent (SGD) is an optimization method typically used in deep learning to train deep neural network (DNN) models. In recent studies for DNN training, pipeline parallelism, a type of model parallelism, is proposed to accelerate SG...

  • Article
  • Open Access
4 Citations
2,019 Views
17 Pages

11 January 2025

With the advancement of autonomous driving technology, vehicle lane-change decision (LCD) has become a critical issue for improving driving safety and efficiency. Traditional deep reinforcement learning (DRL) methods face challenges such as slow conv...

  • Article
  • Open Access
16 Citations
4,043 Views
20 Pages

Stochastic Recognition of Human Physical Activities via Augmented Feature Descriptors and Random Forest Model

  • Sheikh Badar ud din Tahir,
  • Abdul Basit Dogar,
  • Rubia Fatima,
  • Affan Yasin,
  • Muhammad Shafiq,
  • Javed Ali Khan,
  • Muhammad Assam,
  • Abdullah Mohamed and
  • El-Awady Attia

2 September 2022

Human physical activity recognition from inertial sensors is shown to be a successful approach for monitoring elderly individuals and children in indoor and outdoor environments. As a result, researchers have shown significant interest in developing...

  • Article
  • Open Access
29 Citations
5,161 Views
17 Pages

Extracting Knowledge from Big Data for Sustainability: A Comparison of Machine Learning Techniques

  • Raghu Garg,
  • Himanshu Aggarwal,
  • Piera Centobelli and
  • Roberto Cerchione

25 November 2019

At present, due to the unavailability of natural resources, society should take the maximum advantage of data, information, and knowledge to achieve sustainability goals. In today’s world condition, the existence of humans is not possible witho...

  • Article
  • Open Access
5 Citations
3,024 Views
37 Pages

7 November 2021

The central problems of some of the existing Non-Intrusive Load Monitoring (NILM) algorithms are indicated as: (1) higher required electrical device identification accuracy; (2) the fact that they enable training over a larger device count; and (3) t...

  • Article
  • Open Access
9 Citations
5,683 Views
15 Pages

9 March 2022

In recent years, deep neural networks (DNN) have been widely used in many fields. Lots of effort has been put into training due to their numerous parameters in a deep network. Some complex optimizers with many hyperparameters have been utilized to ac...

  • Article
  • Open Access
23 Citations
5,587 Views
22 Pages

Soya Yield Prediction on a Within-Field Scale Using Machine Learning Models Trained on Sentinel-2 and Soil Data

  • Branislav Pejak,
  • Predrag Lugonja,
  • Aleksandar Antić,
  • Marko Panić,
  • Miloš Pandžić,
  • Emmanouil Alexakis,
  • Philip Mavrepis,
  • Naweiluo Zhou,
  • Oskar Marko and
  • Vladimir Crnojević

7 May 2022

Agriculture is the backbone and the main sector of the industry for many countries in the world. Assessing crop yields is key to optimising on-field decisions and defining sustainable agricultural strategies. Remote sensing applications have greatly...

  • Article
  • Open Access
36 Citations
3,882 Views
17 Pages

Analysis of Fault Classifiers to Detect the Faults and Node Failures in a Wireless Sensor Network

  • S. Gnanavel,
  • M. Sreekrishna,
  • Vinodhini Mani,
  • G. Kumaran,
  • R. S. Amshavalli,
  • Sadeen Alharbi,
  • Mashael Maashi,
  • Osamah Ibrahim Khalaf,
  • Ghaida Muttashar Abdulsahib and
  • Theyazn H. H. Aldhyani
  • + 1 author

Technology evaluation in the electronics field leads to the great development of Wireless Sensor Networks (WSN) for a variety of applications. The sensor nodes are deployed in hazardous environments, and they are operated by isolated battery sources....

  • Article
  • Open Access
8 Citations
3,163 Views
22 Pages

23 July 2022

The primary motivation is to address difficulties in data interpretation or a reduction in model accuracy. Although differential privacy can provide data privacy guarantees, it also creates problems. Thus, we need to consider the noise setting for di...

  • Article
  • Open Access
6 Citations
4,407 Views
16 Pages

Stochastic gradient descent is the method of choice for solving large-scale optimization problems in machine learning. However, the question of how to effectively select the step-sizes in stochastic gradient descent methods is challenging, and can gr...

  • Article
  • Open Access
12 Citations
3,122 Views
12 Pages

Damped Newton Stochastic Gradient Descent Method for Neural Networks Training

  • Jingcheng Zhou,
  • Wei Wei,
  • Ruizhi Zhang and
  • Zhiming Zheng

29 June 2021

First-order methods such as stochastic gradient descent (SGD) have recently become popular optimization methods to train deep neural networks (DNNs) for good generalization; however, they need a long training time. Second-order methods which can lowe...

  • Article
  • Open Access
5 Citations
6,188 Views
7 Pages

15 January 2020

This paper is a step towards developing a geometric understanding of a popular algorithm for training deep neural networks named stochastic gradient descent (SGD). We built upon a recent result which observed that the noise in SGD while training typi...

  • Feature Paper
  • Article
  • Open Access
27 Citations
5,045 Views
15 Pages

Parallel Implementation on FPGA of Support Vector Machines Using Stochastic Gradient Descent

  • Felipe F. Lopes,
  • João Canas Ferreira and
  • Marcelo A. C. Fernandes

Sequential Minimal Optimization (SMO) is the traditional training algorithm for Support Vector Machines (SVMs). However, SMO does not scale well with the size of the training set. For that reason, Stochastic Gradient Descent (SGD) algorithms, which h...

  • Article
  • Open Access
7 Citations
3,443 Views
16 Pages

This paper mainly proposes some improved stochastic gradient descent (SGD) algorithms with a fractional order gradient for the online optimization problem. For three scenarios, including standard learning rate, adaptive gradient learning rate, and mo...

  • Proceeding Paper
  • Open Access
3 Citations
2,595 Views
9 Pages

8 November 2023

Machine learning algorithms are integrated into computer-aided design (CAD) methodologies to support medical practitioners in diagnosing patient disorders. This research seeks to enhance the accuracy of classifying malaria-infected erythrocytes (RBCs...

  • Article
  • Open Access
1 Citations
1,339 Views
25 Pages

27 April 2024

The uncertainty in the new power system has increased, leading to limitations in traditional stability analysis methods. Therefore, considering the perspective of the three-dimensional static security region (SSR), we propose a novel approach for sys...

  • Article
  • Open Access
1,276 Views
13 Pages

Background/Objectives: Posture is a significant indicator of health status in older adults. This study aimed to develop an automatic posture assessment tool based on sagittal photographs by validating recognition models using convolutional neural net...

  • Article
  • Open Access
1 Citations
1,701 Views
11 Pages

26 April 2023

The aim of this article is to establish a stochastic search algorithm for neural networks based on the fractional stochastic processes {BtH,t≥0} with the Hurst parameter H∈(0,1). We define and discuss the properties of fractional stochastic p...

  • Article
  • Open Access
1,473 Views
26 Pages

8 August 2025

We study the generalization properties of stochastic optimization methods under adaptive data sampling schemes, focusing on the setting of pairwise learning, which is central to tasks like ranking, metric learning, and AUC maximization. Unlike pointw...

  • Article
  • Open Access
3 Citations
2,467 Views
23 Pages

3 October 2024

Accurate monitoring of estuarine turbidity patterns is important for maintaining aquatic ecological balance and devising informed estuarine management strategies. This study aimed to enhance the prediction of estuarine turbidity patterns by enhancing...

  • Article
  • Open Access
1 Citations
2,434 Views
13 Pages

A Scaling Transition Method from SGDM to SGD with 2ExpLR Strategy

  • Kun Zeng,
  • Jinlan Liu,
  • Zhixia Jiang and
  • Dongpo Xu

24 November 2022

In deep learning, the vanilla stochastic gradient descent (SGD) and SGD with heavy-ball momentum (SGDM) methods have a wide range of applications due to their simplicity and great generalization. This paper uses an exponential scaling method to reali...

  • Review
  • Open Access
215 Citations
26,541 Views
23 Pages

29 January 2023

In the age of artificial intelligence, the best approach to handling huge amounts of data is a tremendously motivating and hard problem. Among machine learning models, stochastic gradient descent (SGD) is not only simple but also very effective. This...

  • Article
  • Open Access
3 Citations
2,011 Views
16 Pages

A Novel Sine Step Size for Warm-Restart Stochastic Gradient Descent

  • Mahsa Soheil Shamaee and
  • Sajad Fathi Hafshejani

6 December 2024

This paper proposes a novel sine step size for warm-restart stochastic gradient descent (SGD). For the SGD based on the new proposed step size, we establish convergence rates for smooth non-convex functions with and without the Polyak–Łoja...

  • Article
  • Open Access
43 Citations
17,765 Views
12 Pages

Learning-Rate Annealing Methods for Deep Neural Networks

  • Kensuke Nakamura,
  • Bilel Derbel,
  • Kyoung-Jae Won and
  • Byung-Woo Hong

22 August 2021

Deep neural networks (DNNs) have achieved great success in the last decades. DNN is optimized using the stochastic gradient descent (SGD) with learning rate annealing that overtakes the adaptive methods in many tasks. However, there is no common choi...

  • Article
  • Open Access
1 Citations
1,568 Views
21 Pages

In this work, we explored, for the first time, to the best of our knowledge, the potential of stochastic gradient descent (SGD) to optimize random phase functions for application in non-iterative phase-only hologram generation. We defined and evaluat...

  • Feature Paper
  • Article
  • Open Access
2 Citations
2,844 Views
9 Pages

This paper proposes a novel mathematical theory of adaptation to convexity of loss functions based on the definition of the condense-discrete convexity (CDC) method. The developed theory is considered to be of immense value to stochastic settings and...

  • Article
  • Open Access
18 Citations
3,510 Views
15 Pages

7 January 2023

In this paper, a deep learning optimization algorithm is proposed, which is based on the Grünwald–Letnikov (G-L) fractional order definition. An optimizer fractional calculus gradient descent based on the G-L fractional order definition (F...

  • Article
  • Open Access
1,342 Views
16 Pages

6 June 2025

In structural health monitoring (SHM), ensuring data completeness is critical for enhancing the accuracy and reliability of structural condition assessments. SHM data are prone to random missing values due to signal interference or connectivity issue...

  • Article
  • Open Access
14 Citations
3,925 Views
17 Pages

Effect of Initial Configuration of Weights on Training and Function of Artificial Neural Networks

  • Ricardo J. Jesus,
  • Mário L. Antunes,
  • Rui A. da Costa,
  • Sergey N. Dorogovtsev,
  • José F. F. Mendes and
  • Rui L. Aguiar

13 September 2021

The function and performance of neural networks are largely determined by the evolution of their weights and biases in the process of training, starting from the initial configuration of these parameters to one of the local minima of the loss functio...

  • Article
  • Open Access
1 Citations
2,747 Views
19 Pages

27 July 2023

Recent meta-learning models often learn priors from observed tasks using a network optimized via stochastic gradient descent (SGD), which usually takes more training steps to convergence. In this paper, we propose an accelerated Bayesian meta-learnin...

  • Article
  • Open Access
15 Citations
9,600 Views
17 Pages

Stochastic Weight Averaging Revisited

  • Hao Guo,
  • Jiyong Jin and
  • Bin Liu

24 February 2023

Averaging neural network weights sampled by a backbone stochastic gradient descent (SGD) is a simple-yet-effective approach to assist the backbone SGD in finding better optima, in terms of generalization. From a statistical perspective, weight-averag...

  • Article
  • Open Access
8 Citations
3,495 Views
16 Pages

A Bounded Scheduling Method for Adaptive Gradient Methods

  • Mingxing Tang,
  • Zhen Huang,
  • Yuan Yuan,
  • Changjian Wang and
  • Yuxing Peng

1 September 2019

Many adaptive gradient methods have been successfully applied to train deep neural networks, such as Adagrad, Adadelta, RMSprop and Adam. These methods perform local optimization with an element-wise scaling learning rate based on past gradients. Alt...

  • Article
  • Open Access
67 Citations
8,892 Views
21 Pages

18 January 2018

Recently, deep learning techniques have had a practical role in vehicle detection. While much effort has been spent on applying deep learning to vehicle detection, the effective use of training data has not been thoroughly studied, although it has gr...

  • Article
  • Open Access
49 Citations
5,790 Views
11 Pages

Averaging Is Probably Not the Optimum Way of Aggregating Parameters in Federated Learning

  • Peng Xiao,
  • Samuel Cheng,
  • Vladimir Stankovic and
  • Dejan Vukobratovic

11 March 2020

Federated learning is a decentralized topology of deep learning, that trains a shared model through data distributed among each client (like mobile phones, wearable devices), in order to ensure data privacy by avoiding raw data exposed in data center...

  • Article
  • Open Access
12 Citations
5,439 Views
15 Pages

17 May 2020

This paper demonstrates a novel approach to training deep neural networks using a Mutual Information (MI)-driven, decaying Learning Rate (LR), Stochastic Gradient Descent (SGD) algorithm. MI between the output of the neural network and true outcomes...

  • Article
  • Open Access
131 Citations
18,031 Views
15 Pages

7 March 2018

This study proposes a modified convolutional neural network (CNN) algorithm that is based on dropout and the stochastic gradient descent (SGD) optimizer (MCNN-DS), after analyzing the problems of CNNs in extracting the convolution features, to improv...

  • Article
  • Open Access
164 Citations
17,739 Views
20 Pages

29 October 2020

The advent of social media, particularly Twitter, raises many issues due to a misunderstanding regarding the concept of freedom of speech. One of these issues is cyberbullying, which is a critical global issue that affects both individual victims and...

  • Article
  • Open Access
124 Citations
10,065 Views
22 Pages

Shallow Landslide Prediction Using a Novel Hybrid Functional Machine Learning Algorithm

  • Dieu Tien Bui,
  • Himan Shahabi,
  • Ebrahim Omidvar,
  • Ataollah Shirzadi,
  • Marten Geertsema,
  • John J. Clague,
  • Khabat Khosravi,
  • Biswajeet Pradhan,
  • Binh Thai Pham and
  • Saro Lee
  • + 5 authors

17 April 2019

We used a novel hybrid functional machine learning algorithm to predict the spatial distribution of landslides in the Sarkhoon watershed, Iran. We developed a new ensemble model which is a combination of a functional algorithm, stochastic gradient de...

  • Article
  • Open Access
3 Citations
2,069 Views
29 Pages

14 March 2024

The ISP (Internet Service Provider) industry relies heavily on internet traffic forecasting (ITF) for long-term business strategy planning and proactive network management. Effective ITF frameworks are necessary to manage these networks and prevent n...

  • Article
  • Open Access
11 Citations
3,857 Views
13 Pages

4 June 2021

Deep learning models are based on a combination of neural network architectures, optimization parameters and activation functions. All of them provide exponential combinations whose computational fitness is difficult to pinpoint. The intricate resemb...

  • Article
  • Open Access
8 Citations
2,802 Views
31 Pages

12 October 2022

A key feature of federated learning (FL) is that not all clients participate in every communication epoch of each global model update. The rationality for such partial client selection is largely to reduce the communication overhead. However, in many...

  • Feature Paper
  • Article
  • Open Access
11 Citations
3,179 Views
12 Pages

Multi-Depth Computer-Generated Hologram Based on Stochastic Gradient Descent Algorithm with Weighted Complex Loss Function and Masked Diffraction

  • Jiale Quan,
  • Binbin Yan,
  • Xinzhu Sang,
  • Chongli Zhong,
  • Hui Li,
  • Xiujuan Qin,
  • Rui Xiao,
  • Zhi Sun,
  • Yu Dong and
  • Huming Zhang

6 March 2023

In this paper, we propose a method to generate multi-depth phase-only holograms using stochastic gradient descent (SGD) algorithm with weighted complex loss function and masked multi-layer diffraction. The 3D scene can be represented by a combination...

  • Article
  • Open Access
177 Citations
10,263 Views
19 Pages

State-of-the-Art CNN Optimizer for Brain Tumor Segmentation in Magnetic Resonance Images

  • Muhammad Yaqub,
  • Jinchao Feng,
  • M. Sultan Zia,
  • Kaleem Arshid,
  • Kebin Jia,
  • Zaka Ur Rehman and
  • Atif Mehmood

Brain tumors have become a leading cause of death around the globe. The main reason for this epidemic is the difficulty conducting a timely diagnosis of the tumor. Fortunately, magnetic resonance images (MRI) are utilized to diagnose tumors in most c...

  • Article
  • Open Access
2 Citations
1,462 Views
11 Pages

Stochastic Gradient Descent for Kernel-Based Maximum Correntropy Criterion

  • Tiankai Li,
  • Baobin Wang,
  • Chaoquan Peng and
  • Hong Yin

17 December 2024

Maximum correntropy criterion (MCC) has been an important method in machine learning and signal processing communities since it was successfully applied in various non-Gaussian noise scenarios. In comparison with the classical least squares method (L...

  • Article
  • Open Access
75 Citations
8,856 Views
16 Pages

Performance Analysis of State-of-the-Art CNN Architectures for LUNA16

  • Iftikhar Naseer,
  • Sheeraz Akram,
  • Tehreem Masood,
  • Arfan Jaffar,
  • Muhammad Adnan Khan and
  • Amir Mosavi

11 June 2022

The convolutional neural network (CNN) has become a powerful tool in machine learning (ML) that is used to solve complex problems such as image recognition, natural language processing, and video analysis. Notably, the idea of exploring convolutional...

  • Article
  • Open Access
3 Citations
2,899 Views
14 Pages

27 September 2022

The traditional High-Resolution Range Profile (HRRP) target recognition method has difficulty automatically extracting target deep features, and has low recognition accuracy under low training samples. To solve these problems, a ship recognition meth...

  • Article
  • Open Access
1,206 Views
14 Pages

Supervised Learning Fuzzy Matrix Based on Input–Output Fuzzy Vectors

  • Meili Ye,
  • Nianliang Wang,
  • Xianfeng Yu,
  • Xiao Wang and
  • Wuniu Liu

9 February 2025

Fuzzy matrices play a crucial role in fuzzy logic and fuzzy systems. This paper investigates the problem of supervised learning fuzzy matrices through sample pairs of input–output fuzzy vectors, where the fuzzy matrix inference mechanism is bas...

  • Article
  • Open Access
19 Citations
4,080 Views
26 Pages

An Improved Bees Algorithm for Training Deep Recurrent Networks for Sentiment Classification

  • Sultan Zeybek,
  • Duc Truong Pham,
  • Ebubekir Koç and
  • Aydın Seçer

26 July 2021

Recurrent neural networks (RNNs) are powerful tools for learning information from temporal sequences. Designing an optimum deep RNN is difficult due to configuration and training issues, such as vanishing and exploding gradients. In this paper, a nov...

  • Article
  • Open Access
13 Citations
4,063 Views
18 Pages

Train Me If You Can: Decentralized Learning on the Deep Edge

  • Diogo Costa,
  • Miguel Costa and
  • Sandro Pinto

6 May 2022

The end of Moore’s Law aligned with data privacy concerns is forcing machine learning (ML) to shift from the cloud to the deep edge. In the next-generation ML systems, the inference and part of the training process will perform at the edge, whi...

  • Article
  • Open Access
3 Citations
2,336 Views
14 Pages

20 June 2021

Nowadays, as the number of items is increasing and the number of items that users have access to is limited, user-item preference matrices in recommendation systems are always sparse. This leads to a data sparsity problem. The latent factor analysis...

of 3