You are currently viewing a new version of our website. To view the old version click .
Applied Sciences
  • Article
  • Open Access

23 November 2022

On the Relative Impact of Optimizers on Convolutional Neural Networks with Varying Depth and Width for Image Classification

,
and
1
Department of Computer Engineering, Federal University of Technology, Minna 920211, Nigeria
2
Department of Electrical Engineering Science, University of Johannesburg, Johannesburg 2092, South Africa
3
Digital Transformation Portfolio, Tshwane University of Technology, Pretoria 0183, South Africa
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Deep Learning Architectures for Computer Vision

Abstract

The continued increase in computing resources is one key factor that is allowing deep learning researchers to scale, design and train new and complex convolutional neural network (CNN) architectures in terms of varying width, depth, or both width and depth to improve performance for a variety of problems. The contributions of this study include an uncovering of how different optimization algorithms impact CNN architectural setups with variations in width, depth, and both width/depth. Specifically in this study, three different CNN architectural setups in combination with nine different optimization algorithms—namely SGD vanilla, with momentum, and with Nesterov momentum, RMSProp, ADAM, ADAGrad, ADADelta, ADAMax, and NADAM—are trained and evaluated using three publicly available benchmark image classification datasets. Through extensive experimentation, we analyze the output predictions of the different optimizers with the CNN architectures using accuracy, convergence speed, and loss function as performance metrics. Findings based on the overall results obtained across the three image classification datasets show that ADAM and NADAM achieved superior performances with wider and deeper/wider setups, respectively, while ADADelta was the worst performer, especially with the deeper CNN architectural setup.

1. Introduction

Neural network optimization algorithms continue to be a well-studied subject among researchers [1,2,3,4,5,6,7]. Training a deep neural network could largely be termed an optimization problem and is usually trained using stochastic gradient descent-based optimization algorithms. The main goals of researchers are to minimize the error function and accelerate convergence to an optimal global solution, with the overall objective of improving the model’s performance and its generalization ability. In this regard, optimizers are an important hyperparameter that affect the training performance of deep neural networks. Hence, it is important to choose the right optimizer for any given dataset problem that is being investigated, since the overall objective of neural network training is to minimize the prediction error by locating the global optimum on the loss surface. However, there is no consensus on this, so researchers are left with the task of experimenting with different optimization algorithms.
Several well-known optimization algorithms are used in training neural networks and can be broadly categorized into two: the classic stochastic gradient descent (SGD), which uses a static learning rate, and the adaptive stochastic learning rate such as ADAM and ADAGrad, which use adaptive learning rates. The learning rate is a key and central hyperparameter used by the optimizers for training neural networks and controls how fast a given model adapts to the problem it seeks to solve by helping the model to respond to the estimated error or loss each time the weights and biases values are updated. For instance, too large a learning rate could result in too fast a weights update, which may lead to instability in the training process and cause the model to converge prematurely to a suboptimal solution, whereas too small a learning rate may result in very tiny updates to the weights and a longer training process that could cause the model to be stuck in the local optima.
Researchers developed the adaptive learning rate optimizers as an improvement to the classic SGD because of the latter’s perceived slowness in arriving at an optimal solution and manually tuning the learning rate. However, one of the challenges still faced by these adaptive learning rate optimization algorithms is being trapped in local minima as against the desirable global minima [8], sometimes even exhibiting inferior performance in comparison to the classic SGD in some machine learning settings and problems [8], which has resulted in using a warm-up heuristic to mitigate these effects [9] and other ongoing improvements published in the literature. Among the adaptive learning rate optimizers, ADAM is the most popular among researchers due to its suitability in most problem cases. Because there is no consensus on the right optimizer to choose for a given task, as shown in a comparative study conducted in [10], researchers continue to work to address the issues with ADAM and other well-known optimization algorithms bordering around learning rate, stabilizing training, faster convergence, and improved generalization [6,11].
There have been two recent developments in optimization research for deep neural networks. First is a study that sought to develop an improved variant of ADAM called rADAM (rectified ADAM) [7]. The authors sought to address the underlying causes of sub-optimal convergence to local optima by rectifying the undesirable large variance of adaptive learning rate usually observed during the early stage of training in various neural network architectures and models. The root cause, they argue, is because of the limited training data at the initial training stage compared to the later stage. They advocated the use of a lower learning rate during the first few training epochs as a solution to alleviate the convergence problem.
Secondly, another recent study saw the development of the Lookahead optimization algorithm published by the authors in [6], which iteratively updates two weights, with the faster weights exploring or ‘looking ahead’, while the slower weights maintain the overall training stability of the neural network. The study was inspired by advances made in neural network loss surface research that allows a robust way to accelerate convergence and training stability All this is achieved with less hyperparameter tuning and minimal computational cost.
The backpropagation algorithm is key to fast learning in neural networks [12]. Backpropagation is a term that is used to compute the partial derivatives ∂C/∂w and ∂C/∂b of the cost function C, concerning any weight w and bias b in the network [13]. It gives insights into the overall behavior of the neural network in terms of how changing the weights and biases minimizes the cost function, which is determined by the gradients of the cost function. One of the factors that determine the stability and speed of network converge is the optimization algorithm; hence, it is worth examining how each optimizer’s update rule could make convergence faster, while at the same time maintaining high accuracy and low loss based on any given dataset-specific problem and the neural network architecture.
This paper extends the work in [14] in which the authors performed a comparative evaluation of the seven most commonly used first-order stochastic gradient-based optimization algorithms on a simple convolutional neural network architecture. In this current work, we go further to probe the impact of optimization algorithms on increasing CNN network sizes to understand their performance behavior. Accordingly, we pose the following questions: (1) How do different optimizers impact the learning performance of CNN architectures with variations in width, depth and width/depth on image classification problems? (2) Are there significant differences in the performance outputs?
Thus, we formulate a hypothesis for this study that given a non-convex problem, different optimization algorithms can find completely different solutions when initialized from the same point. We achieve this through an empirical study of different CNN architectures with varying network sizes to find out if we observe any discernable learning performance.
This leads us to the main contributions of the present paper, summarized as follows:
  • We implemented three different wider, deeper, and wider/deeper simple CNN architectures.
  • We conducted extensive experiments and analyzed the effects of nine different optimizers on increasing deepness, wideness, and deepness/wideness of the CNN architectures on three benchmark image classification datasets (Cats and Dogs, Natural Images, and Fashion MNIST).
  • We provided insights into the effects of optimizers on CNN depth, width and depth/width architectures to inform optimal problem-specific CNN model design.
The remainder of the paper is organized in the following way: Section 2 presents a review of related works; Section 3 describes the different optimizers investigated in this study; and Section 4 presents the methodology of the study, including the experimental setup, summary of the dataset characteristics, and an overview of the CNN architectures studied. The experimental results and discussions are presented in Section 5. Section 6 concludes the paper. For clarity regarding the reminder of this paper, all acronyms adopted and used are provided in Table A1 of Appendix D.

3. Optimization Techniques

Optimizer is an important hyperparameter needed for training deep learning models They are broadly categorized into two for non-convex optimization problems, namely, non-adaptive learning rate (classical SGD), and adaptive learning rate optimization algorithms [5,32]. Although there is also a second-order optimization algorithm for convex problems, the main difference between convex and non-convex optimization problems is that for convex optimization problems there is only one global optimum, with no notion of local optima, whereas the non-convex optimization problem has multiple local optima across the neural network loss surface, with only one global optimum [32]. The focus of this paper is on first-order optimization algorithms for non-convex problems. This is because non-convex optimization problems are the most prevalent in neural network research. The nature of the loss surface for non-convex problems determines the complexity and difficulty in locating the global optimum [33]. An optimizer is therefore used to reduce the cost function which is calculated by cross-entropy, while the loss function is used to calculate the error which is the metric that indicates the efficiency of the model.
A brief overview of the examined optimizers is described as follows:
  • Non-adaptive SGD and its enhanced variants are based on the gradient descent approach. Gradient descent is a means to minimise an objective function f (x) parameterised by a model’s parameters x ℝ, by updating the parameters in the opposite direction of the gradient of the objective function x f ( x ) with regard to the parameters reaching the local minimum which is determined by the learning rate [3]. SGD uses a single training dataset, one row after another, and then followed by iteration adjustment of weights for each row. Oscillation because of stochastic noise in SGD is one of its problems, resulting from updates not capturing curvature information and thereby slowing down SGD when loss surface curvature is high. Momentum is a technique, when applied to SGD, that dampens these oscillations [32,34]. This is achieved by accelerating the gradient descent towards reducing the objective function across iterations through tailoring the accumulated velocity vector towards that objective. Given an objective function f (x), momentum is defined as [9]:
    v t + 1 = μ v t ε x f ( x t )
    x t + 1 = x t + v t + 1
    where ε = learning rate; μ = momentum coefficient; and x f ( x t ) = gradient at xt; when ε > 0 ; μ   ϵ   [ 0 ,   1 ] .
In SGD with momentum, the gradient-based velocity vector is first corrected/updated in the current position x t , and then a big step is taken based on this new velocity vector value. In contrast, in SGD with Nesterov momentum, the first step is taken along the direction of the velocity and then correction and update to the velocity vector value is made based on the new location x t + μ v t x t + 1 . Hence, Nesterov momentum is given (taking into account the new velocity vector update) as [9]:
v t + 1 = μ v t ε x f ( x t + μ v t )
x t + 1 = x t + v t + 1
  • ADAGrad was designed as an improvement to the non-adaptive SGD optimization when the gradient vectors are sparse, particularly in the online learning setting [35]. Generally, with the overall objective of achieving faster convergence time [11], most of the adaptive learning rate techniques developed over the years are variants of ADAGrad, as they seek to address some inherent and notable weaknesses in ADAGrad, such as the rapid decrease in the learning rate in non-convex settings due to the accumulated squared gradient over time [11].
  • RMSProp is a modification of ADAGrad to improve performance in non-convex situations. It is based on the idea that dividing the vector gradient by the root mean square for each weight improves learning [36].
  • ADAM combines the strengths of RMSProp and SGD with momentum [37]. It uses the squared gradient method as in RSMProp for scaling the learning rate as well as the moving average of the gradient leveraging on momentum [37]. ADAM also combines the advantages of ADAGrad and RMSProp, and works well in settings with sparse gradients and online settings [35,36]. This is the reason why ADAM can perform well across numerous problem domains and datasets.
  • ADADelta is a method developed to combat the issue with the accumulation of all past squared gradients in ADAGrad, which affects the learning rate by decreasing it towards zero and eventually terminating training. ADADelta is a remedy to prevent the continual decay of learning rates and to avoid the need to select the global learning rate manually [37,38].
  • ADAMax is a variation of ADAM that uses the L-infinity ( L ) norm rule to update present and past gradients, resulting in stable behavior. Given a generalised L p norm-based update rule, the ADAM algorithm becomes numerically unstable in situations for norms with large p values [37].
  • NADAM is a combination of RMSProp and Nesterov momentum. It builds on the work of [36] by incorporating Nesterov momentum into ADAM, rather than using only momentum to estimate the exponential moving average of the gradient [39], and thus making NADAM an improved variant of ADAM.
A brief explanation of the optimizers update rule is provided in Table 2:
Table 2. A brief explanation of optimizer update rules.

4. Materials and Methods

This section presents the experimental setup, summary, and characteristics of the datasets used, and the configurations of the CNN architectural models considered in the study.

4.1. Experimental Setup

The aim was to establish that different optimization algorithms can find different solutions in a non-convex scenarios. This subsection reports the empirical framework followed to study the impact of nine different optimizers on three different modified CNN architectures with varying depth, width and depth/width on three image classification problems. The experiments were conducted on Kaggle’s two CPU cores, 14 GB RAM with GPU in a Python environment using Keras framework with Tensorflow backend, and other sklearn libraries. The initial learning rate (lr) was fixed as lr = 3 × 10−4 for all the optimizers studied. We studied the optimizers using their default hyperparameter settings to reduce complexities and to observe the effects of these optimizers on the CNN models. The results were benchmarked against a simple CNN architecture studied in [14].

4.2. Dataset

In this work, we considered three popularly used machine learning benchmark image classification datasets—namely Cats and Dogs, Natural Images and Fashion MNIST—to evaluate the different optimization algorithms and the CNN models. A summary of the datasets is given in Table 3, and subsequently, a brief description of the datasets is provided.
Table 3. Summary of the image classification datasets.
  • Cats and Dogs: The dataset is an input size of 64 × 64 pixel colored images of cats and dogs. The training set has 25,000 images, containing 12,500 images of cats and 12,500 images of dogs, and a test set of 12,500 images [40].
  • Natural Images: The dataset consists of 6899 images of 150 × 150 pixel colored images from eight distinct classes, which include aircraft, cars, cats, dogs, flowers, fruit, motorbikes, and people, obtained from various sources used in [41] as the benchmark dataset in their work.
  • Fashion MNIST: The dataset comprises a training set of 60,000 images and a test set of 10,000 images. Each image is a 28 × 28-grayscale image, associated with a label from 10 classes, with 7000 images per class. Fashion MNIST is intended to serve as a direct drop-in replacement for the original MNIST dataset to benchmark machine learning algorithms [42].

4.3. CNN Architectural Models

Several CNN architectures have been designed over the past decade with various modifications aimed at improving the performance of problem-specific applications, beginning with LeNet in 1998 by Yann LeCun and AlexNet in 2012 and continuing to the very recent and modern high-resolution network (HRNetV2) CNN architecture [43,44]. Such modifications include increasing depth or width or both, adding regularization, hyperparameters tuning for optimizations and data augmentation, or using transfer learning [45]. In this section, we show the three different CNN architecture configurations examined in this study. The simple CNN configuration is the building block for the other three CNN configurations (with increasing depth, width and width/depth). The reason we use the simple CNN architecture is to allow us to clearly understand the effects of the different optimizers on the different CNN configurations, and is not necessary designed for performance improvement. The benchmark simple CNN configuration is presented in Table 4, with increasing depth (increasing the number of convolutional layers) in Table 5, with increasing width (increasing number of filters) in Table 6, and with increasing depth/width in Table 7. The CNN architecture is a fully connected CNN of depth d and width w for classification problems with c classes. The model consists of the input image (with RGB image channel number equal to 3 and grayscale channel number equal to 1), convolutional layers with stride 1, input filters and output filters, followed by a nonlinear ReLU activation function applied to the convolution layer output, MaxPooling layers with a pool size of 2 by 2 for down-sampling every feature map sub-sampling layer, leading to a reduction in the network parameters and dropout after each layer to handle over-fitting. Afterwards, the output of the previous layer is flattened. Finally come the FC layers which are the last stage layers that receive low-level features and create high-level abstraction. The probability classification scores are generated using either sigmoid or Softmax depending on the class labels.
Table 4. Configuration summary of the simple CNN denoted as sCNN.
Table 5. Configuration summary of CNN with increasing depth denoted as dCNN.
Table 6. Configuration summary of CNN with increasing width denoted as wCNN.
Table 7. Configuration summary of CNN with increasing depth and width denoted as d/wCNN.

5. Results and Discussions

In this section, we present the empirical results of the effect of different optimization algorithms on varying the network size (width, depth and width/depth) on three image classification datasets. The results obtained in the study by [14] are also presented as the benchmark to the three CNN architectures with varying network sizes. On the results presented in Table 8, Table 9 and Table 10, the best-performing results for each optimizer with the different CNN architectures and dataset problems are the bolded values.
Table 8. Results of optimizers with CNN architectures on Cats and Dogs (epoch = 500).
Table 9. Results of optimizers with CNN architectures on Natural Images (epoch = 150).
Table 10. Results of optimizers with CNN architectures on Fashion MNIST (epoch = 500).

5.1. Effect of Optimizers with Varying Network Sizes on Cats and Dogs

The results obtained for the Cats and Dogs dataset are reported in Table 8. ADAM, in combination with the depth/width CNN model, achieved the best performance in terms of validation accuracy across the three CNN architectures at 91.1%, followed closely by ADAMax (88.9%), RMSProp (87.7%) and NADAM (86.5%), although with variations for convergence time and loss. The other remaining optimizers did not have any significant performance improvements when compared to the simple CNN, except marginal improvements with validation convergence time for NADAM on depth CNN model and better validation loss for RSMSProp on depth/with CNN. The overall worst performers were ADADelta and vanilla SGD. Figure 1 shows plots of train and validation test accuracy and loss against epoch on the best-performing configuration with ADAM and increasing CNN depth/width. More details of all the experimental plots for CNN with varying network sizes for Cats and Dogs are reported in Figure A1 of Appendix A.1, Figure A5 of Appendix B.1 and Figure A7 of Appendix C.1. We noticed no significant improvement after about 500 epochs.
Figure 1. Plot of best-performing ADAM with increasing CNN depth and width on Cats and Dogs.

5.2. Effect of Optimizers with Varying Network Sizes on Natural Images

Table 9 reports the results obtained using the Natural Images dataset. ADAM exhibited superior performance with the wider CNN models in terms of validation accuracy at 98.1%, closely followed by NADAM at 97.6%, but at the expense of higher convergence time. ADADelta was still the worst performer in terms of validation accuracy across the CNN models, and in addition, no improvement was noticed compared to the results that were obtained on the simple CNN model. We noticed improved validation accuracy across all the optimizers on the wider CNN model, notably with ADAMax, SGD momentum and SGD Nesterov momentum. Figure 2 shows plots of accuracy and loss against epoch for the best-performing configuration with ADAM and increasing CNN width. In Figure A2 in Appendix A.2, Figure A5 in Appendix B.2, and Figure A8 in Appendix C.2, there are more details of all the experimental plots for CNN with varying network sizes for Natural Images. We noticed no improvement after about 150 epochs.
Figure 2. Plot of best-performing ADAM with increasing CNN width for Natural Images.

5.3. Effect of Optimizers with Varying Network Sizes on Fashion MNIST

Table 10 reports the results obtained for Fashion MNIST. NADAM was the overall best performer in terms of validation accuracy (83.1%) and loss (0.585) on the depth/width CNN model, followed by ADAM (80.4%) in second position, once again at the expense of higher convergence time. However, ADAM recorded better convergence time when compared to its closest rival, NADAM, on the depth CNN model, with lower accuracy and higher loss. ADADelta was the overall worst performer across all the CNN models for this dataset. Figure 3 shows plots of accuracy and loss against epoch on the best-performing configuration with NADAM and increasing CNN depth/width. In Figure A3 in Appendix A.3, Figure A6 in Appendix B.3, and Figure A9 in Appendix C.3, there are more details of all the experimental plots for CNN with varying network sizes for Fashion MNIST. We noticed no significant improvement after about 500 epochs.
Figure 3. Plot of best-performing NADAM with increasing CNN depth + width on Fashion MNIST.
In this study, the following inferences are drawn based on the results obtained:
  • Overall, all the optimizers showed improved performance, particularly ADAM with an accuracy of 91.1% with the deeper/wider CNN architecture for the two-class Cats and Dogs dataset problem. It is, however, worth noting that ADAMax with an accuracy of 88.9% performed marginally better than NADAM with an accuracy of 86.5% with the deeper/wider CNN architecture on Cats and Dogs. This indicates that the incorporated Nesterov momentum in NADAM had little effect on this dataset problem.
  • For Natural Images with eight classes, the results show that the wider network architecture generally had better validation accuracy across all the optimizers, especially ADAM with an accuracy of 98.1%, when compared to the simple, deeper, and deeper/wider networks. However, the improved validation accuracy was at the expense of higher convergence time, especially for the adaptive learning rate optimizers. The only exception was ADADelta, which generally performed poorly on this dataset.
  • It is worth noting that SGD momentum and SGD Nesterov momentum with the wider CNN architecture performed well on Natural Images with accuracies of 92.9% and 93.7%, respectively. This is consistent with previous studies suggesting that vanilla SGD optimizer can be improved with momentum and Nesterov momentum, and that the improved SGD can be competitive with the adaptive learning rate optimizers for some specific tasks, but with additional time to converge.
  • For Fashion MNIST, NADAM was the better performer with an accuracy of 83.1% on depth/width CNN. However, we noted that for the Fashion MNIST there was generally a reduced performance across all the optimizers combined with the CNN architectures. This could be related to the higher number of classes and the simple network design considered in this study.
  • The initial results obtained suggest that for the two-class problems, ADAM is a better optimizer to consider with a deeper/wider CNN architecture. For the larger number of classes, ADAM and NADAM with wider networks or deeper/wider network architectures irrespective of dataset size could be considered. However, because of the empirical evidence in the literature that the performance of deeper network tends to worsen after a certain critical point, in addition to increased computational cost due to increased parameterization, a careful balance between depth and width should be applied during the design of deeper/wider CNN architectures. For the majority of tasks, we suggest increasing the width while keeping the depth constant, especially when depth has reached a critical performance point.
  • The study demonstrated empirically that significant differences do exist in the performance outputs of different optimizers on varying depth, width and depth/width CNN models, depending on the dataset problem. This is most likely because of their distinctive characteristics and influence on the ability of the CNN models in learning internal representations.
  • Finally, from the results obtained, we observed that the optimizers exhibited significant differences in the performance outputs with the wide, deep and deep/wide models. However, the deeper architecture showed an increased risk of poor performance compared to wider and deeper/wider architecture during validation testing. This finding is consistent with previous studies.

6. Conclusions

The objective of the paper was to assess the effects of different optimizers on varying depth and width of CNN architectures. To this end, we conducted an empirical comparison of nine stochastic gradient-based optimization algorithms against three simple CNN architectures of varying depth, width and depth/width, using three image classification dataset problems. Our study reveals that for the Cats and Dogs data problems, ADAM performed remarkably well, much better than its closest rival, NADAM, even though ADAM had a slightly higher convergence time. For the Natural Images data problem, NADAM consistently performed better across all the CNN models. Overall, the worst performer by far was ADADelta across all the models and datasets. For the Fashion MNIST data problem, NADAM was the better performer, but once again, with a slightly higher convergence time than its closest rival ADAM. On the whole, considering the three image classification datasets evaluated, ADAM and NADAM with wider or deeper/wider CNN models were the better performers in terms of accuracy, but at the expense of higher convergence time, while the deeper CNN models generally showed increased performance risk. In this paper, we only looked at three datasets, default settings for the optimizers, and simple CNN architectures. An interesting future direction to pursue is to consider several image classification dataset problems in terms of structure, size, and the number of classes, in addition to hyperparameter tuning of the optimizers and modern CNN architectures. It will also be of interest to investigate the effect of noise from the data images on the prediction accuracies of CNN models, and how robust the CNN models are against noise. However, from the initial results and empirical evidence obtained in this study, we suggest a careful balance between increasing depth and width during the design of deeper/wider CNN architectures, and to try ADAM, NADAM, ADAMax, and SGD with Nesterov momentum as a starting point for image classification problems.

Author Contributions

Conceptualization, E.M.D.; methodology, E.M.D. and O.J.A.; software, E.M.D. and O.J.A.; validation, E.M.D., O.J.A. and B.T.; formal analysis, E.M.D., O.J.A. and B.T.; investigation, E.M.D., O.J.A. and B.T.; writing—original draft preparation, E.M.D.; writing—review and editing, E.M.D., O.J.A. and B.T.; supervision, B.T.; project administration, E.M.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Experimental Plots of CNN Architecture with Increasing Depth

Appendix A shows plots of Train and Validation Test accuracies and losses of the CNN models with varying depth for Cats and Dogs in Figure A1, Natural Images in Figure A2 and Fashion MNIST in Figure A3.

Appendix A.1. Cats and Dogs

Figure A1. Plots of Train and Validation Test accuracies and losses of the CNN model with varying depth for Cats and Dogs: (a) Vanilla SGD; (b) SGD momentum; (c) SGD Nesterov momentum; (d) ADAGrad; (e) RSMProp; (f) ADAM; (g) ADADelta; (h) ADAMax; (i) NADAM.

Appendix A.2. Natural Images

Figure A2. Plots of Train and Validation Test accuracies and losses of the CNN model with varying depth for Natural Images: (a) Vanilla SGD; (b) SGD momentum; (c) SGD Nesterov momentum; (d) ADAGrad; (e) RSMProp; (f) ADAM; (g) ADADelta; (h) ADAMax; (i) NADAM.

Appendix A.3. Fashion MNIST

Figure A3. Plots of Train and Validation Test accuracies and losses of the CNN model with varying depth for Fashion MNIST: (a) Vanilla SGD; (b) SGD momentum; (c) SGD Nesterov momentum; (d) ADAGrad; (e) RSMProp; (f) ADAM; (g) ADADelta; (h) ADAMax; (i) NADAM.

Appendix B. Experimental Plots of CNN Architecture with Increasing Width

Appendix B shows plots of Train and Validation Test accuracies and losses of the CNN models with varying width for Cats and Dogs in Figure A4, Natural Images in Figure A5, and Fashion MNIST in Figure A6.

Appendix B.1. Cats and Dogs

Figure A4. Plots of Train and Validation Test accuracies and losses of the CNN model with varying width for Cats and Dogs: (a) Vanilla SGD; (b) SGD momentum; (c) SGD Nesterov momentum; (d) ADAGrad; (e) RSMProp; (f) ADAM; (g) ADADelta; (h) ADAMax; (i) NADAM.

Appendix B.2. Natural Images

Figure A5. Plots of Train and Validation Test accuracies and losses of the CNN model with varying width for Natural Images: (a) Vanilla SGD; (b) SGD momentum; (c) SGD Nesterov momentum; (d) ADAGrad; (e) RSMProp; (f) ADADelta; (g) ADAMax; (h) NADAM.

Appendix B.3. Fashion MNIST

Figure A6. Plots of Train and Validation Test accuracies and losses of the CNN model with varying width for Fashion MNIST: (a) Vanilla SGD; (b) SGD momentum; (c) SGD Nesterov momentum; (d) ADAGrad; (e) RSMProp; (f) ADAM; (g) ADADelta; (h) ADAMax; (i) NADAM.

Appendix C. Experimental Plots of CNN Architecture with Increasing Depth/Width

Appendix C shows plots of Train and Validation Test accuracies and losses of the CNN models with varying depth/width for Cats and Dogs in Figure A7, Natural Images in Figure A8, and Fashion MNIST in Figure A9.

Appendix C.1. Cats and Dogs

Figure A7. Plots of Train and Validation Test accuracies and losses of the CNN model with varying depth/width for Cats and Dogs: (a) Vanilla SGD; (b) SGD momentum; (c) SGD Nesterov momentum; (d) ADAGrad; (e) RSMProp; (f) ADADelta; (g) ADAMax; (h) NADAM.

Appendix C.2. Natural Images

Figure A8. Plots of Train and Validation Test accuracies and losses of the CNN model with varying depth/width for Natural Images: (a) Vanilla SGD; (b) SGD momentum; (c) SGD Nesterov momentum; (d) ADAGrad; (e) RSMProp; (f) ADAM; (g) ADADelta; (h) ADAMax; (i) NADAM.

Appendix C.3. Fashion MNIST

Figure A9. Plots of Train and Validation Test accuracies and losses of the CNN model with varying depth/width for Fashion MNIST: (a) Vanilla SGD; (b) SGD momentum; (c) SGD Nesterov momentum; (d) ADAGrad; (e) RSMProp; (f) ADAM; (g) ADADelta; (h) ADAMax.

Appendix D. List of Acronyms

Appendix D shows the list of acronyms adopted and used in this study in Table A1.
Table A1. List of acronyms adopted in this paper.
Table A1. List of acronyms adopted in this paper.
AcronymDescription
VSGDstochastic gradient descent with vanilla
SGDMstochastic gradient descent with momentum
SGDNMstochastic gradient descent with Nesterov momentum
ADAGradAdaptive Gradient
RMSPropRoot Mean Square Propagation
ADAMAdaptive Moment Estimation
ADADeltaAdaptive Delta
ADAMaxAdaptive moment estimation Extension based on infinity norm
NADAMNesterov-accelerated Adaptive Moment Estimation
sCNNSimple Convolutional Neural Network
dCNNConvolutional Neural Network with varying depth
wCNNConvolutional Neural Network with varying width
d/wCNNConvolutional Neural Network with varying depth/width
LARSLayer-wise Adaptive Rate Scaling
LAMBLayer-wise Adaptive Moments optimizer for Batch training

References

  1. Papamakarios, G. Comparison of Stochastic Optimization Algorithms. University of Edinburgh: Edinburgh, Scotland, 2014. [Google Scholar]
  2. Hallen, R. A Study of Gradient-Based Algorithms. 2017. Available online: http://lup.lub.lu.se/student-papers/record/8904399 (accessed on 20 September 2022).
  3. Ruder, S. An overview of gradient descent optimization algorithms. arXiv 2016, arXiv:1609.04747. [Google Scholar]
  4. Shalev-Shwartz, S.; Shamir, O.; Shammah, S. Failures of gradient-based deep learning. In Proceedings of the 34th International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; pp. 3067–3075. [Google Scholar]
  5. Tschoepe, M. Beyond SGD: Recent Improvements of Gradient Descent Methods. Master’s Thesis, Technische Universität Kaiserslautern, Kaiserslautern, Germany, 2019. [Google Scholar] [CrossRef]
  6. Zhang, M.; Lucas, J.; Ba, J.; Hinton, G.E. Lookahead optimizer: K steps forward, 1 step back. In Advances in Neural Information Processing Systems; MIT: Cambridge, MA, USA, 2019; Volume 32. [Google Scholar]
  7. Zhu, A.; Meng, Y.; Zhang, C. An improved Adam Algorithm using look-ahead. In Proceedings of the 2017 International Conference on Deep Learning Technologies, Chengdu, China, 2–4 June 2017; Association for Computing Machinery: New York, NY, USA, 2017; pp. 19–22. [Google Scholar]
  8. Reddi, S.J.; Kale, S.; Kumar, S. On the convergence of adam and beyond. arXiv 2019, arXiv:1904.09237. [Google Scholar]
  9. Sutskever, I.; Martens, J.; Dahl, G.; Hinton, G. On the importance of initialization and momentum in deep learning. In Proceedings of the 30th International Conference on Machine Learning, Atlanta, GA, USA, 16–21 June 2013; pp. 1139–1147. [Google Scholar]
  10. Schaul, T.; Antonoglou, I.; Silver, D. Unit tests for stochastic optimization. arXiv 2013, arXiv:1312.6055. [Google Scholar]
  11. Liu, L.; Jiang, H.; He, P.; Chen, W.; Liu, X.; Gao, J.; Han, J. On the variance of the adaptive learning rate and beyond. arXiv 2019, arXiv:1908.03265. [Google Scholar]
  12. Rumelhart, D.E.; Hinton, G.E.; Williams, R.J. Learning Internal Representations by Error Propagation. In Parallel Distributed Processing: Explorations in the Microstructure of Cognition: Foundations; MITP: Mainz, Germany, 1987; pp. 318–362. [Google Scholar]
  13. Nielsen, M.A. Neural Networks and Deep Learning; Determination Press: San Francisco, CA, USA, 2015; Volume 25. [Google Scholar]
  14. Dogo, E.M.; Afolabi, O.J.; Nwulu, N.I.; Twala, B.; Aigbavboa, C.O. A comparative analysis of gradient descent-based optimization algorithms on convolutional neural networks. In Proceedings of the 2018 International Conference on Computational Techniques, Electronics and Mechanical Systems (CTEMS), Belgaum, India, 21–22 December 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 92–99. [Google Scholar]
  15. Joshua, A.O.; Nelwamondo, F.V.; Mabuza-Hocquet, G. Segmentation of optic cup and disc for diagnosis of glaucoma on retinal fundus images. In Proceedings of the 2019 Southern African Universities Power Engineering Conference/Robotics and Mechatronics/Pattern Recognition Association of South Africa (SAUPEC/RobMech/PRASA), Bloemfontein, South Africa, 28–30 January 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 183–187. [Google Scholar]
  16. Afolabi, O.J.; Mabuza-Hocquet, G.P.; Nelwamondo, F.V.; Paul, B.S. The use of U-Net lite and Extreme Gradient Boost (XGB) for glaucoma detection. IEEE Access 2021, 9, 47411–47424. [Google Scholar] [CrossRef]
  17. Yu, Y.; Liang, S.; Samali, B.; Nguyen, T.N.; Zhai, C.; Li, J.; Xie, X. Torsional capacity evaluation of RC beams using an improved bird swarm algorithm optimised 2D convolutional neural network. Eng. Struct. 2022, 273, 115066. [Google Scholar] [CrossRef]
  18. Yu, Y.; Samali, B.; Rashidi, M.; Mohammadi, M.; Nguyen, T.N.; Zhang, G. Vision-based concrete crack detection using a hybrid framework considering noise effect. J. Build. Eng. 2022, 61, 105246. [Google Scholar] [CrossRef]
  19. Wilson, A.C.; Roelofs, R.; Stern, M.; Srebro, N.; Recht, B. The marginal value of adaptive gradient methods in machine learning. In Advances in Neural Information Processing Systems; MIT: Cambridge, MA, USA, 2017; Volume 30. [Google Scholar]
  20. Belkin, M.; Hsu, D.; Ma, S.; Mandal, S. Reconciling modern machine-learning practice and the classical bias–variance trade-off. Proc. Natl. Acad. Sci. USA 2019, 116, 15849–15854. [Google Scholar] [CrossRef]
  21. Nichani, E.; Radhakrishnan, A.; Uhler, C. Do deeper convolutional networks perform better? In Proceedings of the International Conference on Machine Learning, Online, 18–24 July 2021. [Google Scholar]
  22. Nguyen, T.; Raghu, M.; Kornblith, S. Do wide and deep networks learn the same things? Uncovering how neural network representations vary with width and depth. arXiv 2020, arXiv:2010.15327. [Google Scholar]
  23. Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 6105–6114. [Google Scholar]
  24. Hastie, T.; Tibshirani, R.; Friedman, J.H.; Friedman, J.H. The Elements of Statistical Learning: Data Mining, Inference, and Prediction; Springer: New York, NY, USA, 2009; Volume 2, pp. 1–758. [Google Scholar]
  25. Yang, Z.; Yu, Y.; You, C.; Steinhardt, J.; Ma, Y. Rethinking bias-variance trade-off for generalization of neural networks. In Proceedings of the International Conference on Machine Learning, Online, 13–18 July 2020; pp. 10767–10777. [Google Scholar]
  26. Neyshabur, B. Towards learning convolutions from scratch. In Advances in Neural Information Processing Systems; MIT: Cambridge, MA, USA, 2020; Volume 33, pp. 8078–8088. [Google Scholar]
  27. Urban, G.; Geras, K.J.; Kahou, S.E.; Aslan, O.; Wang, S.; Caruana, R.; Richardson, M. Do deep convolutional nets really need to be deep and convolutional? arXiv 2016, arXiv:1603.05691. [Google Scholar]
  28. Nguyen, Q.; Hein, M. Optimization landscape and expressivity of deep CNNs. In Proceedings of the International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018; pp. 3730–3739. [Google Scholar]
  29. Keskar, N.S.; Socher, R. Improving generalization performance by switching from adam to sgd. arXiv 2017, arXiv:1712.07628. [Google Scholar]
  30. Choi, D.; Shallue, C.J.; Nado, Z.; Lee, J.; Maddison, C.J.; Dahl, G.E. On empirical comparisons of optimizers for deep learning. arXiv 2019, arXiv:1910.05446. [Google Scholar]
  31. Nado, Z.; Gilmer, J.M.; Shallue, C.J.; Anil, R.; Dahl, G.E. A large batch optimizer reality check: Traditional, generic optimizers suffice across batch sizes. arXiv 2021, arXiv:2102.06356. [Google Scholar]
  32. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT: Cambridge, MA, USA, 2016. [Google Scholar]
  33. Li, H.; Xu, Z.; Taylor, G.; Studer, C.; Goldstein, T. Visualizing the loss landscape of neural nets. arXiv 2017, arXiv:1712.09913. [Google Scholar]
  34. Bottou, L. Large-scale machine learning with stochastic gradient descent. In Proceedings of the COMPSTAT’2010: 19th International Conference on Computational Statistics, Paris, France, 22–27 August 2010; Physica-Verlag HD.: Heidelberg, Germany, 2010; pp. 177–186. [Google Scholar]
  35. Duchi, J.; Hazan, E.; Singer, Y. Adaptive subgradient methods for online learning and stochastic optimization. J. Mach. Learn. Res. 2011, 12, 2121–2159. [Google Scholar]
  36. Tieleman, T.; Hinton, G. Lecture 6.5-rmsprop: Divide the gradient by a running average of its recent magnitude. In Neural Networks for Machine Learning; COURSERA: Mountain View, CA, USA, 2012; Volume 4, pp. 26–31. [Google Scholar]
  37. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  38. Zeiler, M.D. Adadelta: An adaptive learning rate method. arXiv 2012, arXiv:1212.5701. [Google Scholar]
  39. Dozat, T. Incorporating Nesterov momentum into Adam. ICLR Workshop 2016, 1, 2013–2016. [Google Scholar]
  40. Elson, J.; Douceur, J.R.; Howell, J.; Saul, J. Asirra: A CAPTCHA that exploits interest-aligned manual image categorization. In Proceedings of the 14th ACM Conference on Computer and Communications Security (CCS), CATS/DOGS, Alexandria, VA, USA, 31 October–2 November 2007; Association for Computing Machinery, Inc.: New York, NY, USA, 2007; Volume 7, pp. 366–374. [Google Scholar]
  41. Roy, P.; Ghosh, S.; Bhattacharya, S.; Pal, U. Effects of degradations on deep neural network architectures. arXiv 2018, arXiv:1807.10108. [Google Scholar]
  42. Xiao, H.; Rasul, K.; Vollgraf, R. Fashion-mnist: A novel image dataset for benchmarking machine learning algorithms. arXiv 2017, arXiv:1708.07747. [Google Scholar]
  43. Khan, A.; Sohail, A.; Zahoora, U.; Qureshi, A.S. A survey of the recent architectures of deep convolutional neural networks. Artif. Intell. Rev. 2020, 53, 5455–5516. [Google Scholar] [CrossRef]
  44. Shrestha, A.; Mahmood, A. Review of deep learning algorithms and architectures. IEEE Access 2019, 7, 53040–53065. [Google Scholar] [CrossRef]
  45. Alzubaidi, L.; Zhang, J.; Humaidi, A.J.; Al-Dujaili, A.; Duan, Y.; Al-Shamma, O.; Farhan, L. Review of deep learning: Concepts, CNN architectures, challenges, applications, future directions. J. Big Data 2021, 8, 1–74. [Google Scholar] [CrossRef] [PubMed]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.