Next Article in Journal
A Fast Hybrid Beamforming Scheme for High-Speed Railway Communications
Previous Article in Journal
Modeling and Analysis of an Electromagnetic Fully Variable Valve Train with a Magnetorheological Buffer
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Evolutionary-Fuzzy-Integral-Based Convolutional Neural Networks for Facial Image Classification

1
Department of Computer Science and Information Engineering, National Chin-Yi University of Technology, Taichung 411, Taiwan
2
Department of Computer Science and Information Engineering, National Cheng Kung University, Tainan 701, Taiwan
3
Department of Electrical Engineering, National Formosa University, Yunlin 632, Taiwan
*
Author to whom correspondence should be addressed.
Electronics 2019, 8(9), 997; https://doi.org/10.3390/electronics8090997
Submission received: 23 July 2019 / Revised: 20 August 2019 / Accepted: 27 August 2019 / Published: 5 September 2019
(This article belongs to the Section Artificial Intelligence)

Abstract

:
Various optimization methods and network architectures are used by convolutional neural networks (CNNs). Each optimization method and network architecture style have their own advantages and representation abilities. To make the most of these advantages, evolutionary-fuzzy-integral-based convolutional neural networks (EFI-CNNs) are proposed in this paper. The proposed EFI-CNNs were verified by way of face classification of age and gender. The trained CNNs’ outputs were set as inputs of a fuzzy integral. The classification results were operated using either Sugeno or Choquet output rules. The conventional fuzzy density values of the fuzzy integral were decided by heuristic experiments. In this paper, particle swarm optimization (PSO) was used to adaptively find optimal fuzzy density values. To combine the advantages of each CNN type, the evaluation of each CNN type in EFI-CNNs is necessary. Three CNN structures, AlexNet, very deep convolutional neural network (VGG16), and GoogLeNet, and three databases, computational intelligence application laboratory (CIA), Morph, and cross-age celebrity dataset (CACD2000), were used in experiments to classify age and gender. The experimental results show that the proposed method achieved 5.95% and 3.1% higher accuracy, respectively, in classifying age and gender.

Graphical Abstract

1. Introduction

Image recognition technology has continued to develop chiefly due to deep learning technology. The origins of convolutional neural networks (CNNs) can be traced back to 1998. LeCun et al. [1] proposed the LeNet-5 model and used the back propagation (BP) algorithm to adjust the parameters of the neural networks; this model is a successful convolutional neural network even today. Later, a deeper network architecture, AlexNet [2], was proposed by Alex Krizhevsky and opened up the development of deep learning.
For the convolutional layers, AlexNet contains 60 million parameters and uses rectified linear unit (ReLU) as the activation function, which is different from LeNet, and it adds dropout, avoiding model overfitting. To enhance the abilities of each convolutional layer, Lin et al. [3] replaced conventional convolutional layers with multilayer perceptron (MLP) convolutional layers. GoogLeNet [4] improved the MLP convolutional layer by using a 1 × 1 convolutional kernel which achieves cross-channel message exchange and reduces dimensions. Due to the degradation problem, these networks cannot learn the features in deeper layer networks. In order to solve this problem, ResNet [5] was proposed to map the low-level features directly to high-level networks, which means that instead of learning from the beginning, the deeper layer networks contain the previous few layers’ representation ability.
The latter part of a CNN is a classifier, which includes the most parameters in the CNN. Based on the concept of dropout, Wan et al. [6] proposed DropConnect, which closes neurons randomly, and its performance is slightly better than the dropout method. Network in Network [3] changed fully connected layers to global average pooling. Since global average pooling has no parameters to train, it can reduce the burden of the networks without the problem of overfitting.
For the activation function, sigmoid, tanh, and other saturation functions were used in the early networks. However, gradient disappearance problems occur in the deeper layer networks; rectified linear unit (ReLU) [2] was therefore proposed as the activation function. It reduces the calculation time and makes the model more resistant to overfitting. However, some neurons cannot be triggered using ReLU; thus, some parameters cannot be updated. There are many improved activation functions used in CNNs, such as Leaky ReLU [7], parametric rectified linear unit (PReLU) [8], randomized leaky ReLU (RReLU), and Swish [9].
The most common optimization method in the training process is stochastic gradient descent (SGD), which randomly selects a set of training samples each time for training. This method usually learns effectively; however, it relies on the learning rate setting and the training time is relatively long. Some improved optimization methods [10,11,12] have been proposed in order to solve these problems. The adaptive learning rate was proposed by Adam [13] and can predict a result faster.
The aforementioned architectures and optimization methods have different advantages and features in different applications. Kornblith et al. [14] questioned whether the results attained from ImageNet can be used in other problems, and the result was confirmed, which proves that different network architectures have their representation abilities. Therefore, maximizing the performance of systems by combining multiple features or multiple classifiers has become a popular topic. Classifier selection and classifier interaction modules within a Bayesian framework were used to recognize an object [15]. A multiple classifier system that incorporates a global optimization technique based on a genetic algorithm which adopts the weighted majority vote by maximizing the performance of the system was tested on a handwritten digit recognition problem [16]. Based on these ideas, in order to leverage advantages from different CNNs, the integration of multiple convolutional neural networks is proposed in this paper. The fuzzy integral is used to evaluate different CNNs in this paper; however, instead of using the conventional fuzzy integral which sets fuzzy density values by heuristic experiments, an evolutionary fuzzy integral is proposed in the paper which chooses the fuzzy density adaptively by particle swarm optimization (PSO), and the final classification results are computed using both Sugeno’s and Choquet’s rules. Three CNN structures, AlexNet, very deep convolutional neural network (VGG16), and GoogLeNet, and three databases, computational intelligence application laboratory (CIA), Morph, and cross-age celebrity dataset (CACD2000), were used in experiments to classify age and gender.
In this study, the major contributions of the proposed evolutionary-fuzzy-integral-based CNNs (EFI-CNNs) are as follows:
  • Integrating multiple CNNs is proposed to leverage advantages from different CNNs.
  • An evolutionary fuzzy integral is proposed to choose optimal fuzzy density values by PSO, avoiding the effects of manual setting.
  • The experimental results indicate that the proposed method exhibited superior accuracy compared to other methods.
The rest of this paper is organized as follows. Section 2 introduces the conventional CNN structure. The proposed EFI-CNNs are described in Section 3. Section 4 presents the experimental results of three facial databases, and Section 5 gives the conclusions.

2. Overview of CNNs

Conventional CNNs consist of two parts, as shown in Figure 1. The first part contains the steps of convolution and pooling which are used for feature extraction. The second part is a classifier which uses fully connected layers. Conventional CNNs still use feature extraction plus a classifier; however, lately, CNNs have replaced fully connected layers with average pooling. This can reduce the large number of required parameters and lower the degree of overfitting as well.
In the following sections, the feature extraction steps, convolutional layer, pooling, and activation function will be introduced.

2.1. Convolutional Layer

Convolution is the concept of a receptive field in the feature extraction process. It is used by many conventional filters, such as Sobel and Gabor. The convolution kernel is used on the input matrix like a sliding window, as shown in Figure 2. A two-dimensional convolution operation with a moving step equal to 1 is presented in Equation (1):
Y I J = i = 0 K w j = 0 K h x ( I + i 1 ) ( J + j 1 ) k i j ,
where Y I J is an output matrix, and K w and K h are the width and height of a convolution kernel, respectively. In general, K w = K h represents a square convolution kernel, x i j is the input matrix, and k i j is the weight of the convolution kernel which needs to be updated during training. In order to maintain a constant size after convolution, zero-padding is used at the edge of the input matrix.

2.2. Pooling

After the convolution operation, the extracted features can theoretically be classified directly. However, this requires a huge number of parametric operations, which makes the training process difficult and prone to overfitting. Pooling is one way to reduce the dimensions.
The pooling process applies a mask operation on the input matrix with a sliding window. During the process, the moving step is equal to the width of the convolution kernel. Each element is calculated only one time. Therefore, an N × N mask can lower the input feature matrix 1/N times to achieve the effect of reducing the dimensions.
There are two pooling operations: Max pooling and average pooling. Max pooling takes the maximum value in a region Rij and ignores other values. Average pooling calculates the average value in a region. The max pooling function is expressed as
a k i j = max ( p , q ) R i j ( a k p q ) ,
and the average pooling function is shown by the following equation:
a k i j = 1 | R i j | ( p , q ) R i j a k p q ,
where akij is the output activation of the kth feature map at (i,j), akpq is the input activation at (p,q) within Rij, and |Rij| is the size of the pooling region [17]. In order to retain the most prominent features, max pooling is the pooling method most often used during the convolution process. In contrast, global average pooling is usually applied in the output layer, because when classifying a target, average pooling contains physical meaning in the calculation process; therefore, a larger target can get a higher output value in the calculation.

2.3. Activation Function

The purpose of the activation function is to get nonlinear outputs from linearly combined networks. The disadvantage of sigmoid is that when the network gets deeper, gradient disappearance problems occur during back propagation. Therefore, rectified linear unit (ReLU) is used herein as the activation function. The definition of ReLU is illustrated as follows:
f ( x ) = { 0 ,   i f   x < 0 x ,   i f   x 0 ,
where if input x is smaller than 0, then the output is 0. If input x is bigger than 0, then the output is x.

3. The Proposed EFI-CNNs

In order to leverage advantages from different CNNs, the integration of multiple convolutional neural network architectures with different optimal methods through the evolutionary fuzzy integral is proposed in this paper. The flowchart of an EFI-CNN is depicted in Figure 3.
First, image preprocessing is required based on the dataset in use. Second, we input images to train the CNN models with different CNN architectures or different optimization methods. Third, we choose optimal fuzzy density values based on the best fitness value via the particle swarm optimization (PSO) method. Later, we use the obtained optimal fuzzy density values to calculate the fuzzy measure. Then, we sort the classifiers and define the set A for calculating Sugeno and Choquet, two fuzzy integral rules. At the end, the result with higher accuracy from either Sugeno or Choquet is chosen.
As Figure 3 shows, the outputs are obtained from the CNNs and treated as the inputs of the evolutionary fuzzy integral. The evolutionary fuzzy integral is described as follows.
Assuming X = { x i } i = 1 : n represents a set of n classifiers (in this study, n was set to 3), the fuzzy measure g ( { x i } ) can be considered as the worth value of subset { x i } , and its value is between 0 and 1. The fuzzy measure must fit the three following conditions:
(1)
g ( X ) = 1 represents that the outputs of all classifiers are consistent; the results can be fully trusted.
(2)
g(∅) = 0 represents that the outputs of all classifiers are not considered; the result has no reference value.
(3)
The fuzzy measure is an increasing monotonic function:
If   A B X , then   0 g ( A ) g ( B ) 1 .
When there is only one element in set X, the fuzzy measure g(X) is called the fuzzy density. Before calculating other fuzzy measures, the fuzzy density must be decided first. In the conventional fuzzy integral, the fuzzy density is set by users based on their experience. In general, the fuzzy density represents the worth value of the outputs in each network classifier. Some studies have used the accuracy of non-training data as the fuzzy density [18]; however, the accuracy of non-training data is usually not the optimal fuzzy density. The relation between fuzzy measure and fuzzy density is shown in Figure 4.
In this paper, the fuzzy density was chosen by particle swarm optimization (PSO) [19], avoiding the effects of manual setting. PSO was derived from the behavior patterns of birds. Flying birds are interpreted as countless particles moving continuously in the solution space and undergoing multiple iterations to find the best solution.
The fuzzy density of the evolutionary fuzzy integral is n × k floating point numbers, where n is the number of classifiers, i.e., the number of CNNs in the system, and k is the number of categories in a classification problem. The dimension of the solution space is n × k and ranges from 0 to 1.
Figure 5 shows the flowchart of the particle swarm optimization (PSO). The PSO is used for determining the fuzzy density in fuzzy integral. Initially, M number of particles p are randomly generated in the solution space, and each particle represents a set of fuzzy densities, as shown in Figure 6.
The fitness value of each particle is evaluated as follows:
F i t n e s s = T P T P + F P ,
where TP is the number of samples that are predicted correctly, and FP is the number of samples that are prejudged to be wrong.
Among the whole particle swarm, the best fitness value among all the particles is denoted G b e s t , and the best fitness value in one particle during an iteration is denoted P b e s t i for i = { 1 , 2 , M } . When updating the position of a particle, the particle’s fitness value will be compared with G b e s t and P b e s t i . If the particle’s current location is better than its historical locations, the particle will be updated and recorded. The location update formula for a particle swarm is as follows:
v i d ( t ) = w v i d ( t 1 ) + c 1 r a n d (   ) ( P b e s t i p i d ( t ) ) + c 2 r a n d (   ) ( G b e s t p i d ( t ) ) ,
p i d ( t + 1 ) = p i d ( t ) + v i d ( t ) ,
where v i d ( t ) represents the velocity of the ith particle in the dth dimension at the tth generation, p i d ( t ) is the location of the ith particle in the dth dimension at the tth generation, and r a n d ( ) is a random number from 0 to 1. The parameter w is linearly decreasing from 0.8 to 0.2 according to the number of generations, and c 1 and c 2 are usually both set to 2.
In Equations (7) and (8), in order to find the optimal fuzzy density, each particle changes its location according to its past experiences and group experiences. The searching process stops when the maximum generation number is reached. After calculating the fuzzy density, the fuzzy measure can be determined. The fuzzy measure formula is as follows:
g ( A B ) = g ( A ) + g ( B ) + λ g ( A ) g ( B ) ,   where   A ,   B X ,
λ + 1 = i = 1 n ( 1 + λ g i ) .
The value of λ is equal to the output dimension in the evolutionary fuzzy integral, which means that each output category can calculate one λ , where λ ( 1 , ) and with the following characteristics:
(1)
If   i = 1 n g i = 1 ,   then   λ = 0 ;
(2)
If   i = 1 n g i < 1 ,   then   λ > 0 ;
(3)
If   i = 1 n g i < 1 ,     then 1 λ > 0 .
There are many methods for calculating the fuzzy integral, such as Sugeno [20], Choquet [21], ordered weighted averaging AND operator (OWA-AND) [22], ordered weighted averaging OR operator (OWA-OR) [23], and Fuzzy min–max [24]. The Sugeno and Choquet methods were chosen in this study. The steps to calculate Sugeno and Choquet fuzzy integrals are as follows:
First, we sort the n classifiers according to their output:
h ( x π 1 ) h ( x π 2 ) h ( x π j ) h ( x π n ) ,
where π j represents the jth largest output belonging to a classifier. For example, π 2 = 3 reveals that the output of the third classifier is the second largest number in all the outputs. h ( x π 1 ) indicates the maximum output value of the classifier.
Second, we define set A i as follows:
A 0 = { } , A i = A i 1 + { x π i } .
Lastly, we calculate the Sugeno and Choquet fuzzy integrals as follows.
Sugeno FI:
Y s = i = 1 N ( h ( x π i ) g ( A i ) ) .
Choquet FI:
Y c = i = 1 n h ( x π i ) ( g ( A i ) g ( A i 1 ) ) .
The two fuzzy integrals are calculated at the same time; the fuzzy integral with higher accuracy is then chosen.

4. Experimental Results

In order to evaluate the proposed method of EFI-CNNs, three CNNs—AlexNet [2], VGG16 [25], and GoogLeNet [4]—were used to experiment on face classification by both gender and age range with three popular gradient descent optimization algorithms—stochastic gradient descent (SGD), adaptive subgradient (Adagrad), and adaptive moment estimation (Adam) [26]. In addition, to prove that the evolutionary fuzzy integral works, the conventional fuzzy integral (FI) was also compared with the experimental results.
To evaluate the recognition rate of fuzzy integrals objectively, the CIA, Morph, and CACD2000 face databases were used in this study. Each database was divided into three parts: Training data, verified data, and test data. Only training data were involved in the convolutional neural networks. Then, the verification data were used to adjust the fuzzy density evolutionally. Finally, we used the test data to compare the final accuracy. In the evolution of the fuzzy density, the maximum number of iterations was set to 1000.

4.1. CIA Database

The CIA database is a small Chinese face database collected by a research laboratory. Some examples from the CIA database are shown in Figure 7. The age distribution is from 6 to 80 years old. The numbers of images per gender and age range are shown in Table 1 and Table 2. Two identified problems, age and gender, were classified using the different CNN architectures and different optimization methods for training. The results are shown in Table 3 and Table 4. Table 5 reveals the results using EFI-CNN to identify age and gender. The results explain that in the examined gender and age ranges, EFI has higher recognition rates.

4.2. Morph Database

The Morph database is a Western-based face database [27]. It contains 55,000 face photos taken of 13,000 people. Figure 8 displays some photos from the Morph database. The age distribution is between 16 and 77 years old. The average interval between photos of each person is 164 days, and there is no continuous shooting. The data on age ranges and gender are shown in Table 6 and Table 7. Table 8 and Table 9 illustrate the age ranges and gender accuracy in classification of the Morph database using various CNNs. Table 10 illustrates the results of EFI-CNN on the Morph database. From the classification results of age ranges, the different CNN architectures with EFI have a better recognition rate, and the gender classification accuracy reached 98.56%.

4.3. CACD2000 Database

The CACD2000 database [28] consists of 160,000 photos taken of 2000 celebrities at different times. Five photos from the CACD2000 database are displayed in Figure 9. Table 11 shows that the age distribution is from 16 to 62 years old. In this database, there is no gender label, so no gender classification was performed in this experiment. The celebrities in the database are mainly non-Asian people. Each photo was taken in different lighting and contains some noise. Table 12 shows the accuracy of CACD2000 database age classification using different CNNs. Since the photo source of the CACD2000 database was not manually filtered, much noise is mixed in, so the age recognition rates were lower than those for the CIA database and Morph. Table 13 reveals the identification results by EFI-CNN on the CACD2000 database. From the age group classification results, CNNs with different architectures combined with EFI had a better identification rate—7% higher than that of GoogLeNet CNNs.
Further experiments with the databases UTKFace, face and gesture recognition network (FGNET), internet movie database – Wikipedia (IMDB-WiKi) were also implemented to support the proposed method. Different architectures LeNet, AlexNet, and GoogLeNet were used but with the same optimization methods mentioned above and the same parameter settings during the experiments. The age range classification results from the three datasets are listed in Table 14. The gender classification results from the UTKFace database and IMDB-WIKI database are displayed in Table 15. Both classification results reveal that integrating multiple CNNs with EFI allows for better accuracy than a single CNN.

4.4. Discussions

From Figure 10, we can see that the method proposed in this paper can effectively improve age and gender identification. Figure 11, Figure 12 and Figure 13 illustrate one of the test images from each of the three databases. The results of these images are shown in Table 16, Table 17 and Table 18, and there are two wrong CNN predictions. This means that if the system uses a conventional voting concept, the prediction will be wrong. Instead of using a voting concept, the FI mechanism can get the correct prediction.
Figure 14 illustrates the accuracy of EFI combined with multiple CNNs. However, combining more than three CNNs does not significantly improve accuracy. This means that we cannot continuously increase the number of CNNs to improve the accuracy of the system, which will greatly increase the demand for hardware.

5. Conclusions

In this paper, we utilized EFI-CNNs for age and gender classification of faces from the CIA, Morph, and CACD2000 databases based on fuzzy integral theory. The outputs acquired from trained CNNs were used as input to the EFI, and the final fuzzy integral rule was chosen from either Sugeno or Choquet during testing. Particle swarm optimization was applied to automatically search for the optimal fuzzy density value. The experimental results show that the proposed method obtains the best result when compared with the CNNs AlexNet, VGG16, and GoogLeNet. The proposed method improved the accuracy rates of age and gender classification by 5.95% and 3.1%, respectively. To further validate the proposed method, experiments and analysis were performed using more databases UTKFace, FGNET, and IMDB-WiKi. The experimental results show that the proposed method has better accuracy than other methods. In the future work, how to automatically determine the number of CNN and adjust the combination of CNN is a research direction that can effectively improve the classification accuracy.

Author Contributions

Conceptualization and methodology, C.-J.L., C.-C.S. and S.-H.W.; software, C.-J.L. and C.-H.L.; writing–original draft preparation, C.-J.L. and C.-H.L.

Funding

This research was funded by the Ministry of Science and Technology of the Republic of China, Taiwan (no. MOST 108-2634-F-005-001).

Conflicts of Interest

The authors declare that there are no conflicts of interest regarding the publication of this manuscript.

References

  1. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based Learning Applied to Document Recognition; Wiley-IEEE Press: Hoboken, NJ, USA, 1998; Volume 86, pp. 2278–2323. [Google Scholar]
  2. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. In Proceedings of the 25th International Conference on Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012; pp. 1097–1105. [Google Scholar]
  3. Lin, M.; Chen, Q.; Yan, S. Network in Network. In Proceedings of the International Conference on Learning Representations, Banff, AB, Canada, 14–16 April 2014. [Google Scholar]
  4. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going Deeper with Convolutions. In Proceedings of the Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  5. Wu, S.; Zhong, S.; Liu, Y. Deep Residual Learning for Image Steganalysis. Multimed. Tools Appl. 2018, 77, 10437–10453. [Google Scholar] [CrossRef]
  6. Wan, L.; Zeiler, M.; Zhang, S.; LeCun, Y.; Fergus, R. Regularization of Neural Networks Using Dropconnect. In Proceedings of the 30th International Conference on Machine Learning, Atlanta, GA, USA, 16–21 June 2013; pp. 1058–1066. [Google Scholar]
  7. Maas, A.L.; Hannun, A.Y.; Ng, A.Y. Rectifier Nonlinearities Improve Neural Network Acoustic Models. In Proceedings of the International Conference on Machine Learning, Atlanta, GA, USA, 16–21 June 2013. [Google Scholar]
  8. He, K.; Zhang, X.; Ren, S.; Sun, J. Delving Deep into Rectifiers: Surpassing Human-level Performance on Imagenet Classification. In Proceedings of the 2015 IEEE International Conference on Computer Vision, Washington, DC, USA, 7–13 December 2015; pp. 1026–1034. [Google Scholar]
  9. Ramachandran, P.; Zoph, B.; Le, Q.V. Searching for Activation Functions. In Proceedings of the 6th International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  10. Asness, C.S.; Moskowitz, T.J.; Pedersen, L.H. Value and Momentum Everywhere. J. Financ. 2013, 68, 929–985. [Google Scholar] [CrossRef] [Green Version]
  11. Duchi, J.; Hazan, E.; Singer, Y. Adaptive Subgradient Methods for Online Learning and Stochastic Optimization. J. Mach. Learn. Res. 2011, 12, 2121–2159. [Google Scholar]
  12. Neural Networks for Machine Learning—Lecture 6a—Overview of Mini-Batch Gradient Descent. Available online: https://www.cs.toronto.edu/~hinton/coursera/lecture6/lec6.pdf (accessed on 27 August 2019).
  13. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. In Proceedings of the International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  14. Kornblith, S.; Shlens, J.; Le, Q.V. Do Better ImageNet Models Transfer Better? In Proceedings of the Computer Vision and Pattern Recognition, San Diego, CA, USA, 16–20 June 2019. [Google Scholar]
  15. Tang, J.; Wen, G. Object Recognition via Classifier Interaction with Multiple Features. In Proceedings of the 8th International Conference on Intelligent Human-Machine Systems and Cybernetics, Hangzhou, China, 27–28 August 2016. [Google Scholar]
  16. De Stefano, C.; Della Cioppa, A.; Marcelli, A. An adaptive weighted majority vote rule for combining multiple classifiers. In Proceedings of the 16th International Conference on Pattern Recognition, Quebec City, QC, Canada, 11–15 August 2002. [Google Scholar]
  17. Williams, T.; Li, R. Wavelet Pooling for Convolutional Neural Networks. In Proceedings of the 6th International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  18. Tong, B.; Liu, Y. An Speech and Face Fusion Recognition Method Based on Fuzzy Integral. In Proceedings of the IEEE International Conference on Robotics and Biomimetics, Qingdao, China, 3–7 December 2016. [Google Scholar]
  19. Pandey, S.; Wu, L.; Guru, S.; Buyya, R. A Particle Swarm Optimization-based Heuristic for Scheduling Workflow Applications in Cloud Computing Environments. In Proceedings of the 24th IEEE International Conference on Advanced Information Networking and Applications, Perth, Australia, 20–23 April 2010. [Google Scholar]
  20. Abbaszadeh, S.; Eshaghi, M.; de la Sen, M. The Sugeno fuzzy integral of log-convex functions. J. Inequal. Appl. 2015, 2015, 362. [Google Scholar] [CrossRef]
  21. Torra, V.; Narukawa, Y. The interpretation of fuzzy integrals and their application to fuzzy systems. Int. J. Approx. Reason. 2006, 41, 43–58. [Google Scholar] [CrossRef] [Green Version]
  22. Cho, S.B. Fuzzy Aggregation of Modular Neural Networks with Ordered Weighted Averaging Operators. Int. J. Approx. Reason. 1995, 13, 359–375. [Google Scholar]
  23. Cheng, C.H.; Chen, C.T.; Huang, S.F. Combining fuzzy integral with order weight average (OWA) method for evaluating financial performance in the semiconductor industry. Afr. J. Bus. Manag. 2012, 6, 6358–6368. [Google Scholar]
  24. Mesiar, R.; Mesiarova, A. Fuzzy integrals and linearity. Int. J. Approx. Reason. 2008, 47, 352–358. [Google Scholar] [CrossRef] [Green Version]
  25. Liu, S.; Deng, W. Very Deep Convolutional Neural Network Based Image Classification Using Small Training Sample Size. In Proceedings of the 3rd IAPR Asian Conference on Pattern Recognition, Kuala Lumpur, Malaysia, 3–6 November 2015. [Google Scholar]
  26. Ruder, S. An overview of gradient descent optimization algorithms. arXiv 2017, arXiv:1609.04747. [Google Scholar]
  27. Morph Database. Available online: https://ebill.uncw.edu/C20231_ustores/web/classic/store_main.jsp?STOREID=4 (accessed on 27 August 2019).
  28. Chen, B.C.; Chen, C.S.; Hsu, W.H. Face recognition and retrieval using cross-age reference coding with cross-age celebrity dataset. IEEE Trans. Multimed. 2015, 17, 804–815. [Google Scholar] [CrossRef]
Figure 1. The structure of a convolutional neural network (CNN).
Figure 1. The structure of a convolutional neural network (CNN).
Electronics 08 00997 g001
Figure 2. Procedure of convolution.
Figure 2. Procedure of convolution.
Electronics 08 00997 g002
Figure 3. Flowchart of an evolutionary-fuzzy-integral (EFI)-based CNN.
Figure 3. Flowchart of an evolutionary-fuzzy-integral (EFI)-based CNN.
Electronics 08 00997 g003
Figure 4. Relation between fuzzy measure and fuzzy density.
Figure 4. Relation between fuzzy measure and fuzzy density.
Electronics 08 00997 g004
Figure 5. Flowchart of the particle swarm optimization.
Figure 5. Flowchart of the particle swarm optimization.
Electronics 08 00997 g005
Figure 6. Encoding of particles.
Figure 6. Encoding of particles.
Electronics 08 00997 g006
Figure 7. Examples from the computational intelligence application laboratory (CIA) face database.
Figure 7. Examples from the computational intelligence application laboratory (CIA) face database.
Electronics 08 00997 g007
Figure 8. Examples from the Morph face database.
Figure 8. Examples from the Morph face database.
Electronics 08 00997 g008
Figure 9. Examples from the cross-age celebrity dataset (CACD2000) face database.
Figure 9. Examples from the cross-age celebrity dataset (CACD2000) face database.
Electronics 08 00997 g009
Figure 10. Accuracy in different CNNs on the three databases.
Figure 10. Accuracy in different CNNs on the three databases.
Electronics 08 00997 g010
Figure 11. An example photo from the CIA database.
Figure 11. An example photo from the CIA database.
Electronics 08 00997 g011
Figure 12. An example photo from the Morph database.
Figure 12. An example photo from the Morph database.
Electronics 08 00997 g012
Figure 13. An example photo from the CACD2000 database.
Figure 13. An example photo from the CACD2000 database.
Electronics 08 00997 g013
Figure 14. Age range recognition accuracy in multiple (1~5) CNNs with EFI.
Figure 14. Age range recognition accuracy in multiple (1~5) CNNs with EFI.
Electronics 08 00997 g014
Table 1. Age ranges in the CIA database.
Table 1. Age ranges in the CIA database.
Class<1213~1920~45>45
Quantity313592934248
Table 2. Genders in the CIA database.
Table 2. Genders in the CIA database.
ClassMaleFemale
Quantity10801008
Table 3. Age classification results of the CIA database by the three CNNs.
Table 3. Age classification results of the CIA database by the three CNNs.
AgeAlexNet [2]VGG16 [25]GoogLeNet [4]
SGD69.2669.8269.31
Adagrad63.4867.6668.43
Adam69.1870.5769.78
SGD—stochastic gradient descent; Adagrad—adaptive subgradient; Adam—adaptive moment estimation.
Table 4. Gender classification results of the CIA database by the three CNNs.
Table 4. Gender classification results of the CIA database by the three CNNs.
GenderAlexNet [2]VGG16 [25]GoogLeNet [4]
SGD84.2984.3584.30
Adagrad81.7882.1083.52
Adam83.2883.4785.02
Table 5. Classification results of the CIA database by the different methods.
Table 5. Classification results of the CIA database by the different methods.
MethodsAgeGender
AlexNet (SGD) [2]69.2684.29
VGG16 (Adam/SGD) [25]70.5784.35
GoogLeNet (Adam) [4]69.7885.02
AlexNet (SGD + SGD + SGD) with EFI72.1786.47
VGG16 (Adam + Adam + Adam) with EFI71.2687.28
GoogLeNet (Adam + Adam + Adam) with EFI72.6186.10
AlexNet (SGD + Adagrad + Adam) with FI71.9285.37
VGG16 (SGD + Adagrad + Adam) with FI72.6785.47
GoogLeNet (SGD + Adagrad + Adam) with FI72.5185.97
AlexNet (SGD) + VGG16 (Adam) + GoogLeNet (Adam) with FI74.3686.66
AlexNet (SGD + Adagrad + Adam) with EFI73.3487.39
VGG16 (SGD + Adagrad + Adam) with EFI74.4888.51
GoogLeNet (SGD + Adagrad + Adam) with EFI73.9189.85
AlexNet (SGD) + VGG16 (Adam) + GoogLeNet (Adam) with EFI76.0290.61
Table 6. Age ranges in the Morph database.
Table 6. Age ranges in the Morph database.
Class<2525~3536~45>45
Quantity16,63214,14115,6098752
Table 7. Genders in the Morph database.
Table 7. Genders in the Morph database.
ClassMaleFemale
Quantity46,6458489
Table 8. Age classification results of the Morph database by the three CNNs.
Table 8. Age classification results of the Morph database by the three CNNs.
AgeAlexNet [2]VGG16 [25]GoogLeNet [4]
SGD70.8967.4970.58
Adagrad69.3568.1269.67
Adam68.0770.4768.63
Table 9. Gender classification results of the Morph database by the three CNNs.
Table 9. Gender classification results of the Morph database by the three CNNs.
GenderAlexNet [2]VGG16 [25]GoogLeNet [4]
SGD96.5597.4096.91
Adagrad97.5796.8197.30
Adam95.7197.3898.04
Table 10. Classification results of the Morph database by the different methods.
Table 10. Classification results of the Morph database by the different methods.
MethodsAgeGender
AlexNet (SGD/Adagrad) [2]70.8997.57
VGG16 (Adam/SGD) [25]70.4797.40
GoogLeNet (SGD/Adagrad) [4]70.5898.04
AlexNet (SGD + SGD + SGD) with EFI71.3197.64
VGG16 (Adam + Adam + Adam) with EFI72.8598.02
GoogLeNet (Adam + Adam + Adam) with EFI72.7798.56
AlexNet (SGD + Adagrad + Adam) with FI72.5297.75
VGG16 (SGD + Adagrad + Adam) with FI72.2798.03
GoogLeNet (SGD + Adagrad + Adam) with FI72.7298.22
AlexNet (SGD) + VGG16 (Adam) + GoogLeNet (Adam) with FI74.4898.28
AlexNet (SGD + Adagrad + Adam) with EFI74.9297.83
VGG16 (SGD + Adagrad + Adam) with EFI74.6698.28
GoogLeNet (SGD + Adagrad + Adam) with EFI74.5298.56
AlexNet (SGD) + VGG16 (Adam) + GoogLeNet (Adam) with EFI76.6898.52
Table 11. Age ranges in the CACD2000 database.
Table 11. Age ranges in the CACD2000 database.
Class<2525~3536~45>45
Quantity29,02938,59542,58853,234
Table 12. Age classification results of the CACD2000 database by the three CNNs.
Table 12. Age classification results of the CACD2000 database by the three CNNs.
AgeAlexNet [2]VGG16 [25]GoogLeNet [4]
SGD59.0858.5363.11
Adagrad61.9557.3160.13
Adam62.1361.3961.41
Table 13. Classification results of the CACD2000 database by the different methods.
Table 13. Classification results of the CACD2000 database by the different methods.
MethodsAge
AlexNet (Adam) [2]62.13
VGG16 (Adam) [25]61.39
GoogLeNet (SGD) [4]63.11
AlexNet (Adam + Adam + Adam) with EFI64.93
VGG16 (Adam + Adam + Adam) with EFI63.58
GoogLeNet (SGD + SGD + SGD) with EFI65.55
AlexNet (SGD + Adagrad + Adam) with FI63.33
VGG16 (SGD + Adagrad + Adam) with FI63.25
GoogLeNet (SGD + Adagrad + Adam) with FI64.53
AlexNet (Adam) + VGG16 (Adam) + GoogLeNet (SGD) with FI66.41
AlexNet (SGD + Adagrad + Adam) with EFI67.04
VGG16 (SGD + Adagrad + Adam) with EFI66.38
GoogLeNet (SGD + Adagrad + Adam) with EFI67.15
AlexNet (Adam) + VGG16 (Adam) + GoogLeNet (SGD) with EFI69.71
Table 14. Age classification results in three databases.
Table 14. Age classification results in three databases.
MethodsUTKFaceFGNETIMDB-WIKI
AlexNet (SGD) [2]69.5963.6360.88
GoogLeNet (SGD) [4]67.0655.5565.12
LeNet (SGD) [1]62.8854.5451.84
AlexNet (SGD) + LeNet (SGD) + GoogLeNet (SGD) with EFI71.7472.7265.15
Table 15. Gender classification results in UTKFace and IMDB-WiKi databases.
Table 15. Gender classification results in UTKFace and IMDB-WiKi databases.
MethodsUTKFaceIMDB-WIKI
AlexNet (Adam) [2]92.8191
GoogLeNet (SGD) [4]92.3291.78
LeNet (Adagrad) [1]90.4585.84
AlexNet (Adam) + LeNet (Adagrad) + GoogLeNet (SGD) with EFI93.7391.9
Table 16. Recognition results of the CIA photo.
Table 16. Recognition results of the CIA photo.
Classes<1213~1920~45>45
Classifier
Ground truth0010
AlexNet [2]0.000.670.300.03
VGG16 [25]0.020.700.220.07
GoogLeNet [4]0.000.210.790.00
Proposed method0.000.580.590.06
Table 17. Recognition results of the Morph photo.
Table 17. Recognition results of the Morph photo.
Classes<2525~3536~45>45
Classifier
Ground truth0001
AlexNet [2]0.050.180.430.35
VGG16 [25]0.140.220.370.27
GoogLeNet [4]0.010.080.260.66
Proposed method0.010.190.400.58
Table 18. Recognition results of the CACD2000 photo.
Table 18. Recognition results of the CACD2000 photo.
Classes<2525~3536~45>45
Classifier
Ground truth0010
AlexNet [2]0.060.110.720.11
VGG16 [25]0.190.430.320.07
GoogLeNet [4]0.210.390.380.02
Proposed method0.060.420.550.03

Share and Cite

MDPI and ACS Style

Lin, C.-J.; Lin, C.-H.; Sun, C.-C.; Wang, S.-H. Evolutionary-Fuzzy-Integral-Based Convolutional Neural Networks for Facial Image Classification. Electronics 2019, 8, 997. https://doi.org/10.3390/electronics8090997

AMA Style

Lin C-J, Lin C-H, Sun C-C, Wang S-H. Evolutionary-Fuzzy-Integral-Based Convolutional Neural Networks for Facial Image Classification. Electronics. 2019; 8(9):997. https://doi.org/10.3390/electronics8090997

Chicago/Turabian Style

Lin, Cheng-Jian, Chun-Hui Lin, Chi-Chia Sun, and Shyh-Hau Wang. 2019. "Evolutionary-Fuzzy-Integral-Based Convolutional Neural Networks for Facial Image Classification" Electronics 8, no. 9: 997. https://doi.org/10.3390/electronics8090997

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop