You are currently viewing a new version of our website. To view the old version click .
Electronics
  • Article
  • Open Access

24 May 2023

GFAM: A Gender-Preserving Face Aging Model for Age Imbalance Data

and
1
Department of Computer Science and Engineering, Center for Advanced Image and Information Technology, Jeonbuk National University, Jeonju 54896, Republic of Korea
2
Department of Computer Science and Engineering, Cangzhou Normal University, Cangzhou 061000, China
*
Author to whom correspondence should be addressed.
This article belongs to the Section Artificial Intelligence

Abstract

The objective of face aging is to generate facial images that present the effects of aging. The existing one-hot encoding method for aging and/or rejuvenation patterns overlooks the personalized patterns for different genders and races, causing errors such as a male beard appearing on an aged female face. A gender-preserving face aging model is proposed to address these issues, termed GFAM. GFAM employs a generative adversarial network and includes several subnetworks that simulate the aging process between two adjacent age groups to learn specific aging effects. Specifically, the proposed model introduces a gender classifier and gender loss function that uses gender information as a self-guiding mechanism for maintaining gender attributes. To maintain the identity information of synthetic faces, the proposed model also introduces an identity-preserving module. Additionally, age balance loss is used to mitigate the impact of imbalanced age distribution and enhance the accuracy of aging predictions. Moreover, we construct a dataset with balanced age distribution for the task of face age progression, referred to as Age_FR. This dataset is expected to facilitate current research efforts. Ablation studies have been conducted to extensively evaluate the performance improvements achieved by our method. We obtained relative improvements of 3.75% higher than the model without the gender preserving module. The experimental results provide evidence of the effectiveness of the proposed method, both through qualitative and quantitative analyses. Notably, the mean face verification accuracy for the age-progressed groups (0–20, 31–40, 41–50, and 51–60) was found to be 100%, 99.83%, 99.79%, and 99.11%, respectively, highlighting the robustness of our approach across various age ranges.

1. Introduction

The task of face aging involves generating natural-looking aged versions of face images while preserving the subject’s identity and distinctive facial features. Face aging has practical applications in fields such as locating missing children, kinship verification, and identifying fugitives, among others, which help to create a safe and secure society. Despite its practical value, face aging remains challenging owing to the need for adequate labeled age data on the same subject. Face aging has made remarkable progress in recent years, with numerous methods proposed. These methods produce more realistic aging effects and fewer ghosting artifacts compared to traditional solutions. The resultant methods can be broadly categorized into two groups: conditional generative adversarial network (cGAN)-based and generative adversarial network (GAN)-based methods. Although both cGAN-based and GAN-based methods have been used for face aging, there are notable differences between the two approaches. While cGAN-based methods have greater flexibility, as noted in references [1,2,3,4,5], GAN-based methods, as reported in references [6,7,8], tend to yield superior results in this field.
Generating accurate and high-quality aged faces is an inherently challenging task. Although several methods have been proposed for face aging, meeting three critical requirements for face aging simultaneously, namely aging accuracy, identity preservation, and gender preservation, remains a difficult problem. The faces generated using current methods do not satisfy all three requirements simultaneously. Owing to the inherent complexity of face aging, these methods may not guarantee the aging smoothness of the synthesized faces, particularly when the original face images are of young children. In such cases, age progression beyond a few years may not be practically achievable, which can affect the aging accuracy of the generated faces. Additionally, there could be situations where the aged appearance of a face appears younger than expected. Consequently, to address such variations in aging, the algorithm is trained to learn group-level patterns of aging and rejuvenation for each specific age group condition, such as the tendency for individuals to develop beards as they age. However, this approach has two drawbacks. Firstly, the use of one-hot encoding to represent age group-level aging and/or rejuvenation patterns disregards the personalized patterns of aging that are specific to an individual’s identity, particularly with regard to gender and race. Secondly, there is the issue of inaccurate depictions of gender-specific features, such as the appearance of a beard on an aged female face, which is especially true when the original face image is of a young child. For instance, toddlers often lack obvious gender characteristics, which can lead to inaccurate sex prediction by the learning model and, hence, inaccurate simulation.
For this reason, one of the primary goals is to achieve high natural and gender preservation in face aging. Although the progressive face aging with generative adversarial network (PFA-GAN) [7] is a skilled progressive facial aging model, it has limitations in capturing gender information during the aging process as it primarily focuses on the general aging pattern. Consequently, some of the generated samples may not retain the original gender information. Our proposed solution is the gender-preserving face aging model (GFAM), a novel deep learning model that incorporates gender- and identity-preserving components. Extending the PFA-GAN architecture, GFAM is specifically designed to tackle the challenges related to gender and identity preservation. Firstly, the image is input to the encoder to obtain the personality characteristics. Then, the personality traits are associated with aging through the decoder. Finally, the identity-preserving module utilizes a method of decomposition to extract the identity-specific features. These features are then employed to maintain the identification details of the synthesized faces. The addition of a gender classifier and gender loss, which utilizes gender information as a guiding mechanism, helps to simulate aging effects vividly while maintaining gender information to a finer extent. To improve the accuracy of age prediction models, we introduce a novel age balance loss, which is designed to address the issue of imbalanced age distribution in the training dataset.
Moreover, we construct a new dataset named Age_FR, which has a balanced age distribution to facilitate ongoing and future studies on face age progression. Age_FR contains 181,824 face images of 9719 individuals annotated with identity, gender, and age information. The effectiveness of the proposed method has been demonstrated through extensive experiments conducted on the Age_FR dataset.
The paper is organized as follows. Section 2 provides a brief literature survey of related work. Section 3 describes our proposed model. Section 4 presents the experimental methods and results. Finally, in Section 5, we conclude with a discussion and the limitations of our approach.

3. Gender-Preserving Face-Aging Model

3.1. Overview

The main study objective is to robustly transform the face age while maximally retaining gender information in the generated face images. We develop a long-term aging model for natural facial features and gender preservation in the faces by combining several identical short-term aging models, each of which aims to learn the differences between two adjacent age groups. This approach provides finer control over the aging process, allowing us to better identify and address specific challenges at each stage. In the following section, we provide a detailed description of our network architectures, which are critical to the successful implementation of our proposed approach.
Suppose we have a set of face images X , and given an image x X , our objective is to train sub-generators G to generate multiple faces of several fixed age groups corresponding to the identity in x . To achieve this, we utilize the age-invariant feature extraction network (AFEN) [32] to maintain the identity information during the generation process. By incorporating an age balance loss into the model training process, the impact of age imbalance is effectively mitigated. In addition, a gender classifier is introduced to utilize gender information as self-guiding data to preserve the gender attributes of the generated faces.
Considering the dataset’s wide range of ages, the interval is divided into six groups: up to 20, 21–30, 31–40, 41–50, 51–60, and above 61 years. Inspired by the methodology presented in previous work [12], which employed an RNN to effectively capture the transformation patterns across various age groups, we adopt five sub-generators to model the transition patterns between adjacent age groups. By applying a skip connection in the sub-network, the network allows easy control over the aging flow, as shown in Figure 1. Assuming the age group i + 1 depends on the age group i , six age groups can be formulated as shown in Equation (1):
X i + 1 = X i + G i X i
where X i is the input image and G i is the i -th sub-generator.
Figure 1. Architecture exploiting several GANs to model the aging patterns.
Finally, given any young face x s from the source age group, we can generate the corresponding aged face x t from the target age group through the aging network.

3.2. GAN-Based Architecture

The GAN-based framework is composed of multiple sub-generators, and each of them is made up of four main components: generator, discriminator, gender classifier, and identity-preserving model. These sub-generators are responsible for learning the aging patterns between two neighboring age groups. To achieve this, each sub-generator undergoes extensive training using a combination of several loss functions, including adversarial loss, identity loss, age balance loss, and gender loss. The ultimate goal of the framework is to generate realistic and high-quality aged faces while preserving the identity information of the input face. The complete framework and optimization process are illustrated in Figure 2, and the following subsections provide a detailed description of the proposed face aging framework.
Figure 2. The sub-generator framework is devised to facilitate the transformation of a young face into an aged one, which falls into the next age group.
The generator uses convolutional layers, with a residual blocks-based generator structure, to learn the age transformation. The input faces from the source age group s are denoted by X s . The generator output from s to the target age group t is represented as X t . To enforce the generator to produce samples toward the decision boundary, the least-squares GANs [33] utilize the least-squares loss function instead of the traditional GAN’s negative log-likelihood loss function. The adversarial loss is defined as follows:
L a d v = 1 2 E X s [ D X t 1 ] 2
where s denotes the source age group and X s refers to the input image; t denotes the target age group and X t represents the generated image. D is the discriminator, which is used to distinguish between the generated image and a real image. E represents the expectation operation.
The PatchDiscriminator [25] is adopted as the discriminator D in this work. Each branch performs binary classification to determine the validity of the image as real or fake with respect to its corresponding age domain. The synthesized face in the target age group t should not be classified as a fake sample by the discriminator D . However, the discriminator lacks the ability to determine whether the learned aging effect is genuine or not based on gender. Therefore, a gender classifier is added to preserve the gender attributes in the aged face.

3.3. Gender Classifier

The dataset may contain images with gender imbalances within the age groups. To address this issue, a gender classifier is used to preserve the gender attributes by considering the target gender as a condition and utilizing the gender label as self-guiding information. Based on the cross-age celebrity dataset (CACD) [34], 79,143 individuals are initially labeled as female and 79,057 individuals are labeled as male. The model was trained using an 80% and 20% split for training and testing, respectively. Subsequently, the corresponding gender labels are created for the training dataset. The gender classifier is then trained using the gender label of the input face. The ResNet18 framework [35] is adopted as the base network, as illustrated in Figure 2. The gender loss is defined by the softmax loss to obtain an adequately pre-trained gender classifier model. Our gender classifier achieved an average accuracy of 90.42%.

3.4. Identity-Preserving Module

Preserving the identity information of the synthesized faces is crucial. However, due to the adversarial loss, the generator can produce samples that adhere to the target data distribution, resulting in generated samples resembling any person in the targeted age group. Figure 2 illustrates the implementation of the proposed decomposition method [19] to extract age- and identity-specific features. This model decomposes all features extracted from a facial image into two separate components using the spherical coordinate system. The identity-specific features are then employed to maintain the identity information of the generated faces. The identity loss is defined as follows:
L i d e n t i t y = 1 N f X s f X t 2
where N is the number of training samples. X s refers to the input image, and X t represents the generated image. f represents features extracted from a specific layer of a pre-trained neural network.
To further ensure that the generated faces fall into the target age group t , the age-specific features are leveraged, and an age classifier is trained to identify the correct age of the generated image. Most of the training images used are from adults aged 20–60, while some are from children under 10 years old and senior adults over 60 years old. Therefore, the models trained using such imbalanced data may perform poorly on underrepresented groups, such as those aged above 60 years. The balanced mean-squared error (MSE) loss [36] is utilized in the model instead of the traditional MSE loss to account for the label imbalances in a statistical manner, thereby mitigating the influence of imbalanced label distribution on the MSE. The balanced MSE calculates the discrepancy between the predicted and target ages. The batch-based Monte Carlo implementation of the balanced MSE loss is employed, where all labels in a training batch are treated as random samples and require no label preprocessing beforehand. For labels in a training batch, B y = y t 1 , y t 2 , , y t N , the loss L a g e _ b a l a n c e can be rewritten like the softmax function to represent the loss between the estimated age y p r e d and the target age y t :
L a g e _ b a l a n c e = l o g e x p y p r e d y t 2 2 2 σ n o i s e 2 y t B y exp y p r e d y t 2 2 2 σ n o i s e 2
where σ n o i s e is a one-dimensional hyperparameter that can be optionally learned during training.

3.5. Optimization

As illustrated in Figure 2, the complete loss function for achieving face aging goals incorporates four key elements: (1) adversarial loss that strives to generate aged facial images of superior quality that cannot be differentiated from real ones; (2) identity loss that seeks to preserve the same identity; (3) age balance loss that is expected to enhance the accuracy of aging while also serving as a potential solution to the issue of age imbalance within datasets; (4) gender loss that helps to improve aging accuracy.
L = α 1 L a d v + α 2 L i d e n t i t y + α 3 L a g e _ b a l a n c e + α 4 L g e n d e r

4. Experiments

We conducted a new large-scale cross-age face dataset based on four published datasets for photorealistic cross-age face synthesis research. Table 2 lists the breakdown of the four published datasets and the new dataset into the different age categories and showcases individuals from diverse age groups. The dataset comprises 182,004 face images from 9719 subjects annotated with identity, gender, and age labels. Compared to the previous CACD dataset, our dataset has a better balance of data, with large age gaps from 0 to 116 years, particularly in terms of images featuring subjects under 10 years old and over 60 years old. This is essential for forcing the model to generate faces that exhibit desirable rejuvenation and/or aging effects in these age groups. Images are taken in real scenes (in the wild) and include not only celebrities but also ordinary people, increasing the diversity of the data.
Table 2. Statistics on the age range distribution for four available cross-age face datasets (CACD, AdienceFaces, CASIA-WebFace, and UTKFaces) and Age_FR. N/A means ‘Not Applicable’.

4.1. Data Collection and Annotation

The AdienceFaces benchmark dataset [37] is designed for gender and age classification with diverse gender, age, and ethnic backgrounds. The images were sourced from multiple platforms, including the Internet and social media. The dataset comprises approximately 16K images of 2284 subjects captured in the wild and labeled into eight ordinal groups of age ranges: babies (0–2), infants (4–6), children (8–13), teenagers (15–20), young adults (25–32), adults (38–45), middle-aged (48–53), and seniors (60 and above years). The dataset’s annotations for age and gender information for every image make it an excellent resource for identifying facial attributes such as age and gender from face images.
The UTKFace dataset [1] is a comprehensive collection of 23,699 facial images that represent a diverse population of individuals with ages ranging from 0 to 116 years. These images were sourced from publicly available platforms and feature people from various parts of the world. Notably, the dataset exhibits substantial diversity in terms of pose, facial expression, illumination, occlusion, and resolution. Additionally, it is annotated with information on ethnicity, gender, and age, which makes it a valuable resource for applications in computer vision, such as gender classification, age estimation, and face recognition.
The CASIA-WebFace dataset [38] has a collection of 10,575 individuals and 494,414 face images. Each person in the dataset is represented by at least two facial images, all captured at 250 × 250 resolution with varying lighting, poses, and expressions. To ensure that the dataset is appropriate for age-invariant face recognition tasks, we carefully reviewed and filtered out any images with unsuitable lighting, pose, expression, or occlusion by sunglasses.
The CACD dataset [34] contains 163,446 face images of 2000 celebrities which were captured under less controlled conditions. In addition to the significant differences in posture, lighting, and facial expressions (also known as PIE variations), the dataset was compiled through a Google image search, which poses a significant challenge due to the discrepancies between the person’s actual face in each image and the provided labels. The data were manually verified to obtain a precise dataset with correct identity labels, resulting in a final dataset consisting of 158,200 facial images. Since age and gender labels are required for model training, they were manually generated.
Data processing on the image was conducted as follows. (1) Data cleaning and normalization. To filter the images with borders, which can reduce the accuracy of face recognition, the input images were preprocessed and normalized. This involves cropping the borders and retaining only the facial portion. The multi-task cascaded convolutional network was utilized to detect the facial landmarks and areas and align the facial images based on the identified eye coordinates. Figure 3 shows some face images that have been aligned and normalized from the Age_FR dataset. (2) Data annotation. The dataset’s original gender and age annotations were collected and identity annotations were created manually. The Age_FR dataset contains annotations including identity, gender, and age information, which renders it well-suited for employment in tasks related to face aging. (3) Manual inspection. After annotation, the accuracy of all images and their related annotations were confirmed through manual inspection.
Figure 3. Example images from the Age_FR dataset. The images in the top row are original images, while those in the bottom row depict images that have been aligned and normalized.
To create the Age_FR dataset, the following data selection procedure was used to maintain age balance. Firstly, images from all age groups were selected from the CACD dataset. Next, face data from UTKFace and AdieneFaces, for individuals aged between 0–20 and over 60 years old, were included in the dataset. Since CASIA-WebFace does not contain age labels, only individuals over 60 years old were included in the dataset, and their age was marked as 65. The Age_FR dataset finally contained 182,004 face images from four parts: (1) CACD, 158,000 images; (2) UTKFace, 5971 images; (3) AdieneFaces, 4105 images; (4) CASIA-WebFace, 13,928 images. The CACD dataset offers paired face data that can effectively train sub-generators G 1 , G 2 , G 3 , and G 4 . In addition, by utilizing the samples from AdienceFaces, UTKFace, and CASIA-WebFace datasets, we can train sub-generator G 5 to capture the aging patterns beyond 60 years of age.

4.2. Implementation Details

The proposed model is initialized using the He initialization method [39] and implemented based on the PyTorch platform. All models are trained with a maximum of 40 epochs on four Nvidia Titan X Pascal GPUs with 48G memory, and the batch size is 16. The model is then optimized using the Adam optimization method [40], and the learning rate is 1.0 × 10−4. The hyperparameters in Equation (5) are empirically set as: α 1 = 100, α 2 = 5.0 × 10−4, α 3 = 0.4, and α 4 = 0.1.

4.3. Qualitative Comparison

Figure 4 shows some sample results of face aging on images from the CACD and FGNET datasets. Although some faces appear unnatural, the results indicate that the proposed method is capable of preserving both identity and gender while generating faithful, diverse aged faces.
Figure 4. Sample results for the proposed model on images from the CACD and FGNET datasets for face aging.
To illustrate the efficiency of the proposed GFAM, comparisons were performed with some of the latest state-of-the-art methods. As depicted in Figure 5, the proposed method demonstrated superior performance in producing aging effects on faces while maintaining identity information compared to the CAAE [1] method. Moreover, it outperformed the GAN-based WGLA-GAN [16] method in generating realistic aging details such as wrinkles and color distortion. Our approach also excelled over the RNN-based RFA [12] method in conserving the original identity information and preserving aging details. In terms of preserving details such as face shape and original skin tone, our method surpassed the cGAN-based approach of dual AcGAN [5]. Although PFA-GAN [7] performed better than our model in reducing ghosting, it is worth noting that our model preserves gender-specific information for women, such as the absence of beard growth.
Figure 5. Performance comparisons with current methods on the CACD and FGNET datasets for face aging. The top row shows the test faces with the real age, and the middle and bottom rows show the generated faces using currently used models and the GFAM model, respectively, for the target age group (51+).

4.4. Quantitative Comparison

Identity Preservation. A face verification experiment was conducted for every synthesized face to ensure that their identity property was accurately preserved. We conducted paired comparisons between the input image and the generated faces for each individual test face (such as test versus aged face 31–40, test versus aged face 41–50, and test versus aged face 51–60). As shown in Table 3, for the CACD dataset, we performed a total of 34,432 face verification tests (8608 per age group), comparing each test face to the corresponding generated faces for the four age ranges (20–30, 31–40, 41–50, and 51–60); the mean verification accuracy for these four age-progressed groups are 100%, 99.83%, 99.79%, and 99.11%, respectively. We performed face verification on test faces and synthetic faces using a commonly used online face recognition API called Face++ [41] to check that the original identity attributes are preserved during aging. The false accept rate (FAR) and threshold were set to 10−5 and 76.5 in Face++ APIs, respectively. Our method outperformed IPCGAN owing to the identity-preserving module. The face verification results confirm that the proposed method is relatively superior at preserving the original identity information.
Table 3. Quantitative comparisons using the face verification rate (VR) on the CACD dataset. The boldface entries represent the best values.

4.5. Ablation Study

Qualitative study on the effect of the gender-preserving module. Experiments were conducted to explore the contributions of the proposed gender-preserving module. All experiments in this category were performed on images from the CACD dataset. Figure 6 displays multiple visual illustrations of facial images produced by the proposed model. The results show that when the gender classifier is not included, the generated results suffer from severe gender errors, indicating that the proposed method preserves gender information.
Figure 6. Samples of the visual results from the ablation study using the gender preserving module. (a) Performance without gender classifier; (b) Performance with gender classifier.
Quantitative study on the effect of the gender-preserving module. Table 4 displays the gender accuracy of two models: one without the gender-preserving module and the other with the module. It was observed that our model achieved a gender accuracy of 79.69%, which is 3.75% higher than the model without the gender-preserving module. These results demonstrate that the inclusion of the gender-preserving module leads to improved gender accuracy compared to models that lack this module.
Table 4. Gender accuracy with and without the gender-preserving module on the CACD dataset.
The effect of the identity-preserving module. By performing an ablation study on face verification, we assessed the models with and without the identity-preserving module, using the same experimental setup as in Section 4.4. Table 5 presents the results, which indicate that incorporating the identity-preserving module results in an improvement in the performance of face verification.
Table 5. Face verification accuracy with and without the identity-preserving module on FG-NET. The bold represents the best value.
The effect of age classifier. Our study focuses on assessing the accuracy of the GFAM with and without an age classifier. We utilized a facial age estimation tool called Face++ to predict the age group of synthetic faces. The FG-NET dataset was chosen as the input image for the model because it provides a larger age range compared to the CACD dataset. The dataset consists of 1002 images of 82 individuals whose ages range from 0 to 69. The 1002 images are categorized into five age groups: 0–20, 21–31, 31–40, 41–50, and 50+. The images are fed into the GFAM, with or without an age classifier, to produce aged faces, which are grouped into four age categories: 21–30, 31–40, 41–50, and 51–60. Afterward, we compared the estimated age groups with the target age groups and calculated the percentage of faces that matched. Table 6 shows the effect of the age classification term and illustrates that incorporating an age classifier improves age classification accuracy compared to models without one.
Table 6. Aging effects with and without an age classifier term on FG-NET. The bold represents the best value.

5. Conclusions

The proposed GFAM framework addresses the challenges associated with generating accurate and high-quality aged faces through a gender-preserving face aging synthesis approach based on the GAN. Facial aging synthesis involves three critical factors, namely aging accuracy, gender preservation, and identity preservation. We proposed a comprehensive framework to generate accurate and high-quality aged faces while preserving gender attributes and identity information. Our approach leverages multiple subnetworks that simulate the aging process from youth to old age, allowing us to capture the unique effects of aging between adjacent age groups. By effectively using a gender classifier and gender loss, our approach can maintain gender attributes during the aging synthesis process while also ensuring that the generated faces retain their original identity with the use of an identity-preserving module. To further enhance the accuracy of aging predictions, we utilized an age balance loss to address the issue of imbalanced age distribution. Furthermore, we created a dataset called Age_FR, which has a well-balanced distribution of ages, to propel the research on face age progression. We demonstrated the efficiency of the suggested approach through both quantitative and qualitative analyses in our experiments. However, currently, it is difficult for GFAM to generate detailed aging predictions for children aged 0–13 due to limited high-quality facial data for those.
To address this challenge, we plan to collect more extensive and targeted data and design a model specifically tailored to accurately predict the aging of missing children. Although this study focuses on the face aging framework, further investigations into the potential of facial rejuvenation may also be worth exploring in the future.

Author Contributions

Conceptualization, S.L. and H.J.L.; methodology, S.L.; software, S.L.; validation, S.L.; formal analysis, S.L.; writing—original draft preparation, S.L.; writing—review and editing, S.L. and H.J.L.; supervision, H.J.L.; project administration, H.J.L.; funding acquisition, H.J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by a project for Joint Demand Technology R&D of Regional SMEs funded by the Korea Ministry of SMEs and Startups in 2023 (No. RS-2023-000207672).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhang, Z.; Song, Y.; Qi, H. Age progression/regression by conditional adversarial autoencoder. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5810–5818. [Google Scholar]
  2. Wang, Z.; Tang, X.; Luo, W.; Gao, S. Face aging with identity-preserved conditional generative adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7939–7947. [Google Scholar]
  3. Despois, J.; Flament, F.; Perrot, M. AgingMapGAN (AMGAN): High-resolution controllable face aging with spatially-aware conditional GANs. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 613–628. [Google Scholar]
  4. Song, J.; Zhang, J.; Gao, L.; Liu, X.; Shen, H.T. Dual Conditional GANs for Face Aging and Rejuvenation. In Proceedings of the 2018 International Joint Conference on Artificial Intelligence (IJCAI), Stockholm, Sweden, 13–19 July 2018; pp. 899–905. [Google Scholar]
  5. Li, Q.; Liu, Y.; Sun, Z. Age progression and regression with spatial attention modules. In Proceedings of the AAAI Conference on Artificial Intelligence 2020, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 11378–11385. [Google Scholar]
  6. Yang, H.; Huang, D.; Wang, Y.; Jain, A.K. Learning face age progression: A pyramid architecture of gans. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 31–39. [Google Scholar]
  7. Huang, Z.; Chen, S.; Zhang, J.; Shan, H. PFA-GAN: Progressive face aging with generative adversarial network. IEEE Trans. Inf. Forensics Secur. 2020, 16, 2031–2045. [Google Scholar] [CrossRef]
  8. Liu, Y.; Li, Q.; Sun, Z.; Tan, T. A 3 GAN: An attribute-aware attentive generative adversarial network for face aging. IEEE Trans. Inf. Forensics Secur. 2021, 16, 2776–2790. [Google Scholar] [CrossRef]
  9. Todd, J.T.; Mark, L.S.; Shaw, R.E.; Pittenger, J.B. The perception of human growth. Sci. Am. 1980, 242, 132–145. [Google Scholar] [CrossRef] [PubMed]
  10. Wu, Y.; Thalmann, N.M.; Thalmann, D. A plastic-visco-elastic model for wrinkles in facial animation and skin aging. In Fundamentals of Computer Graphics: World Scientific, Proceedings of the Second Pacific Conference on Computer Graphics and Applications, Beijing, China, 26–29 August 1994; World Scientific Inc.: River Edge, NJ, USA, 1994; pp. 201–213. [Google Scholar]
  11. Kemelmacher-Shlizerman, I.; Suwajanakorn, S.; Seitz, S.M. Illumination-aware age progression. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 3334–3341. [Google Scholar]
  12. Wang, W.; Yan, Y.; Cui, Z.; Feng, J.; Yan, S.; Sebe, N. Recurrent face aging with hierarchical autoregressive memory. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 41, 654–668. [Google Scholar] [CrossRef] [PubMed]
  13. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial networks. Commun. ACM 2020, 63, 139–144. [Google Scholar] [CrossRef]
  14. Mirza, M.; Osindero, S. Conditional generative adversarial nets. arXiv 2014, arXiv:1411.1784. [Google Scholar]
  15. Li, P.; Hu, Y.; Li, Q.; He, R.; Sun, Z. Global and local consistent age generative adversarial networks. In Proceedings of the 2018 24th International Conference on Pattern Recognition (ICPR), Beijing, China, 20–24 August 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1073–1078. [Google Scholar]
  16. Li, P.; Hu, Y.; He, R.; Sun, Z. Global and local consistent wavelet-domain age synthesis. IEEE Trans. Inf. Forensics Secur. 2019, 14, 2943–2957. [Google Scholar] [CrossRef]
  17. Yao, X.; Puy, G.; Newson, A.; Gousseau, Y.; Hellier, P. High resolution face age editing. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; IEEE: Piscataway, NJ, USA, 2018; pp. 8624–8631. [Google Scholar]
  18. Park, T.; Liu, M.-Y.; Wang, T.-C.; Zhu, J.-Y. Semantic image synthesis with spatially-adaptive normalization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 2337–2346. [Google Scholar]
  19. He, Z.; Kan, M.; Shan, S.; Chen, X. S2gan: Share aging factors across ages and share aging trends among individuals. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9440–9449. [Google Scholar]
  20. Makhmudkhujaev, F.; Hong, S.; Park, I.K. Re-Aging GAN: Toward Personalized Face Age Transformation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 3908–3917. [Google Scholar]
  21. Liu, L.; Wang, S.; Wan, L.; Yu, H. Multimodal face aging framework via learning disentangled representation. J. Vis. Commun. Image Represent. 2022, 83, 103452. [Google Scholar] [CrossRef]
  22. Huang, Z.; Zhang, J.; Shan, H. When Age-Invariant Face Recognition Meets Face Age Synthesis: A Multi-Task Learning Framework and A New Benchmark. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 7917–7932. [Google Scholar] [CrossRef] [PubMed]
  23. Huang, Z.; Chen, S.; Zhang, J.; Shan, H. AgeFlow: Conditional age progression and regression with normalizing flows. arXiv 2021, arXiv:2105.07239. [Google Scholar]
  24. Zhao, Y.; Po, L.; Wang, X.; Yan, Q.; Shen, W.; Zhang, Y.; Liu, W.; Wong, C.-K.; Pang, C.-S.; Ou, W.; et al. ChildPredictor: A Child Face Prediction Framework with Disentangled Learning; IEEE Transactions on Multimedia; IEEE: Piscataway, NJ, USA, 2022. [Google Scholar]
  25. Isola, P.; Zhu, J.-Y.; Zhou, T.; Efros, A.A. Image-to-image translation with conditional adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1125–1134. [Google Scholar]
  26. Zhu, J.-Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2223–2232. [Google Scholar]
  27. Choi, Y.; Uh, Y.; Yoo, J.; Ha, J.-W. Stargan v2: Diverse image synthesis for multiple domains. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 8188–8197. [Google Scholar]
  28. Liu, M.; Ding, Y.; Xia, M.; Liu, X.; Ding, E.; Zuo, W.; Wen, S. Stgan: A unified selective transfer network for arbitrary image attribute editing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3673–3682. [Google Scholar]
  29. Or-El, R.; Sengupta, S.; Fried, O.; Shechtman, E.; Kemelmacher-Shlizerman, I. Lifespan age transformation synthesis. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany; pp. 739–755. [Google Scholar]
  30. Karras, T.; Laine, S.; Aittala, M.; Hellsten, J.; Lehtinen, J.; Aila, T. Analyzing and improving the image quality of stylegan. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 8110–8119. [Google Scholar]
  31. He, S.; Liao, W.; Yang, M.Y.; Song, Y.-Z.; Rosenhahn, B.; Xiang, T. Disentangled lifespan face synthesis. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 3877–3886. [Google Scholar]
  32. Li, S.; Lee, H.J. Effective Attention-Based Feature Decomposition for Cross-Age Face Recognition. Appl. Sci. 2022, 12, 4816. [Google Scholar] [CrossRef]
  33. Mao, X.; Li, Q.; Xie, H.; Lau, R.Y.; Wang, Z.; Smolley, S.P. Least squares generative adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2794–2802. [Google Scholar]
  34. Chen, B.-C.; Chen, C.-S.; Hsu, W.H. Cross-age reference coding for age-invariant face recognition and retrieval. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; Springer: Berlin/Heidelberg, Germany; pp. 768–783. [Google Scholar]
  35. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  36. Ren, J.; Zhang, M.; Yu, C.; Liu, Z. Balanced MSE for Imbalanced Visual Regression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 7926–7935. [Google Scholar]
  37. Eidinger, E.; Enbar, R.; Hassner, T. Age and gender estimation of unfiltered faces. IEEE Trans. Inf. Forensics Secur. 2014, 9, 2170–2179. [Google Scholar] [CrossRef]
  38. Yi, D.; Lei, Z.; Liao, S.; Li, S.Z. Learning face representation from scratch. arXiv 2014, arXiv:1411.7923. [Google Scholar]
  39. He, K.; Zhang, X.; Ren, S.; Sun, J. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1026–1034. [Google Scholar]
  40. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  41. MEGVII Inc. Face++ Research Toolkit. Available online: https://www.faceplusplus.com/faceid-solution/ (accessed on 21 May 2023).
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.