You are currently viewing a new version of our website. To view the old version click .
Electronics
  • Article
  • Open Access

4 May 2023

Face Aging with Feature-Guide Conditional Generative Adversarial Network

,
,
,
and
1
School of Information Science and Technology, North China University of Technology, Beijing 100144, China
2
Baidu Inc., Beijing 100085, China
3
University of Science and Technology Beijing, Beijing 100083, China
*
Author to whom correspondence should be addressed.
This article belongs to the Section Computer Science & Engineering

Abstract

Face aging is of great importance for the information forensics and security fields, as well as entertainment-related applications. Although significant progress has been made in this field, the authenticity, age specificity, and identity preservation of generated face images still need further discussion. To better address these issues, a Feature-Guide Conditional Generative Adversarial Network (FG-CGAN) is proposed in this paper, which contains extra feature guide module and age classifier module. To preserve the identity of the input facial image during the generating procedure, in the feature guide module, perceptual loss is introduced to minimize the identity difference between the input and output face image of the generator, and L2 loss is introduced to constrain the size of the generated feature map. To make the generated image fall into the target age group, in the age classifier module, an age-estimated loss is constructed, during which L-Softmax loss is combined to make the sample boundaries of different categories more obvious. Abundant experiments are conducted on the widely used face aging dataset CACD and Morph. The results show that target aging face images generated by FG-CGAN have promising validation confidence for identity preservation. Specifically, the validation confidence levels for age groups 20–30, 30–40, and 40–50 are 95.79%, 95.42%, and 90.77% respectively, which verify the effectiveness of our proposed method.

1. Introduction

Face aging, also known as age image generation or age regression problem [1,2,3,4], can be defined as the process of aesthetically rendering a face image while making the processed image visually appealing with natural aging or rejuvenation of the human face. Face aging has a broad range of applications in different fields, such as cross-age facial recognition, lost children searching, and audio–visual entertainment.
An ideal face aging algorithm should possess the following key characteristics: authenticity, identity preservation, and accuracy when generating images within the target age group. Previous research on facial aging has primarily focused on two categories of methods: physical model-based [4,5,6,7] and prototype-based [2,8,9]. Physical model-based methods rely on adding or removing age-related features, such as wrinkles, gray hair, or beards, that align with image generation rules. The prototype-based method refers to first counting the average face of an age group, and then using the differences between different age groups as an aging pattern to synthesize aging faces, which does not preserve identity well. However, these methods often lack a deep understanding of facial semantics, resulting in generated images that may not be authentic.
Along with the rapid development of deep neural networks, generative adversarial networks (GANs) have drawn much attention from researchers in the face aging field [10,11,12,13,14,15,16] and have been proven that they can generate images with better quality, identity consistency, and aging accuracy compared with the traditional methods. Many studies [17,18] have used unpaired face aging data to train models. However, these methods mainly focus on face aging itself and ignore other key conditional information of the input face (e.g., facial attributes). As a result, incorrect face attributes may appear in the generated results, and the identity information of the generated face cannot be preserved well. In order to suppress such undesired changes in semantic information during face aging, many recent face aging studies [19,20] have attempted to supervise the output by enforcing identity consistency, which to some extent preserves identity information. However, significant unnatural variation in facial attributes is still observed. This indicates enforcing identity consistency alone is not sufficient to achieve satisfactory face aging performance.
To combat this, Variational Autoencoders (VAEs) [21] are combined with GANs to generate new images. Age estimation is used to help implement the generation of aging face images [15]. The subsequent Pyramid Face Aging-GAN [22] combines a pyramid weight-sharing scheme is combined to ensure that face aging changes slightly between adjacent age groups and dramatically between distant age groups. Most existing GAN-based methods [18,23] usually use pixel-level loss to train the model to preserve identity consistency and background information. However, to minimize the Euclidean distance between the synthesized image and the input image, the aging accuracy of its generated results is not very high, which indicates that preserving good identity information does not mean a reasonable aging result.
To more effectively preserve the identity information of faces in face aging tasks, a new face aging framework called Feature-Guide Conditional Generative Adversarial Network (FG-CGAN) is proposed in this article. Compared with existing methods in the literature, to ensure that the generative model preserves the identity information of face images well, feature-guide methods are introduced. Specifically, it is to constrain the network features generated in the process of image generation, generating the image and the identity information features of the original image and the generated image. At the same time, to make the generated image should fall into the target age group, an age classifier module is attached to the discriminator network. Finally, abundant comparison experiments are conducted. To summarize, the main contributions are as follows:
  • To address identity preservation as well as aging accuracy, a network structure based on a gradient feature guide named FG-CGAN is proposed. Extra sub-modules, including a feature guide module in the generator part as well as an age classifier module combined with the discriminator, are attached to tackle identity reservation as well as aging accuracy.
  • To minimize the distance between the input and output face image identities in the feature guide module, the perceptual loss combined is introduced. During this, L2 loss is introduced to constrain the size of the feature map generated in the generator module.
  • In the age classifier module, to improve classification accuracy, the age-estimated loss is constructed, in which L-Softmax loss is combined to learn the intra-class compactness and inter-class separability between features.

3. Methodology

As illustrated in Figure 1, the proposed framework contains four components, including the generator module, the feature guide module, the age classifier module, and the discriminator module. In order to ensure that the comparisons are fair, age categories are divided into five groups: 10–20, 20–30, 30–40, 40–50, and over 50 years old. One-hot labels are used to indicate the age groups. One-hot is filled with 1 in one dimension to indicate the target age and filled with 0 in other dimensions.
Figure 1. Feature-Guide Conditional Generative Adversarial Network (FG-CGAN). FG-CGAN mainly adds a feature guide module and an age classification module to the basic GAN structure, enabling the original face image to generate new images that preserve identity information and have good visual effects.
The purpose of FG-CGAN is to generate an aging face that conforms to the target age group from the original face image. The generator module is composed of an encoder and a decoder. In the process of generating face, the feature guide module extracts the corresponding feature maps of the encoder and decoder, respectively, and makes them conform to L2 constraints. At the same time, L2 is also used to constrain the identity features extracted by the pre-trained network. The discriminator is used to distinguish whether the generated face is true or false. The age classifier module determines whether the age of generated face is within the target age group. The details of this framework are explained below.

3.1. Base Network

The fundamental architecture of the network comprises a generator and a discriminator that utilize GAN principles. The generator has been modified from the Variational Autoencoder (VAE) [21] architecture to enhance its performance.
Generator module: given an input face image x R h × w × N and several target age groups C g R N , in which, h and w represent the height and width of a feature map, N represents the number of age groups, respectively. To generate a synthetic face image x t within the target age group C t , a generator G is built, which refers to the work of the VAE. The synthetic face image can be shown in Equation (1):
x t = G x ,   C t ,
G consists of both an encoder and a decoder.
The encoder module aims to encode the high-dimensional input x into a low-dimensional latent vector, thereby forcing the neural network to learn the most informative features. The encoder module in this paper is constructed with a full convolution neural network. In order to preserve the semantic information of the image, the full connection layer of the full convolution network is replaced with the convolution layer in our proposed network. The input pipes’ number of the input layer is adjusted to be consistent with the image matrix dimension after adding one-hot coding. Using this encoder structure allows the generator to obtain the expression learning ability in the hidden space so that the data can be manipulated at the semantic level through interpolation or conditional embedding on the hidden variable space [21].
The decoder module aims at restoring the latent vector of the hidden layer to the initial dimension and making the output x l x . The fractional step convolution structure is introduced to construct the decoder network.
Discriminator module: this module is used to determine the authenticity of the input synthetic image x t , which is functionally consistent with the discriminator model in the original generative adversarial network, that is, max D x t ,   y . D is a discriminator to identify the facticity of the synthetic face and y is the real target image belonging to the target age group C t .
Adversarial loss: for the problem that the original GANs cannot generate images with specific attributes. The core of CGANs [26] is to integrate attribute information into generator G and discriminator D, where attribute information can be any label information. The objective function of CGANs can be expressed as:
m i n G m a x D V D , G = E x ~ p x x l o g D x | C t + E y ~ p y y log 1 D G y | C t ,
The distributions of x and y are represented by p x x and p y y , respectively, where p x x and p y y are denoting the corresponding probability distributions.
However, CGANs share the same drawbacks as the original GANs as they employ cross entropy as the loss function, which results in generated samples being distant from the decision boundary and only achieving a small loss. This instability in the training process leads to a low-quality output from the generator. In contrast, LSGAN [36] aims to minimize the distance between the generated and real faces, thereby making it difficult for the discriminator to distinguish between them. FG-CGAN adopts the conditional LSGAN function as its adversarial loss:
L D = 1 2 E x ~ p x x D ( G ( x | C t ) 1 2 + 1 2 E y ~ p y y D ( G ( y | C t ) ) 2 ,
L G = 1 2 E y ~ p y y D ( G ( y | C t ) 1 ) 2

3.2. Feature Guide Module

Preserving the identity information of the input face is a critical requirement in the face aging process. However, only utilizing the adversarial loss to make the generated sample and target sample distributions similar may not adequately preserve identity information. To address this issue, we introduce the identity information preservation function to supervise the image generation process by using the features extracted from the network. Specifically, this function is called the feature guide module.
This module obtains the feature maps of the encoder and decoder separately and compares the corresponding feature maps. This requires that the encoder and decoder be symmetrical when designing the network. The advantage of this is that the extracted corresponding feature maps are of the same size, and they can be fed into the same pre-trained network, ensuring that the final output is a one-dimensional vector for metric comparison based on L2 normal form. Thus, to make the corresponding feature maps the same size, L l a y e r is defined in the following Equation (5):
L l a y e r = i k f e n c o d e r i f d e c o d e r i 2 ,
Here k represents the total number of network layers, and f e n c o d e r i and f d e c o d e r i represent the feature maps of the encoder and the decoder at the i - t h layer, respectively.
For identity consistency, the perceptual loss is introduced to minimize the distance between the input and output face image identities of the generator, as shown in Equation (6):
L i d = x p x x h i d x t h i d x 2 ,
where h i d x · represents features extracted from a specific feature layer in the pre-trained model with x as input. The difference metric of the paired feature maps can preserve the identity information between the original and generated images. The pre-trained network uses ResNet-34 as the basic network structure to classify the age of the generated images. The reason why the formulas use L2 instead of L1 as the metric is that L1 is pixel-based and strongly supervises each pixel, leading to a conservative approach to generating the original image. Consequently, the generative network may lack diversity in the generated images. The overall loss function for identity preservation is:
l i d e n t i t y = l i d + l l a y e r

3.3. Age Classifier Module

The age classifier module is used to classify the age group C t of the generated face image x t . This module is primarily structured as Resnet-34.
In addition to meeting the requirements of visual perception, the synthetic face image must also satisfy the target age group condition. To achieve this, the generator utilizes the age classifier module to regulate the age distribution of the synthetic image through loss estimation. This enables the generator to generate a synthetic image x t that conforms to the target age condition by comparing it with the target image y , the loss function of the age classifier module is:
l a g e = 1 M s = 1 M j = 1 N s i g n × l o g P ,
where M represents the number of samples. s i g n represents the sign function, if the age group of the sample is equal to the true group C t , take 1; otherwise, take 0.
The L-Softmax loss is introduced to learn the intra-class compactness and inter-class separability between features. This loss function enhances the distinguishability of sample boundaries for different categories by adjusting the inter-class angle boundary constraints. By multiplying the preset constant with the angle between the sample and the ground truth class, an angular margin is created. The strength of the margin around the ground truth category is determined by the preset constant, allowing the L-Softmax loss to be customized according to the task requirements.

3.4. Overall Objective Function

The final integrated objective function can be obtained by combining the aforementioned equations:
l f i n a l G = λ G l G + λ i d e n t i t y l i d e n t i t y + λ a g e l a g e ,
l f i n a l D = l D ,
where λ G , λ i d e n t i t y , and λ a g e are hyper-parameters used to balance the weight of objective function.

4. Experiments and Evaluation

4.1. Dataset

To ensure fair comparisons, the CACD dataset [37] is introduced to evaluate face generation based on identity preservation. This dataset comprises over 160,000 face images with variations in pose, illumination, and expression obtained from 2000 celebrities aged between 16–62. All the images are age-annotated, although not very accurately. We first use target detection to calibrate the face position and then perform various data enhancement processes on the input images, including adjusting saturation and brightness, horizontally flipping the image, randomly rotating the angle, and normalizing the image. The final images used comprise approximately 146,794 images with a resolution of 400 × 400 pixels, which is split into two parts for training and validation, with 90% and 10% of the images, respectively. The face images are divided into five age groups: 10–20, 20–30, 30–40, 40–50, and over 50 years old, with the number of samples in each age group being 8656, 36,662, 38,736, 35,768, and 26,972, respectively. To further validate the effectiveness of our method, we utilize the Morph dataset [38] for testing. Table 1 shows the comparison information of both data sets.
Table 1. Specific data of CACD-ours and Morph.

4.2. Experimental Details

In this section, the FG-CGAN method is compared with IPCGANs [15], acGANs [39], and CAAE [19], which can generate real face images of specific age groups with identity conditions as constraints.
The image size of the input generation network is 128 × 128 × 3. For the training parameters of the network, refer to the strategy of Wang et al. [25]. The age recognition network proposed in Section 3.4 is used as a feature extractor to extract the feature map between symmetric network layers. During the training phase, the batch size is set to 32 and the learning rate to 0.001. For the selection of optimizer, Adam is used as the optimizer, taking into account the characteristics that it is difficult to train and optimize the generated countermeasure network itself. In the end, the training iteration of the whole network setup is 50,000 rounds. The hyperparameter settings are consistent with Fang et al.’s work [40].

4.3. Classification Accuracy Influence for Age Classifier Module

To supervise the training process of the generative adversarial network, we utilize the age classifier module with varying classification accuracy to identify the optimal classifier network for the current generative adversarial network. During the training process of the age classifier module, we preserve the models with training accuracy of 62, 72, and 82, respectively. These three models are then incorporated into the model training process. The experimental results for different accuracies are presented in Figure 2.
Figure 2. The results from different age classification accuracies: (a) the effect of a classification accuracy of 82; (b) the effect of a classification accuracy of 62.
Figure 2a depicts the final training output when using the age classifier module with an accuracy of 82. The generated image in Figure 2a displays a trend where the generative adversarial network is less likely to alter the face image due to the larger penalty imposed by the age classifier module. A higher classification accuracy can reduce diversity in age groups, causing the generated images to remain similar or unchanged from the original images. Conversely, using a model with low classification accuracy, as seen in Figure 2b, results in obvious artifacts and unrealistic features in the generated images. A lower classification accuracy leads to less punishment during training, causing the generative adversarial network to generate more images with a chance-based approach. Based on these results, a model with a classification accuracy of 72 is selected as the age classifier module. The weights of the age classifier module are set according to the experimental findings in IPCGANs.

4.4. Intuitive Visual Display

Visual effect: Figure 3 shows the output of images randomly selected from Morph dataset at the age of 11–20 and 21–30 as inputs. It can be directly seen that the algorithm proposed in this paper has a good effect on the preservation of face identity information and the change of age diversity. Among them, these include changes in hair color, an increase in facial wrinkles, and an enlarged jaw with age. These changes are in line with people’s intuitive understanding of changes when the face ages.
Figure 3. The generated aged faces by FG-CGAN. The images demonstrate the direct visual effect of FG-CGAN on eight randomly selected input images from the CACD and Morph datasets. As shown, all of these images exhibit a high-quality visual effect of facial aging, showcasing the effectiveness of the model.
Intuitive comparison: based on previous work, we compare the effects of different methods of face generation. We randomly extract four face images from the CACD dataset. Different models are then used to generate images corresponding to age groups to visually compare the realism and information preservation of the images. Since IPCGANs do not provide an official implementation, we re-implement the IPCGANs network and use the acGANs algorithm model available on the network. We evaluate the effects of different algorithms and explain the problems and advantages of algorithms in generating images. Figure 4a–d shows the comparison results:
Figure 4. Performance comparison with prior works on CACD dataset: (ad). Using the same input face images with acGANs, IPCGANs, and FG-CGANs, the synthetic aging images for the age groups of 20–30, 30–40, and 40–50 are displayed.
Figure 4a–d provides a comprehensive comparison of different face-generation methods. The first column of the figure represents the selected original face image, and the last three columns are arranged by age order of 20–30, 30–40, and 40–50. In Figure 4a, the age change of the face generated by the acGANs algorithm is not very evident, and a certain level of smoothing effect is observed. In contrast, the age change of the images generated by IPCGANs and the method proposed in this paper is more apparent, and the identity information of the face is well preserved.
In Figure 4c, acGANs still exhibit a noticeable smoothing effect, and the images suffer from blurred backgrounds. While IPCGANs have a considerable impact on age change and identity preservation, noise appears, resulting in unclear images. The approach proposed in this paper preserves the identity characteristics of the face relatively well and generates images that meet the target age range. Moreover, the images are clearer than those generated by IPCGANs.
In Figure 4b, acGANs result in a certain degree of face identity information loss, and the significant smoothing effect even makes it difficult to recognize the gender of the person in the image. Both IPCGAN and the method proposed in this paper perform well.
In Figure 4d, acGANs still suffer from problems related to face identity information loss, and the age change is not very evident. The IPCGANs differ from the method proposed in this paper for the direction of image aging. IPCGANs prefer to add beards and other elements to show the age difference, while the algorithm in this paper highlights the change characteristics of age from the wrinkles and changes in hair color that occur after face aging.
After conducting the above comparison analysis, it is evident that the method proposed in this paper produces more realistic and reasonable age changes than IPCGANs. The algorithm highlights the features that change with age, such as wrinkles, more frequently. However, both approaches perform well in preserving the identity information of the face. In contrast, the images generated by acGANs exhibit severe dermabrasion and smoothing effects, causing the loss of identity information in the generated images. It is difficult to identify whether the images before and after the generation correspond to the same person or even estimate the gender of the original face. In summary, the approach proposed in this paper and IPCGANs generate images that appear more reliable and realistic as compared to acGANs.
Intuitively looking at the images generated by the three models, the images generated by the work in this paper can not only preserve the identity information features of the face intact but also make the change of age more obvious and intuitive.
Effectiveness of our algorithm: the generation effect verification presented above primarily focuses on gradually generating aging images from young images, but this approach does not fully demonstrate the algorithm’s effectiveness. To address this, we also verify the spread from middle age to both sides. Specifically, we randomly select the faces of individuals aged 30–45 years and input them into our network to generate images of both young and old age. The experiment is illustrated in Figure 5a–d, and we evaluate the generation effect of each image.
Figure 5. Display of face aging results on middle-aged faces by FG-CGAN: (ad). Inputting randomly selected face images of individuals aged 30–45 into the network, the resulting images of the same face at a younger and older age are shown. This comparison illustrates the ability of the model to transform a face’s appearance across a range of ages.
Figure 5a depicts the original faces of individuals aged 41 and 42 as input, with the generated facial appearance displaying a gradual change in wrinkles and graying of hair.
Similarly, Figure 5b shows the increased wrinkles, graying of hair, and enlargement of the jaw, while preserving the identity information features of the face. The generated images meet the expectations of the experiment. Figure 5c also demonstrates a similar effect to the two groups in Figure 5a,b.
However, in Figure 5d, a small number of whiskers appear on the female face in the first row of images. This can be attributed to the principle of generative adversarial networks, which only fit the distribution of data and may struggle to ensure the strict distinction between the male and female sexes throughout the training process.

4.5. Quality Comparison

Face aging aims to convert the face of the input image to the target age while preserving personal identity. Therefore, the face aging model can be evaluated from two perspectives: (1) how well the identity of the original image is preserved; (2) what is the quality of the age classification of the generated aging images.
Identity preservation: we randomly select 500 samples and calculate the face verification confidence by comparing the input images with the images generated for each age group. This approach can comprehensively evaluate the performance of each generative model in terms of preserving identity and achieving accurate age classification. CACE, acGAN, and IPCGAN are measured in comparison.
To evaluate identity preservation in our face aging experiments, we conducted face verification experiments using FaceNet. For each input image of a young face, not only are the original input and the generated faces compared but also the generated faces themselves. In addition, we verified the face verification rate for each age group using different methods on the dataset.
Table 2 reports verification confidences between synthetic aging images from different age groups of the same face, where high verification confidence indicates consistent preservation of identity information. Notably, as the face ages, authentication confidence decreases, indicating that face aging changes appearance. Table 3 also reports the results of the face verification rate, where the threshold is set to 1.0 in FaceNet. Although IPCGAN has an identity retention module, our proposed feature preservation module can better improve identity preservation.
Table 2. Face verification confidence quantitative results.
Table 3. Face verification rate quantitative results.
Accurate age classification: at this stage, volunteers are asked to make an age prediction of the images. One hundred images from the Morph and CACD datasets are randomly selected as input to the generative model. Then, different generative models are used to generate images of the target age groups, resulting in each model containing the original 100 images and 300 generated images. Based on 100 original images, we randomly select 20 images and randomly select 20 images from the corresponding 300 generated images to distribute to volunteers for discrimination in two directions.
For face verification, we generate three images for each input image based on the age label. These images are then grouped into three pairs: (input image, age label 0 image), (age label 1 image, age label 2 image), and (age label 3 image, image randomly selected from other generated face images). The first two pairs are compared to verify if the generated images belong to the same person, while the third pair is used to verify if the generated images are similar to other faces. Volunteers complete the face verification task, and we compare the results using different methods. To calculate the accuracy, using the acc formula,
acc = k p + k n N p + N n
N p and N n represent the total number of sample pairs in the first and second groups, and the number of sample pairs in the third group. k p and k n represent the number of people considered to be the same in the first and second groups and the number of people considered not to be the same in the third group.
For age classification, volunteers need to estimate the age of the image they receive, that is, vote for the age group to which the image belongs. After scoring multiple votes using different models, record the percentage of face images that report a target age that is consistent with the user’s estimated result.
On the other hand, to ensure objectivity in the comparison of effects. We use the face-comparing service provided by Face++ to evaluate how well the generated image matches the face information of the original image. Similarly, we refer to previous work using the VGG-face score to measure image quality. Table 4 shows the result of these experiments.
Table 4. Generate image quality assessment results.

5. Conclusions

Face-based human identification is still a complex problem when it encounters images of the person of different age groups. Cross-age applications such as finding the track missing children after many years, surveillance, and so on are challenging mainly due to the lack of labeled cross-age faces dataset. Hence generating visually realistic faces with GAN-based methods seems to be a promising way. However, identity preservation, as well as aging accuracy, are the essential characteristics of most cross-age applications. In this paper, we tackle these issues with a proposed Feature-Guide Conditional Generative Adversarial Network (FG-CGAN), which is composed of four sub-modules, including the generator module, the feature guide module, the age classifier module, and the discriminator module. In the process of face generation, perceptual loss combined with L2 constraint is introduced to minimize the distance between the input and output face image identities in the feature guide module. In the age classifier module, to improve classification accuracy, the age-estimated loss is constructed, in which L-Softmax loss is combined to learn the intra-class compactness and inter-class separability between features. Sufficient experiments are conducted on the widely used face aging dataset CACD and Morph. Encouraging results are obtained, which verify the effectiveness of our proposed method.
In subsequent work, we will address some of the limitations of this method. For example, this method does not consider facial differences between different races. In addition, although the paper introduced L-Softmax losses in experiments to improve classification accuracy, further validation is needed to evaluate the applicability of this method to other data sets and scenarios. Therefore, the universality and applicability of this method need to be further evaluated, and we hope that future research can more effectively address these issues.

Author Contributions

Conceptualization, X.L. and G.Y.; methodology, C.L., Y.L., Z.W. and X.L.; software, Y.L. and Z.W.; investigation, G.Y.; writing—original draft preparation, C.L., Y.L. and Z.W.; writing—review and editing, C.L., Y.L. and X.L.; supervision, X.L. and G.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This paper is supported by the Research Project of the Beijing Young Topnotch Talents Cultivation Program (Grand No. CIT&TCD201904009), partially by the National Natural Science Foundation of China (Grand No. 62172006 and 61977001) and the Great Wall Scholar Program (CIT&TCD20190305).

Data Availability Statement

Data will be made available on request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Tazoe, Y.; Gohara, H.; Maejima, A.; Morishima, S. Facial aging simulator considering geometry and patch-tiled texture. In ACM SIGGRAPH 2012 Posters; Association for Computing Machinery: New York, NY, USA, 2012; p. 1. [Google Scholar]
  2. Kemelmacher-Shlizerman, I.; Suwajanakorn, S.; Seitz, S.M. Illumination-aware age progression. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 3334–3341. [Google Scholar]
  3. Lanitis, A.; Taylor, C.J.; Cootes, T.F. Toward automatic simulation of aging effects on face images. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 442–455. [Google Scholar] [CrossRef]
  4. Suo, J.; Zhu, S.C.; Shan, S.; Chen, X. A compositional and dynamic model for face aging. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 32, 385–401. [Google Scholar]
  5. Suo, J.; Chen, X.; Shan, S.; Gao, W.; Dai, Q. A concatenational graph evolution aging model. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 2083–2096. [Google Scholar] [PubMed]
  6. Ramanathan, N.; Chellappa, R. Modeling shape and textural variations in aging faces. In Proceedings of the 2008 8th IEEE International Conference on Automatic Face & Gesture Recognition, Amsterdam, The Netherlands, 17–19 September 2008; IEEE: Piscataway, NJ, USA, 2008; pp. 1–8. [Google Scholar]
  7. Ramanathan, N.; Chellappa, R. Modeling age progression in young faces. In Proceedings of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’06), New York, NY, USA, 17–22 June 2006; IEEE: Piscataway, NJ, USA, 2006; Volume 1, pp. 387–394. [Google Scholar]
  8. Liu, C.; Yuen, J.; Torralba, A. SIFT Flow: Dense Correspondence across Scenes and Its Applications. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 978–999. [Google Scholar] [CrossRef] [PubMed]
  9. Rowland, D.; Perrett, D. Manipulating facial appearance through shape and color. IEEE Comput. Graph. Appl. 1995, 15, 70–76. [Google Scholar] [CrossRef]
  10. Zhu, H.; Huang, Z.; Shan, H.; Zhang, J. Look globally, age locally: Face aging with an attention mechanism. In Proceedings of the ICASSP 2020—2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1963–1967. [Google Scholar]
  11. Peng, F.; Yin, L.P.; Zhang, L.B.; Long, M. CGR-GAN: CG facial image regeneration for Antiforensics based on generative adversarial network. IEEE Trans. Multimed. 2019, 22, 2511–2525. [Google Scholar] [CrossRef]
  12. Yi, Z.; Zhang, H.; Tan, P.; Gong, M. Dualgan: Unsupervised dual learning for image-to-image translation. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2849–2857. [Google Scholar]
  13. Kim, T.; Cha, M.; Kim, H.; Lee, J.K.; Kim, J. Learning to discover cross-domain relations with generative adversarial networks. In Proceedings of the International Conference on Machine Learning. PMLR, Sydney, Australia, 6–11 August 2017; pp. 1857–1865. [Google Scholar]
  14. Zhu, J.Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2223–2232. [Google Scholar]
  15. Wang, Z.; Tang, X.; Luo, W.; Gao, S. Face aging with identity-preserved conditional generative adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7939–7947. [Google Scholar]
  16. Yang, H.; Huang, D.; Wang, Y.; Jain, A.K. Learning Continuous Face Age Progression: A Pyramid of GANs. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 499–515. [Google Scholar] [CrossRef] [PubMed]
  17. Li, P.; Hu, Y.; Li, Q.; He, R.; Sun, Z. Global and local consistent age generative adversarial networks. In Proceedings of the 2018 24th International Conference on Pattern Recognition (ICPR), Beijing, China, 20–24 August 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1073–1078. [Google Scholar]
  18. Yang, H.; Huang, D.; Wang, Y.; Jain, A.K. Learning face age progression: A pyramid architecture of gans. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 31–39. [Google Scholar]
  19. Zhang, Z.; Song, Y.; Qi, H. Age progression/regression by conditional adversarial autoencoder. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5810–5818. [Google Scholar]
  20. Antipov, G.; Baccouche, M.; Dugelay, J.L. Face aging with conditional generative adversarial networks. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 2089–2093. [Google Scholar]
  21. Kingma, D.P.; Welling, M. Auto-Encoding Variational Bayes. In Proceedings of the International Conference on Learning Representations, Scottsdale, AZ, USA, 2–4 May 2013. [Google Scholar]
  22. Pantraki, E.; Kotropoulos, C. Face Aging Using Global and Pyramid Generative Adversarial Networks. Mach. Vision Appl. 2021, 32, 82. [Google Scholar] [CrossRef]
  23. Isola, P.; Zhu, J.Y.; Zhou, T.; Efros, A.A. Image-to-Image Translation with Conditional Adversarial Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  24. Duong, C.N.; Luu, K.; Quach, K.G.; Bui, T.D. Longitudinal Face Modeling via Temporal Deep Restricted Boltzmann Machines. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  25. Wang, W.; Cui, Z.; Yan, Y.; Feng, J.; Yan, S.; Shu, X.; Sebe, N. Recurrent face aging. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2378–2386. [Google Scholar]
  26. Mirza, M.; Osindero, S. Conditional Generative Adversarial Nets. arXiv 2014, arXiv:1411.1784. [Google Scholar]
  27. Sharma, N.; Sharma, R.; Jindal, N. Prediction of face age progression with generative adversarial networks. Multimed. Tools Appl. 2021, 80, 33911–33935. [Google Scholar] [CrossRef] [PubMed]
  28. Makhmudkhujaev, F.; Hong, S.; Kyu Park, I. Re-Aging GAN: Toward Personalized Face Age Transformation. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 11–17 October 2021; pp. 3888–3897. [Google Scholar]
  29. Shen, Y.; Yang, C.; Tang, X.; Zhou, B. InterFaceGAN: Interpreting the Disentangled Face Representation Learned by GANs. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 2004–2018. [Google Scholar] [CrossRef] [PubMed]
  30. Liu, Y.; Li, Q.; Sun, Z. Attribute-Aware Face Aging With Wavelet-Based Generative Adversarial Networks. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 11869–11878. [Google Scholar]
  31. Liu, Y.; Li, Q.; Sun, Z.; Tan, T. A3GAN: An Attribute-Aware Attentive Generative Adversarial Network for Face Aging. IEEE Trans. Inf. Forensics Secur. 2021, 16, 2776–2790. [Google Scholar] [CrossRef]
  32. Levi, G.; Hassncer, T. Age and gender classification using convolutional neural networks. In Proceedings of the 2015 IEEE 146 Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Boston, MA, USA, 7–12 June 2015; pp. 34–42. [Google Scholar]
  33. Rothe, R.; Timofte, R.; Van Gool, L. Deep Expectation of Real and Apparent Age from a Single Image without Facial Landmarks. Int. J. Comput. Vision 2018, 126, 144–157. [Google Scholar] [CrossRef]
  34. Li, P.; Hu, Y.; Wu, X.; He, R.; Sun, Z. Deep Label Refinement for Age Estimation. Pattern Recogn. 2020, 100, 107178. [Google Scholar] [CrossRef]
  35. Xia, M.; Zhang, X.; Liu, W.; Weng, L.; Xu, Y. Multi-Stage Feature Constraints Learning for Age Estimation. IEEE Trans. Inf. Forensics Secur. 2020, 15, 2417–2428. [Google Scholar] [CrossRef]
  36. Mao, X.; Li, Q.; Xie, H.; Lau, R.Y.; Wang, Z.; Paul Smolley, S. Least squares generative adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2794–2802. [Google Scholar]
  37. Chen, B.C.; Chen, C.S.; Hsu, W.H. Cross-age reference coding for age-invariant face recognition and retrieval. In Proceedings of the European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2014; pp. 768–783. [Google Scholar]
  38. Ricanek, K.; Tesafaye, T. Morph: A longitudinal image database of normal adult age-progression. In Proceedings of the 7th International Conference on Automatic Face and Gesture Recognition (FGR06), Southampton, UK, 10–12 April 2006; IEEE: Piscataway, NJ, USA, 2006; pp. 341–345. [Google Scholar]
  39. Odena, A.; Olah, C.; Shlens, J. Conditional image synthesis with auxiliary classifier gans. In Proceedings of the International conference on machine learning. PMLR, Sydney, Australia, 6–11 August 2017; pp. 2642–2651. [Google Scholar]
  40. Fang, H.; Deng, W.; Zhong, Y.; Hu, J. Triple-GAN: Progressive Face Aging with Triple Translation Loss. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 14–19 June 2020; pp. 3500–3509. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.