Next Article in Journal
Opportunistic Interference Alignment for Spectrum Sharing between Radar and Communication Systems
Previous Article in Journal
User Oriented Transmit Antenna Selection in Massive Multi-User MIMO SDR Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A High-Performance Face Illumination Processing Method via Multi-Stage Feature Maps

1
National Key Laboratory of Fundamental Science on Synthetic Vision, Sichuan University, Chengdu 610065, China
2
Department of Information Technology, Neijiang Vocational&Technical College, Neijiang 641000, China
3
College of Computer Science, Sichuan University, Chengdu 610065, China
4
School of Aeronautics and Astronautics, Sichuan University, Chengdu 610065, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2020, 20(17), 4869; https://doi.org/10.3390/s20174869
Submission received: 30 July 2020 / Revised: 26 August 2020 / Accepted: 26 August 2020 / Published: 28 August 2020
(This article belongs to the Section Intelligent Sensors)

Abstract

:
In recent years, Generative Adversarial Networks (GANs)-based illumination processing of facial images has made favorable achievements. However, some GANs-based illumination-processing methods only pay attention to the image quality and neglect the recognition accuracy, whereas others only crop partial face area and ignore the challenges to synthesize photographic face, background and hair when the original face image is under extreme illumination (Image under extreme illumination (extreme illumination conditions) means that we cannot see the texture and structure information clearly and most pixel values tend to 0 or 255.). Moreover, the recognition accuracy is low when the faces are under extreme illumination conditions. For these reasons, we present an elaborately designed architecture based on convolutional neural network and GANs for processing the illumination of facial image. We use ResBlock at the down-sampling stage in our encoder and adopt skip connections in our generator. This special design together with our loss can enhance the ability to preserve identity and generate high-quality images. Moreover, we use different convolutional layers of a pre-trained feature network to extract varisized feature maps, and then use these feature maps to compute loss, which is named multi-stage feature maps (MSFM) loss. For the sake of fairly evaluating our method against state-of-the-art models, we use four metrics to estimate the performance of illumination-processing algorithms. A variety of experimental data indicate that our method is superior to the previous models under various illumination challenges in illumination processing. We conduct qualitative and quantitative experiments on two datasets, and the experimental data indicate that our scheme obviously surpasses the state-of-the-art algorithms in image quality and identification accuracy.

1. Introduction

As is known to all, the performance of computer vision tasks will degrade when the image sensor is under poor light conditions. As shown in Figure 1, many reasons, such as the excessive exposure and the lack of exposure of the image sensor, the intensity and direction of the light, could make the lighting conditions complicated. Face appearances can change dramatically due to illumination variations [1]. Therefore, illumination processing of facial image under various illumination conditions is highly desired especially in face recognition, expression recognition and so on, due to its wide application in security, health-care, marketing and so on, more and more people are doing research in this field.
In recent decades, in order to solve the illumination problem, experts around the world have come up with various solutions. Most works concentrate on the illumination processing of gray image [2,3,4,5,6,7]. In contrast, research on illumination processing of color image has developed slowly in this field. The data distribution of color image is more complex than that of the gray image is the primary reason to hinder the development of relevant research. As a pioneering work, Faisal et al. [8] combine Phong’s lighting model and a 3D face model to process illumination of color face. Unfortunately, due to the requirement for 3D point clouds and a large amount of computation, this method has limited practical application.
With the developments of hardware and deep learning, illumination processing is gradually evolved from traditional ways to deep learning-based techniques. Ma et al. [9] first use Generative Adversarial Networks (GANs) to process illumination of facial images. Then Ma et al. [10] deal with face illumination by combining triplet loss and GANs. Han et al. [11] put forward asymmetric joint GANs to process facial illumination. Zhang et al. present IL-GAN [12] model based on variational auto-encoder and GANs for processing face illumination.
However, some GANs-based illumination-processing methods only pay attention to the image quality and neglect the recognition accuracy, whereas others only crop partial face area and ignore the challenges to synthesize photographic face, background and hair when the original face image is under extreme illumination conditions. Moreover, the recognition accuracy is low when the face image under extreme illumination conditions. We can use image-to-image translation technique to accomplish the illumination processing of face images. The poor-lighted face images belong to a domain, whereas the standard illumination face images belong to another domain. For these reasons, and inspired by the success of GANs on image-to-image translation, we consider the illumination-processing problem similar to the way of image translation. Our purpose is not only to synthesize photographic face, background and hair when the original face image is under extreme illumination conditions but also preserve identity effectively. The following items are our primary contributions:
  • First, we present a multi-stage feature maps (MSFM) loss that uses different convolutional layers of pre-trained feature network to extract varisized feature maps, and then use these feature maps to compute loss. MSFM loss and our elaborately designed generator are conducive to generating high-quality images and preserving identity effectively.
  • Secondly, our method can effectively synthesize photographic face, background, hair and preserve identity when the original face image is under extreme illumination conditions.
  • Finally, we conduct qualitative and quantitative experiments on two databases and a variety of experimental data indicate that our method significantly surpasses the state-of-the-art methods in image quality and identification accuracy.
We organize our paper as follows. We describe the related work on illumination processing in Section 2. In Section 3, we introduce the proposed method in detail. In Section 4, we show the qualitative and quantitative experimental data. Finally, we draw conclusions in Section 5.

2. Related Work

At the beginning of this section, the traditional illumination-processing algorithms are briefly introduced first, and the state-of-the-art GANs such as CycleGAN [13], DMIT [14], EDIT [15], and Pix2Pix [16] that are relevant to our method are in the second part. In the end, we present deep-learning-based illumination-processing methods.

2.1. Traditional Illumination-Processing Methods

Over recent decades, numerous works have been put forward for solving the illumination problem. In 1987, Pizer et al. [2] proposed adaptive histogram equalization to enhance image contrast. Afterward, many researchers extend the histogram equalization algorithm. For instance, Shan et al. [17] propose region-based histogram equalization to deal with illumination. Xie et al. [18] put forward block-based histogram equalization for illumination processing. To encode rich information on the edge orientations, Lee et al. [19] present orientated local histogram equalization to compensate illumination.
In 1999, Shashua et al. propose the quotient image method [20] that provides an invariant approach to deal with the illumination. Afterward, many researchers extend the quotient image algorithm. For instance, Shan et al. [17] develop gamma intensity correction by introducing an intensity mapping and quotient image relighting. Wang et al. [21] put forward self-quotient image. Chen et al. [22] produce the TV-based quotient image model for illumination processing. Srisuk et al. [23] propose Gabor quotient image by extending the self-quotient image. An et al. [24] propose a decomposed image under L1 and L2 norm constraint, then obtain illumination invariant large-scale part by region-based histogram equalization and get illumination invariant small-scale part by self-quotient image.
Adini et al. [1] propose logarithmic transformation, directional gray-scale derivation, and Laplacian of Gaussian for illumination processing. Single-scale retinex [25] is put forward by Jobson et al. for processing illumination. W et al. [26] propose Gaussian high pass to process illumination. Local processing technology [4] proposed by Xie et al. can improve the uneven illumination effectively. Chen et al. [27] propose a lighting processing method based on the generic intrinsic illumination subspace. Du et al. [28] present wavelet-based illumination-processing method. Chen et al. [5] propose logarithmic total variation for processing illumination. Chen et al. [3] put forward a new method named logarithmic discrete cosine transformation for illumination compensation and processing. Tan and Triggs [7] process illumination by combining some existing methods such as gamma correction, difference of Gaussian filtering, contrast equalization and masking, which is called TT in the literature [29]. Fan et al. [30] propose a method named homomorphic filtering-based illumination processing. The filter’s key component is a difference of Gaussian.
Wang et al. [31] propose illumination processing based on Weber’s Law. Zhao et al. [32] process illumination by using self-lighting ratio to suppress illumination differences in the frequency domain. A linear representation-based face illumination-processing method is put forward by Li et al. [33]. BimaSenaBayu et al. [34] propose an adaptive contrast ratio based on appearance estimation model and shadow coefficient model. Goel et al. [35] put forward an illumination-processing method based on discrete wavelet transformation and discrete cosine transformation. Vishwakarma [36] proposes a fuzzy filter applied over the low-frequency discrete cosine transformation coefficients method for illumination processing. Zhao et al. [37] use ambient, diffuse, and specular lighting maps to decompose lighting effect and estimate the face albedo. Tu et al. [38] use an energy minimization framework to process illumination. Ahmad et al. [39] use independent component analysis and filtering to process illumination. Zhang et al. [40] use patch-based dictionary learning (DL) to process face illumination. Zheng et al. [41] combine difference of Gaussian filters and difference of bilateral filters for illumination processing. Zhang et al. [42] generate the chromaticity intrinsic image (CII) in a log chromaticity space that is robust to illumination variations by combining Phong’s Model and Lambertian Model. Liu et al. [43]. use fusion-based descattering and color tone correction to enhance the illumination of underwater image.

2.2. Illumination-Processing Methods Based on Deep Learning

The developments of machine learning and deep learning accelerate the rapid progress of GANs. The development of GANs [44] brings extraordinary vitality to the image generation. With the help of the combination of GANs and CNN, DCGAN [45] makes a great leap in the ability of image generation. By specifying the input conditions, conditional GAN [46] can generate the specific target photos. With the development of GANs, image translation has also achieved impressive development. Isola et al. [16] propose Pix2Pix for a wide range of supervised domain translation tasks. Since obtaining a mass of paired data is not easy and impractical for many domain translation tasks, DualGAN [47], DiscoGAN [48], CycleGAN [13], DMIT [14] and EDIT [15] are proposed to learn two cross-domain translation models that obey the cycle consistent rule from unpaired data.
Recently, illumination-processing of image based on deep learning has made favorable achievements. Ma et al. [49] use deep convolutional neural network and HSI color space to enhance low-light image. Ma et al. [9] first use Generative Adversarial Nets to process illumination of facial image. Then Ma et al. [10] process face illumination by combining triplet loss and GANs. Han et al. [11] put forward asymmetric joint GANs to process facial illumination. Their method contains two GANs, one of which is employed to process illumination, the other is to maintain personalized facial structures. Zhang et al. put forward IL-GAN [12] based on variational auto-encoder and GANs for processing face illumination. AJGAN [11] shows that it is difficult to obtain a favorable illumination-processing result by unsupervised illumination-processing methods. Therefore, we still use weakly supervised method to process illumination of color and gray faces in this paper.
Although the above illumination-processing methods can deal with illumination effectively and preserve identity well, there are some problems. For example, Ma et al. [9,10] only pay attention to image quality and neglect recognition accuracy. AJGAN [11] and IL-GAN [12] not only conduct face recognition experiments but also illustrate preferable image quality, but their methods only crop partial face area and ignore the challenges to synthesize photographic face, background and hair when the original face image is under extreme illumination conditions. Moreover, the face recognition methods in their paper are outmoded. The main problem is that the recognition accuracy is low when the faces under extreme illumination conditions. For these reasons, we put a new scheme to process the illumination of color and gray faces. We use advanced ResNet-50 [50] pre-trained on VGGFace2 dataset [51], Light-CNN-9 and Light-CNN-29 [52] pre-trained on CASIA-WebFace and MS-Celeb-1M dataset for face identification and the structural similarity (SSIM) [53] index, the visual information fidelity (VIF) [54] and the feature similarity (FSIM) [55] to evaluate our method comprehensively.

3. Proposed Method

We define X as poor-lighted faces and Y as standard illumination faces. Given sets x i j X , y i Y , i denotes identity and j means light type. We expect H ( G ( x i j ) ) = H ( y i ) , which means that after translating various light into a standard one, the synthesized face images G ( x i j ) and the corresponding standard illumination y i have the same identity i. Identity preservation is very important in various image-to-image translation about face images [56]. In this paper, H means feature extractor such as ResNet-50, Light-CNN-9 and Light-CNN-29. Next, we denote x i j and y i with x, y for short.

3.1. Overall Framework

At the beginning of this section, we first present our illumination-processing method of face image in detail. Figure 2 shows the block diagram of the proposed method, which takes a set of poor-lighted face images as input and outputs a set of well-lighted face images in an end-to-end way. From the Figure 2, we can learn that the core of our approach is made up of a generator, a loss network for extracting varisized feature maps and using these feature maps to compute multi-stage feature maps (MSFM) loss. An encoder and a decoder make up our generator together. In the testing phase, we just use the generator to transform poor-lighted face images into well-lighted images. We use 3 loss items: adversarial loss, multi-stage feature maps (MSFM) loss and L1 distance loss. F is introduced in detail when we narrate multi-stage feature maps (MSFM) loss in Section 3.3.

3.2. Generator and Discriminator Architecture

The generator of our method is inspired by the components of residual network [50] and U-net [57]. Our generator consists of 9 convolutional layers, 6 residual blocks and 3 up-sample layers, each of which is equipped with a ReLU as activation. Details of the generator are illustrated in Figure 3. From the Figure 3, we can know that the input size of the generator is designed to be a 128 × 128 color image. The output resolution of our generator is 128 × 128 pixels in size. The dotted lines in Figure 3 are skip connections that are conducive to feature retention. In the middle 6 convolutional layers, we use dropout to avoid over-fitting and special up-sample blocks in the decoder of our generator for enhancing the synthetic ability of our model. For this special design, which further enhances the ability of feature retention. Because InstanceNorm [58] has the characteristics of preventing instance-specific mean and covariance shift simplifying the learning process, we use InstanceNorm after each convolutional layers. InstanceNorm can be computed by:
y t i j k = x t i j k μ t i σ t i 2 + ε ,
μ t i = 1 H W l = 1 W m = 1 H x t i l m ,
σ t i 2 = 1 H W l = 1 W m = 1 H ( x t i l m μ t i ) 2 ,
where x R T × C × W × H is a tensor including T images. x t i j k mean its t i j k -th element, k and j are spatial dimensions, i denotes color channel, t is the index of the image in the batch.
The discriminator of our method is inspired by the components of Pix2Pix [16]. The detailed structure of our discriminator is shown in Figure 4. The ReLU is used as activation after the left four convolutional layers and we replace BatchNorm with InstanceNorm. The input size of the discriminator is designed to be a 128 × 128 paired color images such as ( x , y ) and ( x , G ( x ) ) . We use InstanceNorm after the left four convolutional layers.

3.3. Objective Function

Three terms make up our objective function together: a multi-stage feature maps loss for making the ground truth and the generative results more similar, an adversarial loss for making the real distribution and the distribution of synthesized images more similar. A L1 distance loss for improving the performance of our method further.
Adversarial Loss: The adversarial process is made up of the generator G and the discriminator D. D attempts to discriminate the generated fake image G(x) and the ground truth image y whereas G strives to generate fake image G(x) to fool discriminator D. The objective function is as follows:
L G A N ( G , D , x , y ) = E y P Y [ l o g D ( y ) ] +   E x P X [ l o g ( 1 D ( G ( x ) ) ) ] ,
where x denotes input image (poor-lighted face), whereas y is target image (standard illumination).
Multi-Stage Feature Maps Loss: Experimental data indicate that early layers of loss network F (VGG-16) pre-trained on the ImageNet dataset [59] tend to produce smooth facial images, whereas the content and the overall spatial structure can be preserved by the higher layers of F, but the color, the texture and the exact shape are not preserved effectively. Therefore, we choose middle layers of F as our loss network for synthesizing high-quality facial images.
As narrated before, we hope the synthesized image G ( x ) and its ground truth y to be similar in illumination and to have same identity features. Meanwhile, the generator is not only to generate a well-lighted face image but also to fool the discriminator as soon as possible. We use different layers of pre-trained feature network for extracting varisized feature maps, and then use these feature maps to compute feature loss, named multi-stage feature maps (MSFM) loss. The multi-stage feature maps (MSFM) loss can be computed by:
L M S F M ( G , x , y ) = i = 1 3 λ i × | | F i ( y ) F i ( G ( x ) ) | | 2 2 = i = 1 3 λ i × | | F i ( y ) F i ( y ^ ) | | 2 2 ,
F is the VGG-16 network that has 13 convolutional layers. F 1 means 1 to 5 convolutional layers of VGG-16 [60], F 2 means 1 to 6 convolutional layers from of VGG-16, F 3 means 1 to 7 convolutional layers of VGG-16. λ i is weight parameter, y ^ means the output result of our generator, whereas y is target image (standard illumination). F 1 ( y ) means that we use the 1 to 5 convolutional layers of VGG-16 to obtain the feature maps of y.
L1 Distance Loss: Although MSFM loss and adversarial loss make our method gain favorable performance in image quality and recognition accuracy, for the sake of improving the performance of our method further, we add L1 distance loss to our loss function. Because L2 result in blurry synthesized image [16], we choose L1 instead of L2. In this paper, L1 is used to compute the sum of absolute difference between the ground truth image y and the generated image y ^ . L1 distance can be computed by:
L L 1 ( G , x , y ) = | | y G ( x ) | | 1 = | | y y ^ | | 1 ,
where y is the ground truth, and y ^ is the synthesized result.
The final objective function is:
L ( G , D , x , y ) = L G A N ( G , D , x , y ) +   L M S F M ( G , x , y ) +   α × L L 1 ( G , x , y ) ,
where α is weight parameter.

4. Experimental Results

In this section, the datasets, the implementation details and the qualitative and quantitative results of our algorithm are illustrated in detail. Our experiments are conducted on two sets, one is color set, the other is gray set. We compare with the state-of-the-art unsupervised and supervised deep learning methods on the color set. On the gray set, we mainly make comparison between classical illumination-processing methods and ours.

4.1. Datasets

In our paper, we first perform the experiments on the MultiPIE [61] database and then make a comparison with some state-of-the-art methods such as CycleGAN [13], DMIT [14], EDIT [15], Pix2Pix [16], and then compare with the classical methods such as LDCT [3], LN [4], LTV [5], SQI [6], TT [7] and famous CycleGAN [13], Pix2Pix [16] and ours on the Extended YaleB datasets. Finally, we randomly choose some face images from FRGC [62] dataset to verify the generalization performance of our algorithm.
MultiPIE database: The MultiPIE [61] database has been extensively used in illumination processing and face recognition. We choose frontal faces, 20 illumination conditions and natural expression, without glasses from session 1 of MultiPIE as our dataset. We detect and crop faces from the dataset with single shot scale-invariant face detector (S3FD) [63] and resize to 128 × 128 as our training and test set. The 07 illumination faces are chosen as standard faces (standard illumination) and the rest are selected as poor-lighted facial images. In the training dataset, 99 individuals are chosen. For making the experiments more challenging, we use gamma correction to make the image become darker. When g a m m a = 2 , we think it is not challenging enough. When g a m m a = 4 , the mean value of some faces is 0.47 and the max pixel value is 9. The zero values are more than three quarters. Due to the lack of effective information, it is difficult to restore meaningful images. So, we set g a m m a = 3 and we use gamma correction to process all the images, which is not only challenging but also being able to restore meaningful images. In actual application, we can set more gamma values such as 1.1, 1.2 … and so on for obtaining more training data and improving the performance of the method. Thus, the training set has 99 × 19 × 2 + 99 = 3861 images. The test set has 30 individuals and 30 × 19 × 2 = 1140 images. We use 30 individuals under 38 lighting conditions (19 original lighting conditions and 19 analog lighting conditions obtained from the former’s images by conducting gamma correction when g a m m a = 3 ) to test CycleGAN, DMIT, EDIT, Pix2Pix and our method.
Extended YaleB: The Extended YaleB that has 38 subjects under 64 illumination conditions is widely used to evaluate different illumination-processing methods. We divide this dataset according to the literature [64,65], 1 to 28 (1792 images) are used to train all the deep learning-based approaches, 29 to 38 (630 images) are used for test. Next, we call it YaleB for short.
FRGC v2 dataset: The Face Recognition Grand Challenge (FRGC) [62] dataset contains 3D images and high resolution controlled and uncontrolled stills under various illumination conditions. We randomly choose some 2D faces to verify the generalization performance of our algorithm. Next, we call it FRGC for short.

4.2. Implementation Details

As to the encoder and decoder, we use ReLU for activation first. For gradient descent, we use Adam [66] optimizer, and choose a learning rate of 0.0002 with momentum parameters β 1 = 0.95, β 2 = 0.999. We set b a t c h s i z e = 8 during training period. 100 epochs have completed within about 200 min on the MultiPIE dataset. Moreover, during the training period, we use random cropping for data enhancement. In this work, we set α = 0.1 . By setting different values for λ 1 , λ 2 and λ 3 , we get 4 combinations of loss items. We set λ 1 = 1.0, λ 2 = 0.0 and λ 3 = 0.0 to train a model using 1 to 5 convolutional layers of VGG-16 for extracting feature maps and computing loss. We choose λ 1 = 0.0, λ 2 = 1.0 and λ 3 = 0.0 to train a model using 1 to 6 convolutional layers of VGG-16 for extracting feature maps and computing loss. We set λ 1 = 0.0, λ 2 = 0.0 and λ 3 = 1.0 to train a model using 1 to 7 convolutional layers of VGG-16 for extracting feature maps and computing loss. For balancing the image quality and identity feature, we set λ 1 = 0.91, λ 2 = 0.08 and λ 3 = 0.01 and train a model using multi-stage feature maps loss.

4.3. Metrics

At present, most literature evaluates the performance of illumination-processing algorithms from two aspects: one is to compare recognition accuracy, the other is to illustrate some face images before and after processing by various methods. We make use of cosine similarity of feature vectors for face recognition. For the purpose of more comprehensively estimating the performance of various illumination-processing algorithms, except for comparing recognition accuracy and illustrating examples, we also adopt the structural similarity (SSIM) [53] index, visual information fidelity (VIF) [54] and the feature similarity (FSIM) [55] to evaluate the performance of various illumination-processing methods. Cosine similarity, VIF, SSIM and FSIM are briefly introduced as follows:
  • The cosine similarity is a metric to determine how similar the two vectors are. It computes the cosine of the angle between two vectors. The smaller the angle is, the higher similarity is. The cosine similarity is computed by:
    cos ( θ ) = i = 1 n A i × B i i = 1 n ( A i ) 2 × i = 1 n ( B i ) 2 ,
    where A and B are the feature vectors which are extracted by the ResNet-50 [50] pre-trained on VGGFace2 [51], the Light-CNN-9 [52] and the Light-CNN-29 [52] pre-trained on CASIA-WebFace and MS-Celeb-1M. A i denotes i-th element of vector A, B i denotes i-th element of vector B.
  • The visual information fidelity (VIF) [54] is a metric to measure the information fidelity between the ground truth image and the generated image. VIF is defined as:
    V I F = j s u b b a n d s I ( C N , j ; F N , j | S N , j ) j s u b b a n d s I ( C N , j ; E N , j | S N , j ) ,
    where I ( C N , j ; F N , j | S N , j ) and I ( C N , j ; E N , j | S N , j ) denote the information that can ideally be extracted by the brain from a particular sub-band in the ground truth image and the generated image respectively.
  • The structural similarity (SSIM) [53] is a widely used metric to measure the level of similarity in structure between the ground truth and the synthesized image. The two images are exactly equal when the SSIM values of two images is equal to 1. SSIM is defined as:
    S S I M ( x , y ) = ( 2 μ x μ y + C 1 ) ( 2 σ x y + C 2 ) ( μ x 2 + μ y 2 + C 1 ) ( σ x 2 + σ y 2 + C 2 ) ,
    where x and y means two images, μ x is the mean value of image x, μ y is the mean value of image y. σ x and σ y are standard variances of image x and y. C 1 and C 2 are two constants.
  • The feature similarity (FSIM) [55] is a widely used metric to measure the feature similarity between the ground truth and the synthesized image. The closer the FSIM value is to 1, the more similar feature the two images have. FSIM is defined as:
    F S I M = x Ω S L ( x ) P C m ( x ) x Ω P C m ( x ) ,
    where Ω is the whole image spatial domain, P C m ( x ) is the phase congruency of image x, and S L ( x ) is the overall similarity of two images.
In summary, the closer the values of SSIM, VIF and FSIM are to 1, the higher quality the image.

4.4. Qualitative Comparisons

In this section, we illustrate some poor-lighted faces processed by CycleGAN [13], DMIT [14], EDIT [15], Pix2Pix [16] and our method. As can be observed from the Figure 5, in general, all these methods can process various illumination conditions and obtain better visual effects, but some methods have drawbacks. As illustrated in Figure 5, the first column is the original poor-lighted faces under various illumination conditions. From the second to the sixth columns are faces processed by CycleGAN, DMIT, EDIT, Pix2Pix and ours. From the second column, we can see that CycleGAN obtains favorable face images, but the first and third faces are not photo-realistic enough, moreover, the background of the first faces has some shadow and the hair of the third faces is translated into background. From the third and the fourth columns of Figure 5, we can see clearly that DMIT and EDIT cannot discriminate background and hair well. Although Pix2Pix obtains preferable faces, its outputs have some blemishes. For instance, the first face of the fifth column is not photo-realistic and natural enough, moreover, from the third and fourth faces, we can know that some hair areas are not bright enough. From the sixth column of Figure 5, we can learn that our method cannot only translate poor-lighted faces under various illumination conditions into well-lighted and photo-realistic faces but also discriminate background and hair well.
Figure 6 illustrates the illumination-processing results of the eight methods on the YaleB database. The synthesized faces of LDCT [3] are blurry. The output results of LN [4] are still dark and have many shadows. The LTV [5] method obtains smooth synthesized face images. The SQI [6] and TT [7] methods get noisy output results and cannot process cast shadows well. From the seventh column of Figure 6, we can see the faces generated by CycleGAN [13] are not natural and some face area is too bright. At the eighth column of Figure 6, we can see that some faces synthesized by Pix2Pix [16] have 3 eyes and some faces are distorted. Our method can not only process illumination effectively but also obtain high-quality synthesized face images with good visual effects. In the next section, we will evaluate all the aforementioned methods with four metrics objectively.

4.5. Quantitative Evaluation

Next, we conduct face identification and full reference image quality assessment (FR-IQA) on the MultiPIE database and the YaleB database separately.

4.5.1. Quantitative Evaluation on the MultiPIE Database

As to FR-IQA, we conduct experiments on the MultiPIE database. When we compute the FR-IQA of the synthesized faces, the faces with standard illumination are chose as reference images. We compute FR-IQA by three metrics: (1) the first is the structural similarity (SSIM) index [53]; (2) the second is the visual information fidelity (VIF) index [54]; (3) the third is the feature similarity (FSIM) [55]. The FR-IQA values of the faces synthesized by CycleGAN, Pix2Pix, EDIT, DMIT and our method are shown in Figure 7, which shows the higher FR-IQA values are, the higher percentage of images on the horizontal axis is. It is obvious that our algorithm obtains higher FR-IQA values than the other four approaches. In Table 1, we illustrate the average FR-IQA of various algorithms, which demonstrates that our method is superior to the others.
As shown in Table 2, we use ResNet-50 [50] pre-trained on VGGFace2 [51], Light-CNN-9 [52] and Light-CNN-29 [52] pre-trained on CASIA-WebFace and MS-Celeb-1M to evaluate the recognition accuracy of original poor-lighted face images and faces processed by CycleGAN [13], DMIT [14], EDIT [15], Pix2Pix [16] and our method. From the second column of Table 2, we can know when we use ResNet-50 to recognize the original poor-lighted faces, the recognition accuracy is only 93.86%. After processing by CycleGAN, the recognition accuracy is 98.51% and improved by 4.65%. After processing by DMIT, the recognition accuracy is 93.77% and decreased by 0.09%. After processing by EDIT, its recognition accuracy is 98.25% and improved by 4.39%. After processing by Pix2Pix, the recognition accuracy is 97.89% and improved by 4.03%. After processing by our method, the recognition accuracy is 99.91% and improved by 6.05%. From the third column of Table 2, we can learn that the identification accuracy is improved by 5.26% when we use Light-CNN 9 to recognize the faces processed by our method, whereas CycleGAN, DMIT, EDIT and Pix2Pix only improved by 3.69%, 3.16%, 4.12% and 3.33% separately. From the fourth column of Table 2, we can know that the identification accuracy is improved by 3.33%, 2.72%, 3.51% and 3.16% separately when we use Light-CNN 29 to recognize the faces processed by CycleGAN, DMIT, EDIT and Pix2Pix, whereas our method improves the recognition accuracy by 3.95% and reaches to 100%. It is obvious that recognition accuracy of original poor-lighted faces processed by these methods listed in the table has been improved more less. When we use ResNet-50 to recognize, except for DMIT, all other algorithms improve the recognition accuracy obviously. Compared to other method listed in Table 2, our approach obtains the maximal recognition accuracy.
For the sake of further evaluating the performance of the above methods, we also draw ROC curves. As shown in Figure 8, the figure illustrates the Verification Rate (VR) and False Acceptance Rate (FAR). In Table 2, we show the results of top-1 and VR@FAR = 0.1%. From the Figure 8 and Table 2, we can see that our method is better than the others in identity preservation.

4.5.2. Quantitative Evaluation on the YaleB Database

We choose the standard illumination faces as reference images when we conduct full reference image quality assessment (FR-IQA) experiments on the YaleB database. We use three full reference image quality assessment (FR-IQA) metrics to evaluate the performance of the classical illumination-processing methods and ours: (1) the first is the structural similarity (SSIM) index [53]; (2) the second is the visual information fidelity (VIF) index [54]; (3) the third is the feature similarity (FSIM) [55]. The FR-IQA values of the faces synthesized by LDCT [3], LN [4], LTV [5], SQI [6], TT [7] and our algorithm are shown in Figure 9, which shows the higher FR-IQA values are, the higher percentage of images on the horizontal axis is. Except for a few of our synthesized images are worse than CycleGAN, Pix2Pix in image quality. It is obvious that our algorithm obtains higher average FR-IQA values than the other seven approaches. In Table 3, we illustrate the average FR-IQA of various algorithms, which demonstrates that our method is superior to the others on average FR-IQA values on the YaleB database.
Table 4 shows the top-1 identification accuracy and verification accuracy at 0.1% FAR on the YaleB database using various feature extractor. In the second column of Table 4, we can know that the top-1 identification accuracy of original poor-lighted face images is only 51.26%. The identification rates of the 5 classical illumination-processing methods and Pix2Pix are lower than the original poor-lighted face images. The foremost reason is that the illumination of the face images from the YaleB dataset is too dire. The dire illumination causes the images generated by LDCT, LTV, TT to loss details and to become smooth. As shown in the red box of Figure 6, the verification accuracy of Pix2Pix decreases significantly due to the poor identity retention and distorted faces caused by the extreme illumination conditions. However, our method obtains 89.52% identification rate. From the fourth column of Table 4, we can see that the identification rate reaches 94.29% after illumination processing by our method then using Light-CNN-29 to identify. However, the identification rates of state-of-the-art algorithms such as IL-GAN [12] and AJGAN [11] are 89.61% and 87.20% separately.
For the sake of further evaluating the performance of the above methods, we also draw ROC curves. As shown in Figure 10, the figure illustrates the Verification Rate (VR) and False Acceptance Rate (FAR). In Table 4, we show the results of top-1 and VR@FAR = 0.1%. From the Figure 10 and Table 4, we can see that our algorithm precedes the others in identity preservation.

4.6. Application to Unseen Dataset

For the sake of verifying the generalization performance of our method further, we perform illumination-processing experiment on the facial images of FRGC dataset. As illustrated in Figure 11, we can see that the faces in the first row are under various illumination conditions, the second row is the synthesized face images of the first row by our algorithm. It is obvious that our algorithm can not only improve the illumination of faces under various poor-lighted conditions but also keep corresponding identities effectively.

4.7. Ablation Studies

Theoretically, different layer of deep convolution network can extract features of different granularity. Inspired by this, we use different layers of VGG-16 for extracting varisized feature maps, and then use these feature maps to compute multi-stage feature maps loss.
As shown in Table 5, we demonstrate the SSIM, VIF and FSIM score of face images processed by our method trained under various loss items. In the first row, conv1-5 means we use 1 to 5 convolutional layers of VGG-16 for extracting feature and computing loss, conv1-6 means we use 1 to 6 convolutional layers of VGG-16 for extracting feature and computing loss, conv1-7 means we use 1 to 7 convolutional layers of VGG-16 for extracting feature and computing loss, combination means to combine conv1-5, conv1-6 and conv1-7 for computing multi-stage feature maps loss. For convenience, we use loss A, B, C, D to indicate the aforementioned 4 losses. Although our method trained under loss A, B and C can obtain high SSIM, VIF and FSIM separately, we increase them by combining loss A, B and C further.
As illustrated in Figure 12, we illustrate some facial images processed by our method trained under various loss items. The first poor-lighted image is the original image. The second one is obtained by our method trained under 1 to 5 convolutional layers of VGG-16 for extracting feature and computing loss. The third one is generated by our method trained under 1 to 6 convolutional layers of VGG-16 for extracting feature and computing loss. The fourth one is obtained by our method trained under 1 to 7 convolutional layers of VGG-16 for extracting feature and computing loss. Although the former 3 schemes can process illumination effectively and preserve identifies well, the combination scheme can improve the image quality further.

5. Conclusions

In this work, we use multi-stage feature maps (MSFM) loss and an elaborately designed architecture based on convolutional neural network and GANs for illumination processing of face images. Our method can synthesize photographic face, background and hair when the original face image is under extreme illumination conditions. Furthermore, our method can not only process illumination of face image under extreme illumination conditions, but also preserve identity information effectively. We discover that face images with favorable quality does not guarantee high recognition accuracy, only the face images that preserve identity features and structure well can obtain high recognition accuracy. Our method will fail when there is no clear boundaries between the clothes and the background, specifically when the background and clothes have the same color, and when the mean value is less than 1 and most pixel values are equal to zero. Although our illumination-processing algorithm achieves preferable results, there is a lot of future work and potential applications here worth continuing to study:
  • To improve our network structure for preserving more texture details.
  • To train a feature extractor and classifier for the facial images after illumination processing by our method.
  • Using our method to process the illumination of other images, such as landscape and medical images.
  • Applying our method to the preprocessing stage of other visual analysis tasks, such as face detection, head pose estimation, facial landmark detection and face alignment, to improve the performance in these tasks.

Author Contributions

Conceptualization, S.L.; methodology, S.L.; resources, S.L. and Y.L.; validation, S.L.; formal analysis, S.L.; writing—original draft preparation, S.L.; writing—review and editing, S.L., Y.L., K.F., D.Y. and P.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Joint Funds of the National Natural Science Foundation of China under Grant U1833128.

Acknowledgments

We would like to thank Zhisheng You for his strong supports on this study.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
MSFMMulti-stage Feature Maps
GANsGenerative Adversarial Networks
SSIMStructural Similarity
VIFVisual Information Fidelity
FSIMFeature Similarity
LADLeast Absolute Deviations
LAELeast Absolute Errors
FRGCFace Recognition Grand Challenge
FR-IQAFull Reference Image Quality Assessment
FARFalse Acceptance Rate
LTLogarithmic Transform
HEHistogram Equalization
SQISelf-quotient Image
LDCTLogarithmic Discrete Cosine Transform
LTVLogarithmic Total Variation
LNLocal Normalization

References

  1. Adini, Y.; Moses, Y.; Ullman, S. Face Recognition: The Problem of Compensating for Changes in Illumination Direction. IEEE Trans. Pattern Anal. Mach. Intell. 1997, 19, 721–732. [Google Scholar] [CrossRef] [Green Version]
  2. Pizer, S.M.; Amburn, E.P.; Austin, J.D.; Cromartie, R.; Geselowitz, A.; Greer, T.; ter Haar Romeny, B.M.; Zimmerman, J.J. Adaptive histogram equalization and its variations. Comput. Vision Graph. Image Process. 1987, 39, 355–368. [Google Scholar] [CrossRef]
  3. Chen, W.; Er, M.J.; Wu, S. Illumination compensation and normalization for robust face recognition using discrete cosine transform in logarithm domain. Syst. Man Cybern. 2006, 36, 458–466. [Google Scholar] [CrossRef] [PubMed]
  4. Xie, X.; Lam, K.M. An efficient illumination normalization method for face recognition. Pattern Recognit. Lett. 2006, 27, 609–617. [Google Scholar] [CrossRef]
  5. Chen, T.; Yin, W.; Zhou, X.S.; Comaniciu, D.; Huang, T.S. Total variation models for variable lighting face recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2006, 28, 1519–1524. [Google Scholar] [CrossRef]
  6. Wang, H.; Li, S.; Wang, Y. Face recognition under varying lighting conditions using self quotient image. In Proceedings of the Sixth IEEE International Conference on Automatic Face and Gesture Recognition, Seoul, Korea, 17–19 May 2004; pp. 819–824. [Google Scholar]
  7. Tan, X.; Triggs, B. Enhanced Local Texture Feature Sets for Face Recognition Under Difficult Lighting Conditions. IEEE Trans. Image Process. 2010, 19, 1635–1650. [Google Scholar]
  8. Al-Osaimi, F.R.; Bennamoun, M.; Mian, A.S. Illumination Normalization for Color Face Images. In International Symposium on Visual Computing; Springer: Berlin/Heidelberg, Germany, 2006. [Google Scholar]
  9. Ma, W.; Xie, X.; Yin, C.; Lai, J.H. Face Image Illumination Processing Based on Generative Adversarial Nets. In Proceedings of the 2018 24th International Conference on Pattern Recognition (ICPR), Beijing, China, 20–24 August 2018; pp. 2558–2563. [Google Scholar]
  10. Ma, W.; Xie, X.; Lai, J.; Zhu, J. Face Image Illumination Processing Based on GAN with Dual Triplet Loss. In Chinese Conference on Pattern Recognition and Computer Vision (PRCV); Springer: Cham, Switzerland, 2018; pp. 150–161. [Google Scholar]
  11. Han, X.; Yang, H.; Xing, G.; Liu, Y. Asymmetric Joint GANs for Normalizing Face Illumination from a Single Image. IEEE Trans. Multimed. 2019, 22, 1619–1633. [Google Scholar] [CrossRef]
  12. Zhang, Y.; Hu, C.; Lu, X. IL-GAN: Illumination-invariant representation learning for single sample face recognition. J. Vis. Commun. Image Represent. 2019, 59, 501–513. [Google Scholar] [CrossRef]
  13. Zhu, J.Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired Image-to-Image Translation using Cycle-Consistent Adversarial Networkss. In Proceedings of the 2017 IEEE International Conference Computer Vision (ICCV), Venice, Italy, 22–29 October 2017. [Google Scholar]
  14. Yu, X.; Chen, Y.; Li, T.H.; Liu, S.; Li, G. Multi-mapping Image-to-Image Translation via Learning Disentanglement. In Proceedings of the NeurIPS, Vancouver, BC, Canada, 8–14 December 2019. [Google Scholar]
  15. Fu, Y.; Ma, J.; Ma, L.; Guo, X. EDIT: Exemplar-Domain Aware Image-to-Image Translation. arXiv 2019, arXiv:1911.10520. [Google Scholar]
  16. Isola, P.; Zhu, J.Y.; Zhou, T.; Efros, A.A. Image-to-Image Translation with Conditional Adversarial Networks. In Proceedings of the 2017 IEEE Conference Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  17. Shan, S.; Gao, W.; Cao, B.; Zhao, D. Illumination normalization for robust face recognition against varying lighting conditions. In Proceedings of the 2003 IEEE International SOI Conference, Newport Beach, CA, USA, 29 September–2 October 2003; pp. 157–164. [Google Scholar]
  18. Xie, X.; Lam, K.M. Face recognition under varying illumination based on a 2D face shape model. Pattern Recognit. 2005, 38, 221–230. [Google Scholar] [CrossRef]
  19. Lee, P.H.; Wu, S.W.; Hung, Y.P. Illumination Compensation Using Oriented Local Histogram Equalization and its Application to Face Recognition. IEEE Trans. Image Process. 2012, 21, 4280–4289. [Google Scholar] [CrossRef] [PubMed]
  20. Shashua, A.; Riklin-Raviv, T. The Quotient Image: Class-Based Re-Rendering and Recognition with Varying Illuminations. IEEE Trans. Pattern Anal. Mach. Intell. 1999, 23, 129–139. [Google Scholar] [CrossRef] [Green Version]
  21. Wang, H.; Li, S.Z.; Wang, Y. Generalized quotient image. In Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Washington, DC, USA, 2–27 June 2004; Volume 2. [Google Scholar]
  22. Chen, T.; Yin, W.; Zhou, X.S.; Comaniciu, D.; Huang, T.S. Illumination normalization for face recognition and uneven background correction using total variation based image models. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; Volume 2, pp. 532–539. [Google Scholar]
  23. Srisuk, S.; Petpon, A. A Gabor Quotient Image for Face Recognition under Varying Illumination. In Proceedings of the ISVC, Las Vegas, NV, USA, 1–3 December 2008. [Google Scholar]
  24. An, G.; Wu, J.; Ruan, Q. An illumination normalization model for face recognition under varied lighting conditions. Pattern Recognit. Lett. 2010, 31, 1056–1067. [Google Scholar] [CrossRef]
  25. Jobson, D.; Rahman, Z.; Woodell, G. Properties and performance of a center/surround retinex. IEEE Trans. Image Process. 1997, 6, 451–462. [Google Scholar] [CrossRef] [PubMed]
  26. Fitzgibbon, A.W.; Zisserman, A. On Affine Invariant Clustering and Automatic Cast Listing in Movies. In Proceedings of the ECCV, Copenhagen, Denmark, 28–31 May 2002. [Google Scholar]
  27. Chen, C.P.; Chen, C.S. Lighting normalization with generic intrinsic illumination subspace for face recognition. In Proceedings of the Tenth IEEE International Conference on Computer Vision (ICCV’05), Beijing, China, 17–21 October 2005; Volume 2, pp. 1089–1096. [Google Scholar]
  28. Du, S.; Ward, R.K. Wavelet-based illumination normalization for face recognition. In Proceedings of the IEEE International Conference on Image Processing, Genoa, Italy, 11–14 September 2005; Volume 2, p. 954. [Google Scholar]
  29. Han, H.; Shan, S.; Chen, X.; Gao, W. A comparative study on illumination preprocessing in face recognition. Pattern Recognit. 2013, 46, 1691–1699. [Google Scholar] [CrossRef]
  30. Fan, C.N.; Zhang, F.Y. Homomorphic filtering based illumination normalization method for face recognition. Pattern Recognit. Lett. 2011, 32, 1468–1479. [Google Scholar] [CrossRef]
  31. Wang, B.; Li, W.; Yang, W.; Liao, Q. Illumination Normalization Based on Weber’s Law with Application to Face Recognition. IEEE Signal Process. Lett. 2011, 18, 462–465. [Google Scholar] [CrossRef]
  32. Zhao, X.; Shah, S.K.; Kakadiaris, I.A. Illumination Normalization Using Self-lighting Ratios for 3D2D Face Recognition. In Proceedings of the ECCV Workshops, Florence, Italy, 7–13 October 2012. [Google Scholar]
  33. Li, Y.; Meng, L.; Feng, J. Lighting Coefficients Transfer Based Face Illumination Normalization. In Proceedings of the CCPR, Beijing, China, 24–26 September 2012. [Google Scholar]
  34. BimaSenaBayu, D.; Miura, J. Fuzzy-based illumination normalization for face recognition. In Proceedings of the 2013 IEEE Workshop on Advanced Robotics and Its Social Impacts, Tokyo, Japan, 7–9 November 2013; pp. 131–136. [Google Scholar]
  35. Goel, T.; Nehra, V.; Vishwakarma, V.P. Illumination normalization using down-scaling of low-frequency DCT coefficients in DWT domain for face recognition. In Proceedings of the 2013 Sixth International Conference on Contemporary Computing (IC3), Noida, India, 9–11 August 2010; pp. 295–300. [Google Scholar]
  36. Vishwakarma, V.P. Illumination normalization using fuzzy filter in DCT domain for face recognition. Int. J. Mach. Learn. Cybern. 2015, 6, 17–34. [Google Scholar] [CrossRef]
  37. Zhao, X.; Evangelopoulos, G.; Chu, D.; Shah, S.K.; Kakadiaris, I.A. Minimizing Illumination Differences for 3D to 2D Face Recognition Using Lighting Maps. IEEE Trans. Cybern. 2014, 44, 725–736. [Google Scholar] [CrossRef]
  38. Tu, X.; Yang, F.; Xie, M.; Ma, Z. Illumination Normalization for Face Recognition Using Energy Minimization Framework. IEICE Trans. 2017, 100-D, 1376–1379. [Google Scholar] [CrossRef] [Green Version]
  39. Ahmad, F.; Khan, A.; Islam, I.U.; Uzair, M.; Ullah, H. Illumination normalization using independent component analysis and filtering. Imaging Sci. J. 2017, 65, 308–313. [Google Scholar] [CrossRef]
  40. Zhang, Y.; Wang, L.; Guan, X.; Wei, H. Illumination Normalization for Face Recognition via Jointly Optimized Dictionary-Learning and Sparse Representation. IEEE Access 2018, 6, 66632–66640. [Google Scholar] [CrossRef]
  41. Zheng, C.; Wu, S.; Xu, W.; Xie, S. Illumination Normalization via Merging Locally Enhanced Textures for Robust Face Recognition. arXiv 2019, arXiv:1905.03904. [Google Scholar]
  42. Zhang, W.; Zhao, X.; Morvan, J.M.; Chen, L. Improving Shadow Suppression for Illumination Robust Face Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 41, 611–624. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  43. Liu, Y.; Xu, H.; Shang, D.; Li, C.; Quan, X. An Underwater Image Enhancement Method for Different Illumination Conditions Based on Color Tone Correction and Fusion-Based Descattering. Sensors 2019, 19, 5567. [Google Scholar] [CrossRef] [Green Version]
  44. Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.C.; Bengio, Y. Generative Adversarial Nets. Adv. Neural Inf. Process. Syst. 2014, 27, 2672–2680. [Google Scholar]
  45. Radford, A.; Metz, L.; Chintala, S. Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks. In International Conference on Learning Representations; Springer: Cham, Switzerland, 2016. [Google Scholar]
  46. Mirza, M.; Osindero, S. Conditional Generative Adversarial Nets. arXiv 2014, arXiv:1411.1784. [Google Scholar]
  47. Yi, Z.; Zhang, H.; Tan, P.; Gong, M. DualGAN: Unsupervised Dual Learning for Image-to-Image Translation. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2868–2876. [Google Scholar]
  48. Kim, T.; Cha, M.; Kim, H.; Lee, J.K.; Kim, J. Learning to discover cross-domain relations with generative adversarial networks. In Proceedings of the ICML’17 34th International Conference on Machine Learning, Sydney, NSW, Australia, 6–11 August 2017; Volume 70, pp. 1857–1865. [Google Scholar]
  49. Ma, S.; Ma, H.; Xu, Y.; Li, S.; Lv, C.; Zhu, M. A Low-Light Sensor Image Enhancement Algorithm Based on HSI Color Model. Sensors 2018, 18, 3583. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  50. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  51. Cao, Q.; Shen, L.; Xie, W.; Parkhi, O.M.; Zisserman, A. VGGFace2: A Dataset for Recognising Faces across Pose and Age. In Proceedings of the 2018 13th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2018), Xi’an, China, 15–19 May 2018; pp. 67–74. [Google Scholar]
  52. Wu, X.; He, R.; Sun, Z.; Tan, T. A light CNN for deep face representation with noisy labels. IEEE Trans. Inf. Forensics Secur. 2018, 13, 2884–2896. [Google Scholar] [CrossRef] [Green Version]
  53. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [Green Version]
  54. Sheikh, H.R.; Bovik, A.C. Image information and visual quality. IEEE Trans. Image Process. 2006, 15, 430–444. [Google Scholar] [CrossRef] [PubMed]
  55. Zhang, L.; Zhang, L.; Mou, X.; Zhang, D. FSIM: A Feature Similarity Index for Image Quality Assessment. IEEE Trans. Image Process. 2011, 20, 2378–2386. [Google Scholar] [CrossRef] [Green Version]
  56. Lin, Y.; Ling, S.; Fu, K.; Cheng, P. An Identity-Preserved Model for Face Sketch-Photo Synthesis. IEEE Signal Process. Lett. 2020, 27, 1095–1099. [Google Scholar] [CrossRef]
  57. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Medical Image Computing and Computer Assisted Intervention; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  58. Ulyanov, D.; Vedaldi, A.; Lempitsky, V.S. Instance Normalization: The Missing Ingredient for Fast Stylization. arXiv 2016, arXiv:1607.08022. [Google Scholar]
  59. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. ImageNet Large Scale Visual Recognition Challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar]
  60. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  61. Gross, R.; Matthews, I.; Cohn, J.; Kanade, T.; Baker, S. Multi-PIE. Image Vis. Comput. 2010, 28, 807–813. [Google Scholar] [CrossRef] [PubMed]
  62. Phillips, P.; Flynn, P.; Scruggs, T.; Bowyer, K.; Chang, J.; Hoffman, K.; Marques, J.; Min, J.; Worek, W. Overview of the face recognition grand challenge. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; Volume 1, pp. 947–954. [Google Scholar]
  63. Zhang, S.; Zhu, X.; Lei, Z.; Shi, H.; Wang, X.; Li, S.Z. S3FD: Single Shot Scale-Invariant Face Detector. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 192–201. [Google Scholar]
  64. Tang, Y.; Salakhutdinov, R.; Hinton, G.E. Deep Lambertian Networks. In Proceedings of the ICML, Edinburgh, Scotland, UK, 26 June–1 July 2012. [Google Scholar]
  65. Gao, S.; Zhang, Y.; Jia, K.; Lu, J.; Zhang, Y. Single Sample Face Recognition via Learning Deep Supervised Autoencoders. IEEE Trans. Inf. Forensics Secur. 2015, 10, 2108–2118. [Google Scholar] [CrossRef]
  66. Kingma, D.P.; Ba, J.L. Adam: A Method for Stochastic Optimization. In International Conference on Learning Representations; Springer: Cham, Switzerland, 2015. [Google Scholar]
Figure 1. Some poor-lighted faces under various illumination conditions.
Figure 1. Some poor-lighted faces under various illumination conditions.
Sensors 20 04869 g001
Figure 2. Our overall network framework. The core of our method mainly contains 3 parts: an elaborately designed generator, a discriminator and a loss network.
Figure 2. Our overall network framework. The core of our method mainly contains 3 parts: an elaborately designed generator, a discriminator and a loss network.
Sensors 20 04869 g002
Figure 3. The detailed structure of our generator. 7, 128 × 128 × 64 means k e r n e l _ s i z e = 7 , feature maps with 128 values in width, 128 values in height and 64 channels.
Figure 3. The detailed structure of our generator. 7, 128 × 128 × 64 means k e r n e l _ s i z e = 7 , feature maps with 128 values in width, 128 values in height and 64 channels.
Sensors 20 04869 g003
Figure 4. The detailed structure of our discriminator. The input images are a 128 × 128 paired color images such as ( x , y ) and ( x , G ( x ) ) . 4, 64 × 64 × 64 means k e r n e l _ s i z e = 4 , feature maps with 64 values in width, 64 values in height and 64 channels.
Figure 4. The detailed structure of our discriminator. The input images are a 128 × 128 paired color images such as ( x , y ) and ( x , G ( x ) ) . 4, 64 × 64 × 64 means k e r n e l _ s i z e = 4 , feature maps with 64 values in width, 64 values in height and 64 channels.
Sensors 20 04869 g004
Figure 5. Comparisons between the CycleGAN [13], the DMIT [14], the EDIT [15], the Pix2Pix [16] and ours on the MultiPIE database. The first column: original poor-lighted face images. The second to sixth columns: synthesized results of the original poor-lighted face images.
Figure 5. Comparisons between the CycleGAN [13], the DMIT [14], the EDIT [15], the Pix2Pix [16] and ours on the MultiPIE database. The first column: original poor-lighted face images. The second to sixth columns: synthesized results of the original poor-lighted face images.
Sensors 20 04869 g005
Figure 6. Comparisons between the LDCT [3], the LN [4], the LTV [5], the SQI [6], the TT [7], the CycleGAN [13], the Pix2Pix [16] and ours on the YaleB database. The first column: original poor-lighted face images. The second to ninth columns: synthesized results of the original poor-lighted face images.
Figure 6. Comparisons between the LDCT [3], the LN [4], the LTV [5], the SQI [6], the TT [7], the CycleGAN [13], the Pix2Pix [16] and ours on the YaleB database. The first column: original poor-lighted face images. The second to ninth columns: synthesized results of the original poor-lighted face images.
Sensors 20 04869 g006
Figure 7. FR-IQA values of the face images synthesized by various algorithms on the MultiPIE dataset. (a) SSIM values of generated faces; (b) VIF values of synthesized faces; (c) FSIM values of synthesized faces.
Figure 7. FR-IQA values of the face images synthesized by various algorithms on the MultiPIE dataset. (a) SSIM values of generated faces; (b) VIF values of synthesized faces; (c) FSIM values of synthesized faces.
Sensors 20 04869 g007
Figure 8. The ROC curves of face verification using ResNet-50, Light-CNN-9 and Light-CNN-29 on the MultiPIE dataset. (a) ROC curves of ResNet-50; (b) ROC curves of Light-CNN-9; (c) ROC curves of Light-CNN-29.
Figure 8. The ROC curves of face verification using ResNet-50, Light-CNN-9 and Light-CNN-29 on the MultiPIE dataset. (a) ROC curves of ResNet-50; (b) ROC curves of Light-CNN-9; (c) ROC curves of Light-CNN-29.
Sensors 20 04869 g008
Figure 9. FR-IQA values of the face images synthesized by various algorithms on the YaleB dataset. (a) SSIM values of synthesized faces; (b) VIF values of synthesized faces; (c) FSIM values of synthesized faces.
Figure 9. FR-IQA values of the face images synthesized by various algorithms on the YaleB dataset. (a) SSIM values of synthesized faces; (b) VIF values of synthesized faces; (c) FSIM values of synthesized faces.
Sensors 20 04869 g009
Figure 10. The ROC curves of face verification using ResNet-50, Light-CNN-9 and Light-CNN-29 on the YaleB dataset. (a) ROC curves of ResNet-50; (b) ROC curves of Light-CNN-9; (c) ROC curves of Light-CNN-29.
Figure 10. The ROC curves of face verification using ResNet-50, Light-CNN-9 and Light-CNN-29 on the YaleB dataset. (a) ROC curves of ResNet-50; (b) ROC curves of Light-CNN-9; (c) ROC curves of Light-CNN-29.
Sensors 20 04869 g010
Figure 11. Some poor-lighted faces of FRGC dataset before and after processing by our algorithm. The first row is original facial images under various illumination conditions. The second row is the illumination-processing results of our method.
Figure 11. Some poor-lighted faces of FRGC dataset before and after processing by our algorithm. The first row is original facial images under various illumination conditions. The second row is the illumination-processing results of our method.
Sensors 20 04869 g011
Figure 12. The synthesized results of our method trained under various loss items. Conv1-5 means to use 1 to 5 convolutional layers of VGG-16 for extracting feature and computing loss. Conv1-6 means to use 1 to 6 convolutional layers of VGG-16 for extracting feature and computing loss. Conv1-7 means to use 1 to 7 convolutional layers of VGG-16 for extracting feature and computing loss. Combination denotes to combine the former three loss items for extracting multi-stage feature maps and computing multi-stage feature maps loss.
Figure 12. The synthesized results of our method trained under various loss items. Conv1-5 means to use 1 to 5 convolutional layers of VGG-16 for extracting feature and computing loss. Conv1-6 means to use 1 to 6 convolutional layers of VGG-16 for extracting feature and computing loss. Conv1-7 means to use 1 to 7 convolutional layers of VGG-16 for extracting feature and computing loss. Combination denotes to combine the former three loss items for extracting multi-stage feature maps and computing multi-stage feature maps loss.
Sensors 20 04869 g012
Table 1. Average FR-IQA values of the face images synthesized by various illumination-processing methods on the MultiPIE dataset.
Table 1. Average FR-IQA values of the face images synthesized by various illumination-processing methods on the MultiPIE dataset.
MetricsSSIM ↑VIF ↑FSIM ↑
Methods
Original0.32890.16690.7165
CycleGAN [13]0.74770.33400.8644
DMIT [14]0.70320.28450.8446
EDIT [15]0.72160.31320.8591
Pix2Pix [16]0.72650.32210.8548
Ours0.78990.37880.8865
Table 2. Top-1 identification accuracy and verification accuracy at 0.1% FAR on the MultiPIE dataset using ResNet-50, Light-CNN-9 and Light-CNN-29.
Table 2. Top-1 identification accuracy and verification accuracy at 0.1% FAR on the MultiPIE dataset using ResNet-50, Light-CNN-9 and Light-CNN-29.
FRTop-1VR@FAR = 0.1%
VerifiersResNet-50 [50]Light-CNN-9 [52]Light-CNN-29 [52]ResNet-50 [50]Light-CNN-9 [52]Light-CNN-29 [52]
Methods
Original0.93860.94560.96050.91220.91840.9552
CycleGAN [13]0.98510.98250.99380.96660.97360.9859
DMIT [14]0.93770.97720.98770.91140.96490.9728
EDIT [15]0.98250.98680.99560.95000.97010.9929
Pix2Pix [16]0.97890.97890.99210.95870.97190.9754
Ours0.99830.9991100.0000.98590.99380.9974
Table 3. Average FR-IQA values of the face images synthesized by various illumination-processing methods on the YaleB dataset.
Table 3. Average FR-IQA values of the face images synthesized by various illumination-processing methods on the YaleB dataset.
MetricsSSIM ↑VIF ↑FSIM ↑
Methods
Original0.35690.14130.7410
LDCT [3]0.36230.05030.7370
LN [4]0.35260.14350.7370
LTV [5]0.39140.04090.6602
SQI [6]0.38910.07750.7092
TT [7]0.30360.05330.6365
CycleGAN [13]0.58490.20330.8318
Pix2Pix [16]0.60310.21280.8316
Ours0.69460.26210.8660
Table 4. Top-1 identification accuracy and verification accuracy at 0.1% FAR on the YaleB dataset using ResNet-50, Light-CNN-9 and Light-CNN-29.
Table 4. Top-1 identification accuracy and verification accuracy at 0.1% FAR on the YaleB dataset using ResNet-50, Light-CNN-9 and Light-CNN-29.
FRTop-1VR@FAR = 0.1%
VerifiersResNet-50 [50]Light-CNN-9 [52]Light-CNN-29 [52]ResNet-50 [50]Light-CNN-9 [52]Light-CNN-29 [52]
Methods
Original0.51260.73020.80640.42850.45870.6492
SQI [6]0.50320.83170.90000.30310.50630.6952
LDCT [3]0.15230.59040.58890.07140.30470.3571
LN [4]0.49680.73330.82060.42380.45390.6650
LTV [5]0.13170.66670.56350.03170.33490.3253
TT [7]0.10640.58730.67620.03010.26030.4269
CycleGAN [13]0.68890.77620.80150.59680.63330.7285
Pix2Pix [16]0.18890.20160.10480.02220.04920.0016
Ours0.89520.90160.94290.87140.87460.9317
Table 5. Average FR-IQA values of poor-lighted face images processed by our methods trained by various loss items on the MultiPIE dataset.
Table 5. Average FR-IQA values of poor-lighted face images processed by our methods trained by various loss items on the MultiPIE dataset.
VGG-16Conv1-5Conv1-6Conv1-7Combination
Metrics
SSIM [53]0.77620.77560.77150.7814
VIF [54]0.36420.36510.36120.3685
FSIM [55]0.88280.88300.88190.8837

Share and Cite

MDPI and ACS Style

Ling, S.; Lin, Y.; Fu, K.; You, D.; Cheng, P. A High-Performance Face Illumination Processing Method via Multi-Stage Feature Maps. Sensors 2020, 20, 4869. https://doi.org/10.3390/s20174869

AMA Style

Ling S, Lin Y, Fu K, You D, Cheng P. A High-Performance Face Illumination Processing Method via Multi-Stage Feature Maps. Sensors. 2020; 20(17):4869. https://doi.org/10.3390/s20174869

Chicago/Turabian Style

Ling, Shenggui, Ye Lin, Keren Fu, Di You, and Peng Cheng. 2020. "A High-Performance Face Illumination Processing Method via Multi-Stage Feature Maps" Sensors 20, no. 17: 4869. https://doi.org/10.3390/s20174869

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop