Next Article in Journal
Automatic Identification of Ultrasound Images of the Tibial Nerve in Different Ankle Positions Using Deep Learning
Previous Article in Journal
Artificial Neural Networks for IoT-Enabled Smart Applications: Recent Trends
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

DVDR-SRGAN: Differential Value Dense Residual Super-Resolution Generative Adversarial Network

1
School of Electronics and Information Engineering, Liaoning University of Technology, Jinzhou 121001, China
2
College of Science, Liaoning University of Technology, Jinzhou 121001, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(10), 4854; https://doi.org/10.3390/s23104854
Submission received: 21 February 2023 / Revised: 5 May 2023 / Accepted: 15 May 2023 / Published: 18 May 2023
(This article belongs to the Section Sensing and Imaging)

Abstract

:
In the field of single-image super-resolution reconstruction, GAN can obtain the image texture more in line with the human eye. However, during the reconstruction process, it is easy to generate artifacts, false textures, and large deviations in details between the reconstructed image and the Ground Truth. In order to further improve the visual quality, we study the feature correlation between adjacent layers and propose a differential value dense residual network to solve this problem. We first use the deconvolution layer to enlarge the features, then extract the features through the convolution layer, and finally make a difference between the features before being magnified and the features after being extracted so that the difference can better reflect the areas that need attention. In the process of extracting the differential value, using the dense residual connection method for each layer can make the magnified features more complete, so the differential value obtained is more accurate. Next, the joint loss function is introduced to fuse high-frequency information and low-frequency information, which improves the visual effect of the reconstructed image to a certain extent. The experimental results on Set5, Set14, BSD100, and Urban datasets show that our proposed DVDR-SRGAN model is improved in terms of PSNR, SSIM, and LPIPS compared with the Bicubic, SRGAN, ESRGAN, Beby-GAN, and SPSR models.

1. Introduction

Super-resolution (SR) reconstruction of images is a technique for obtaining high-resolution images from single or multiple low-resolution images. In real life, limited by factors such as digital imaging equipment and hardware storage capacity, the image resolution obtained cannot meet people’s needs, especially in some specific fields such as surveillance, remote sensing, military, medicine, and so on. Using super-resolution image reconstruction technology to restore and reconstruct low-resolution images can efficiently improve image details and image quality.
Image reconstruction methods can be roughly divided into three categories, which are interpolation-based [1], reconstruction-based [2], and learning-based [3]. With the development of the times, the two methods based on interpolation and reconstruction have some shortcomings, such as reconstructing the image blur, insufficient computational power, and serious detail loss. In recent years, the development of deep learning has made the learning-based image reconstruction technology gradually become a research hotspot, and interpolation-based and reconstruction-based technologies are usually used to preprocess images.
In recent years, with the rapid rise of artificial intelligence, image super-resolution reconstruction methods based on the convolutional neural network (CNN) and generative adversarial network (GAN) have been widely used because their reconstruction performance far exceeds that of traditional algorithms. Dong et al. [4] proposed the super-resolution convolutional neural network (SRCNN), which uses three convolutional layers for reconstruction and greatly improves the speed of reconstruction compared to traditional methods. Kim et al. [5] studied SRCNN and constructed a reconstruction model with deep-level networks, which used VGG network architecture to construct a deep convolution network with 20 layers and greatly improved the network convergence speed via a very high learning rate during the training process. Jiang et al. [6] proposed a new differential network in image super-resolution reconstruction to increase the image reconstruction effect and used the multi-path supervised reconstruction blocks to monitor the reconstruction process, making the network more stable. Lim et al. [7] proposed an enhanced depth super-resolution network (EDSR), which removes unnecessary batch normalization (BN) layers in the residual blocks, resulting in a significant performance improvement in the reconstruction. Zhang et al. [8] proposed RCAN, which constructed channel attention to focus on improving the PSNR value. Hu et al. [9] proposed Meta-SR to achieve the effect of up-sampling images to arbitrary sizes. Li et al. [10] proposed a feedback framework for gradually refining the super-resolved results.
The CNN models have achieved excellent results in traditional detection, but they rely too much on the objective function that minimizes the mean square error (MSE). In this way, the PSNR is high, but the reconstructed images are too smooth, and the high-frequency details of the high-resolution images are missing. As the depth of the training network increases, convergence becomes slower and slower, resulting in disappearing gradients, instability, etc. The performance of the generative adversarial network [11] model in image reconstruction surprised researchers. Huang et al. [12] and Goodfellow et al. [13] discussed and compared the importance of GAN models and their variants in generating samples. Creswell et al. [14] provided a survey summary of GAN evaluation methods and training problems. Ledig et al. [15] first applied a generative adversarial network to the field of image super-resolution reconstruction and proposed a generative adversarial network super-resolution reconstruction model (SRGAN), which uses perceptual loss and adversarial loss to jointly recover texture details of images. Bulat et al. [16] designed a GAN based on a High-to-Low network to obtain more natural LR images from HR images to simulate real low-resolution data. Based on SRGAN, Wang et al. [17] proposed an enhanced super-resolution generative adversarial network (ESRGAN) under the condition of improving the network architecture, adversarial loss, and perceptual loss. The network uses dense residual blocks instead of residual blocks in the original generator and uses the relative discriminator to further improve the quality of the reconstructed images. Zhang et al. [18] believed that the process of constructing sample images via down-sampling HR lost the details and accuracy of the original data, so they proposed a super-resolution reconstruction method for real scenes, which has an important impact on subsequent super-resolution reconstruction methods. The SRFlow method proposed by Lugmayr et al. [19] learns image super-resolution through normalized flow and can generate several different SR images from LR images. Rad et al. [20] made adjustments to the composition of perceptual loss and proposed adjusting a target perceptual loss based on object, background, and boundary labels. In particular, Li et al. [21] considered that one-to-one supervision was not the most reasonable way and thus proposed the Beby-GAN with one-to-many supervision. Ma et al. [22] believed that the structure information is very important in the super-resolution problem and proposed implementing an additional gradient branch network. This network generates a high-resolution gradient map as an additional structure prior, making the generated images clearer.
However, most existing networks ignore the feature correlation of adjacent layers and the utilization of some low-frequency information, which leads to the insufficient use of features at different levels. In view of the problems of artifacts and insufficient detail processing in existing algorithms, we propose a differential value dense residual block as the basic unit of the generator to solve the above problems on the basis of improving the generator network structure. At the same time, we use the relative discriminator rather than the standard discriminator, which can further enhance the performance of the generator. Additionally, the average absolute error loss, adversarial loss, and the new joint perceptual loss function are used to train the generator so that the reconstruction performance of the network can reach the best.
The main contributions of this paper are itemized as follows:
  • The use of the traditional convolutional neural networks for image reconstruction neglects the utilization between adjacent layers of the network, which leads to poor image quality. In order to solve this problem, we propose a differential value dense residual network to improve the utilization between adjacent layers of the network;
  • In order to make the training process of the GAN stable and the perceptual quality of reconstructed images better, we introduce a joint perceptual loss function. The loss function uses different layers of the VGG network to extract information with different concerns and fuses high-frequency information and low-frequency information to more effectively guide the network training;
  • To verify the validity and wideness of the proposed method in this paper, we used the proposed model to conduct quantitative and qualitative studies under four datasets and to perform relevant ablation studies.

2. Methods

Based on ESRGAN, we propose a differential value dense residual super-resolution reconstruction network model (DVDR-SRGAN), which consists of an improved generator and a discriminator. The generator network is mainly composed of multiple differential value dense residual blocks (DVDB). Each DVDB is composed of differential value structure (DV) and dense residual block (DB). Therefore, the generator can extract the data features from two perspectives. On the one hand, the differential value structure is used to obtain the lost information of each convolution layer; on the other hand, the dense residual block is used to enrich the input information of each deconvolution layer. Then, the information extracted is combined to generate the data that can “cheat” discriminator under the effects of joint perceptual loss, adversarial loss, and mean absolute error loss.

2.1. Network Model of DVDR-SRGAN Generator

Based on the idea of deconvolution and differential values, the network structure of the generator network model is improved, and the generator network model is shown in Figure 1, which is divided into three parts: shallow feature extraction block, differential value dense residual block set, and image reconstruction.

2.1.1. The Extraction Block of Shallow Feature

The extraction block of shallow feature consists of a convolutional layer, which is represented by Conv. We first input a low-resolution image I L R , and then obtain the shallow features of the I L R image by extracting block of shallow features, as shown in the following Formula (1):
F i n = f T ( I L R ) ,
where f T ( ) represents the operation of the shallow feature extraction block, and F i n represents the shallow feature of the I L R image.

2.1.2. A Set of Differential Value Dense Residual Blocks

Convolutional neural network extract features from images by means of layer-by-layer sampling. If the area mapped by the pixel on the output feature of the convolution layer on the input is too small, that is, the receptive field is too small, then you can only focus on the local feature. If the receptive field is too large, it contains too much useless information. In order to effectively use the features of each layer and strengthen the feature transmission between each layer, we introduce the differential value dense residual technology into the convolutional network. The dvdb network structure is designed as shown in Figure 2, where Deconv represents deconvolution layer, Conv represents convolution layer, and LRelu represents the Leaky ReLu activation function. First, the dense residual technique is used so that the input of each deconvolution layer is a stitching of the input of the previous deconvolution layer and the output of the corresponding convolution layer. The introduction of this structure allows each layer to make direct use of the gradient information and input information of the loss function, thus alleviating the gradient disappearance phenomenon to a certain extent and helping to train deeper networks. Secondly, in order to avoid only paying attention to local features during feature extraction, a differential value structure is introduced. Feature extraction is carried out according to the differential value, and the extracted features are given corresponding weights, thus enhancing the ability of information extraction and improving the network performance and accuracy. Based on the dvdb network structure, a set of differential value dense residual blocks is designed, which is composed of multiple DVDB modules. Each DVDB module contains three dvdb modules.
In the dvdb network structure, the low-resolution image x is first sent to Deconv, and then Deconv amplifies the features of x to obtain more high-frequency information. Next, the output F d e c o n i ( x ) of Deconv is input into conv, and the size of the output F c o n i ( x ) of conv is guaranteed to be the same as the size of x. In order to prevent information loss and network failure to train as the network increases in depth, we adopt dense residual technology [23]. This technique compensates for lost information by fusing x and the output F c o n i ( x ) of conv. As the fused information becomes more reflective of the original high-frequency information, some of the lower-frequency information about the image texture is ignored. In order to retain low-frequency information, we make a difference between x and F c o n i ( x ) to obtain the differential value between them. This difference in value can enrich the acquired image information. We sum the differential values of each group of Deconv input and conv output in dvdb and then pass the sum value through a conv and an LRelu successively. The result obtained is weighted with the dense residual result to obtain the final result of dvdb.
F d-value = i = 1 2 F i n i 1 F c o n i ( F d e c o n i ( F i n i 1 ) )
F d v d b = α F d-value + β F d b
F o u t = F d v d b ( F i n ) + F i n
where F d-value ( ) represents the sum of differential values in dvdb. F d e c o n i ( ) and F c o n i ( ) represent the output values of the ith deconvolution layer and the ith convolution layer, respectively. F d b ( ) represents the result of dense residual, and F o u t ( ) represents the output of dvbd.

2.1.3. Image Reconstruction

First, the output of the differential value dense residual block set is upsampled, then the upsampled result is passed through two convolution layers in turn, and, finally, the required size SR image is output. This reconstruction method reduces network complexity.

2.2. Relative Discriminator Network

The traditional discriminator of GAN can only determine the probability that the input image is real and natural under the same network structure, while the relative discriminator is introduced in ESRGAN to try to predict the probability that the real image is more real than the false image, as shown in Figure 3.
In adversarial training, using this discriminator can help the network learn clearer edges and finer textures. Therefore, this paper adopts the relative discriminator RaD of ESRGAN. The specific network model is shown in Figure 4, where Conv represents the convolution layer, LRelu represents the Leaky ReLu activation function, BN represents the batch normalization layer, and Dense represents the full connection layer.
In the relative discriminator network model, Leaky ReLU and BN layers are used simultaneously, and pooling operation is not used. The relative discriminator network consists of 8 convolution layers, all of which are 3 × 3 convolutional kernels in size. The number of convolutional kernels has doubled from 64 to 512. The discriminator network alternately uses convolution layers with step size, which is 1 and 2, respectively. When the number of features doubles, the convolution layer with step size of 2 is used to reduce the image resolution. After the convolution layer obtains advanced features, the probability is finally obtained through two full connection layers and the sigmoid activation function.
The loss function of the relative discriminator and the adversarial loss function of the generator are shown in Formulas (5) and (6):
L D R a = E x r [ log ( D R a ( x r , x f ) ) ] E x f [ log ( 1 D R a ( x f , x r ) ) ]
L G R a = E x r [ log ( 1 D R a ( x r , x f ) ) ] E x f [ log ( D R a ( x f , x r ) ) ]
D R a ( x r , x f ) = σ ( C ( x r ) E x f [ C ( x f ) ] )
where x f represents the input LR image; x r represents the input HR image; E x f [ ] represents the average of the data generated by all generators; σ is a sigmoid function; L D R a ( ) represents the loss of discriminator; L G R a ( ) represents the generator adversarial loss. It can be seen that the adversarial loss of the generator include x r and x f . Therefore, in adversarial training, our generator benefits from the gradient between generated data and real data, while in traditional GAN, only the generated data is effective.

2.3. Loss Function

The traditional mean square error loss (MSE) is introduced in SRGAN as a part of the loss, while the relative discriminator is introduced in ESRGAN to cancel the mean square error loss, and the pre-activated features are used to construct the perceptual loss to obtain more information. Research [17] shows that the use of pre-activated features will cause the brightness of the reconstructed image to be inconsistent with the real image, so this paper continues to use the pre-activated features. On this basis, a new joint perceptual loss function is proposed, which combines the mean absolute error and adversarial loss to train the network.

2.3.1. Joint Perceived Loss Function

VGG-54 is defined in the deep network, which extracts the feature map of high-frequency features. These feature maps pay more attention to the content. VGG-22 is defined in the shallow network, and most of the extracted features are low-frequency features such as contour and shape [24]. Only using VGG-54 to define the loss will make the texture of the reconstructed image too sharp, which results in distortion of the details and produces noise and artifacts, etc. On the basis of ESRGAN, this paper uses the pre-activation features of VGG-54 and VGG-22 to construct the joint perceptual loss function so that the reconstructed image has the feature of smooth details, offsets the part noise generated, and can also have a good subjective visual effect. The improved loss is shown in Formula (8) as follows:
l V G G S R = α l V G G 22 + β l V G G 54 ,
where l V G G 22 is the feature map before the 2nd convolutional layer before the 2nd pooling in the VGG19 network, which is the ultra-low frequency feature; l V G G 54 is the feature map before the 4th convolutional layer before the 5th pooling in the VGG19 network, which is the high-frequency feature; α and β are pre-set parameters. We will conduct the related experiments on the parameter values in Section 3.

2.3.2. Mean Absolute Error Loss

The mean absolute error (MAE) is the average of the absolute error between the predicted value and the observed value. MAE can avoid the error offsetting each other, so it can accurately reflect the actual prediction error. Therefore, this paper combines the average absolute error loss with the joint perceptual loss to increase the prediction effect of the model. The formula of MAE is shown as follows.
l M A E = 1 m i = 1 m | G ( I i L R ) I i H R | ,
where l M A E represents the mean absolute error loss function, m is the number of iterations, i is the HR serial number, I i H R is the distribution of the real image, and G ( I i L R ) is the distribution of the high-resolution image generated by the generator.
We fuse the joint perceived loss proposed in Section 2.3.1, the average absolute error, and the adversarial loss given in Section 2.3.2. The improved loss function is shown in the following Formula (10):
l S R = l V G G S R + η l M A E + γ l G e n S R ,
where l G e n S R represents adversarial loss, which can make the images reconstructed by generator network deceive the discriminator network as much as possible and enhance the expressive force of the reconstructed image in terms of visual perceptual. η and γ are coefficients that balance different loss terms. We will conduct relevant experiments on parameter values in Section 3.4.

3. Experiments

3.1. Datasets and Evaluation Metrics

The training set used in this paper is the DIV2K dataset [25], which contains 800 images. These images are divided into various types, such as city, architecture, landscape, nature, and so on. These images are all from the real world and have high complexity and authenticity. The test datasets include Set5 [26], Set14 [27], BSD100 [28], and Urban100 [29]. Set5 and Set14 are low-complexity single-image super-resolution datasets. BSD100 contains various degraded images, such as images with noise, blur, and lossless compression. Urban100 is an image super-resolution dataset for urban environments, which contains 100 high-resolution images. These four datasets are commonly used for performance testing of super-resolution tasks. When training the network, first, the high-resolution images in the training set are randomly flipped horizontally or vertically, and then the images are clipped to obtain 128 × 128 high-quality image blocks. Finally, we perform the bicubic interpolation on these image blocks to obtain the down-sampling 4 × LR images for training.
In this paper, we use three evaluation metrics, namely peak signal-to-noise ratio (PSNR), structural similarity (SSIM) [30], and learned perceptual image patch similarly (LPIPS) [31] to evaluate the performance of algorithms. As shown in Formula (11), PSNR can evaluate the quality of the image by comparing the gray value difference of the pixels corresponding to the two images. The higher the PSNR value is, the smaller the distortion is. As shown in Formula (12), SSIM evaluates the similarity of the two images from brightness, contrast, and structure, and the closer the SSIM value is to 1, the more similar the structure of the reconstructed image is to the original image, and the better the reconstructed effect. As shown in Formula (13), the LPIPS metric can measure the difference between two images. The lower the value of LPIPS, the more similar the two images are.
P S N R ( X , Y ) = 10 lg 255 2 w h c m = 1 w n = 1 h z = 1 c [ X ( m , n ) Y ( m , n ) ] 2 ,
where X represents the original HR images; Y represents the reconstructed SR images by the generator; c represents the number of channels in the images; w and h represent the width and height of the images; m represents the m-th pixel on the width of the image; n represents the n-th pixel in the height of the image; z represents the z-th channel of three primary color channels.
S S I M ( X , Y ) = ( 2 μ X μ Y + C 1 ) ( 2 σ X Y + C 2 ) ( μ X 2 + μ Y 2 + C 1 ) ( σ X 2 + σ Y 2 + C 2 ) ,
where σ X represents the average value of X, and μ Y represents the average value of Y; σ X 2 represents the variance of X, σ Y 2 represents the variance of Y, and σ X Y represents the covariance between X and Y; C 1 = ( 0.001 L ) 2 , C 2 = ( 0.003 L ) 2 are the two variables used to maintain stability, and L represents the dynamic range of image pixels.
L P I P S ( x , x 0 ) = l 1 H l W l h , w | | w l ( y h w l y 0 h w l ) | | 2 2 .
The real image x and reconstructed image x 0 are sent to the neural network for feature extraction. The output of each layer is activated and normalized, denoted as y l , y 0 l R H l × W l × C l . Then, the L2 distance is calculated by multiplying the weight points of the w layer, and the average distance is obtained.

3.2. Training Details

The parameters of hardware equipment for experiment conduction are Intel(R) Xeon(R) CPU E5-2680 v4, instance memory: 28G, core: 28, 3080 Ti-12G; Experimental environment: Linux, PyTorch 1.12.0(Python3.8), Cuda11.3.
The training process is divided into two stages. First, a PSNR-oriented model is trained using the average absolute error as the loss function. The initial learning rate is set to 2 × 10−4. Next, the learning rate reduces to 0.5 times after each 5 × 104 iteration. Then, the perceptual loss and adversarial loss are introduced into the PSNR-oriented model to obtain the final model. The initial learning rate is set to 1 × 10−4. Next, the learning rate reduces to 0.5 times after each 5 × 104 iteration. In the training process, the model uses the Adam optimizer, the batch_ size is set to 16, and the generator network uses 18 DVDB feature extraction blocks. The trained PSNR-oriented model is used as the pre-trained model. One reason is to make the GAN more stable and avoid the remaining local optima in the generator, and the other reason is to ensure that the images received by the discriminator have high resolution, which helps the discriminator to focus more on texture recognition.
In order to obtain the values of α and β in Formula (8), we conducted the experiments on public test sets Set5 and Set14, and the experimental results are shown in Table 1. When the values of α and β are 0.2 and 0.8, respectively, the performance of PSNR, SSIM, and LPIPS reaches the best.

3.3. Comparative Experiments

In this paper, our proposed algorithm DVDR-SRGAN is compared with Bicubic, SRGAN [15], ESRGAN [17], Beby-GAN [21], and SPSR [22] on the public sets Set5, Set14, BSD100, and Urban100. In this paper, we evaluate the algorithm performance based on the quantitative results and qualitative results, as follows.

3.3.1. Quantitative Results

Table 2 shows the comparison results of the proposed algorithm with other comparative algorithms on multiple evaluation metrics. The algorithm DVDR-SRGAN has the highest PSNR and SSIM values on the Set5, Set14, and BSD100 datasets, and is also outstanding in LPIPS. Only when facing the Urban100 dataset, the metric results are slightly lower than the algorithm SPSR. The algorithm SPSR based on gradient loss also has better PSNR and SSIM values, but the LPIPS value is not as good as the algorithm DVDR-SRGAN, which shows that the visual quality of the image generated by the algorithm SPSR is much worse than that of the algorithm DVDR-SRGAN. The algorithm Beby-GAN performs poorly in reconstructing large images. The main reason is that the algorithm Beby-GAN uses a method with one-to-many supervision. All the values of evaluation metrics of the algorithm ESRGAN are not as good as algorithm DVDR-SRGAN, which indicates that the algorithm ESRGAN is at the expense of authenticity to optimize the visual quality of the reconstructed images. All the values of evaluation metrics of algorithm SRGAN and algorithm Bicubic are poor, indicating that satisfactory results cannot be obtained when facing tasks of high-quality image super-resolution. In summary, the algorithm DVDR-SRGAN can effectively solve the problem of the general lack of authenticity in adversarial perception methods and ensure the visual quality as much as possible while increasing image details.

3.3.2. Qualitative Results

In order to more intuitively feel the visual effect of the algorithm DVDR-SRGAN, we take SRGAN, ESRGAN, Beby-GAN, and SPSR as the comparison algorithms and carry out the comparison experiment based on the four test sets mentioned above. Figure 5, Figure 6, Figure 7, Figure 8, Figure 9 and Figure 10 show some of the results of reconstructed images. In order to better illustrate the reconstruction effect, we take a portion of each image for comparison, as shown in the red box in the image.
From the visual perspective, Figure 5 highlights the image of a parrot’s foot grasping a branch. It can be observed that except for the images of the parrot’s foot generated by DVDR-SRGAN and SPSR, which are more obvious, the images generated by other methods have a certain degree of distortion. However, compared to SPSR, DVDR-SRGAN can still make the branch image in the backward position clear. Figure 6 highlights the butterfly’s back and wing image. The reconstructed image using the algorithm DVDR-SRGAN is closest to GT, while the images generated by other algorithms either have a poor shape or have too many unrealistic artifacts. Figure 7 highlights the hair image near the boy’s temple. For the most obvious strand of hair in the middle, only the image reconstructed by algorithm DVDR-SRGAN is closest to GT, and other algorithms cannot do it. Figure 8 highlights the image of the man’s hand. The reconstructed image using the algorithm DVDR-SRGAN is closest to GT, while the images generated by other algorithms either have a poor shape or have too many unrealistic artifacts. Figure 9 highlights the image of the elephant’s ears. Only the algorithm DVDR-SRGAN can generate dense and clear cross stripes, while other algorithms cannot do it. Figure 10 highlights the pattern of the building. The reconstructed image using the algorithm DVDR-SRGAN is closest to GT, while other algorithms generated unrealistic artifacts when reconstructing the left wall.
In summary, the images generated by the algorithm SRGAN have serious detail loss. In addition to producing too many artifacts in the reconstruction process, the images generated by algorithm ESRGAN also have a loss of details. Due to the use of the method with one-to-many supervision in the algorithm Beby-GAN, it is not possible to perform better image reconstruction when dealing with large-sized images. Algorithm SPSR restores high-resolution gradient maps via a gradient branch to provide additional structure priors for the SR process and introduces the loss function of gradient, which alleviates the problem of image reconstruction distortion. However, a better reconstruction effect cannot be achieved for image areas with small color differences. The algorithm DVDR-SRGAN introduces deconvolution and differential value dense residual methods. This algorithm not only focuses on high-frequency information but also learns and extracts effective features from low-frequency information with weak color to make the reconstructed image more realistic.

3.4. Ablation Study

In order to verify the necessity of each part of our proposed network model, we conduct an ablation study on loss function via superposition. Figure 11 shows the relevant visual effects. With the superposition of the loss function, the structure of the reconstructed images is gradually clear, and the artifacts are eliminated, thus improving the authenticity and visual quality. When η = 10 2 , γ = 10 3 , the reconstructed images have the best effect.
As algorithm DVDR-SRGAN is proposed based on model ESRGAN, we conducted an ablation study on network modules. We designed three algorithms to compare them. The first algorithm (DVDB no JPL) is trained without using joint perceived loss, but the DVDB network module is applied in the network. The second algorithm (DVDB no dvdb) has the same network model as ESRGAN and is trained using joint perceived loss. The third algorithm (DVDB) uses our proposed model. The experimental results are shown in Table 3.
From Table 3, it can be seen that compared with ESRGAN, the network with DVDB has significantly improved the LPIPS performance, which proves the effectiveness of the proposed differential value dense residual in improving the perceptual quality. In addition, the experiment results of the network model with joint perceptual loss show that joint perceptual loss can significantly improve the performance of images in PSNR and SSIM. For algorithm, DVDR-SRGAN, all the values of evaluation metric are better than algorithm ESRGAN on different test sets. Therefore, the effectiveness of our proposed method has been clearly verified.

4. Conclusions

Aiming at the super-resolution task with high visual quality, this paper first proposes DVDR-SRGAN, a differential value dense residual network, which can obtain more effective information in the process of feature extraction and improve learning efficiency. Additionally, a relative discriminator has also been introduced to replace traditional discriminator, which can obtain a more realistic probability of real images compared to false images, thereby promoting the reconstructed images via the generator to have more detailed textures. Then, the joint loss function is used to merge the high-frequency information and low-frequency information, which improves the visual effect of the reconstructed images to some extent. Finally, a large number of quantitative and qualitative experiments verify the effectiveness of the proposed method, and the necessity of the proposed method is verified by the ablation study. In future work, considering the complexity of the network, we will focus on optimizing the computational complexity of the network and trying to build a high-performance lightweight network.

Author Contributions

Conceptualization, H.Q. and H.Y.; methodology, H.Q. and H.Y.; software, H.Q.; validation, Y.S. and J.L.; formal analysis, H.Q.; investigation, H.Q.; resources, H.Y.; data curation, H.Q.; writing—original draft preparation, H.Q.; writing—review and editing, H.Q. and Y.S.; visualization, H.Q.; supervision, J.L.; project administration, H.Y.; funding acquisition, H.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation for youth scientists of China (No. 62203201), the Foundation Research Project of the Educational Department of Liaoning Province (No. LJKZZ20220085), and the Cooperation Innovation Plan of Yingkou for Enterprise and Doctor (No. 2022-13).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Our training set DIV2K datasets can be obtained from available online: https://data.vision.ee.ethz.ch/cvl/DIV2K/ (accessed on 21 November 2022). Set5, Set14, BSD100, Urban100 can be obtained from: https://arxiv.org/abs/1909.11856/ (accessed on 21 November 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhu, S.; Zeng, B.; Liu, G. Image interpolation based on non-local geometric similarities. In Proceedings of the IEEE International Conference on Multimedia and Expo (ICME), Turin, Italy, 29 June–3 July 2015. [Google Scholar]
  2. Papyan, V.; Elad, M. Multi-Scale Patch-Based Image Restoration. IEEE Trans. Image Process. 2016, 25, 249–261. [Google Scholar] [CrossRef] [PubMed]
  3. Dong, C.; Loy, C.; Tang, X. Accelerating the Super-Resolution Convolutional Neural Network. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Springer: Cham, Switzerland, 2016; pp. 391–407. [Google Scholar]
  4. Dong, C.; Loy, C.; He, K. Image super resolution using deep convolutional networks. IEEE Trans. Pattern Anal. Mach. 2016, 38, 295–307. [Google Scholar] [CrossRef] [PubMed]
  5. Kim, J.; Lee, J.; Lee, K. Deeply-Recursive Convolutional Network for Image Super-Resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  6. Jiang, Z.; Pi, K.; Huang, Y.; Qian, Y.; Zhang, S. Difference Value Network for Image Super-Resolution. IEEE Signal Process. Lett. 2021, 28, 1070–1074. [Google Scholar] [CrossRef]
  7. Lim, B.; Son, S.; Kim, H. Enhanced Deep Residual Networks for Single Image Super-Resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  8. Zhang, Y.; Li, K.; Li, K.; Wang, L.; Zhong, B.; Fu, Y. Image Super-Resolution Using Very Deep Residual Channel Attention Networks. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
  9. Hu, X.; Mu, H.; Zhang, X.; Wang, Z.; Tan, T.; Sun, J. Meta-SR: A Magnification-Arbitrary Network for Super-Resolution. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  10. Li, Z.; Yang, J.; Liu, Z.; Yang, X.; Wu, W. Feedback Network for Image Super-Resolution. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  11. Salimans, T.; Goodfellow, I.; Zaremba, W. Improved techniques for training gans. In Proceedings of the Annual Conference on Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016. [Google Scholar]
  12. Huang, H.; Yu, S.; Wang, C. An introduction to image synthesis with generative adversarial nets. arXiv 2018, arXiv:1803.04469. [Google Scholar]
  13. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M. Generative adversarial networks. Commun. ACM 2020, 63, 139–144. [Google Scholar] [CrossRef]
  14. Creswell, A.; White, T.; Dumoulin, V. Generative Adversarial Networks: An Overview. IEEE Signal Process. Mag. 2018, 35, 53–65. [Google Scholar] [CrossRef]
  15. Ledig, C.; Theis, L.; Huszár, F. Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  16. Bulat, A.; Yang, J.; Tzimiropoulos, G. To learn image super-resolution, use a GAN to learn how to do image degradation first. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018. [Google Scholar]
  17. Wang, X.; Yu, K.; Wu, S. Esrgan: Enhanced Super-Resolution Generative Adversarial Networks. In Proceedings of the European Conference on Computer Vision (ECCV) Workshops, Munich, Germany, 8–14 September 2018. [Google Scholar]
  18. Zhang, X.; Chen, Q.; Ng, R. Zoom to Learn, Learn to Zoom. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  19. Lugmayr, A.; Danelljan, M.; Gool, V. Srflow: Learning the super-resolution space with normalizing flow. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Cham, Switzerland, 2020. [Google Scholar]
  20. Rad, M.S.; Bozorgtabar, B.; Marti, U.V.; Basler, M.; Ekenel, H.K.; Thiran, J.P. SROBB: Targeted Perceptual Loss for Single Image 299 Super-Resolution. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
  21. Li, W.; Zhou, K.; Qi, L.; Lu, L.; Jiang, N.; Lu, J.; Jia, J. Best-Buddy GANs for Highly Detailed Image Super-Resolution. In Proceedings of the 36th AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 22 February–1 March 2022. [Google Scholar]
  22. Ma, C.; Rao, Y.; Cheng, Y. Structure-Preserving Super Resolution with Gradient Guidance. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  23. Zhang, Y.; Tian, Y.; Kong, Y. Residual Dense Network for Image Super-Resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  24. Zeiler, D.; Fergus, R. Visualizing and Understanding Convolutional Networks. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; Springer: Cham, Switzerland, 2014. [Google Scholar]
  25. Agustsson, E.; Timofte, R. Ntire 2017 Challenge on Single Image Super-Resolution: Dataset and Study. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  26. Bevilacqua, M.; Roumy, A.; Guillemot, C. Neighbor embedding based single-image super-resolution using semi-nonnegative matrix factorization. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Kyoto, Japan, 25–30 March 2012. [Google Scholar]
  27. Yuan, Y.; Liu, S.; Zhang, J. Unsupervised Image Super-Resolution Using Cycle-in-Cycle Generative Adversarial Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  28. Martin, D.; Fowlkes, C.; Tal, D. A Database of Human Segmented Natural Images and Its Application to Evaluating Segmentation Algorithms and Measuring Ecological Statistics. In Proceedings of the Eighth IEEE International Conference on Computer Vision, Vancouver, BC, Canada, 7–14 July 2001. [Google Scholar]
  29. Huang, B.; Singh, A.; Ahuja, N. Single Image Super-Resolution from Transformed Self-Exemplars. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  30. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image Quality Assessment: From Error Visibility to Structural Similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  31. Zhang, R.; Isola, P.; Efros, A. The Unreasonable Effectiveness of Deep Features as a Perceptual Metric. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
Figure 1. Network model of DVDR-SRGAN generator.
Figure 1. Network model of DVDR-SRGAN generator.
Sensors 23 04854 g001
Figure 2. dvdb network structure diagram.
Figure 2. dvdb network structure diagram.
Sensors 23 04854 g002
Figure 3. The difference between standard discriminator and relative discriminator.
Figure 3. The difference between standard discriminator and relative discriminator.
Sensors 23 04854 g003
Figure 4. Relative discriminator network model.
Figure 4. Relative discriminator network model.
Sensors 23 04854 g004
Figure 5. Algorithms 4× reconstruction results. Image “bird” from Set5.
Figure 5. Algorithms 4× reconstruction results. Image “bird” from Set5.
Sensors 23 04854 g005
Figure 6. Algorithms 4× reconstruction results. Image “butterfly” from Set5.
Figure 6. Algorithms 4× reconstruction results. Image “butterfly” from Set5.
Sensors 23 04854 g006
Figure 7. Algorithms 4× reconstruction results. Image “face” from Set14.
Figure 7. Algorithms 4× reconstruction results. Image “face” from Set14.
Sensors 23 04854 g007
Figure 8. Algorithms 4× reconstruction results. Image “man” from Set14.
Figure 8. Algorithms 4× reconstruction results. Image “man” from Set14.
Sensors 23 04854 g008
Figure 9. Algorithms 4× reconstruction results. Image “296059” from BSDS100.
Figure 9. Algorithms 4× reconstruction results. Image “296059” from BSDS100.
Sensors 23 04854 g009
Figure 10. Algorithms 4× reconstruction results. Image “026” from Urban100.
Figure 10. Algorithms 4× reconstruction results. Image “026” from Urban100.
Sensors 23 04854 g010
Figure 11. Visual evaluation of different loss functions.
Figure 11. Visual evaluation of different loss functions.
Sensors 23 04854 g011
Table 1. Comparison of values of α and β . The best performance is highlighted in red (1st best).
Table 1. Comparison of values of α and β . The best performance is highlighted in red (1st best).
ParameterSet5Set14
αβPSNRSSIMLPIPSPSNRSSIMLPIPS
0.10.928.400.7960.06124.690.6020.107
0.20.828.520.8250.05224.810.7020.094
0.30.727.930.7670.06824.560.6530.110
0.40.627.850.7590.07324.530.6210.121
Table 2. Comparison of PSNR, SSIM, and LPIPS values of 4× reconstruction results of each algorithm. The best performance is highlighted in red (1st best) and blue (2nd best).
Table 2. Comparison of PSNR, SSIM, and LPIPS values of 4× reconstruction results of each algorithm. The best performance is highlighted in red (1st best) and blue (2nd best).
DatasetMetricBicubicSRGANESRGANBeby-GANSPSRDVDR-SRGAN
Set5PSNR26.6626.9127.3527.8228.4428.52
SSIM0.7900.8040.8060.8010.8240.825
LPIPS0.3640.1310.1080.1180.0870.064
Set14PSNR24.2323.8723.6124.6924.7524.81
SSIM0.6850.6770.6500.7010.6960.702
LPIPS0.3870.1420.1250.1090.1060.098
BSD100PSNR22.6522.6723.3324.1324.2124.21
SSIM0.6010.6360.6130.6350.6550.667
LPIPS0.4450.1630.1430.1190.1190.118
Urban100PSNR21.7021.7721.8222.7523.2422.91
SSIM0.6520.6770.6790.6950.7370.706
LPIPS0.4350.1530.1350.1090.1060.108
Table 3. Comparison of methods under different network modules.
Table 3. Comparison of methods under different network modules.
MethodSet14BSD100Urban100
LPIPSPSNRSSIMLPIPSPSNRSSIMLPIPSPSNRSSIM
ESRGAN0.13823.610.6500.14723.330.6130.17821.820.679
DVDB no JPL0.12824.250.6480.13823.590.6010.17121.990.671
DVDB no dvdb0.13124.520.6570.14224.040.6230.17622.290.681
DVDR-SRGAN0.12224.810.6680.13224.210.6270.16822.910.706
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Qu, H.; Yi, H.; Shi, Y.; Lan, J. DVDR-SRGAN: Differential Value Dense Residual Super-Resolution Generative Adversarial Network. Sensors 2023, 23, 4854. https://doi.org/10.3390/s23104854

AMA Style

Qu H, Yi H, Shi Y, Lan J. DVDR-SRGAN: Differential Value Dense Residual Super-Resolution Generative Adversarial Network. Sensors. 2023; 23(10):4854. https://doi.org/10.3390/s23104854

Chicago/Turabian Style

Qu, Hang, Huawei Yi, Yanlan Shi, and Jie Lan. 2023. "DVDR-SRGAN: Differential Value Dense Residual Super-Resolution Generative Adversarial Network" Sensors 23, no. 10: 4854. https://doi.org/10.3390/s23104854

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop