Next Article in Journal
A Non-Invasive Technique for Long-Term Monitoring of Gastroesophageal Reflux—A Pilot Study
Next Article in Special Issue
A Lightweight Model for Real-Time Detection of Vehicle Black Smoke
Previous Article in Journal
Contactless Heart and Respiration Rates Estimation and Classification of Driver Physiological States Using CW Radar and Temporal Neural Networks
Previous Article in Special Issue
Automatic Detection Method for Black Smoke Vehicles Considering Motion Shadows
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Multi-Scale Recursive Attention Feature Fusion Network for Image Super-Resolution Reconstruction Algorithm

The Key Laboratory of Industrial Automation of Shaanxi Province, Shaanxi University of Technology, Hanzhong 723000, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(23), 9458; https://doi.org/10.3390/s23239458
Submission received: 31 October 2023 / Revised: 21 November 2023 / Accepted: 26 November 2023 / Published: 28 November 2023
(This article belongs to the Special Issue Computer Vision Sensing and Pattern Recognition)

Abstract

:
In recent years, deep convolutional neural networks (CNNs) have made significant progress in single-image super-resolution (SISR) tasks. Despite their good performance, the single-image super-resolution task remains a challenging one due to problems with underutilization of feature information and loss of feature details. In this paper, a multi-scale recursive attention feature fusion network (MSRAFFN) is proposed for this purpose. The network consists of three parts: a shallow feature extraction module, a multi-scale recursive attention feature fusion module, and a reconstruction module. The shallow features of the image are first extracted by the shallow feature extraction module. Then, the feature information at different scales is extracted by the multi-scale recursive attention feature fusion network block (MSRAFFB) to enhance the channel features of the network through the attention mechanism and fully fuse the feature information at different scales in order to improve the network’s performance. In addition, the image features at different levels are integrated through cross-layer connections using residual connections. Finally, in the reconstruction module, the upsampling capability of the deconvolution module is used to enlarge the image while extracting its high-frequency information in order to obtain a sharper high-resolution image and achieve a better visual effect. Through extensive experiments on a benchmark dataset, the proposed network model is shown to have better performance than other models in terms of both subjective visual effects and objective evaluation metrics.

1. Introduction

Image super-resolution reconstruction (SR) is an important image processing technique that has far-reaching significance in the field of computer vision. Super-resolution techniques enhance low-resolution (LR) images by transforming them into high-resolution (HR) counterparts [1,2,3], thereby improving the quality of the images. Image super-resolution techniques are widely used in medical imaging [4,5], video security imaging [6], satellite remote sensing [7,8], image spectroscopy [9,10], and other fields with important applications. However, image super-resolution reconstruction is an ill-posed problem, as a low-resolution image can be obtained from multiple HR images by various methods; solving this problem requires high performance on the part of the algorithm [11]. To solve this problem, researchers have proposed various super-resolution reconstruction methods. There are currently three main types of super-resolution reconstruction methods: interpolation-based methods [12], reconstruction-based methods [13], and deep learning-based methods [14,15]. Interpolation-based approaches upsample the LR image through a simple interpolation technique; however, they cannot obtain the real-time detail information of the image. Reconstruction-based methods reconstruct LR images by means of image modeling or optimization problems, however, they consume a large amount of computing power. The learning-based approach uses deep learning networks to learn and predict HR images from a large number of HR/LR image pairs with good reconstruction and fast speed, and is currently the research direction with the most potential.
Recently, deep learning-based super-resolution reconstruction algorithms have become a hot research topic in academia. Dong et al. [16] first proposed a CNN-based SISR method called SRCNN, which is an end-to-end three-layer neural network. However, this simple network structure does not handle information-rich images well. Therefore, many researchers have started to propose deeper and wider neural networks to improve network performance [17,18,19,20,21,22,23,24,25,26,27,28,29]. For example, RCAN [17] is a deep network with a depth of more than 400 layers. Although it shows superior performance, the significant increase in network depth gives rise to problems such as increased network complexity and computation cost, which increase the performance requirements of hardware devices.
Researchers have been able to reduce the network parameters while deepening the network structure to achieve better performance. For example, IDN [18] has been proposed as a fast and accurate SISR network with information distillation. DSRN [19] uses a dual-state recurrent network, which has advantages in terms of performance. Recursive networks are an effective network shrinkage model that can effectively extract and fuse features in images to obtain more accurate and comprehensive feature representation while improving the image reconstruction capability of the network. However, the above-mentioned SISR methods seem to lack flexibility in dealing with spatial contexts of different scales, and cannot effectively extract and use local and global features, making it necessary to capture both local and global contextual information to make full use of the extracted feature information.
In order to deal with these problems, we propose a multi-scale recursive attention feature fusion network (MSRAFFN) with the aim of improving the performance of image super-resolution reconstruction through effective use of multi-scale features. The proposed network uses a multi-scale recursive attention feature fusion block (MSRAFFB) consisting of a multi-scale feature extraction unit (MSFE) and an attention feature fusion unit (AFF), with the MSFE unit consisting of multiple branches for extracting features at different scales. The AFF unit is used to adaptively fuse multi-scale features. Meanwhile, residual connections are used in the backbone network to integrate the layered features from the multi-scale feature extraction module with the deeper features through cross-layer connections, thereby preserving more detailed information of the image. In summary, MSRAFFN can better recover details in HR images by extracting more contextual information from LR images, resulting in improved image reconstruction quality. The contributions of this paper are as follows:
(1)
An MSRAFFN is proposed that utilizes a multi-scale recursive network structure and attention mechanism to govern network information exchange and effectively handle feature information at different scales.
(2)
An MSFE block is proposed to extract feature information from different levels in the form of parallel branch stacked connections, then fuse features between each level layer-by-layer to obtain a reconstructed image with richer texture information.
(3)
The AFF block is designed to integrate feature information from each branch within the MSRAFFB and learn the importance of the weights of the features of different branches using the attention mechanism.
(4)
The proposed network directly extracts feature information from the LR image without interpolation, uses local residuals to connect neighboring layers within the MSRAFFB to facilitate information transfer within a single block, and utilizes recursion and global residuals to facilitate information exchange between blocks.
The rest of this work is organized as follows: Section 2 briefly describes existing network structures related to the proposed structure; Section 3 describes the proposed method in detail; Section 4 provides the results and analysis of different experiments; finally, Section 5 presents our conclusions and recommendations for future work.

2. Related Work

2.1. CNN-Based Super-Resolution of Single Images

SRCNN is the first CNN-based super-resolution method with a simple network model; however, its use of interpolation operations may affect the performance of the network. FSRCNN [20] uses a deconvolution layer on top of SRCNN to enlarge the image and improve its resolution. VDSR [21] improves performance by stacking deeper convolutional layers, although the resulting increase in the number of network layers leads to greater computational complexity. ESPCN [22] contains a sub-pixel convolutional layer at the end of the network to handle the image super-resolution reconstruction problem with different scale factors, and additionally offers faster processing speeds.
LapSRN [23] improves reconstruction quality through the residual pyramid frame, although it requires transformation of the pyramid frame into multiple parts for processing, which increases computational complexity. In contrast, CMSC [24] can improve both computational efficiency and reconstruction quality; however, it is not effective for image edge detail reconstruction. To further improve the reconstruction accuracy of edge details, MSRN [25] uses a multi-scale residual structure and a densely connected layer to reconstruct the edge details of images more accurately. However, because MSRN requires more computational complexity, a trade-off between computational efficiency and reconstruction accuracy is needed when selecting a suitable model in practical applications.
Among the common recurrent networks, DRCN [26] and DRRN [27] are suitable for input images of different sizes, the former enhancing the feature representation by recursive connectivity and upsampling and the latter avoiding loss of feature information by introducing residual and cyclic structures. DSRN [19] and DRUDN [28] can both process input images of different sizes, with the former predicting high-resolution images through Laplace pyramids and the latter improving the sharpness and quality of images through recursive upsampling and denoising techniques. However, recursive connections in DRCN leads to an increase in computation, and the residual structure in DRRN can only handle local features. Additionally, the Laplace pyramid in DSRN requires a large amount of computation and storage space, while the denoising technique in DRUDN may eliminate some useful detailed information.

2.2. Multi-Scale Networks

Multi-scale networks can achieve multi-scale feature representation of images by processing blocks of images at different scales, which improves the accuracy and efficiency of the network [30]. In super-resolution tasks, the loss of detail information in low-resolution images is severe, limiting the quality of the reconstructed images [31]. To capture the features of low-resolution images more comprehensively and accurately, a multi-scale network can be used to extract features at different scales and reconstruct images with rich details and clear textures. Figure 1 shows the simplified structure of the LapSRN multi-scale network module.
LapSRN incorporates a Laplace pyramid into the CNN structure, with low-resolution images being input directly into the network and clear HR images reconstructed via gradual upsampling. The structure of the LapSRN can be thought of as having multiple stages, with each stage accomplishing a two-fold upsampling operation in which features are first extracted through a number of convolutional layers and then the sizes of the extracted features undergo twofold upsampling through the deconvolutional layers. In each stage, the features are first extracted using a number of convolutional layers. In LapSRN, subnetworks of different scales share parameters, allowing the number of parameters of the model to be greatly reduced and the training speed of the model to be improved. However, the model is relatively complex, and this structure can only accomplish the task of super-resolution reconstruction to the power of two. Despite the shortcomings of LapSRN, the multiscale idea allows the model to process information at different scales simultaneously, which helps to improve performance on image processing tasks.

2.3. Recursive-Based SISR Method

Recursive networks are widely used in SISR in order to reduce the number of parameters in SISR methods. This section focuses on the three recurrent networks most relevant to the proposed method: DRRN [27], DSRN [19], and DRUDN [28]. Figure 2 shows the simplified structures of these recurrent networks.
DRRN: In Figure 2a, the green dashed boxes indicate the number of residual blocks in the recursive network, while the black boxes indicate the number of recursive networks. The network learns through the design idea of deep residual networks by stacking multiple residual blocks together to form a deep network structure. It incorporates both local and global residual learning to construct a deep network, thereby reducing the burden of image information in the deep network structure and making the network easier to train. In addition, recursive learning is introduced in the recursive block to keep the model compact and reduce the number of parameters. DRRN uses local normalization and global normalization to enhance the generalization ability of the network.
DSRN: In Figure 2b, the green dashed boxes indicate the number of residual units and the black boxes indicate the number of recursive networks. DSRN is different from many other networks in that the LR images are input directly into the network instead of interpolating images, thereby reducing its computational complexity. Another difference between DSRN and DRRN is that DSRN uses the output of the previous recursive unit as the local residual instead of the global input, ignoring the hierarchical feature information.
DRUDN: In Figure 2c, the black boxes in the figure represent recursive upsampling and downsampling blocks. DRUDN employs a strategy of feeding low-resolution (LR) images directly into the network and extends the sensory field of the entire network using convolutional and deconvolutional layers, which effectively improves the performance of super-resolution reconstruction. The core of this method lies in the fact that by introducing convolution and deconvolution operations, the network is able to capture image features more comprehensively, which in turn enhances its ability to learn details. Th use of an attention mechanism to process the important information in the image has been proposed.

2.4. Attention Mechanism

Attention mechanisms have attracted the attention of many scholars because they allow models to focus more flexibly on certain features or regions when processing input data, improving the performance and capability of the resulting models [32,33,34,35]. First, Hu et al. [32] proposed the SE attention mechanism to achieve significant performance gains in a variety of tasks, greatly facilitating the application of attention mechanism in computer vision tasks. RCAN integrated channel attention (CA) into the residual block, which pays more attention to the high-frequency channel features to improve the reconstruction performance. Wang et al. [33] proposed a combination of spatial attention and channel attention in the Convolutional Block Attention Module (CBAM), which captures more informative features at all locations by learning channel and spatial correlation features at each layer. The Holistic Attention Network (HAN) [34] introduced a Layer Attention Module (LAM) to learn the correlation between features in multi-scale layers. Attention mechanisms in CNNs are able to redistribute the network’s attention to different parts or features in order to enhance the processing of the information that is most of interest while improving network performance.
Squeeze-and-Excitation Networks (SEnet), which aim to focus on the relationship between different channels, automatically learn the importance of different channel features. They selectively enhance those features containing useful information and suppress useless features based on the global information. The basic structure of SEnet is shown in Figure 3, in which a feature map X with channel number C , height H , and width W is input and a feature map U with channel number C, height H, and width W is obtained by one or more standard convolution operations.
The steps of SEnet are as follows. First, the squeeze operation uses global average pooling to encode the entire spatial feature on a channel as a global feature, i.e., the input of H × W × C is converted to an output of 1 × 1 × C . The excitation operation is a nonlinear transformation of the result after the squeeze operation using two fully connected neural networks to obtain the magnitude of the importance of different channels. Finally, the scale operation, i.e., the weighting operation, weights the output weights after the excitation operation to the previous features by multiplying them channel-by-channel to complete the rescaling of the original features in the channel dimension and obtain the feature map X ˜ with dimensions H × W × C . Using the SEnet model, the overall channel features are enhanced and the network performance is improved.

3. Proposed Method

3.1. Overview of the Network Model

MSRAFFN contains three parts: a shallow feature extraction module, a multi-scale recursive attention feature fusion module, and a reconstruction module. In this structure, the shallow feature extraction unit extracts global features through two convolutional layers, which lays the foundation for subsequent deep feature extraction. MSRAFFB is used to extract the deep features, which contains multiple branches; each branch outputs features with different scales and applies SEnet to learn deep feature information and then fuse the feature information of different scales for improved network performance. The residual connection method in the backbone network can integrate shallow features and deep features via cross-layer connection, allowing the layered features from the multi-scale feature extraction module to be effectively used to extract richer and more accurate feature information from low-resolution images while retaining more detailed information for better reconstruction of high-resolution images. Finally, the reconstruction module generates residual images using general convolution and deconvolution operations to further improve the quality of image reconstruction. The structure of MSRAFFN is shown in Figure 4.
Table 1 shows the detailed architecture of the network, illustrating the number of convolutional layers used in the CNN network. The input low-resolution image is first processed by two 3 × 3 convolutions, then we several MSRAFFBs with the same structure are used to form the backbone of the network and fuse the processed features of the front-end network using a 1 × 1 convolution after each MSRAFFB. The MSFE contains several branches internally, with the first one containing only one 3 × 3 convolutional layer and the others containing one 3 × 3 convolutional layer and a 1 × 1 convolutional layer each. The AFF unit includes two ordinary convolutional layers, one pooling layer, and two fully connection layers. The reconstruction module includes two ordinary convolutional layers and one deconvolution layer.

3.2. Shallow Feature Extraction Module

In order to speed up the training of super-resolution models, mainstream shallow feature extraction methods adopt the strategy of directly inputting small-sized low-resolution images into the shallow feature extraction module instead of first amplifying the images by interpolation before inputting them into the network. This strategy reduces the computational complexity of the network and accelerates the training speed. Inspired by this idea, our method uses a similar approach to extract shallow image features; specifically, the shallow feature extraction module in the model uses two 3 × 3 convolutional layers (the padding and stride are both 1) to extract the shallow features of the image. This simple feature extraction method can effectively reduce the training time of the super-resolution model, retain all the details and complexity in the original image as much as possible, and make the generated high-resolution image more similar to the ground truth image. The input LR image is denoted as I L R , while the extracted shallow features are denoted as F 0 :
F 0 = H F E P ( I L R )
where H F E P (•) denotes the shallow feature extraction part.

3.3. Multi-Scale Recursive Attention Feature Fusion Module

The MSRAFFB is an important module of the designed algorithm to extract deep feature information. As shown in Figure 5, MSRAFFB is composed of two parts, MSFE and AFF. The AFF unit takes the output of all branches as input, adaptively fuses multi-scale features, and learns deep feature information through SEnet to improve network performance. In the backbone network, the shallow features are integrated with the deep features by means of cross-layer connections using residual connections and the hierarchical features of MSRAFFB are used effectively in a layer-wise manner, meaning that we have
F D F = H M R A ( F 0 ) .
where F D F denotes the deep feature information and H M R A (•) denotes the multi-scale recursive attention feature fusion part.
The multi-scale feature extraction unit mainly consists of 3 × 3 (where the padding and stride are both 1) and 1 × 1 (padding 0, stride 1) convolutional layers, and contains numerous parallel branches inside which connect the input of MSRAFFB and the output of the previous branch as the input of the next branch through the concatenation operator. The output of each branch has a different perceptual field, as follows:
F b 1 1 = f 3 × 3 ( F b 1 )
F b 1 2 = f 3 × 3 ( f 1 × 1 ( c o n c a t ( F b 1 , F b 1 c ) ) )
F b 1 c 1 = f 3 × 3 ( f 1 × 1 ( c o n c a t ( F b 1 , F b 1 c 2 ) ) )
F b 1 c = f 3 × 3 ( f 1 × 1 ( c o n c a t ( F b 1 , F b 1 c 1 ) ) ) .
where F b 1 1 represents the output of the bth MSRFFB block. f 3 × 3 is the 3 × 3 convolution, f 1 × 1 is the 1 × 1 convolution, c o n c a t (•) is the concatenation operator, and F b 1 c represents the output of the c-th branch of the b-th MSRFFB block. The attention feature fusion unit connects the different scale features in each branch by the connection operator, then enhances the overall channel features using SEnet to improve the network performance; thus, the output of this unit is
F b * = f 3 × 3 ( f S E ( f 1 × 1 ( c o n c a t ( F b 1 1 , F b 1 2 , , F b 1 c ) ) ) ) .
In addition, a residual connection is used to connect the initial input features to the output of the attention feature fusion unit, making the final output of this module
F b = F b * + F b 1 .

3.4. Reconstruction Module

The reconstruction module is shown in Figure 4; it contains two 3 × 3 convolutions and one deconvolution. The final reconstructed image is obtained by using the residual module to concatenate the image after double triple interpolation and fuse it with the convolution output. The first ordinary convolutional layer with 64 convolutional kernels is used for mapping the feature map to a super-resolution image. Deconvolution is used to upsample the image by scaling the low-resolution image to a high-resolution level. The last ordinary convolutional layer is used to recover the number of channels in order to generate a three-channel RGB image. The deconvolution layer parameters are set as follows: upscale factors of × 2 , stride of 2, padding of 2, and convolution kernel size of 6. For upscale factors of × 3 , the stride is 3, the padding is 2, and the convolution kernel size is 7. For upscale factors of × 4 , the stride is 4, the padding is 2, and the convolution kernel size is 8. When the scaling factor is 2, the deconvolution model is as shown in Figure 6, which first expands the input image to twice the original size and sets all the new pixel values to 0. Then, the deconvolution operation is performed using the convolution (padding 0, stride 1) to upsample the low-resolution image in order to obtain a higher-resolution feature map.

3.5. Loss Function

The L 2 loss function may over-penalize the squared value of the error, resulting in the loss of high-frequency detail information in the image. On the other hand, the L 1 loss function can effectively penalize the absolute error between the predicted and true values, causing the network to be more concerned with the recovery of details in the image. Therefore, the L 1 loss function is chosen for use in the design of the super-resolution network. If there are N training samples, the L 1 loss function can be expressed as follows:
L L o s s = 1 N i = 1 N F ( X i ) Y i
where Y i denotes the original high-resolution image, F ( x i ) denotes the reconstructed super-resolution image, F ( x i ) Y i 1 denotes the loss of the i-th original image and the reconstructed image, and the loss is averaged over all images.

4. Experimental Results and Analysis

The experimental system was an Ubuntu 20.04 server with an NVIDIA GP102 GPU; the model was implemented using the PyCharm compilation platform in Python language. In order to objectively evaluate the performance of the proposed method, we used two metrics that are recognized as authoritative in the field of image processing, namely, the peak signal-to-noise ratio (PSNR) [36] and the structural similarity (SSIM) [37]. These objective evaluation metrics allow for a more objective performance assessment of the proposed method.

4.1. Experiment Details

The DIV2K [38] dataset, which contains 800 high-quality images, was used for training in the experiments. During the training phase, randomly cropped 48 × 48 LR image patches were used as input to the network model and the patch size of HR images was determined based on the scale factor. The patches were randomly added using horizontal, vertical, flipped, and 90° rotated data enhancement techniques to expand the diversity and number of samples in the dataset. In addition, the Adam optimizer [39] was used with an initial learning rate set to 0.0002 and a 50% reduction every 200 epoch. In the network structure, the number of output channels of the last convolutional layer was set to 3 in order to adapt to the reconstruction of RGB images. The number of output channels for the rest of the 3 × 3 convolutional layers was set to 64, with the PReLU activation function, padding of 0, and stride of 1. For the proposed multi-scale network model, we used five benchmark datasets in the performance evaluation: Set5 [40], Set14 [41], B100 [42], Urban100 [43], and Manga109 [44]. The Set5, Set14, and BSD100 datasets consist of natural scenes, the Urban100 dataset contains urban scenes such as roads and buildings, and the Manga109 dataset contains many manga pictures. These datasets are widely used in the field of image super-resolution, and allow for comprehensive performance testing of the proposed algorithms in different scenarios.

4.2. Network Model Parameters Analysis

In this section, we explore the effect on network performance of the number of MSRAFFB modules B in MSRAFFN and the number of branches C in MSFE. Experiments were performed on the DIV2K dataset, setting the scaling factor to × 4 and the number of iterations to 100. the values of B were 4, 6, and 8, while the values of C were 6, 8, and 10. To test the effect of the C parameter, the value of B was fixed and the value of C was set to 6, 8, and 10. The experimental results are shown in Figure 7.
From Table 2, it can be seen that when B = 4, 6, and 8, the PSNR on all five datasets increases and then decreases with the increase of C. This indicates that increasing the number of branches in the module does not always ensure that the performance of the network is improved; based on this comparison, C should first be set to 8. In order to explore the effect of the value of B on network performance, C was set to 8 and B was set to 4, 6, and 8. The performance of the model did not always improve with higher values; it can be seen that when the value of B is increased from 8 to 10, both the PSNR and SSIM metrics decrease. These experimental results show that stacking a greater number of modules does not always lead to improved network performance. Thus, the value of B was set to 6.
From Figure 7a, it can be seen that when B = 4, the lowest LOSS value is obtained when C = 8, and that the largest PSNR and SSIM values are obtained in this case. In Figure 7, when B = 6, the PSNR value is the lowest when C is taken as 6. When C is taken as 8 or 10, the difference between the two in terms of performance is not obvious. The network performance when B = 8 is shown in Figure 7c. It can be seen that the LOSS value is the lowest at C = 10. When the number of iterations increases to a certain level, the difference between the PSNR and SSIM values at C = 6 and C = 10 becomes smaller.
While the network performance of B4C8, B6C8, B6C10, B8C6, and B8C10 may be better when the value of B is fixed, considering the overall network complexity, B4C8, B6C8, and B8C6 were finally selected.
As shown in Figure 8, by comparing the corresponding LOSS values of B4C8, B6C8 and B8C6, it is apparent that B6C8 and B8C6 have the lowest LOSS values and that their PSNR and SSIM values are higher. Compared to B4C8, B6C8 and B8C6 both have better network performance. Based on the analysis of Table 2 and Figure 7 and Figure 8, which show the trade-off relationship between network performance and model size, B = 6 and C = 8 were finally selected.

4.3. Comparison with State-of-the-Art Methods

To enhance the comprehensiveness of this study, Table 3 shows the main differences between our proposed algorithm and typical SR algorithms in terms of upsampling methods, including different recursive networks, residual networks, and loss functions. The upsampling methods are categorized into three types: bicubic interpolation, deconvolution, and subpixel convolution, while the loss functions can be categorized into L 1 , L 2 , MSE, and char loss functions. Our proposed algorithm uses deconvolution to enlarge the target image, utilizes the ideas of recursive and residual networks to improve the network’s ability to extract feature information, and uses the L 1 loss function to optimize the network model.
To test the performance advantages and disadvantages of the MSRAFFA model, the super-resolution reconstruction model proposed in this paper was compared with existing super-resolution reconstruction models on five benchmark test sets. The tested reconstruction models were SRCNN, FSRCNN, VDSR, DRCN, LapSRN, DRRN, DSRN, IDN, and MSRN.
PSNR and SSIM are authoritative evaluation metrics that can be used to objectively assess image quality. These two metrics were compared at scale factors of ×2, ×3, and ×4. For the IDN method, this experiment was conducted using only the metrics in its original text for comparison, and the network model was not evaluated on the Manga109 dataset. The results are shown in Table 4, where the best performance is indicated in bold and the second-best by underlining. It can be seen in the table that with ×2 SR, the performance of our proposed method is not much different from that of MSRN compared to the previous methods. With ×3 SR, the PSNR values of our proposed method are improved by 0.19 dB and 0.37 dB compared to those of MSRN on the Urban100 and Manga109 datasets, respectively. With ×4 SR, the PSNR values of our proposed method are improved by 0.19 dB and 0.79 dB compared to those of MSRN on the Urban100 and Manga109 test sets, respectively. Finally, the PSNR values of our proposed method are improved by 0.19 dB and 0.79 dB on the Urban100 and Manga109 datasets, respectively, compared to the PSNR values of MSRN. The results of these experiments prove the superior performance of our proposed method.
In order to show the difference in subjective visual effects between the proposed method and existing methods, we chose a scaling factor of 4 and selected a representative image in the Set14, BSD100, Manga109, and UrBan100 datasets for a local magnification comparison. The visual comparison of image effects is shown in Figure 9. A comparison of the four reconstructed images with local enlarged parts shows that the reconstructed images from Bicubic, SRCNN, and FSRCNN are blurred and do not reconstruct the clear detailed texture contours of the images. The VDSR, LapSRN, and IDN methods are more or less able to reconstruct the texture contours of images. Compared with the above methods, our proposed model can reconstruct more texture contour information, and the reconstructed images are clearer and sharper with better visual perception.
As can be seen from the example of Set14: zebra, our proposed model is able to reproduce the textured parts of the animal’s body more clearly. From the example of Urban100: img016, although our proposed model does not completely reconstruct all the details of the building outline, the outline of the reconstructed building is clearer and more textured.

4.4. Ablation Study

To test the effectiveness of MSFE and AFF in improving the performance of the model, we conducted an ablation experiment. Performance was examined using a common convolution instead of the corresponding module. To eliminate any performance gain caused by different parameters, it was ensured that all networks contained the same number of MSRAFFBs and used same training strategy. The models for the different ablation studies were as follows:
(1)
Normal convolution replacing the corresponding MSFE module, denoted as Non MSFE.
(2)
A 3 × 3 standard convolution replacing the corresponding AFF module, denoted as Non-AFF.
(3)
Both the MSFE and AFF modules are present, denoted as MSFE + AFF.
The training results after 500 iterations are shown in Table 5 and Figure 10.
It can be visually observed in Figure 10 that the model performance degrades very significantly without the multi-scale feature extraction module, with lower PSNR and SSIM values and larger LOSS values. The attention feature extraction module has a similar impact. As can be seen from Table 5, the network with the multi-scale feature extraction module obtains better results compared to the reconstruction with the ordinary convolution kernel. The average PSNR values on the test set are improved by 0.31 dB, 0.26 dB, 0.18 dB, 0.51 dB, and 1.06 dB on Set5, Set14, B100, Urban100m and Manga109, respectively. With the AFF module, the average PSNR values on the Set5, Set14, B100, Urban100, and Manga109 test sets improve by 0.03 dB, 0.03 dB, 0.02 dB, 0.07 dB, and 0.41 dB, respectively, compared with normal convolution.
In Figure 11, it can be seen that the eyes of the bird reconstructed by MSFE+AFF are clearer and have more detail. The results of the ablation experiments show that deeper feature information can be extracted using MSFE and AFF, and that this rich feature information can be fully utilized to improve the quality of the reconstructed images.

4.5. Model Performance Analysis

While our method shows promising results regarding image quality metrics, it is essential to consider its computational efficiency as well. To further demonstrate the excellence of the MSRAFFN algorithm, we compared the number of parameters of each network using the PSNR evaluation metrics; the selected dataset was the Set14 dataset with a scaling factor of 4. The results are shown in Figure 12. It is evident from the figure that the MSRAFFN model has a larger number of parameters compared to algorithms such as DRRN and VDSR; however, it achieves significantly higher reconstruction performance. It is worth noting that even though IDN has a larger number of parameters than our proposed MSRAFFN algorithm, its reconstruction performance is much lower.

5. Conclusions and Future Prospects

In this paper, we have proposed MSRAFFN for super-resolution image reconstruction. The advantage of this model is that MSRAFFB is embedded in the network to achieve the extraction of features at different scales, then fully fuse the features at different scales to recover the rich details of the image. A channel attention mechanism is introduced on the basis of feature fusion to enhance the channel features of the overall network. In addition, residual connectivity is used to integrate shallow and deep features via cross-layer connectivity, which reduces the loss of shallow feature information in the process of propagation to deeper levels while fully utilizing the features at each intermediate level. Finally, the low-resolution image is restored to the high-resolution level using deconvolution.
Through extensive experiments and analyses on five benchmark datasets, we have shown that our proposed model has certain advantages over current advanced models in terms of objective indicators. It shows better reconstruction effects in terms of subjective vision, and most of the reconstructed images have clearer and more realistic texture contours. In the future, we intend to devote our research efforts to seeking a more general and effective image reconstruction model. Notably, the proposed model is mainly intended for use in processing natural images; extending it to other scenes, such as medical images, satellite images, etc., is a problem that needs to be explored in future research.

Author Contributions

Methodology, X.H.; Software, X.H.; Writing—original draft, X.H.; Writing—review & editing, X.H.; Supervision, L.W., X.W., P.Z. and H.X.; Funding acquisition, L.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (No.62176146), the National Social Science Foundation of China (21XTY012), the National Education Science Foundation of China (No.BCA200083), and Key Project of Shaanxi Provincial Natural Science Basic Research Program (No.2023-JC-ZD-34).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset is available upon request from the authors via email at hxw_1405@163.com.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Anwar, S.; Khan, S.; Barnes, N. A deep journey into super-resolution: A survey. ACM Comput. Surv. (CSUR) 2020, 53, 1–34. [Google Scholar] [CrossRef]
  2. Wang, Z.; Chen, J.; Hoi, S.C. Deep learning for image super-resolution: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 3365–3387. [Google Scholar] [CrossRef] [PubMed]
  3. Bai, Y.; Zhang, Y.; Ding, M.; Ghanem, B. Sod-mtgan: Small object detection via multi-task generative adversarial network. In Proceedings of the European Conference On Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 206–221. [Google Scholar]
  4. Bing, X.; Zhang, W.; Zheng, L.; Zhang, Y. Medical image super resolution using improved generative adversarial networks. IEEE Access 2019, 7, 145030–145038. [Google Scholar] [CrossRef]
  5. Isaac, J.S.; Kulkarni, R. Super resolution techniques for medical image processing. In Proceedings of the 2015 International Conference on Technologies for Sustainable Development (ICTSD), Mumbai, India, 4–6 February 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 1–6. [Google Scholar]
  6. Sajjadi, M.S.; Vemulapalli, R.; Brown, M. Frame-recurrent video super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6626–6634. [Google Scholar]
  7. Shamsolmoali, P.; Zareapoor, M.; Jain, D.K.; Jain, V.K.; Yang, J. Deep convolution network for surveillance records super-resolution. Multimed. Tools Appl. 2019, 78, 23815–23829. [Google Scholar] [CrossRef]
  8. Dong, R.; Zhang, L.; Fu, H. RRSGAN: Reference-based super-resolution for remote sensing image. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–17. [Google Scholar] [CrossRef]
  9. Shabairou, N.; Cohen, E.; Wagner, O.; Malka, D.; Zalevsky, Z. Color image identification and reconstruction using artificial neural networks on multimode fiber images: Towards an all-optical design. Opt. Lett. 2018, 43, 5603–5606. [Google Scholar] [CrossRef] [PubMed]
  10. Malka, D.; Berke, B.; Tischler, Y.; Zalevsky, Z. Improving Raman spectra of pure silicon using super-resolved method. J. Opt. 2019, 21, 075801. [Google Scholar] [CrossRef]
  11. Zhao, J.; Chen, C.; Zhou, Z.; Cao, F. Single image super-resolution based on adaptive convolutional sparse coding and convolutional neural networks. J. Vis. Commun. Image Represent. 2019, 58, 651–661. [Google Scholar] [CrossRef]
  12. Cao, F.; Cai, M.; Tan, Y. Image interpolation via low-rank matrix completion and recovery. IEEE Trans. Circuits Syst. Video Technol. 2014, 25, 1261–1270. [Google Scholar]
  13. Xu, H.; Zhai, G.; Yang, X. Single image super-resolution with detail enhancement based on local fractal analysis of gradient. IEEE Trans. Circuits Syst. Video Technol. 2013, 23, 1740–1754. [Google Scholar] [CrossRef]
  14. Arsalan Bashir, S.M.; Wang, Y.; Khan, M.; Niu, Y. A Comprehensive Review of Deep Learning-based Single Image Super-resolution. arXiv 2021, arXiv:2102.09351. [Google Scholar]
  15. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
  16. Dong, C.; Loy, C.C.; He, K.; Tang, X. Learning a deep convolutional network for image super-resolution. In Proceedings of the Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Proceedings, Part IV 13. Springer: Berlin/Heidelberg, Germany, 2014; pp. 184–199. [Google Scholar]
  17. Zhang, Y.; Li, K.; Li, K.; Wang, L.; Zhong, B.; Fu, Y. Image super-resolution using very deep residual channel attention networks. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 286–301. [Google Scholar]
  18. Hui, Z.; Wang, X.; Gao, X. Fast and accurate single image super-resolution via information distillation network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 723–731. [Google Scholar]
  19. Han, W.; Chang, S.; Liu, D.; Yu, M.; Witbrock, M.; Huang, T.S. Image super-resolution via dual-state recurrent networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 1654–1663. [Google Scholar]
  20. Dong, C.; Loy, C.C.; Tang, X. Accelerating the super-resolution convolutional neural network. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part II 14. Springer: Berlin/Heidelberg, Germany, 2016; pp. 391–407. [Google Scholar]
  21. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  22. Shi, W.; Caballero, J.; Huszár, F.; Totz, J.; Aitken, A.P.; Bishop, R.; Rueckert, D.; Wang, Z. Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 1874–1883. [Google Scholar]
  23. Lai, W.S.; Huang, J.B.; Ahuja, N.; Yang, M.H. Deep laplacian pyramid networks for fast and accurate super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 624–632. [Google Scholar]
  24. Hu, Y.; Gao, X.; Li, J.; Huang, Y.; Wang, H. Single image super-resolution via cascaded multi-scale cross network. arXiv 2018, arXiv:1802.08808. [Google Scholar]
  25. Li, J.; Fang, F.; Mei, K.; Zhang, G. Multi-scale residual network for image super-resolution. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 517–532. [Google Scholar]
  26. Kim, J.; Lee, J.K.; Lee, K.M. Deeply-recursive convolutional network for image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 1637–1645. [Google Scholar]
  27. Tai, Y.; Yang, J.; Liu, X. Image super-resolution via deep recursive residual network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3147–3155. [Google Scholar]
  28. Li, Z.; Li, Q.; Wu, W.; Yang, J.; Li, Z.; Yang, X. Deep recursive up-down sampling networks for single image super-resolution. Neurocomputing 2020, 398, 377–388. [Google Scholar] [CrossRef]
  29. Shocher, A.; Cohen, N.; Irani, M. “zero-shot” super-resolution using deep internal learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3118–3126. [Google Scholar]
  30. Qin, J.; Huang, Y.; Wen, W. Multi-scale feature fusion residual network for single image super-resolution. Neurocomputing 2020, 379, 334–342. [Google Scholar] [CrossRef]
  31. Ha, V.K.; Ren, J.; Xu, X.; Zhao, S.; Xie, G.; Vargas, V.M. Deep learning based single image super-resolution: A survey. In Proceedings of the Advances in Brain Inspired Cognitive Systems: 9th International Conference, BICS 2018, Xi’an, China, 7–8 July 2018; Proceedings 9. Springer: Berlin/Heidelberg, Germany, 2018; pp. 106–119. [Google Scholar]
  32. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  33. Woo, S.; Park, J.; Lee, J.; Kweon, I. Cbam: Convolutional block attention module. In Proceedings of the European Conference On Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  34. Niu, B.; Wen, W.; Ren, W.; Zhang, X.; Yang, L.; Wang, S.; Zhang, K.; Cao, X.; Shen, H. Single image super-resolution via a holistic attention network. In Proceedings of the Computer Vision—ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part XII 16. Springer: Berlin/Heidelberg, Germany, 2020; pp. 191–207. [Google Scholar]
  35. Wu, H.; Zou, Z.; Gui, J.; Zeng, W.; Ye, J.; Zhang, J.; Liu, H.; Wei, Z. Multi-grained attention networks for single image super-resolution. IEEE Trans. Circuits Syst. Video Technol. 2020, 31, 512–522. [Google Scholar] [CrossRef]
  36. Wang, Z.; Bovik, A.C. A universal image quality index. IEEE Signal Process. Lett. 2002, 9, 81–84. [Google Scholar] [CrossRef]
  37. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  38. Agustsson, E.; Timofte, R. Ntire 2017 challenge on single image super-resolution: Dataset and study. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 126–135. [Google Scholar]
  39. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  40. Bevilacqua, M.; Roumy, A.; Guillemot, C.; Alberi-Morel, M.L. Low-complexity single-image super-resolution based on nonnegative neighbor embedding. In Proceedings of the British Machine Vision Conference, Surrey, UK; 2012; pp. 3–7. [Google Scholar]
  41. Zeyde, R.; Elad, M.; Protter, M. On single image scale-up using sparse-representations. In Proceedings of the Curves and Surfaces: 7th International Conference, Avignon, France, 24–30 June 2010; Revised Selected Papers 7. Springer: Berlin/Heidelberg, Germany, 2012; pp. 711–730. [Google Scholar]
  42. Martin, D.; Fowlkes, C.; Tal, D.; Malik, J. A database of human segmented natural images and its application to evaluating segmentation algorithms and measuring ecological statistics. In Proceedings of the Eighth IEEE International Conference on Computer Vision, ICCV 2001, Vancouver, BC, Canada, 7–14 July 2001; IEEE: Piscataway, NJ, USA, 2001; Volume 2, pp. 416–423. [Google Scholar]
  43. Huang, J.B.; Singh, A.; Ahuja, N. Single image super-resolution from transformed self-exemplars. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 5197–5206. [Google Scholar]
  44. Matsui, Y.; Ito, K.; Aramaki, Y.; Fujimoto, A.; Ogawa, T.; Yamasaki, T.; Aizawa, K. Sketch-based manga retrieval using manga109 dataset. Multimed. Tools Appl. 2017, 76, 21811–21838. [Google Scholar] [CrossRef]
Figure 1. Schematic diagram of LapSRN network; red arrow: convolution for feature extraction, blue arrow: inverse convolution for resolution enhancement, green arrow: pixel-by-pixel summation.
Figure 1. Schematic diagram of LapSRN network; red arrow: convolution for feature extraction, blue arrow: inverse convolution for resolution enhancement, green arrow: pixel-by-pixel summation.
Sensors 23 09458 g001
Figure 2. The structures of recursive networks: simplified network models for (ac) DRRN, DSRN, and DRUDN, respectively.
Figure 2. The structures of recursive networks: simplified network models for (ac) DRRN, DSRN, and DRUDN, respectively.
Sensors 23 09458 g002
Figure 3. SEnet network structure.
Figure 3. SEnet network structure.
Sensors 23 09458 g003
Figure 4. Network structure of the multi-scale recursive attention feature fusion network.
Figure 4. Network structure of the multi-scale recursive attention feature fusion network.
Sensors 23 09458 g004
Figure 5. Structure of the multi-scale attention feature fusion module; the pink shaded portion is the MSFE block and the blue shaded portion is the AFF block.
Figure 5. Structure of the multi-scale attention feature fusion module; the pink shaded portion is the MSFE block and the blue shaded portion is the AFF block.
Sensors 23 09458 g005
Figure 6. Deconvolution model at × 2 SR.
Figure 6. Deconvolution model at × 2 SR.
Sensors 23 09458 g006
Figure 7. Effect of different values of C on network performance: (a) network convergence analysis for B = 4; (b) network convergence analysis for B = 6; and (c) network convergence analysis for B = 8.
Figure 7. Effect of different values of C on network performance: (a) network convergence analysis for B = 4; (b) network convergence analysis for B = 6; and (c) network convergence analysis for B = 8.
Sensors 23 09458 g007
Figure 8. Performance comparison of B4C8, B6C8, and B8C6.
Figure 8. Performance comparison of B4C8, B6C8, and B8C6.
Sensors 23 09458 g008
Figure 9. Comparison of the proposed method with other methods in terms of subjective visualization on the Set14, B100, Urban100, and Manga109 datasets at ×4 SR.
Figure 9. Comparison of the proposed method with other methods in terms of subjective visualization on the Set14, B100, Urban100, and Manga109 datasets at ×4 SR.
Sensors 23 09458 g009
Figure 10. The performance comparison of Non-MSFE, Non-AFF, and MSFE + AFF on the Set5 (×4) dataset.
Figure 10. The performance comparison of Non-MSFE, Non-AFF, and MSFE + AFF on the Set5 (×4) dataset.
Sensors 23 09458 g010
Figure 11. Visual comparison of the Non-MSFE, Non-AFF, and MSFE + AFF on the Set5 (×4) dataset.
Figure 11. Visual comparison of the Non-MSFE, Non-AFF, and MSFE + AFF on the Set5 (×4) dataset.
Sensors 23 09458 g011
Figure 12. Model performance compared to the number of parameters on the Set14 dataset with a ×4 scaling factor.
Figure 12. Model performance compared to the number of parameters on the Set14 dataset with a ×4 scaling factor.
Sensors 23 09458 g012
Table 1. Detailed network architecture.
Table 1. Detailed network architecture.
LayerKernel Size
Shallow feature extraction module3 × 3 × 64, 3 × 3 × 64
Multi-scale feature extraction unit3 × 3 × 64, 3 × 3 × 64, 1 × 1 × 128
3 × 3 × 64, 1 × 1 × 128, 3 × 3 × 64, 1 × 1 × 128
3 × 3 × 64, 1 × 1 × 128, 3 × 3 × 64, 1 × 1 × 128
3 × 3 × 64, 1 × 1 × 128, 3 × 3 × 64, 1 × 1 × 128
Attention feature fusion unit1 × 1 × 512
Average_pool, fc1, fc2
3 × 3 × 64
1 × 1 × 128
Reconstruction module3 × 3 × 64
Deconvolution
3 × 3 × 64 × 3
Table 2. Mean PSNR/SSIM on the five datasets(×4 SR) when B is taken 4, 6, and 8 and C is taken 6, 8, and 10. The best and second-best results are indicated by bold and underline, respectively.
Table 2. Mean PSNR/SSIM on the five datasets(×4 SR) when B is taken 4, 6, and 8 and C is taken 6, 8, and 10. The best and second-best results are indicated by bold and underline, respectively.
Different MethodsSet5Set14B100Urban100Manga109
PSNR/SSIMPSNR/SSIMPSNR/SSIMPSNR/SSIMPSNR/SSIM
B4C631.99/0.892428.43/0.777727.45/0.732225.74/0.774730.03/0.9029
B4C832.08/0.893628.48/0.779027.48/0.733125.83/0.777930.12/0.9045
B4C1031.36/0.889028.26/0.774727.33/0.728425.57/0.769329.41/0.8962
B6C632.05/0.892628.49/0.778127.49/0.732225.82/0.775830.08/0.9036
B6C832.28/0.895428.66/0.782827.61/0.736826.23/0.789530.96/0.9108
B6C1032.12/0.893928.48/0.779527.48/0.733825.84/0.778530.18/0.9050
B8C632.10/0.894328.53/0.779527.50/0.733225.86/0.778230.24/0.9061
B8C832.09/0.893928.47/0.779127.50/0.733325.85/0.778030.13/0.9045
B8C1031.74/0.889428.34/0.776827.37/0.729825.54/0.776230.18/0.9050
Table 3. Comparison with classical methods in terms of upsampling method, use of recurrent and residual networks, and loss function. The √ means that a method was used, while a blank space means that it was not used.
Table 3. Comparison with classical methods in terms of upsampling method, use of recurrent and residual networks, and loss function. The √ means that a method was used, while a blank space means that it was not used.
MethodsUpsampling MethodsRecursive NetworkResidual NetworkLoss Function
SRCNNbicubic L 2
FSRCNNdeconvolution L 2
VDSRbicubic MSE
DRCNbicubic L 2
LapSRNdeconvolution Char
IDNdeconvolution MSE
MSRNsubpixel convolution L 1
Oursdeconvolution L 1
Table 4. Quantitative evaluation (PSNR(dB)/SSIM) of different SR algorithms with ×2, ×3 and ×4; the best results are indicated by bold and the second-best results by underline.
Table 4. Quantitative evaluation (PSNR(dB)/SSIM) of different SR algorithms with ×2, ×3 and ×4; the best results are indicated by bold and the second-best results by underline.
MethodScaleSet5Set14B100Urban100Manga109
PSNR/SSIMPSNR/SSIMPSNR/SSIMPSNR/SSIMPSNR/SSIM
Bicubic 33.66/0.929930.24/0.866829.56/0.843126.88/0.840330.80/0.9339
SRCNN [16] 36.66/0.954232.45/0.906731.36/0.887929.50/0.894635.60/0.9663
FSRCNN [20] 37.00/0.955832.63/0.908831.53/0.892029.88/0.902036.67/0.9710
VDSR [21] 37.53/0.958733.03/0.912431.90/0.896030.76/0.914037.22/0.9750
DRCN [26] 37.63/0.958833.04/0.911831.85/0.894230.75/0.913337.55/0.9732
LapSRN [23]× 237.52/0.959132.99/0.912431.80/0.894930.41/0.910137.53/0.9740
DRRN [27] 37.74/0.959133.23/0.913632.05/0.897331.23/0.918837.88/0.9749
DSRN [19] 37.66/0.959033.15/0.913032.10/0.897030.97/0.9160-/-
IDN [18] 37.83/0.960033.30/0.914832.08/0.898531.27/0.9196-/-
MSRN [25] 38.08/0.960533.74/0.917032.23/0.901332.22/0.932638.82/0.9868
Proposed Method 38.03/0.960733.58/0.917832.18/0.899832.22/0.928938.83/0.9776
Bicubic 30.39/0.868227.55/0.774227.21/0.738524.46/0.734926.95/0.8556
SRCNN [16] 32.75/0.909029.30/0.821528.41/0.786326.24/0.798930.48/0.9117
FSRCNN [20] 33.18/0.914029.37/0.824028.53/0.791026.43/0.808031.10/0.9210
VDSR [21] 33.66/0.921329.77/0.831428.82/0.797627.14/0.827932.01/0.9340
DRCN [26] 33.82/0.922629.76/0.831128.80/0.796327.15/0.827632.24/0.9343
LapSRN [23]× 333.82/0.922729.79/0.832028.82/0.797327.07/0.827132.21/0.9350
DRRN [27] 34.03/0.924429.96/0.834928.95/0.800427.53/0.837532.71/0.9379
DSRN [19] 33.88/0.922030.26/0.837028.81/0.797027.16/0.8280-/-
IDN [18] 34.11/0.925329.99/0.835428.95/0.801327.42/0.8359-/-
MSRN [25] 34.38/0.926230.34/0.839529.08/0.804128.08/0.855433.44/0.9427
Proposed Method 34.48/0.928230.35/0.842129.11/0.805928.27/0.854333.81/0.9460
Bicubic 28.42/0.810426.00/0.702725.96/0.667523.14/0.657724.89/0.7866
SRCNN [16] 30.48/0.862827.50/0.751326.90/0.710124.52/0.722127.58/0.8555
FSRCNN [20] 30.72/0.866027.61/0.755026.98/0.715024.62/0.728027.90/0.8610
VDSR [21] 31.35/0.883828.01/0.767427.29/0.725125.18/0.752428.83/0.8870
DRCN [26] 31.53/0.885428.02/0.767027.23/0.723325.14/0.751028.93/0.8854
LapSRN [23]× 431.54/0.886628.09/0.769427.32/0.726425.21/0.755329.09/0.8900
DRRN [27] 31.68/0.888828.21/0.772127.38/0.728425.44/0.763829.45/0.8946
DSRN [19] 31.40/0.883028.07/0.770027.25/0.724025.08/0.7470-/-
IDN [18] 31.82/0.890328.25/0.773027.41/0.729725.41/0.7632-/-
MSRN [25] 32.07/0.890328.60/0.775127.52/0.727326.04/0.789630.17/0.9034
Proposed Method 32.28/0.895428.66/0.782827.61/0.736826.23/0.789530.96/0.9108
Table 5. Average PSNR and SSIM of the three models on the five datasets (×4).
Table 5. Average PSNR and SSIM of the three models on the five datasets (×4).
MethodSet5Set14B100Urban100Mnaga109
PSNR/SSIMPSNR/SSIMPSNR/SSIMPSNR/SSIMPSNR/SSIM
MSFE + AFF32.28/0.895428.66/0.782827.61/0.736826.23/0.789530.96/0.9108
Non MSFE31.97/0.891828.40/0.777227.43/0.731225.72/0.773229.90/0.9015
Non AFF32.25/0.896028.63/0.782727.59/0.736926.16/0.788630.55/0.9098
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Han, X.; Wang, L.; Wang, X.; Zhang, P.; Xu, H. A Multi-Scale Recursive Attention Feature Fusion Network for Image Super-Resolution Reconstruction Algorithm. Sensors 2023, 23, 9458. https://doi.org/10.3390/s23239458

AMA Style

Han X, Wang L, Wang X, Zhang P, Xu H. A Multi-Scale Recursive Attention Feature Fusion Network for Image Super-Resolution Reconstruction Algorithm. Sensors. 2023; 23(23):9458. https://doi.org/10.3390/s23239458

Chicago/Turabian Style

Han, Xiaowei, Lei Wang, Xiaopeng Wang, Pengchao Zhang, and Haoran Xu. 2023. "A Multi-Scale Recursive Attention Feature Fusion Network for Image Super-Resolution Reconstruction Algorithm" Sensors 23, no. 23: 9458. https://doi.org/10.3390/s23239458

APA Style

Han, X., Wang, L., Wang, X., Zhang, P., & Xu, H. (2023). A Multi-Scale Recursive Attention Feature Fusion Network for Image Super-Resolution Reconstruction Algorithm. Sensors, 23(23), 9458. https://doi.org/10.3390/s23239458

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop