Next Article in Journal
Design and Implementation of HD Mapping, Vehicle Control, and V2I Communication for Robo-Taxi Services
Next Article in Special Issue
Hint-Based Image Colorization Based on Hierarchical Vision Transformer
Previous Article in Journal
Can the Supido Radar Be Used for Measuring Ball Speed during Soccer Kicking? A Reliability and Concurrent Validity Study of a New Low-Cost Device
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Scale Attention-Guided Non-Local Network for HDR Image Reconstruction

School of Computing, Gachon University, Seongnam 13120, Korea
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(18), 7044; https://doi.org/10.3390/s22187044
Submission received: 21 July 2022 / Revised: 31 August 2022 / Accepted: 4 September 2022 / Published: 17 September 2022
(This article belongs to the Special Issue Applications of Video Processing and Computer Vision Sensor II)

Abstract

:
High-dynamic-range (HDR) image reconstruction methods are designed to fuse multiple Low-dynamic-range (LDR) images captured with different exposure values into a single HDR image. Recent CNN-based methods mostly perform local attention- or alignment-based fusion of multiple LDR images to create HDR contents. Depending on a single attention mechanism or alignment causes failure in compensating ghosting artifacts, which can arise in the synthesized HDR images due to the motion of objects or camera movement across different LDR image inputs. In this study, we propose a multi-scale attention-guided non-local network called MSANLnet for efficient HDR image reconstruction. To mitigate the ghosting artifacts, the proposed MSANLnet performs implicit alignment of LDR image features with multi-scale spatial attention modules and then reconstructs pixel intensity values using long-range dependencies through non-local means-based fusion. These modules adaptively select useful information that is not damaged by an object’s movement or unfavorable lighting conditions for image pixel fusion. Quantitative evaluations against several current state-of-the-art methods show that the proposed approach achieves higher performance than the existing methods. Moreover, comparative visual results show the effectiveness of the proposed method in restoring saturated information from original input images and mitigating ghosting artifacts caused by large movement of objects. Ablation studies show the effectiveness of the proposed method, architectural choices, and modules for efficient HDR reconstruction.

1. Introduction

High-dynamic-range (HDR) imaging techniques extend the range of brightness in an image such that a photograph taken by a camera becomes as similar as possible to the scene as it would be observed by the naked eye. HDR images have been used in a wide range of applications such as photography, film, and games [1,2] as they contain more information from a given scene and can provide a better visual experience. However, because camera image sensors generally have a narrow dynamic range, capturing scenes with more widely varying ranges of brightness using a single exposure value is considered relatively difficult [3,4]. Recent advances in imaging technology have enabled the acquisition of HDR images with only a single-sensor camera, but these special cameras are often very expensive for consumers in general. Therefore, the primary method used to obtain HDR imaging is to continuously photograph and overlap multiple images with a low dynamic range (LDR) with different exposure values and then select or fuse the best photo pixels or segments to reconstruct a single HDR image [5,6,7].
Traditional methods mostly depend on rejection-based approaches [8,9,10,11,12,13,14,15,16,17], rigid and non-rigid registration-based approaches [18,19,20,21,22,23] and patch-based optimization methods [24,25] for reconstructing HDR images. Rejection-based HDR reconstruction methods produce high-quality HDR contents for static scenes with LDR contents for moving regions, thus limiting their scope for static scenes mostly. Registration-based traditional approaches work plausibly well with static scenes and small movements, however, fail in complex and dynamic scenes with large motions and occluded pixels. Patch-based optimization methods, though capable of high-quality HDR image reconstruction, suffer from high computational complexity and often fail to produce plausible HDR images in run-time.
To alleviate the HDR reconstruction problem, recently, various deep learning methods based on convolutional neural network (CNN) have been utilized to perform multi-exposure-based multi-frame HDR imaging [26,27,28,29,30,31]. CNN-based methods have shown significant promise in producing HDR contents with fewer LDR images (i.e., typically three LDR images) for comparatively complex scenes. Early studies focused on pre-processing methods to align LDR images beforehand [32,33,34,35] using optical flow [22,23,28,29] or homography [29,34,36,37], and then input the data to convolutional neural network (CNN) models. However, although properly coping with small movements of objects in this way has been shown to be possible, the performance of these methods is limited by the accuracy of the alignment. Moreover, these methods often fail in scenes with large motions and saturated regions.
Recent CNN models are focused on performing feature-level attention-based implicit alignment mechanisms that use contextual information from the LDR images to adaptively select pixels/regions for efficient fusion [26,27,30,38]. These methods enabled the movement of an object to be properly processed while maintaining important information for each input image. The key challenge of such a multi-frame HDR imaging approach is to properly handle the “ghosting artifacts” caused by inappropriate image fusion due to objects moving within a scene or camera instability. For example, as shown in Figure 1, if images captured continuously at different times are improperly processed and fused, different objects appear to overlap in the resulting image, i.e., ghosting artifacts, due to the motion of the objects.
In this study, we propose a deep neural network, called MSANLnet, that uses the multi-scale attention mechanism and non-local means technique to effectively alleviate the ghosting artifact problem in HDR image reconstruction. The proposed MSANLnet has two distinctive parts—(a) multi-scale attention for object motion and saturation mitigation, and (b) non-local means-based fusion. Existing spatial-wise attention only captures important regions on a single scale, which is not effective for distinguishing both object movements and saturated regions in the case of producing HDR contents. In the proposed method, LDR features are aligned based on multi-scale spatial attention modules. During this process, spatial attention is performed for each scale of features to expand the size of the receptive field and progressively correct and fuse important information, such as large movements and saturation, which is found to be effective in producing high-quality HDR content. Again, the non-local means module looks at the whole image and only fuses the correlated global contents with the reference image. Hence, the multi-scale attention mechanism can effectively capture the local movements of the objects and the non-local means-based fusion can reduce ghosting artifacts on a global level by looking at the whole image, which effectively mitigates ghosting artifacts for large object motions and global shifts such as camera translation.
This two-stage design effectively selects different exposure values for each input frame and useful information that is not altered by motion and then fuses images to minimize locality to reduce noise and mitigate the ghosting artifact problem. We validated that the proposed method performs better than the existing methods quantitatively and qualitatively through the experiments on a publicly available dataset that has been mainly used for HDR image restoration. We summarize our contributions as follows.
  • We propose MSANLnet, a multi-scale attention-guided non-local network for HDR image reconstruction, that extracts important features from the LDR features using the multi-scale spatial attention and adaptively fuses the contextual features to obtain HDR images.
  • We show that the multi-scale spatial attention, combined with the non-local means-based fusion, can effectively alleviate the “ghosting artifact” and produce aesthetic HDR images.
  • Our proposed method outperforms the existing methods in both qualitative and quantitative measures, validating the efficacy of the attention modules, non-local means-based fusion, and architectural choices.
The remainder of this study is organized as follows. Related work is discussed in Section 2. The proposed MSANLnet is presented along with detailed descriptions of each module in Section 3. The experimental setup and results are presented in Section 4. Finally, Section 5 presents our conclusions and suggests some possible avenues for further research.

2. Related Work

First proposed by Madden [40] and Mann [41] and popularized by Debevec and Malik [42] for digital imaging, multi-exposure-fusion-based HDR reconstruction methods can be classified into two major classes—(i) traditional approaches, and (ii) deep learning-based approaches. For brevity, we restrict our related work only to multi-exposure-fusion-based approaches.
Rejection-based methods try to find the pixels/regions with motions and select only pixels/regions using a reference image or the static contents of the LDR images [5,8,12], i.e., these methods reject the moving contents. Rejection mechanisms are typically implemented by performing patch-wise comparison [9], illumination consistency and linear relationships among pixels of the LDR images [16,43,44], thresholding [5,11,12], background probability map [45], super-pixel comparison [15] etc. Such approaches can produce fast results, however, are limited to producing LDR contents in the dynamic regions.
Alignment-based traditional methods typically select one reference exposure LDR image and register the remaining LDR images either using rigid registration mechanisms (e.g., SIFT-based [18], SURF-based [10], Median Threshold Bitmap-based [19] methods, etc.) or non-rigid registration mechanisms (e.g., optical flow-based methods [20,22,23,46]). These methods can produce plausible results for static scenes and small motions, however, perform poorly in recovering saturated pixels.
Patch-based optimization methods typically synthesize LDR patches based on the structure reference patches by finding dense correspondence among the LDR images and fuse the synthesized LDR images to produce HDR contents [24,25,47]. These methods can produce high-quality HDR content but suffer from computational complexity. Though most of the traditional methods produce plausible HDR contents for static scenes, they suffer from poor synthesis/misalignment of LDR contents in dynamic scenes with large motions and recovering saturated pixels/regions from the input LDR images.
Deep learning-based methods are capable of synthesizing novel HDR contents based on the input LDR images, making them suitable for complex dynamic scenes. Early methods employed CNN-based methods, e.g., optical flow estimation [28,29], multi-scale feature extraction [38], etc., to compensate ghosting artifacts. Subsequently, attention-based alignment methods are used to remove ghosting artifacts and recover saturated regions [26,27,39,48].
Deep learning-based multi-scale attention mechanisms have also shown significant promise in various image reconstruction tasks, e.g., semantic image synthesis [49,50], generative image synthesis [51,52,53], image inpainting [54,55,56,57], segmentation [58,59,60,61,62,63,64], super-resolution [65,66,67,68,69,70,71,72,73,74], image enhancement [75], etc. Though several HDR approaches utilize multi-scale feature fusion [26,38,39,76,77,78,79,80,81], the removal of object motion and the mitigation of saturation using multi-scale attention mechanisms remain an under-explored research area.
Recently, generative adversarial networks (GANs) have been employed to supervise the HDR reconstruction process, which can effectively correct large motions with higher efficiency [82]. Despite the success of deep learning-based approaches, removing ghosting artifacts in HDR images for large motions is still an active area of research.

3. Methods

3.1. Overview of MSANLnet

The MSANLnet takes three input LDR tensors as inputs. Each tensor consists of an LDR image and the tone-mapped HDR image of the corresponding LDR image. First, the three LDR images in the input LDR set (i.e., L = { L 1 , L 2 , L 3 } , where L i = { 1 , 2 , 3 } R H × W × 3 ) are arranged according to their exposure lengths t i , i.e., short exposure ( L 1 ), intermediate exposure ( L 2 ), and long exposure length ( L 3 ). Then, a mapping process is performed on the input LDR image set L to convert them into the input HDR image set, H = { H 1 , H 2 , H 3 } , H i = { 1 , 2 , 3 } R H × W × 3 . Here, the gamma correction ( γ = 2.2 , following [26]) is used for the mapping process as follows:
H i = ( L i ) γ t i ; i = 1 , 2 , 3 ,
Then, a pair of input tensors X i for each L i is obtained by concatenating the gamma-corrected H i and the corresponding L i in the channel dimension, i.e., X i = { H i , L i } , where X i R H × W × 6 . Among these tensors, X 2 with the intermediate exposure length is selected as a reference tensor. The set of input tensors X is then fed to MSANLnet to produce an HDR image.
We define the process of generating the final HDR image H f by fusing the three tensors, X = { X 1 , X 2 , X 3 } , as follows,
H f = f ( X 1 , X 2 , X 3 ; θ ) ,
where f represents the proposed MSANLnet and θ is a parameter of the corresponding network.
The proposed MSANLnet model uses a CNN-based encoder-decoder structure and is largely composed of multi-scale spatial attention modules and non-local means-based fusion modules. Figure 2 shows the overview of the proposed network. As shown in Figure 2, the encoder extracts spatial attention features based on the scale of LDR tensor set X i through the multi-scale spatial attention module (please refer to Section 3.2 and Supplementary Materials). The multi-scale spatial attention features contribute to compensating motions and recovering feature values from the saturated regions adaptively. The decoder uses a transposed convolution block to restore the reduced resolution to the original resolution while concatenating spatial attention features and reference image features. Finally, the final HDR image is output through a non-local means-based fusion mechanism (please refer to Section 3.3 and Supplementary Materials) which considers the global relationships among feature values, thus effectively removing ghosting artifacts for larger motions during the HDR image reconstruction.

3.2. Multi-Scale Spatial Attention Module

The proposed approach is designed to effectively identify areas to be used in images with long exposure values and images with short exposure values. Specifically, a multi-scale attention mechanism is devised in the encoder to implicitly align the two images with the reference image, as shown in Figure 2.
First, three 6-channel input tensors X 1 , X 2 , X 3 are encoded in spatial multi-scales, respectively, via a convolutional layer to extract features. During this process, max pooling is applied to reduce the spatial size of the feature map extracted at each step by half from the previous step. Thus, the receptive field is increased to capture a larger foreground movement in the multi-scale attention module. Note that when we decrease the spatial size of a feature map by half, we increase its number of channels by double. For example, in our experiment, we used 64 channels for the first scale and 128 channels for the second scale.
As shown in Figure 2, a spatial attention block is executed for each scale to extract the spatial attention maps of the non-reference images. A spatial attention block is shown in Figure 3. In the spatial attention block, the input is the concatenation of the non-reference image features, X i ; i 2 , and the reference image X 2 features. Then, the spatial attention map is extracted at each scale s as follows,
A i ( s ) = g ( X i ( s ) , X 2 ( s ) ) ; i = 1 , 3 ; s = 1 , 2 , , N ,
where s denotes the number of spatial scales and g ( ) is the spatial attention map extractor. The attention map extractor consists of two consecutive convolutions and sigmoid function, as used in a previous study [26]. Note that, in our experiment, we used N=2 (i.e., two scales as depicted in Figure 2). An element of the spatial attention map A i is in a range of [ 0 1 ] and the spatial size of the generated A i ( s ) is equal to X i ( s ) .
Finally, an element-wise multiplication between the input feature X i ( s ) and the extracted A i ( s ) is performed to generate the final spatial attention feature X i ( s ) of the images.
X i ( s ) = A i ( s ) · X i ( s ) ; i = 1 , 3 ; s = 1 , 2 , , N .
As seen in Figure 4, two spatial attention maps focus on important regions for implicit alignment of the input LDR images to the reference image. As the network is provided with explicit multi-scale contextual information, it can compensate for large foreground motions and can effectively mitigate the ghosting artifacts. Moreover, as the spatial attention maps represent the differences among the LDR images with varying exposures, the network implicitly learns to recover saturated pixel values.

3.3. Non-Local Means-Based Fusion

In some local areas of the image feature, sufficient information is not available due to occlusion or saturation caused by the movement of objects [26]. Therefore, in image fusion networks, useful information must be extracted into a large receptive field. However, a single CNN convolution filter can be used in only a limited area of the receptive field. Previous studies have shown the importance of short and long-range dependencies among the pixel/feature values for various computer vision tasks [54,83,84,85,86]. Non-local means is a method of calculating long-range dependencies and restoring pixels through weighted averages based on these dependencies, intuitively beneficial for removing ghosting artifacts. To this end, we applied a non-local block to image fusion so that global information is utilized and locality is reduced to effectively alleviate ghosting artifacts.
First, we utilize the residual dense convolutional block [87] before applying the non-local block, such that the feature Z recovered to the original size through the decoding process learns sufficient local information. By concatenating all layers, useful local information is adaptively extracted through the features of the previous layer and the features of the current layer. The extracted X enters a non-local block input and outputs Y of the same size, H × W × C , as a result. Figure 5 illustrates the structure of a residual dense convolutional block composed of three convolutional blocks.
For the non-local based fusion, we adopt the asymmetric pyramid non-local block (APNB) [88] to construct the non-local blocks. In APNB, spatial pyramid pooling is applied to a standard non-local block to reduce computation by sampling part of the feature map based on meaningful global information. Figure 6 shows the structure of an asymmetric pyramidal non-local block, including spatial pyramid pooling.
First, the input feature map X is converted to different embeddings, Φ , θ , and γ , respectively, through three 1 × 1 convolutions, W Φ , W θ and W γ .
Φ = W Φ ( X ) ; θ = W θ ( X ) ; γ = W γ ( X ) .
The spatial pyramid pooling module is composed of four pooling layers that derive the results of different sizes in parallel. Such a pyramid pooling mechanism improves the expressive power of global features as previous studies have shown the effectiveness of the global and multi-scale representations for capturing scene semantics. The sampling process through pyramid pooling P θ and P γ are represented by the following equation.
θ P = P θ ; γ P = P γ .
The number of spatially sampled anchor points S can be expressed as,
S = 110 = n 1 , 3 , 6 , 8 n 2 ,
where S denotes the number of sampled anchor points and n denotes the width of the output features processed through the pooling layer. Thereafter, based on the standard non-local block method [89], the pseudo-matrix V P of θ P and Φ is calculated as follows,
V P = Φ T × θ P .
Instance normalization is then applied to V P to generate a normalized pseudo-matrix V P , and the Softmax function of the self-attention mechanism is used to derive the attention result as follows,
Z = V P × γ P T .
The final output of the asymmetric pyramidal non-local block is obtained as follows,
Y = c a t ( W o ( Z ) , X ) ,
where W o is composed of 1 × 1 convolution and can be learned by weighting the parts where parameters are important in the non-local operation. Subsequently, the network generates a final 3-channel HDR image through a convolutional layer.

4. Experimental Results

4.1. Dataset, Objective Function, Implementation Details, and Evaluation Design

4.1.1. Dataset

We used an open dataset [28] for HDR performance comparison to train and evaluate the proposed method. Specifically, out of a total of 89 image sets, 74 image sets were used as a training set, 5 were used as a validation set, and the remaining 10 were used as a testing set. Each image set comprised three LDR images and one HDR ground-truth image. Here, three LDR images were captured using exposure bias { 2 , 0 , + 2 } or { 3 , 0 , + 3 } . As in an existing study on HDR [26], in the training stage, a total of 1775 image patches were used to perform training by cropping LDR images with a size of 1000 × 1500 into 256 × 256 patch sizes and using them with a stride of 128 through data augmentation to mitigate over-fitting.

4.1.2. Objective Function

As proposed in a previous work [28], HDR images were mainly presented after being tone-mapped, and thus, it was more effective to optimize the network in the tone-mapping domain than in the HDR domain. A tone-mapping process using a μ -law was performed on the HDR image H and ground-truth image GT derived in this study as follows.
T ( X ) = l o g ( 1 + μ X ) l o g ( 1 + μ ) ,
where μ is a parameter indicating the degree of compression, and T ( X ) denotes a tone-mapped image. We set μ = 5000 [28], and the range of the resulting T ( X ) is [ 0 , 1 ] . Subsequently, to minimize the error value between T ( H ) and T ( GT ) by applying per-pixel L 1 -Loss, the network was trained after defining the loss function L as follows,
L = | T ( GT ) T ( H ) | 1 .

4.1.3. Implementation Details

In the training stage of the proposed method, the number of epochs and batch size were set to 200 and 4, respectively. The learning rate was initially set to 0.0001, and then the learning rate was reduced by multiplying by 0.1 at 100 epochs. We used the Adam optimizer [90], and both learning and evaluation were implemented using the Pytorch framework [91]. Our experiments are performed on a single NVIDIA 2080Ti GPU. The number of model parameters is 2.8 million (2,772,707) and the average inference time per image is 0.0181 s.

4.1.4. Evaluation Design

To evaluate the efficiency of the proposed method, we compared it with five state-of-the-art methods, both in qualitative and quantitative manner. Specifically, we compared our method with AHDR [26], Wu [28], AD [27], NHDRR [30], and DAHDR [39]. AHDR [26] introduced spatial attention modules in HDR imaging to guide the merging according to the reference image. AD [27] used spatial attention module for feature-level attention and a multi-scale alignment module (i.e., PCD alignment [92]) to align the images in the feature-level. DAHDR [39] exploits both spatial attention and feature channel attention to achieve ghost-free merging. NHDRR [30] is based on U-Net-based architecture with non-local means to produce HDR images. All of the selected models are deep learning-based methods for generating an HDR image using multiple LDR images. In addition, all of these models used LDR images with large motions as training datasets, which were considered appropriate to objectively compare performance with our method. All models were trained in the same way and in the same environment as the proposed approach for a fair comparison.

4.2. Quantitative Evaluation

To quantitatively evaluate the performance of the model, we measured the peak signal-to-noise ratio (PSNR) and structural similarity index map (SSIM) on the testing set. PSNR is used to evaluate the impact of lost information on the quality of a generated or compressed image and represents the power of received noise with respect to the maximum power of the received signal. SSIM is a method designed to evaluate human visual quality differences rather than numerical loss and evaluates quality by comparing luminance, contrast, and structure values that make up images, rather than only performing comparisons between pixels. We used the PSNR-l and PSNR- μ indicators, which measure PSNR in linear and tone-mapped domains, respectively. For the SSIM, we used values measured in tone-mapped domains.
Table 1 shows the average value of the quantitative indicators, and the larger values correspond to the higher quality. The proposed model achieved higher values than other models in all three quantitative performance indicators. The proposed method achieved better performance in terms of PSNR-l, PSNR- μ , and SSIM measurements compared to AHDR [26], Wu [28], AD [27], NHDRR [30], and DAHDR [39]. This occurred because the proposed method can restore details and effectively eliminate ghosting artifacts through non-local means-based fusion modules. NHDRR [30] also merges images using non-local means modules, but MSANLnet, which extracts features using scale-specific spatial attention, exhibited numerically better performance.

4.3. Qualitative Evaluation

Figure 7 and Figure 8 visually compare the results of HDR image generation via the proposed method and existing state-of-the-art models [26,27,28,30,39]. In particular, Figure 7 qualitatively compares the experimental results of images with saturated backgrounds and large foreground movements. Input images referenced by the network by exposure value are shown in Figure 7a, and the tone-mapped HDR result images produced by the proposed method are shown in Figure 7b. The corresponding testing set image had a saturated background area and large foreground movements. Areas with large movements and saturation levels were cut into patches as shown in Figure 7b,c, respectively. The details of the background were obscured by the movement of an object in an image with a small exposure length, and the information was lost even in an image with long or medium exposure lengths. Because the image was saturated, learning models were likely to use artifacts and distorted information from the first image with the background occluded for HDR imaging, as seen in Figure 7c.
As seen in Figure 7, Wu [28] produced an excessively smooth image and failed to restore the details of the saturated area and removed ghosting artifacts. Although AHDR [26] selectively incorporated useful areas through spatial attention modules to restore obscured or saturated details, they still exhibited ghosting artifacts. NHDRR [30] removed ghosting artifacts to some extent, but output some blurring artifacts and missed the details of many areas. AD [27] uses PCD alignment modules [92], which reduce ghosting artifacts to some extent. However, the over-exposed area was not fully recovered. In DAHDR [39], the afterimage of the finger remains intact, which is not effective in removing ghosting artifacts. Furthermore, it can be seen that the hidden area cannot be restored as well. The results of the proposed MSANLnet expressed more details compared to the existing methods, while also eliminating ghosting artifacts, as may be observed from the figures. The resulting images demonstrate that the proposed MSANLnet expresses more details of saturated regions through the scale-specific spatial attention modules, and effectively reduces ghosting artifacts through non-local means-based fusion.
Figure 8 also shows a comparison of images with motion and supersaturation regions in the testing set. The proposed MSANLnet restored details more clearly compared to other models, and also removed visual artifacts to produce higher-quality HDR images.

4.4. Ablation Study

4.4.1. Ablation on the Network Structure

We performed an ablation study of the proposed network structure and analyzed the results. We compared the proposed MSANLnet with the following variant models to identify the importance of the individual components. Table 2 shows the quantitative comparison among the baselines and the proposed MSANLnet.
  • ANLnet: The multi-scale attention module was removed from this version; i.e., this was a variant model using a single-scale attention module of the baseline model AHDR [26] and a non-local means-based fusion module.
  • MSANet: The non-local fusion module was removed from this version. That is, this variant adopted a multi-scale attention module and dilated residual dense block (DRDB)-based fusion as used in baseline models.
  • MSANLnet: The proposed MSANLnet model, which includes a multi-scale attention module and a non-local means-based fusion module.
As shown in Table 2, when a multi-scale attention module was used, the performance in terms of PSNR was improved by 1.41 dB compared to the use of a single-scale module. As seen in Figure 9, an afterimage of arm movement remains in MSAnet, and the difference cannot be repaired naturally, creating boundaries. However, it can be seen that the proposed model with multi-scale attention effectively eliminates such artifacts. These results confirm that expanding the receptive field helps reduce ghosting artifacts and restore details.

4.4.2. Comparison of Degradation of HDR Restoration for Global Motion

We have performed additional experiments with the translation applied to the input LDR images and visually compared the HDR restoration performance between the proposed and existing methods. Specifically, we globally shifted the first input LDR image to the right by 50 pixels, kept the reference LDR image (i.e., the second image) unchanged, and shifted the third input LDR image to the left by 50 pixels, to simulate the global translation operation. The visual results are summarized in Figure 10.
As seen in Figure 10, most of the existing methods fail to mitigate ghosting artifacts in the case of such global motion caused by the camera’s translation. However, our proposed method can effectively produce HDR image contents without severe quality degradation, even though we do not employ any explicit alignment mechanisms. This is because the multi-scale attention mechanism can effectively capture the local movements of the objects and the non-local means-based fusion reduces ghosting artifacts on a global level by looking at the whole image, which essentially mitigates ghosting artifacts for large object motions and global shifts such as translation.

4.4.3. Generalization Ability on Different Dataset

We performed additional experiments on the generalization ability of the proposed method on a different test dataset. Specifically, we used test dataset provided in [93] and summarized the visual results in Figure 11. As seen in the figure, our method can generalize to unseen test datasets and performs comparatively better than the existing methods in restoring HDR contents, even though our method was not trained on the dataset.

5. Conclusions

In this study, we have proposed MSANLnet as a non-local network based on multi-scale attention for effective HDR image restoration. In the encoder part, implicit alignment of features is performed at various resolutions with multi-scale spatial attention modules. In the decoder part, image restoration is performed by adaptively incorporating useful information utilizing a long-range dependency with a non-local means-based fusion module. The results show that the proposed method exhibited better performance than existing deep learning methods. The results of this study have confirmed the importance and impact of the expansion of the receptive field in HDR image restoration based on CNN models. In the future, we plan to further improve the quality of HDR images by applying and developing vision transformer-based [94] modules with long-range dependencies.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/s22187044/s1, Table S1: Architecture of the Encoder. Table S2: Architecure of the Decoder.

Author Contributions

Conceptualization, H.Y. and Y.J.J.; methodology, H.Y. and Y.J.J.; software, H.Y.; validation, H.Y., S.M.N.U. and Y.J.J.; formal analysis, H.Y., S.M.N.U. and Y.J.J.; investigation, H.Y., and Y.J.J.; resources, Y.J.J.; data curation, H.Y.; writing—original draft preparation, H.Y., S.M.N.U. and Y.J.J.; writing—review and editing, S.M.N.U. and Y.J.J.; visualization, H.Y.; supervision, Y.J.J.; project administration, Y.J.J.; funding acquisition, Y.J.J. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Gachon University research fund of 2022 (GCU-202206860001).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Anderson, E.F.; McLoughlin, L. Critters in the classroom: A 3D computer-game-like tool for teaching programming to computer animation students. In Proceedings of the International Conference on Computer Graphics and Interactive Techniques, SIGGRAPH 2007, San Diego, CA, USA, 5–9 August 2007; Educators Program. Swanson, J., Ed.; ACM: New York, NY, USA, 2007; p. 7. [Google Scholar] [CrossRef]
  2. Rempel, A.G.; Heidrich, W.; Li, H.; Mantiuk, R. Video viewing preferences for HDR displays under varying ambient illumination. In Proceedings of the 6th Symposium on Applied Perception in Graphics and Visualization, APGV 2009, Chania, Crete, Greece, 30 September–2 October 2009; Mania, K., Riecke, B.E., Spencer, S.N., Bodenheimer, B., O’Sullivan, C., Eds.; ACM: New York, NY, USA, 2009; pp. 45–52. [Google Scholar] [CrossRef]
  3. Nayar, S.K.; Mitsunaga, T. High dynamic range imaging: Spatially varying pixel exposures. In Proceedings of the 2000 Conference on Computer Vision and Pattern Recognition (CVPR 2000), Hilton Head, SC, USA, 13–15 June 2000; Computer Society: Washington, DC, USA, 2000; pp. 1472–1479. [Google Scholar] [CrossRef]
  4. Tumblin, J.; Agrawal, A.K.; Raskar, R. Why I want a gradient camera. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR 2005), San Diego, CA, USA, 20–26 June 2005; Computer Society: Washington, DC, USA, 2005; pp. 103–110. [Google Scholar] [CrossRef]
  5. Reinhard, E.; Ward, G.; Pattanaik, S.N.; Debevec, P.E.; Heidrich, W. High Dynamic Range Imaging—Acquisition, Display, and Image-Based Lighting, 2nd ed.; Academic Press: Cambridge, MA, USA, 2010. [Google Scholar]
  6. Granados, M.; Ajdin, B.; Wand, M.; Theobalt, C.; Seidel, H.; Lensch, H.P.A. Optimal HDR reconstruction with linear digital cameras. In Proceedings of the The Twenty-Third IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2010, San Francisco, CA, USA, 13–18 June 2010; Computer Society: Washington, DC, USA, 2010; pp. 215–222. [Google Scholar] [CrossRef]
  7. Yan, Q.; Sun, J.; Li, H.; Zhu, Y.; Zhang, Y. High dynamic range imaging by sparse representation. Neurocomputing 2017, 269, 160–169. [Google Scholar] [CrossRef]
  8. Eden, A.; Uyttendaele, M.; Szeliski, R. Seamless image stitching of scenes with large motions and exposure differences. In Proceedings of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR 2006), New York, NY, USA, 17–22 June 2006; IEEE Computer Society: Washington, DC, USA, 2006; pp. 2498–2505. [Google Scholar] [CrossRef]
  9. Gallo, O.; Gelfandz, N.; Chen, W.C.; Tico, M.; Pulli, K. Artifact-free high dynamic range imaging. In Proceedings of the 2009 IEEE International Conference on Computational Photography (ICCP), San Francisco, CA, USA, 16–17 April 2009; pp. 1–7. [Google Scholar] [CrossRef]
  10. Granados, M.; Kim, K.I.; Tompkin, J.; Theobalt, C. Automatic noise modeling for ghost-free HDR reconstruction. ACM Trans. Graph. 2013, 32, 1–10. [Google Scholar] [CrossRef]
  11. Grosch, T. Fast and robust high dynamic range image generation with camera and object movement. In Proceedings of the Vision, Modeling and Visualization, RWTH Aachen, Aachen, Germany, 22–24 November 2006; pp. 277–284. [Google Scholar]
  12. Min, T.; Park, R.; Chang, S. Histogram based ghost removal in high dynamic range images. In Proceedings of the 2009 IEEE International Conference on Multimedia and Expo, ICME 2009, New York, NY, USA, 28 June–2 July 2009; pp. 530–533. [Google Scholar] [CrossRef]
  13. Pece, F.; Kautz, J. Bitmap movement detection: HDR for dynamic scenes. J. Virtual Real. Broadcast. 2013, 10, 1–13. [Google Scholar]
  14. Prabhakar, K.R.; Babu, R.V. Ghosting-free multi-exposure image fusion in gradient domain. In Proceedings of the 2016 IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2016, Shanghai, China, 20–25 March 2016; pp. 1766–1770. [Google Scholar] [CrossRef]
  15. Raman, S.; Chaudhuri, S. Reconstruction of high contrast images for dynamic scenes. Vis. Comput. 2011, 27, 1099–1114. [Google Scholar] [CrossRef]
  16. Wu, S.; Xie, S.; Rahardja, S.; Li, Z. A robust and fast anti-ghosting algorithm for high dynamic range imaging. In Proceedings of the International Conference on Image Processing, ICIP 2010, Hong Kong, China, 26–29 September 2010; pp. 397–400. [CrossRef]
  17. Zhang, W.; Cham, W. Reference-guided exposure fusion in dynamic scenes. J. Vis. Commun. Image Represent. 2012, 23, 467–475. [Google Scholar] [CrossRef]
  18. Tomaszewska, A.; Mantiuk, R. Image Registration for Multi-Exposure High Dynamic Range Image Acquisition. In Proceedings of the 15th International Conference in Central Europe on Computer Graphics, Visualization and Computer Vision, WSCG 2007, Bory, Czech Republic, 29 January–1 February 2007; pp. 49–56. [Google Scholar]
  19. Ward, G. Fast, robust image registration for compositing high dynamic range photographs from hand-held exposures. J. Graph. GPU Game Tools 2003, 8, 17–30. [Google Scholar] [CrossRef]
  20. Gallo, O.; Troccoli, A.J.; Hu, J.; Pulli, K.; Kautz, J. Locally non-rigid registration for mobile HDR photography. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition Workshops, CVPR Workshops 2015, Boston, MA, USA, 7–12 June 2015; IEEE Computer Society: Washington, DC, USA, 2015; pp. 48–55. [Google Scholar] [CrossRef]
  21. Hu, J.; Gallo, O.; Pulli, K. Exposure stacks of live scenes with hand-held cameras. In Proceedings of the Computer Vision—ECCV 2012—12th European Conference on Computer Vision, Florence, Italy, 7–13 October 2012; Lecture Notes in Computer Science. Fitzgibbon, A.W., Lazebnik, S., Perona, P., Sato, Y., Schmid, C., Eds.; Springer: Berlin/Heidelberg, Germany, 2012; Volume 7572, pp. 499–512. [Google Scholar] [CrossRef]
  22. Kang, S.B.; Uyttendaele, M.; Winder, S.A.J.; Szeliski, R. High dynamic range video. ACM Trans. Graph. 2003, 22, 319–325. [Google Scholar] [CrossRef]
  23. Zimmer, H.; Bruhn, A.; Weickert, J. Freehand HDR imaging of moving scenes with simultaneous resolution enhancement. Comput. Graph. Forum 2011, 30, 405–414. [Google Scholar] [CrossRef]
  24. Hu, J.; Gallo, O.; Pulli, K.; Sun, X. HDR deghosting: How to deal with saturation? In Proceedings of the 2013 IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; IEEE Computer Society: Washington, DC, USA, 2013; pp. 1163–1170. [Google Scholar] [CrossRef]
  25. Sen, P.; Kalantari, N.K.; Yaesoubi, M.; Darabi, S.; Goldman, D.B.; Shechtman, E. Robust patch-based hdr reconstruction of dynamic scenes. ACM Trans. Graph. 2012, 31, 1–11. [Google Scholar] [CrossRef]
  26. Yan, Q.; Gong, D.; Shi, Q.; van den Hengel, A.; Shen, C.; Reid, I.D.; Zhang, Y. Attention-guided network for ghost-free high dynamic range imaging. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2019, Long Beach, CA, USA, 16–20 June 2019; pp. 1751–1760. [Google Scholar] [CrossRef]
  27. Liu, Z.; Lin, W.; Li, X.; Rao, Q.; Jiang, T.; Han, M.; Fan, H.; Sun, J.; Liu, S. ADNet: Attention-guided deformable convolutional network for high dynamic range imaging. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, CVPR Workshops 2021, virtual, 19–25 June 2021; pp. 463–470. [Google Scholar] [CrossRef]
  28. Kalantari, N.K.; Ramamoorthi, R. Deep high dynamic range imaging of dynamic scenes. ACM Trans. Graph. 2017, 36, 1–12. [Google Scholar] [CrossRef]
  29. Wu, S.; Xu, J.; Tai, Y.; Tang, C. Deep high dynamic range imaging with large foreground motions. In Proceedings of the Computer Vision—ECCV 2018—15th European Conference, Munich, Germany, 8–14 September 2018; Proceedings, Part II; Lecture Notes in Computer Science. Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y., Eds.; Springer: Berlin/Heidelberg, Germany, 2018; Volume 11206, pp. 120–135. [Google Scholar] [CrossRef]
  30. Yan, Q.; Zhang, L.; Liu, Y.; Zhu, Y.; Sun, J.; Shi, Q.; Zhang, Y. Deep HDR imaging via a non-local network. IEEE Trans. Image Process. 2020, 29, 4308–4322. [Google Scholar] [CrossRef] [PubMed]
  31. Ma, K.; Duanmu, Z.; Zhu, H.; Fang, Y.; Wang, Z. Deep guided learning for fast multi-exposure image fusion. IEEE Trans. Image Process. 2020, 29, 2808–2819. [Google Scholar] [CrossRef] [PubMed]
  32. Tursun, O.T.; Akyüz, A.O.; Erdem, A.; Erdem, E. The state of the art in HDR deghosting: A survey and evaluation. Comput. Graph. Forum 2015, 34, 683–707. [Google Scholar] [CrossRef]
  33. Yan, Q.; Zhu, Y.; Zhang, Y. Robust artifact-free high dynamic range imaging of dynamic scenes. Multim. Tools Appl. 2019, 78, 11487–11505. [Google Scholar] [CrossRef]
  34. Hartley, R.; Zisserman, A. Multiple View Geometry in Computer Vision; Cambridge University Press: Cambridge, UK, 2004. [Google Scholar] [CrossRef] [Green Version]
  35. Baker, S.; Scharstein, D.; Lewis, J.P.; Roth, S.; Black, M.J.; Szeliski, R. A database and evaluation methodology for optical flow. Int. J. Comput. Vis. 2011, 92, 1–31. [Google Scholar] [CrossRef]
  36. Szpak, Z.L.; Chojnacki, W.; Eriksson, A.P.; van den Hengel, A. Sampson distance based joint estimation of multiple homographies with uncalibrated cameras. Comput. Vis. Image Underst. 2014, 125, 200–213. [Google Scholar] [CrossRef]
  37. Szpak, Z.L.; Chojnacki, W.; van den Hengel, A. Robust multiple homography estimation: An ill-solved problem. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 2132–2141. [Google Scholar]
  38. Yan, Q.; Gong, D.; Zhang, P.; Shi, Q.; Sun, J.; Reid, I.D.; Zhang, Y. Multi-scale dense networks for deep high dynamic range imaging. In Proceedings of the IEEE Winter Conference on Applications of Computer Vision, WACV 2019, Waikoloa Village, HI, USA, 7–11 January 2019; pp. 41–50. [Google Scholar] [CrossRef]
  39. Yan, Q.; Gong, D.; Shi, Q.J.; van den Hengel, A.; Shen, C.; Reid, I.D.; Zhang, Y. Dual-attention-guided network for ghost-free high dynamic range imaging. Int. J. Comput. Vis. 2022, 130, 76–94. [Google Scholar] [CrossRef]
  40. Madden, B.C. Extended Intensity Range Imaging; Technical Reports (CIS); University of Pennsylvania: Philadelphia, PA, USA, 1993; p. 248. [Google Scholar]
  41. Mann, S.; Picard, R. Beingundigital’with Digital Cameras; MIT Media Lab Perceptual: Cambridge, MA, USA, 1994; Volume 1, p. 2. [Google Scholar]
  42. Debevec, P.E.; Malik, J. Recovering high dynamic range radiance maps from photographs. In Proceedings of the 24th Annual Conference on Computer Graphics and Interactive Techniques, SIGGRAPH 1997, Los Angeles, CA, USA, 3–8 August 1997; ACM Press/Addison-Wesley Publishing Co.: New York, NY, USA; pp. 369–378. [Google Scholar] [CrossRef]
  43. Kao, W.; Hsu, C.; Chen, L.; Kao, C.; Chen, S. Integrating image fusion and motion stabilization for capturing still images in high dynamic range scenes. IEEE Trans. Consumer Electron. 2006, 52, 735–741. [Google Scholar] [CrossRef]
  44. Sidibé, D.; Puech, W.; Strauss, O. Ghost detection and removal in high dynamic range images. In Proceedings of the 17th European Signal Processing Conference, EUSIPCO 2009, Glasgow, Scotland, UK, 24–28 August 2009; pp. 2240–2244. [Google Scholar]
  45. Khan, E.A.; Akyüz, A.O.; Reinhard, E. Ghost removal in high dynamic range images. In Proceedings of the International Conference on Image Processing, ICIP 2006, Atlanta, GA, USA, 8–11 October 2006; pp. 2005–2008. [Google Scholar] [CrossRef]
  46. Bogoni, L. Extending dynamic range of monochrome and color images through fusion. In Proceedings of the 15th International Conference on Pattern Recognition, ICPR-2000, Barcelona, Spain, 3–7 September 2000; IEEE Computer Society: Washington, DC, USA, 2000; pp. 3007–3016. [Google Scholar] [CrossRef]
  47. Ma, K.; Li, H.; Yong, H.; Wang, Z.; Meng, D.; Zhang, L. Robust multi-exposure image fusion: A structural patch decomposition approach. IEEE Trans. Image Process. 2017, 26, 2519–2532. [Google Scholar] [CrossRef]
  48. Yu, G.; Zhang, J.; Ma, Z.; Wang, H. Efficient Progressive High Dynamic Range Image Restoration via Attention and Alignment Network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, CVPR Workshops 2022, New Orleans, LA, USA, 19–20 June 2022; pp. 1123–1130. [Google Scholar] [CrossRef]
  49. Tang, H.; Bai, S.; Sebe, N. Dual Attention GANs for Semantic Image Synthesis. In Proceedings of the MM ’20: The 28th ACM International Conference on Multimedia, Seattle, WA, USA, 12–16 October 2020; Chen, C.W., Cucchiara, R., Hua, X., Qi, G., Ricci, E., Zhang, Z., Zimmermann, R., Eds.; ACM: New York, NY, USA, 2020; pp. 1994–2002. [Google Scholar] [CrossRef]
  50. Lv, Z.; Li, X.; Niu, Z.; Cao, B.; Zuo, W. Semantic-shape Adaptive Feature Modulation for Semantic Image Synthesis. arXiv 2022, arXiv:2203.16898. [Google Scholar]
  51. Sun, W.; Wu, T. Learning Spatial Pyramid Attentive Pooling in Image Synthesis and Image-to-Image Translation. arXiv 2019, arXiv:1901.06322. [Google Scholar]
  52. Ren, Y.; Wu, Y.; Li, T.H.; Liu, S.; Li, G. Combining Attention with Flow for Person Image Synthesis. In Proceedings of the MM ’21: ACM Multimedia Conference, Virtual Event, 20–24 October 2021; Shen, H.T., Zhuang, Y., Smith, J.R., Yang, Y., Cesar, P., Metze, F., Prabhakaran, B., Eds.; ACM: New York, NY, USA, 2021; pp. 3737–3745. [Google Scholar] [CrossRef]
  53. Zheng, H.; Liao, H.; Chen, L.; Xiong, W.; Chen, T.; Luo, J. Example-Guided Image Synthesis Using Masked Spatial-Channel Attention and Self-supervision. In Proceedings of the Computer Vision—ECCV 2020—16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part XIV; Lecture Notes in Computer Science. Vedaldi, A., Bischof, H., Brox, T., Frahm, J., Eds.; Springer: Berlin/Heidelberg, Germany, 2020; Volume12359, pp. 422–439. [Google Scholar] [CrossRef]
  54. Uddin, S.M.N.; Jung, Y.J. SIFNet: Free-form image inpainting using color split-inpaint-fuse approach. Comput. Vis. Image Underst. 2022, 221, 103446. [Google Scholar] [CrossRef]
  55. Wang, N.; Li, J.; Zhang, L.; Du, B. MUSICAL: Multi-Scale Image Contextual Attention Learning for Inpainting. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, IJCAI 2019, Macao, China, 10–16 August 2019; Kraus, S., Ed.; pp. 3748–3754. [Google Scholar] [CrossRef]
  56. Qin, J.; Bai, H.; Zhao, Y. Multi-scale attention network for image inpainting. Comput. Vis. Image Underst. 2021, 204, 103155. [Google Scholar] [CrossRef]
  57. Wang, X.; Chen, Y.; Yamasaki, T. Spatially adaptive multi-scale contextual attention for image inpainting. Multim. Tools Appl. 2022, 81, 31831–31846. [Google Scholar] [CrossRef]
  58. Xie, B.; Yang, Z.; Yang, L.; Luo, R.; Wei, A.; Weng, X.; Li, B. Multi-Scale Fusion with Matching Attention Model: A Novel Decoding Network Cooperated with NAS for Real-Time Semantic Segmentation. IEEE Trans. Intell. Transp. Syst. 2022, 23, 12622–12632. [Google Scholar] [CrossRef]
  59. Wang, W.; Wang, S.; Li, Y.; Jin, Y. Adaptive multi-scale dual attention network for semantic segmentation. Neurocomputing 2021, 460, 39–49. [Google Scholar] [CrossRef]
  60. Chen, J.; Tian, Y.; Ma, W.; Mao, Z.; Hu, Y. Scale channel attention network for image segmentation. Multim. Tools Appl. 2021, 80, 16473–16489. [Google Scholar] [CrossRef]
  61. Sagar, A.; Soundrapandiyan, R. Semantic Segmentation with Multi Scale Spatial Attention For Self Driving Cars. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, ICCVW 2021, Montreal, BC, Canada, 11–17 October 2021; pp. 2650–2656. [Google Scholar] [CrossRef]
  62. Tao, A.; Sapra, K.; Catanzaro, B. Hierarchical Multi-Scale Attention for Semantic Segmentation. arXiv 2020, arXiv:2005.10821. [Google Scholar]
  63. Liu, M.; Zhang, C.; Zhang, Z. Multi-Scale Deep Convolutional Nets with Attention Model and Conditional Random Fields for Semantic Image Segmentation. In Proceedings of the 2nd International Conference on Signal Processing and Machine Learning, SPML 2019, Hangzhou, China, 27–29 November 2019; ACM: New York, NY, USA, 2019; pp. 73–78. [Google Scholar] [CrossRef]
  64. Yang, S.; Peng, G. Attention to Refine Through Multi Scales for Semantic Segmentation. In Proceedings of the Advances in Multimedia Information Processing—PCM 2018—19th Pacific-Rim Conference on Multimedia, Hefei, China, 21–22 September 2018; Proceedings, Part II; Lecture Notes in Computer Science. Hong, R., Cheng, W., Yamasaki, T., Wang, M., Ngo, C., Eds.; Springer: Berlin/Heidelberg, Germany, 2018; Volume 11165, pp. 232–241. [Google Scholar] [CrossRef]
  65. Zhang, M.; Wang, H.; Zhang, Z.; Chen, Z.; Shen, J. Lightweight Multi-Scale Asymmetric Attention Network for Image Super-Resolution. Micromachines 2021, 13, 54. [Google Scholar] [CrossRef]
  66. Mehta, N.; Murala, S. MSAR-Net: Multi-scale attention based light-weight image super-resolution. Pattern Recognit. Lett. 2021, 151, 215–221. [Google Scholar] [CrossRef]
  67. Pang, S.; Chen, Z.; Yin, F. Lightweight multi-scale aggregated residual attention networks for image super-resolution. Multim. Tools Appl. 2022, 81, 4797–4819. [Google Scholar] [CrossRef]
  68. Sun, Y.; Qin, J.; Gao, X.; Chai, S.; Chen, B. Attention-enhanced multi-scale residual network for single image super-resolution. Signal Image Video Process. 2022, 16, 1417–1424. [Google Scholar] [CrossRef]
  69. Li, W.; Li, J.; Li, J.; Huang, Z.; Zhou, D. A lightweight multi-scale channel attention network for image super-resolution. Neurocomputing 2021, 456, 327–337. [Google Scholar] [CrossRef]
  70. Wang, L.; Shen, J.; Tang, E.; Zheng, S.; Xu, L. Multi-scale attention network for image super-resolution. J. Vis. Commun. Image Represent. 2021, 80, 103300. [Google Scholar] [CrossRef]
  71. Liu, H.; Cao, F.; Wen, C.; Zhang, Q. Lightweight multi-scale residual networks with attention for image super-resolution. Knowl. Based Syst. 2020, 203, 106103. [Google Scholar] [CrossRef]
  72. Xiong, C.; Shi, X.; Gao, Z.; Wang, G. Attention augmented multi-scale network for single image super-resolution. Appl. Intell. 2021, 51, 935–951. [Google Scholar] [CrossRef]
  73. Soh, J.W.; Cho, N.I. Lightweight Single Image Super-Resolution with Multi-Scale Spatial Attention Networks. IEEE Access 2020, 8, 35383–35391. [Google Scholar] [CrossRef]
  74. Cao, F.; Liu, H. Single image super-resolution via multi-scale residual channel attention network. Neurocomputing 2019, 358, 424–436. [Google Scholar] [CrossRef]
  75. Zhang, X.; Wang, X. MARN: Multi-Scale Attention Retinex Network for Low-Light Image Enhancement. IEEE Access 2021, 9, 50939–50948. [Google Scholar] [CrossRef]
  76. Choi, S.; Cho, J.; Song, W.; Choe, J.; Yoo, J.; Sohn, K. Pyramid inter-attention for high dynamic range imaging. Sensors 2020, 20, 5102. [Google Scholar] [CrossRef]
  77. Chen, J.; Yang, Z.; Chan, T.N.; Li, H.; Hou, J.; Chau, L. Attention-Guided Progressive Neural Texture Fusion for High Dynamic Range Image Restoration. IEEE Trans. Image Process. 2022, 31, 2661–2672. [Google Scholar] [CrossRef] [PubMed]
  78. Xiao, J.; Ye, Q.; Liu, T.; Zhang, C.; Lam, K. Multi-scale Sampling and Aggregation Network For High Dynamic Range Imaging. arXiv 2022, arXiv:2208.02448. [Google Scholar] [CrossRef]
  79. Ye, Q.; Suganuma, M.; Xiao, J.; Okatani, T. Learning Regularized Multi-Scale Feature Flow for High Dynamic Range Imaging. arXiv 2022, arXiv:2207.02539. [Google Scholar] [CrossRef]
  80. Lee, J.; Shin, J.; Lim, H.; Paik, J. Deep High Dynamic Range Imaging without Motion Artifacts Using Global and Local Skip Connections. In Proceedings of the IEEE International Conference on Consumer Electronics, ICCE 2022, Las Vegas, NV, USA, 7–9 January 2022; pp. 1–2. [Google Scholar] [CrossRef]
  81. Hu, J.; Wu, L.; Li, N. High dynamic range imaging with short- and long-exposures based on artificial remapping using multiscale exposure fusion. J. Vis. Commun. Image Represent. 2022, 87, 103585. [Google Scholar] [CrossRef]
  82. Niu, Y.; Wu, J.; Liu, W.; Guo, W.; Lau, R.W.H. HDR-GAN: HDR image reconstruction from multi-exposed LDR images with large motions. IEEE Trans. Image Process. 2021, 30, 3885–3896. [Google Scholar] [CrossRef] [PubMed]
  83. Uddin, S.M.N.; Jung, Y.J. Global and Local Attention-Based Free-Form Image Inpainting. Sensors 2020, 20, 3204. [Google Scholar] [CrossRef] [PubMed]
  84. Nadim Uddin, S.M.; Ahmed, S.H.; Jung, Y.J. Unsupervised Deep Event Stereo for Depth Estimation. IEEE Trans. Circuits Syst. Video Technol. 2022, 1. [Google Scholar] [CrossRef]
  85. Ahmed, S.H.; Jang, H.W.; Uddin, S.M.N.; Jung, Y.J. Deep event stereo leveraged by event-to-image translation. In Proceedings of the Thirty-Fifth AAAI Conference on Artificial Intelligence, AAAI 2021, Thirty-Third Conference on Innovative Applications of Artificial Intelligence, AAAI 2021, The Eleventh Symposium on Educational Advances in Artificial Intelligence, EAAI 2021, Virtual Event, 2–9 February 2021; AAAI Press: Palo Alto, CA, USA, 2021; pp. 882–890. [Google Scholar]
  86. Jang, H.W.; Jung, Y.J. Deep color transfer for color-plus-mono dual cameras. Sensors 2020, 20, 2743. [Google Scholar] [CrossRef]
  87. Zhang, Y.; Tian, Y.; Kong, Y.; Zhong, B.; Fu, Y. Residual dense network for image super-resolution. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018, Salt Lake City, UT, USA, 18–22 June 2018; pp. 2472–2481. [Google Scholar] [CrossRef]
  88. Zhu, Z.; Xu, M.; Bai, S.; Huang, T.; Bai, X. Asymmetric non-local neural networks for semantic segmentation. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision, ICCV 2019, Seoul, Korea, 27 October–2 November 2019; pp. 593–602. [Google Scholar] [CrossRef]
  89. Wang, X.; Girshick, R.B.; Gupta, A.; He, K. Non-local neural networks. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7794–7803. [Google Scholar] [CrossRef]
  90. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. In Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  91. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. PyTorch: An imperative style, high-performance deep learning library. In Proceedings of the Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, NeurIPS 2019, Vancouver, BC, Canada, 8–14 December 2019; pp. 8024–8035. [Google Scholar]
  92. Wang, X.; Chan, K.C.K.; Yu, K.; Dong, C.; Loy, C.C. EDVR: Video restoration with enhanced deformable convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, CVPR Workshops 2019, Computer Vision Foundation/IEEE, Long Beach, CA, USA, 16–20 June 2019; pp. 1954–1963. [Google Scholar] [CrossRef]
  93. Zheng, J.; Li, Z.; Zhu, Z.; Wu, S.; Rahardja, S. Hybrid Patching for a Sequence of Differently Exposed Images with Moving Objects. IEEE Trans. Image Process. 2013, 22, 5190–5201. [Google Scholar] [CrossRef]
  94. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. In Proceedings of the 9th International Conference on Learning Representations, ICLR 2021, Virtual Event, 3–7 May 2021. [Google Scholar]
Figure 1. Visual comparison of HDR image restoration results obtained using the proposed method and state-of-the-art methods. From left, Wu [28], AHDR [26], AD [27], NHDRR [30], DAHDR [39], Proposed MSANLnet, and Ground Truth.
Figure 1. Visual comparison of HDR image restoration results obtained using the proposed method and state-of-the-art methods. From left, Wu [28], AHDR [26], AD [27], NHDRR [30], DAHDR [39], Proposed MSANLnet, and Ground Truth.
Sensors 22 07044 g001
Figure 2. Overview of the proposed MSANLnet for HDR image reconstruction.
Figure 2. Overview of the proposed MSANLnet for HDR image reconstruction.
Sensors 22 07044 g002
Figure 3. Overview of a spatial attention block.
Figure 3. Overview of a spatial attention block.
Sensors 22 07044 g003
Figure 4. Visualization of spatial attention maps. The attention maps for the first scale are shown in (a,b). The attention maps for the second scale are shown in (c,d).
Figure 4. Visualization of spatial attention maps. The attention maps for the first scale are shown in (a,b). The attention maps for the second scale are shown in (c,d).
Sensors 22 07044 g004
Figure 5. A residual dense convolutional block is composed of dense concatenations based on skip-connections between a series of convolutional blocks consisting of convolutions and ReLU activation functions.
Figure 5. A residual dense convolutional block is composed of dense concatenations based on skip-connections between a series of convolutional blocks consisting of convolutions and ReLU activation functions.
Sensors 22 07044 g005
Figure 6. Structure of the asymmetric pyramidal non-local block with spatial pyramid pooling [88].
Figure 6. Structure of the asymmetric pyramidal non-local block with spatial pyramid pooling [88].
Sensors 22 07044 g006
Figure 7. Qualitative comparison of HDR result images of the proposed method (MSANLnet) and state-of-the-art models. From left, Wu [28], AHDR [26], AD [27], NHDRR [30], DAHDR [39], Proposed MSANLnet, and Ground Truth.
Figure 7. Qualitative comparison of HDR result images of the proposed method (MSANLnet) and state-of-the-art models. From left, Wu [28], AHDR [26], AD [27], NHDRR [30], DAHDR [39], Proposed MSANLnet, and Ground Truth.
Sensors 22 07044 g007
Figure 8. Visual comparison of HDR images restored from the dataset provided in [28]. From left, Wu [28], AHDR [26], NHDRR [30], AD [27], DAHDR [39], Proposed MSANLnet, and Ground Truth.
Figure 8. Visual comparison of HDR images restored from the dataset provided in [28]. From left, Wu [28], AHDR [26], NHDRR [30], AD [27], DAHDR [39], Proposed MSANLnet, and Ground Truth.
Sensors 22 07044 g008
Figure 9. Qualitative comparison of the performance of the proposed method (MSANLnet) and its baseline variants.
Figure 9. Qualitative comparison of the performance of the proposed method (MSANLnet) and its baseline variants.
Sensors 22 07044 g009
Figure 10. Visual comparison of degradation of HDR restoration for global motion (i.e., when applied translation). From left, Wu [28], AHDR [26], NHDRR [30], AD [27], DAHDR [39], and Proposed MSANLnet.
Figure 10. Visual comparison of degradation of HDR restoration for global motion (i.e., when applied translation). From left, Wu [28], AHDR [26], NHDRR [30], AD [27], DAHDR [39], and Proposed MSANLnet.
Sensors 22 07044 g010
Figure 11. Visual comparison of additional HDR images from the dataset provided in [93]. From left, Wu [28], AHDR [26], NHDRR [30], AD [27], DAHDR [39], and Proposed MSANLnet.
Figure 11. Visual comparison of additional HDR images from the dataset provided in [93]. From left, Wu [28], AHDR [26], NHDRR [30], AD [27], DAHDR [39], and Proposed MSANLnet.
Sensors 22 07044 g011
Table 1. Quantitative comparison of the performance of the proposed method (MSANLnet) and state-of-the-art models.
Table 1. Quantitative comparison of the performance of the proposed method (MSANLnet) and state-of-the-art models.
Method/MetricPSNR-lPSNR- μ SSIM ↑
NHDRR [30]33.697039.24610.9382
Wu [28]36.426641.98380.9618
AHDR [26]38.094839.59370.9757
AD [27]37.928740.81020.9812
DAHDR [39]36.591540.07080.9753
MSANL (proposed)40.437042.74660.9821
Table 2. Quantitative comparison of the performance of the proposed method (MSANLnet) and its baseline variants.
Table 2. Quantitative comparison of the performance of the proposed method (MSANLnet) and its baseline variants.
Method/MetricPSNR-lPSNR- μ SSIM
MSAnet37.195839.56460.9699
ANLnet39.019340.68100.9781
MSANLnet (Proposed)40.437042.74660.9821
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yoon, H.; Uddin, S.M.N.; Jung, Y.J. Multi-Scale Attention-Guided Non-Local Network for HDR Image Reconstruction. Sensors 2022, 22, 7044. https://doi.org/10.3390/s22187044

AMA Style

Yoon H, Uddin SMN, Jung YJ. Multi-Scale Attention-Guided Non-Local Network for HDR Image Reconstruction. Sensors. 2022; 22(18):7044. https://doi.org/10.3390/s22187044

Chicago/Turabian Style

Yoon, Howoon, S. M. Nadim Uddin, and Yong Ju Jung. 2022. "Multi-Scale Attention-Guided Non-Local Network for HDR Image Reconstruction" Sensors 22, no. 18: 7044. https://doi.org/10.3390/s22187044

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop