Next Article in Journal
Exploring Multi-Channel GPS Receivers for Detecting Spoofing Attacks on UAVs Using Machine Learning
Previous Article in Journal
Driving Pattern Analysis, Gear Shift Classification, and Fuel Efficiency in Light-Duty Vehicles: A Machine Learning Approach Using GPS and OBD II PID Signals
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Single-Image Super-Resolution via Cascaded Non-Local Mean Network and Dual-Path Multi-Branch Fusion

School of Computer and Control Engineering, Yantai University, Yantai 264005, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(13), 4044; https://doi.org/10.3390/s25134044 (registering DOI)
Submission received: 7 May 2025 / Revised: 26 June 2025 / Accepted: 26 June 2025 / Published: 28 June 2025
(This article belongs to the Section Sensing and Imaging)

Abstract

Image super-resolution (SR) aims to reconstruct high-resolution (HR) images from low-resolution (LR) inputs. It plays a crucial role in applications such as medical imaging, surveillance, and remote sensing. However, due to the ill-posed nature of the task and the inherent limitations of imaging sensors, obtaining accurate HR images remains challenging. While numerous methods have been proposed, the traditional approaches suffer from oversmoothing and limited generalization; CNN-based models lack the ability to capture long-range dependencies; and Transformer-based solutions, although effective in modeling global context, are computationally intensive and prone to texture loss. To address these issues, we propose a hybrid CNN–Transformer architecture that cascades a pixel-wise self-attention non-local means module (PSNLM) and an adaptive dual-path multi-scale fusion block (ADMFB). The PSNLM is inspired by the non-local means (NLM) algorithm. We use weighted patches to estimate the similarity between pixels centered at each patch while limiting the search region and constructing a communication mechanism across ranges. The ADMFB enhances texture reconstruction by adaptively aggregating multi-scale features through dual attention paths. The experimental results demonstrate that our method achieves superior performance on multiple benchmarks. For instance, in challenging ×4 super-resolution, our method outperforms the second-best method by 0.0201 regarding the Structural Similarity Index (SSIM) on the BSD100 dataset. On the texture-rich Urban100 dataset, our method achieves a 26.56 dB Peak Signal-to-Noise Ratio (PSNR) and 0.8133 SSIM.

1. Introduction

Image resolution refers to the amount of spatial detail captured in an image, typically quantified by the number of pixels per unit area. Higher resolution enables finer visual details to be distinguished, which is essential for accurate interpretation and analysis in various computer vision and image processing tasks. In many real-world applications—such as medical diagnostics, remote sensing, security surveillance, and autonomous driving—HR images are important. They help to identify subtle patterns, detect small objects, and improve the reliability of downstream algorithms [1,2,3,4]. For instance, in medical imaging, a higher resolution can reveal microstructures that are vital for early disease detection; in autonomous driving, it can enhance object recognition accuracy for better decision-making.
However, directly capturing HR images through physical sensors remains challenging due to multiple factors. These include limitations in sensor size and pixel density, optical aberrations and diffraction, motion-induced blur, bandwidth constraints, storage requirements, and cost [5,6,7]. For example, miniaturized imaging devices often rely on compact sensors with small pixel sizes, which restrict light sensitivity and increase noise. Optical imperfections further degrade image sharpness, while high-speed scenes may introduce motion blur. Additionally, transmitting or storing HR images is often infeasible in resource-constrained environments. Due to these limitations, many imaging systems produce low-LR outputs, limiting their effectiveness in practical applications.
To address this problem, SR techniques have been developed as computational alternatives to hardware-based solutions. SR aims to reconstruct HR images from one or more LR observations, enhancing visual fidelity and structural details without changing the image acquisition hardware. Compared to sensor upgrades, SR methods are more cost-effective and flexible, and have become an essential preprocessing step for tasks like object detection [8], segmentation [9], and scene understanding [5].
The early SR methods addressed this ill-posedness by introducing prior knowledge of natural image statistics into the reconstruction process [10]. These approaches typically added regularization terms to the objective function to favor plausible solutions [11,12]. The common regularization techniques include L1 regularization (encouraging sparsity in the image gradient), L0 regularization (promoting even sparser gradients), mixed hybrid L1–L0 regularization, and sparse representation (assuming that image patches can be represented as a sparse linear combination of atoms from a learned dictionary) [11,13]. While these methods offered some degree of improvement in image resolution, they often failed to capture the complex nonlinear patterns present in natural scenes. This limitation frequently resulted in artifacts such as blurring, ringing, and the loss of fine textures.
The introduction of deep learning, particularly convolutional neural networks (CNNs), advanced the performance of SISR. CNNs can learn hierarchical features and model complex nonlinear mappings. They have achieved much better results than traditional approaches [14,15,16,17,18,19,20,21,22,23], largely because they can automatically extract informative patterns from large datasets. CNNs also exploit spatial correlations within small image regions, making them effective for local feature extraction [14,16,18,19]. Due to their relatively small number of parameters compared to fully connected networks, CNNs are also more computationally efficient. However, a limitation of standard CNNs is their limited receptive field. This limits their ability to capture long-range dependencies, which are important for understanding global context, identifying repeated textures, and reconstructing structured objects.
To overcome this, researchers have turned to Transformers. Their multi-head self-attention mechanism allows each pixel to interact with all the other pixels in the image, regardless of spatial distance [24,25,26,27,28]. This enables modeling of long-range dependencies and global structure. Despite their strengths, Transformers face scalability issues. The self-attention mechanism has quadratic complexity with respect to image size, making standard Transformers computationally expensive. This high cost requires large training datasets and powerful Graphics Processing Units (GPUs), which can limit their usability in practical scenarios, particularly on resource-constrained devices. Although many lightweight Transformer variants have been proposed to mitigate these computational costs [29,30,31,32], they often trade performance for efficiency. As a result, they may struggle to reconstruct fine details and textures.
Given the respective strengths and limitations of CNNs and Transformers, hybrid CNN–Transformer architectures have emerged as a promising research direction for SISR. However, the existing hybrid methods still struggle to effectively integrate long-range dependencies with fine-grained local details. To address this gap, we propose a novel hybrid architecture called PSNLMN. The key innovations of our approach are as follows:
  • We introduce a novel PSNLM module, inspired by the traditional NLM algorithm, which effectively captures fine-grained long-range dependencies by combining the strengths of NLM and self-attention mechanisms;
  • We design an ADMFB that enhances both scale diversity and feature discriminability by simultaneously extracting and fusing hierarchical features through parallel attention pathways;
  • We integrate a PSNLM and ADMFB into a CNN framework, achieving competitive performance on standard benchmarks, demonstrating the effectiveness of our hybrid design.
To the best of our knowledge, this is the first attempt to combine the principles of NLM with self-attention mechanisms for the task of image super-resolution. To clearly present our approach and findings, the remainder of this paper is organized as follows: Section 1 introduces the background, motivation, and contributions of the study. Section 2 provides a detailed review of the related work, including the existing SISR methods and their limitations. Section 3 elaborates on our proposed PSNLMN, including the network framework and the design details and principles of the NLMB, PSNLM, and ADMFB. Section 4 describes the experimental settings, ablation studies, and performance evaluations. Section 5 concludes the paper by summarizing the proposed method and contributions.

2. Related Work

SR has been an active research area for decades, with a wide range of approaches proposed to address this challenging problem. These methods are typically divided into two categories: traditional approaches and deep learning-based methods.

2.1. Traditional Methods

Traditional SR techniques can be broadly categorized into interpolation-based, reconstruction-based, and example-based methods. Interpolation-based methods, such as bicubic interpolation [33] and Lanczos resampling [34], are computationally efficient. However, they assume local continuity in the image signal, which often leads to oversmoothing. Consequently, these methods tend to blur edges and lose high-frequency details [35], limiting their ability to preserve fine textures.
Reconstruction-based methods were developed to address the shortcomings of interpolation. They formulate SISR as an inverse problem and incorporate image priors to regularize the solution. These priors are introduced through objective functions that balance data fidelity with regularization terms, such as Tikhonov regularization, Total Variation (TV), or sparsity-promoting norms like l 1 and l 2 minimization [12,36,37]. Although these methods often produce sharper results, they are sensitive to the choice in regularization parameters, which can be difficult to tune and may not generalize well to different image types [38]. Moreover, the use of hand-crafted priors limits their ability to capture the complex and nonlinear structures found in natural images, often resulting in artifacts and limited texture reconstruction.
Example-based SR represents a learning-based alternative. These methods attempt to learn mappings from low-resolution to high-resolution patches using either external dictionaries (e.g., sparse coding [13]) or internal image similarities (e.g., anchored neighborhood regression [39]). Although capable of producing visually pleasing details, example-based methods face two main challenges: (1) high computational cost during patch matching and (2) strong dependence on the quality and relevance of the training data or the presence of repetitive patterns within the input image.

2.2. Deep Learning-Based Super-Resolution Methods

The development of deep learning has brought significant progress to the field of SR. The methods based on deep learning can be broadly divided into three categories: CNN-based, Transformer-based, and hybrid CNN–Transformer approaches.

2.2.1. CNN-Based Methods

Dong et al. [14] proposed the Super-Resolution Convolutional Neural Network (SRCNN), which was the first to apply CNNs to SISR. The SRCNN uses a shallow three-layer network to learn a nonlinear mapping from LR to HR image patches. Although it outperforms traditional methods, the SRCNN is limited by its small receptive field, which hinders its ability to capture long-range dependencies within images.
To improve performance, later works introduced deeper CNN architectures. Kim et al. [40] presented the Very Deep Super-Resolution (VDSR) network, which uses a 20-layer structure with residual learning to enhance reconstruction accuracy and ease training. Tai et al. [41] proposed the Deep Recursive Residual Network (DRRN), which employs recursive blocks to deepen the network without significantly increasing the number of parameters.
Deeper networks generally improve reconstruction quality, but they can also be harder to train. To address this and to enhance feature representation, attention mechanisms have been incorporated. Zhang et al. [19] introduced the Residual Channel Attention Network (RCAN), which applies channel attention to adaptively reweight feature maps. This helps the network to focus on more informative channels.
CNN-based methods are effective in extracting local features. However, due to the localized nature of convolution, they face challenges in modeling long-range dependencies, which are important for capturing global image context.

2.2.2. Transformer-Based Methods

Motivated by the success of Transformers in natural language processing, researchers have adapted them for image processing tasks, including SR. Chen et al. [42] proposed the Image Processing Transformer (IPT), a large-scale pre-trained model that applies the Transformer architecture to various low-level vision problems. IPT demonstrated strong performance in SR tasks. However, the computational cost of the standard Transformer scales quadratically with image resolution, making it both memory- and computation-intensive.
To mitigate this, Liang et al. [26] introduced SwinIR, which is built on the Swin Transformer [25]. It employs shifted windows to reduce computational complexity while retaining the ability to model long-range dependencies. In addition, several lightweight Transformer architectures have been proposed to further reduce resource demands. Despite improvements in efficiency, these models often face trade-offs in reconstruction quality, especially in recovering fine details and textures.

2.2.3. Hybrid CNN–Transformer Methods

To leverage the strengths of both CNNs and Transformers, recent studies have explored hybrid architectures that combine local feature extraction with global context modeling. These approaches aim to integrate the localized processing capabilities of CNNs with the broader receptive fields of Transformers. This combination enhances reconstruction quality by addressing both local and global information.
For instance, the ESRT [31] embedded Transformer modules within a CNN framework to balance computational efficiency with contextual awareness. LBNet [32] introduced bimodal blocks that capture both local patterns and non-local dependencies. Methods such as SwinIR [26] and its variants [28] have shown good performance on several benchmark datasets, highlighting the potential of hybrid designs.
Despite these advancements, many hybrid models still struggle with the reconstruction of high-frequency details and complex textures. Although Transformers help in capturing global structure, the current fusion mechanisms or attention designs may not fully preserve local fine-grained features. As a result, reconstructed images can lack detail clarity, particularly in areas containing dense textures or repetitive patterns.
Therefore, there is a continued need for hybrid architectures that better integrate global and local information. Such designs should aim to improve the fidelity of texture and detail reconstruction without compromising efficiency.

3. Methods

3.1. Network Structure

The proposed network framework is illustrated in Figure 1a. It consists of two main stages: feature extraction and image reconstruction. The feature extraction stage is further divided into shallow feature extraction and deep feature extraction.
Shallow Feature Extraction: Initially, the input image I L R R H × W × C i n undergoes shallow feature extraction using a single convolutional layer with a kernel size of 9 × 9 . This layer aims to capture low-level spatial information and enhance the initial representation of the input image. The process can be represented as
F s = σ p r l e u ( C 9 × 9 ( I L R ) )
where F s R H × W × C denotes the shallow features, σ p r l e u is the PReLU activation function, and C 9 × 9 ( · ) represents a convolutional operation with a 9 × 9 kernel. The 9 × 9 kernel provides a large receptive field for capturing contextual information at an early stage.
Deep Feature Extraction: The shallow features F s serve as input to the deep feature extraction module. This module adopts a cascaded structure of k non-local multi-branch modules (NLMBs). These modules are designed to effectively capture both global and local features at multiple scales. The deep feature extraction can be formulated as
F d = C 3 × 3 ( N L M B k ( N L M B k 1 ( N L M B 1 ( F s ) ) ) )
where N L M B i represents the i-th NLMB block in the cascade and C 3 × 3 ( · ) represents the 3 × 3 convolutional operation. The output feature F d is then combined with the input shallow features through a residual connection:
F d = F s + F d
where F d R H × W × C is the final output of the deep feature extraction stage. This residual connection facilitates gradient flow during training and allows the network to learn residual information, further enhancing the feature representation.
Image Reconstruction: The image reconstruction stage focuses on upsampling deep feature maps F d to the desired output resolution. This is achieved through the use of upsampling block (UPB), as detailed in Figure 1b. The upsampling operation can be represented as
F u p = U P B ( F d ) = σ p r l e u ( P i x e l S h u f f l e ( C 3 × 3 ( F d ) ) )
where P i x e l S h u f f l e operation efficiently upsamples feature maps by rearranging channels into spatial resolution [43], F u p R ( r H ) × ( r W ) × C is the output of the UPB, and r is the upscaling factor for super-resolution.
Following the upsampling process, a final convolutional layer with a 9 × 9 kernel is applied to refine the reconstructed image. This final convolutional layer mirrors the initial shallow feature extraction layer, allowing for effective feature fusion and adaptation of the reconstructed features to the desired output characteristics:
I H R = C 9 × 9 ( F u p )
where I H R R ( r H ) × ( r W ) × C i n is the final reconstructed HR image.

3.2. The Non-Local Multi-Branch Module (NLMB)

The NLMB, depicted in Figure 2, is a core component designed to extract and fuse both global and local features. The module consists of three primary parts: initial feature extraction, PSNLM, and ADMFB.
In initial feature extraction, input features X R H × W × C are processed through a 3 × 3 convolutional layer and followed by batch normalization (BN) and a PReLU activation function to enhance the representation, as shown in Equation (6). This initial processing provides a feature foundation for subsequent stages. The resulting feature map is then fed into a PSNLM, which captures long-range dependencies and global contextual information; more details of PSNLM will be given in Section 3.3. The output F P S N L M of the PSNLM block is combined with the initial feature map using a residual connection and is followed by layer normalization (LN) to stabilize the training, as described by Equation (7).
F i n = σ p r l e u ( B N ( C 3 × 3 ( X ) ) )
F L N = L N ( X + F P S N L M )
F N L M B = F A D M F B + F L N
In the final stage, the normalized features are fed into the ADMFB block (further details will be provided in Section 3.4), which extracts and fuses multi-scale features across branches. A final skip connection refines the module output, as shown in Equation (8), where F N L M B R H × W × C represents the final output feature map of the NLMB module. This skip connection allows the module to learn residual features and enhance the output representation.

3.3. Pixel-Wise Self-Attention-Based Non-Local Mean Module (PSNLM)

The architecture of PSNLM is illustrated in Figure 3. By integrating NLM with pixel-wise attention, it enables efficient modeling of long-range dependencies. Traditional NLM algorithms perform a pixel-wise search across the entire image to compute similarities, resulting in substantial computational redundancy and high complexity. This is particularly problematic for high-resolution images. To mitigate this, we adopt a window-based partitioning strategy that limits the search scope, reducing computational burden and enabling efficient high-resolution processing. Furthermore, the use of fixed-size windows facilitates parallel computation, improving the overall efficiency of the algorithm.
Specifically, as illustrated in Figure 4, the input feature map F i n is initially divided into n s w non-overlapping equally sized sub-feature maps W s = { W s w | W s w R s w × s w × C } via a window partitioning operation. Each sub-feature map, of size s w × s w × C , is treated as an independent search window, where s w represents the search window size.
Within each search window, we adapt the core concept of NLM, where similarities are assessed based on the similarity between local neighborhoods centered around individual pixels. We draw inspiration from the Transformer architecture by incorporating learned embeddings for these neighborhoods. Specifically, we first define a local neighborhood window W n w R n w × n w × C . Depth-wise convolutional and point-wise convolutional operations are then employed to extract features from these neighborhoods. The embedding for each neighborhood patch is obtained as follows:
f w j = F ( σ l ( C p o i n t ( C d e p t h ( W n w j ) ) ) )
where F is the flatten operation, σ l denotes the LeakyReLU activation function, C p o i n t represents a point-wise 1 × 1 convolutional operation, C d e p t h represents a depth n w × n w convolutional operation, and f w j R 1 × d is the embedded features of the j-th neighborhood patch, where d represents the number of kernels in the convolution operation, and the maximum value of j is s w × s w , which reflects the number of domain blocks within the search window.
For the center point X j R 1 × 1 × C within each neighborhood patch, we extract a feature vector using a point-wise convolutional operation:
c j = F ( σ l ( C p o i n t ( x j ) ) )
where c j R 1 × C is the embedding vector representing the center point.
Given a query point X q within a search window W s w l W s , the similarity computation proceeds as follows: the embedding vector f w q for the neighborhood patch and the embedding vector c q for the center point X q are computed using Equations (9) and (10), respectively. Notably, zero-padding with a width of n w 2 is applied around the search window to ensure that query points positioned at or near the window boundaries maintain a complete n w × n w neighborhood, thereby guaranteeing robust algorithm execution and processing. Similarly, we extract the features W k = { f w j } for all patches within W s w l using Equation (9), with corresponding center points W v = { c j } . Next, these embedded features are distributed across n attention heads, as follows:
f w q = [ f w q 1 , f w q 2 , , f w q n ] , W k = [ W k 1 , W k 2 , , W k n ] , W v = [ W v 1 , W v 2 , , W v n ]
where f w q i R 1 × d n , W k i , W v i R ( s w · s w ) × d n , and the superscript i denotes the i-th attention head. The similarity between all points C = { X j | j = 1 , 2 , , s w · s w } in W s w i and the query point X q is then computed using a dot product operation:
s i ( x q , C ) = s i ( f w q , W k ) = S o f t m a x ( f w q i · ( W k i ) T d / n )
where d / n is the scaling factor, S o f t m a x is the normalized function, and s i R 1 × ( s w · s w ) represents the similarity score between the query point and all pixels. Then, we weight W v by similarity as follows:
o i = s i ( x q , C ) · W v = j = 1 s w × s w s ( x q , x j ) · c j
where O i is the weighted output of the i-th attention head. Subsequently, we fuse the multiple heads using a set of weights to obtain the final attention output:
c q o u t = C o n c a t ( o 1 , o 2 , , o i , , o n ) W o
where c q o u t R 1 × C , C o n c a t denotes the concatenation function employed to concatenate the outputs from multiple attention heads along a specified dimension, and W o is a linear layer function used to aggregate the outputs of multiple heads and subsequently map them back to the original channel dimension. Finally, a spatial restoration transformation F 1 is utilized to reconstruct the output x q o u t R 1 × 1 × C to its original spatial dimensions:
x q o u t = F 1 ( c q o u t )
By iterating this process for all query points within the search window, we obtain the output feature map W s o u t for each window. These feature maps are then normalized using layer normalization. Next, they are processed by a feedforward network consisting of convolutional layers and a DropPath layer, producing an intermediate result. The normalization stabilizes the feature distribution, while the convolutional layers capture local spatial patterns. The DropPath mechanism is incorporated to enhance regularization and prevent overfitting.
To enable information exchange between different windows, we adopt the Shifted Windows Multi-Head Self-Attention (SW-MSA) mechanism [25]. The output of SW-MSA is further processed by another sequence of layer normalization and a convolutional feedforward network, generating the final feature maps. This repeated structure maintains consistency across the network. Finally, the output feature maps of all search windows are stitched, as shown in Figure 4, forming a global feature map with the same spatial dimensions as the input.
Note that we use the dot product in the attention mechanism to measure similarity instead of the nonlinear Gaussian kernel (used in NLM algorithm). This choice is based on the observation that the features undergo multiple nonlinear transformations throughout the network, allowing the dot product to implicitly model complex relationships.

3.4. Adaptive Dual-Path Multi-Branch Fusion Block (ADMFB)

To improve multi-scale feature representation in SR, we propose ADMFB, a novel architecture that aggregates features at different scales. As depicted in Figure 5, the ADMBFM comprises two components: a multi-branch feature extraction (MBFE) block and an adaptive dual-path fusion (ADPF) block. The MBFE uses multiple branches with dilated convolutions at different rates. This design captures contextual information across scales, addressing scale variation in image processing. The subsequent ADPF module is responsible for adaptively fusing the features extracted by the MBFE unit. This fusion is achieved through a dual-path approach to leverage complementary information. The specific details are as follows:
The MBFE begins with distributing the input feature Y R H × W × C to N parallel branches. Each branch extracts scale-specific features using dilated convolutions Y i = ϕ ( Y ; d i ) , where Y i R H × W × C , d i represents the dilation rate of the i-th branch and ϕ ( · ) describes dilated convolution operation. The multi-branch structure generates feature maps enriched with diverse scales.
In ADPF, the feature maps captured from MFBE are sent to average pooling and maximum pooling layers; the results are addressed as P a v g and P m a x R 1 × 1 × C , respectively. P a v g and P m a x are then passed through two parallel attention generation paths to compute adaptive weights:
Path 1: The global features are processed by two shared 1 × 1 convolution layers, followed by the element-wise addition and an activation function, generating channel attention weights:
W i ( 1 ) = σ f 1 b f 1 a P i a v g + f 1 b f 1 a P i m a x
where f 1 a ( · ) and f 1 b ( · ) represent two distinct sequential 1 × 1 convolution operations. The former reduces the channel dimensionality to C / r 1 , while the latter expands it back to C, W i ( 1 ) R 1 × 1 × C , and σ ( · ) denotes the sigmoid activation function. This path employs channel attention mechanisms to selectively weight each branch of the MBFE, thereby emphasizing the informative features and suppressing irrelevant noise.
Path 2: The pooled features are first concatenated and then compressed through a 1 × 1 convolution. Subsequently, an additional convolution is applied, followed by normalization using the softmax function to generate another set of weights.
W i ( 2 ) = f 2 b f 2 a C o n c a t P i a v g , P i m a x j = 1 N f 2 b f 2 a C o n c a t P j a v g , P j m a x
where W i ( 2 ) R 1 × 1 × C is the result of normalizing with respect to the other weights, N denotes the number of branches, f 2 a ( · ) and f 2 b ( · ) represent other group 1 × 1 convolution operations, the first 1 × 1 convolution reduces the channel dimension from 2 C to 2 C r 2 , and the second 1 × 1 convolution reduces it to C. Path 2 is designed to calculate fusion weights that govern the contribution of each branch during feature aggregation.
Ultimately, the weights from the two paths are multiplied ( W i = W i ( 1 ) W i ( 2 ) ) to compute the final attention weight for each branch, producing the weighted feature Y i = Y i W i . A final 3 × 3 convolution is applied to further refine and integrate the details, and to generate the output feature Y o u t p u t = C 3 × 3 ( i = 1 N Y i ) .
In summary, ADMFB allows the block to dynamically prioritize branches based on their relevance to the specific input, mitigating the limitations of static fusion strategies. By combining these two paths, the ADMFB effectively achieves a more nuanced and context-aware fusion process, leading to enhanced feature representation and improved performance in reconstruct tasks. This dual-path design combines channel-wise selection and adaptive branch weighting, enabling the network to model dependencies across scales.

4. Experiments

4.1. Experimental Settings

We trained the PSNLMN on an NVIDIA RTX 3090 GPU using 120K randomly sampled images from the CoCo2014 dataset [44]. To evaluate model performance, we conducted experiments on four widely used benchmark datasets: Set5 [45], Set14 [46], BSD100 [47], and Urban100 [48]. The super-resolution performance was assessed by calculating the PSNR and SSIM on the luminance (Y) channel. The detailed training configuration and parameter settings are described as follows:
During the training, LR images were generated by downsampling HR images using bicubic interpolation, followed by data augmentation through random horizontal flips and orthogonal rotations (90° and 270°). LR patches were extracted at 48 × 48 pixel resolution. Network parameters were initialized via Kaiming initialization with an initial learning rate of 0.0002. Optimization employed the AdamW optimizer with momentum coefficients β 1 = 0.9 and β 2 = 0.999 , minimizing the L 2 loss function. Models were trained for 300 epochs with a batch size of 64, maintaining a constant learning rate for the initial 150 epochs before applying linear decay. After each epoch, model performance was validated and weights archived. The optimal model was selected post-training through comprehensive evaluation of loss convergence trajectories and validation metrics.
Testing comprised (1) loading preprocessed LR test data into the best-trained model; (2) generating super-resolved outputs; and (3) computing quantitative metrics against ground-truth HR references. A batch size of 1 was implemented during testing to eliminate spatial variance in metric computation.
The proposed network employs a base configuration with 64 channels (C = 64), incorporating 4 attention heads, an 8 × 8 search window, and a neighborhood window size of 3. The ADFB block utilizes reduction factors r = 16. For × 4 super-resolution, the architecture stacks 10 NLMBs, where the ADMFB features three parallel branches with dilation rates of 1, 2, and 3, respectively. When the scaling factor is reduced to × 2 or × 3 , we adjust the number of NLMBs to 12 and expand the ADMFB to four parallel branches with dilation rates of 1, 2, 3, and 4, thereby enhancing multi-scale feature extraction capabilities.

4.2. Ablation Studies

In this section, we conduct ablation studies on the two core components of our proposed PSNLMN—PSNLM and ADMFB—to validate their effectiveness. To ensure fairness, all experiments use the same parameters (learning rate, optimizer, batch size, network architecture, etc.) and data preprocessing techniques as used in the original model training. For efficiency, the models are trained for only 100 epochs with a magnification factor of 4 and evaluated on the same test dataset. Although the reported metrics may be lower than those of the fully trained models in the quantitative result analysis, this accelerated evaluation enables rapid assessment of different configurations and reveals the contributions of individual components. The observed trends and relative performance differences are consistent, meaningful, and indicative of each component’s value.

4.2.1. Ablation Study of Module PSNLM

To establish a baseline, we replaced the NLMB in the deep feature extraction stage with a classical residual block (RB) [16] and named it Base_Model. The RB consists of two convolutional layers and a residual connection, as shown in Figure 6a.
To verify the effectiveness of the proposed PSNLM module, we replace the second convolutional layer in the RB module with our PSNLM and add a layer normalization, as shown in Figure 6c. This modified block is named NLMB_P and is used to construct the model Model_PSNLM. In addition, to investigate the impact of the window interaction mechanism, we replace the PSNLM in NLMB_P with a variant that does not include window interaction, referred to as PSNLM_L, as shown in Figure 6b. We name this block NLMB_PL, which is used to construct the model Model_PSNLM_L.
According to Table 1, Model_PSNLM_L, which includes an incomplete version of PSNLM, achieves modest improvements in both PSNR and SSIM over the baseline model across all four benchmark datasets. For example, on the Urban100 dataset, Model_PSNLM_L improves the PSNR from 25.670 dB to 26.098 dB (+0.428 dB) and SSIM from 0.7870 to 0.7994 (+0.0124). These results suggest that the core design of PSNLM, even in its simplified form, is more effective than traditional residual blocks for capturing and reconstructing image features, particularly in images with complex structures. Furthermore, comparing Model_PSNLM_L with Model_PSNLM, which includes the complete PSNLM module, we observe further performance improvements. For example, on the BSD100 dataset, PSNR increases from 27.561 dB to 27.613 dB (+0.052 dB).
Overall, these results demonstrate that the model incorporating the complete PSNLM module offers comprehensive performance advantages over the baseline.

4.2.2. Ablation Study of Module ADMFB

To effectively capture multi-scale features, an effective strategy involves the parallel employment of dilated convolution layers with varying dilation rates. This approach has been successfully implemented in the Wide-Focus [27], which integrates contextual information across different ranges by combining standard convolution with multi-branch dilated convolutions, followed by additive fusion. However, this architecture has two limitations: (1) insufficient channel-wise selectivity within individual branches, and (2) the assumption of fixed contribution weights across different scales due to the simplistic element-wise summation used during fusion. The proposed ADMFB addresses these shortcomings through two key innovations: Firstly, Path 1 enables adaptive enhancement of informative channels and suppression of redundancy within each branch through channel-wise attention mechanisms, thereby optimizing intra-scale feature representation before fusion. Secondly, Path 2 replaces conventional additive fusion with content-aware adaptive weighting, allowing dynamic adjustment of cross-scale contribution weights based on input characteristics.
To quantitatively assess the contribution of each path, we conducted ablation studies. We constructed variants by incorporating simplified modules containing only the Path 1 components or only the Path 2 components into the baseline architecture, designated Base_Model_P1 and Base_Model_P2, respectively (illustrated by components in Figure 7c). Furthermore, to validate the overall efficacy of our proposed ADMFB against the original Wide-Focus, we established comparative baseline models: Base_Model_WF, which incorporates the original Wide-Focus block (Figure 7a), and Base_ADMFB, incorporating our complete ADMFB block (Figure 7b). The corresponding experimental results are presented in Table 2.
The ablation studies conducted across four benchmark datasets quantitatively validate the effectiveness of our adaptive dual-path fusion strategy. Compared to the baseline Wide-Focus model (Base_Model_WF), the incorporation of Path 1’s adaptive channel attention mechanism (Base_Model_P1) yields consistent, although moderate, improvements across all datasets. These enhancements are particularly pronounced in texture-rich Urban100. This demonstrates its capability in enhancing intra-scale feature discriminability through channel-wise feature refinement. Path 2’s dynamic fusion mechanism (Base_Model_P2) exhibits more substantial performance gains, especially in complex urban scenarios (Urban100) and natural imagery (BSD100), confirming the critical role of context-aware cross-scale integration. The complete ADMFB (Base_ADMFB) achieves optimal performance through synergistic combination of both pathways, showing progressive metric improvements that positively correlate with dataset complexity. Notably, the most significant performance gap between ADMFB and Wide-Focus emerges in Urban100, where structural complexity necessitates precise channel selection and adaptive multi-scale fusion. The consistent performance gains across datasets—from simple (Set5) to complex (Urban100)—highlight the complementary strengths of our dual-path design. Path 1 enhances intra-scale representation via adaptive channel recalibration, while Path 2 prioritizes cross-scale context using learnable weight allocation. This hierarchical design enables selective integration of the most relevant multi-scale features, enhancing image reconstruction quality. The experimental evidence suggests that the dual-path collaboration effectively addresses the limitations of fixed fusion paradigms by jointly improving intra-scale discrimination and cross-scale context modeling.

4.2.3. Ablation Study on the Combined Modules

The previous sections have confirmed the individual contributions of PSNLM and ADMFB. To further explore the overall gain achieved by their combination, we conducted joint ablation experiments, as summarized in Table 3, to assess the combined effect of both modules on super-resolution performance. Specifically, PSNLMN is constructed by replacing the RB module in Model_ADMFB with NLMB_P. In addition, we investigate how varying the number of branches (num_branch) in the MBFB module of ADMFB affects the performance of the combined model. The corresponding results are also summarized in Table 3.
Comparison between Model_ADMFB, Model_PSNLM, and the integrated model PSNLM_3 (three-branch configuration) shows that the integrated model PSNLMN_3, incorporating both ADMFB and PSNLM modules, achieved the highest PSNR/SSIM metrics across all four benchmark datasets. Specifically, on the Set14 dataset, the PSNR performance of PSNLMN_3 exhibited improvements of 0.322 dB over Model_PSNLM and 0.359 dB over Model_ADMFB. This finding underscores that the integration of both modules demonstrates effective complementarity in capturing multi-scale features and modeling long-range dependencies.
Furthermore, systematic comparisons of reconstruction performance under varying branch numbers in MBFB reveal that the 3-branch configuration (PSNLMN_3) consistently achieves optimal PSNR/SSIM metrics across all test datasets. Increasing the number of branches from 2 to 3 led to notable performance gains, particularly on Set14 and Urban100. The 3-branch configuration also showed more stable or slightly better performance than the 4-branch configuration. This pattern suggests that the 3-branch architecture optimally balances feature extraction efficiency and fusion effectiveness in the PSNLMN framework. Too few branches can limit multi-scale feature representation, while too many may introduce redundancy or training instability without meaningful performance gains, sometimes even causing slight degradation.
Collectively, the experiments validate the efficacy of combining ADMFB with PSNLM, while quantitatively establishing the 3-branch MBFB configuration as the optimal architectural choice under current magnification factors.

4.3. Performance Evaluation

To assess the effectiveness of the proposed PSNLMN, we conducted comparisons with several state-of-the-art SR methods, including MemNet [15], SRMDNF [17], CARN [18], IMDN [20], RFDN [49], ShuffleMixer [21], RLFN [49], ESRT [31], LBNet [32], SAFMN [50], and NGswin [28].

4.3.1. Quantitative Results Analysis

Table 4 presents the quantitative comparison of our proposed PSNLMN with a range of representative methods on four benchmark datasets under scale factors ×2, ×3, and ×4. PSNLMN demonstrates a considerable improvement in performance across all settings.
Scale ×2: On Set14, PSNLMN achieves a PSNR of 34.27 dB, outperforming the best baseline NGswin (33.79 dB) by +0.48 dB. On Urban100, SSIM improves from 0.9324 (NGswin) to 0.9447, a gain of +0.0123. PSNLMN also achieves the highest SSIM on all four datasets under this scale. Although the PSNR on Set5 (38.06 dB) and Urban100 (32.55 dB) is slightly lower than the best values (38.10 dB by BSRN and 32.58 dB by ESRT), PSNLMN surpasses them in SSIM by a clear margin, indicating improved structural similarity and visual fidelity.
Scale ×3: PSNLMN achieves substantial improvements in SSIM across all datasets. For example, on Set14, SSIM increases from 0.8456 (NGswin) to 0.8689 and, on BSD100, from 0.8078 to 0.8266, representing gains of +0.0233 and +0.0188, respectively. On Urban100, SSIM improves from 0.8603 to 0.8740 (+0.0137), and PSNR from 28.52 dB to 28.71 dB (+0.19 dB). While PSNR improvements on Set5 and BSD100 are relatively modest, the SSIM consistently ranks first, highlighting better texture and detail preservation.
Scale ×4: At this scale, PSNLMN consistently achieves the highest SSIM across all datasets. On Set5, the SSIM improves from 0.8966 (BSRN) to 0.9059 and, on Urban100, from 0.7963 (NGswin) to 0.8133, representing a notable improvement of 0.017. On BSD100, the SSIM increases by 0.0201, reaching 0.7597. Although the PSNR on Set5 (32.33 dB) and Set14 (28.74 dB) is slightly lower than that of the best-performing methods by 0.02–0.04 dB, the consistently superior SSIM indicates enhanced structural fidelity and perceptual quality.
In summary, PSNLMN delivers significant SSIM improvements across all datasets and scales, especially on texture-rich scenes like Set14 and Urban100. Although PSNR is not always the highest, it remains competitive, and the consistent SSIM superiority indicates better structural fidelity and perceptual quality of the reconstructed images. These results validate the effectiveness of our approach in enhancing both objective metrics and visual realism.

4.3.2. Qualitative Results Analysis

Figure 8 shows a visual comparison of reconstructed images by different methods, all under the challenging ×4 upscaling setting. We focus on the ×4 scale because it presents the greatest difficulty among all tested magnification levels. At this scale, artifacts and detail loss are more apparent, making it a better indicator of a model’s ability to recover fine structures. Therefore, it serves as a representative case for evaluating the perceptual quality and detail preservation of super-resolution algorithms.
In the first example, both PSNLMN and NGswin accurately reconstruct the striped texture on the bookshelf, maintaining clear and continuous lines. In contrast, methods such as IMDM and ShuffleMixer produce blurred or broken stripe patterns, failing to preserve the original structure. In the second example, PSNLMN recovers the dense line patterns of the bridge with higher precision, while the other methods show visible distortion or smearing. In the third example, which involves a tunnel ceiling with grid-like structures, PSNLMN preserves the regularity and sharpness of the grid, whereas competing methods tend to distort the geometry or lose detail. These visual results further demonstrate the superior ability of PSNLMN to retain fine-grained textures and structural integrity, especially in complex scenes with rich high-frequency content.

5. Conclusions

In this paper, we propose a novel architecture for SISR. The architecture integrates two key components: (1) a PSNLM grounded in NLM theory and multi-head self-attention mechanisms; and (2) an ADMFB to enhance multi-scale feature representation capabilities. This hybrid architecture leverages the complementary properties of CNNs and Transformers, aiming to mitigate their respective limitations. Specifically, the PSNLM module leverages a pixel-level self-attention mechanism to explicitly model long-range dependencies and non-local similarities. It reformulates the conventional NLM framework within a learnable self-attention paradigm, which enables dynamic weighting of image pixels based on domain patches with similar scenarios within the image. This mechanism helps to recover repetitive textures and long-range structures, whereas standard CNNs often miss due to limited receptive fields. Concurrently, the ADMFB employs a dual-path attention mechanism within a hierarchical feature fusion strategy. It adaptively fuses multi-scale features extracted from parallel branches possessing varying receptive fields. This design not only enriches feature diversity but also optimizes the trade-off between computational efficiency and multi-scale representation capabilities.
By cascading the PSNLM and ADMFB modules within an overarching CNN framework, our proposed architecture synergistically leverages both efficient local feature processing and comprehensive global and multi-scale feature representation. The experimental validation on benchmark datasets (Set5, Set14, BSD100, and Urban100) demonstrates that the proposed architecture outperforms most state-of-the-art methods across three distinct SISR tasks, measured by quantitative metrics (PSNR/SSIM) and qualitative visual effects, yielding more natural-looking reconstructions.
The practical applicability of the proposed architecture is particularly relevant for resource-constrained sensor systems where hardware upgrades are often infeasible, yet high-resolution image reconstruction remains critical for downstream applications such as medical diagnosis, autonomous navigation, and remote sensing analysis. By combining local and global modeling, the framework offers a balance between accuracy and efficiency for real-world super-resolution tasks.
However, the proposed framework still has several limitations. First, the incorporation of attention-based modules increases computational overhead compared to lightweight architectures, which may hinder its real-time deployment on edge devices. Second, its robustness under complex degradation scenarios remains to be further validated.
In future work, we plan to extend the proposed architecture to handle more complex and realistic degradation models. We also aim to investigate its generalization ability across different domains and image restoration tasks.

Author Contributions

Conceptualization, Y.X. and Y.W.; methodology, Y.X. and Y.W.; software, Y.X.; validation, Y.X. and Y.W.; formal analysis, Y.W.; investigation, Y.X.; resources, Y.X.; data curation, Y.X.; writing—original draft preparation, Y.X.; writing—review and editing, Y.X. and Y.W.; visualization, Y.X.; supervision, Y.W.; project administration, Y.X. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Doctoral Scientific Research Foundation of Yantai University, grant number JS22B56.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used in this study are publicly available.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Li, Y.; Sixou, B.; Peyrin, F. A review of the deep learning methods for medical images super resolution problems. IRBM 2021, 42, 120–133. [Google Scholar] [CrossRef]
  2. Dong, R.; Zhang, L.; Fu, H. RRSGAN: Reference-based super-resolution for remote sensing image. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5601117. [Google Scholar] [CrossRef]
  3. Shan, T.; Wang, J.; Chen, F.; Szenher, P.; Englot, B. Simulation-based lidar super-resolution for ground vehicles. Robot. Auton. Syst. 2020, 134, 103647. [Google Scholar] [CrossRef]
  4. Anwar, S.; Khan, S.; Barnes, N. A deep journey into super-resolution: A survey. Acm Comput. Surv. CSUR 2020, 53, 1–34. [Google Scholar] [CrossRef]
  5. Xu, X.; Ma, Y.; Sun, W. Towards real scene super-resolution with raw images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 1723–1731. [Google Scholar]
  6. Holst, G.C. Electro-Optical Imaging System Performance; SPIE Press Monograph; SPIE Press: Bellingham, WA, USA, 2008; Volume PM187. [Google Scholar]
  7. Sze, V.; Chen, Y.H.; Emer, J.; Suleiman, A.; Zhang, Z. Hardware for machine learning: Challenges and opportunities. In Proceedings of the 2017 IEEE Custom Integrated Circuits Conference (CICC), Austin, TX, USA, 30 April–3 May 2017; pp. 1–8. [Google Scholar]
  8. Wang, Y.; Bashir, S.M.A.; Khan, M.; Ullah, Q.; Wang, R.; Song, Y.; Guo, Z.; Niu, Y. Remote sensing image super-resolution and object detection: Benchmark and state of the art. Expert Syst. Appl. 2022, 197, 116793. [Google Scholar] [CrossRef]
  9. Wang, L.; Li, D.; Zhu, Y.; Tian, L.; Shan, Y. Dual super-resolution learning for semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2020), Seattle, WA, USA, 14–19 June 2020; pp. 3774–3783. [Google Scholar]
  10. Kim, K.I.; Kwon, Y. Single-image super-resolution using sparse regression and natural image prior. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 1127–1133. [Google Scholar] [PubMed]
  11. Dong, W.; Zhang, L.; Shi, G.; Wu, X. Image deblurring and super-resolution by adaptive sparse domain selection and adaptive regularization. IEEE Trans. Image Process. 2011, 20, 1838–1857. [Google Scholar] [CrossRef] [PubMed]
  12. Yue, L.; Shen, H.; Li, J.; Yuan, Q.; Zhang, H.; Zhang, L. Image super-resolution: The techniques, applications, and future. Signal Process. 2016, 128, 389–408. [Google Scholar] [CrossRef]
  13. Yang, J.; Wright, J.; Huang, T.S.; Ma, Y. Image super-resolution via sparse representation. IEEE Trans. Image Process. 2010, 19, 2861–2873. [Google Scholar] [CrossRef] [PubMed]
  14. Dong, C.; Loy, C.C.; He, K.; Tang, X. Image super-resolution using deep convolutional networks. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 38, 295–307. [Google Scholar] [CrossRef] [PubMed]
  15. Tai, Y.; Yang, J.; Liu, X.; Xu, C. Memnet: A persistent memory network for image restoration. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4539–4547. [Google Scholar]
  16. Ledig, C.; Theis, L.; Huszár, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Aitken, A.; Tejani, A.; Totz, J.; Wang, Z.; et al. Photo-realistic single image super-resolution using a generative adversarial network. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 4681–4690. [Google Scholar]
  17. Zhang, K.; Zuo, W.; Zhang, L. Learning a single convolutional super-resolution network for multiple degradations. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3262–3271. [Google Scholar]
  18. Ahn, N.; Kang, B.; Sohn, K.A. Fast, accurate, and lightweight super-resolution with cascading residual network. In Proceedings of the Computer Vision—ECCV 2018, Munich, Germany, 8–14 September 2018; pp. 252–268. [Google Scholar]
  19. Zhang, Y.; Li, K.; Li, K.; Wang, L.; Zhong, B.; Fu, Y. Image super-resolution using very deep residual channel attention networks. In Proceedings of the Computer Vision—ECCV 2018, Munich, Germany, 8–14 September 2018; pp. 286–301. [Google Scholar]
  20. Hui, Z.; Gao, X.; Yang, Y.; Wang, X. Lightweight image super-resolution with information multi-distillation network. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 2024–2032. [Google Scholar]
  21. Sun, L.; Pan, J.; Tang, J. Shufflemixer: An efficient convnet for image super-resolution. Adv. Neural Inf. Process. Syst. 2022, 35, 17314–17326. [Google Scholar]
  22. Ji, J.; Zhong, B.; Wu, Q.; Ma, K.K. A channel-wise multi-scale network for single image super-resolution. IEEE Signal Process. Lett. 2024, 31, 805–809. [Google Scholar] [CrossRef]
  23. Ke, G.; Lo, S.L.; Zou, H.; Liu, Y.F.; Chen, Z.Q.; Wang, J.K. CSINet: A Cross-Scale Interaction Network for Lightweight Image Super-Resolution. Sensors 2024, 24, 1135. [Google Scholar] [CrossRef] [PubMed]
  24. Dosovitskiy, A. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  25. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 10012–10022. [Google Scholar]
  26. Liang, J.; Cao, J.; Sun, G.; Zhang, K.; Van Gool, L.; Timofte, R. Swinir: Image restoration using swin transformer. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW), Montreal, BC, Canada, 11–17 October 2021; pp. 1833–1844. [Google Scholar]
  27. Tragakis, A.; Kaul, C.; Murray-Smith, R.; Husmeier, D. The fully convolutional transformer for medical image segmentation. In Proceedings of the 2023 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV 2023), Waikoloa, HI, USA, 2–7 January 2023; pp. 3660–3669. [Google Scholar]
  28. Choi, H.; Lee, J.; Yang, J. N-gram in swin transformers for efficient lightweight image super-resolution. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2023), Vancouver, BC, Canada, 18–22 June 2023; pp. 2071–2081. [Google Scholar]
  29. Liu, Y.; Xue, J.; Li, D.; Zhang, W.; Chiew, T.K.; Xu, Z. Image recognition based on lightweight convolutional neural network: Recent advances. Image Vis. Comput. 2024, 146, 105037. [Google Scholar] [CrossRef]
  30. Yang, F.; Yang, H.; Fu, J.; Lu, H.; Guo, B. Learning texture transformer network for image super-resolution. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2020), Seattle, WA, USA, 14–19 June 2020; pp. 5791–5800. [Google Scholar]
  31. Lu, Z.; Li, J.; Liu, H.; Huang, C.; Zhang, L.; Zeng, T. Transformer for single image super-resolution. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), New Orleans, LA, USA, 19–20 June 2022; pp. 457–466. [Google Scholar]
  32. Gao, G.; Wang, Z.; Li, J.; Li, W.; Yu, Y.; Zeng, T. Lightweight Bimodal Network for Single-Image Super-Resolution via Symmetric CNN and Recursive Transformer. In Proceedings of the 31st International Joint Conference on Artificial Intelligence, Vienna, Austria, 23–29 July 2022; pp. 913–919. [Google Scholar]
  33. Keys, R. Cubic convolution interpolation for digital image processing. IEEE Trans. Acoust. Speech Signal Process. 1981, 29, 1153–1160. [Google Scholar] [CrossRef]
  34. Duchon, C.E. Lanczos filtering in one and two dimensions. J. Appl. Meteorol. 1962–1982 1979, 18, 1016–1022. [Google Scholar] [CrossRef]
  35. Siu, W.C.; Hung, K.W. Review of image interpolation and super-resolution. In Proceedings of the 2012 Asia Pacific Signal and Information Processing Association Annual Summit and Conference, Hollywood, CA, USA, 3–6 December 2012; pp. 1–10. [Google Scholar]
  36. Rudin, L.I.; Osher, S.; Fatemi, E. Nonlinear total variation based noise removal algorithms. Phys. D Nonlinear Phenom. 1992, 60, 259–268. [Google Scholar] [CrossRef]
  37. Donoho, D.L. Compressed sensing. IEEE Trans. Inf. Theory 2006, 52, 1289–1306. [Google Scholar] [CrossRef]
  38. Wang, Z.; Chen, J.; Hoi, S.C. Deep learning for image super-resolution: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 3365–3387. [Google Scholar] [CrossRef] [PubMed]
  39. Timofte, R.; De Smet, V.; Van Gool, L. Anchored neighborhood regression for fast example-based super-resolution. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 1920–1927. [Google Scholar]
  40. Kim, J.; Lee, J.K.; Lee, K.M. Accurate image super-resolution using very deep convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1646–1654. [Google Scholar]
  41. Tai, Y.; Yang, J.; Liu, X. Image super-resolution via deep recursive residual network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3147–3155. [Google Scholar]
  42. Chen, H.; Wang, Y.; Guo, T.; Xu, C.; Deng, Y.; Liu, Z.; Ma, S.; Xu, C.; Xu, C.; Gao, W. Pre-trained image processing transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 12299–12310. [Google Scholar]
  43. Shi, W.; Caballero, J.; Huszár, F.; Totz, J.; Aitken, A.P.; Bishop, R.; Rueckert, D.; Wang, Z. Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1874–1883. [Google Scholar]
  44. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In Proceedings of the 13th European Conference, Zurich, Switzerland, 6–12 September 2014; pp. 740–755. [Google Scholar]
  45. Bevilacqua, M.; Roumy, A.; Guillemot, C.; Alberi-Morel, M.L. Low-complexity single-image super-resolution based on nonnegative neighbor embedding. In Proceedings of the British Machine Vision Conference, Guildford, UK, 3–7 September 2012; pp. 135.1–135.10. [Google Scholar]
  46. Zeyde, R.; Elad, M.; Protter, M. On single image scale-up using sparse-representations. In Proceedings of the 7th International Conference, Avignon, France, 24–30 June 2010; pp. 711–730. [Google Scholar]
  47. Martin, D.; Fowlkes, C.; Tal, D.; Malik, J. A database of human segmented natural images and its application to evaluating segmentation algorithms and measuring ecological statistics. In Proceedings of the Eighth IEEE International Conference on Computer Vision, ICCV 2001, Vancouver, BC, Canada, 7–14 July 2001; Volume 2, pp. 416–423. [Google Scholar]
  48. Huang, J.B.; Singh, A.; Ahuja, N. Single image super-resolution from transformed self-exemplars. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Boston, MA, USA, 7–12 June 2015; pp. 5197–5206. [Google Scholar]
  49. Liu, J.; Tang, J.; Wu, G. Residual feature distillation network for lightweight image super-resolution. In Proceedings of the Computer Vision—ECCV 2020 Workshops, Glasgow, UK, 23–28 August 2020; pp. 41–55. [Google Scholar]
  50. Sun, L.; Dong, J.; Tang, J.; Pan, J. Spatially-adaptive feature modulation for efficient image super-resolution. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops (ICCVW), Paris, France, 2–6 October 2023; pp. 13190–13199. [Google Scholar]
  51. Li, Z.; Liu, Y.; Chen, X.; Cai, H.; Gu, J.; Qiao, Y.; Dong, C. Blueprint separable residual network for efficient image super-resolution. In Proceedings of the Computer Vision and Pattern Recognition, New Orleans, LO, USA, 18–24 June 2022; pp. 833–843. [Google Scholar]
Figure 1. The overall overview of the proposed PSNLMN. (a) The architecture of PSNLMN; (b) upsampling block.
Figure 1. The overall overview of the proposed PSNLMN. (a) The architecture of PSNLMN; (b) upsampling block.
Sensors 25 04044 g001
Figure 2. The non-local multi-branch module (NLMB).
Figure 2. The non-local multi-branch module (NLMB).
Sensors 25 04044 g002
Figure 3. Pixel-wise self-attention-based non-local mean module (PSNLM).
Figure 3. Pixel-wise self-attention-based non-local mean module (PSNLM).
Sensors 25 04044 g003
Figure 4. Window partition and window stitching. Different colors represent different channels.
Figure 4. Window partition and window stitching. Different colors represent different channels.
Sensors 25 04044 g004
Figure 5. Adaptive dual-path multi-branch fusion block (ADMFB).
Figure 5. Adaptive dual-path multi-branch fusion block (ADMFB).
Sensors 25 04044 g005
Figure 6. The schematic diagram of the structures of RB, NLMB_PL, and NLMB_P.
Figure 6. The schematic diagram of the structures of RB, NLMB_PL, and NLMB_P.
Sensors 25 04044 g006
Figure 7. The schematic diagram of the structures of Wide-Focus and the basic blocks used in the experiment.
Figure 7. The schematic diagram of the structures of Wide-Focus and the basic blocks used in the experiment.
Sensors 25 04044 g007
Figure 8. Visual comparison of ×4 SR methods on the benchmark test set (The area used for comparison is marked with a red box).
Figure 8. Visual comparison of ×4 SR methods on the benchmark test set (The area used for comparison is marked with a red box).
Sensors 25 04044 g008
Table 1. Results of ablation study on PSNLM. The best results are highlighted in bold.
Table 1. Results of ablation study on PSNLM. The best results are highlighted in bold.
ModelRBNLMB_PLNLMB_PSet5Set14BSD100Urban100
PSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIM
Base_Model31.6070.898427.9420.802927.4190.750225.6700.7870
Model_PSNLM_L31.9290.902428.2170.808627.5610.754926.0980.7994
Model_PSNLM32.0210.903628.2500.809827.6130.756326.1610.8021
Table 2. Results of ablation study on ADMFB. The best results are highlighted in bold.
Table 2. Results of ablation study on ADMFB. The best results are highlighted in bold.
ModelRBWide-FocusPath1Path2Set5Set14BSD100Urban100
PSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIM
Base_Model_WF31.7940.901028.0300.805027.4060.753525.8030.7947
Base_Model_P131.8350.901128.1020.806527.4760.753925.8970.7951
Base_Model_P231.8980.901428.1750.806827.5150.754125.9370.7956
Model_ADMFB31.9620.902628.2130.807727.5980.756026.0840.7989
Table 3. Results of ablation study on combined modules. The best results are highlighted in bold.
Table 3. Results of ablation study on combined modules. The best results are highlighted in bold.
ModelNum_branchNLMB_PPath1Path2Set5Set14BSD100Urban100
PSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIM
Model_PSNLM332.0210.903628.2500.809827.6130.756326.1610.8021
Model_ADMFB331.9620.902628.2130.807727.5980.756126.0840.7989
PSNLMN_2232.0240.903128.2710.808827.5730.755226.040.7986
PSNLMN_4332.1320.903728.5420.810827.6430.756826.2370.8029
PSNLMN_3432.1440.903928.5720.811027.6500.757126.2430.8030
Table 4. Quantitative comparison of image SR with other methods (PSNR/SSIM). The best and second-best results are highlighted using bold and underlined, respectively.
Table 4. Quantitative comparison of image SR with other methods (PSNR/SSIM). The best and second-best results are highlighted using bold and underlined, respectively.
MethodScaleSet5Set14BSD100Urban100
PSNR/SSIM PSNR/SSIM PSNR/SSIMPSNR/SSIM
MemNet [15] × 2 37.78/0.959733.28/0.914232.08/0.897831.31/0.9195
SRMDNF [17] × 2 37.79/0.96033.32/0.91532.05/0.89831.33/0.920
CARN [18] × 2 37.76/0.959033.52/0.916632.09/0.897831.92/0.9256
IMDN [20] × 2 38.00/0.960533.63/0.917732.19/0.899632.17/0.9283
RFDN-L [49] × 2 38.08/0.960633.67/0.919032.18/0.899632.24/0.9290
ShuffleMixer [21] × 2 38.01/0.960633.63/0.918032.17/0.899531.89/0.9257
ESRT [31] × 2 38.03/0.960033.75/0.918432.25/0.900132.58/0.9318
LBNet [32] × 2 38.05/0.960733.65/0.917732.16/0.899432.30/0.9291
SAFMN [50] × 2 38.00/0.960533.54/0.917732.16/0.899531.84/0.9256
BSRN [51] × 2 38.10/0.961033.74/0.919332.24/0.900632.34/0.9303
NGswin [28] × 2 38.05/0.961033.79/0.919932.27/0.900832.53/0.9324
PSNLMN(our) × 2 38.06/0.965434.27/0.934932.31/0.911932.55/0.9447
MemNet [15] × 3 34.09/0.924830.00/0.835028.96/0.800127.56/0.8376
SRMDNF [17] × 3 34.12/0.92530.04/0.83728.97/0.80327.57/0.840
CARN [18] × 3 34.29/0.925530.29/0.840729.06/0.803428.06/0.8493
IMDN [20] × 3 34.36/0.927030.32/0.841729.09/0.804628.17/0.8519
RFDN-L [49] × 3 34.47/0.928030.35/0.842129.11/0.805328.32/0.8547
ShuffleMixer [21] × 3 34.40/0.927230.37/0.842329.12/0.805128.08/0.8498
ESRT [31] × 3 34.42/0.926830.43/0.843329.15/0.806328.46/0.8574
LBNet [32] × 3 34.47/0.927730.38/0.841729.13/0.806128.42/0.8559
SAFMN [50] × 3 34.34/0.926730.33/0.841829.08/0.804827.95/0.8474
BSRN [51] × 3 34.46/0.927730.47/0.844929.18/0.806828.39/0.8567
NGswin [28] × 3 34.52/0.928230.53/0.845629.19/0.807828.52/0.8603
PSNLMN(our) × 3 34.56/0.936930.60/0.868929.23/0.826628.71/0.8740
MemNet [15] × 4 31.74/0.889328.26/0.772327.40/0.728125.50/0.7630
SRMDNF [17] × 4 31.96/0.89328.35/0.77727.49/0.73425.68/0.773
CARN [18] × 4 32.13/0.893728.60/0.780627.58/0.734926.07/0.7837
IMDN [20] × 4 32.21/0.894828.58/0.781127.56/0.735326.04/0.7838
RFDN-L [49] × 4 32.28/0.895728.61/0.781827.58/0.736326.20/0.7883
ShuffleMixer [21] × 4 32.21/0.895328.66/0.782727.61/0.736626.08/0.7835
ESRT [31] × 4 32.19/0.894728.69/0.783327.69/0.737926.39/0.7962
LBNet [32] × 4 32.29/0.896028.68/0.783227.62/0.738226.27/0.7906
SAFMN [50] × 4 32.18/0.894828.60/0.781327.58/0.735925.97/0.7809
BSRN [51] × 4 32.35/0.896628.73/0.784727.65/0.738726.27/0.7908
NGswin [28] × 4 32.33/0.896328.78/0.785927.66/0.739626.45/0.7963
PSNLMN(our) × 4 32.33/0.905928.74/0.813227.70/0.759726.56/0.8133
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xu, Y.; Wang, Y. Single-Image Super-Resolution via Cascaded Non-Local Mean Network and Dual-Path Multi-Branch Fusion. Sensors 2025, 25, 4044. https://doi.org/10.3390/s25134044

AMA Style

Xu Y, Wang Y. Single-Image Super-Resolution via Cascaded Non-Local Mean Network and Dual-Path Multi-Branch Fusion. Sensors. 2025; 25(13):4044. https://doi.org/10.3390/s25134044

Chicago/Turabian Style

Xu, Yu, and Yi Wang. 2025. "Single-Image Super-Resolution via Cascaded Non-Local Mean Network and Dual-Path Multi-Branch Fusion" Sensors 25, no. 13: 4044. https://doi.org/10.3390/s25134044

APA Style

Xu, Y., & Wang, Y. (2025). Single-Image Super-Resolution via Cascaded Non-Local Mean Network and Dual-Path Multi-Branch Fusion. Sensors, 25(13), 4044. https://doi.org/10.3390/s25134044

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop