Next Article in Journal
A Miniaturized Design for a Terahertz Tri-Mirror CATR with High QZ Characteristics
Previous Article in Journal
AGEN: Adaptive Error Control-Driven Cross-View Geo-Localization Under Extreme Weather Conditions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

ICAFormer: An Image Dehazing Transformer Based on Interactive Channel Attention

1
Hubei Key Laboratory of Optical Information and Pattern Recognition, School of Electrical and Information Engineering, Wuhan Institute of Technology, Wuhan 430205, China
2
School of Artificial Intelligence and Automation, Huazhong University of Science and Technology, Wuhan 430072, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(12), 3750; https://doi.org/10.3390/s25123750
Submission received: 28 April 2025 / Revised: 7 June 2025 / Accepted: 9 June 2025 / Published: 15 June 2025
(This article belongs to the Section Sensing and Imaging)

Abstract

Single image dehazing is a fundamental task in computer vision, aiming to recover a clear scene from a hazy input image. To address the limitations of traditional dehazing algorithms—particularly in global feature association and local detail preservation—this study proposes a novel Transformer-based dehazing model enhanced by an interactive channel attention mechanism. The proposed architecture adopts a U-shaped encoder–decoder framework, incorporating key components such as a feature extraction module and a feature fusion module based on interactive attention. Specifically, the interactive channel attention mechanism facilitates cross-layer feature interaction, enabling the dynamic fusion of global contextual information and local texture details. The network architecture leverages a multi-scale feature pyramid to extract image information across different dimensions, while an improved cross-channel attention weighting mechanism enhances feature representation in regions with varying haze densities. Extensive experiments conducted on both synthetic and real-world datasets—including the RESIDE benchmark—demonstrate the superior performance of the proposed method. Quantitatively, it achieves PSNR gains of 0.53 dB for indoor scenes and 1.64 dB for outdoor scenes, alongside SSIM improvements of 1.4% and 1.7%, respectively, compared with the second-best performing method. Qualitative assessments further confirm that the proposed model excels in restoring fine structural details in dense haze regions while maintaining high color fidelity. These results validate the effectiveness of the proposed approach in enhancing both perceptual quality and quantitative accuracy in image dehazing tasks.

1. Introduction

Haze is a natural atmospheric phenomenon caused by the presence of numerous suspended particles, such as water droplets and dust. Under hazy weather conditions, atmospheric light is absorbed and scattered by these particles during propagation, leading to a gradual attenuation of light intensity before reaching the imaging sensor. This reduction in light intensity significantly impairs the visibility of the surrounding environment, resulting in a degradation of image quality to varying degrees. Images captured under hazy conditions often suffer from diminished clarity, contrast, and color vibrancy. In many practical applications—including public transportation safety, satellite image analysis, outdoor video surveillance, and autonomous driving—high-quality images are crucial for accurate information extraction and decision making. Narasimhan and Nayar et al. [1] formalized the atmospheric scattering model as follows:
I x = J x t x + A 1 t x
where I(x) represents the image affected by haze, J(x) represents the clear, haze-free image, A denotes the global atmospheric light, and t(x) represents the transmission rate. In the case of a homogeneous atmosphere, the transmission rate can be expressed as follows:
t x = e β d x
where β represents the scattering coefficient and d(x) represents the scene depth. Equation (2) indicates that when d(x) approaches infinity, t(x) approaches zero. Combining the two equations above, we have
A = I x ,   d x
In the practical imaging process of a distant view, d(x) cannot be infinite, but rather a long distance, which results in a very low transmission rate t0. Instead of relying on Equation (3) to obtain the global atmospheric light A, a more accurate global atmospheric light A can be derived using the following equation:
A = m a x y x t ( x ) t 0 I ( y )
The above formulation highlights that the core challenge in image dehazing lies in accurately estimating the transmission map. By obtaining a reliable transmission map, the original haze-free image can be effectively restored. However, the precise and direct recovery of the transmission map typically requires additional depth information, which is often unavailable in practical scenarios.
It can be seen that image dehazing is an ill-posed problem. Early image dehazing methods used prior constraints to solve the problem space. They typically estimate A and t(x) separately to reduce the complexity of the problem, and then use Equation (1) to obtain the result. These prior-based methods can generate images with good visual clarity. However, these images are visually different from clear images, and may introduce artifacts in regions that do not satisfy the prior conditions.
In 2011, He et al. [2] discovered that in a natural image, there is at least one region where the pixel values are very low, which is referred to as the dark channel. Based on this discovery, they proposed the dark channel prior dehazing algorithm. In 2014, Fattal et al. [3] found that in natural images, colors exhibit a one-dimensional distribution within local regions. Building on this, they utilized features with lighter colors and incorporated the smoothness of the transmission map to enhance the image visibility. In 2016, Ren et al. [4] constructed a multi-scale convolutional neural network to model the mapping between hazy image features and transmission rate. This approach significantly improved the scene adaptability of dehazing algorithms. In the same year, Cai et al. [5] designed an end-to-end deep learning model, which directly predicts the transmission map and innovatively couples it with the classic atmospheric scattering model, achieving physically interpretable image restoration. The following year, Li et al. [6] made groundbreaking improvements to the traditional physical model. Through mathematical derivation, they merged the two key parameters, atmospheric light intensity and transmission rate, into a unified variable, creating a simplified parameter model for hazy imaging. While this approach improved the dehazing effect compared with previous methods, the simultaneous estimation of these two parameters introduced cumulative errors in image restoration, which negatively affected the dehazing results. In 2018, Zhang et al. [7] innovatively constructed a dense skip-connected encoder–decoder network with a U-shaped symmetric topology design and proposed a dual-branch collaborative optimization mechanism to jointly estimate the haze concentration parameters. The dehazed image was obtained by solving the atmospheric scattering equation in reverse. The study used a deep learning framework to achieve the end-to-end prediction of transmission rate and atmospheric light parameters in the traditional physical model. In 2021, Zhang et al. [8] improved and optimized the transmission rate, atmospheric light, and dehazed results, and integrated the refined dehazing results with the reconstructed clear image perception, producing more natural and clearer dehazing results. Although dehazing methods based on physical prior constraints have demonstrated significant restoration effects in conventional scenarios after long iterations and evolution, they still experience performance fluctuations under extreme conditions, such as complex lighting and uneven haze distribution, limiting the expansion of their practical application range. The current models still require improvements in their ability to model the nonlinear coupling relationships between atmospheric parameters, which has become a major technical bottleneck limiting the robustness of the algorithms.
In recent years, deep learning has become very popular in the field of computer vision, and researchers have proposed a large number of image dehazing methods based on deep convolutional networks. These methods can be divided into three types: The first type combines the concept of the atmospheric scattering model with deep learning [5,6,7], using deep learning techniques to correct the transmission rate and atmospheric light, and then incorporating them into the model to obtain a clear image. The second type is based on the idea of Generative Adversarial Networks (GANs) [9,10,11], introducing adversarial loss, L1 loss, and perceptual loss to effectively remove haze from degraded images. The third type utilizes a large dataset of synthetic images [12] as the input to learn image features for dehazing, without the need to estimate the transmission rate and atmospheric light.
Generative Adversarial Networks (GANs), as an innovative unsupervised learning paradigm, have a core architecture consisting of two modules, a generator and a discriminator, forming an adversarial training framework. The model creates a dynamic game mechanism between the generator and the discriminator, where the generator aims to synthesize samples that closely approximate the distribution of real data, while the discriminator continually optimizes its ability to distinguish between real and synthetic data. This adversarial training process drives both networks to reach a Nash equilibrium in the parameter space, ultimately enabling the generative model to accurately model complex data distributions. GANs have achieved great success in image processing. Traditional convolutional neural network algorithms use pixel-level traditional losses, which can lead to blurry output images. GANs introduce high perceptual loss in the feature space, improving the quality of dehazed images. Goodfellow et al. [9] proposed GANs, generating clean images from input images with random noise. However, due to the instability of the training process, synthesized images are easily affected by color shifts and noise. Mirza et al. [10] proposed Conditional GANs (cGAN), which improve the stability of the training process by adding conditional information and variables. Li et al. [13] applied cGAN to single image dehazing, achieving dehazing by introducing L1 regularization gradients and VGG features. Compared with traditional cGANs, Runde’s proposed encoder–decoder structure is an end-to-end design that generates better results. However, unsupervised GAN-based algorithms also have some drawbacks. First, these GAN algorithms require substantial computational resources and time during the training process. Second, their training is unstable and requires adjusting a large number of parameters. Finally, when applied to dehazing, GANs have limitations in local feature extraction and cannot effectively perceive global feature information of the image.
This study presents a novel single image dehazing model built upon the Transformer architecture. The core design philosophy draws inspiration from the remarkable success of Transformers in the field of natural language processing. Recent advancements have demonstrated that vision Transformers exhibit superior performance across various computer vision tasks, posing a paradigm shift from traditional convolutional neural network (CNN)-based architectures. Guo et al. [14] leveraged Transformer-based architectures to enhance the modeling of long-range dependencies in images, thereby improving dehazing performance. Similarly, Zhao et al. [15] proposed an end-to-end single image dehazing approach that integrates the Transformer architecture with a texture-enhanced attention mechanism, further advancing feature representation capabilities. Specifically, in traditional methods, the feature representation is constructed through hierarchical convolutional operations, which, when compared with the global attention mechanism of Transformers, reveal clear limitations.

2. Related Work

2.1. Image Dehazing Based on CNNs

In recent years, Convolutional Neural Networks (CNNs) have made significant progress in the field of visual restoration, with image dehazing technology emerging as a particularly prominent application. Existing research primarily follows two major technical approaches: physics-based model-driven methods and end-to-end data-driven methods. The first approach is based on the atmospheric scattering model for constructing a mathematical framework, in which neural networks are employed to estimate the medium transmission rate and atmospheric light parameters. Typical examples include the DehazeNet [5], DCPDN [7], and AOD-Net [6] architectures. These methods perform well in scenes that align with physical assumptions, but their dehazing performance significantly deteriorates when there are deviations in the estimation of atmospheric optical parameters. The second approach adopts an end-to-end mapping strategy, directly regressing a clear image from a hazy input. While this method eliminates the reliance on physical models, it inherently struggles to model long-range pixel dependencies effectively.

2.2. Vision Transformer

In 2020, the ViT architecture [16] demonstrated its potential as an alternative to CNNs and achieved significant breakthroughs in the field of computer vision. SwinIR [17] combined the Swin Transformer layers with convolutional layers to form a residual module, which was successfully applied to image restoration tasks. The Swin Transformer [18] divides an image into windows and performs the self-attention mechanism within these windows, ensuring that the computational cost grows linearly. Restormer [19] further improved image restoration by incorporating multi-head transposed attention modules and gated feed-forward networks, enabling high-resolution image restoration. Additionally, ViT has achieved state-of-the-art (SOTA) performance in various tasks, including image enhancement [20,21,22], image segmentation [23,24,25], object detection [26,27,28], human pose estimation [29,30,31], and image classification [32,33,34].
Swin Transformer is one of the pioneers in applying Transformers to low-level vision tasks. It constructs a large residual block consisting of stacked Swin Transformer blocks and convolutional layers. Here, we will briefly review the Swin Transformer. Given a feature map X, it projects X into Q, K, and V (query, key, and value) through a linear layer, then divides it into different windows, where the partitioning regions of Q, K, and V differ for each window. The calculation formula for the self-attention mechanism is as follows:
A t t e n t i o n ( Q , K , V ) = S o f t m a x Q K T d V
where d is the dimension of the query and key vectors, used to scale the dot product for numerical stability and to prevent excessively large gradients.
The Swin Transformer adopts a partitioned self-attention mechanism, leveraging a windowing strategy to improve computational efficiency while maintaining the ability to model long-range dependencies. This approach enhances the flexibility of the network while preserving computational efficiency.

2.3. Efficient Attention Mechanism

The attention mechanism has been widely adopted in the field of computer vision due to its ability to highlight the relative importance of different features by assigning them adaptive weights. Depending on the focus of the operation, attention mechanisms can be broadly categorized into channel attention, spatial attention, and self-attention, among others. The attention mechanism proposed in this paper is the channel attention mechanism. Compared with other attention mechanisms, the channel attention mechanism is capable of capturing specific patterns or features within an image while minimizing computational complexity, particularly in the task of image dehazing. The most classic implementation of channel attention is the SE (Squeeze-and-Excitation) attention mechanism, proposed by Hu [35] in 2018. Its core idea is to explicitly model the dependencies between channels and adaptively adjust the feature responses of each channel. The formation process of the SE attention mechanism can be divided into three steps: Squeeze, Excitation, and Scale. Let us review the formation process of the SE attention mechanism: Let the input feature map be U R H × W × C , where H, W, and C represent the height, width, and the number of channels, respectively. The spatial dimensions H × W are compressed by performing global average pooling on each channel c.
z c = 1 H × W i = 1 H j = 1 W u c ( i , j )
Each element zc represents the global information of the c-th channel. The dimensionality of zc is reduced through a fully connected layer as follows:
s = W 2 R e L U ( W 1 z c + b 1 ) + b 2
where W 1 , W 2 R C r × C are the weight matrices, and b1, b2 are the biases. After normalizing the weight s, it is multiplied with the input feature map U channel-wise to obtain the weighted feature map U ~ R H × W × C . The channel attention mechanism proposed in this paper similarly enhances the feature representation ability of convolutional neural networks.

3. The Proposed Method

Our objective is to develop an efficient and lightweight Transformer-based network for image dehazing. To reduce computational complexity while enhancing the extraction of fine-grained image features, we adopt a Swin Transformer-based framework. Furthermore, we introduce a multi-level convolutional attention mechanism to effectively capture features across different representation levels.

3.1. The Overall Architecture of the Network

U-Net [36] is one of the commonly used networks in image processing, capable of capturing multi-scale image features. Inspired by this, the main framework of our model is shown in Figure 1, with a hierarchical design consisting of four stages. The input image I R H × W × C is received through a 3 × 3 convolution, which extracts local texture features and outputs a feature map I R H × W × C with the same resolution. Global dependencies are modeled on the feature map to capture long-range spatial relationships. The resolution is halved and the number of channels is doubled through a downsampling operation, resulting in I R H 2 × W 2 × 2 C , where more complex features are modeled in a higher dimensional space. Another downsampling operation is applied, halving the resolution again and doubling the channels, producing D R H 4 × W 4 × 4 C , which enhances the retention of global feature details. The last two stages are the symmetric processes of the first two stages. D R H 4 × W 4 × 4 C is upsampled to obtain J R H 2 × W 2 × 2 C , and upsampled again to obtain J R H × W × C . We assign weights to the features at each step and fuse the features from different layers to achieve the dehazing effect.
The Transformer block is designed by replacing the standard feed-forward network with a haze-specific feature extraction module, enabling the capture of haze-related features across multiple scales.
Following average pooling and activation via the Sigmoid function, the feature fusion module employs an interactive channel attention mechanism to extract and integrate features across multiple hierarchical levels.
The ICA module is developed based on an interactive channel attention mechanism. We introduce this channel attention mechanism to enhance the representation capability of features at different scales. This design is incorporated into the decoder part of the U-Net architecture (as illustrated in Figure 1), corresponding to the integration of multi-scale features.
Considering that multiple operations within the Transformer block may result in the loss of some image detail information, we incorporate both local and global residual connections. Within the Transformer block, a standard residual connection structure is employed to ensure stable gradient propagation, thus preventing the vanishing gradient problem during Transformer training. Additionally, cross-layer global residual connections are implemented between each basic layer throughout the network. This operation mitigates the vanishing gradient issue in deep networks and facilitates the fusion of features at different scales.

3.2. Transformer Block

To effectively enhance feature representation in the single image dehazing task, this paper proposes an efficient visual Transformer block module, as illustrated in Figure 2. This module improves upon the Swin Transformer block by integrating a dual-path design that simultaneously captures long-range dependencies and restores local spatial details.
The input features are normalized through the RescaleNorm layer. Unlike traditional LayerNorm, RescaleNorm preserves the global mean and variance information of the input features, preventing the loss of overall image brightness and contrast due to normalization in low-level vision tasks.
R e s c a l e N o r m ( x ) = s x μ σ ( x ) + b
where σ(x) denotes the standard deviation of the input feature x, s is the learnable scaling coefficient, μ is the mean value, and b is the bias term. The formula for computing σ(x) is:
σ ( x ) = 1 n i = 1 n x i μ 2 + ε
where, as in Equation (8), μ is the mean value, and ε is a constant introduced for numerical stability.
The features are processed through two distinct paths. The first path utilizes the standard Multi-Head Self-Attention (MHSA) mechanism. The input features are first mapped into query, key, and value vectors, where queries and keys are generated through linear layers. A dot product attention calculation (using MatMul and SoftMax operations) is then performed, followed by the merging and fusion of the attention head outputs through a linear layer to obtain global context information. The second path employs convolutional operations, which complement the local modeling capability of MHSA and enhance the extraction of spatial structural information. This convolutional branch directly operates on the value branch and runs in parallel with the self-attention path, aiding in the recovery of detail and texture information.
To further enhance the model’s feature representation capability, an affine transformation is applied after the fusion of the multi-head attention and convolution branches’ outputs, improving the model’s adaptability to the feature space. The fused features are then fed into a Multi-Layer Perceptron (MLP) module, which consists of two linear layers with a ReLU activation function in between. Compared with the commonly used GELU activation function, ReLU is more suitable for low-level vision tasks, such as image restoration, as it effectively improves the model’s stability and reconstruction quality.
Moreover, the Transformer block incorporates multiple residual connections, which ensure efficient feature transfer between different modules and maintain gradient stability. This design accelerates model convergence during training and enhances the overall dehazing performance.

3.3. Interactive Channel Attention Mechanism

In recent years, the SE channel attention mechanism has been widely used in the field of image dehazing, achieving impressive results. The fully connected layers in the SE channel attention mechanism capture global information, enhancing the dehazing network’s ability to focus on key features, thereby improving performance. However, this approach lacks interaction between global and local information, leading to inaccurate feature weight distribution in image dehazing tasks. To efficiently integrate global and local information and achieve a more reasonable weight distribution, we introduce an interactive channel attention mechanism. The framework structure is shown in Figure 3. Our mechanism facilitates bidirectional interaction between local textures and global context during feature fusion, enabling more precise attention weight assignment across hierarchical features.
The interactive channel attention mechanism is primarily utilized during the feature fusion stage of the backbone network. In the process of network construction, effectively leveraging both global and local information is crucial for enhancing the performance of image dehazing.
As shown in Figure 3, given a feature map X R H × W × C , where C, H, and W denote the number of channels, height, and width, respectively. Global information from the channels is first extracted through global average pooling as follows:
U c = G A P ( X c ) = 1 H × W i = 1 H j = 1 W X c ( i , j )
where Xc(i,j) is the feature value of the c-th channel and GAP(x) is the global average pooling function. This function allows the feature map to be transformed from C × H × W to C × 1 × 1.
Uc represents the statistical measure of the c-th channel, while U denotes the global average pooled feature vector, which can be expressed as U = [U1,U2,…,Uc]T.
To extract the global information of the channels, we obtain a portion of the information from the globally average pooled feature U, which primarily focuses on the weight of its own channel without considering the interactions between channels as follows:
U g c = d i a g ( U ) = i = 1 c U d i
where diag(x) denotes the diagonalization operation, which is used to extract the global information of the channel features, and di represents the global weight vector of the i-th channel, which models the overall inter-channel relationships. During the attention computation, it is essential to incorporate dependencies between adjacent channels to enable effective information interaction across channels. Additionally, a portion of the information is extracted from U as follows:
U l c = b a n d ( U ) = i = 1 k U b i
where band(x) represents the band-like operation used to extract local channel correlations and bi represents the i-th banded matrix convolution kernel.
In summary, the diag(x) operation performs channel-wise weighting through a fully connected transformation across channels, modeling global inter-channel dependencies and generating the output Ugc. In contrast, the band(x) operation applies channel-wise weighting via a localized convolution across channels, capturing local inter-channel dependencies and producing the output Ulc.
Next, matrix multiplication is performed to obtain the interaction information between channels as follows:
S g c = σ j c U g c U l c T i , j , S l c = σ j c U l c U g c T i , j , i 1 , 2 , 3 c
where σ(x) represents the activation function applied to x, ensuring that the attention weights lie within the range (0, 1), i denotes the global channel index corresponding to Ugc, and j denotes the local channel index corresponding to Ulc. The operation above sums these two matrices to compute the inductive information for each channel.
Finally, the calculated attention weights W are applied to the original input features X as follows:
X * = X × W
where X is the input feature map and X* is the final output feature map with a shape of C × 1 × 1. This allows for the weighted adjustment of each channel of the input F, enabling the network to focus more on the key channels and enhancing the feature representation capability.

3.4. Loss Functions

To effectively enhance the model’s ability to recover fine image details and structural information, this paper introduces a combination of Mean Absolute Error (MAE) loss, perceptual loss, and edge loss. These loss functions guide the model in learning the differences between hazy and clear images. The total loss is defined as a weighted combination of these three components:
L = λ 1 L 1 + λ 2 L p e r + λ 3 L edge
where λ1, λ2, and λ3 represent the respective weighting coefficients assigned to each loss function. Based on the results obtained from controlled experiments, the optimal values were empirically determined to be λ1 = 0.9, λ2 = 0.05, and λ3 = 0.05.
The mean absolute error loss (L1 loss) function is specified as follows:
L 1 = 1 N i = 1 N I i ^ I i
where I i ^ denotes the model’s predicted value for the i-th pixel, Ii represents the corresponding pixel value in the ground-truth clear image, and N indicates the total number of pixels in the image. Compared with L2 loss optimization, the L1 loss effectively reduces oversmoothing artifacts in image dehazing while demonstrating superior detail preservation capability. Furthermore, the inherent insensitivity to outliers in L1 loss enhances training stability and model robustness.
The Lper loss function is implemented using VGG-16 [37] as the loss network, and its formulation is as follows:
L p e r = i = 1 3 1 C j H j W j ϕ j I ϕ J 2
where Cj, Hj, and Wj denote the channel dimension, height, and width of ϕj( ), respectively, with ϕj(I) representing the feature maps of the dehazed image and ϕj(J) corresponding to the feature maps of the ground-truth clear image.
The Ledge loss function is calculated by applying convolution operations with the Laplacian operator to extract the edge maps of both the clear image and the dehazed image. The loss is then defined as the difference between these edge maps. The formulation is specified as follows:
L e d g e = i , j x g i , j x r i , j
where ∇xg(i,j) denotes the gradient of the generated image at position (i,j) and ∇xr(i,j) represents the gradient of the real image at the corresponding spatial coordinates.

4. Experiments and Results

4.1. Dataset

The dataset used in this paper is sourced from the RESIDE dataset [38], which consists of multiple subsets for comprehensive evaluation: a synthetic indoor image pair collection (Indoor Training Set, ITS) containing 13,990 samples, an outdoor training subset (Outdoor Training Set, OTS) with 72,135 images, and a Synthetic Objective Testing Set (SOTS) comprising 500 indoor and 500 outdoor hazy images for quantitative assessment. To ensure the fairness and reliability of the experimental results, the dataset was split into a training set and a testing set at a ratio of 9:1.

4.2. Evaluation Metrics

The evaluation metrics employed in this study are Peak Signal-to-Noise Ratio (PSNR) and Structural Similarity Index Measure (SSIM). PSNR is an objective metric calculated based on pixel-wise error that is used to quantify the difference between the restored image and the reference image. Essentially, it represents the signal-to-noise ratio derived from a logarithmic transformation of the mean squared error (MSE). A higher PSNR value indicates greater similarity between the restored and reference images, reflecting better image quality. SSIM, on the other hand, is an image quality metric that aligns more closely with human visual perception, comprehensively evaluating the brightness, contrast, and structural information of the image.
Due to the lack of corresponding ground-truth references in real-world images, which are required for the calculation of PSNR and SSIM, we selected FADE [39]—one of the representative methods among numerous no-reference perceptual haze density estimation techniques [40,41]. FADE is capable of predicting haze density directly from a single hazy image without the need for a corresponding haze-free reference. The haze density is computed as follows:
F A D E = D f D f f + 1
where Df denotes the Mahalanobis distance between the input image and the statistical model of haze-free images, while Dff represents the Mahalanobis distance between the input image and the statistical model of hazy images. The Mahalanobis distance is calculated as follows:
D = 1 N i = 1 N x i μ T 1 x i μ
where μ denotes the mean feature vector of haze-free or hazy images, Σ represents the corresponding covariance matrix of the feature vectors for haze-free or hazy images, and xi denotes the 12-dimensional feature vector extracted from the i-th image patch. The higher the FADE score, the more the image resembles a hazy image.

4.3. Comparison Experiment of Synthetic Dataset

The data presented in Table 1 demonstrate that the traditional Dark Channel Prior (DCP) algorithm underperforms in dehazing metrics, with its PSNR (18.7 dB) and SSIM (0.76) significantly below the benchmark values of 20 dB and 0.8, respectively. In contrast, classical deep learning-based models like DehazeNet and AOD-Net have achieved substantial improvements in both metrics. Notably, the MSBDN model enhanced PSNR to 29.73 dB/27.75 dB and SSIM to 0.954/0.935 through architectural optimizations. Of particular significance, the proposed method in this study achieves substantial improvements of 0.53 dB (indoor) and 1.64 dB (outdoor) in PSNR, along with 1.4% and 1.7% gains in SSIM, respectively, while maintaining model compactness. These results validate that our novel dehazing network more accurately reconstructs textural details and color distribution in haze-free images compared with MSBDN [42].
The performance comparison between the proposed algorithm and existing methods on the synthetic SOTS dataset is illustrated in Figure 4. The first three experimental groups showcase the dehazing results in indoor scenarios, while the following four groups present the outcomes in outdoor environments. The experimental analysis indicates that images processed by the DCP algorithm suffer from widespread brightness attenuation and are prone to color distortion, particularly in the high-light regions of the original hazy images. Although DehazeNet effectively removes haze, residual haze remains visible in some samples, accompanied by noticeable chromatic aberrations. AOD-Net improves upon these methods but tends to cause overexposure in darker regions. Among the MSBDN [42], FSAD-Net [43], and PAN [44] algorithms, MSBDN demonstrates superior performance in both subjective and objective evaluations. While MSBDN achieves better dehazing performance, comparative analysis with our proposed algorithm and haze-free ground-truth images reveals its limitations in restoring color in local regions. Both subjective and objective evaluations confirm that the proposed algorithm outperforms the comparative methods in terms of visual quality and quantitative metrics, providing restoration results that more closely resemble real-world scenes. Notably, our method demonstrates exceptional performance in preserving color naturalness.
To emphasize the superiority of the proposed algorithm in dehazing performance, Figure 5 uses a local magnification comparative approach for validation. High-resolution image comparisons reveal significant discrepancies at the microscopic level in traditional methods, with the DCP algorithm showing structural distortions, as specifically illustrated in the magnified regions. Although DehazeNet and AOD-Net effectively remove haze to some extent, their generated images still exhibit insufficient color saturation compared with haze-free ground-truth samples, particularly in high-density haze regions where haze interference is not fully eliminated. In contrast, the proposed algorithm demonstrates remarkable advantages in restoring texture details while maintaining chromatic consistency. It not only effectively removes haze but also preserves visual fidelity in complex scenarios.

4.4. Comparison Experiment of Real Dataset

To systematically validate the effectiveness of the proposed method, a multi-dimensional comparative experimental framework was constructed under real-world hazy conditions. As illustrated in Figure 6, the experimental results highlight the key performance limitations of existing algorithms. The DCP algorithm exhibits noticeable deficiencies in color fidelity, including pronounced color gamut compression in sky and road regions (Figure 6b,c), along with residual haze along edge contours (Figure 6a). DehazeNet demonstrates partial dehazing effectiveness under light-haze conditions but fails to fully remove haze in dense regions (Figure 6b). Although AOD-Net shows generally strong performance, it introduces localized artifacts in areas with complex road textures (Figure 6c). The MSBDN algorithm struggles to handle non-uniform haze distributions, resulting in abnormal haze concentration gradients, as observed in Figure 6a,d. The FSAD-Net algorithm tends to produce excessive hierarchical differences, leading to overexposure in regions with significant color gamut variations (Figure 6a,d). The PAN algorithm, on the other hand, fails to achieve satisfactory dehazing performance in the edge regions of the image (Figure 6a,b). Subjective evaluations further confirm that the proposed method outperforms both traditional approaches and state-of-the-art deep learning models in critical aspects such as detail preservation and chromatic consistency, demonstrating its practical superiority in real-world scenarios.
To better perform quantitative evaluation on real-world images, we introduced the FADE algorithm. A lower FADE evaluation value indicates that the image is closer to a haze-free image. As shown in Table 2, our method achieves the best performance on Figure 6a,b,d, and the second-best performance on Figure 6c. This further confirms the superior performance of our algorithm in practical applications.

4.5. Comparative Experiment on Params and FLOPs

To assess the practical applicability of the proposed algorithm, we conducted a comparative analysis of the parameter count and floating point operations (FLOPs) across various dehazing models, as summarized in Table 3. Among the existing models, DehazeNet and AOD-Net demonstrate lower computational complexity due to their relatively small parameter and FLOP counts. However, this simplicity may compromise their ability to effectively handle haze with varying densities. On the other hand, models like MSBDN and FSAD-Net involve significantly higher parameter counts and FLOPs, which can limit their deployment in resource-constrained scenarios. Although the PAN algorithm exhibits a similar computational footprint to our method, its dehazing performance remains suboptimal. In contrast, the proposed model achieves a more favorable balance between restoration quality and computational efficiency, underscoring its practical advantages for real-world applications.

4.6. Comparative Experiment on Attention Mechanisms

To evaluate the optimization efficacy of the interactive attention mechanism, this study established a comparative analysis framework based on the SOTS dataset. In a controlled variable experimental design, four representative attention modules—SE [35], ECA [45], CBAM [46], and SK [47]—were embedded into a unified network architecture for benchmark testing. The quantitative evaluation data in Table 4 demonstrate that the ICA module achieves a peak structural similarity index (SSIM) of 0.968 and a suboptimal peak signal-to-noise ratio (PSNR) of 30.26 dB. Although ICA yields a slightly lower PSNR compared with SK, it offers greater flexibility for subsequent image processing tasks. Compared with mainstream attention mechanisms, the ICA mechanism exhibits significant advantages in haze-line estimation accuracy and detail reconstruction capability through its multi-scale feature interaction strategy.

5. Conclusions

This paper presents ICAFormer, a novel Transformer-based dehazing framework that integrates an interactive channel attention (ICA) mechanism to address the challenges of feature correlation in single image dehazing. By introducing a bidirectional cross-layer attention structure, the model effectively balances global context modeling with local texture enhancement. Additionally, the proposed multi-scale feature fusion strategy further strengthens the network’s ability to preserve fine details and restore color fidelity in dense haze regions.
Extensive experiments on both synthetic and real-world datasets demonstrate the effectiveness of our approach, showing notable improvements over existing state-of-the-art methods in terms of visual quality and quantitative metrics. Our method particularly excels in challenging conditions where detail preservation and contrast recovery are critical.
While the model achieves strong performance, we also acknowledge the remaining challenges under extreme conditions such as highly non-uniform haze or complex lighting. These limitations provide valuable directions for future research, which we aim to explore through lightweight architectures and video-based extensions as outlined in the next section.

6. Future Work

In future work, we plan to integrate attention mechanisms from Transformer architectures into more lightweight models to enhance the applicability of the proposed framework. We aim to investigate its implementation in real-time inference for camera-captured images and further explore its extension to video dehazing tasks, thereby providing robust technical support for practical deployment. This direction will emphasize optimizing computational efficiency while maintaining restoration fidelity, particularly addressing temporal consistency and dynamic scene adaptation in video-based applications.
Moreover, we recognize that extreme dehazing scenarios—such as those involving non-uniform haze distribution, complex illumination conditions, or dense occlusions—remain challenging. To this end, we intend to explore improved domain adaptation techniques, robustness-oriented training strategies, and hybrid modeling approaches that combine physical priors with deep features. These advancements are expected to improve generalization in uncontrolled real-world environments and further enhance the reliability of the model under diverse operational conditions.

Author Contributions

Conceptualization, T.Y. and H.H.; methodology, T.Y.; software, T.Y.; validation, T.Y., H.H. and Y.C.; formal analysis, T.Y. and Y.C.; investigation, T.Y. and Y.Z.; resources, T.Y. and T.L.; data curation, T.Y. and Y.L.; writing—original draft preparation, T.Y.; writing—review and editing, T.Y., Y.C. and P.A.; visualization, T.Y.; supervision, T.Y. and Y.C.; project administration, T.Y. and Y.C. All authors have read and agreed to the published version of the manuscript.

Funding

This study was funded by the National Natural Science Foundation of China (62176189) and the Graduate Innovative Fund of Wuhan Institute of Technology (No: CX2024549).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The RESIDE dataset is made publicly available for research purposes. For more information, please refer to the website: https://sites.google.com/site/boyilics/website-builder/reside/ (accessed on 10 April 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Narasimhan, S.G.; Nayar, S.K. Vision and the atmosphere. Int. J. Comput. Vis. 2002, 48, 233–254. [Google Scholar] [CrossRef]
  2. He, K.; Sun, J.; Tang, X. Single image haze removal using dark channel prior. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 33, 2341–2353. [Google Scholar] [PubMed]
  3. Fattal, R. Dehazing using color-lines. ACM Trans. Graph. (TOG) 2014, 34, 13. [Google Scholar] [CrossRef]
  4. Ren, W.; Liu, S.; Zhang, H.; Pan, J.; Cao, X.; Yang, M.H. Single image dehazing via multi-scale convolutional neural networks. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Springer International Publishing: Cham, Switzerland, 2016; pp. 154–169. [Google Scholar]
  5. Cai, B.; Xu, X.; Jia, K.; Qing, C.; Tao, D. Dehazenet: An end-to-end system for single image haze removal. IEEE Trans. Image Process. 2016, 25, 5187–5198. [Google Scholar] [CrossRef] [PubMed]
  6. Li, B.; Peng, X.; Wang, Z.; Xu, J.; Feng, D. Aod-net: All-in-one dehazing network. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4770–4778. [Google Scholar]
  7. Zhang, H.; Patel, V.M. Densely connected pyramid dehazing network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 3194–3203. [Google Scholar]
  8. Zhang, X.; Wang, J.; Wang, T.; Jiang, R. Hierarchical feature fusion with mixed convolution attention for single image dehazing. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 510–522. [Google Scholar] [CrossRef]
  9. Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. In Proceedings of the 28th Annual Conference on Neural Information Processing Systems (NIPS 2014), Montreal, QC, Canada, 8–13 December 2014. [Google Scholar]
  10. Mirza, M.; Osindero, S. Conditional generative adversarial nets. arXiv 2014, arXiv:1411.1784. [Google Scholar]
  11. Dong, Y.; Liu, Y.; Zhang, H.; Chen, S.; Qiao, Y. FD-GAN: Generative adversarial networks with fusion-discriminator for single image dehazing. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 10729–10736. [Google Scholar]
  12. Qin, X.; Wang, Z.; Bai, Y.; Xie, X.; Jia, H. FFA-Net: Feature fusion attention network for single image dehazing. Proc. AAAI Conf. Artif. Intell. 2020, 34, 11908–11915. [Google Scholar] [CrossRef]
  13. Li, R.; Pan, J.; Li, Z.; Tang, J. Single image dehazing via conditional generative adversarial network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 8202–8211. [Google Scholar]
  14. Guo, C.L.; Yan, Q.; Anwar, S.; Cong, R.; Ren, W.; Li, C. Image dehazing transformer with transmission-aware 3d position embedding. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 5812–5820. [Google Scholar]
  15. Zhao, X.; Xu, F.; Liu, Z. TransDehaze: Transformer-enhanced texture attention for end-to-end single image dehaze. Vis. Comput. 2025, 41, 1621–1635. [Google Scholar] [CrossRef]
  16. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Houlsby, N. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  17. Liang, J.; Cao, J.; Sun, G.; Zhang, K.; Van Gool, L.; Timofte, R. Swinir: Image restoration using swin transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 1833–1844. [Google Scholar]
  18. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  19. Zamir, S.W.; Arora, A.; Khan, S.; Hayat, M.; Khan, F.S.; Yang, M.H. Restormer: Efficient transformer for high-resolution image restoration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 5728–5739. [Google Scholar]
  20. Zhou, L.; Tao, Y.; Huang, H.; Zhu, Y. Etc-Net: A space-adaptive swin transformer-based method for underwater image enhancement integrating edge sharpening and color correction. Int. J. Mach. Learn. Cybern. 2025, 16, 3311–3327. [Google Scholar] [CrossRef]
  21. Peng, L.; Zhu, C.; Bian, L. U-shape transformer for underwater image enhancement. IEEE Trans. Image Process. 2023, 32, 3066–3079. [Google Scholar] [CrossRef] [PubMed]
  22. Cai, Y.; Bian, H.; Lin, J.; Wang, H.; Timofte, R.; Zhang, Y. Retinexformer: One-stage retinex-based transformer for low-light image enhancement. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 1–6 October 2023; pp. 12504–12513. [Google Scholar]
  23. Jain, J.; Li, J.; Chiu, M.T.; Hassani, A.; Orlov, N.; Shi, H. Oneformer: One transformer to rule universal image segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 2989–2998. [Google Scholar]
  24. Cheng, B.; Misra, I.; Schwing, A.G.; Kirillov, A.; Girdhar, R. Masked-attention mask transformer for universal image segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 1290–1299. [Google Scholar]
  25. Huang, X.; Deng, Z.; Li, D.; Yuan, X. Missformer: An effective medical image segmentation transformer. arXiv 2021, arXiv:2109.07162. [Google Scholar] [CrossRef] [PubMed]
  26. Tu, D.; Min, X.; Duan, H.; Guo, G.; Zhai, G.; Shen, W. End-to-end human-gaze-target detection with transformers. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 2192–2200. [Google Scholar]
  27. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-end object detection with transformers. In European Conference on Computer Vision; Springer International Publishing: Cham, Switzerland, 2020; pp. 213–229. [Google Scholar]
  28. Li, Y.; Mao, H.; Girshick, R.; He, K. Exploring plain vision transformer backbones for object detection. In European Conference on Computer Vision; Springer Nature Switzerland: Cham, Switzerland, 2022; pp. 280–296. [Google Scholar]
  29. Xu, Y.; Zhang, J.; Zhang, Q.; Tao, D. Vitpose: Simple vision transformer baselines for human pose estimation. Adv. Neural Inf. Process. Syst. 2022, 35, 38571–38584. [Google Scholar]
  30. Li, W.; Liu, H.; Tang, H.; Wang, P.; Van Gool, L. Mhformer: Multi-hypothesis transformer for 3d human pose estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 13147–13156. [Google Scholar]
  31. Zheng, C.; Zhu, S.; Mendieta, M.; Yang, T.; Chen, C.; Ding, Z. 3d human pose estimation with spatial and temporal transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 11656–11665. [Google Scholar]
  32. Chen, C.F.R.; Fan, Q.; Panda, R. Crossvit: Cross-attention multi-scale vision transformer for image classification. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 357–366. [Google Scholar]
  33. Bhojanapalli, S.; Chakrabarti, A.; Glasner, D.; Li, D.; Unterthiner, T.; Veit, A. Understanding robustness of transformers for image classification. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10231–10241. [Google Scholar]
  34. Lanchantin, J.; Wang, T.; Ordonez, V.; Qi, Y. General multi-label image classification with transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 16478–16488. [Google Scholar]
  35. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  36. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; Springer International Publishing: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  37. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  38. Li, B.; Ren, W.; Fu, D.; Tao, D.; Feng, D.; Zeng, W.; Wang, Z. Benchmarking single-image dehazing and beyond. IEEE Trans. Image Process. 2018, 28, 492–505. [Google Scholar] [CrossRef] [PubMed]
  39. Choi, L.K.; You, J.; Bovik, A.C. Referenceless prediction of perceptual fog density and perceptual image defogging. IEEE Trans. Image Process. 2015, 24, 3888–3901. [Google Scholar] [CrossRef] [PubMed]
  40. Guan, T.; Li, C.; Gu, K.; Liu, H.; Zheng, Y.; Wu, X.J. Visibility and distortion measurement for no-reference dehazed image quality assessment via complex contourlet transform. IEEE Trans. Multimed. 2022, 25, 3934–3949. [Google Scholar] [CrossRef]
  41. Guan, T.; Li, C.; Zheng, Y.; Wu, X.; Bovik, A.C. Dual-stream complex-valued convolutional network for authentic dehazed image quality assessment. IEEE Trans. Image Process. 2023, 33, 466–478. [Google Scholar] [CrossRef] [PubMed]
  42. Dong, H.; Pan, J.; Xiang, L.; Hu, Z.; Zhang, X.; Wang, F.; Yang, M.H. Multi-scale boosted dehazing network with dense feature fusion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 2157–2167. [Google Scholar]
  43. Zhou, Y.; Chen, Z.; Li, P.; Song, H.; Chen, C.P.; Sheng, B. FSAD-Net: Feedback spatial attention dehazing network. IEEE Trans. Neural Netw. Learn. Syst. 2022, 34, 7719–7733. [Google Scholar] [CrossRef] [PubMed]
  44. Liu, J.; Liu, P.; Zhang, Y. Multi-scale feature fusion pyramid attention network for single image dehazing. IET Image Process. 2023, 17, 2726–2735. [Google Scholar] [CrossRef]
  45. Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient channel attention for deep convolutional neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 11534–11542. [Google Scholar]
  46. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  47. Li, X.; Wang, W.; Hu, X.; Yang, J. Selective kernel networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 510–519. [Google Scholar]
Figure 1. The overall architecture of the network.
Figure 1. The overall architecture of the network.
Sensors 25 03750 g001
Figure 2. Schematic diagram of Transformer block structure.
Figure 2. Schematic diagram of Transformer block structure.
Sensors 25 03750 g002
Figure 3. Interactive attention mechanism.
Figure 3. Interactive attention mechanism.
Sensors 25 03750 g003
Figure 4. Comparison chart of indoor dehazing methods using ITS (first three sets of photos). Comparison chart of outdoor dehazing methods using OTS (last three sets of photos). From left to right are the original image, DCP, DehazeNet, AOD-Net, MSBDN, FSAD-Net, PAN, our algorithm, and ground truth.
Figure 4. Comparison chart of indoor dehazing methods using ITS (first three sets of photos). Comparison chart of outdoor dehazing methods using OTS (last three sets of photos). From left to right are the original image, DCP, DehazeNet, AOD-Net, MSBDN, FSAD-Net, PAN, our algorithm, and ground truth.
Sensors 25 03750 g004
Figure 5. Partial zoom in comparison chart of dehazing methods based on SOTS.
Figure 5. Partial zoom in comparison chart of dehazing methods based on SOTS.
Sensors 25 03750 g005
Figure 6. Comparison chart of real-world dehazing methods. From top to bottom are (ad). From left to right are the original image, DCP, DehazeNet, AOD-Net, MSBDN, FSAD-Net, PAN, and our algorithm.
Figure 6. Comparison chart of real-world dehazing methods. From top to bottom are (ad). From left to right are the original image, DCP, DehazeNet, AOD-Net, MSBDN, FSAD-Net, PAN, and our algorithm.
Sensors 25 03750 g006
Table 1. Quantitative comparison of various algorithms on RESIDE SOTS testing set.
Table 1. Quantitative comparison of various algorithms on RESIDE SOTS testing set.
MethodsIndoorOutdoor
PSNRSSIMPSNRSSIM
DCP18.720.76919.130.774
DehazeNet22.750.88621.290.877
AOD-Net24.510.91923.750.903
MSBDN29.730.95427.750.935
FSAD-Net23.410.93726.390.929
PAN23.920.92325.390.917
Ours30.260.96829.390.952
Table 2. Quantitative comparison of various algorithms on the FADE metric.
Table 2. Quantitative comparison of various algorithms on the FADE metric.
MethodsFigure 6aFigure 6bFigure 6cFigure 6d
Input1.19801.25940.96341.0366
DCP0.18900.17800.17880.1800
DehazeNet0.32370.57980.42770.3296
AOD-Net0.18900.26570.23200.2332
MSBDN0.20350.28070.21950.2762
FSAD-Net0.22740.26060.19340.1875
PAN0.26750.32640.20760.3417
Ours0.11710.14770.18300.1566
Table 3. Comparison of Params and FLOPs for various algorithms on RESIDE SOTS testing set.
Table 3. Comparison of Params and FLOPs for various algorithms on RESIDE SOTS testing set.
MethodsParams/×106FLOPs/×109
DCP--
DehazeNet0.0090.581
AOD-Net0.0020.115
MSBDN31.3541.54
FSAD-Net11.2750.46
PAN2.61152.20
Ours2.51725.79
Table 4. Comparison of various attention mechanisms on RESIDE SOTS testing set.
Table 4. Comparison of various attention mechanisms on RESIDE SOTS testing set.
AttentionsPSNRSSIM
SE23.420.906
ECA22.160.914
CBAM25.880.931
SK30.540.962
ICA30.260.968
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, Y.; Yue, T.; An, P.; Hong, H.; Liu, T.; Liu, Y.; Zhou, Y. ICAFormer: An Image Dehazing Transformer Based on Interactive Channel Attention. Sensors 2025, 25, 3750. https://doi.org/10.3390/s25123750

AMA Style

Chen Y, Yue T, An P, Hong H, Liu T, Liu Y, Zhou Y. ICAFormer: An Image Dehazing Transformer Based on Interactive Channel Attention. Sensors. 2025; 25(12):3750. https://doi.org/10.3390/s25123750

Chicago/Turabian Style

Chen, Yanfei, Tong Yue, Pei An, Hanyu Hong, Tao Liu, Yangkai Liu, and Yihui Zhou. 2025. "ICAFormer: An Image Dehazing Transformer Based on Interactive Channel Attention" Sensors 25, no. 12: 3750. https://doi.org/10.3390/s25123750

APA Style

Chen, Y., Yue, T., An, P., Hong, H., Liu, T., Liu, Y., & Zhou, Y. (2025). ICAFormer: An Image Dehazing Transformer Based on Interactive Channel Attention. Sensors, 25(12), 3750. https://doi.org/10.3390/s25123750

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop