Next Article in Journal
Sensor Arrays: A Comprehensive Systematic Review
Previous Article in Journal
Secure LoRa Drone-to-Drone Communication for Public Blockchain-Based UAV Traffic Management
Previous Article in Special Issue
Estimation of Flood Inundation Area Using Soil Moisture Active Passive Fractional Water Data with an LSTM Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Scale Image Defogging Network Based on Cauchy Inverse Cumulative Function Hybrid Distribution Deformation Convolution

College of Aeronautics, Nanjing University of Aeronautics and Astronautics, Nanjing 210006, China
*
Authors to whom correspondence should be addressed.
Sensors 2025, 25(16), 5088; https://doi.org/10.3390/s25165088
Submission received: 23 June 2025 / Revised: 12 August 2025 / Accepted: 12 August 2025 / Published: 15 August 2025

Abstract

The aim of this study was to address the issue of significant performance degradation in existing defogging algorithms under extreme fog conditions. Traditional Taylor series-based deformable convolutions are limited by local approximation errors, while the heavy-tailed characteristics of the Cauchy distribution can more successfully model outliers in fog images. The following improvements are made: (1) A displacement generator based on the inverse cumulative distribution function (ICDF) of the Cauchy distribution is designed to transform uniform noise into sampling points with a long-tailed distribution. A novel double-peak Cauchy ICDF is proposed to dynamically balance the heavy-tailed characteristics of the Cauchy ICDF, enhancing the modeling capability for sudden changes in fog concentration. (2) An innovative Cauchy–Gaussian fusion module is proposed to dynamically learn and generate hybrid coefficients, combining the complementary advantages of the two distributions to dynamically balance the representation of smooth regions and edge details. (3) Tree-based multi-path and cross-resolution feature aggregation is introduced, achieving local–global feature adaptive fusion through adjustable window sizes (3/5/7/11) for parallel paths. Experiments on the RESIDE dataset demonstrate that the proposed method achieves a 2.26 dB improvement in the peak signal-to-noise ratio compared to that obtained with the TaylorV2 expansion attention mechanism, with an improvement of 0.88 dB in heavily hazy regions (fog concentration > 0.8). Ablation studies validate the effectiveness of Cauchy distribution convolution in handling dense fog and conventional lighting conditions. This study provides a new theoretical perspective for modeling in computer vision tasks, introducing a novel attention mechanism and multi-path encoding approach.

1. Introduction

Single-image defogging, as a highly challenging image restoration task in computer vision, aims to reconstruct clear scene radiations from foggy observation images. Early methods based on physical priors, such as atmospheric scattering models [1] and dark channel priors [2], achieved promising preliminary results but were limited due to simplifications of assumptions about complex atmospheric conditions. With the rise of deep learning, Convolutional Neural Network (CNN) [3,4,5] architectures have gradually pushed the boundaries of dehazing performance through technological innovations such as multi-scale fusion [6], deformable convolutions [7], and attention mechanisms [8]. Recent advances in feature extraction architectures, including Multiscale DenseNet with bidirectional recurrent networks (Bi-RNN) [9] and graph-based approaches like the Neighborhood Adaptive Graph Isomorphism Network [10], have demonstrated remarkable success in handling complex spatial–spectral relationships, offering valuable insights for fog density estimation in defogging tasks. However, existing methods still face two core challenges when handling non-uniform haze (e.g., strong fog lines or regions with abrupt depth changes): (1) the fixed sampling grid of traditional convolutions struggles to adapt to the spatial heterogeneity of fog concentration; (2) the quadratic computational complexity of standard self-attention models limits their application in pixel-level dense prediction tasks.
Recently, breakthroughs in the performance of visual self-attention models [11] in high-level visual tasks has led to explorations of their application in fog removal. However, direct transplantation faces two challenges: first, the original self-attention mechanism has a spatial complexity of O (n2), and even window partitioning strategies [12] sacrifice global context modeling capabilities; second, existing visual self-attention models [13] typically use fixed-size convolution kernels to generate feature units, which cannot flexibly adapt to the dynamic distribution of multi-scale features in foggy images. An improved method for the traditional Retinex algorithm [14] was proposed by constructing a single-parameter homomorphic filter function using the Sigmoid function, effectively correcting the lighting in foggy images. However, due to the Retinex algorithm’s limitations, improvements in its performance were still limited. Yang, P, W [15] propose a lightweight image defogging method that combines pixel and frequency domain information by improving the upsampling and downsampling mechanisms through discrete wavelet transform and utilizing frequency domain features for auxiliary processing, achieving an outstanding performance of 41.11 dB on the SOTS-Indoor dataset. The Multi-Branch Taylor Transformer V2 (MB-TaylorFormerV2) [16,17] linearizes self-attention through Taylor expansion, partially alleviating the computational bottleneck. However, there is still room for improvement in terms of approximation error compensation and local–global feature synergy.
In response to the aforementioned challenges, the technical contributions of this paper can be summarized as follows: (1) For the first time, the inverse Cauchy cumulative distribution function is used in the offset generation process of deformation convolution, providing a mathematical foundation for handling non-uniform haze. A novel double-peak inverse Cauchy cumulative function is proposed, which, compared to the single-peak function, not only exhibits the advantage of heavy tails but also explicitly models multi-modal offsets, significantly enhancing the method’s expressive capability in complex deformation scenarios. (2) An innovative hybrid Cauchy–High-Speed distribution self-attention linearization method is proposed, achieving a performance similar to the original self-attention model while maintaining O (n) complexity. Innovative dynamically learning hybrid coefficients based on fog concentration are proposed, combining the complementary advantages of the two distributions to dynamically balance the representation of smooth regions and edge details. (3) A multi-path collaborative tree-based processing pipeline is designed to achieve adaptive extraction and fusion of multi-scale features. Through the above novel and improved strategies, efficient and high-quality defogging capabilities are achieved.

2. Theory and Methods

2.1. Cauchy and Inverse Convolution Deformation Model

2.1.1. Mathematical Foundations

The core limitations of traditional CNN convolution kernels lie in their rigid locality constraints: 1. Spatial Restrictions: their fixed grid sampling (e.g., 3 × 3) only captures neighborhood interactions, failing to model large-scale geometric deformations (e.g., rotation and perspective), leading to artifacts in depth discontinuity regions of foggy images. 2. Single-Scale Processing: each convolutional kernel operates on a fixed scale, requiring manually designed multi-scale architectures to increase complexity. 3. Channel Isolation: independent computation across channels ignores spectral/semantic correlations, limiting accuracy in tasks like hyperspectral classification. The Cauchy convolution kernel overcomes these limitations through probabilistic deformation modeling: 1. Heavy-Tailed Offsets: this generates continuous spatial offsets (e.g., ±15 pixels) via the Cauchy inverse CDF, supporting extreme deformations like 60° rotation. 2. Dynamic Multi-Scale Adaptation: the scale parameter adaptively adjusts the sampling density, enabling multi-scale feature processing in a single layer. 3. Cross-Channel Collaboration: the offset generation network implicitly learns inter-channel dependencies and upgrades discrete linear transformations into a probability-driven continuous geometry processor.
The Cauchy inverse cumulative distribution function [18] is integrated into the offset generation process of deformable convolutions. The long-tail characteristics of the Cauchy distribution are utilized to enhance the model’s ability to model extreme deformations such as large rotations and perspective transformations, while adaptive adjustment is achieved through distribution parameters.
The probability density function of the Cauchy distribution is calculated as follows:
f x ; x 0 , γ = 1 π γ 1 + x x 0 γ 2
where x 0 is the position parameter, which controls peak position; γ is the scale parameter, which controls the distribution width.
The Cauchy inverse cumulative distribution function is as follows:
F 1 p ; x 0 , γ = x 0 + γ · t a n ( π ( p 0.5 ) )
where p ( 0 , 1 ) is the uniformly distributed probability values, with the output being random variables obeying the Cauchy distribution, with heavy-tail characteristics.
The advantages of this are as follows:
  • Polynomial decay is slower.
At x , f ( x ) 1 π γ · 1 ( x / γ ) 2 = γ π x 2 O x 2 , and the exponential decay is slower than that of the Gaussian distribution. At x = 5 γ , f ( x ) value f ( 5 γ ) = 1 26 π γ 0.012 f ( 0 ) , with a significant probability density (compared to that of Gaussian distribution, which is only 2.9 × 10 7 f ( 0 )).
2.
Better dimensional flexibility.
The distribution shape is dynamically adjusted through learnable parameters.
A small γ < 0.4 is suitable for fine texture restoration and high-precision edge detection; a large γ > 0.6 , with a flat distribution and a flat heavy tail, is suitable for large-scale fog line tracking.
3.
Better gradient stability.
The Cauchy distribution gradient expression is as follows:
x f ( x ) = 2 x x 0 π γ 3 1 + x x 0 γ 2 2
where x f ( x ), the maximum value appears at x = x 0 ± γ 3 , and the amplitude is always less than x f m a x = 3 3 8 π γ 2 . This avoids the problem of gradient vanishing when x .

2.1.2. Innovative Solutions

The traditional single-peak Cauchy is extended to a double-peak form as follows:
f x ; x 0 1 , x 0 2 , γ 1 , γ 2 , w = w · 1 π γ 1 1 + x x 0 1 γ 1 2 + ( 1 w ) · 1 π γ 2 1 + x x 0 2 γ 2 2
where x 0 1 , x 0 2 is a double peak position, which is a learnable parameter; x 0 1 x 0 2 controls the peak spacing; γ 1 , γ 2 is the scale parameter, which controls the degree of overlap between the two peaks; and w [ 0 , 1 ] is the mode weight (dynamically predicted by the input features).
The sampling of the inverse cumulative distribution function of Cauchy distribution is improved by achieving double-peak sampling through mixed probability:
F 1 ( p ; x 0 1 , x 0 2 , γ 1 , γ 2 , w ) = w · x 0 1 + γ 1 t a n ( π ( p 0.5 ) ) + ( 1 w ) · x 0 2 + γ 2 t a n ( π ( p 0.5 ) )
The bimodal Cauchy distribution retains the advantages of a single peak and heavy tail while incorporating a dynamic shift mechanism. The main peak x 0 1 models the physical scattering of fog layers, while the secondary peak x 0 2 captures local fog concentration fluctuations and fog penetration at building edges. The adaptive weight w is dynamically adjusted through a fog concentration weight sensing module. By explicitly modeling multi-modal shifts, the method significantly improves its expressive capability in complex deformable scenes. As shown in Figure 1a, the bimodal Cauchy kernel excels due to its dual-peak heavy-tailed properties: Multimodal Modeling: With α = 0.6 (green dash-dot line in the plot shows dual inflection points at p ≈ 0.2 and 0.8), it captures two distinct feature patterns simultaneously, ideal for complex data (e.g., multi-object detection). Heavy-Tail Robustness: Unlike the Gaussian PPF (black line) with steep tail decay, the bimodal Cauchy’s F−1 (p) declines gently (e.g., F−1 (0.001) ≈ −4), enhancing sensitivity to outliers/extreme events. Adaptability: Adjusting α controls peak separation, balancing local feature extraction and long-range dependency modeling. As shown in Figure 1b, (1) double-peak feature extraction (orange dotted line in the figure) simultaneously captures local and mid-range features through the main peak (0 position) and secondary peak (+2 position), making it particularly suitable for processing periodic structures or symmetrical targets; (2) Controllable non-locality, with a valley design in the −2 to +1 interval to avoid feature blurring, expanding the receptive field while maintaining position sensitivity; (3) Adaptive weight balancing: The normalized weight of 0.6 for the double peak outperforms the smooth distribution of the single-peak Cauchy (blue dashed line), and its adjustable α parameter ( α = 0.6 in the figure) can flexibly adapt to feature detection requirements at different scales.

2.1.3. Operating Network Design

Given an input feature map X R B × C × H × W , the offset generation process of Cauchy deformation convolution consists of three steps, where B is the batch size, which is the number of samples processed simultaneously; C is the number of channels, indicating the depth of the feature map; H is the height (number of pixels) of an image or feature map; and W is the width (number of pixels) of an image or feature map:
  • Basic offset: the initial coarse offset is obtained by sampling through the inverse convolution function:
    F b a s e 1 p ; x 0 1 , x 0 2 , γ 1 , γ 2 , w = w · x 0 1 + γ 1 t a n ( π ( p 0.5 ) ) + ( 1 w ) · x 0 2 + γ 2 t a n ( π ( p 0.5 ) )
  • Adaptive correction: a lightweight CNN with 3/5/7 layers (number of parameters < 1 k) is used to predict pixel-wise correction coefficients.
    F final 1 p ; x 0 1 , x 0 2 , γ 1 , γ 2 , w = F b a s e 1 p ; x 0 1 , x 0 2 , γ 1 , γ 2 , w ( 1 + α ) + β
    where α , β are predicted by the correction network and constrained within the range [ 0.5 , 0.5 ] , α adjusts the amplitude of the offset and enhances adaptability to high-frequency details, β compensates for system bias and solves the center offset problem caused by the symmetry of the Cauchy distribution; and denotes element-wise multiplication.
  • Design of adaptive truncation strategy constraints.
To ensure training stability and prevent numerical explosion of the tan (·) function when p 0 / 1 , controllable deformation is performed, with a threshold τ and double-peak scale γ 1 , γ 2 parameters that are dynamically associated to ensure that, in low fog concentration regions ( γ ), larger deformation exploration is allowed, and in high fog concentration regions ( γ ), constraints are strengthened to avoid overfitting. Therefore, an adaptive truncation strategy is designed:
F clipped 1 p ; x 0 1 , x 0 2 , γ 1 , γ 2 , w = F final 1 p ; x 0 1 , x 0 2 , γ 1 , γ 2 , w F final 1 p ; x 0 1 , x 0 2 , γ 1 , γ 2 , w 2 τ τ F final 1 p ; x 0 1 , x 0 2 , γ 1 , γ 2 , w F final 1 p ; x 0 1 , x 0 2 , γ 1 , γ 2 , w 2 otherwise
where the threshold τ = 2 ( w · γ 1 + 1 w · γ 2 ) is dynamically adjusted during training. The L2 norm of the output F 1 p ; x 0 1 , x 0 2 , γ 1 , γ 2 , w of the inverse cumulative distribution function is calculated as follows: F 1 p ; x 0 1 , x 0 2 , γ 1 , γ 2 , w 2 = w · x 0 1 + γ 1 tan π p 0.5 + ( 1 w ) · [ x 0 2 + γ 2   tan ( π ( p 0.5 ) ) ] 2 .

2.2. Theory and Implementation of the Cauchy–Gaussian Hybrid Distribution Attention Mechanism

2.2.1. Mathematical Foundations

The hybrid distribution attention mechanism adopts a dual-branch parallel structure, which dynamically fuses global context and local detail features through weight fusion:
Attention hybrid = λ · Attention Cauchy   + ( 1 λ ) · Attention Gauss
where λ 0 , 1 is a learnable hybrid coefficient, which is adaptively adjusted through Formula (19).
This design addresses the specific requirements of the fog removal task: 1. The Cauchy branch models long-range dependencies to address the discontinuity of the fog concentration spatial distribution. 2. The Gaussian branch captures local texture details to maintain edge sharpness. 3. Dynamic fusion automatically adjusts weights based on the fog concentration distribution.

2.2.2. Innovative Cauchy Attention Branch

(1)
Linearization Implementation Method
By approximating the Cauchy kernel function using a Taylor series expansion [19], the original O n 2 complexity is reduced to O ( n ):
K Cauchy x i , x j = 1 1 + x i x j 2 η 2 k = 0 m   ( 1 ) k η 2 k x i x j 2 k
where m is the Taylor expansion order, balancing accuracy and efficiency; when m = 2 , the associative property of matrix multiplication is used to achieve linear computation:
Attention Cauchy = S o f t m a x Q K T K Q T η 4 2 Q K T η 2 + 1 · V
where η is the scale parameter, controlling the attention decay rate, with an initial value of 0.5; Q R B × h × H × W × d is the query matrix, representing the features of the target location; K R B × h × H × W × d is the key matrix, representing the features of the source positions; V R B × h × H × W × d is the value matrix, representing the feature information to be aggregated; S o f t m a x ( x ) = e x i i   e x i is the activation function used for multi-class classification problems.
Q K T K Q T η 4 captures the global second-order statistics of the feature space, enhancing the model’s ability to perceive complex patterns; 2 Q K T η 2 models local linear correlations, with the negative sign ensuring complementarity with the higher-order terms; +1 prevents the kernel function values from becoming too small, which could cause the gradients to vanish.
The value of η varies as shown in Equation (12):
η = S o f t p l u s w η T P o o l ( Q + K ) + b η
where w η T R h × d is the weight vector; d is the feature dimension of each attention head, where d = C / h C is the total number of channels; h is the number of attention heads, initialized as a Xavier normal distribution N ( 0 , 1 / d ) and which performs a linear projection on the pooled features; P o o l   ( Q + K ) R B × h × H × W × d is the global average pooling function 1 L i = 1 L   Q i + K i ; b η is the bias term, adjusting the baseline value of the parameters, and is a scalar; and S o f t p l u s R B × h is the activation function ln (1 + e x ).
(2)
Gradient constraint conditions
The upper bound of the Cauchy kernel gradient is as follows:
η K Cauchy 2 η 3 k = 1 m   k 1 1 η 2 k 1
When η > 1 , this ensures that the gradient is bounded and avoids training oscillations.
Stability condition:
η Attention   4 η 3 Q K + 4 η 5 Q K T K

2.2.3. Innovative Gaussian Attention Branch

(1)
Multi-Scale Modulation Mechanism
A pyramid attention mechanism is constructed using a three-level Gaussian kernel [20] (σ = 1, 3, 5):
Attention Gauss = s { 1,3 , 5 }   w s · S o f t m a x   Q K T σ s d k V
where w s is the mixed weight of the s-th level Gaussian kernel, which is dynamically generated by the input features, where w s = 1 .
A novel gradient-aware weight allocation parameter is dynamically generated and introduced [21]:
w s = S i g m o i d   ( M L P   ( a v g p o o l   ( Q ) , Q ) )
where a v g p o o l   ( Q ) is the global average pooling of the query matrix a v g p o o l ( Q ) = 1 L i = 1 L   Q i ; Q is the gradient magnitude of the query matrix Q = ( Q / x ) 2   +   ( Q / y ) 2 ;   M L P · is a multi-layer perceptron that maps the pooled features to the scale weight space R 2 d R 3 ; S i g m o i d compresses the MLP output to the [ 0 , 1 ] interval R 3 [ 0 , 1 ] 3 .
σ s = 1 : this focuses on a 3 × 3 local window, primarily achieving edge sharpening and texture restoration; σ s = 3 : this covers a 7 × 7 region, primarily for medium-range structures; σ s = 5 : this ensures consistent fog density through global correlation.
(2)
Gaussian Receptive Field Constraint Strategy
The effective range of Gaussian attention is restricted to the local neighborhood via a masking mechanism (15):
M i j = 1 i f p i p j 2 r 0 otherwise
where p i , p j are spatial position coordinates and pixel coordinates; the receptive field radius r is scale-adaptive; M i j = 1 indicates that position i is allowed to focus on position j , conforming to the locality principle; and M i j = 0 indicates that distant irrelevant regions are masked, reducing the computational redundancy and noise interference.
An innovative dynamic radius adjustment strategy is introduced. In high-entropy regions, where textures are complex, the radius is expanded to capture more context; in low-entropy regions, which are flat areas, the radius is reduced to minimize redundant computations:
r = 3 σ s ( 1 + α · E n t r o p y Q l o c a l )
where α is the entropy adjustment intensity coefficient, which controls the influence of local texture complexity on the radius, and is a learnable parameter through Sigmoid activation obtained by dividing into local blocks through the unfold operation; Q l o c a l is the local query feature block centered at the current position, R B × h × 3 × 3 × d ; E n t r o p y Q l o c a l is the information entropy of local features, which measures the texture complexity, R B × h × H × W ; the formula to calculate this is k = 1 9   p k l o g p k , p k = S o f t m a x Q local ( k ) .

2.2.4. Cauchy–Gauss Branch Dynamic Fusion Module

(1)
Hybrid Coefficient Generation
The innovative fog concentration-aware weight prediction mechanism is as follows:
λ = σ ( C o n v 3 × 3   ( R e L U   C o n v 1 × 1   F m a x ; F a v g ) )
where F m a x and F a v g are the maximum/average pooling features of the input feature map F (along the spatial dimension), mathematically expressed as R C × H × W ;   C o n v 1 × 1   i s   1 × 1 convolution, with dimension reduction and fusion of global–local features, with mathematical form R 2 C × H × W R C × H × W ;   R e L U is a nonlinear activation function that enhances feature separability, with mathematical form R C × H × W R C × H × W ; C o n v 3 × 3 is 3 × 3 convolution that captures spatial context relationships, mathematically expressed as R C × H × W R 1 × H × W ;   σ ( · ) is the Sigmoid function activation, which constrains the output to probability weights, and is mathematically expressed as Sigmoid R H × W [ 0 , 1 ] H × W .
(2)
Gradient problem analysis of mixed operations
Gradient distribution for mixed operations:
L λ = Attention Cauchy Attention Gauss · L Output
where L λ is the gradient of the dynamic weight λ , guiding its update direction, with the mathematical form R B × H × W ; Attention Cauchy is the output of the Cauchy attention branch, which excels at modeling long-range fog concentration dependencies, with the mathematical form R B × C × H × W ; Attention Gauss is the output of the Gaussian attention branch, which excels at capturing local details, and is mathematically expressed as R B × C × H × W ; L Output is the gradient of the downstream task (such as dehazing loss) on the mixed output and is mathematically expressed as R B × C × H × W and obtained through backpropagation of the loss function.
When the Cauchy branch performs better than the Gaussian branch, ( Attention Cauchy Attention Gauss ) > 0, the forward gradient increases λ to enhance the contribution of the Cauchy branch; otherwise, λ is reduced to favor the Gaussian branch.
To control gradient competition, the Cauchy branch is prevented from being blocked by λ and suppressing the Gaussian branch, and to improve the training stability, a new method is proposed to prevent λ from over-competing through the stop-gradient strategy.
L λ = sg   ( Attention Cauchy ) Attention Gauss · L Output
where sg   ( x ) = x is the forward propagation, s g ( x ) x = 0 is the backward propagation, the input and output are completely the same during forward propagation, and no values are changed. During backward propagation, the gradient is forced to zero, and the upstream gradient is truncated here.

2.2.5. Operating Network Design

The mixed distribution attention mechanism combines the advantages of the Cauchy distribution and the Gaussian distribution to optimize the processing of input feature maps X R B × C × H × W . The specific process is as follows:
  • Dual-branch parallel processing:
The Cauchy branch uses a linearization approximation method to convert the nonlinear Cauchy distribution into a piecewise linear representation through Taylor expansion, significantly reducing the computational complexity. Simultaneously, gradient constraints are introduced to ensure numerical stability during backpropagation and to avoid gradient explosion.
2.
The Gaussian branch:
Extracts local features using multi-scale convolution kernels (such as 3 × 3, 5 × 5, 7 × 7), followed by batch normalization layers for each scale branch to satisfy local receptive field constraints. The receptive field is expanded without increasing the number of parameters using dilated convolution. The dynamic fusion mechanism is as follows:
A lightweight gated network is designed to generate fusion weights λ [ 0 , 1 ] , with the input being the element-wise product of the two-branch features. Dual attention (channel + spatial attention) optimizes the weight distribution, with the formula λ = σ C o n v 3 × 3   R e L U   C o n v 1 × 1   F m a x ; F a v g .
A gradient clipping strategy is introduced to constrain the gradient norm in the fusion stage to not exceed the threshold τ, and projection gradient descent is used to ensure training stability.
3.
Feature reconstruction output:
The final feature Attention hybrid = λ · Attention Cauchy + ( 1 λ ) · Attention Gauss is output as R R B × C × H × W after residual connection and layerNorm.

2.3. Theory and Implementation of Tree-like Multi-Path Coding Structure

2.3.1. Principles of Bio-Inspired Design

Inspired by the hierarchical processing mechanism of the primate visual cortex V1–V4 regions [22], as shown in Figure 2, visual cortex target recognition mechanism diagram., these are divided into V1 for edge detection, V2 for shape integration, V3 for shape abstraction, and V4 for color and texture synthesis.
Based on the different functions of the V1, V2, V3, and V4 regions of the primate visual cortex, as shown in Figure 3, innovation lies in the following aspects: 1. Parallel multi-scale processing: each path corresponds to different receptive fields (3 × 3 to 11 × 11); 2. Heterogeneous feature extraction: each path independently learns the prior distribution of fog concentration; 3. Dynamic feature fusion: the brain’s attention selection mechanism is simulated.
The multi-pathway collaboration mechanism includes Path 1 (3 × 3 receptive field), which corresponds to the V1 region, extracts local fog line edges, and performs edge detection; Path 2 (5 × 5 receptive field), which corresponds to the V2 region, integrates medium-scale fog cluster structures, and completes shape integration; Path 3 (5 × 5 receptive field), which corresponds to the V3 region and performs 3D shape abstraction; Path 4 (7 × 7 receptive field), which performs color and texture acquisition; and Path 5 (11 × 11 receptive field), which corresponds to the IT region, captures cross-regional semantic associations, and completes object recognition.

2.3.2. Innovative Cauchy Deformation Bimodal Convolution

Adaptive path deformation parameters are based on different functional paths:
Δ p k = γ k · t a n h   π U 1 τ Narrow   Peak + η k · t a n   π U 2 0.5 + β · N   ( 0 , 1 ) Wide   Peak   +   Noise
where γ k = γ 0 + k · Δ γ controls the deformation intensity of the narrow peak; U 1 ~ U n i f o r m ( 0 , 1 ) is a random variable that generates the base deformation field; the τ hyperparameter is 0.3~0.5; t a n h ( · ) is a hyperbolic tangent activation function [23] with parameter range [−1, 1]; η k = η 0 · e k λ is a learnable parameter controlling the wide-peak deformation intensity; U 2 ~ U n i f o r m ( 0 , 1 ) is a random variable; t a n ( · ) is periodic function, ( , + ), generating unbounded deformation (actually constrained to ± 3 η k ); β is a hyperparameter, with a value of 0.05~0.15, and is used to control noise; N ( 0 , 1 ) is a random variable with a standard normal distribution.
A path-adaptive mechanism is adopted: In path 1 (3 × 3 receptive field), corresponding to the V1 region, a narrow peak dominant strategy is used with γ k and η k values of 0.10 and 0.50, and the τ value is 0.30, each primarily achieving edge detection. In Paths 2 and 3 (5 × 5 receptive field), corresponding to the V2 and V3 regions, a dual-peak balanced strategy is adopted. The V2 regions have γ k and η k values of 0.15 and 0.67, and the τ value is 0.40; the V3 regions have γ k and η k values of 0.175 and 0.88, and a τ value of 0.40. Each primarily achieves shape integration. In Path 4 (7 × 7 receptive field), corresponding to the V4 region, a wide-peak dominant strategy is adopted, with γ k and η k values of 0.20 and 0.90, and a τ value of 0.45, respectively, primarily achieving region fusion. Path 5 (11 × 11 receptive field) corresponds to the IT region, using a wide peak and noise as the dominant peaks, with γ k and η k values of 0.25 and 1.20, and a τ value of 0.50, respectively, to complete target recognition.

2.3.3. Dynamic Gate Fusion

(1)
Fog concentration estimation module.
A lightweight fog concentration estimator is constructed as follows:
M haze = S i g m o i d C o n v 3 × 3 R e L U C o n v 1 × 1 I in
where I i n R B × C × H × W is the input image; M haze [ 0 , 1 ] H × W is the output resolution, consistent with the input ( H × W × 1 ).
The depth cue extraction formula is as follows:
d ( x ) = 1 β l o g   1 M haze   ( x )
where β is the atmospheric scattering coefficient (default 0.8); d ( x ) is the depth map normalized to [0, 1].
The color correction formula is as follows:
I corrected c ( x ) = I in   ( x ) · 1 + α · 1 M haze   ( x ) 1 / γ
where c { R , G , B } is the color channel; α is the gain intensity, with a default of 1.5, and is set to 2.0 for the blue channel; γ is the Gamma correction parameter, with a default of 2.2.
The detail mask formula is as follows:
D detail = S o b e l I i n 1 M haze + λ · T V M haze
where S o b e l is the edge detection operator; T V is the total variation regularization term; λ is 0.1 by default.
(2)
Gate coefficient generation.
Path weight is calculated as follows:
g k = e w k T M h a z e i = 0 N 1   e w i T M h a z e · 1 1 + e α g k β g M h a z e
where w k is a learnable weight vector, with dimensions = H × W , and a spatial weight matrix of path k with an initial value N ( 0 , 2 / ( H + W ) ) using gradient backpropagation; α g is the slope factor (initial value 1.0), a scalar that can be learned; and β g is the fog concentration sensitivity (initial value 0.5), a scalar that can be learned.

2.3.4. Multi-Path Constraint Strategy

(1)
Multi-path gradient allocation.
Gradients from each path are weighted by the gate coefficient:
L θ k = g k · L   Output + α τ · g k θ k
where α τ = 0.1 is the gradient balancing coefficient.
(2)
Orthogonality Constraint Between Paths
A regularization term is added to promote path differentiation:
L orth = i < j   [ F i T F j F i 2 F j 2 2 Cosine   similarity   penalty + 0.1 · T V F i F j Gradient   correlation   penalty ]
where F i is the output feature map of path i and the dimension is R B × C × H × W ; F j is the output feature map of path i and the dimension is R B × C × H × W ;   F i 2 and F j 2 are scalars with L2 norm; F i T F j F i 2 F j 2 2 is the square operation; F i and F j are the spatial gradients of the features of paths i and j and the dimension is R B × C × H × W × 2 ;   a n d   is the Hadamard product (element-wise multiplication).

2.3.5. Detailed Implementation Steps

  • Input Feature Division
The input channels are divided evenly between the paths using 1 × 1 convolution:
I i n k = I i n : , k C N : ( k + 1 ) C N , k = 0 , 1 , 2 , 3 , 4
where I i n is the input feature map with dimensions is R B × C × H × W ;   C is the total number of input channels; and I i n k is the feature slice assigned to the k path.
2.
Independent processing of paths
The data processing formula for each path is as follows:
F k = P a t h B l o c k k I i n k = G r o u p N o r m S h u f f l e C o n v C a u c h y D e f o r m C o n v I i n k
where C a u c h y D e f o r m C o n v is an innovative Cauchy deformation double-peak convolution function, calculated with Formula (32):
C a u c h y D e f o r m C o n v I i n k = i = 1 N k   w i · I i n k p 0 + Δ p i · Δ m i
where I i n k p 0 + Δ p i is the eigenvalue of the input feature map I i n at p 0 + Δ p i ; p 0 is the integer coordinate of the current convolution center point in the two-dimensional coordinate ( x , y ); Δ p i is the two-dimensional coordinate offset using Formula (22) and the offset of the I i n sampling point; Δ m i = S i g m o i d M L P I i n k : , p 0 is the modulation factor, with an initial value of 1, constrained by the Sigmoid function; m i { 0 , 1 } , m i = 1 if   Δ p i close   to   + δ 0 if   Δ p i close   to   δ ,   δ is the double-peak spacing, w i = S o f t m a x C o n v 1 × 1 I i n k : , p 0 i is a dynamic scalar and is the weight of the i sampling point, i = 1 N k   w i = 1 ; N k is the total number of sampling points, N k = 9 for a 3 × 3 convolution kernel.
S h u f f l e C o n v is a channel shuffle convolution that enhances cross-group information flow, as shown in Formula (33):
S h u f f l e C o n v ( I i n c k ) = R e L U ( C o n v ( P ( I i n c k ; G ) , K H , groups = G ) )
where P ( I i n c k ;   G ) is the channel shuffle operation (divides C channels into G  groups and permutes them); K H is the convolution kernel.
G r o u p N o r m is the group normalization function, defined as (34):
G r o u p N o r m I i n c s k = γ g n · I i n c s k μ g σ g 2 + ϵ + β g n
where μ g and σ g are the mean/variance of each group of channels; γ g n , β g n R C are learnable affine parameters; and ϵ = 10 5 is a numerical stability term.
3.
Dynamic feature fusion
The output result is the weighted sum of the features of each path:
I out = k = 0 N k 1   g k · C o n v 1 × 1 F k
where N k is the number of paths plus 1; F k is the output feature of the k path; g k is the dynamically generated weight of the k path.

3. Operational Process and Innovative Advantages

3.1. Operational Process

The model adopts a U-Net-shaped encoder–decoder structure, achieving multi-scale feature extraction through four-stage downsampling and upsampling. It innovatively introduces the Cauchy distribution into deformation convolution, constructing a three-stage processing workflow: “Cauchy deformation convolution—multi-scale Cauchy–Gaussian attention—tree-like multi-path encoding hybrid distribution fusion.”
As shown in Figure 4, the algorithm runs as follows (Algorithm 1):
Algorithm 1 EnhancedMB-CauchyFormer Pseudocode
  I—Input: Image tensor (B, C, H, W),—Apply 3x3 convolution with stride 1 and padding 1, O—Output: Embedded patches (B, embed_dim, H, W)
1: function EnhancedMB-CauchyFormer(I):
2:    x = I // Receive raw image input
3:    x = OverlapPatchEmbed(x) // Initial feature extraction [B, C0, H, W]
4:    // --- Level 1 ---
5:    x1_list = Patch_Embed_stage_Cauchy(x, num_path=N1) // Generate N1 paths [B, C0, H, W] × N1
6:    x1 = EnhancedMHCA_stage(x1_list) + x // Tree-based multi-head attention + residual connection
7:    // --- Level 2 ---
8:    x2 = Downsample(x1) // Downsampling [B, C1, H/2, W/2]
9:    x2_list = Patch_Embed_stage_Cauchy(x2, num_path=N2)
10:   x2 = EnhancedMHCA_stage(x2_list) + x2
11:   // --- Level 3 ---
12:   x3 = Downsample(x2) // Downsampling [B, C2, H/4, W/4]
13:   x3_list = Patch_Embed_stage_Cauchy(x3, num_path=N3)
14:   x3 = EnhancedMHCA_stage(x3_list) + x3
15:   // --- Level 4 (Latent) ---
16:   x4 = Downsample(x3) // Downsampling [B, C3, H/8, W/8]
17:   x4_list = Patch_Embed_stage_Cauchy(x4, num_path=N4)
18:   x4 = EnhancedMHCA_stage(x4_list) + x4
19:   // ======== 4. Decoder Upsampling Path ========
20:   // --- Level 3 ---
21:   x3_up = Upsample(x4) // Upsampling [B, C2, H/4, W/4]
22:   x3_up = Concat(x3_up, x3) // Concatenate with encoder features
23:   x3_up = ReduceChannels(x3_up) // 1 × 1 convolution for dimension reduction
24:   x3_up_list = Patch_Embed_stage_Cauchy(x3_up, num_path=N3)
25:   x3_up = EnhancedMHCA_stage(x3_up_list) + x3_up
26:   // --- Level 2 ---
27:   x2_up = Upsample(x3_up) // Upsampling [B, C1, H/2, W/2]
28:   x2_up = Concat(x2_up, x2)
29:   x2_up = ReduceChannels(x2_up)
30:   x2_up_list = Patch_Embed_stage_Cauchy(x2_up, num_path=N2)
31:   x2_up = EnhancedMHCA_stage(x2_up_list) + x2_up
32:   // --- Level 1 ---
33:   x1_up = Upsample(x2_up) // Upsampling [B, C0, H, W]
34:   x1_up = Concat(x1_up, x1)
35:   x1_up_list = Patch_Embed_stage_Cauchy(x1_up, num_path=N1)
36:   x1_up = EnhancedMHCA_stage(x1_up_list) + x1_up
37:   // ======== 5. Refinement Stage ========
38:   x_refine_list = Patch_Embed_stage_Cauchy(x1_up, num_path=N1)
39:   x_refine = EnhancedMHCA_stage(x_refine_list) + x1_up
40:   // ======== 6. Output Processing ========
41:   if dual_pixel_task:
42:      x_refine = x_refine + SkipConv(x) // Skip connection
43:     O = OutputConv(x_refine) // 3-channel output
44:   else:
45:      O = OutputConv(x_refine) + I // Residual output
46: return O

3.2. Methodological Advantages

This network proposes a bio-inspired multi-path defogging architecture, the core algorithm of which consists of three innovative modules. The workflow can be divided into three stages: feature decoupling, multi-scale fusion, and dynamic reconstruction.

3.2.1. Cauchy Deformable Convolution

Cauchy deformable convolution (CDC), as an innovative module of the Enhanced Multi-Branch Cauchy Transformer Network (EnhancedMB-CauchyFormer) algorithm, offers the following significant advantages over traditional deformable convolution (DCN) and standard convolution: 1. In image restoration tasks (such as dehazing and denoising), the displacement of edges and texture regions often suddenly changes. The Cauchy inverse accumulation function can better adapt to such irregular deformations. The double-peak Cauchy inverse accumulation distribution retains the advantages of a single-peak heavy tail while explicitly modeling multi-modal displacement, significantly enhancing the method’s expressive capability in complex deformation scenarios. 2. The heavy-tail characteristic of the Cauchy inverse accumulation distribution function makes it more robust to outliers. 3. An adaptive correction method is used for continuous learning to enhance the algorithm’s adaptability.

3.2.2. Innovative Dual-Distribution Hybrid Mechanism

As described in Section 2.2, by integrating the Cauchy inverse cumulative distribution function and the Gaussian distribution, and automatically adjusting the weights of the two distributions through learnable parameters λ, we propose a differentiable dual-distribution sampling mechanism that combines mixed probability distributions with feature fusion for the first time, and derive a Cauchy–Gaussian hybrid gradient formula suitable for deep learning.
A lightweight fuzzy detection network (only two convolutional layers) is designed to achieve feature enhancement with degraded perception. Gradient stabilization techniques are developed to address the numerical instability of the Cauchy distribution during backpropagation. Compared with other attention fusion methods, this approach offers the best robustness. The fuzzy region is explicitly modeled, resulting in strong anti-fuzzy performance. Training is stable and does not require fine-tuning, as the system automatically learns to balance parameters, and the computational overhead is relatively small.

3.2.3. Innovative Tree-Based Self-Attention Model Block

For parallel branch processing, the total computation of four branches is only one-fourth that of the standard self-attention model; window attention locality avoids global attention matrix computation; blur mask guidance automatically enhances high-frequency features in motion blur regions; and thick-tail distribution robustness enables Cauchy branches to effectively suppress blur artifacts.

4. Experimental Setup and Result Analysis

4.1. Algorithm Setup

We selected the Dark Channel Prior (DCP) [1], Improved All-in-One Dehazing Network (AOD-Net) [24,25], Multi-Scale Boosted Dehazing Network with Dynamic Feature Fusion (MSBDN-DFF) [26], Feature Fusion Attention Network (FFA-Net) [27], Revitalizing Convolutional Network (ConvIR) [28], Frequency Domain Assistance and Detailed Brightness Information Guidance (FIGD-Net) [15], Multi-Branch Taylor Transformer V2 (MB-TaylorFormerV2) [16,17], and Enhanced Multi-Branch Cauchy Transformer Network (EnhancedMB-CauchyFormer) for the comparative analysis. The input parameters of the algorithms were uniformly configured.
Model Architecture: EnhancedMB-CauchyFormer, number of parameters: 15.8 M; computational complexity: 96.8GFlops; employs 5-branch dynamic weight allocation + multi-scale Cauchy feature fusion. We set the initial learning rate to 2 × 10−4 and gradually reduced it to 1 × 10−6 using sine annealing [29].
Training Configuration Data Augmentation: random cropping to 256 × 256 pixels (overlap rate ≥ 30%); probabilistic horizontal flipping (p = 0.5); color space perturbation (ΔHSV∈ [0.1, 0.3]).

4.2. Experiment on Synthetic Blur and Real Blur Images

Using the data set in Table 1, the test results are shown in Table 2. In the RESIDE benchmark test, the EnhancedMB-CauchyFormer demonstrated excellent performance improvements: its Cauchy–Gaussian hybrid attention mechanism maintained robustness even under extreme fog concentrations (Fog concentration >0.8). In synthetic haze scenes, the EnhancedMB-CauchyFormer achieved a peak signal-to-noise ratio (PSNR) of 45.12 dB and a Structural Similarity Index Measure (SSIM) of 0.997 with 15.80 million parameters in indoor scenes (SOTS-Indoor), representing a 5.27% improvement in the PSNR compared to that with MB-TaylorFormerV2, as shown in the experimental comparison in Figure 5, where GT denotes the original image. This demonstrates the modeling advantages of the Cauchy distribution for synthetic haze, enabling better image restoration. In outdoor scenes (SOTS-Outdoor) under complex weather conditions, the adaptability to outdoor scenes was improved. EnhancedMB-CauchyFormer achieved a PSNR of 39.55 dB, surpassing that of MB-TaylorFormerV2 (39.25 dB) by 0.25 dB, as shown in Figure 6, demonstrating robustness to changes in outdoor lighting. In real haze scenes, on the O-HAZE dataset, the PSNR value decreased by 1.14, and the SSIM value decreased by 0.05, as shown in the experimental comparison in Figure 7. This dataset maintains its structural similarity advantage. In the Dense-Haze dataset, the PSNR improved by 0.12 dB, representing a 0.68% increase, as shown in Figure 8, indicating that the proposed method has superior structural restoration capabilities in high-concentration haze compared to those of existing techniques. The lowest performance-to-computation ratio (PSNR/MACs) among the state-of-the-art methods was 0.466, outperforming MB-TaylorFormerV2’s 0.498, demonstrating superior computational efficiency compared to that of similar attention-based models.

4.3. Ablation Studies

All of the models were trained on the SOTS-Indoor dataset, with an input consisting of randomly cropped 256 × 256 pixel patches, a batch size of 32, and the Adam with Weight Decay optimizer ( β 1 = 0.9 , β 2 = 0.999 ). The base learning rate was set to 2 × 10−4 and was decayed to 1 × 10−6 over 500 epochs using sine annealing scheduling. Based on the EnhancedMB-CauchyFormer architecture, the effects of the Cauchy variable convolution layer, hybrid attention mechanism, and tree-like multi-branch architecture were analyzed sequentially, and the synergistic effects of the modules were examined.

4.3.1. Cauchy Deformable Convolution Layer

As can be seen from Table 3, the Cauchy deformable convolutions achieved a further gain of 0.25 dB over the Gaussian deformable convolutions, and their depth-separable Cauchy convolutions more accurately modeled the physical distribution characteristics of fog.

4.3.2. Cauchy–Gauss Hybrid Attention Mechanism

As can be seen from Table 4, the Cauchy–Gauss kernel approximation, compared to the standard attention mechanism, improves the PSNR by 1.88 dB and the SSIM from 0.991 to 0.995, demonstrating its superior structural preservation capability. The Cauchy kernel approximation error, achieved through a third-order Taylor expansion, is only 0.021, significantly outperforming linear attention (0.148), making it closer to the ideal attention concentration modeling capability. The long-range modeling of the Cauchy distribution’s long-tail characteristics achieves a cross-image block dependency strength of 0.92, which is 2.9 times that of linear attention, making it particularly beneficial for global consistency restoration in heavily fogged regions.

4.3.3. Tree-like Multi-Path Branch Architecture

As shown in Table 5, the effectiveness of multi-path processing and fusion is demonstrated. The five branches represent the five paths of the tree, and adaptive feature selection is achieved through a gating mechanism. The coupling of edge detection (L1) and shape integration (L2) improves the contour PSNR by 2.3% (44.98 dB vs. 43.95 dB), validating the guiding role of mid-frequency information in edge restoration. At the color and texture recognition (L4) turning point, after introducing color and texture paths, the SSIM increased from 0.994 to 0.995, indicating that color consistency can mitigate color bias issues.

4.3.4. Module Synergy Analysis

Table 6 shows a comparison of the performance of the four image processing models, illustrating the gradual improvements achieved through algorithm optimization. The baseline model (standard convolution + attention) achieves a PSNR of 40.71 dB, an SSIM of 0.991, and a computational load of just 3.2 GFLOPs, offering a lightweight advantage. Cauchy–Fourier transform introduces a frequency domain analysis, improving the PSNR by 1.16 dB (a relative increase of 26.3%) while increasing the computational complexity to 7.8 GFLOPs. This significantly enhances the defogging performance while maintaining low computational costs. The hybrid distribution attention combining Cauchy narrow peaks and Gaussian wide peaks achieves a PSNR reaching 42.59 dB (an increase of 1.88 dB) and an SSIM of 0.995, enhancing the adaptability to complex fog distributions. The final tree-based multi-branch complete model performs optimally, achieving a PSNR of 45.12 dB (an increase of 4.41 dB, or a 57.4% relative improvement) and an SSIM of 0.996, demonstrating robust multi-scale feature fusion capabilities. Although the computational load increased to 96.8 GFLOPs, the model achieved a strong improvement in accuracy and is suitable for high-precision defogging tasks. Overall, the algorithm has been progressively optimized in terms of the PSNR, SSIM, and adaptability, while maintaining a reasonable balance between computational efficiency and performance improvements.

4.4. Real Target Recognition and Verification

As shown in the target recognition heat map in Figure 9, EnhancedMB-CauchyFormer significantly improves the explainability of target recognition: the red area (0.8–1.0) accurately covers key decision areas such as the front of cars, bus doors, and the upper bodies of pedestrians; the yellow area (0.5–0.8) effectively captures auxiliary features such as wheels, windows, and the limbs of pedestrians; and the background suppression (blue < 0.5) reduces false positives.
As shown in Figure 10, for image recognition under real foggy weather conditions, the algorithm recognizes small cars, pedestrians, and buses in foggy weather conditions. It accurately identifies small pedestrian targets on bridges and precisely distinguishes between small cars and buses, demonstrating the algorithm’s advantage of multi-branch fusion to adapt to targets of different scales. The Cauchy variable convolution enhances edges and small features as well as long-tail features. Dynamic weight learning automatically adapts to simple or complex scenes, and the verification results align with the research findings.

5. Conclusions

In this paper, we propose a novel defogging algorithm inspired by a bio-inspired multi-path defogging architecture, which addresses the performance bottlenecks of traditional methods in dense fog edges and strong illumination fog lines through Cauchy inverse cumulative function variable convolution and multi-scale fusion. Based on an innovative multi-peak Cauchy inverse cumulative distribution offset generator, the algorithm precisely models sudden changes in fog concentration using heavy-tailed characteristics; a hybrid Cauchy–Gaussian module dynamically balances the representation of smooth regions and edge details; and a tree-based multi-path model identification achieves adaptive local–global feature fusion. The algorithm’s superiority was validated through comparisons across various datasets, and its effectiveness was demonstrated through target recognition tests in real-world dense fog conditions.

Author Contributions

Conceptualization, L.J. and C.C.; methodology, L.J.; software, L.J.; validation, L.J. and C.C.; formal analysis, L.J.; investigation, L.J.; resources, L.J.; data curation, L.J.; writing—original draft preparation, L.J.; writing—review and editing, L.J.; visualization, L.J.; supervision, L.J.; project administration, L.J.; funding acquisition, L.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data are contained within the article.

Acknowledgments

We sincerely thank Chao Chen for his insightful guidance and support. Additionally, we acknowledge the use of Python (version 3.10) for data analysis and visualization in this study. The authors have carefully reviewed and edited the output and assume full responsibility for the content of this publication.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ICDFInverse Cumulative Distribution Function
CNNConvolutional Neural Network
MB-TaylorFormerMulti-Branch Taylor Transformer
DCPDark Channel Prior
AOD-NetAll-in-One Dehazing Network
MSBDNMulti-Scale Boosted Dehazing Network
FFA-NetFeature Fusion Attention Network
RestormerRestoration Transformer
DehamerDeep Hybrid Atmospheric-scattering Model Guided Network
EnhancedMB-CauchyFormerEnhanced Multi-Branch Cauchy Transformer Network
PSNRPeak Signal-to-Noise Ratio
SSIMStructural Similarity Index Measure

References

  1. Narasimhan, S.G.; Nayar, S.K. Contrast restoration of weather degraded images. IEEE Trans. Pattern Anal. Mach. Intell. 2003, 25, 713–724. [Google Scholar] [CrossRef]
  2. He, K.M.; Sun, J.; Tang, X.O. Single Image Haze Removal Using Dark Channel Prior. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 2341–2353. [Google Scholar]
  3. Liu, Z.H.; Zhao, S.J.; Wang, X. Research on Driving Obstacle Detection Technology in Foggy Weather Based on GCANet and Feature Fusion Training. Sensors 2023, 23, 2822. [Google Scholar] [CrossRef] [PubMed]
  4. Zai, W.J.; Yan, L.S. Multi-Patch Hierarchical Transmission Channel Image Dehazing Network Based on Dual Attention Level Feature Fusion. Sensors 2023, 23, 7026. [Google Scholar] [CrossRef]
  5. Zamanidoost, Y.; Ould-Bachir, T.; Marterl, S. OMS-CNN: Optimized Multi-Scale CNN for Lung Nodule Detection Based on Faster R-CNN. IEEE J. Biomed. Health Inform. 2024, 20, 2148–2160. [Google Scholar] [CrossRef] [PubMed]
  6. Liu, Z.H.; Yan, J.; Zhang, J.Z. Research on a Recognition Algorithm for Traffic Signs in Foggy Environments Based on Image Defogging and Transformer. Sensors 2024, 24, 4370. [Google Scholar] [CrossRef]
  7. Zhang, H.; Sindagi, V.; Patel, V.M. Multi-scale Single Image Dehazing Using Perceptual Pyramid Deep Network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–22 June 2018; pp. 1015–1023. [Google Scholar]
  8. Anvari, Z.; Athitsos, V. Dehaze-GLCGAN: Unpaired Single Image De-hazing via Adversarial Training. arXiv 2020, arXiv:2008.06632. [Google Scholar]
  9. Liang, L.H.; Zhang, S.Q.; Li, J. Multiscale DenseNet Meets with Bi-RNN for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 5401–5405. [Google Scholar] [CrossRef]
  10. Zhang, J.; Tu, B.; Liu, B.; Li, J.; Plaza, A. Hyperspectral Image Classification via Neighborhood Adaptive Graph Isomorphism Network. IEEE Trans. Geosci. Remote Sens. 2025, 63, 5515717. [Google Scholar] [CrossRef]
  11. Xia, Z.F.; Pan, X.R.; Song, S.J.; Li, L.E.; Huang, G. Vision Transformer with Deformable Attention. arXiv 2022, arXiv:2201.00520. [Google Scholar] [CrossRef]
  12. Liu, Z.; Hu, H.; Lin, Y.T.; Yao, Z.L.; Xie, Z.D.; Wei, Y.X. Swin Transformer V2: Scaling Up Capacity and Resolution. arXiv 2022, arXiv:2111.09883. [Google Scholar] [CrossRef]
  13. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv 2022, arXiv:2010.11929. [Google Scholar]
  14. Liu, W.N.; Zhang, J.H. Research on Image Defogging Algorithm Combining Homomorphic Filtering and Retinex. In Proceedings of the International Symposium on Computer Technology and Information Science, Xi’an, China, 12–14 July 2024; pp. 596–599. [Google Scholar]
  15. Yang, P.W.; Wang, L. Multi-Scale Dehaze Network Based on Frequency Domain Assistance and Detailed Brightness Information Guidance. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Hyderabad, India, 6–11 April 2025; pp. 1–5. [Google Scholar]
  16. Qiu, Y.W.; Zhang, K.H.; Wang, C.X.; Luo, W.H.; Li, H.D.; Jin, Z. MB-TaylorFormer: Multi-branch Efficient Transformer Expanded by Taylor Formula for Image Dehazing. In Proceedings of the IEEE International Conference on Computer Vision, Paris, France, 2–3 October 2023; pp. 12756–12767. [Google Scholar]
  17. Jin, Z.; Qiu, Y.W.; Zhang, K.H.; Li, H.D.; Luo, W.H. MB-TaylorFormer V2: Improved Multi-Branch Linear Transformer Expanded by Taylor Formula for Image Restoration. arXiv 2025, arXiv:2501.04486. [Google Scholar] [CrossRef] [PubMed]
  18. Wang, M.; Wang, J.S.; Li, X.D.; Zhang, M.; Hao, W.K. Harris Hawk Optimization Algorithm Based on Cauchy Distribution Inverse Cumulative Function and Tangent Flight Operator. Appl. Intell. 2021, 52, 10999–11026. [Google Scholar] [CrossRef]
  19. Wang, S.; Gao, Y.B.; Li, S.; Lv, C.; Cai, X.; Li, C.K.; Yuan, H.; Zhang, J.L. MetricGrids: Arbitrary Nonlinear Approximation with Elementary Metric Grids based Implicit Neural Representation. arXiv 2025, arXiv:2503.10000. [Google Scholar]
  20. Zhang, Y.F.; Wang, C.Y.; Wang, X.G.; Zeng, W.J.; Liu, W.Y. FairMOT: On the Fairness of Detection and Re-Identification in Multiple Object Tracking. arXiv 2021, arXiv:2004.01888. [Google Scholar] [CrossRef]
  21. Dubey, S.R.; Singh, S.K.; Chaudhuri, B.B. Activation functions in deep learning: A comprehensive survey and benchmark. Nerocomputing 2022, 6, 92–108. [Google Scholar] [CrossRef]
  22. Wang, T.; Dai, W.F.; Wu, Y.J. Nonuniform and pathway-specific laminar processing of spatial frequencies in the primary visual cortex of primates. Nat. Commun. 2024, 15, 4005–4021. [Google Scholar] [CrossRef]
  23. Luo, W.J.; Li, Y.J.; Urtasun, R.; Zemel, R. Understanding the Effective Receptive Field in Deep Convolutional Neural Networks. arXiv 2017, arXiv:1701.04128. [Google Scholar] [CrossRef]
  24. Li, B.Y.; Peng, X.L.; Wang, Z.Y.; Xu, J.Z.; Feng, D. AOD-Net: All-in-One Dehazing Network. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4780–4788. [Google Scholar]
  25. Zheng, T.Y.; Xu, T.Y.; Li, X.D.; Zhao, X.W. Improved AOD-Net Dehazing Algorithm for Target Image. In Proceedings of the International Conference on Computer Engineering and Intelligent Control, Guangzhou, China, 11–13 October 2024; pp. 333–337. [Google Scholar]
  26. Wang, Y.; Yu, Z.; Guo, C.; Zhao, J. Layer Separation Network with Contrastive Loss for Robust Dehazing. In Proceedings of the International Conference on Cloud Computing and Big Data Analytics, Chengdu, China, 26–28 April 2023; pp. 290–296. [Google Scholar]
  27. Zheng, Y.J.; Liu, S.C.; Bruzzone, L. An Attention-Enhanced Feature Fusion Network (AeF2N) for Hyperspectral Image Classification. IEEE Geosci. Remote Sens. Lett. 2023, 20, 5511005. [Google Scholar] [CrossRef]
  28. Cui, Y.N.; Ren, W.Q.; Cao, X.C.; Knoll, A. Revitalizing Convolutional Network for Image Restoration. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 46, 9423–9438. [Google Scholar] [CrossRef] [PubMed]
  29. Faramarzi, F.; Heidarinejad, M.; Mirjalili, S.; Gandomi, A.H. Marine predators algorithm: A nature-inspired metaheuristic. Expert Syst. Appl. 2020, 152, 1929–1957. [Google Scholar] [CrossRef]
  30. Li, B.Y.; Ren, W.Q.; Fu, D.P.; Tao, D.C.; Feng, D.; Zeng, W.J. Benchmarking Single-Image Dehazing and Beyond. IEEE Trans. Image Process. 2019, 28, 492–505. [Google Scholar] [CrossRef] [PubMed]
  31. Ancuti, C.O.; Ancuti, C.; Timofte, R.; Vleeschouwer, C.D. O-haze: A dehazing benchmark with real hazy and haze-free outdoor images. In Proceedings of the CVF Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–22 June 2018; pp. 754–762. [Google Scholar]
  32. Ancuti, C.O.; Ancuti, C.; Timofte, R.; Vleeschouwer, C.D. Dense-haze: A benchmark for image dehazing with dense-haze and haze-free images. In Proceedings of the IEEE International Conference on Image Processing, Taipei, Taiwan, 22–25 September 2019; pp. 1014–1018. [Google Scholar]
Figure 1. Double-peak inverse cumulative distribution functions of Cauchy distribution.
Figure 1. Double-peak inverse cumulative distribution functions of Cauchy distribution.
Sensors 25 05088 g001
Figure 2. Visual cortex target recognition mechanism diagram.
Figure 2. Visual cortex target recognition mechanism diagram.
Sensors 25 05088 g002
Figure 3. Diagram of the innovative parallel multi-scale processing principle.
Figure 3. Diagram of the innovative parallel multi-scale processing principle.
Sensors 25 05088 g003
Figure 4. Algorithm execution flowchart.
Figure 4. Algorithm execution flowchart.
Sensors 25 05088 g004
Figure 5. SOTS-Indoor image comparison experiment. The red large frame is enlarged by 1 times compared to the red small frame.
Figure 5. SOTS-Indoor image comparison experiment. The red large frame is enlarged by 1 times compared to the red small frame.
Sensors 25 05088 g005
Figure 6. SOTS-Outdoor image comparison experiment. The red large frame is enlarged by 4 times compared to the red small frame.
Figure 6. SOTS-Outdoor image comparison experiment. The red large frame is enlarged by 4 times compared to the red small frame.
Sensors 25 05088 g006
Figure 7. O-HAZE image comparison experiment. The red large frame is enlarged by 4 times compared to the red small frame.
Figure 7. O-HAZE image comparison experiment. The red large frame is enlarged by 4 times compared to the red small frame.
Sensors 25 05088 g007
Figure 8. Dense-haze image comparison experiment. The red large frame is enlarged by 4 times compared to the red small frame.
Figure 8. Dense-haze image comparison experiment. The red large frame is enlarged by 4 times compared to the red small frame.
Sensors 25 05088 g008
Figure 9. Different target heat maps.
Figure 9. Different target heat maps.
Sensors 25 05088 g009
Figure 10. Real target recognition and probability map.
Figure 10. Real target recognition and probability map.
Sensors 25 05088 g010
Table 1. Dataset configuration table.
Table 1. Dataset configuration table.
DatasetTraining Sample CountTest Sample CountFog Concentration DistributionEvaluation Subset
RESIDE-ITS [30]13,990500Low → high (indoor)SOTS
RESIDE-OTS [30]313,950500Uniform (outdoor)SOTS
O-HAZE [31]405Medium (real scene)Last 5
Dense-HAZE [32]505Extremely high (heavy fog)Last 5
Table 2. Quantitative comparison of different methods in fog removal benchmark testing.
Table 2. Quantitative comparison of different methods in fog removal benchmark testing.
MethodSOTS-IndoorSOTS-OutdoorO-HAZEDense-HazeComputational Efficiency
-PSNR ↑ SSIM ↑PSNR ↑ SSIM ↑PSNR ↑ SSIM ↑PSNR ↑ SSIM ↑Params (M) MACs (G)
Based on physical models-----
DCP [1]16.62 0.81819.13 0.81516.78 0.65312.72 0.442- 0.78
Improved AOD-Net [24,25]24.61 0.85422.356 0.93420.577 0.72313.59 0.5440.025 2.3
MSBDN-DFF [26]34.29 0.98333.65 0.97024.87 0.75816.34 0.63831.45 43.82
Self-attention model-----
FFA-Net [27]36.39 0.98933.57 0.98422.12 0.77015.70 0.5494.46 287.8
ConvIR [28]41.53 0.99437.95 0.99025.36 0.78416.86 0.6005.53 42.1
SOTA-----
FIGD-Net [15]41.11 0.99638.19 0.999226.12 0.80517.71 0.608124.24 70.76
MB-TaylorFormerV2 [16,17]42.86 0.99539.25 0.99225.43 0.79216.95 0.6217.29 86.0
Ours (SOTA breakthrough)-----
EnhancedMB-CauchyFormer45.12 0.99739.55 0.99424.98 0.80017.83 0.58415.80 96.8
Here, “-” indicates that the result is not available. The best and second best results are highlighted in bold and underlined, respectively. The core meaning of the upward arrow (↑) Indicates that “the higher the value, the better the performance”: The arrow ↑ clearly indicates that the larger the value of the indicator (such as PSNR, SSIM), the better the algorithm effect.
Table 3. Comparison of different convolution layer strategies.
Table 3. Comparison of different convolution layer strategies.
MethodPSNR (dB)SSIMParameters (M)MACs (G)
Standard convolution40.710.9911.83.2
Gaussian deformable convolution41.620.9922.55.1
Cauchy deformable convolution41.870.9932.77.8
Table 4. Comparison of the approximation error and performance of different attention mechanisms.
Table 4. Comparison of the approximation error and performance of different attention mechanisms.
Attention TypePSNR (dB)SSIMApproximationLong-Range Dependency Strength
Standard attention40.710.99101.0
Linear attention36.120.9730.1480.32
Gaussian kernel approximation40.050.9890.0870.75
Cauchy–Gaussian Kernel Approximation42.590.9950.0210.92
Table 5. Progressive performance of branch combinations.
Table 5. Progressive performance of branch combinations.
Activate BranchPSNR (dB)SSIMMain Domain
L143.950.994Edge texture
L1 + L2 + L344.980.993Contour shape integration
L1 + L2 + L3 + L445.020.995Global color/texture shape modeling
L1 + L2 + L3 + L4 + L545.080.996Target recognition
Full branch + dynamic gate control45.120.996Cross-scale feature fusion
Table 6. Module synergy analysis.
Table 6. Module synergy analysis.
Module CombinationPSNR (dB)SSIMΔPSNR vs. BaselineContribution DecompositionMACs (G)
Baseline (standard convolution+ attention)40.710.9913.2
Cauchy–Fourier transform41.870.993+1.1626.3%7.8
Mixed distribution attention42.590.995+1.8816.3%58.8
Tree-like multi-branching (complete model)45.120.996+4.4157.4%96.8
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ji, L.; Chen, C. Multi-Scale Image Defogging Network Based on Cauchy Inverse Cumulative Function Hybrid Distribution Deformation Convolution. Sensors 2025, 25, 5088. https://doi.org/10.3390/s25165088

AMA Style

Ji L, Chen C. Multi-Scale Image Defogging Network Based on Cauchy Inverse Cumulative Function Hybrid Distribution Deformation Convolution. Sensors. 2025; 25(16):5088. https://doi.org/10.3390/s25165088

Chicago/Turabian Style

Ji, Lu, and Chao Chen. 2025. "Multi-Scale Image Defogging Network Based on Cauchy Inverse Cumulative Function Hybrid Distribution Deformation Convolution" Sensors 25, no. 16: 5088. https://doi.org/10.3390/s25165088

APA Style

Ji, L., & Chen, C. (2025). Multi-Scale Image Defogging Network Based on Cauchy Inverse Cumulative Function Hybrid Distribution Deformation Convolution. Sensors, 25(16), 5088. https://doi.org/10.3390/s25165088

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop