1. Introduction
Traditionally, haze is an atmospheric condition in which visibility and clarity are degraded by suspended dry particulates such as smoke, dust, and other pollutants. This phenomenon arises from various sources, including air pollution [
1], forest fires [
2], and other environmental disturbances. Haze not only diminishes visibility [
3], but also creates significant challenges in critical situations, such as slowing down vehicles or obstructing the ability to identify individuals during emergencies like fires. Numerous models have been proposed in the past to address this problem by manipulating parameters of the hazy image equation—particularly focusing on atmospheric light constants or transmission maps [
4]. However, despite these efforts, existing methods often lack robustness, efficiency, and accuracy when applied under diverse and extreme real-world conditions.
Real-time haze and smoke removal has therefore become an urgent research priority due to its relevance across multiple safety-critical domains. For example, in the automotive industry, effective dehazing enhances driver visibility and strengthens the performance of Advanced Driver Assistance Systems (ADAS). In firefighting operations, real-time dehazing technologies can enable clearer visibility for victim identification and hazard analysis under smoky conditions. Similarly, in surveillance and security, maintaining clarity in polluted or foggy environments is essential for reliable monitoring. The inefficiencies of previous methods and the widespread demand for strong real-time dehazing solutions in such applications were the primary reasons for designing and developing our proposed model.
To address these challenges, this paper proposes a novel image dehazing model that integrates multiple aspects of the hazy image equation with innovative design strategies. Unlike traditional models that optimize only one parameter, our method simultaneously incorporates the atmospheric light constant (A) and the transmission map (), with particular emphasis on the scattering coefficient () in the transmission formulation. A distinctive feature of this model is its approach to depth map calculation, which leverages a reference image of the target environment. This reference-based methodology allows for accurate recovery of depth information and enhances the fidelity of haze removal. Furthermore, because haze is inherently non-uniform across images, the model first identifies the haziest region using grayscale conversion and a quad-tree-based method. The atmospheric constant is then estimated based on this localized haze concentration, and combined with regression-based estimation of the scattering coefficient and the depth map for precise reconstruction of the dehazed image.
The effectiveness of the proposed model is validated through extensive comparisons against state-of-the-art methods, including CLAHE, AOD-Net, Dark Channel Prior, FFA, and Single U-Net. Across multiple quantitative metrics such as SSIM, PSNR, VIF, FADE, UIQI, and C-H Ratio, our approach consistently demonstrates superior performance. Notably, it achieves the highest SSIM (0.99), PSNR (22.25), and VIF (1.08), indicating remarkable structural preservation, image quality retention, and visual fidelity; while FADE values are slightly higher (63.87), they reflect the model’s effective balance between enhancement and detail preservation, ensuring practicality in real-world scenarios. Comparative analyses with visual examples further illustrate the superiority of our method over prior solutions such as Single U-Net, FFA-Net, AOD-Net, and Dark Channel. Summarizing the above contributions, the novelty of the work is as follows.
It combines the transmission map () with the atmospheric light constant (A) in the dehazing equation, while dynamically estimating the scattering coefficient () using a linear regression model, unlike standard single-parameter methods.
It employs a quad-tree-based method to identify the haziest regions of an image, enabling precise measurement of atmospheric light (A) and improving dehazing performance in unevenly hazy conditions.
It utilizes grayscale conversion, depth estimation, and linear regression within a unified dehazing system. It incorporates a reference image to generate an accurate depth map (), thereby improving transmission map reliability with prior scene information.
It achieves state-of-the-art results, including SSIM = 0.99, PSNR = 22.25, and VIF = 1.08, surpassing classic methods such as Single-U-Net, Double-U-Net, and Dark Channel Prior while preserving structural details and overall image fidelity.
It demonstrates wide-ranging applications, such as enhanced road safety, improved surveillance and emergency response capabilities, higher-quality aerial/satellite imagery for environmental monitoring, and safer autonomous navigation in foggy environments.
2. Related Works
The restoration of visibility in hazy images is one of the primary challenges in image processing, and many dehazing techniques rely on airlight estimates. One technique [
4] improves transmission map estimates by employing a global dynamic template and a positive depth of field to overcome DCP’s shortcomings in situations with objects that resemble atmospheric light. For improved transmission map and airlight estimation, a hybrid approach [
5] that combines histogram equalization with DCP performs better on the D-Hazy dataset. A rapid airlight estimate approach [
6] decreases computational burden by 92.82%, and Down-sampling DCP (DS-DCP) [
7] reduces computational complexity by 98% while retaining an error rate of 0.22%. The paper [
8] suggests a Modified DCP (MDCP) approach to overcome common DCP problems, such as artifacts, halos, color distortions, and high computational cost. It utilizes a pixel-based dark channel as the reference image in filtering and incorporates adaptive scaling for airlight estimates, resulting in improved assessments and a 5.12 times increase in processing speed. These developments show notable gains in DCP’s precision and effectiveness, fixing long-standing problems and boosting the algorithm’s performance in various applications.
A U-Net-based segmentation network in [
9] enhances the quality of medium transmission map estimation from a hazy image by employing a modified dark channel prior approach to calculate global atmospheric light. This method works well and consistently across a variety of datasets. Another technique [
10] achieves a thirty-fold speedup in processing huge images by estimating the transmission function using a linear model with a quadtree search. The paper [
11] presents a robust single-image dehazing method that optimizes the transmission map using image features such as contrast energy, entropy, and sharpness to estimate the extinction coefficient more accurately. They also introduced an adaptive atmospheric light model to handle non-uniform illumination, unlike conventional homogeneous light assumptions. A heuristic model for an ideal transmission map is proposed by [
12], guaranteeing depth consistency for precise dehazing. Outperforming state-of-the-art methods, ref. [
13] enhances transmission map estimate by merging foreground and sky regions, maintaining features, and reducing artifacts. The paper [
14] presents SIDGAN, a U-Net-based GAN for single-picture dehazing that improves performance by utilizing color consistency and perceptual loss. Using convolution modules, channel attention, and gate mechanism residual blocks, ref. [
15] introduces zUNet, a lightweight network for real-time systems that outperforms current latency, parameter count, and PSNR techniques. This research shows how important the transmission map is for dehazing and provides practical ways to enhance image quality.
A unique atmospheric scattering model (NASM) with a scattering compensation coefficient addresses color cast in dehazing by introducing the IDACC method [
16], which produces better results. Another study suggests a single-picture dehazing technique that uses average saturation prior and sky recognition to independently estimate transmission in sky and non-sky regions for high-quality haze removal [
17]. In terms of efficiency and performance, ref. [
18] surpasses state-of-the-art approaches by implementing color attenuation prior to employing a linear model for depth restoration. The paper [
19] uses adaptive gamma correction and k-means clustering to improve visual quality in areas with a lot of sky. By addressing color-related problems collectively, these techniques provide creative answers for enhanced dehazing and effectiveness.
Integration of the atmospheric scattering model (ASM) and segmentation to ensure precise recovery in [
20] is introduced. The paper [
21] improves dehazing in areas with abrupt depth changes by estimating ambient light using quad-tree subdivision and linear transformation in the minimum channel. For better performance with less complexity, ref. [
22] models a linear relationship between the depth map and the minimum channel of the hazy image. In paper [
23] PSNR and SSIM improves the performance, especially in bright areas, by refining Dark Channel Prior (DCP) with multiple linear regression, which lowers estimation errors. These strategies show how effective regression techniques improve dehazing efficiency and accuracy in foggy situations.
The technique presented in [
24] utilizes Repeated Averaging Filters to address halo artifacts and improve radiance recovery when estimating ambient light from a single foggy image. The paper [
25] offers a quick single-image dehazing technique that balances speed and quality for real-time systems by utilizing gray projection and the atmospheric scattering model. Demonstrates efficacy in severe weather without causing an increase in noise by introducing a combined dehazing and denoising technique that lessens noise amplification in foggy circumstances [
26]. The paper [
27] demonstrates exceptional performance on dehazing benchmarks by proposing a deep learning-based dehazing method with a pre-dehazer for directing haze removal. By passing transmission map estimates and learning a nonlinear mapping from hazy to dehazed images directly, the deep residue learning (DRL) method in [
28] outperforms existing approaches in terms of both objective and subjective quality. This paper [
29] introduces a depth estimation method that evaluates dehazing techniques by integrating geometry and edge information, utilizing a synthetic outdoor dataset. This paper [
30] provides the hyperspectral multi-level hazy picture dataset SHIA to test dehazing techniques and determine how well they handle multi-level hazy images.
In order to improve restoration quality, ref. [
31] adds a dynamic scattering coefficient for vision across hazy levels. The paper [
32] rapidly and accurately estimates atmospheric light, particularly in bright regions, using pixel-based dark and bright channel limitations. In the paper [
33], the Non-Homogeneous RESIDE dataset is used to improve processing speed and dehazing precision. These techniques prioritize performance in the actual world and visibility enhancement. AEDNet, an attention-based model focusing on fine detail retention, is presented in [
34]. It has a channel shuffling mechanism to maintain picture features during dehazing. Utilizing adversarial learning and removing the requirement for paired training datasets, ref. [
35] uses CycleGAN for unsupervised dehazing. These methods differ in their strategies, from physical models to attention processes and adversarial learning. Still, they aim to increase image quality, estimate atmospheric light, and process data in real-time.
Several techniques demonstrate the importance of depth maps in enhancing dehazing techniques. A physics-based network used a simulated depth map from virtual environments [
36] to provide consistent training and improved benchmark performance without real-world data. We observe that in the paper [
37], object detection is enhanced by combining CNN-based transmission map estimation with adaptive color correction and rephrasing it as a depth estimation problem, and [
38] enhanced photos for underwater scenarios by employing haze as a depth cue, surpassing traditional techniques. In [
39], a GAN-based architecture combines depth-guided refinement with physical restoration to overcome unpaired learning restrictions and restore distant details. This paper [
40] enhanced depth maps using a second-order variational framework to accomplish robust haze removal across conditions and maintain structures.
Advances in picture dehazing in recent years have investigated various methods to improve visual restoration in multiple settings. Perceptual loss functions such as PSNR-HVS and HaarPSI, for example, perform better in CNN-based dehazing models, as [
41] demonstrates. In [
42], a minimal DMCGF network is suggested to employ gate fusion modules and multi-scale feature extraction for real-time dehazing. Color correction and dark channel enhancement are used in [
43] to correct color variation in dusty photos. This paper [
44] presents a fog density fusion technique for less halo effects and seamless transitions. ZRD-Net, introduced in [
45], pioneeringly addresses dehazing in an unsupervised, zero-shot way, removing the need for large datasets. This paper [
46] illustrates a hybrid U-Net and AOD-Net method with adaptive loss for improved accuracy. This paper [
47] achieves effective image restoration by integrating cutting-edge methods for haze reduction in SAR imaging, such as Multi-Scale Dehazing and guided filters. For robust dehazing, this research highlights advancements in loss functions, structures, and adaptive techniques.
In addition to the classical and early deep-learning-based dehazing algorithms discussed above, several recent studies have extended these architectures with improved feature fusion, transformer integration, and cycle-consistent training mechanisms. For instance, Dudhane and Murala proposed RYF-Net [
48], a deep fusion network that estimates transmission maps by combining RGB and YCbCr representations through separate sub-networks and a fusion stage. Jain [
49] proposed an enhanced Feature Fusion Attention Network combined with CycleGAN for multi-weather restoration tasks, demonstrating improved adaptability across haze, snow, and rain. Li et al. [
50] introduced UTCR-Dehaze, a transformer-augmented U-Net and cycle-consistent generative model for unpaired remote sensing image dehazing, which effectively captures both global and local contextual information. Furthermore, Majid and Is a [
51] presented an updated CLAHE variant that adapts contrast enhancement at a local level for low-visibility conditions, reaffirming the continued relevance of traditional histogram-based approaches in modern enhancement pipelines. These contemporary developments, all from 2025, represent the evolution of earlier classical, CNN-based, and attention-driven frameworks, and therefore were adopted as baselines for comparison in
Table 1.
In conclusion, the reviewed literature demonstrates significant progress in image dehazing, focusing on improving image quality, processing efficiency, and handling complex scenarios. Among the key advancements are deep learning models, enhanced Dark Channel Prior methods, and the utilization of depth information for more precise transmission map estimation. Lightweight networks and hybrid approaches offer real-time applications without compromising quality. These efforts pave the way for future developments in real-world scenarios by demonstrating the importance of combining deep learning methods with physical models for more effective and successful dehazing.
3. Methodology
Before detailing the methodology, we provide an overview of the proposed algorithm for training the dehazing model. The complete workflow is clearly represented in
Figure 1, which outlines the major steps from pre-processing to evaluation. The algorithm begins with pre-processing of clear and hazy images, including resizing and normalization, followed by conversion to HSV color space to extract brightness (
) and saturation (
) features. A depth map is then computed from the difference between clear and hazy images, and the atmospheric scattering coefficient (
) is estimated. A regression model is trained on [
,
] features to predict
, which is subsequently used to compute the transmission map
. Finally, the hazy image is restored using the estimated transmission map, followed by brightness and contrast adjustments, and the quality of the dehazed output is evaluated through metrics such as SSIM and PSNR.
The suggested regression-based dehazing approach uses a mathematical regression-based formulation for parameter estimation rather than iterative deep learning optimization. A linear regression model trained on the extracted brightness–saturation features (
) and their related scattering coefficients (
) was used to obtain the coefficients
, which describe the scattering relationship. The least squares minimization method outlined in Equation (
8) was used to solve this regression analytically, guaranteeing convergence in a single computer step.
A grid search mechanism was used during the dehazing process to identify the haziest regions within an image by creating quadrant-based subdivisions (quadruples), not for hyperparameter tuning. This allowed for better restoration in non-uniform haze distributions and precise assessment of localized atmospheric light.
Every experiment was carried out in the Kaggle cloud GPU environment, which only needed simple mathematical calculations rather than complex deep learning structures. The derivation of the coefficients and associated regression parameters was the only machine learning component. This resulted in a very low average processing time of about 0.038 s per image for the model, which is significantly faster than traditional deep-learning-based dehazing techniques. It should be noted, nonetheless, that the processing time during real-time deployment may differ based on the hardware configuration and computational resources, including embedded GPUs or processors.
3.1. Dataset Description
Image dehazing tasks are frequently evaluated using the RESIDE dataset [
52,
53], and it includes both the RESIDE-Indoor (ITS) and RESIDE-Outdoor (OTS) subsets. Furthermore, the RESIDE-6K collection includes synthetic photos from indoor and outdoor environments, providing a wide range of foggy circumstances. The test data in RESIDE, known as the Synthetic Objective Testing Set (SOTS), is divided into indoor and outdoor subsets for performance analysis. The inclusion of actual foggy photographs improves the dataset’s relevance to real-world situations, while the synthetic images are created using a physical scattering model as specified in the standardized dataset, guaranteeing consistency in haze simulation.
The dataset is divided into two principal folders: one for training and validation, and another for testing. The subfolders of each folder contain a combination of indoor and outdoor locations, as well as ground truth and fuzzy photographs. Training, validation, and testing subsets of the dataset utilized in this study were separated using the conventional 70:20:10 ratio. The regression model was trained using 70% of the data, with 20% being used for validation and parameter adjustment purposes, while the remaining 10% was used for independent performance evaluation. This ratio was chosen to guarantee a fair trade-off between avoiding overfitting to particular haze patterns and learning generalizable regression coefficients (). The model’s performance is not affected by particular data divisions, as demonstrated by the empirical validation of the selected split, which produced consistent findings across several random seeds. The observed changes in SSIM and PSNR were negligible (within ()) when the dataset division was changed (e.g., 80:10:10 or 60:20:20), indicating the stability and robustness of the regression-based approach. Therefore, the 70:20:10 split offers the best possible compromise between evaluation reliability and model generalization.
3.2. Data Pre-Processing
The data preprocessing for the image dehazing task involves implementing a custom dataset class, Dehazing-Dataset, which efficiently handles paired hazy and clean images. The dataset is loaded from specified directories, filtering only valid image formats (e.g., .jpg, .jpeg, .png) and ensuring alignment by truncating excess images to maintain an equal number of hazy and clean image pairs. A filtering mechanism validates that only image pairs with identical dimensions are retained, ensuring consistency in input-output mappings during model training. The rgb_loader function is integral to the pipeline, loading images and converting them to RGB format for compatibility with standard machine learning frameworks. Additionally, the show_grayscale_image method offers a utility for visualizing grayscale images, aiding in inspecting and validating preprocessing results. The images are further processed with optional transformations, such as resizing or augmentation, to standardize the dataset and enhance its variability. This comprehensive preprocessing pipeline ensures high-quality, well-aligned, and properly formatted data, facilitating the adequate training and evaluation of image dehazing models.
Feature extraction is performed by converting the image from RGB to HSV color space and computing the mean brightness and saturation, from which the parameters and are derived. In the initial stage, the image is converted from RGB to HSV (hue, saturation, value) color format, as HSV is more suitable for color quality analysis since it distinguishes between luminance (value) and color saturation (saturation). The brightness (V channel) and saturation (S channel) are extracted, as haze affects these components differently: brightness determines how light or dark the image is, while saturation reflects how vivid the colors appear. The mean brightness and saturation are then calculated for the entire image, and two parameters are derived: , which indicates the proportion of brightness to the total brightness and saturation and reflects how much light is influencing the scene, and , which indicates the proportion of saturation to the total brightness and saturation and characterizes the available color information relative to brightness.
3.3. Mathematical Formulations for Image Dehazing
To systematically recover clear images from degraded observations, the dehazing process is formulated through a sequence of mathematical steps. Each step builds upon the previous one, starting from the fundamental hazy image equation and progressing toward brightness–saturation analysis, regression-based parameter estimation, and final image restoration. The entire process is summarized in Algorithm 1.
- Step 1.
Hazy Image Equation
where
represents the perceived hazy image,
denotes the scene radiance,
is the transmission map indicating the amount of light reaching the camera without scattering,
A is the global atmospheric illumination, and
x corresponds to the pixel location in the image.
- Step 2.
Calculate Mean Brightness and Saturation
Let
V and
S represent the visibility and saturation channels in the HSV color space, respectively. The mean brightness and mean saturation are calculated as
Here, and correspond to the pixel intensities in the value and saturation channels of the HSV image, respectively, while N represents the total number of pixels in the image.
- Step 3.
Brightness Sum
The total brightness is the aggregate of the mean brightness and mean saturation.
- Step 4.
Ratios and
The ratios
and
, which represent the relative contributions of brightness and saturation to the overall image characteristics, are defined as
- Step 5.
Determining Optimal Values for a, b, and c Through Regression Analysis
The optimal values of
a,
b, and
c are determined by solving a linear regression problem. Define the feature matrix
X and the target vector
y as follows:
The parameters
can thereafter be approximated by minimizing the aggregate of squared residuals:
where
represents the transpose of
X, and the expression
denotes the inverse of the matrix
.
This yields the values of a, b, and c that best fit the data.
- Step 6.
Refining the Transmission Map
The transmission map
for pixel
x is computed using the estimated
as
where
denotes the depth map, calculated as the absolute difference between the hazy and clear images at each pixel.
- Step 7.
Identifying the Haziest Region
The haziest region in the image is determined using a haze score, which considers the uniformity of pixel intensities and the lack of edges. For a region
R, the haze score is calculated as
where
denotes the standard deviation of pixel intensities within region
R and edge count represents the quantity of edges within the region
R, detected using the canny edge detector.
Regions are recursively divided into quadrants using a quadtree algorithm until a maximum depth is reached or the region size is minimal. The area with the highest haze score is designated as the most hazy.
- Step 8.
Estimating the Airlight Value A
To identify the airtight value A, the most luminous pixels in the haziest area are examined. The procedure is outlined as follows:
Normalize the pixel intensities of the region to
:
Identify the top 20% brightest pixels within the region:
Determine the average intensity of the brightest pixels in order to estimate
A:
The estimated A represents the intensity of the scattered light dominating the haziest region.
- Step 9.
Restoration of the Dehazed Image
The transmission map
and the airlight value
A are employed to restore the dehazed image
:
where
is the observed hazy image and
is a small constant that is included to prevent division by zero.
- Step 10.
Brightness and Contrast Adjustment
The restored image is multiplied by 255 to transform it from a floating-point to an 8-bit format. Brightness and contrast adjustments are performed using the parameters and :
Brightness adjustment: the parameter shifts pixel values by adding or subtracting a constant.
Contrast adjustment: the parameter scales pixel values, enhancing contrast by making dark regions darker and bright regions brighter.
- Step 11.
Gamma Correction
Gamma correction is used to fine-tune brightness and compensate for the nonlinear way the human eye perceives light. Gamma correction involves boosting pixel values to the power of
. The value of
determines the image’s overall contrast: greater values make it darker, while lower values make it brighter. After removing the haze, this procedure guarantees that the image seems natural to the human eye.
| Algorithm 1 Model training of proposed approach. |
Require: Training data, uniform dimensions, normalized pixel values Ensure: Dehazed images
- 1:
Load clear and hazy images, resize, normalize - 2:
Convert images to HSV color space - 3:
Compute brightness and saturation adjustment factor ( and ): - 4:
- 5:
- 6:
Train a linear regression model using as input features and corresponding values as labels. The model outputs where - 7:
Compute transmission map using :
where . - 8:
- 9:
Adjust brightness and contrast of restored images. - 10:
Evaluate dehazing quality using performance metrics like SSIM, PSNR, etc. - 11:
Save and compare dehazed images with clear and hazy images.
|
4. Results
This study suggests a novel dehazing model that combines regression-based depth map estimates with atmospheric scattering coefficients to overcome these constraints. By concentrating on determining the most hazardous areas and utilizing sophisticated regression methods, the suggested model seeks to offer a more precise and effective resolution to the issue of haze elimination. The suggested model’s resilience was examined across a range of haze densities. It maintained high SSIM scores in every case, continuously outperforming alternative approaches. This demonstrates its versatility and efficiency in harsh environments like thick fog or haze. The normalized pixel intensity distributions for clean (red), dehazed (green), and hazy (blue) are shown in
Figure 2. There is a noticeable peak in the lower intensity range of the fuzzy image, which suggests less contrast and visibility. With a discernible movement toward higher intensity values and a wider distribution, the dehazed image shows a notable improvement in contrast and detail recovery. As a reference, the clear image exhibits clear peaks at both low and high intensities, signifying ideal clarity. The effectiveness of the dehazing algorithm in restoring visibility is demonstrated by the close alignment of the dehazed image with the clear image.
As visible there is an apparent shift to the right in the dehazed image’s pixel intensity distribution, especially for brightness levels greater than 200. This change is a direct result of the suggested regression-based dehazing technique, which uses the estimated transmission map and subsequent brightness-contrast adjustments to compensate for haze attenuation and compute the restored picture . The recovery of genuine scene radiance that was previously reduced by scattering is reflected in the elevation in higher intensity locations. It is worth noting that the increase in brightness results from radiometrically consistent restoration rather than overexposure or noise amplification, as evidenced by the high SSIM value and preserved color fidelity. With a discernible movement toward higher intensity values and a wider distribution, the dehazed image shows a notable improvement in contrast and detail recovery. As a reference, the clear image exhibits clear peaks at both low and high intensities, signifying ideal clarity.
We compare the performance of the suggested dehazing model to that of current techniques in this section. The results are assessed using multiple quantitative measures, including the structural similarity index (SSIM), peak signal-to-noise ratio (PSNR), visual information fidelity (VIF), fog aware density evaluator (FADE), color–haze ratio (C–H Ratio), universal image quality index (UIQI), and the average processing time. These parameters collectively evaluate perceptual quality, fidelity, sharpness, color balance, and computational efficiency. Furthermore, we provide visual comparisons to illustrate the qualitative improvements our method achieves.
Table 1 illustrates the quantitative results obtained for various dehazing methods, including our proposed model. To ensure a comprehensive and balanced evaluation, five representative baseline methods were selected to encompass the primary methodological categories in single-image dehazing. The Contrast-Limited Adaptive Histogram Equalization (CLAHE) method from paper [
51] represents a classical, non-learning image-processing approach widely used for contrast enhancement in degraded imagery. The Dark Channel Prior (DCP) model proposed in paper [
44] serves as a physics-based prior baseline, offering a traditional benchmark for transmission and atmospheric light estimation methods. The AOD-Net model from [
46] exemplifies a lightweight, end-to-end convolutional framework designed for computational efficiency through reformulation of the atmospheric scattering model. The Feature Fusion Attention Network (FFA-Net) introduced in [
49] represents a modern deep learning baseline utilizing attention and feature-fusion mechanisms for enhanced perceptual quality. Lastly, the UTCR-Dehaze model described in [
50], a transformer-augmented U-Net variant, illustrates the encoder–decoder architecture commonly adopted in contemporary dehazing research. Together, these baselines collectively represent classical, physics-prior, lightweight, attention-based, and encoder–decoder paradigms, establishing a robust comparative foundation for assessing the effectiveness and generalizability of the proposed regression-based dehazing framework.
The findings indicate that our model achieves superior performance across most metrics, evidenced by an SSIM score of 0.9904 and a PSNR of 22.25 dB, while also delivering competitive VIF and FADE values. Importantly, it attains the lowest processing time (0.038 s), underscoring its efficiency. These results demonstrate that the proposed model outperforms state-of-the-art approaches such as Single U-Net, FFA-Net, AOD-Net, and the Dark Channel Prior method, both in quantitative accuracy and computational practicality. Moreover, the quality of the images produced, compared to other standard models, can be clearly seen in
Figure 3. The suggested regression-based model achieves a PSNR of 22.25 dB, which is noticeably higher than the compared techniques, as indicated by the quantitative data presented in
Table 1. This notable improvement arises from the intrinsic design of the model rather than bias in the dataset or selected image conditions. In contrast to deep learning frameworks that depend on end-to-end feature abstraction, our approach employs regression-driven estimation of the scattering coefficient (
) and refinement of the transmission map (
) to directly model the physical scattering process. By combining brightness–saturation cues with a pixel-level regression, the model successfully restores the true scene radiance without overfitting to specific patterns or textures.
As shown in
Table 2, the top 10 regression coefficient triplets
obtained during the dehazing model optimization are reported, along with their corresponding quality evaluation metrics (SSIM, PSNR, VIF, FADE, C–H ratio, and UIQI). The results highlight that the optimal coefficients, particularly around
100,628.8,
100,634.1, and
100,637.8, yield the highest SSIM of
, thereby demonstrating the stability of regression-based parameter selection for image dehazing.
To ensure the robustness of the proposed model, furthermore, we compared it with the RYF-Net model discussed in [
48]. The comparative results across different datasets (i.e., Indoor SOTS [
52], D-HAZY [
54], (OHI) ImageNet [
55], and HazeRD [
56]), under consistent experimental conditions, are presented in
Table 3. The D-HAZY dataset uses depth data from the Middlebury stereo set to synthesize hazy indoor and outdoor images. The large-scale OHI (ImageNet) dataset extends ImageNet with natural and hazy image pairs for improved generalization. HazeRD contains real foggy images captured under diverse atmospheric conditions, offering realistic complexity for robustness evaluation. Indoor-SOTS, part of the RESIDE benchmark, provides paired clear and hazy indoor images for unbiased dehazing assessment. For a fair comparison, some representative images are presented in (
Figure 4), and the results of the proposed method are evaluated under the same experimental settings and protocols as those employed in [
48]. As evident from the tabulated and visual outcomes, the proposed model outperforms RYF-Net in indoor environments compared to outdoor scenarios. The observed behavior can be attributed to the proposed model’s regression-based estimation of the scattering coefficient and transmission map using brightness–saturation cues from the HSV color space, which makes it highly effective in controlled and uniformly illuminated indoor environments. This physically grounded approach enables precise recovery of fine structural and color details, leading to superior SSIM and PSNR values for indoor datasets. However, in outdoor scenes with complex illumination and depth variations, deep fusion networks like RYF-Net, which integrate multi-scale contextual features, handle non-uniform haze more effectively.
5. Discussion
The proposed framework, in contrast to current methods that depend exclusively on prior-based assumptions or deep neural networks, employs an analytical method to estimate the scattering coefficient () by making use of brightness–saturation cues () that are obtained from the HSV color space. This makes it possible to describe the differences in haze density throughout the various parts of the image in an adaptive manner, which enables precise reconstruction even in scenarios when the haze is not uniformly distributed over the image. We conducted a comprehensive comparative evaluation of our suggested regression-based dehazing model versus well-established and state-of-the-art techniques, such as DCP, Single-U-Net, and Double-U-Net. Classical approaches such as DCP offered computational simplicity; nevertheless, they demonstrated limited performance when it came to handling dense or spatially uneven haze. This frequently led to color distortion and the loss of small details, which was reflected in poorer SSIM and PSNR scores. In a similar vein, sophisticated deep architectures like as Single-U-Net and Double-U-Net displayed increased performance through encoder–decoder feature learning and hierarchical refinement. However, these designs continued to require a significant amount of processing resources and struggled to generalize successfully under extreme meteorological fluctuations. These findings provide evidence that the regression-based architecture is both efficient and robust. The proposed image dehazing model markedly outperforms current techniques, attaining high levels of both quantitative and qualitative performance. It recorded an average SSIM of 0.99, PSNR of 22.25 dB, and VIF of 1.08, in addition to having the shortest processing time of around 0.038 s per image. As a consequence, the model provides a solution that is scalable, interpretable, and hardware-efficient for dehazing applications used in physical environments.
Furthermore, existing state-of-the-art models, such as the RYF-Net model [
48], rely heavily on large paired datasets, making them less adaptable to unseen haze conditions. Additionally, their multi-stream architecture increases computational cost, limiting real-time use. Moreover, RYF-Net lacks physical interpretability since it does not explicitly estimate parameters such as the scattering coefficient
or atmospheric light
, often leading to color distortions under non-uniform haze conditions. In contrast, our pixel-dehaze approach explicitly models these physical parameters using a regression-based estimation built on brightness–saturation cues, enabling better interpretability and adaptability. It employs localized atmospheric light estimation through quad-tree search and a lightweight least-squares regression for
, reducing complexity while maintaining high image quality. Furthermore, when a reference clear image is available, pixel-dehaze refines depth estimation for improved detail recovery. Overall, our method achieves higher SSIM and PSNR with faster inference, making it more efficient and reliable for real-time applications where RYF-Net’s data and computational demands are limiting. This new methodology is particularly beneficial in important applications, including real-time surveillance, firefighting, and autonomous driving, demonstrating its capacity to tackle visibility issues in various contexts. Moreover, the model’s efficacy and versatility underscore its potential for extensive practical use, providing a dependable solution for dehazing in intricate and safety-sensitive situations.
6. Conclusions
Haze removal is a vital image processing problem with significant implications for real-world applications such as autonomous driving, surveillance, and emergency response. This study compared the performance of various dehazing approaches, including Dark Channel Prior, Single-U-Net, Double-U-Net, and the proposed model that integrates atmospheric scattering coefficients with a regression-based depth estimation strategy, while traditional methods like Dark Channel Prior exhibited limited effectiveness in handling non-uniform haze and restoring fine details, leading to lower SSIM values, advanced architectures such as Single-U-Net and Double-U-Net demonstrated improved results but were still surpassed by the proposed approach. The suggested model achieved the highest SSIM value of 0.99 along with superior PSNR, VIF, and processing time, thereby outperforming state-of-the-art techniques in both quantitative and qualitative evaluations. By effectively leveraging atmospheric scattering coefficients and regression-based depth estimation, it successfully addressed the challenges of non-uniform haze distribution and restored clarity even in heavily degraded regions. These results confirm that the proposed model is a reliable and efficient solution for the dehazing process, offering consistently high performance and strong practical utility in applications requiring enhanced visibility, including autonomous navigation, security systems, and disaster management.
Future research could expand the model’s capability to handle real-time video streams, hence increasing its utility in dynamic environments. Incorporating other data modalities, such as LiDAR or infrared images, may also improve the model’s performance under harsh conditions and broaden its applicability to a broader range of applications. Furthermore, the focus will be on trying to execute the operations in real-time with minimal lag due to processing so that they can be efficiently used in cases of emergencies.