1. Introduction
Sea fog is a common hazardous weather phenomenon, typically occurring in the atmosphere near the sea surface. During the warm season, sea fog tends to form over sea surface temperature (SST) minima in shallow water areas, under conditions of a stable atmosphere [
1]. In regions affected by sea fog, the horizontal visibility is reduced to less than 1 km [
2], posing serious threats to the safety of coastal transportation and maritime navigation. The atmosphere near the sea surface is more stable at night than during daytime, and the absence of solar radiation facilitates both the formation and prolonged persistence of sea fog. Therefore, research on nighttime sea fog detection methods is of great significance. Observational data shows that sea fog occurs most frequently in the northwestern Pacific during summer. Notably, the Bohai Sea and the Yellow Sea experience frequent sea fog events during spring and summer [
1,
3], significantly impacting maritime safety in China’s coastal waters. Traditional sea fog detection relies on in situ observational data from limited coastal stations, ships, buoys, and other platforms [
1,
2,
4,
5]. These data are spatially scattered and insufficient for large-scale sea fog detection requirements. Advances in satellite remote sensing technology have facilitated the development of sea fog detection methods, which have now become the primary means for sea fog detection.
Traditional nighttime sea fog detection methods primarily utilize the dual-channel brightness temperature difference technique (DCD). Hunt et al. [
6] first demonstrated in 1973 that fog and low clouds exhibit lower emissivity in the Mid-Infrared (MIR) band than in the Thermal Infrared (TIR) band, laying the foundation for satellite-based remote sensing of sea fog. Eyre et al. pioneered a nighttime fog detection method using 3.7 μm and 10.8 μm infrared band data from the Advanced Very High Resolution Radiometer (AVHRR). Subsequently, Wu et al. [
7] validated the effectiveness of the dual-infrared channel brightness temperature difference method using Moderate Resolution Imaging Spectroradiometer (MODIS) satellite data. However, the DCD-based methods depend on manual threshold selection, and the high variability of sea fog makes it challenging to determine appropriate thresholds. Ellrod et al. [
8] pointed out the poor detection performance of the DCD method for thin fog and stratus. Subsequent researchers have proposed various refinements. Cermak et al. [
9] and Chaurasia et al. [
10] incorporated the local standard deviation of the TIR band into the infrared brightness temperature difference method, further enhancing the algorithm’s performance. Amani et al. [
11] added the difference between TIR brightness temperature and sea surface temperature (SST), improving the accuracy of nighttime sea fog detection. Despite these substantial improvements, traditional infrared detection methods still face difficulties with manual threshold selection, which remains a major constraint on further advancing detection accuracy.
On the other hand, Miller et al. [
12] utilized Cloud-Aerosol Lidar and Infrared Pathfinder Satellite Observations (CALIPSO) and Visible Infrared Imaging Radiometer Suite (VIIRS)/Day-Night Band (DNB) data to demonstrate that traditional DCD-based methods are prone to false alarms for low clouds/fog under certain conditions. Specifically, in the presence of a temperature inversion near the sea surface (characterized by cooler sea surface temperatures overlain by a warm, moist atmospheric layer), DCD-based methods tend to overestimate sea fog coverage. This issue is prevalent in coastal upwelling zones, river estuaries, and oceanic frontal regions, significantly undermining the reliability of DCD-based methods [
12]. The VIIRS/DNB sensor, with its high spatial resolution and accurate radiometric calibration, provides visible-like imagery during nighttime, offering new opportunities for improving sea fog detection. Miller et al. [
12] found that incorporating VIIRS/DNB data can help mitigate false alarms in traditional infrared detection methods during temperature inversion events. Given these advantages, several researchers have incorporated DNB data into fog detection, developing multi-channel threshold detection algorithms [
13,
14,
15]. However, the fog cases selected in these studies were relatively limited, and the focus has been predominantly on land fog rather than sea fog. Furthermore, the highest average probability of detection achieved was merely 0.86 [
14]. In summary, the current state of traditional nighttime sea fog detection research can be outlined as follows: (1) Infrared data-based sea fog detection methods have achieved considerable success through years of development. Furthermore, compared to VIIRS/DNB data, infrared data offer broader data availability and higher temporal resolution. However, their accuracy is considerably compromised in the presence of sea surface temperature inversions. (2) The VIIRS/DNB’s unique capability to provide visible-like imagery at night shows great potential in mitigating false alarms associated with DCD techniques, offering significant advantages for sea fog detection. Nevertheless, related research remains limited, indicating substantial research significance and potential. This study aims to develop a novel nighttime sea fog detection method leveraging VIIRS/DNB data to address these existing gaps.
In recent years, the rapid advancement of artificial intelligence (AI) technology has provided new opportunities for improving sea fog detection. Its outstanding nonlinear fitting capabilities offer a promising approach to address the threshold selection problem inherent in DCD-based methods and increase sea fog detection accuracy. In both infrared and visible satellite imagery, sea fog typically presents as a smooth, homogeneous gray-white cloud layer that AI models can effectively learn to achieve accurate identification. Several studies have begun exploring AI-based approaches for sea fog detection. Hu et al. [
16] utilized Himawari-8 satellite data and neural networks to achieve effective classification of sea fog and clouds in China’s Bohai Sea and Yellow Sea regions, although their method could not effectively detect the spatial extent of sea fog. Yi et al. [
17] applied a fully convolutional neural network to FY-4A infrared data for sea fog detection during the dawn period, but the label data were still derived from the DCD algorithm, potentially decreasing detection accuracy. Considering the impact of label data on sea fog detection accuracy, some researchers have adopted weakly supervised or unsupervised AI methods to reduce reliance on pixel-level labels. Among these, Shin and Kim et al. [
18] implemented sea fog identification using the Expectation-Maximization algorithm, based on infrared channel data from the Communication, Ocean and Meteorological Satellite (COMS) and the Operational Sea Surface Temperature and Sea Ice Analysis (OSTIA) sea surface temperature dataset. Huang et al. [
19] proposed a novel sea fog detection method using Class Activation Mapping with Himawari-8 data, effectively alleviating the labor-intensive manual labeling requirements for traditional supervised methods. These studies demonstrate that AI methods can effectively eliminate the dependence on fixed thresholds for sea fog detection. These methods have achieved excellent results in infrared data-based sea fog detection, significantly improving detection accuracy. Considering the distinct textural characteristics of sea fog in DNB imagery, this study introduces an AI approach for nighttime sea fog detection using VIIRS/DNB data.
This study aims to develop a novel deep learning model based on Generative Adversarial Networks (GANs) for detecting nighttime sea fog using VIIRS/DNB data. The VIIRS/DNB exhibits high radiometric calibration accuracy, enabling exceptional capabilities in generating visible-light cloud imagery under low-illumination conditions such as nighttime. This capability effectively mitigates false alarms caused by sea surface temperature inversion in traditional infrared detection algorithms [
12]. Furthermore, the VIIRS/DNB channel offers unique advantages in detecting sea fog beneath thin clouds. Thin cirrus clouds are optically thick in the thermal infrared, preventing satellite infrared channels from detecting underlying sea fog. In contrast, lunar reflectance exhibits stronger scattering characteristics within the visible light spectrum. Consequently, the VIIRS/DNB satellite channel can effectively penetrate semi-transparent thin clouds for sea fog detection [
20], highlighting its significant potential for sea fog detection applications. Unlike other AI methods (e.g., convolutional neural networks, Vision Transformer) that utilize pixel-wise loss function calculations, GANs emphasize structural and textural consistency between generated outputs and ground truth labels. This leads to segmentation results with improved structural integrity and semantic coherence. Sea fog, composed of spatially continuous distributed micro-droplets, exhibits more homogeneous grayscale and textural characteristics in VIIRS/DNB imagery. These distinctive characteristics make GANs particularly suitable for image segmentation tasks in sea fog detection. In summary, this study proposes training a GANs model using VIIRS/DNB data to achieve effective detection of nighttime sea fog.
The remainder of this paper is organized as follows.
Section 2 introduces the study area, satellite data, and proposed method.
Section 3 presents the experimental results and provides a detailed analysis.
Section 4 discusses relevant aspects of this research. Finally,
Section 5 summarizes the findings of this study.
3. Experiments and Results
Based on the constructed sea fog detection dataset, this study proposes a novel generative adversarial network model. To comprehensively evaluate the performance of the proposed model, we reproduced the latest operational infrared detection algorithm for comparison. Sea fog detection can also be considered a specialized image segmentation task. To evaluate the performance of the proposed model, we selected several widely adapted deep learning models in related fields for comparative analysis, including R2U-Net [
27], Attention R2U-Net, and DA-TransUNet [
28]. Among these, Attention R2U-Net is a variant of R2U-Net [
27] that incorporates Attention Gate [
29] into its skip connections. DA-TransUNet [
28] employs a hybrid architecture integrating convolutional neural networks with Vision Transformer [
30], which has achieved remarkable performance in the field of image segmentation. Additionally, to examine the contribution of individual modules within the proposed method, we designed three corresponding ablation experiments. This section will provide a detailed description of the experimental setup and present a specific analysis of selected sea fog case study results.
3.1. Experimental Setup
This study first partitioned the dataset, which contains sea fog data collected from 2017 to 2024. Given that the sea fog cases acquired in 2023 are more comprehensive and temporally cover the peak seasons of sea fog occurrence (spring and summer), thus being more representative, we designated the 2023 sea fog data as the test set, with the remaining data serving as the training set. Random flipping was applied for image augmentation during the training process.
The SEGAN model was implemented based on the PyTorch (torch 2.4.0 + cuda 11.8) library and trained and tested on the aforementioned sea fog dataset. During training, the generator and discriminator were trained alternately for multiple epochs until final convergence was achieved. In practice, the Adam optimizer with an initial learning rate of 0.0029 was utilized. All other deep learning models were similarly trained within the PyTorch framework, with consistent training environment configurations to ensure a fair comparison with SEGAN. To evaluate the performance of SEGAN, this study reproduced the state-of-the-art operational infrared detection algorithm for comparison, employing Amani et al.’s [
11] method (Auto-DCD) as the benchmark algorithm. Notably, while the GEOS-16 infrared channel data used by Amani et al. has a spatial resolution of 2 km, the data used in this study has a spatial resolution of 750 m. To adapt the Auto-DCD method to the spatial resolution of the VIIRS channels, we modified the 3 × 3 kernel used by Amani et al. to a 9 × 9 kernel. Infrared data from the VIIRS M12 and M15 bands corresponding to a total of 162 sea fog cases were statistically processed to derive the final Auto-DCD test results. Fully connected Conditional Random Fields (Dense CRFs) constitute a statistical modeling method employed for pixel-level prediction tasks. To enhance the detection accuracy of the output results from the proposed model, this study incorporated Dense CRFs for post-processing. The approximate inference algorithm proposed by Philipp et al. [
31] was used to accelerate the inference process. Specifically, the confidence of the initial segmentation result was set to 0.9, and 12 iterations of inference were performed to obtain the final output.
3.2. Evaluation Metrics
To quantitatively evaluate the detection performance of the proposed method, we employed six widely adopted evaluation metrics for quantitative analysis of the test results: probability of detection (
POD), false alarm rate (
FAR), Precision,
F1-score, critical success index (
CSI), and mean Intersection over Union (
mIoU). These metrics assess the detection results from different perspectives, and their corresponding calculation formulas are as follows.
True Positives (TP), False Negatives (FN), False Positives (FP), and True Negatives (TN) are determined based on the test results, with their actual definitions as follows.
- (1)
TP: Sea fog occurs and is detected.
- (2)
FN: Sea fog occurs but is not detected.
- (3)
FP: Sea fog does not occur but is detected.
- (4)
TN: Sea fog does not occur and is not detected.
3.3. Experimental Results
This section presents a comparative analysis between the proposed SEGAN model, the conventional infrared threshold-based method (Auto-DCD), and several established image segmentation models (including R2U-Net, Attention R2U-Net, and DA-TransUNet). The evaluation encompasses comprehensive statistical results across all selected metrics, along with detailed case studies of specific sea fog events. Through both quantitative and qualitative analyses, a comprehensive assessment of the superior performance of the proposed method is conducted. Specific test results are shown in
Table 2.
We constructed a novel generative adversarial network model that effectively leverages the textural features of sea fog within VIIRS/DNB visible-light imagery data to accurately identify sea fog.
Table 2 presents a comparison between SEGAN and the latest operational infrared threshold-based detection method (Auto-DCD), as well as three other deep learning models from related fields. Comprehensive evaluation metrics demonstrate that our proposed SEGAN model significantly outperforms other comparative methods. Regarding individual metrics, SEGAN improves the POD by 0.0632 and Precision by 0.1658 compared to Auto-DCD, while reducing the false alarm rate by 0.0287. In terms of composite metrics, SEGAN attained an F1-score and mIoU both exceeding 0.83, and a CSI surpassing 0.73, highlighting its superior performance. Collectively, these metrics demonstrate that the proposed method achieves a significant enhancement in overall detection accuracy compared to conventional infrared-based approaches.
In addition, comparative analyses were conducted with several widely adapted deep learning models in the image segmentation domain. As shown in
Table 2, SEGAN outperforms the other three deep learning models across three comprehensive evaluation metrics (F1-score, CSI, and mIoU) as well as in POD, which indicates that our specifically designed SEGAN model is better suited for sea fog detection tasks compared to general-purpose models. Moreover, SEGAN surpasses DA-TransUNet in detection performance, achieving high-precision sea fog detection with a lightweight network architecture, thereby fully exhibiting its exceptional capabilities. To comprehensively evaluate the detection performance of SEGAN, we now present a detailed analysis of selected sea fog case studies.
Figure 5 illustrates a sea fog event recorded by the Suomi NPP satellite on 10 March 2023. The DNB image data reveal that portions of the sea fog were obscured by thin cloud cover. Conventional infrared threshold-based methods cannot identify such sub-cloud sea fog, leading to partial detection omissions. This limitation arises because clouds exhibit high optical thickness in infrared bands, preventing the detection of underlying sea fog. Conversely, moonlight reflectance possesses stronger scattering properties in the visible spectrum. Consequently, partial sea fog information beneath thin clouds is observable in DNB imagery (as seen in the raw image of
Figure 5). Leveraging this inherent advantage of DNB data, SEGAN effectively learns the continuous textural features of extensive sea fog within visible-light imagery, accurately identifying its spatial distribution. This enhances detection accuracy under thin cloud cover conditions. Notably, SEGAN also demonstrates effective detection of thin fog in the lower portion of the
Figure 5 case study. These findings fully demonstrate that, compared to traditional infrared threshold methods, SEGAN’s utilization of DNB imagery significantly enhances sea fog detection accuracy in the presence of thin cloud cover.
As illustrated in
Figure 5, both SEGAN and the other three deep learning models exhibit varying degrees of missed detection in this case study. Specifically, R2U-Net, and Attention R2U-Net all demonstrate a POD notably lower than 0.7 (as shown in
Table 3.). Closer examination reveals that the missed detection areas predominantly occur in the upper-left region of the case, where the sea fog concentration is relatively low. This phenomenon may be attributed to the low concentration of sea fog in the missed areas, resulting in less distinct grayscale features in the imagery data that consequently impair model detection performance. Despite these challenges, SEGAN outperforms all other models across three comprehensive evaluation metrics—F1-score, CSI, and mIoU, achieving relatively superior detection results. Unlike conventional models that optimize pixel-wise loss functions, SEGAN’s adversarial training paradigm emphasizes semantic coherence in its output, rendering it particularly suitable for sea fog detection tasks and endowing it with enhanced robustness across varying sea fog concentrations.
Figure 6 presents a sea fog event recorded by the NOAA-20 satellite on 4 March 2023. The Auto-DCD method exhibited partial detection omissions in this case. This is likely attributable to suboptimal threshold selection within the Auto-DCD algorithm, leading to missed detections in certain areas and consequently reduced sea fog detection accuracy. In contrast, SEGAN accurately identified most of the sea fog extent. Despite SEGAN’s commendable performance, our analysis revealed remaining limitations. Within the regions marked by red boxes in
Figure 6 and
Figure 7, the SEGAN model demonstrated lower detection probability for spatially limited, isolated sea fog patches compared to large, continuous sea fog areas. Conversely, the Auto-DCD method performed better than SEGAN in these regions. This suggests that the infrared brightness temperature difference method possesses unique advantages in such scenarios. Additionally, the two cases in
Figure 6 and
Figure 7 indicate that SEGAN underperforms Attention R2U-Net in detecting scattered patches marked by rectangular boxes. This suggests that while SEGAN prioritizes global semantic coherence, it remains less effective in capturing localized features within small, fragmented regions of the imagery data. Analysis indicates that the proposed SEGAN model effectively learns the textural features of sea fog within visible-light imagery, enabling accurate identification. Its AI-based approach eliminates the need for manual threshold selection, thereby improving overall sea fog detection accuracy. Crucially, SEGAN’s utilization of DNB imagery significantly enhances detection accuracy under partial thin cloud cover conditions. Although SEGAN yields promising results, its detection efficacy for certain isolated, small-scale sea fog patches requires further improvement. Future improvements will focus on two key aspects: (1) optimizing the model architecture to enhance its feature extraction capacity, (2) incorporating infrared brightness temperature difference as auxiliary input data to enrich the model’s informational context and improve detection accuracy.
3.4. Ablation Experimental Results
To comprehensively evaluate the performance of the proposed model, we designed three ablation experiments. This subsection provides a detailed description of the experimental specifics. First, the settings for ablation experiment 1 are introduced. The generator of SEGAN is proposed based on U-Net [
24]. Its skip connection structure preserves sufficient low-level features for the image segmentation task within the decoder path. However, preserving excessive low-level features may lead to information redundancy, potentially impairing the model’s overall performance. To fully exploit the model’s superior capabilities, we first designed ablation experiment 1 to investigate the number of skip connections. Specifically, while the original generator incorporates four skip connections, we incrementally reduced this number for testing, ultimately selecting the optimal parameter configuration. Given that high-level features contain richer semantic information and should be retained, we sequentially removed the skip connection modules closest to the input layer. Each configuration was trained until convergence, and the best test results were selected for comparison. The experimental results are presented in
Table 4.
A comprehensive analysis of the evaluation metrics in
Table 4 reveals that the model performance reached its optimal level when the number of skip connections was set to 2. At this configuration, the model also maintained a relatively moderate parameter count. Consequently, a skip connection count of 2 was ultimately selected. Further analysis of
Table 4 indicates that model test performance was nearly identical when the skip connection count was less than 3, with particularly similar results observed between configurations of 2 and 3 skip connections. However, a relatively significant decline in POD occurred when the skip connection count was increased to 4. This decline may be attributed to the increased model parameter size coupled with information redundancy, collectively impairing model performance. Therefore, within this study, the generative adversarial model configuration with 2 skip connections delivered the best performance. As illustrated in
Figure 8, selected test cases were analyzed to qualitatively assess model performance. The figure visually demonstrates the output results of the different model configurations. While the test results were generally similar across models, the configuration with 2 skip connections yielded the best overall detection effectiveness, consistent with the conclusions drawn from the quantitative evaluation metrics.
To enhance feature information fusion, we incorporated a spatial attention mechanism into the skip connection structure. To evaluate the efficacy of the spatial attention mechanism within SEGAN, we conducted a comparative experiment using the Hybrid Attention Mechanism (HAM) [
32], a widely adopted approach in the neural network domain. The experimental results are presented in
Table 5.
A comprehensive analysis of
Table 5 reveals that the generative adversarial model incorporating HAM exhibited a certain degree of reduction in the false alarm rate (FAR) and an improvement in Precision compared to SEGAN. However, its probability of detection (POD) showed a significant decline, indicating a marked decrease in sea fog detection performance. Furthermore, based on composite metrics such as the critical success index (CSI), F1-score, and mean Intersection over Union (mIoU), the detection performance of the proposed SEGAN model is demonstrably superior to the variant incorporating HAM.
Figure 9 presents results from selected test cases, clearly demonstrating that the variant with HAM exhibited partial detection omissions, whereas SEGAN achieved more accurate detection of sea fog.
To enhance the model’s feature extraction capability, this study introduced the SE-Net module into the generator. Ablation experiment 3 was designed to validate the contribution of the SE-Net module. Specifically, we trained both SEGAN and a variant excluding the SE-Net module, ensuring both reached convergence. The best test results corresponding to their respective optimal hyperparameters were selected for comparison. The results are presented in
Table 6.
SEGAN surpassed the variant without the SE-Net module across key metrics including probability of detection (POD), F1-score, critical success index (CSI), and mean Intersection over Union (mIoU). Notably, POD increased by 0.0557, demonstrating that the SE-Net module effectively enhances the model’s feature extraction capability. As shown in
Figure 10, selected sea fog cases reveal that SEGAN’s sea fog identification performance is superior to that of the variant without the SE-Net module, particularly in regions characterized by thin fog patches. The SE-Net module improves feature extraction by modeling dependencies among feature channels, thus explaining SEGAN’s enhanced sea fog detection effectiveness.
4. Discussion
This study proposes a novel generative adversarial network model (SEGAN) that utilizes VIIRS/DNB data to detect nighttime sea fog. Compared to conventional operational infrared detection methods, the proposed approach eliminates the need for complex threshold selection processes while improving detection accuracy. Although SEGAN demonstrates favorable detection performance, certain limitations remain. These limitations will be discussed in detail in this section.
4.1. Limitations of VIIRS Data
VIIRS/DNB data possess exceptional capability for nighttime visible-light cloud imaging. However, the actual imaging performance is influenced by lunar phase conditions. Under low moonlight fraction, the limited intensity of the lunar radiation source results in detected reflectance values being indistinguishable from noise, preventing effective cloud discrimination. Therefore, in this study, the proposed method is applicable only under conditions where the moonlight fraction is ≥70. Additionally, since DNB measures top-of-atmosphere reflectance, it cannot detect features beneath cloud layers. DNB operates within a wavelength range of 0.5 to 0.9 μm. Thick clouds significantly attenuate its optical signals, thereby limiting the detection capability primarily to radiance reflected from cloud tops. Consequently, sea fog completely obscured by high-level clouds remains undetectable.
4.2. Limited Applicability to Chinese Offshore Waters and the Sea of Japan
The SEGAN method demonstrates effective performance in the target maritime regions (China’s coastal waters and the Sea of Japan). However, its generalizability to other sea regions remains unverified due to insufficient data. Limited by the scarcity of meteorological records for sea fog events, this study collected sea fog occurrences only from China’s coastal waters and the Sea of Japan between 2017 and 2024. Furthermore, sea fog exhibits dynamic complexity and significant spatial heterogeneity. This limitation arises from variations in fundamental sea surface conditions (such as sea surface temperature and coastal topography), atmospheric environmental factors (such as surface wind conditions and humidity profiles), and regional meteorological regimes all influence the formation and persistence of sea fog. These variations may result in distinct sea fog characteristics across different sea regions, consequently limiting the generalization capability of the proposed model. To comprehensively evaluate the detection performance of the model, future work should involve collecting additional sea fog data from other maritime regions (e.g., Grand Banks, Newfoundland) for further validation.
4.3. Missed Detection in Certain Areas
As demonstrated by the test cases in
Figure 6 and
Figure 7, SEGAN exhibits detection omissions for fragmented sea fog patches. The detection probability of SEGAN in these small-scale areas is statistically lower than in extensive sea fog regions, indicating inadequate learning of textural features characterizing small-scale sea fog. The texture patterns of small-scale sea fog in visible-light imagery lack the homogeneity observed in large-scale counterparts. This inherent complexity, further compounded by the limited dataset volume, likely prevented SEGAN from effectively learning the discriminative textural features of small-scale sea fog.
4.4. Comparison of Average Inference Time
To evaluate the inference performance of the proposed model, we recorded the average time required for different methods to process individual sea fog cases (as shown in
Table 7). All deep learning models were inferred using a single NVIDIA GeForce RTX 4090 GPU with 24 GB VRAM, while the traditional infrared threshold-based method (Auto-DCD) was computed on a single CPU (Intel Core i9-14900HX). The results presented in
Table 7 were obtained under these configurations. As evidenced in
Table 7, the average inference time of deep learning models is less than 5 s, significantly reducing the computational time compared to Auto-DCD’s 1199.0961 s. Furthermore, when combined with the inference times presented in
Table 7, it is evident that SEGAN achieves inference speeds comparable to R2U-Net and its variants, while reducing inference time by 1.1779 s compared to DA-TransUNet. This demonstrates that our lightweight architectural design effectively minimizes computational overhead and accelerates the inference process. This efficiency enables SEGAN to be integrated with satellite data for rapid detection in practical scenarios, delivering real-time and reliable sea fog information to support ship navigation. Furthermore, SEGAN’s inference time is comparable to that of R2U-Net, indicating that the lightweight network architecture adopted in this study demands relatively few computational resources and can deliver inference results rapidly.
4.5. Future Research Direction
In future work, we will enhance the sea fog detection performance of the proposed methodology through the following aspects. (1) Dataset Expansion: The sea fog dataset will be extended by acquiring global sea fog records and collecting corresponding VIIRS/DNB satellite data. This expansion aims to improve the model’s generalization capability. (2) Integration of Multi-source Meteorological Data: Diverse meteorological data sources, such as infrared channel data, lidar data, will be incorporated. This integration seeks to enhance applicability across diverse scenarios. As SEGAN currently utilizes only single-channel satellite imagery, detecting sea fog beneath cloud layers remains challenging. Lidar, providing vertically resolved high-resolution water vapor profiles throughout the atmospheric column, can effectively identify low clouds and sea fog. Future efforts could explore leveraging lidar data as auxiliary information to guide the proposed model towards detecting sub-cloud sea fog. (3) Incorporation of Sea Surface Temperature (SST): SST information will be considered to further boost detection capability. Given sea fog’s proximity to the sea surface, its temperature closely resembles the SST. Utilizing the brightness temperature difference between the top of the atmosphere and the SST could aid in distinguishing high-level clouds, thereby improving sea fog detection. (4) Model Architecture Optimization: The model structure will be optimized to enhance feature learning capacity, specifically targeting improved detection capability for small-scale sea fog patches.
5. Conclusions
Sea fog is a relatively common hazardous weather that significantly impacts ship navigation and carrier-based aircraft takeoff and landing. The persistence of sea fog within the nighttime marine boundary layer is prolonged, resulting in greater impacts compared to daytime occurrences. However, existing detection methods face challenges in threshold selection and suffer from false alarms concerning spatial extent. To address these issues, this study developed a novel generative adversarial network model (SEGAN) by integrating the SE-Net module [
25] and a spatial attention mechanism based on the U-Net architecture [
24]. SEGAN is trained on the constructed VIIRS/DNB dataset. To validate the model’s performance, the latest operational infrared sea fog detection algorithm (Auto-DCD) and three widely adapted image segmentation models (R2U-Net, Attention R2U-Net, and DA-TransUNet) were selected for comparison. The results demonstrate that SEGAN achieved a detection probability of 0.8708, representing an improvement of 0.0632 over the Auto-DCD method. Simultaneously, the false alarm rate decreased by 0.0287, and the CSI increased by 0.1587, fully illustrating the superior detection performance of SEGAN. Furthermore, SEGAN reduces inference time by 1.1779 s compared to DA-TransUNet. Moreover, SEGAN surpasses DA-TransUNet in detection performance, achieving high-precision sea fog detection with a lightweight network architecture, thereby fully exhibiting its exceptional capabilities. Several case studies were selected for qualitative analysis, confirming that SEGAN’s detection performance is overall better than that of comparative methods. Notably, in cases where sea fog is partially obscured by thin cloud layers, SEGAN exhibits more reliable detection compared to Auto-DCD, effectively improving detection accuracy. Unlike conventional models that rely on pixel-wise loss optimization, SEGAN’s adversarial training paradigm emphasizes semantic coherence in its output, rendering it particularly suitable for sea fog detection tasks and endowing it with enhanced robustness across varying sea fog concentrations.
SEGAN holds significant importance for ensuring the safety of ship navigation and carrier-based aircraft operations. By accurately identifying the spatial extent and distribution of sea fog, it provides reliable support for ship route planning, enabling vessels to avoid fog-affected areas effectively. Furthermore, sea fog can be utilized to conceal the movements of aircraft carriers. From a scientific perspective, sea fog is a crucial component of boundary layer clouds and holds a significant influence on the global radiation budget. nighttime sea fog events exhibit relatively high frequency and prolonged duration, exerting considerable influence on radiative processes. Accurate spatial distribution information of sea fog can provide reliable data for theoretical research on the global radiation balance, thereby contributing to the advancement of studies on global climate change.
Although SEGAN has achieved promising results, further analysis reveals certain limitations. First, the quality of DNB data is affected by lunar phase conditions. SEGAN can only be used to detect sea fog under lunar fraction conditions of 70 or higher. Second, the collected sea fog data are currently limited to offshore areas of China, and the lack of data from other sea areas may constrain the model’s generalization capability. Third, the detection of scattered small-scale sea fog regions remains problematic, with the current detection probability for such areas being relatively low. To address these issues, we plan to implement improvements in the following aspects: (1) Collecting global sea fog data to expand the dataset and enhance the model’s generalization ability. (2) Incorporating multi-source input data such as lidar, satellite infrared channels, and sea surface temperature reanalysis data to broaden the applicability of the proposed method to diverse scenarios. (3) Optimizing the model structure to improve its feature extraction capabilities.