Next Article in Journal
The Effective Baryon–Baryon Potential with Configuration Mixing in Quark Models
Next Article in Special Issue
Radio Spectrum Observations and Studies of the Solar Broadband Radio Dynamic Spectrometer (SBRS)
Previous Article in Journal
Editorial to the Special Issue “Universe: Feature Papers 2023—Cosmology”
Previous Article in Special Issue
Calibration of Polarization Data for Vector Magnetographs at the Huairou Solar Observing Station over the Past Four Decades
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automated High-Precision Recognition of Solar Filaments Based on an Improved U2-Net

1
Nanjing Institute of Astronomical Optics & Technology, Chinese Academy of Sciences, Nanjing 210042, China
2
CAS Key Laboratory of Astronomical Optics & Technology, Nanjing Institute of Astronomical Optics & Technology, Nanjing 210042, China
3
University of Chinese Academy of Sciences, Beijing 100049, China
*
Author to whom correspondence should be addressed.
Universe 2024, 10(10), 381; https://doi.org/10.3390/universe10100381
Submission received: 13 August 2024 / Revised: 6 September 2024 / Accepted: 11 September 2024 / Published: 29 September 2024

Abstract

:
Solar filaments are a significant solar activity phenomenon, typically observed in full-disk solar observations in the H-alpha band. They are closely associated with the magnetic fields of solar active regions, solar flare eruptions, and coronal mass ejections. With the increasing volume of observational data, the automated high-precision recognition of solar filaments using deep learning is crucial. In this study, we processed full-disk H-alpha solar images captured by the Chinese H-alpha Solar Explorer in 2023 to generate labels for solar filaments. The preprocessing steps included limb-darkening removal, grayscale transformation, K-means clustering, particle erosion, multiple closing operations, and hole filling. The dataset containing solar filament labels is constructed for deep learning. We developed the Attention U2-Net neural network for deep learning on the solar dataset by introducing an attention mechanism into U2-Net. In the results, Attention U2-Net achieved an average Accuracy of 0.9987, an average Precision of 0.8221, an average Recall of 0.8469, an average IoU of 0.7139, and an average F1-score of 0.8323 on the solar filament test set, showing significant improvements compared to other U-net variants.

1. Introduction

Solar filaments are a typical solar activity phenomenon in the solar atmosphere. These filaments consist of plasma on the solar surface with temperatures ranging from 6000K to 8000K, approximately 100 times cooler than the surrounding corona. The composition of filaments is similar to that of the chromosphere, and they form ring-like structures that extend into the corona. When observed against the space background at the solar limb, they are referred to as prominences [1]. Solar filaments sometimes align with the polarity inversion lines of the solar photospheric magnetic field. Filaments located in active regions are often unstable and prone to eruption. Eruptive filaments are closely associated with solar flares and coronal mass ejections (CMEs) [2], and are considered manifestations of the same physical process of magnetic energy release at different times and heights in the solar atmosphere [3]. The eruption of solar filaments is not only regarded as the core of CMEs but also as central to the research of CMEs [4]. CMEs can cause damage to satellites and other spacecraft, disrupt communication, navigation, and remote sensing on Earth, and damage power transmission equipment, leading to large-scale power outages. Furthermore, CMEs may jeopardize the normal operation of space stations and pose a threat to the safety of astronauts [5].
Solar filaments are often observed in full-disk H-alpha images. Many solar telescopes around the world can observe the H-alpha wavelength, such as those in the global high-resolution H-alpha network, which includes the Big Bear Solar Observatory in California, Kanzelhöhe Solar Observatory in Austria, the Catania Astrophysical Observatory (CAO) in Italy, Meudon and Pic du Midi Observatories in France, and the Huairou Solar Observing Station (HSOS) and the Yunnan Astronomical Observatory (YNAO) in China [6]. Other notable telescopes include the Daniel K. Inouye Solar Telescope [7] and the planned European Solar Telescope [8]. To obtain accurate details of solar filaments in the H-alpha band, solar telescopes have been launched into heliocentric orbit. The Chinese H-alpha Solar Explorer (CHASE), launched in 2021, carries the H-alpha Imaging Spectrograph, which provides solar filament images with clear details in the H-alpha wavelength [9,10].
With the dramatically increasing volume of solar observational data, achieving automated high-precision recognition and segmentation of solar filaments, while preserving as much detail as possible, is crucial. This allows for the investigation of the formation, deformation, and eruption of solar filaments. Understanding the dynamics and evolution of solar filaments and their relationship with the magnetic fields in solar active regions, as well as predicting space weather phenomena such as solar flares and coronal mass ejections, are of paramount importance [1].
Artificial intelligence technologies, represented by convolutional neural networks [11], have broad applications in image recognition and segmentation, including in fields such as biomedical imaging [12], satellite remote sensing [13], and astronomical image processing [14]. In solar physics, deep learning-based image segmentation techniques are frequently applied to the identification of solar filaments, sunspots, and solar plages [15]. Zhu [16] proposed a solar filament segmentation method based on the U-net neural network. Liu et al. [17] implemented solar filament segmentation using a U-net improved with generative adversarial networks. Xin et al. [18] introduced attention mechanisms, deep supervision networks, and Inception modules into V-Net to achieve solar filament segmentation. Guo et al. [19] proposed a deep learning model based on CondInst for the detection and classification of solar filaments, facilitating the classification of isolated and non-isolated filaments. Zheng et al. [20] utilized U-net for solar filament segmentation and employed the CSRT algorithm to track their evolution.

2. Data Preprocessing and Dataset Construction

The CHASE telescope’s H-alpha Imaging Spectrograph captures full-disk solar images, storing the data in FITS format. Each dataset includes high-quality solar grayscale images obtained at 118 wavelength nodes within the 655.97 nm to 656.59 nm range. This range is centered around the H-alpha wavelength node (656.28 nm), where the detailed features of solar filaments can be clearly observed. The following data preprocessing steps apply to all 118 solar grayscale images within each dataset.

2.1. Limb-Darkening Removal

Solar filaments are plasma structures with lower temperatures than the corona. When located at the solar limb, they are referred to as prominences and appear brighter against the space background. The Sun also exhibits limb darkening, a phenomenon where the brightness decreases from the center of the solar disk to the edge. The darker regions at the edge may resemble the morphology of solar filaments, which can hinder the precise identification of filament details using K-means clustering. Therefore, to minimize the occurrence of false negatives during deep learning training, it is necessary to fit the solar disk’s edge using a second-order cosine function. This approach helps retain most of the brighter regions of the solar surface, excluding the edge. Finally, a solar image of size 2048 ∗ 2048 is obtained, with the sun center located at the center of the image and the full solar disk radius of 950.

2.2. Grayscale Transformation and K-Means Clustering

The grayscale values of the solar images are stretched to the range of 0–255. Since K-means clustering is sensitive to regions with concentrated pixel distributions, stretching the grayscale values of the solar H-alpha monochromatic images after limb-darkening removal enhances the contrast between solar filaments and the solar surface. This makes it easier for K-means clustering to identify filaments accurately and efficiently.
K-means clustering [21] is an unsupervised deep learning method primarily used to partition a dataset into K pre-specified clusters. The goal of K-means clustering is to maximize the similarity of samples within each cluster while maximizing the difference between samples from different clusters. This method effectively distinguishes solar filaments, sunspots, and solar plages, and is widely used in solar physics. After limb-darkening removal and grayscale stretching, K-means clustering analysis is performed on the solar images at all 118 wavelength nodes, with the number of clusters K set to 50. A lower K value would be insufficient to differentiate solar filament signals from noise. The cluster that best matches the detailed features of solar filaments is manually selected from the K clusters obtained through K-means clustering.

2.3. Particle Erosion, Multi-Closing Operations, and Hole Filling

The initial segmented images of solar filaments obtained from K-means clustering may contain small background particles, as well as holes and fragments within the filaments. These small particles may originate from incomplete differentiation of tiny chromospheric fibers, while the holes and fragments may result from the imperfect recognition of detailed features of solar filaments by K-means clustering. Using OpenCV, an image processing library, particles with a pixel area smaller than 60 are firstly eroded from the initial segmented images. Secondly, multi-closing operations are applied to connect fragmented pieces of solar filaments. Finally, the contours are filled to eliminate holes within the solar filaments. This process yields the final labeled images of solar filaments, effectively minimizing noise while preserving the filament signals as much as possible.

2.4. Dataset Construction

In the study, we utilized 150 full-disk solar images captured in the H-alpha wavelength band by the Chinese H-alpha Solar Explorer’s imaging spectrometer in 2023, downloaded from the Solar Science Data Center of Nanjing University [22]. The images underwent preprocessing steps, including limb-darkening removal, grayscale transformation, K-means clustering, particle erosion, multiple closing operations, and hole filling, to generate corresponding labeled images. Figure 1 illustrates the specific steps of data preprocessing, while Figure 2 illustrates the results of the solar full-disk H-alpha images before and after preprocessing. All images were uniformly resized to 512 × 512 pixels to ensure consistency. Randomly, a total of 560 selected images were assigned to the training set, 22 images to the validation set, and the other 18 images were used as a test set to evaluate the efficiency of the neural network in deep learning. Due to the lack of symmetry in convolutional neural networks when processing images, the dataset was augmented through automatic rotations, horizontal flips, and vertical flips in the training set.

3. Deep Learning

3.1. Neural Networks

3.1.1. U-Net Family

U-Net is a neural network architecture proposed by Ronneberger et al. [12] for semantic segmentation in deep learning, originally applied in biomedical image segmentation. Inspired by an encoder–decoder structure, U-Net exhibits a symmetrical U-shaped appearance, hence its name. The network consists of two main parts: the encoder and the decoder. The encoder consists of a series of convolutional and pooling layers designed to gradually reduce the spatial dimensions of the input image and extract features. With each convolutional operation, the number of channels increases progressively, while pooling layers reduce the spatial resolution of the feature maps while retaining important features. The symmetric decoder consists of a series of upsampling and convolutional layers. Upsampling layers are used to restore the spatial resolution of the feature maps to the original input image size.
A distinctive feature of U-Net is its use of skip connections, which connect corresponding features from the encoder to the decoder. This mechanism helps address issues of information loss and spatial inconsistency in segmentation tasks, enabling the network to learn richer features and improving overall performance. The architecture of U-Net is illustrated in Figure 3.
Ozan et al. [23] proposed Attention U-Net based on the U-Net architecture, incorporating the attention mechanism [24] into the upsampling path for medical image segmentation. This enhancement improves the neural network’s capability of feature selection, thereby enhancing segmentation accuracy, training efficiency, and model generalization. The architecture of Attention U-Net is illustrated in Figure 4. Building on the U-Net architecture, Qin [25] proposed U2-Net, which utilizes a two-level nested U-Net structure. The classical U2-Net employs residual U-blocks (RSU) to replace the standard convolutional blocks used as encoders or decoders in U-Net. Each RSU block consists of an encoder and decoder with a structure similar to U-Net, combining different receptive field sizes. Each RSU block can be viewed as a “small U-Net”, incorporating residual connections to merge local and multi-scale features through summation. The height of the RSU blocks ranges from 7 to 4, decreasing with the height of the U2-Net encoder or decoder. This design captures more contextual information at different scales, significantly increasing the network’s depth and enhancing training and generalization efficiency. Consequently, U2-Net is widely applicable to various image segmentation tasks in solar observation environments. The architecture of the residual U-block is shown in Figure 5. The architecture of U2-Net is illustrated in Figure 6.
Li et al. [26] proposed AM-SegNet, a model designed for segmentation and quantification of high-resolution X-ray images. AM-SegNet combines a lightweight convolutional block with a custom attention mechanism to perform image segmentation and feature quantification on X-ray imaging data collected from various synchrotron experiments. AM-SegNet outperforms other state-of-the-art segmentation models in terms of accuracy, speed, and robustness.

3.1.2. Attention Gate

Attention mechanisms are widely used in machine translation and computer vision [23]. In the field of semantic segmentation, the attention mechanism is incorporated into U-Net in the form of attention gates, creating Attention U-Net. Attention gates receive a set of low-level features (typically high-resolution feature maps) from the encoder path and a set of high-level features (typically low-resolution feature maps) from the decoder path. Through convolution operations and ReLU activation functions, the input feature maps are mapped to a common space.
Specifically, a 1 × 1 convolution is applied to the low-level features to reduce the number of channels, and a 1 × 1 convolution is also applied to the high-level features to map both sets of features to the same dimension. The mapped features are then added together to obtain a new feature map. This combined feature map is processed using the ReLU activation function to set all negative values to zero, adding nonlinearity. Subsequently, a 1 × 1 convolution and a sigmoid activation function are applied to compute the attention weights for each pixel location. Finally, these attention weights are multiplied with the original low-level feature map to produce the weighted feature map.
Attention gates can enhance the response to target regions in semantic segmentation, suppress irrelevant background noise, and dynamically adjust based on the input image’s features. This enhances the neural network’s robustness when handling complex backgrounds or small target images. The attention mechanism is shown in Figure 7.

3.1.3. Attention U2-Net

Building upon U2-Net and Attention U-Net, this work adopts the U-Net encoder–decoder architecture as the backbone network. Attention gates are introduced into the upsampling stages of the RSU blocks, replacing the standard convolutional blocks of U-Net with RSU blocks enhanced by attention gates. Additionally, attention gates are incorporated into the upsampling stages of the backbone network, resulting in the creation of Attention U2-Net. This architecture not only captures more contextual information at different scales but also enhances the neural network’s robustness in handling complex backgrounds or small object images. The architecture of Attention U2-Net is illustrated in Figure 8.
Sigmoid is defined in Equation (1):
f ( x ) = 1 ( 1 + e x )
ReLU is defined in Equation (2):
f ( x ) = 0 , x < 0 x , x 0

3.2. Experimental Setup and Training Parameters

The neural network training was conducted on a Windows 10 operating system using Python 3.11. The software environment included the parallel computing architecture CUDA 12.1, the deep learning library PyTorch 2.1, and the GPU acceleration library cuDNN 9.0.10. The hardware environment comprised an Intel Xeon Platinum 8173M CPU with 128 GB RAM and an NVIDIA Quadro RTX4000 GPU with 8 GB VRAM. During training, the learning rate was set to 0.0001, the batch size was set to 2, and the number of iterations was set to 150. At the beginning of each training session, the loss functions for both the training set and validation set are recorded and compared after each iteration. Training is terminated when the loss functions for both sets gradually decrease and converge, in order to prevent the model from overfitting on the test set.

3.3. Optimization Function and Loss Function

After selecting U2-Net as the neural network for image segmentation in this study, it was essential to determine the most effective optimizer and loss function to facilitate network convergence. The Adam optimizer provided by PyTorch was chosen in the study [27]. Adam is an adaptive learning rate optimization algorithm that adjusts the learning rate dynamically using estimates of the first and second moments of the gradients. The loss function selected for this study was the binary cross entropy loss (BCELoss) function.
The Adam optimizer is defined in Equation (3):
g = θ k 1 L ( θ ) m k = β 1 m k 1 + ( 1 β 1 ) g v k = β 2 v k 1 + ( 1 β 2 ) g g m ^ k = m k 1 β 1 k v ^ k = v k 1 β 2 k θ k = θ k 1 η v ^ k + ϵ m ^ k
The BCELoss function is defined in Equation (4):
H p ( q ) = 1 N i = 1 N y i · l o g ( p ( y i ) ) + ( 1 y i ) · l o g ( 1 p ( y i ) )

3.4. Evaluation Metrics

This study evaluates and compares the performance of four neural networks: U-Net, Attention U-Net, U2-Net, and Attention U2-Net. The distinctions among these networks are detailed in Table 1. Precision, Recall, IoU [28], F1 score, and Accuracy are used to evaluate the performance of the models. TN, TP, FP, and FN represent true negative, true positive, false positive, and false negative measurements, respectively. The calculation methods for each evaluation metric are shown in the Table 2.

3.5. Result

A qualitative comparison of the segmentation results of solar H-alpha grayscale images U-Net, Attention U-Net, U2-Net, and Attention U2-Net is found in Figure 9. The average values of evaluation metrics Precision, Recall, IoU, and F1 for the segmentation of solar filaments on the test set by the four neural networks, showing a quantitative comparison of the segmentation results are provided in Table 3. Figure 10 shows the Precision–Recall curve of four networks. Figure 11 and Figure 12 show the loss function curves for the validation sets over each iteration during the learning process of the four neural networks.

4. Discussion

Based on the comparison of evaluation metrics for U-Net, Attention U-Net, U2-Net, and Attention U2-Net on the test set, Attention U2-Net achieved an average Accuracy of 0.9987, an average Precision of 0.8221, an average Recall of 0.8469, an average IoU of 0.7139, and an average F1 score of 0.8323. Except for the Accuracy value, the other four metrics, which are used to evaluate the inference accuracy of the neural network, showed significant improvement compared to the other neural networks. The results demonstrate that all U-Net series networks achieve high-precision segmentation of different solar filaments. However, when it comes to multi-scale features, Attention U2-Net exhibits superior segmentation accuracy compared to U-Net, Attention U-Net, and U2-Net. This improvement is evident in both the complex contour details of larger filaments and the fine-grained features of smaller filaments. Moreover, under the same dataset and with random data augmentation, Attention U2-Net demonstrates a faster iteration speed. These advantages can likely be attributed to the combined effect of the residual U-Block and attention gate.

5. Conclusions

In this study, we utilize full-disk H-alpha solar images from the CHASE solar telescope for the automatic and high-precision recognition of solar filaments based on deep learning. The preprocessing steps include limb-darkening removal, grayscale transformation, K-means clustering, grain erosion, multiple closing operations, and hole filling, resulting in binarized label images of solar filaments. By incorporating the attention mechanism into U2-Net, we developed Attention U2-Net. Compared to other methods, Attention U2-Net significantly improves the recognition accuracy of full-disk H-alpha solar filaments over other U-Net variants. The high-precision automatic recognition of the detailed features of solar filaments is crucial for understanding the dynamics and evolution of solar filaments, their connection with solar active region magnetic fields, and predicting solar flares and coronal mass ejections.

Author Contributions

Conceptualization, W.J. and Z.L.; methodology, W.J.; data collection, W.J.; experiments, W.J.; writing—original draft preparation, W.J.; visualization, W.J.; writing—review and editing, Z.L.; funding acquisition, Z.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was financially supported by the National Key R&D Program of China (grant no. 2022YFC2807300).

Data Availability Statement

The data used in this research are publicly available and can be found at www.github.com/wdjiang-karl/solardataset (accessed on 10 September 2024).

Acknowledgments

The authors thank the reviewers very much for the valuable comments and suggestions that helped improve the manuscript. We acknowledge the use of data from CHASE mission supported by China National Space Administration.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Chen, P.F.; Xu, A.; Ding, M. Some interesting topics provoked by the solar filament research in the past decade. Res. Astron. Astrophys. 2020, 20, 166. [Google Scholar] [CrossRef]
  2. Gilbert, H.R.; Holzer, T.E.; Burkepile, J.T.; Hundhausen, A.J. Active and Eruptive Prominences and Their Relationship to Coronal Mass Ejections. Astrophys. J. 2000, 537, 503. [Google Scholar] [CrossRef]
  3. Gopalswamy, N.; Shimojo, M.; Lu, W.; Yashiro, S.; Shibasaki, K.; Howard, R.A. Prominence Eruptions and Coronal Mass Ejection: A Statistical Study Using Microwave Observations. Astrophys. J. 2003, 586, 562. [Google Scholar] [CrossRef]
  4. Chen, P.F.; Harra, L.K.; Fang, C. Imaging and Spectroscopic Observations of a Filament Channel and the Implication to the Nature of Counter-streamings. Astrophys. J. 2014, 784, 50. [Google Scholar] [CrossRef]
  5. Pulkkinen, T. Space Weather: Terrestrial Perspective. Living Rev. Sol. Phys. 2007, 4, 1. [Google Scholar] [CrossRef]
  6. Steinegger, M.; Hanslmeier, A.; Otruba, W.; Freislich, H.; Denker, C.; Goode, P.R.; Marquette, W.M.; Varied, J.; Wang, H.; Luo, G.; et al. An Overview of the New Global High-Resolution H-alpha Network. Hvar Obs. Bull. 2000, 24, 179. [Google Scholar]
  7. Wöger, F.; Rimmele, T.; Ferayorni, A. The Daniel K. Inouye Solar Telescope (DKIST)/Visible Broadband Imager (VBI). Solar Phys. 2021, 296, 1–25. [Google Scholar] [CrossRef]
  8. Noda, C.Q.; Schlichenmaier, R.; Rubio, L.B.; Löfdahl, M.G.; Khomenko, E.; Jurčák, J.; Leenaarts, J.; Kuckein, C.; Manrique, S.G.; Gunár, S.; et al. The European Solar Telescope. Astron. Astrophys. 2022, 666, A21. [Google Scholar] [CrossRef]
  9. Li, C.; Fang, C.; Li, Z. Chinese H-alpha Solar Explorer (CHASE)—A complementary space mission to the ASO-S. Res. Astron. Astrophys. 2019, 19, 165–170. [Google Scholar] [CrossRef]
  10. Li, C.; Fang, C.; Li, Z. The Chinese Ha Solar Explorer (CHASE) mission: An overview. Sci. China Phys. Mech. Astron. 2022, 65, 289602. [Google Scholar] [CrossRef]
  11. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25. [Google Scholar] [CrossRef]
  12. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. Int. Conf. Med. Image Comput. Comput.-Assist. Interv. 2015, 9351, 234–241. [Google Scholar]
  13. Iglovikov, V.; Mushinskiy, S.; Osin, V. Satellite Imagery Feature Detection using Deep Convolutional Neural Network: A Kaggle Competition. arXiv 2017, arXiv:1706.06169. [Google Scholar]
  14. Vojtekova, A.; Lieu, M. Learning to Denoise Astronomical Images with U-Nets. Mon. Not. R. Astron. Soc. 2020, 503, 3204–3215. [Google Scholar] [CrossRef]
  15. Asensio Ramos, A.; Cheung, M.C.M.; Chifu, I.; Gafeira, R. Machine learning in solar physics. Living Rev. Sol. Phys. 2023, 20, 4. [Google Scholar] [CrossRef]
  16. Zhu, G.; Lin, G.; Wang, D.; Liu, S.; Yang, X. Solar Filament Recognition Based on Deep Learning. Sol. Phys. 2019, 294, 117. [Google Scholar] [CrossRef]
  17. Liu, D.; Wei, S.; Lin, G.; Wang, H. Solar Filament Segmentation Based on Improved U-Nets. Sol. Phys. 2021, 296, 176. [Google Scholar] [CrossRef]
  18. Xin, Z.; Shang, Z. Solar Filament Recognition Based on Improved VNet. Astron. Res. Technol. 2022, 9, 54–64. [Google Scholar]
  19. Guo, X.; Yang, Y.; Feng, S.; Bai, X.; Liang, B.; Dai, W. Solar-Filament Detection and Classification Based on Deep Learning. Sol. Phys. 2007, 297, 104. [Google Scholar] [CrossRef]
  20. Zheng, Z.; Hao, Q.; Qiu, Y.; Hong, J.; Li, C.; Ding, M.D. Developing an Automated Detection, Tracking, and Analysis Method for Solar Filaments Observed by CHASE via Machine Learning. Astrophys. J. 2024, 965, 150. [Google Scholar] [CrossRef]
  21. MacQueen, J. Some methods for classification and analysis of multivariate observations. In Proceedings of the Fifth Berkeley Symposium on Mathematical Statistics and Probability, Oakland, CA, USA, 21 June–18 July 1965; University of California Press: Berkeley, CA, USA, 1967. Volume I—Theory of Statistics. p. 281. [Google Scholar]
  22. The Solar Science Data Center of Nanjing University. Available online: https://www.ssdc.nju.edu.cn/home (accessed on 1 August 2023).
  23. Oktay, O.; Schlemper, J.; Folgoc, L.L.; Lee, M.; Heinrich, M.; Misawa, K.; Mori, K.; McDonagh, S.; Hammerla, N.Y.; Kainz, B.; et al. Attention U-Net: Learning Where to Look for the Pancreas. In Proceedings of the 1st Conference on Medical Imaging with Deep Learning, Amsterdam, The Netherlands, 11 April 2018. [Google Scholar]
  24. Vaswani, A.; Shazeer, N.; Parmar, N. Attention is all you need. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  25. Qin, X.; Zhang, Z.; Huang, C.; Dehghan, M.; Zaiane, O.R.; Jagersand, M. U2-Net: Going Deeper with Nested U-Structure for Salient Object Detection. Pattern Recognit. 2020, 106, 107404. [Google Scholar] [CrossRef]
  26. Li, W.; Lambert-Garcia, R.; Getley, A.C.M.; Kim, K.; Bhagavath, S.; Majkut, M.; Leung, C.L.A. AM-SegNet for additive manufacturing in situ X-ray image segmentation and feature quantification. Virtual Phys. Prototyp. 2024, 19, e2325572. [Google Scholar] [CrossRef]
  27. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. In Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  28. Rezatofighi, H.; Tsoi, N.; Gwak, J.; Sadeghian, A.; Reid, I.; Savarese, S. Generalized Intersection over Union: A Metric and a Loss for Bounding Box Regression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019. [Google Scholar]
Figure 1. Step of data preprocessing diagram.
Figure 1. Step of data preprocessing diagram.
Universe 10 00381 g001
Figure 2. The results of the solar full-disk H-alpha images before and after preprocessing. (a) Grayscale image of the solar H-alpha band after limb-darkening correction. (b) Preliminary solar segmentation image obtained after grayscale inversion and stretching followed by K-means clustering analysis. (c) Label of solar filaments after particle erosion, multiple closing operations, and hole filling. (d) Mapping of segmented solar filament contours onto the original image.
Figure 2. The results of the solar full-disk H-alpha images before and after preprocessing. (a) Grayscale image of the solar H-alpha band after limb-darkening correction. (b) Preliminary solar segmentation image obtained after grayscale inversion and stretching followed by K-means clustering analysis. (c) Label of solar filaments after particle erosion, multiple closing operations, and hole filling. (d) Mapping of segmented solar filament contours onto the original image.
Universe 10 00381 g002
Figure 3. The architecture of U-Net.
Figure 3. The architecture of U-Net.
Universe 10 00381 g003
Figure 4. The architecture of Attention U-Net.
Figure 4. The architecture of Attention U-Net.
Universe 10 00381 g004
Figure 5. The architecture of residual U-block.
Figure 5. The architecture of residual U-block.
Universe 10 00381 g005
Figure 6. The architecture of Attention U-Net.
Figure 6. The architecture of Attention U-Net.
Universe 10 00381 g006
Figure 7. The architecture of attention gates.
Figure 7. The architecture of attention gates.
Universe 10 00381 g007
Figure 8. The architecture of Attention U2-Net.
Figure 8. The architecture of Attention U2-Net.
Universe 10 00381 g008
Figure 9. Qualitative comparison of the segmentation of each neural network on the test set: (a) image, (b) GT, (c) U-Net, (d) Attention U-Net, (e) U2-Net, (f) Attention U2-Net.
Figure 9. Qualitative comparison of the segmentation of each neural network on the test set: (a) image, (b) GT, (c) U-Net, (d) Attention U-Net, (e) U2-Net, (f) Attention U2-Net.
Universe 10 00381 g009
Figure 10. Precision–Recall curve comparison.
Figure 10. Precision–Recall curve comparison.
Universe 10 00381 g010
Figure 11. Validation loss per epoch of U-Net and Attention U-Net.
Figure 11. Validation loss per epoch of U-Net and Attention U-Net.
Universe 10 00381 g011
Figure 12. Validation loss per epoch of U2-Net and Attention U2-Net.
Figure 12. Validation loss per epoch of U2-Net and Attention U2-Net.
Universe 10 00381 g012
Table 1. Four neural networks.
Table 1. Four neural networks.
NetNon Attention GateWith Attention Gate
Classic U-NetU-NetAttention U-Net
RSU Nested U-NetU2-NetAttention U2-Net
Table 2. Five evaluation metrics.
Table 2. Five evaluation metrics.
Evaluation MetricsFormula
IoU T P T P + F N + F P
Precision T P T P + F P
Recall T P T P + F N
F1 2 P r e c i s i o n R e c a l l P r e c i s i o n + R e c a l l
Accuracy T P + T N T P + F P + F N + T N
Table 3. Average metrics of each neural network on the test set.
Table 3. Average metrics of each neural network on the test set.
NetPrecisionRecallIoUF1Accuracy
U-Net0.80410.81980.67900.80770.9985
Attention U-Net0.80720.83360.68830.81430.9985
U2-Net0.81370.82700.69240.81710.9986
Attention U2-Net (ours)0.82210.84690.71390.83230.9987
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Jiang, W.; Li, Z. Automated High-Precision Recognition of Solar Filaments Based on an Improved U2-Net. Universe 2024, 10, 381. https://doi.org/10.3390/universe10100381

AMA Style

Jiang W, Li Z. Automated High-Precision Recognition of Solar Filaments Based on an Improved U2-Net. Universe. 2024; 10(10):381. https://doi.org/10.3390/universe10100381

Chicago/Turabian Style

Jiang, Wendong, and Zhengyang Li. 2024. "Automated High-Precision Recognition of Solar Filaments Based on an Improved U2-Net" Universe 10, no. 10: 381. https://doi.org/10.3390/universe10100381

APA Style

Jiang, W., & Li, Z. (2024). Automated High-Precision Recognition of Solar Filaments Based on an Improved U2-Net. Universe, 10(10), 381. https://doi.org/10.3390/universe10100381

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop