Next Article in Journal
On Classification of the Human Emotions from Facial Thermal Images: A Case Study Based on Machine Learning
Previous Article in Journal
SGRiT: Non-Negative Matrix Factorization via Subspace Graph Regularization and Riemannian-Based Trust Region Algorithm
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automatic Detection of Equatorial Plasma Bubbles in Airglow Images Using Two-Dimensional Principal Component Analysis and Explainable Artificial Intelligence

1
Department of Space Environment, Institute of Basic and Applied Sciences, Egypt-Japan University of Science and Technology (E-JUST), New Borg El-Arab City 21934, Alexandria, Egypt
2
Department of Electronics and Communication Engineering, Egypt-Japan University of Science and Technology (E-JUST), P.O. Box 179, New Borg El-Arab City 21934, Alexandria, Egypt
3
Institute for Space-Earth Environmental Research (ISEE), Nagoya University, Nagoya 464-8601, Japan
4
Department of Physics, Faculty of Science, Helwan University, Helwan, Cairo 11795, Egypt
*
Author to whom correspondence should be addressed.
Mach. Learn. Knowl. Extr. 2025, 7(1), 26; https://doi.org/10.3390/make7010026
Submission received: 25 January 2025 / Revised: 2 March 2025 / Accepted: 10 March 2025 / Published: 16 March 2025
(This article belongs to the Section Learning)

Abstract

:
Equatorial plasma bubbles (EPBs) are regions of depleted electron density that form in the Earth’s ionosphere due to Rayleigh–Taylor instability. These bubbles can cause signal scintillation, leading to signal loss and errors in position calculations. EPBs can be detected in images captured by All-Sky Imager (ASI) systems. This study proposes a low-cost automatic detection method for EPBs in ASI data that can be used for both real-time detection and classification purposes. This method utilizes Two-Dimensional Principal Component Analysis (2DPCA) with Recursive Feature Elimination (RFE), in conjunction with a Random Forest machine learning model, to create an Explainable Artificial Intelligence (XAI) model capable of extracting image features to automatically detect EPBs with the lowest possible dimensionality. This led to having a small-sized and extremely fast-trained model that could be used to identify EPBs within the captured ASI images. A set of 2458 images, classified into two categories—Event and Empty—were used to build the database. This database was randomly split into two subsets: a training dataset (80%) and a testing dataset (20%). The produced XAI model demonstrated slightly higher detection accuracy compared to the standard 2DPCA model while being significantly smaller in size. Furthermore, the proposed model’s performance has been evaluated and compared with other deep learning baseline models (ResNet18, Inception-V3, VGG16, and VGG19) in the same environment.

1. Introduction

Equatorial plasma bubbles (EPBs) are common phenomena in the low-latitude nighttime ionospheric F region, characterized by sharp and sudden depletions in plasma densities [1]. The Rayleigh–Taylor instability is widely recognized as a key mechanism in the formation of plasma bubbles [2]. These bubbles cause significant scintillation of radio waves transmitted from satellites and the resulting turbulence can severely impact communication and navigation systems, such as the Global Navigation Satellite System (GNSS). This can lead to delays and the overall degradation of system performance [3,4,5,6]. Optical and satellite observations reveal that EPBs are magnetic-field-aligned regions of depleted ionospheric plasma density, significantly lower than the surrounding plasma [7,8,9,10].
As these irregularities ascend in altitude, they extend towards the poles along the magnetic field lines, appearing as depletions in optical intensity (OI 630.0 nm) in airglow images. These depletions are commonly referred to as equatorial plasma bubbles (EPBs) [11]. The OI 630.0 nm emission, which can be captured by CCD cameras, originates in the lower F region (250–300 km) due to the dissociative recombination of O 2 + with electrons according to the following mechanism:
O + + O 2 + e O 2 + + O
O 2 + + e O + O ( 1 D )
O ( 1 D ) O + h v ( 630.0   nm )
where O ( 1 D ) indicates an excited state [12,13].
Many studies have employed optical devices and CCDs to image equatorial plasma bubbles (EPBs) [10,14,15,16,17,18,19,20,21,22,23,24,25,26,27]. Gentile et al. [28] conducted a study using plasma density data from DMSP to identify EPBs within the surrounding plasma and determine EPB occurrence rates based on total orbits from 1989 to 2004. Huang and Roddy [29] utilized ion density data from C/NOFS, covering 2008 to 2014, to detect EPBs. Wan et al. [30] employed Swarm electron density data to detect and study EPB formation. Adkins and England [31] utilized 135.6 nm data from the Global-scale Observations of the Limb and Disk (GOLD) satellite to develop a method for automatically detecting and tracking EPBs. Lastly, Chakrabarti et al. [32] introduced a deep learning framework to accurately detect plasma bubble occurrences in ASI images.
Two-dimensional principal component analysis (2DPCA) has been utilized to reduce dimensionality while preserving essential structural information, thereby enhancing the detection and recognition features [33]. It has been effectively employed across various fields for detection and recognition tasks, leveraging its capability to process two-dimensional data matrices directly. Lin [34,35] and Lin et al. [36] utilized 2DPCA to identify total electron content (TEC) anomalies in the ionosphere preceding earthquakes in various regions. Yu and Liu [37] introduced a 2DPCA-based convolutional autoencoder designed for wafer map defect recognition. In computer vision, Sun et al. [38] proposed a randomized nonlinear 2DPCA network aimed at object recognition tasks by integrating 2DPCA with random Fourier mappings and activation functions. Also, it has been applied in bioinformatics with Rotation Forest to predict protein–protein interactions [39] and used for the prediction of lung cancer genetic mutations [40]. These diverse applications underscore the versatility and effectiveness of 2DPCA in handling two-dimensional data structures across multiple detection scenarios.
Explainable Artificial Intelligence (XAI) has multiple techniques, with two primary approaches: post hoc and ante hoc methods. Post hoc methods, such as SHAP [41] and LIME [42], provide explanations after a model has been trained. While they are versatile, their approximate nature can make them less trustworthy in certain scenarios. In contrast, ante hoc methods integrate explainability directly into the model’s design, offering inherent transparency but potentially limiting flexibility and model complexity. The Retzlaff et al. study [43] provides timely guidance in this regard, offering design guidelines that clarify the distinctions and practical applications of post hoc and ante hoc methods. It emphasizes that post hoc methods are best suited for black-box models requiring external interpretability, whereas ante hoc methods are superior in scenarios where interpretable decision making is a priority from the beginning [43]. These distinctions are crucial for bridging the gap between theory and practice in XAI research.
In this paper, we introduce a new method for automatically detecting EPB occurrences in ASI data using an ante hoc XAI model that utilizes Two-Dimensional Principal Component Analysis (2DPCA) with Recursive Feature Elimination (RFE) in conjunction with the Random Forest model. This method offers significantly lower computational costs and requires much simpler specifications. Additionally, it has a very short training time compared to deep learning models in the same environment, facilitating faster categorization of archived data and enabling real-time detection of the plasma bubbles in low-cost systems.

2. Methodology

2.1. Images Pre-Processing

Before building the dataset, the images went through a pre-processing stage. To enhance the visual information within the images to choose the proper images for each dataset, a histogram equalization [44] was performed according to the following process:
First, identify the histogram of the image using Equation (4). The image contains pixel intensities ranging from r = 0 , representing the darkest intensity, to r = L 1 , representing the brightest intensity, and L is the total number of the possible intensity levels:
h r k = n k n
where k is referring to the intensity level, n k   is the number of pixels with intensity r k , and n is the total number of the image’s pixels.
Then, compute the Cumulative Distribution Function (CDF) through the following equation:
c r k = j = 0 k h r j
Finally, use the transformation function to obtain the new pixel intensities through Equation (6):
s k = c r k × ( L 1 )
where s k is the new pixel intensity level and L is the total number of the possible intensity levels.
Figure 1a shows a sample raw image captured at the Bom Jesus da Lapa (BJL) observatory (Geo 13.3 °S, 43.5 °W) on 8 October 2021 at 00:02:25. The histogram corresponding to the image is shown in Figure 1b, where most of the pixel intensity levels are concentrated on the darker side of the histogram. In contrast, Figure 1c displays the same image after enhancement, revealing its features using the mentioned method. The histogram corresponding to the enhanced image is shown in Figure 1d, where the pixel intensities are distributed across the full possible range, enhancing visualization.

2.2. Dataset Preparation

The ASI dataset used in this study was collected from the BJL (Geo 13.3°S, 43.5°W) and São João do Cariri observatory (CA) (Geo 7.4°S, 36.6°W). These stations are part of the Embrace/INPE network. The data collection period spans from January 2020 to January 2022. The dataset considered seasonal and local time variations. The set comprises 2458 ASI raw images, manually classified based on their corresponding pre-enhanced images. The dataset was divided into two categories. The first category is the ‘Event’ Class, which includes images containing visible plasma bubbles characterized by distinct depletions in brightness and elongated structures along Earth’s magnetic field lines, originating from the magnetic equator and extending toward the poles, as shown in Figure 2. The second category is the ‘Empty’ Class, which consists of images without any apparent plasma bubble structures. The first category contains 1242 images while the second contains 1216 images. The dataset was then divided randomly into two subsets: training set and test set, following the split of 80% training set (1967 images) and 20% test set (491 images). It is crucial to ensure enough samples for training while avoiding too few samples for testing.

2.3. Radon Transform

To facilitate the detection process, the chosen datasets have been transformed using the Radon transform. The Radon transform is a mathematical technique that involves transforming a two-dimensional function (image) f ( x , y ) into a set of its projections, where x = 0 and y = 0 lay at the center of the image. The resulting projection is the sum of the pixel intensities in each direction, essentially forming an integral line taken along various directions. The result is a new image R ρ , θ [45].
ρ = x c o s   θ + y s i n   θ
where ρ is the distance of the line from the origin and θ is the angle between the perpendicular to the line and the x-axis. The transformation process is depicted in Figure 3.
Radon transform is defined mathematically using the following equations:
R ρ , θ = f x , y δ ρ x c o s   θ y s i n   θ d x d y
where δ is the Dirac delta function.
Figure 4 illustrates examples of original images f x , y of size 512 × 512 pixels and their corresponding Radon-transformed images R ρ , θ of size 729 × 180 pixels for samples from each class. Figure 4a shows the original image of the “Empty” class, while Figure 4b presents its Radon-transformed counterpart. Similarly, Figure 4c displays the original image of the “Event” class and Figure 4d depicts its transformed version.

2.4. Two-Dimensional Principal Component Analysis

The main concept of Two-Dimensional Principal Component Analysis (2DPCA) is to obtain an m-dimensional projected vector that contains the image features through the following linear transformation:
Y = A X
where A is an image of the dimensions m × n , X is a unitary column vector, and Y is the projected feature vector of image A [33]. In the case of the raw airglow images (Figure 4c) at BJL, the size of A is 512 × 512. Meanwhile, for the Radon-transformed images (Figure 4d), the size of A is 729 × 180.
It starts with computing a non-negative n × n definite matrix called the image covariance matrix G t using the training dataset:
G t = 1 M j = 1 M A j A ¯ T A j A ¯
where M is the number of the training images samples, A j is the j th image of the training dataset, and A ¯ is the average image of all training set images.
To determine a good projection vector X , we measure its discriminatory power using the total scatter of the projected samples. The trace of the covariance matrix of the projected feature vectors characterizes this total scatter. A higher trace indicates a better projection vector as it implies greater variance and better discrimination between projected samples.
The optimal projection axis X o p t is the unitary vector that maximizes the generalized total scatter of the projected samples, which corresponds to the eigenvector of the covariance matrix G t with the largest eigenvalue [46]. However, typically, more than one optimal projection axis is needed. A set of projection axes X 1 , , X d is selected. These optimal projection axes are the orthonormal eigenvectors of G t corresponding to the largest d eigenvalues. This process ensures that the most important image feature information is contained within the smallest amount of data. Figure 5 shows the graphical representation of the 2DPCA implementation.
In this study, d was set to a dynamic range to determine the best accuracy based on models trained with different numbers of principal components. However, the last ten eigenvectors accounted for most of the image features, as illustrated in Figure 6.
The feature extraction of image A is conducted using the optimal projection vectors X 1 , , X d :
Y k = A X k , k = 1,2 , , d .
where Y k refers to the k th vector of the projected feature vectors Y 1 , , Y d , which are called the principal component vectors of Image A .
The feature matrices of the training dataset images are saved as the database for classification purposes. A 2DPCA transformation is performed on each image of the testing dataset. Classification is then conducted by finding the nearest neighbor classifier through calculating the distance between the testing image feature matrix and every matrix in the database:
d ( B i , B j ) = k = 1 d Y k ( i ) Y k ( j ) 2
where B i , B j are two arbitrary feature matrices testing and training images and Y k ( i ) Y k ( j ) 2 is the Euclidean distance between two principal component vectors Y k ( i ) and Y k ( j ) .
Assume that the training samples are B 1 , B 2 , , B M where M is the total number of training samples. Each sample is assigned a specific class ω k . For the test sample B , if d   B , B t = min B , B j and B t ω k , then the result will be B ω k . In this study, we used a collective decision based on the first three min B , B j , as it helped to enhance classification accuracy.

2.5. Explainable AI (XAI) Model

The output 2DPCA matrix of each image in the dataset was flattened to be introduced into a vector. This vector represented the features derived from the 2DPCA and a last column was added to represent the class. The class column was labeled either 0 or 1, where the first refers to the Empty class and the latter refers to the Event class. These vectors were combined into a single matrix where each row corresponded to an image and the last column denoted the associated class label. This matrix is used as an input to a machine learning model that performs Recursive Feature Elimination (RFE) in conjunction with Random Forest to determine the contribution of each principal component in the classification process.

2.5.1. Random Forest

Random Forest [48] is a widely used ensemble learning method in machine learning, particularly effective for classification and regression tasks. It constructs a collection of tree-structured classifiers, where each tree is grown based on a random vector Θ k that is sampled independently and with the same distribution for all trees in the forest. The model aggregates predictions from all trees to enhance accuracy and robustness. For classification, Random Forests determine the predicted class through majority voting and, for regression, they compute the average of the tree outputs. The principles of Random Forests are based on determining the margin function, the strength of individual classifiers, and the correlation between trees, as well as their impact on generalization error. These foundational principles enable Random Forests to achieve high accuracy, avoid overfitting, and provide estimates of variable importance, making them a robust and versatile machine learning method.
The margin function, which quantifies the confidence of the classification, is defined as:
m r X , Y = P Θ h X , Θ = Y m a x j Y P Θ ( h X , Θ = j )
where h   X , Θ is the prediction of a tree, P Θ is the probability measure over the random vectors Θ , X is the input, and Y is the true class label. A larger margin indicates higher confidence in the classification decision.
The strength of the forest is the expected value of the margin function:
s = E X , Y [ m r X , Y ]
where s measures the average accuracy of the individual trees in the forest.
The generalization error, which represents the probability of misclassification, is shown to have the upper bound:
P E * ρ ¯ ( 1 s 2 ) s 2
where ρ ¯ is the average correlation between the classifiers in the forest. This equation highlights the balance between the individual trees’ strength and their correlation, underscoring that the effectiveness of Random Forests relies on achieving high tree strength while maintaining low correlation between them.

2.5.2. Recursive Feature Elimination (RFE)

RFE is a feature selection method, particularly suitable for small sample classification problems [49]. It was initially applied to microarray-based cancer classification, where training samples are fewer than 100, while the number of features can reach several thousand, proving to be a powerful technique for feature selection. RFE ranks the features by iteratively fitting a model (logistic regression in our case) and removing the least important features at each step until the desired number of features is reached. Then, it gives the ranking for all original features.

2.5.3. Algorithm Breakdown

The dataset was divided into training and testing subsets. The process begins by loading the training dataset and applying 2DPCA to reduce the high-dimensional image data into a set of principal components, retaining the most significant features in a compact form. The accuracy of 2DPCA is evaluated by testing a dynamic range of principal components with the highest eigenvalues to identify the group responsible for the highest accuracy. The trained data corresponding to the principal components with the highest accuracy are then flattened and fed into RFE in conjunction with Random Forests. The Random Forest classifier is used as the base estimator for RFE. RFE iteratively eliminates the least important features based on Random Forest’s feature importance metrics, progressively refining the feature set to the desired number of principal components. For each iteration, feature importance is aggregated across the components and their overall contribution to the model’s decision making is calculated. The final output includes ranked feature importance scores and visualizations showing feature contributions and their percentage impact on the classification. These steps ensure the model not only achieves high accuracy but also provides clear insights into which principal components drive its predictions, aligning with the principles of XAI. Finally, the top three contributing components are selected to create the final trained model, achieving the highest accuracy with the smallest possible size. The algorithm flowchart is shown in Figure 7.

2.6. Evaluation Parameters

The evaluation parameters for the results of the predictions of the models are defined as follows:
  • Accuracy measures the overall performance of the model by computing the ratio of correct predictions to the total number of tested samples:
    Accuracy = True Positive + True Negative Total number of predictions
    where True Positive is the number of images in which the model correctly detected an EPB when it was actually present. True Negative is the number of images in which the model correctly classified an image as not containing an EPB when it was absent;
  • Precision evaluates the accuracy of positive predictions by calculating the proportion of correctly detected EPBs ( True Positives ) out of all assumed detected EPBs:
    Precision = True Positive True Positive + False Positive
    where False Positive is the number of images in which the model incorrectly detected an EPB when none were present;
  • Recall assesses the model’s ability to identify all positive instances by calculating the ratio of True Positives to the sum of True Positives and False Negatives :
    Recall = True Positive True Positive + False Positive
    where False Negatives is the number of images in which the model failed to detect an EPB when one was actually present (a missed detection);
  • F1-score provides a balanced measure of the model’s performance by combining both precision and recall into a single metric:
    F 1   score = 2 × Precision × Sensitvity Precision + Sensitvity

3. Results and Discussion

The standard proposed 2DPCA model without applying Radon transform was trained multiple times using the training dataset, each time using a different number of the highest principal components, which corresponded to varying cuts of the total eigenvalues within the dataset. This cut was set to be equal to 95% or higher. The model was then tested using the testing dataset, which was selected randomly and kept consistent across all tests. The testing results showed unexpected fluctuations in detection accuracy, unlike the expectations, as increasing the percentage of eigenvalues was assumed to correlate with a higher number of trained features (according to Yang and Yang) [46]. This may be due to the dataset’s dependency as the dataset consists of airglow images taken under different atmospheric and geomagnetic conditions. Variations in cloud cover, background noise, and observational conditions may introduce inconsistencies that affect the model’s performance across different test sets.
Table 1 shows the number of the selected eigenvectors with the highest eigenvalues, their percentages to the total of eigenvalues, the detection accuracy, and the training time consumed using the training dataset, which contains 1967 images.
The 2DPCA model was then trained again using the same criteria after applying Radon transform. Re-processing the images using Radon transform helped to change the image resolution from 512 × 512 pixels to 729 × 180, which resulted in getting a smaller covariance matrix with only 180 eigenvectors in total instead of 512. This resulted in getting a smaller number of eigenvectors that correspond to the same cut of eigenvalues in the previous test, hence noticeably reducing the training time and the model size. However, applying Radon transform to raw images caused a decrease in the detection accuracy. Table 2 shows the percentage of the contributed features, the training time, and the accuracy for different selected numbers of the highest eigenvectors.
Figure 8 provides a visual comparison of the results from Table 1 and Table 2. The x-axis represents the cut-off percentages of the eigenvalues of principal components that were used to train the model to the total eigenvalue sum while the y-axis indicates the detection accuracy corresponding to each percentage. Additionally, the number of principle components used to train the 2DPCA model is written above each point. The plot indicates that all trained models using raw images achieve better accuracy than those using Radon-transformed images. Furthermore, the models yielding the highest accuracy for each training type were as follows: for raw images, the model utilizing the top thirty principal components achieved a detection accuracy of 97.96% while, for Radon-transformed images, the model using the top ten principal components achieved a detection accuracy of 96.1%. These models are marked with green circles.
RFE, in conjunction with Random Forest, was then applied to the models responsible for the highest detection accuracy at each case to determine the contribution of each principal component in the image classification process. Figure 9a shows the eigenvalues of the principal components. Meanwhile, Figure 9b shows the percentage of contribution of each principal component in determining the class of the testing images based on RFE and Random Forest. The comparison between both panels reveals that principal components with the highest eigenvalues do not necessarily contribute the most to the classification process. For instance, the 30th principal component, despite having the smallest eigenvalue, was the highest contributor to the classification process. Similarly, Figure 9c,d illustrate the eigenvalues of the principal components for the model utilizing Radon-transformed images and their corresponding percentage contributions to the classification process. Although the tenth and ninth eigenvectors have the smallest eigenvalues, their principal components ranked as the first and third highest contributors in the classification process, respectively.
The XAI model ranks the principal components based on their contributions. It then utilizes 2DPCA, progressively incorporating the principal components into the training process according to their new ranking, and then applies it to the testing subset. Next, it stores the number of principal components used and the detection accuracy of the model. Finally, it selects the model that achieves the highest detection accuracy while maintaining the smallest possible size. This process is illustrated in Figure 10. The new models were based on the top thirteen and top three principal component contributors for the raw and Radon-transformed models, respectively. These models achieved an increase in detection accuracy of 0.21% for the raw image-based model, reaching 98.17%, and an increase of 1.22% for the Radon-transformed image-based model, reaching 97.35%. Furthermore, they had significantly smaller sizes compared to the original models, with size reductions of 56.6% and 70% for the raw and Radon-transformed models, respectively.
The proposed model was compared with other deep learning baseline models, including ResNet18, Inception-V3, VGG16, and VGG19 [50,51], to evaluate its performance. These CNN models were trained using the aforementioned split database, without applying any transfer learning, while preserving their original architecture. To ensure a fair comparison, all models were trained in a similar environment. Moreover, the same hyperparameters were applied across all models, with a batch size of 16. The epoch limit was set at 10 to reduce training time with the possibility of retrieving the epoch with the best validation accuracy and lowest training loss. The original training dataset was split into training and validation subsets, with 80% allocated for training and 20% for validation. All datasets were kept consistent across all models. The performance of each model was assessed using the evaluation parameters mentioned in Section 2. The results of these evaluation parameters for all models are presented in Table 3.
The results from Table 3 are visualized in Figure 11. Figure 11a highlights the differences among the proposed models, illustrating the relationship between detection accuracy and elapsed training time. Figure 11b presents the same relationship, with the addition of deep learning CNN models for comparison. The figure shows that the best detection accuracy was achieved by the XAI model based on the raw data, achieving 98.17% and overpassing all other proposed models and CNN deep learning models, while the lowest accuracy was obtained by VGG16. However, the XAI model using Radon-transformed images demonstrated the best performance in terms of a balance between accuracy and elapsed time. While it achieved the second-best detection accuracy with only a 0.82% difference from the top model, it required significantly less time, taking just 35% of the time consumed by the first model to achieve the highest accuracy. Furthermore, the proposed method of using 2DPCA required significantly less training time in the same environment compared to the deep learning baseline models. While the incorporation of XAI slightly increased the training time compared to 2DPCA, it contributed to an improvement in detection accuracy and a reduction in the trained model size. Using the proposed XAI model with a small elapsed time enables the creation of compact and fast-trained detection models that consume low computational power while providing high accuracy. Such models are suitable for integration with real-time detection techniques and for deployment in low-cost systems. However, the model may be sensitive to partially cloudy sky conditions, where a single image contains both plasma bubbles and clouds. This can lead to misclassification or failure to recognize actual plasma bubbles due to obstructions in real-time detection. Despite this challenge, equatorial plasma bubbles remain in the field of view longer than clouds. This allows for a more reliable approach based on collective decisions from multiple ASI images, reducing false detections and improving accuracy.

4. Conclusions

In this study, a low-computational-cost XAI model was developed for the automatic detection of equatorial plasma bubbles in ASI data. This model was based on using 2DPCA with RFE, in conjunction with a Random Forest machine learning model. A dataset of 2458 ASI images was collected and classified manually into two classes: “Empty” class when the image does not contain a plasma bubble and “Event” class when a plasma bubble appears in the image. This dataset was split into two subsets: the training subset (80%) and the testing subset (20%). These subsets remained constant among all the classification tests. The first training instance was conducted using only 2DPCA applied to the raw images, achieving a detection accuracy of 97.96%. Applying Radon transform in the further testing helped in decreasing the computational time and the size of the trained model by almost 65% but reduced the detection accuracy to 96.1%.
The XAI model was then utilized by using the eigenvectors from the models with the highest detection accuracy to determine the actual contribution of each principal component in the classification process. These principal components were then ranked based on the contribution percentage using RFE in conjunction with Random Forest. Based on their updated ranking, the principal components were gradually incorporated into the 2DPCA training process to develop a model that maximizes detection accuracy while minimizing its size. Although the implementation of the XAI model slightly increased the processing time compared to 2DPCA, it improved detection accuracy for the raw image-based model by 0.21%, achieving 98.17%, and caused an increase of 1.22% for the Radon-transformed image-based model, reaching 97.35%. It reduced the size of the trained detection models compared to the standard 2DPCA, with a reduction of 56.6% for the raw image-based model and 70% for the Radon-transformed model. Additionally, the comparison with the deep learning CNN models showed the superiority of the proposed models in the detection accuracy within the raw ASI images and for the training time. The highest detection accuracy among the CNN baseline models was achieved by ResNet18 at 91.45%, placing it 4.68% to 6.72% behind the proposed models. Additionally, ResNet18 was the fastest deep learning model, requiring 38.72 min to train over 10 epochs, whereas the proposed models required only 2.18 to 8.71 min to be trained in the same environment. The proposed models are well-suited for integration with real-time detection techniques and deployment in low-cost systems. Although the model may be sensitive to partially cloudy conditions, which can lead to potential misclassification, EPBs remain in the field of view of the ASI for a long time. This can enhance accuracy through the collective decision making of multiple ASI images, reducing false detections and improving overall performance.

Author Contributions

Conceptualization, M.Y. and A.M.; Methodology, M.Y., M.A. and A.M.; Software, M.Y. and M.A.; Validation, M.Y. and K.S.; Formal analysis, M.Y. and K.S.; Investigation, M.Y.; Resources, M.A.; Data curation, M.Y.; Writing—original draft, M.Y.; Writing—review and editing, M.Y., M.A., K.S. and A.M.; Supervision, K.S. and A.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

ASI data are available through the Embrace/INPE site (https://embracedata.inpe.br/imager/) accessed on 3 October 2024.

Acknowledgments

This work was supported by the SCOSTEP Visiting Scholar (SVS) program of the Scientific Committee on Solar-Terrestrial Physics (SCOSTEP), for which the first author stayed for three months at the Institute for Space-Earth Environmental Research, Nagoya University. Furthermore, the authors thank the Embrace/INPE Space Weather Program for their dedicated efforts to provide substantial data to study the ionosphere around the magnetic equator.

Conflicts of Interest

The authors of this research paper declare that they have no financial or personal relationships with any individuals or organizations that could inappropriately influence or bias their work. Additionally, it is hereby stated that there are no interests to declare.

Abbreviations

The following abbreviations are used in this manuscript:
2DPCATwo-Dimensional Principal Component Analysis
ASIAll-Sky Imager
BJLBom Jesus da Lapa Observatory
CASão João do Cariri Observatory
CCDCharged Coupled Device
CDFCumulative Distribution Function
CNNConvolutional Neural Network
EPBsEquatorial Plasma Bubbles
RFRandom Forest
RFERecursive Feature Elimination

References

  1. Hysell, D.L.; Woodman, R.F. Imaging coherent backscatter radar observations of topside equatorial spread F. Radio Sci. 1997, 32, 2309–2320. [Google Scholar] [CrossRef]
  2. Kelley, M.C. The Earth’s Ionosphere; Academic Press: San Diego, CA, USA, 1989. [Google Scholar]
  3. Basu, S.; Kudeki, E.; Basu, S.; Valladares, C.E.; Weber, E.J.; Zengingonul, H.P.; Espinoza, J. Scintillations, plasma drift, and neutral winds in the equatorial ionosphere after sunset. J. Geophys. Res. 1996, 101, 26795–26809. [Google Scholar] [CrossRef]
  4. Kintner, P.M.; Kil, H.; Ledvina, B.M.; Paula, E.D. Fading timescales associated with GPS signals and potential consequences. Radio Sci. 2001, 36, 731–743. [Google Scholar] [CrossRef]
  5. Straus, P.R.; Anderson, P.C.; Danaher, J.E. GPS occultation sensor observations of ionospheric scintillation. Geophys. Res. Lett. 2003, 30, 1436. [Google Scholar] [CrossRef]
  6. Park, J.; Noja, M.; Stolle, C.; Luhr, H. The ionospheric bubble index deduced from magnetic field and plasma observations onboard swarm. Earth Planets Space 2013, 65, 1333–1344. [Google Scholar] [CrossRef]
  7. Weber, E.J.; Brinton, H.C.; Buchau, J.; Moore, J.G. Coordinated airborne and satellite measurements of equatorial plasma depletions. J. Geophys. Res. 1982, 87, 10503–10513. [Google Scholar] [CrossRef]
  8. Tinsley, B.A.; Rohrbaugh, R.P.; Hanson, W.B.; Broadfoot, A.L. Images of transequatorial F region bubbles in 630- and 777-nm emissions compared with satellite measurements. J. Geophys. Res. 1997, 102, 2057–2077. [Google Scholar] [CrossRef]
  9. Su, S.-Y.; Yeh, H.C.; Heelis, R.A. ROCSAT-1 ionospheric plasma and electrodynamics instrument observations of equatorial spread F: An early transitional scale result. J. Geophys. Res. 2001, 106, 29153–29159. [Google Scholar] [CrossRef]
  10. Kelley, M.C.; Makela, J.J.; Ledvina, B.M.; Kintner, P.M. Observations of equatorial spread-F from Haleakala, Hawaii. Geophys. Res. Lett. 2002, 29, 2003. [Google Scholar] [CrossRef]
  11. Makela, J.J.; Miller, E.S. Optical observations of the growth and day-to-day variability of equatorial plasma bubbles. J. Geophys. Res. Space Phys. 2008, 113, 1–7. [Google Scholar] [CrossRef]
  12. Sobral, J.H.; Takahashi, H.; Abdu, M.A.; Muralikrishna, P.; Sahai, Y.; Zamlutti, C.J.; De Paura, R.E.; Batista, P.P. Determination of the quenching rate of the O(1D) by O(3P) from rocket-borne optical (630 nm) and electron density data. J. Geophys. Res. 1993, 98, 7791–7798. [Google Scholar] [CrossRef]
  13. Pimenta, A.A.; Fagundes, P.R.; Bittencourt, J.A.; Sahai, Y.; Gobbi, D.; Medeiros, A.F.; Taylor, M.J.; Takahashi, H. Ionospheric plasma bubble zonal drift: A methodology using OI 630 nm all-sky imaging systems. Adv. Space Res. 2001, 27, 1219–1224. [Google Scholar] [CrossRef]
  14. Weber, E.J.; Buchau, J.; Eather, R.H.; Mende, S.B. North-south aligned equatorial depletions. J. Geophys. Res. 1978, 83, 712–716. [Google Scholar] [CrossRef]
  15. Mendillo, M.; Baumgardner, J.; Sultan, P.J. Airglow characteristics of equatorial plasma depletions. J. Geophys. Res. 1982, 87, 7641–7652. [Google Scholar] [CrossRef]
  16. Taylor, M.J.; Eccles, J.V.; Labelle, J.; Sobral, J.H.A. High-resolution OI (630 nm) image measurements of F-region depletion drifts during the Guara campaign. Geophys. Res. Lett. 1997, 24, 1699–1702. [Google Scholar] [CrossRef]
  17. Fagundes, P.R.; Sahai, Y.; Batista, I.S.; Abdu, M.A.; Bittencourt, J.A.; Takahashi, H. Observations of day-to-day variability in precursor signatures to equatorial F-region plasma depletions. Ann. Geophys. 1999, 17, 1053–1063. [Google Scholar] [CrossRef]
  18. Makela, J.J.; Ledvina, B.M.; Kelley, M.C.; Kintner, P.M. Analysis of the seasonal variations of equatorial plasma bubble occurrence observed from Haleakala. Hawaii. Ann. Geophys. 2004, 22, 3109–3121. [Google Scholar] [CrossRef]
  19. Mendillo, M.; Zesta, E.; Shodhan, S.; Doe, R.; Sahai, Y.; Baumgardner, J. Observations and modeling of the coupled latitude-altitude patterns of equatorial plasma depletions. J. Geophys. Res. 2005, 110, A09303. [Google Scholar] [CrossRef]
  20. Rajesh, P.K.; Liu, J.Y.; Sinha, H.S.S.; Banerjee, S.B.; Misra, R.N.; Dutt, N.; Dadhania, M.B. Observations of plasma depletions in 557.7-nm images over Kavalur. J. Geophys. Res. Space Phys. 2007, 112, A07307. [Google Scholar] [CrossRef]
  21. Liu, J.Y.; Tsai, H.F.; Jung, T.K. Airglow observations over the equatorial ionization anomaly zone in Taiwan. Ann. Geophys. 2011, 29, 749–757. [Google Scholar] [CrossRef]
  22. Shiokawa, K.; Otsuka, Y.; Lynn, K.J.; Wilkinson, P.; Tsugawa, T. Airglow-imaging observation of plasma bubble disappearance at geomagnetically conjugate points. Earth Planet Space 2015, 67, 43. [Google Scholar] [CrossRef]
  23. Fukushima, D.; Shiokawa, K.; Otsuka, Y.; Nishioka, M.; Kubota, M.; Tsugawa, T.; Nagatsuma, T.; Komonjinda, S.; Yatini, C.Y. Geomagnetically conjugate observation of plasma bubbles and thermospheric neutral winds at low latitudes. J. Geophys. Res. Space Phys. 2015, 120, 2222–2231. [Google Scholar] [CrossRef]
  24. Okoh, D.; Rabiu, B.; Shiokawa, K.; Otsuka, Y.; Segun, B.; Falayi, E.; Onwuneme, S.; Kaka, R. First study on the occurrence frequency of equatorial plasma bubbles over West Africa using an all-sky airglow imager and GNSS receivers. J. Geophys. Res. Space Phys. 2017, 122, 12430–12444. [Google Scholar] [CrossRef]
  25. Ghodpage, R.N.; Gurav, O.B.; Taori, A.; Sau, S.; Patil, P.T.; Erram, V.C.; Sripathi, S. Observation of an intriguing equatorial plasma bubble event over Indian sector. J. Geophys. Res. Space Phys. 2021, 126, e2020JA028308. [Google Scholar] [CrossRef]
  26. Wrasse, C.M.; Figueiredo, C.A.O.B.; Barros, D.; Takahashi, H.; Carrasco, A.J.; Vital, L.F.R.; Rezende, L.C.A.; Egito, F.; Rosa, G.M.; Sampaio, A.H.R. Interaction between equatorial plasma bubbles and a medium-scale traveling ionospheric disturbance, observed by OI 630 nm airglow imaging at Bom Jesus de Lapa, Brazil. Earth Planet. Phys. 2021, 5, 397–406. [Google Scholar] [CrossRef]
  27. Liu, J.-Y.; Rajesh, P.K.; Liao, Y.-A.; Chum, J.; Kan, K.-W.; Lee, I.-T. Airglow observed by a full-band imager together with multi-instruments in Taiwan during nighttime of 1 November 2021. Adv. Space Res. 2024, 73, 663–671. [Google Scholar] [CrossRef]
  28. Gentile, L.C.; Burke, W.J.; Rich, F.J. A global climatology for equatorial plasma bubbles in the topside ionosphere. Ann. Geophys. 2006, 24, 163–172. [Google Scholar] [CrossRef]
  29. Huang, C.-S.; Roddy, P.A. Effects of solar and geomagnetic activities on the zonal drift of equatorial plasma bubbles. J. Geophys. Res. Space Phys. 2016, 121, 628–637. [Google Scholar] [CrossRef]
  30. Wan, X.; Xiong, C.; Rodriguez-Zuluaga, J.; Kervalishvili, G.N.; Stolle, C.; Wang, H. Climatology of the occurrence rate and amplitudes of local time distinguished equatorial plasma depletions observed by Swarm satellite. J. Geophys. Res. Space Phys. 2018, 123, 3014–3026. [Google Scholar] [CrossRef]
  31. Adkins, V.J.; England, S.L. Automated detection and tracking of equatorial plasma bubbles utilizing Global-Scale Observations of the Limb and Disk (GOLD) 135.6 nm Data. Earth Space Sci. 2023, 10, e2023EA002935. [Google Scholar] [CrossRef]
  32. Chakrabarti, S.; Patgiri, D.; Rathi, R.; Dixit, G.; Krishna, M.V.S.; Sarkhel, S. Optimizing a deep learning framework for accurate detection of the Earth’s ionospheric plasma structures from all-sky airglow images. Adv. Space Res. 2024, 73, 5990–6005. [Google Scholar] [CrossRef]
  33. Yang, J.; Zhang, D.; Frangi, A.F.; Yang, J. Two-dimensional PCA: A new approach to appearance-based face representation and recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2004, 26, 131–137. [Google Scholar] [CrossRef]
  34. Lin, J.-W. Ionospheric anomaly related to M=6.6, 26 August 2012, Tobelo earthquake near Indonesia: Two-dimensional principal component analysis. Acta Geod. Geophys. 2013, 48, 247–264. [Google Scholar] [CrossRef]
  35. Lin, J.-W. Generalized two-dimensional principal component analysis and two artificial neural network models to detect traveling ionospheric disturbances. Nat. Hazards 2022, 111, 1245–1270. [Google Scholar] [CrossRef]
  36. Lin, J.-W.; Chiou, J.-S.; Chao, C.-T. Detecting all possible ionospheric precursors by kernel-based two-dimensional principal component analysis. IEEE Access 2019, 7, 53650–53666. [Google Scholar] [CrossRef]
  37. Yu, J.; Liu, J. Two-dimensional principal component analysis-based convolutional autoencoder for wafer map defect detection. IEEE Trans. Ind. Electron. 2021, 68, 8789–8797. [Google Scholar] [CrossRef]
  38. Sun, Z.; Shao, Z.; Shang, Y.; Li, B.; Wu, J.; Bi, H. Randomized nonlinear two-dimensional principal component analysis network for object recognition. Mach. Vis. Appl. 2023, 34, 21. [Google Scholar] [CrossRef]
  39. Wang, L.; You, Z.H.; Yan, X. Using two-dimensional principal component analysis and rotation forest for prediction of protein-protein interactions. Sci. Rep. 2018, 8, 12874. [Google Scholar] [CrossRef]
  40. Abdelwahab, M.M.; Abdelrahman, S.A. Four layers image representation for prediction of lung cancer genetic mutations based on 2DPCA. In Proceedings of the 2017 IEEE 60th International Midwest Symposium on Circuits and Systems (MWSCAS), Boston, MA, USA, 6–9 August 2017; 2017; pp. 599–602. [Google Scholar] [CrossRef]
  41. Lundberg, S.M.; Lee, S.-I. A Unified Approach to Interpreting Model Predictions. Adv. Neural Inf. Process. Syst. (NeurIPS) 2017, 30. [Google Scholar] [CrossRef]
  42. Ribeiro, M.T.; Singh, S.; Guestrin, C. “Why should I trust you?” Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 1135–1144. [Google Scholar]
  43. Retzlaff, C.O.; Angerschmid, A.; Saranti, A.; Schneeberger, D.; Müller, D.; Holzinger, A. Post-hoc vs ante-hoc explanations: xAI design guidelines for data scientists. Cogn. Syst. Res. 2024, 86, 101243. [Google Scholar] [CrossRef]
  44. Gonzalez, R.C.; Wintz, P. Digital Image Processing, 3rd ed.Addison-Wesley: Reading, MA, USA, 1997. [Google Scholar]
  45. Carsten, H. The Radon Transform; Aalborg University: Aalborg, Denmark, 2007; VGIS, 07gr721. [Google Scholar]
  46. Yang, J.; Yang, J.Y. From image vector to matrix: A straightforward image projection technique—IMPCA vs. PCA. Pattern Recognit. 2002, 35, 1997–1999. [Google Scholar] [CrossRef]
  47. Yin, X.; Chen, L. A Cross-Modal Image and Text Retrieval Method Based on Efficient Feature Extraction and Interactive Learning CAE. Sci. Program. 2022, 1, 7314599. [Google Scholar] [CrossRef]
  48. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  49. Guyon, I.; Weston, J.; Barnhill, S.; Vapnik, V. Gene selection for cancer classification using support vector machines. Mach. Learn. 2002, 46, 389–422. [Google Scholar] [CrossRef]
  50. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
  51. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
Figure 1. Raw image from the BJL station (a) with its corresponding histogram (b) while (c) shows the same image after enhancement and (d) displays the enhanced histogram.
Figure 1. Raw image from the BJL station (a) with its corresponding histogram (b) while (c) shows the same image after enhancement and (d) displays the enhanced histogram.
Make 07 00026 g001
Figure 2. Visual examples of the detected EPBs.
Figure 2. Visual examples of the detected EPBs.
Make 07 00026 g002
Figure 3. The summation line is given by the equation y = a x + b . Using trigonometry, the inclination a is found to be a = c o s   θ s i n   θ and the intersection with the y axis is b = ρ s i n   θ . This aligns with Equation (4). These parameters are calculated for each combination of θ and ρ . The maximum ρ is set to the length of the image’s diagonal line [45].
Figure 3. The summation line is given by the equation y = a x + b . Using trigonometry, the inclination a is found to be a = c o s   θ s i n   θ and the intersection with the y axis is b = ρ s i n   θ . This aligns with Equation (4). These parameters are calculated for each combination of θ and ρ . The maximum ρ is set to the length of the image’s diagonal line [45].
Make 07 00026 g003
Figure 4. Example of original images f (x, y) and their corresponding Radon-transformed images R ρ , θ for samples from each class: (a) Original image of the “Empty” class, (b) Transformed image of the “Empty” class, (c) Original image of the “Event” class, (d) Transformed image of the “Event” class.
Figure 4. Example of original images f (x, y) and their corresponding Radon-transformed images R ρ , θ for samples from each class: (a) Original image of the “Empty” class, (b) Transformed image of the “Empty” class, (c) Original image of the “Event” class, (d) Transformed image of the “Event” class.
Make 07 00026 g004
Figure 5. Graphical representation of the 2DPCA implementation (after Yin, Xiuye, and Chen) [47].
Figure 5. Graphical representation of the 2DPCA implementation (after Yin, Xiuye, and Chen) [47].
Make 07 00026 g005
Figure 6. Eigenvectors of G t with their corresponding eigenvalues. The last ten eigenvectors have the highest eigenvalues, with a noticeable difference compared to the other eigenvectors.
Figure 6. Eigenvectors of G t with their corresponding eigenvalues. The last ten eigenvectors have the highest eigenvalues, with a noticeable difference compared to the other eigenvectors.
Make 07 00026 g006
Figure 7. Flowchart of the XAI model with a table of matrix sizes in the upper left corner.
Figure 7. Flowchart of the XAI model with a table of matrix sizes in the upper left corner.
Make 07 00026 g007
Figure 8. Detection accuracy versus eigenvalue cut-off using different numbers of principal components to train 2DPCA models with raw images (red) and Radon-transformed images (blue). The highest detection accuracies are highlighted with a green circle.
Figure 8. Detection accuracy versus eigenvalue cut-off using different numbers of principal components to train 2DPCA models with raw images (red) and Radon-transformed images (blue). The highest detection accuracies are highlighted with a green circle.
Make 07 00026 g008
Figure 9. (a) Eigenvalues of each principal component in the XAI model. (b) Actual contribution of each principal component to the classification process. (c,d) Represent the same as (a,b) but for the XAI (Radon) model.
Figure 9. (a) Eigenvalues of each principal component in the XAI model. (b) Actual contribution of each principal component to the classification process. (c,d) Represent the same as (a,b) but for the XAI (Radon) model.
Make 07 00026 g009
Figure 10. Flowchart of the combination process of 2DPCA and Random Forest in conjunction with RFE.
Figure 10. Flowchart of the combination process of 2DPCA and Random Forest in conjunction with RFE.
Make 07 00026 g010
Figure 11. (a) Differences among the proposed models, showing detection accuracy plotted against elapsed training time. (b) The same comparison, including deep learning CNN models.
Figure 11. (a) Differences among the proposed models, showing detection accuracy plotted against elapsed training time. (b) The same comparison, including deep learning CNN models.
Make 07 00026 g011
Table 1. Detection accuracy based on 2DPCA without using Radon transform.
Table 1. Detection accuracy based on 2DPCA without using Radon transform.
Number of Principal ComponentsEigenvalues Cut (%)Accuracy (%)Elapsed Time (s)
2097.5796.95301.692
3098.4397.96301.748
4098.8696.54301.705
5099.1396.54301.896
6099.3197.15301.766
7099.4497.35301.918
8099.5496.95301.916
9099.6297.15301.879
Table 2. Detection accuracy based on 2DPCA using Radon transform.
Table 2. Detection accuracy based on 2DPCA using Radon transform.
Number of Principal ComponentsEigenvalues Cut (%)Accuracy (%)Elapsed Time (s)
399.4894.30130.660
499.6494.70130.678
599.7394.50130.673
1099.9096.13130.683
1599.9593.89130.793
2099.9793.89130.751
2599.9893.89130.873
Table 3. Model evaluation results for the EPB detection of different deep learning models, the proposed 2DPCA, and XAI models.
Table 3. Model evaluation results for the EPB detection of different deep learning models, the proposed 2DPCA, and XAI models.
ModelAccuracy (%)Precision (%)Sensitivity (%)F1-Score (%)Elapsed Time (min)
ResNet1891.4592.7691.3692.0538.72
Inception-V389.4190.7889.3290.04114.30
VGG1685.3488.3585.1986.74215.70
VGG1989.4190.3289.3389.83255.96
2DPCA97.9698.0397.9597.995.03
2DPCA (Radon)96.1396.3796.1796.272.18
XAI98.1798.1898.1698.178.71
XAI (Radon)97.3597.4397.3897.403.03
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yacoub, M.; Abdelwahab, M.; Shiokawa, K.; Mahrous, A. Automatic Detection of Equatorial Plasma Bubbles in Airglow Images Using Two-Dimensional Principal Component Analysis and Explainable Artificial Intelligence. Mach. Learn. Knowl. Extr. 2025, 7, 26. https://doi.org/10.3390/make7010026

AMA Style

Yacoub M, Abdelwahab M, Shiokawa K, Mahrous A. Automatic Detection of Equatorial Plasma Bubbles in Airglow Images Using Two-Dimensional Principal Component Analysis and Explainable Artificial Intelligence. Machine Learning and Knowledge Extraction. 2025; 7(1):26. https://doi.org/10.3390/make7010026

Chicago/Turabian Style

Yacoub, Moheb, Moataz Abdelwahab, Kazuo Shiokawa, and Ayman Mahrous. 2025. "Automatic Detection of Equatorial Plasma Bubbles in Airglow Images Using Two-Dimensional Principal Component Analysis and Explainable Artificial Intelligence" Machine Learning and Knowledge Extraction 7, no. 1: 26. https://doi.org/10.3390/make7010026

APA Style

Yacoub, M., Abdelwahab, M., Shiokawa, K., & Mahrous, A. (2025). Automatic Detection of Equatorial Plasma Bubbles in Airglow Images Using Two-Dimensional Principal Component Analysis and Explainable Artificial Intelligence. Machine Learning and Knowledge Extraction, 7(1), 26. https://doi.org/10.3390/make7010026

Article Metrics

Back to TopTop