7.2. Classifier Performance (Cross-Validation)
Before GAN augmentation, a binary classification task was performed for each minority class versus the majority class (none), and various evaluation metrics were compared.
Table 6,
Table 7,
Table 8,
Table 9,
Table 10,
Table 11,
Table 12,
Table 13 and
Table 14 present the results of both pre-GAN and post-GAN augmentation, comparing macro-averaged Accuracy, Precision, F1-score, and AUC-ROC across 5-fold stratified cross-validation for five classifiers: Logistic Regression, SVM, KNN, Decision Tree, and Random Forest. For the SVM classifier, AUC-ROC values are not reported because probability calibration was not applied. Under extreme class sparsity, reliable calibration is unstable, rendering AUC-ROC values unreliable.
To prevent data leakage, GAN-based augmentation was performed independently within each cross-validation fold. For each fold, the GAN was trained exclusively on minority-class samples from the training split, and synthetic samples were generated solely to augment the corresponding training data. Validation folds contained only real samples and were never used during training or data generation, ensuring strict separation between training and evaluation data.
For several extremely underrepresented attack classes, pre-GAN evaluation was infeasible because at least one cross-validation contained no minority-class samples. In such cases, results are denoted by “-”, indicating that the corresponding metric could not be reliably computed and was therefore intentionally omitted.
7.2.1. Minority-Class Performance Improves Pos-GAN
Across most attack categories and classifiers, we see improvements in Precision, Recall, and F1-score after GAN augmentation. The most significant gains are in Logistic Regression, where the F1-score rises from about 0.62 before GAN to 1.00 after GAN (
Table 6), and in KNN, with initial F1-scores of 0.60 for Lateral Movement (
Table 9), increasing to a perfect 1.00 after GAN.
7.2.2. Consistent Accuracy Across All Tasks
The overall accuracy of all classifiers remained at 1.00, even after the addition of synthetic data to the dataset. This shows that as the detection rate for minority classes increased, the performance of the majority class (none) remained unchanged, a key property of real-world intrusion detection systems. This stability in accuracy, along with better macro-Recall and F1-scores, indicates that improvements in minority classes did not reduce the performance of the majority (“none”) class.
7.2.3. AUC-ROC Indicates Improved Separation
Where applicable, the AUC-ROC scores reached 1.00 or improved after augmentation. For example, in
Table 6 for Logistic Regression with Credential Access, the AUC-ROC score rose from 0.9970 to 1.000. These improvements demonstrate increased separation between attack traffic and normal (none) traffic after augmentation.
7.2.4. Patterns by Classifier
Table 6,
Table 7,
Table 8,
Table 9,
Table 10,
Table 11,
Table 12,
Table 13 and
Table 14 show the patterns by classifiers. Across classifiers, Decision Tree and Random Forest perform well even without GAN, and GAN further enhances their performance through augmentation. Meanwhile, Logistic Regression and KNN exhibit significant fluctuations in recall and F1-scores due to their sensitivity to imbalanced data. In contrast, SVM generally begins with high F1-scores in the mid-to-high 0.9 range and often reaches 0.99–1.00 after augmentation.
In
Table 6,
Table 7,
Table 8,
Table 9,
Table 10,
Table 11,
Table 12,
Table 13 and
Table 14, “-” indicates that the corresponding metric could not be reliably computed in the pre-GAN setting because there were insufficient minority-class samples in at least one cross-validation fold. AUC-ROC is not reported for the SVM classifier because probability estimates were not enabled; under extreme class sparsity, probability calibration may be unreliable.
7.2.5. Tactic-Specific Observations
Credential Access
Before augmentation, classifiers such as Logistic Regression had relatively low F1-scores despite high recall. After augmentation, all classifiers achieved an F1-score of 1.00, indicating that the synthetic samples improved the models’ ability to generalize and distinguish the minority class.
Privilege Escalation
This tactic type showed strong recall among classifiers in the pre-GAN phase, but its precision was occasionally weaker (e.g., Logistic Regression Precision = 0.8917). After augmentation, all metrics, including precision, improved across all classifiers, resulting in a perfect F1-score.
Exfiltration
Pre-GAN Logistic Regression and KNN perform slightly worse with F1-scores between 0.85 and 0.93 due to moderate precision. After using a GAN, all classifiers achieved an F1-score of 1.00, underscoring the effectiveness of data augmentation with synthetic samples.
Lateral Movement
KNN performance was weak before augmentation, with an F1-score of 0.6000. Also, Logistic Regression was not evaluated pre-GAN, likely due to the extremely small number of training samples for this class. After GAN augmentation, KNN performance improved significantly (F1-score from 0.6000 to 1.0000), indicating that the model benefited from synthetic data augmentation.
Resource Development
Pre-GAN metrics indicate that several classifiers, such as KNN and Logistic Regression, either performed poorly or returned unavailable results due to limited sample sizes. For example, KNN achieved an F1-score of 0.7000. After GAN augmentation, all classifiers reported an F1-score of 1.0000, indicating that augmentation improved performance and corrected underperforming classifiers due to data scarcity.
Reconnaissance
Classifiers such as Logistic Regression and SVMs were not evaluated pre-GAN due to insufficient training samples for this class. After augmentation, GAN-generated samples improved classifier performance, with all classifiers achieving an F1-score of 1.00. This demonstrates the effectiveness of GAN-generated data augmentation in improving model performance.
Defense Evasion
As with Reconnaissance, Defense Evasion lacked evaluation results from the pre-GAN phase for Logistic Regression and SVM. Post-augmentation showed significant improvement across all classifiers, confirming the effectiveness of GAN in managing underrepresented classes.
Initial Access
Pre-GAN evaluations for Logistic Regression and SVM were missing, likely due to the limited sample size for this class. After augmentation, all classifiers achieved an F1-score of 1.00, demonstrating the benefit of synthetic data generation for minority classes.
Persistence
Pre-GAN evaluations for Logistic Regression and Support Vector Machine (SVM) were unavailable, likely due to the small sample size in this class. After data augmentation, all classifiers achieved an F1-score of 1.00, demonstrating the benefits of synthetic data generation for minority classes.
7.3. Confusion Matrices
The confusion matrix results for each evaluated ATT&CK tactic are presented in
Table 15,
Table 16,
Table 17,
Table 18,
Table 19,
Table 20 and
Table 21. In intrusion detection systems, false negatives are particularly critical because undetected attacks can lead to severe security consequences; therefore, reduction in false negatives after augmentation represents a meaningful improvement in IDS effectiveness.
In
Table 18,
Table 19,
Table 20,
Table 21,
Table 22 and
Table 23, a dash (“-”) in the pre-GAN confusion matrix denotes cases where evaluation could not be performed due to extreme minority-class sparsity. For several attack tactics, the number of real minority samples was insufficient to support stable classifier training or stratified cross-validation, particularly for linear and margin-based models such as Logistic Regression and SVM. As a result, confusion matrices were undefined prior to augmentation. After GAN-based augmentation, sufficient minority samples are available, enabling stable training and evaluation across all classifiers.
7.3.1. Credential Access
Before GAN augmentation, confusion matrices for several classifiers indicate difficulty distinguishing the Credential Access class from the majority class, as shown in
Table 15. Logistic Regression tends to favor the majority class (none), leading to skewed predictions and poor generalization for the minority class. In logistic regression, false positives are prevalent: 642 benign “none” samples are misclassified as Credential Access.
Following GAN augmentation, minority detection improved across all classifiers. Logistic Regression showed the most significant improvement in detection rates for both minority and majority classes, as shown in
Table 15. The classifier perfectly distinguished Credential Access from the majority class (none). There were no false negatives; all 107,212 Credential Access samples were correctly classified, and no false positives or misclassified “none” samples occurred. This indicates that GAN-generated samples helped the classifier learn better class boundaries and created a more representative, balanced training dataset.
Decision Trees, already high-performing pre-GAN, benefited from the introduction of GAN-generated Credential Access samples, which eliminated misclassification errors. The model retained its perfect precision on the majority class while overcoming the single false negative in the pre-GAN. The result confirms that classifiers that already perform well can benefit from synthetic augmentation by eliminating rare misclassifications.
Random Forest maintained an F1-score of 1.00 even before augmentation, showcasing its strong resistance to class imbalance. After GAN-based augmentation, the model continues to classify both classes flawlessly, confirming that GAN-based data augmentation does not weaken high-performing models. This result also highlights the generalizability and stability of the proposed pipeline for generating additional training data, particularly for minority classes affected by class imbalance.
7.3.2. Privilege Escalation
GAN-based augmentation substantially improved classifier performance on the Privilege Escalation class, as shown in
Table 16. Before augmentation, the Logistic Regression model was effective at identifying Privilege Escalation samples but misclassified four benign records. After augmentation, the classifier improved the decision boundary by reducing false positives, achieving an F1 Score of 1.00.
The linear SVM model missed one Privilege Escalation case. After augmentation, the classifier misclassified only 4 out of over 10,700 Privilege Escalation instances, maintaining near-perfect performance on this substantially expanded minority set. This shows that GAN augmentation improves the SVM’s generalization to underrepresented minority attack types in the cybersecurity dataset.
KNN struggled with both types of errors (false positives and false negatives), likely due to the sparsity of data points for the minority class. Post-GAN, although two false positives remained, the increased diversity of synthetic samples allowed the model to predict all Privileged Escalation instances, with only 5 cases missed out of 107,194. By enriching the local neighborhoods, GAN augmentation makes KNN a more robust classifier, especially for minority classes.
The Decision Tree also improved through augmentation. Before adding synthetic samples, the model exhibited two false negatives due to limited exposure to minority classes. After incorporating more than 100,000 synthetic samples generated by a GAN, the model correctly identified all but one Privilege Escalation case. These findings indicate that decision tree models benefit from GAN-enhanced class balancing, thereby improving their generalizability and stability.
Random Forest, already strong before GAN, still failed to identify two Privilege Escalation instances. Post-GAN, the model eliminated misclassification thanks to the diverse set of synthetic minority-class samples. These results demonstrate the potential of GAN-based augmentation not only to improve underperforming classifiers but also to enhance the reliability of high-performing models, such as Random Forests, in imbalanced cybersecurity detection tasks.
7.3.3. Exfiltration
GAN-based augmentation improved minority class detection for Exfiltration, as summarized in
Table 17. Before augmentation, the model had both false positives and false negatives. After applying GAN augmentation, the model correctly classifies all benign cases and accurately detects most Exfiltration instances, with only three false negatives. This shows that GAN-based augmentation improves minority-class detection for linear classifiers.
SVM classifiers also benefited from GAN-based augmentation. In the pre-GAN phase, the model showed perfect precision but missed two Exfiltration samples due to limited training data for that class. After augmentation, the model recovers nearly all Exfiltration instances with only four false negatives. This demonstrates that GAN augmentation improved the decision boundary, enabling the model to generalize more effectively to minority classes in highly imbalanced settings.
KNN’s performance on Exfiltration class detection also improved with GAN-based augmentation. Before augmentation, the model was accurate but had limited recall due to the shortage of samples. After adding over 100,000 synthetic Exfiltration samples, the model recovers nearly all true positives, with only three false negatives. Although two false positives were introduced, the augmentation significantly improved overall classification performance.
The Decision Tree classifier performed very well on Exfiltration, even before augmentation, achieving 100% accuracy on the small minority-class sample set. After adding more than 100,000 synthetic Exfiltration samples, the model still achieved nearly perfect performance, misclassifying only one sample. This indicates that GAN-based augmentation helps maintain high performance, preventing the decision tree from overfitting to the limited dataset in the pre-GAN stage.
The Random Forest classifier, already highly accurate in detecting Exfiltration Pre-GAN, misclassifies only one instance, and continues to perform well after GAN-based augmentation. The model continues to misclassify only a single Exfiltration instance and maintains zero false positives, confirming that GAN-based augmentation preserves its already strong performance.
7.3.4. Lateral Movement
GAN-based augmentation significantly improved the classifiers’ reliability for Lateral Movement, as shown in
Table 18. Pre-GAN classifiers were limited by a very small number of positive samples. For example, SVM achieved perfect classification, but its confidence was based on just four positive samples, risking overfitting. After GAN augmentation, greater data variability was introduced, enabling the SVM to achieve only three misclassifications among over 100,000 synthetic samples.
KNN particularly benefits from GAN augmentation. Before augmentation, the classifier failed to detect any Lateral Movement instances but correctly identified the majority class. After adding over 100,000 synthetic samples generated by a GAN, the classifier achieved excellent recall and nearly perfect precision. This demonstrates that GAN-based augmentation enhances models such as KNN in their ability to recognize previously undetectable minority-class instances.
Decision Trees also reveal this trend. The model performed well on a very small minority class; however, after using GAN, it demonstrates perfect discrimination with the addition of over 100,000 synthetic Lateral Movement samples. This suggests that Decision Trees can clearly learn class boundaries while avoiding overfitting when GAN-augmented samples are included.
Although Random Forest already achieves perfect classification with very limited minority data, it benefits greatly from GAN-based augmentation. After incorporating GAN, the classifier correctly distinguished a large, balanced dataset without errors. This demonstrates that Random Forest is well-suited to learning from synthetically balanced data, thereby helping to address real-world class imbalance in cybersecurity datasets.
7.3.5. Resource Development
GAN augmentation proved valuable for Resource Development, where pre-GAN evaluation was infeasible due to the limited number of minority samples, particularly for Logistic Regression. As summarized in
Table 19, after augmentation, the classifier correctly identifies nearly all Resource Development and “none” samples, with only three misclassifications in total. This highlights a key contribution of the GAN augmentation framework: before it, evaluation was nonexistent due to severely limited minority-class instances.
SVM also experienced a notable improvement. With only one true positive, the model failed to generalize well. Before using GAN, the classifier lacked sufficient samples to learn effectively. After applying a GAN, the classifier achieved near-perfect performance, demonstrating the value of GAN-generated synthetic data.
KNN highlights the challenge of handling extreme class imbalance. Before GANs, the classifier was ineffective at predicting the minority class because it lacked sufficient neighbors. Post-GAN, with synthetic data generated by the GAN, it nearly achieved near-perfect recall and precision, with only three false negatives among more than 100,000 synthetic Resource Development samples, confirming the usefulness of GANs in addressing locality-sparse issues in KNN.
Decision Trees, which previously misclassified a minority instance, achieved perfect classification after using GAN. After employing GAN, the model accurately classifies all instances without sacrificing performance on the majority class.
Without augmentation, even ensemble models such as Random Forests struggle to generalize from a few minority-class samples. With GAN-based augmentation, the model improves its detection of the minority class while avoiding false positives.
7.3.6. Reconnaissance
Table 20 shows that most classifiers could not be meaningfully evaluated before augmentation. However, after introducing GAN-generated samples, all models achieved near-perfect classification, demonstrating the effectiveness of synthetic data in creating a learnable distribution where none previously existed.
Logistic Regression and SVM couldn’t evaluate pre-GAN, likely due to too few minority-class samples. After augmentation, both achieved perfect classification; all Reconnaissance samples were correctly identified, with no false positives or negatives. This highlights the substantial impact of GAN augmentation in generating learnable data distributions that did not previously exist.
Insufficient data limits the GAN’s performance, preventing the KNN classifier from learning meaningful patterns for the Reconnaissance class. After using a GAN, the model achieves near-perfect results, misclassifying only 2 Reconnaissance instances, and successfully captures the structure of the Reconnaissance class with a balanced training dataset. This demonstrates that GAN augmentation can improve the performance of classifiers such as KNN in detecting rare attack classes.
Before GAN, decision trees successfully predicted Reconnaissance samples despite the very small sample size. However, this could risk overfitting. After GAN augmentation, the model generalizes well, achieving perfect recall and precision, confirming the effectiveness of GAN-based augmentation.
The Random Forest shows perfect accuracy; however, this performance is based on only two positive samples, providing minimal insight into the model’s true ability to generalize for the Reconnaissance class. After applying GAN-based augmentation, the model was tested on a larger, more balanced dataset, yet it still achieved 100% accuracy, demonstrating robustness in learning to recognize the Reconnaissance attack type.
7.3.7. Defense Evasion
Defense Evasion, one of the dataset’s most underrepresented classes, renders pre-GAN evaluation infeasible for classifiers such as Logistic Regression and SVM. With only one misclassification,
Table 21 shows how GAN-based augmentation affects classifiers such as Logistic Regression, which correctly classify all other samples.
The linear SVM model benefited from synthetic samples generated by a GAN. It accurately identifies nearly every instance of Defense Evasion without affecting its ability to classify the majority class. As with Logistic Regression, this classifier shows no significant pre-GAN performance, likely due to the extreme imbalance or the lack of Defense Evasion samples in the original training data.
Before GAN augmentation, KNN failed to recognize the minority class (Defense Evasion), likely due to its underrepresentation in the training data. After GAN augmentation, the KNN model improves in correctly identifying nearly all Defense Evasion cases with only one misclassification. This highlights the benefit of GAN augmentation for models like KNN in detecting minority classes in cybersecurity tasks with class imbalance.
The Decision Tree model, before augmentation, fails to identify the minority class (Defense Evasion), probably because of severe class imbalance and bias toward the majority class. Following GAN augmentation, the model achieves nearly perfect classification performance on the previously underrepresented class. This also demonstrates how GANs can effectively enhance models such as decision trees that may struggle with extreme class imbalance.
Random Forest exhibited the same pattern; it did not detect the minority class before GAN but showed strong recovery after GAN, with only one mistake. This confirms that ensemble models benefit from GAN augmentation when the original dataset is too sparse to provide practical training examples.
7.3.8. Initial Access
Initial Access is also among the most underrepresented classes in the dataset, preventing meaningful pre-GAN evaluation for most classifiers. As summarized in
Table 22, none of the models detected the minority class before augmentation. No pre-GAN confusion matrix was produced for Logistic Regression due to the extreme imbalance in the Initial Access dataset before augmentation. Post-GAN, the classifier correctly identifies nearly all Initial Access samples, with only one misclassification. The model also maintains complete accuracy for the majority class while avoiding false positives.
As with Logistic Regression, SVM indicates that no pre-GAN evaluation for Initial Access was likely due to the severe class imbalance in the training dataset. After applying the GAN, the model shows high accuracy, correctly identifying only one Initial Access sample. The majority class is also correctly classified with zero false positives. Both Logistic Regression and SVM misclassified one Initial Access sample, confirming their consistent performance across multiple linear models after the GAN.
Before GAN, KNN failed to detect Initial Access entirely, likely due to insufficient training data for the minority class. In a single instance, KNN could not establish any neighborhood boundary. After GAN augmentation, the classifier correctly identified nearly all Initial Access samples. This demonstrates that GAN-generated samples constitute a representative, diverse dataset, thereby enabling KNN to form stable neighborhoods.
The Decision Tree classifier completely failed to identify the single Initial Access sample due to severe class imbalance. After augmentation, the classifier accurately identified all Initial Access samples, misclassifying only 1 of over 100,000. This demonstrates the substantial improvement that GAN-augmentation samples provide to the classifier, narrowing the detection gap between the two classes.
Random Forest initially failed to identify the minority class (Initial Access) before using GAN. Due to a single example and strong class imbalance, the classifier defaulted to labeling everything as the majority class. After GAN-based augmentation, the classifier correctly identified nearly all minority class instances while still accurately classifying the majority class.
7.3.9. Persistence
Persistence represented another minority class with very few samples, which impeded meaningful pre-GAN evaluation for most classifiers. As shown in
Table 23, no classifier reliably detected the minority class before augmentation; Logistic Regression and SVM yielded no evaluable results, and all other models failed to identify any positive instances.
Logistic Regression and SVM, which could not be meaningfully evaluated before GAN, both achieved near-perfect classification after augmentation, misclassifying only one minority instance while perfectly classifying the majority class.
KNN failed before GAN because it lacked sufficient neighbors and relied solely on the majority class. After applying GAN augmentation, the model correctly identified nearly all Persistence samples, demonstrating that synthetic samples helped it establish effective neighborhood boundaries.
Decision Trees and Random Forest showed similar patterns, both failing to identify Persistence pre-GAN. However, after augmentation, they detected nearly all instances, with just one misclassification. These results confirm that GAN-based augmentation consistently improves a classifier’s ability to recognize rare attack classes.
The complete set of confusion matrices from
Table 15,
Table 16,
Table 17,
Table 18,
Table 19,
Table 20,
Table 21,
Table 22 and
Table 23 shows a significant improvement in classification performance after applying GAN-based augmentation across all attack categories. Before augmentation, classifiers had difficulty identifying minority classes, often completely misclassifying them as the majority class (none). In several cases, models failed to recognize or produce evaluation results for a very small number of minority class samples.
After augmentation, the classifiers showed significant improvements in correctly identifying minority classes while maintaining high accuracy on the majority class. The augmented data enabled the models to learn more effective decision boundaries, reducing false negatives and improving overall detection performance.
Therefore, the confusion matrices demonstrate the effectiveness of the GAN-based augmentation method in improving models’ ability to distinguish minority classes from the majority class. Consequently, classification performance improved significantly, with a substantial reduction in bias toward the majority class.
For completeness, the detailed differences in pre- and post-GAN confusion matrices (ΔTP, ΔFN, ΔTN, ΔFP) for each classifier and attack type are presented in
Appendix A Table A1,
Table A2,
Table A3,
Table A4,
Table A5,
Table A6,
Table A7,
Table A8 and
Table A9. These tables provide precise numerical evidence of the improvements, particularly in reducing false positives and false negatives for underrepresented classes.
7.4. t-SNE Visualizations
To evaluate the distribution and quality of synthetic samples, t-SNE dimensionality reduction was applied to both real minority-class samples and GAN-generated samples. The t-SNE visualizations are provided for qualitative insight into the alignment between real and synthetic samples. They are not used for interpretation or as a quantitative measure of class separability.
In all visualizations, blue points denote GAN-generated synthetic samples, and red points denote real minority samples. This is shown in
Figure 7, where the blue dots indicate synthetic samples generated by the GAN for Credential Access. In contrast, the red dots represent real minority-class samples from the original dataset.
The graph shows very few red dots, indicating an extremely low number of original sample counts for the Credential Access class, underscoring the class imbalance. The dense blue dots represent many synthetic samples generated via GAN to address this imbalance. The synthetic points form a broad, coherent region around the few real samples, which is consistent with the GAN learning a meaningful approximation of the minority-class structure rather than simply memorizing individual examples.
The t-SNE plot in
Figure 8 shows the distribution of real and synthetic minority samples for the Privilege Escalation class after GAN augmentation. Real minority samples (red dots) are few and scattered, consistent with the original data imbalance. The synthetic minority samples (blue dots) are densely packed and form smooth clusters, indicating that the GAN-based augmentation method has successfully generated many samples. Although the real and synthetic samples occupy nearby regions that do not overlap perfectly, many synthetic points lie close to the real samples, suggesting that the GAN has captured key aspects of the Privilege Escalation feature space despite the limited number of real observations.
The t-SNE plot in
Figure 9 shows the distribution of real and synthetic Exfiltration samples after GAN-based data augmentation. The blue dots representing the synthetic minority samples form well-defined clusters, indicating that the GAN generated a large, diverse sample set. Only a few genuine Exfiltration samples are visible, underscoring the class imbalance before augmentation. The synthetic samples surround the real samples and appear in similar regions, suggesting that the GAN has effectively learned the overall latent distribution of the minority class. The visual similarity between real and synthetic samples supports the conclusion that the GAN-generated data approximate the structure of the real minority class, as reflected in the improved results shown in the confusion matrices in
Table 10,
Table 11,
Table 12,
Table 13,
Table 14,
Table 15,
Table 16 and
Table 17.
The t-SNE plot in
Figure 10 shows the Lateral Movement class after GAN-based augmentation with real and synthetic minority-class samples. The synthetic data points (blue) dominate the visualization and are densely and smoothly spread out, indicating that the GAN effectively learned a diverse representation of the minority class. This implies high-quality synthetic data generation, which helps reduce overfitting and enhances classifier performance, as demonstrated in the confusion matrices in
Table 15,
Table 16,
Table 17,
Table 18,
Table 19,
Table 20,
Table 21,
Table 22 and
Table 23. A small number of real Lateral Movement samples (red) are visible, highlighting the class imbalance.
Figure 11 shows the t-SNE visualization of Resource Development after GAN-based augmentation of the minority class. A dense and well-spread distribution of synthetic samples dominates the plot. Their broad coverage indicates that the GAN generated a diverse and representative set of synthetic data for this class. The real samples are limited and mixed with synthetic samples, indicating substantial overlap between the real and generated data.
Figure 12 shows the t-SNE plot for the Reconnaissance class after GAN-based data augmentation. The synthetic data forms several distinct subclusters, suggesting that the GAN has captured complex structure within the Reconnaissance feature space, with the few real samples embedded within these regions.
Figure 13 shows the t-SNE visualization for the Defense Evasion class after GAN-based data augmentation. The synthetic samples form a large, tight cluster with a smooth density spread across the t-SNE space. This pattern is consistent with stable data generation without apparent collapse into a single mode while exhibiting diversity in the generated samples.
Figure 14 displays the t-SNE plot for Initial Access after GAN augmentation, showing the distribution of both synthetic and real minority samples. The artificial data points form a dense, compact cluster with extensive coverage, indicating that the generator can produce diverse yet consistent samples. Only one real minority point is visible, reflecting the limited number of genuine samples available in this class.
Figure 15 displays the t-SNE plot for Persistence after GAN-based augmentation, comparing synthetic and real minority samples. The artificial data points form a dense, well-structured cluster that spans the local region of the feature space associated with Persistence, indicating that the generator can produce both diverse and coherent samples. In contrast, only one real Persistence point is visible, emphasizing the extreme imbalance of this class in the original dataset. The proximity of the real point to the synthetic cluster suggests that the GAN approximated the underlying distribution of the minority class despite limited training data.
In conclusion, the t-SNE visualizations for the minority classes, shown in
Figure 5,
Figure 6,
Figure 7,
Figure 8,
Figure 9,
Figure 10,
Figure 11,
Figure 12 and
Figure 13, illustrate the effectiveness of the GAN-based augmentation strategy in generating realistic samples for underrepresented attack classes. Across all figures, the synthetic minority samples (blue) form dense, continuous clusters, consistent with the GAN capturing key aspects of the minority-class distributions. The t-SNE visualizations collectively reinforce the conclusions from the confusion matrices: GAN-based augmentation balances class distribution and maintains feature-space structure, enhancing classifier generalization and detection accuracy across minority cyber threat classes.