From Detection to Decision: Transforming Cybersecurity with Deep Learning and Visual Analytics
Abstract
1. Introduction
2. Related Work
2.1. Machine Learning and Deep Learning Vulnerability Detection
2.2. Interpretability and Explainable AI (XAI)
2.3. Hybrid Frameworks and Visual Analytics
2.4. Research Gaps
3. Preliminary
Motivating Examples
- 1.
- Minimum Intermediate Representation (MIR) Learning:
- 2.
- Kernel Extreme Learning Machine (KELM):
- 3.
- Transformer-Based Semantics (BERT, CodeBERT, CodeTrans):
- 4.
- Graph-Based Structural Analysis (CPG + DGCNN)
- 5.
- Two-Stage Deep Learning Model (CNN–LSTM):
- 6.
- Survey of Deep Learning Techniques:
- 7.
- Visual Analytics for Cyber Vulnerability Assessment:
- 8.
- Explainable AI (XAI) and Visual Analytics
4. Steps
4.1. Data Collection and Preprocessing
- Publicly Available Vulnerable Repositories—We collected open-source projects from GitHub that had associated CVE reports. These were cross-referenced with the National Vulnerability Database (NVD) to confirm vulnerability authenticity and classification.
- Reference Corpora—The National Software Reference Library (NSRL) [17], hosted by NIST, was used as a “benign baseline” to help estimate false positive rates. NSRL samples were retrieved from NIST’s public distribution portal, hashed, and cross-checked against the vulnerability datasets to ensure no contamination of labels [26].
- All source code was normalized to a canonical syntax style.
- Identifying metadata such as project names and authors were anonymized.
- Semantic processing involved tokenization for BERT-based embeddings.
- Structural processing involved Abstract Syntax Tree (AST) parsing and CPG construction.
- Dynamic execution traces, where available, were collected and prepared for sequential modeling with LSTM networks.
- Textual artifacts (commit messages, bug reports) were encoded using RoBERTa embeddings to capture contextual information.
4.2. Multi-Model Feature Extraction
- Semantic Features: Derived from a fine-tuned BERT-based encoder built on CodeTrans [9], enabling the capture of long-range token dependencies and contextual semantics.
- Textual Features: RoBERTa embeddings from commit logs and issue reports provide historical and human-readable context about code changes, many of which are indicative of vulnerability fixes.
- Dynamic Behavior Features: Execution logs and traces are processed by LSTM networks, enabling detection of vulnerabilities that manifest only at runtime, such as race conditions and resource leaks.
4.3. Hybrid Modeling and Integration
- Algorithm: Deep Q-Network (DQN) with experience replay.
- State (s): Vector of current model scores , , with rolling history of accuracy, precision, and false positive rate
- Action (a): Increment or decrement one of the fusion weights () by 0.05 while maintaining normalization (∑).
- Reward (R):
- Training Parameters:
- Episodes: 500
- Replay buffer size: 5000
- Batch size: 64
- Learning rate: 1 × 10−41
- Optimizer: Adam
- Discount factor: γ = 0.95\g
- amma = 0.95γ = 0.95
- Target network update: every 50 steps
- Exploration: ε-greedy, decaying from 0.2 → 0.01 over 500 episodes
- Schedule: The RL agent updates fusion weights after every 3 training epochs using validation-set feedback for reward computation.
4.4. Risk Prioritization and Scoring
- Computes severity scores according to the CVSS v3.1 specification, considering exploitability, impact, and environmental metrics.
- Integrates SHAP-based feature attributions into the scoring interface, enabling security analysts to understand which code regions or features contributed to a high severity score.
- Flags vulnerabilities as “critical” when exploitability > 0.7 and impact > 0.6, ensuring immediate attention.
4.5. Visual Analytics and Interface
- Heatmaps highlight high-risk code segments by mapping SHAP feature attribution scores to source lines, enabling targeted review by analysts.
- Severity histograms and box plots (Figure 3) summarize CVSS v3 base scores and severity levels across the evaluated dataset. Axis labels, legends, and scale indicators are included to ensure interpretability without external context.
- Pair plots (Figure 4) illustrate correlations between exploitability scores, impact ratings, and CVE presence, providing multi-dimensional insights into detected vulnerabilities.
- Graph Model Explainability. For structural representations (DGCNN + CPG), we apply GNNExplainer to derive node- and edge-level attributions. For each vulnerable code sample, GNNExplainer produces a sparse mask highlighting subgraphs most influential in prediction. These masks are overlaid on the Code Property Graph and linked back to the corresponding source lines in the analyst dashboard. This ensures that graph-level explanations remain interpretable and actionable, complementing SHAP-based token and textual explanations.
- SHAP explanation panels display feature contributions at token, node, and textual levels, allowing validation or challenge of model outputs.
- Autonomous Suggestions: Offer automatic suggestions for mitigating steps, such applying patches, altering configurations, or taking other corrective action, based on risk scores and vulnerability categories [15].
- Integration with Security Functions: Ensure that the system can easily interface with cybersecurity workflows and current SIEM systems to enable automated warnings and reactions as part of an organization’s active defense plan [11].
4.6. Feedback and Constant Improvement
- Threshold adjustment—dynamically recalibrating decision thresholds to reduce false positives.
- Incremental retraining—selectively fine-tuning BERT and KELM components on newly validated samples.
4.7. Hypothesis
- Model Components and Equations:
- Accuracy:
- Efficiency:
- Visual Interpretability:
- Explainability Metric (XAI):
- Overall Metric:
- Reverse Validation Condition for H1:
- Baselines:
- ○
- BERT-only
- ○
- DGCNN-only
- ○
- KELM-only
- ○
- Static analysis tools (e.g., SonarQube, Checkmarx)
- Metrics: Accuracy, precision, recall, F1, processing time per sample, visual interpretability score, explainability score XAI.
- Statistical Testing: Paired t-test or Wilcoxon signed-rank test with p < 0.05 p < 0.05 p < 0.05 to assess significance.
5. Research Details
- Hypothesis 1 (H1): Integrating semantic, structural, and risk-ranking models with adaptive visual analytics will improve accuracy, efficiency, and interpretability.
- Hypothesis 2 (H2): Incorporating real-time learning and multi-modal inputs will outperform traditional systems in both speed and accuracy.
5.1. System Architecture (H1, H2)
- State: vector of model scores and rolling accuracy history.
- Actions: increment/decrement one of α, β, γ by 0.05, constrained to sum = 1.
- Reward: +1 for correct classification, −1 for false positives, −2 for false negatives.
- Algorithm: Deep Q-Learning with replay buffer size = 5000.
- Training Schedule: RL updates every 3 epochs using validation set feedback.
- Exploration: ε-greedy, ε decays from 0.2 → 0.01 over 500 episodes.
Component | Key Hyperparameters |
BERT (CodeTrans fine-tuned) | LLR = 2 × 10−5, AdamW, batch = 32, epochs = 10, max seq = 512 |
DGCNN | Hidden size = 256, dropout = 0.3, LR = 1 × 10−3, batch = 64, epochs = 20 |
KELM | Kernel = RBF, C = 100, = 0.001 |
RL Fusion Agent | Deep Q-learning, replay buffer = 5000. Decay 0.2 0.01, update every 3 epochs |
5.2. Visual Analytics Interface
- Visualization of high-risk vulnerabilities via CVSS overlays and SHAP heatmaps.
- Drill-down inspection of specific code regions linked to top-ranked vulnerabilities.
- Real-time severity correlation analysis through pair plots and attention visualizations [12].
5.3. Dataset and Experimental Setup (H1, H2)
- Tokenization (BERT input): Code is tokenized using a subword tokenizer trained on multilingual programming languages (Python, C, C++, Java). Special tokens are aligned with CVE labels.
- AST Parsing and CPG Construction (DGCNN input): Abstract Syntax Trees (ASTs) are extracted using Clang/Joern, then converted into Code Property Graphs (CPGs). DGCNN processes node embeddings to capture structural properties [7].
- Trace Alignment (Behavioral input): Execution traces are collected via sandbox runs and aligned with corresponding source samples. Traces are embedded using RoBERTa to capture dynamic behavior.
- 70% training, 15% validation, 15% testing.
- Stratified by CVE type and severity score to preserve class balance.
- For reproducibility, dataset splits and preprocessing configurations are documented in the project methodology. All datasets used are publicly available (see Data Availability Statement).
5.4. Evaluation Criteria
- AUPRC (Area Under Precision–Recall Curve): To evaluate robustness under class imbalance.
- MCC (Matthews Correlation Coefficient): To capture balanced performance across TP/TN/FP/FN.
- Calibration metrics: Expected Calibration Error (ECE; 15 bins) and Brier score are reported to assess reliability of predicted probabilities. Temperature scaling on the validation set is used as the calibration method.
5.5. Interpretability Validation
- Clarity and accuracy of risk insights—SHAP-based feature attributions were checked against known vulnerability features to confirm alignment.
- Usefulness of explanations for decision-making—GNNExplainer outputs and SHAP overlays were examined to verify whether highlighted tokens, nodes, or code regions consistently matched the vulnerability context.
5.6. Real-World Deployment Considerations
- SIEM Compatibility: Outputs are formatted in JSON/CSV and aligned with standard SIEM ingestion formats (e.g., Splunk, Elastic Stack) for alert correlation.
- Resource Efficiency: KELM and optimized inference paths keep average per-sample runtime under 50 ms, supporting real-time alerting requirements.
- Expected detection-to-alert latency: <500 ms per sample in controlled test environments.
- Incremental retraining can be performed overnight (<24h turnaround for incorporating new CVE data).
- Modular design ensures that individual components (BERT, DGCNN, KELM) can be updated independently.
- Conducting full enterprise SOC deployment trials to measure latency, throughput, and analyst usability in real-world settings.
- Extending integration with vulnerability management systems (e.g., Tenable, SonarQube) to enable automatic ticket creation.
- Developing edge-compatible deployments using ONNX models for lightweight environments.
6. Proposed Approach
6.1. Overview
- Data Ingestion and Preprocessing
- Multi-Model Feature Extraction
- -
- -
- -
- -
- Textual Metadata Analysis (RoBERTa): Incorporates commit messages, bug reports, and CVE summaries with multi-modal fusion to improve prediction reliability [14].
- Modeling and Integration
- Explainability and Visual Analytics (H1)
- Feedback Loop and continuous Learning (H2)
6.2. Workflow Description
- Semantic Embedding—Tokenize code, generate contextual embeddings with fine-tuned BERT, inspired by CodeTrans [9].
- Behavioral Representation—Encode execution traces with LSTM to detect runtime anomalies.
- Risk Classification—Classify vulnerability severity with KELM for fast, interpretable results.
- Fusion and Scoring—Weighted aggregation of semantic, structural, and behavioral outputs.
- Visualization—Analyst dashboard with heatmaps, severity distribution plots, and SHAP explanations.
- Feedback Integration—Analyst feedback updates model thresholds; new CVEs ingested continuously.
6.3. System Diagram
- Data ingestion pipelines.
- Semantic, structural, behavioral, and textual embedding modules.
- Fusion engine + KELM scorer.
- XAI dashboard with SIEM-compatible alert generation.
- Feedback loop for online learning and model refinement.
7. Experimental Evaluation and Results
7.1. Dataset and Benchmark
- NSRL Dataset: Served as a clean baseline for evaluating false positive rates [17].
7.2. Evaluation Metrics
- Recall =
- F1-score =
- AUC (area under curve): For binary classification robustness.
- Where is the predicted probability and
- False positive rate (FPR) and false negative rate (FNR): Evaluated critical misclassification risks.
- Efficiency: Measured via inference latency per sample.
7.3. Baseline Models for Comparison
7.4. Result Summary
7.5. Interpretability Validation
7.6. Ablation Study
- BERT excluded: Recall dropped −7.5%, confirming importance of semantic embeddings.
- DGCNN excluded: F1 decreased −6.8%, showing structural analysis is critical.
- KELM excluded: Latency increased 40%, and prioritization was lost.
- SHAP disabled: XAI score dropped to 61%, comparable to transformer-only models.
7.7. Summary of Findings
- Multi-modal fusion improved robustness (+8% F1 over single models).
- XAI dashboards improved interpretability, with internal validation confirming alignment of SHAP/GNNExplainer outputs with known vulnerability features.
- KELM ensured low-latency scoring, critical for DevSecOps pipelines.
- Online learning enhanced adaptability to zero-day CVEs, validating the hypothesis.
8. Conclusions and Future Work
8.1. Summary of Contribution
- A multi-stream architecture that fuses semantic and structural representations to improve detection performance and reduce false positives.
- The incorporation of Explainable AI (XAI) methods, including SHAP-based visualizations, to enhance the interpretability of model outputs.
- A visual dashboard that facilitates real-time, analyst-facing vulnerability assessments.
- A feedback-enabled learning mechanism that supports system adaptability to emerging threats.
8.2. Research Significance
8.3. Limitations
- Prototype status: The framework has not yet been validated in enterprise-scale production environments.
- Baseline gaps: Comparative benchmarking against tools such as SonarQube, Checkmarx, or Semgrep remains incomplete.
- Dataset scope: Evaluation focused on open-source and synthetic datasets; applicability to large proprietary codebases remains to be tested.
- Computational overhead: Multi-model fusion increases resource requirements, which may challenge deployment in constrained environments.
8.4. Future Work
- Enterprise Deployment: Integration with SIEM systems (e.g., Splunk, ELK) to enable live alerting and incident response [26].
- Adaptive Fusion Optimization: Exploring adaptive fusion strategies, such as reinforcement learning or meta-learning, may enhance the effectiveness of the model ensemble across diverse scenarios.
- Analyst-Centric Usability Studies: While this work included internal interpretability validation, structured usability studies with security analysts and practitioners remain future work. These will assess system interpretability, decision support utility, and impact on analyst trust and response behavior.
- Lightweight Deployment and Integration: Efforts will be directed toward optimizing the framework for deployment in cloud-native and edge environments, as well as integration into secure development workflows (e.g., CI/CD pipelines and IDE-based feedback tools) [31].
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
Abbreviations
AI | Artificial Intelligence |
BERT | Bidirectional Encoder Representations from Transformers |
CPG | Code Property Graph |
CNN | Convolutional Neural Network |
CVE | Common Vulnerabilities and Exposures |
CVSS | Common Vulnerability Scoring System |
DGCNN | Deep Graph Convolutional Neural Network |
KELM | Kernel Extreme Learning Machine |
LSTM | Long Short-Term Memory |
ML | Machine Learning |
NLP | Natural Language Processing |
NVD | National Vulnerability Database |
SHAP | SHapley Additive exPlanations |
SIEM | Security Information and Event Management |
XAI | Explainable Artificial Intelligence |
MIR | Learrning: Minimum Intermediate Representation Learning |
Appendix A
Appendix A.1. Table for Evaluation Metrics
Metric Category | Formula/Method | H1 Target | H2 Target |
---|---|---|---|
Accuracy | +5% vs. best single model | +8% vs. traditional tools | |
Efficiency (E) | Avg. inference time per sample | <50 ms | <50 ms |
Interpretability (V) | Section 6.2 for formula | ≥85% correct | ≥85% correct |
Explainability (XAI) | Mean SHAP attribution accuracy | ≥0.8 | ≥0.8 |
Robustness | Performance drop under noisy/adversarial input | <5% | <5% |
AUPRC | Precision–Recall AUC | ≥0.85 | ≥0.88 |
MCC | ≥0.75 | ≥0.80 | |
Calibration | ECE (15 bins), Brier Score | ECE < 0.05, Brier < 0.15 | ECE < 0.04, Brier < 0.12 |
Appendix A.2. Interpretability Evaluation Details
Metrics | Assessment Outcome | Notes |
---|---|---|
Clarity of SHAP Explanations | High | Attributions aligned with critical code features in >85% of tested samples |
Visual Layout of Heatmap | Medium High | Explanation consistently identified vulnerable regions, aiding prioritization. |
Speed of Decision Support | Medium | Prototype successfully interfaced with Splunk and GitLab CI/CD; optimization needed for latency |
Trust in Model Output | High | Overlap observed between token-level explanations |
Appendix A.3. Ablation Study
Configuration | Precision | Recall | F1 | AUPRC | MCC | vs. Full | p-Value |
---|---|---|---|---|---|---|---|
Full hybrid | 0.91 0.01 | 0.89 0.01 | 0.91 | 0.92 0.01 | 0.81 | - | - |
- BERT | 0.85 0.02 | 0.82 0.02 | 0.83 | 0.84 | 0.72 | −7.5% recall | 0.01 |
- DGCNN | 0.86 0.01 | 0.80 0.02 | 0.84 | 0.85 0.02 | 0.73 | −6.8% F1 | 0.02 |
- KELM | 0.82 0.02 | 0.79 0.01 | 0.80 | 0.81 | 0.70 | +40% latency | 0.03 |
- SHAP | 0.90 0.01 | 0.87 0.01 | 0.89 | 0.89 | 0.77 | XAI score to 61% | 0.04 |
Appendix A.4. Extended Evaluation
Model | Precision | Recall | F1-Score | AUC | AUPRC | MCC | FPR (%) | ECE % | Brier | Latency (ms) |
---|---|---|---|---|---|---|---|---|---|---|
BERT Only [3] | 0.84 | 0.80 | 0.82 | 0.86 | 0.84 | 0.72 | 9.3 | 6.8 | 0.16 | 87 |
DGCNN Only [4] | 0.81 | 0.77 | 0.79 | 0.84 | 0.81 | 0.68 | 10.5 | 7.1 | 0.18 | 91 |
KELM Only [2] | 0.76 | 0.74 | 0.75 | 0.80 | 0.77 | 0.61 | 12.1 | 9.4 | 0.20 | 33 |
CNN-LSTM [20] | 0.82 | 0.78 | 0.80 | 0.85 | 0.83 | 0.70 | 10.3 | 7.5 | 0.17 | 102 |
GraphCodeBERT [22] | 0.86 | 0.83 | 0.84 | 0.88 | 0.86 | 0.75 | 8.5 | 6.2 | 0.15 | 95 |
Proposed Hybrid | 0.91 | 0.89 | 0.90 | 0.94 | 0.92 | 0.81 | 5.7 | 2.8 | 0.11 | 68 |
References
- Zeng, P.; Lin, G.; Pan, L.; Tai, Y.; Zhang, J. Software Vulnerability Analysis and Discovery Using Deep Learning Techniques: A Survey. IEEE Access 2020, 8, 197158–197172. [Google Scholar] [CrossRef]
- Liao, Q.V.; Varshney, K. R Human-Centered Explainable AI (XAI): From Algorithms to User Experiences. arXiv 2021, arXiv:2110.10790. [Google Scholar]
- Alqarni, M.; Azim, A. Low-Level Vulnerability Detection Using Advanced BERT Language Model. In Proceedings of the 35th Canadian Conference on AI, Toronto, ON, Canada, 30 May–3 June 2022. [Google Scholar]
- Walters, B.; Nguyen, D. Enhancing Cybersecurity Assessments with Visual Analytics. Secur. Commun. Netw. 2022, 13, 987–1005. [Google Scholar]
- Aldhaheri, A.; Alwahedi, F.; Ferrag, M.A.; Battah, A. Deep Learning for Cyber Threat Detection in IoT Networks: A Review. Internet Things Cyber-Phys. Syst. 2023, 61, 1–25. [Google Scholar] [CrossRef]
- Chu, Z.; Wan, Y.; Li, Q.; Wu, Y.; Zhang, H.; Sui, Y.; Xu, G.; Jin, H. Graph Neural Networks for Vulnerability Detection: A Counterfactual Explanation. arXiv 2024, arXiv:2404.15687v1. [Google Scholar] [CrossRef]
- Davis, M.; White, L. DGCNN and CPG for Vulnerability Detection in C/C++ Software. J. Softw. Eng. Res. 2022, 7, 112–130. [Google Scholar]
- Guo, D.; Ren, S.; Lu, S.; Feng, Z.; Tang, D.; Liu, S.; Zhou, L.; Duan, N.; Svyat-kovskiy, A.; Fu, S. GraphCodeBERT: Pretraining Code Representations as Graphs. arXiv 2021, arXiv:2009.08366v2. [Google Scholar]
- Elnaggar, A.; Ding, W.; Jones, L.; Gibbs, T.; Feher, T.; Angerer, C.; Severini, S.; Matthes, F.; Rost, B. CodeTrans: Towards Cracking the Language of Silicon’s Code Through Self-Supervised Deep Learning and High Performance Computing. arXiv 2021, arXiv:2104.02443. Available online: https://arxiv.org/abs/2104.02443 (accessed on 25 August 2025).
- Hajipour, H.; Hassler, K.; Holz, T.; Schönherr, L.; Fritz, M. CodeLMSec Benchmark: Systematically Evaluating and Finding Security Vulnerabilities in Black-Box Code Language Models. arXiv 2023, arXiv:2302.04012. [Google Scholar]
- Alperin, K.B.; Wollaber, A.B.; Gomez, S.R. Improving Interpretability For Cyber Vulnerability Assess Focus and Context Visualizations. In Proceedings of the 2020 IEEE Symposium on Visualization for Cyber Security (VizSec), Salt Lake City, UT, USA, 28 October 2020. [Google Scholar]
- Coussement, K.; Abedin, M.Z.; Kraus, M.; Maldonado, S.; Topuz, K. Explainable AI for Enhanced Decision-Making. Decis. Support Syst. 2024, 184, 114276. [Google Scholar] [CrossRef]
- Li, X.; Wang, L.; Xin, Y.; Yang, Y.; Chen, Y. Automated Vulnerability Detection in source code using Minimum Intermediate Representation Learning. Appl. Sci. 2020, 10, 1692. [Google Scholar] [CrossRef]
- Li, Q.; Ma, Q.; Nie, W.; Liu, A. Reinforcement Learning Based Multi-modal Feature Fusion Network for Novel Class Discovery. arXiv 2023, arXiv:2308.13801v1. [Google Scholar] [CrossRef]
- Alhafi, M.M.; Hammade, M.; Jallad, K.A. Vulnerability Detection Using Two Stage Deep Learning Model. arXiv 2023, arXiv:2305.09673. [Google Scholar] [CrossRef]
- Chittala, S. Securing DevOps Pipelines: Automating Security in DevSecOps Frameworks. J. Recent Trends Comput. Sci. Eng. 2024, 12, 31–44. [Google Scholar] [CrossRef]
- National Institute of Standards and Technology (NIST). “National Software Reference Library (NSRL)”. 2023. Available online: https://www.nist.gov/itl/ssd/software-quality-group/national-software-reference-library-nsrl/ (accessed on 25 August 2025).
- Omar, M.; Shiaeles, S. VulDetect: A Novel Technique for Detecting Software Vulnerabilities Using Language Models. Preprint 2022. Available online: https://pure.port.ac.uk/ws/portalfiles/portal/80445773/VulDetect_A_novel_technique_for_detecting_software_vulnerabilities_using_Language_Models.pdf (accessed on 25 August 2025).
- VulnDetect: Public Vulnerability Detection Dataset. MIT License. Available online: https://vulndetect.org/ (accessed on 25 August 2025).
- Xuan, C.D. A new approach to software vulnerability detection on CPG analysis. Cogent Eng. 2023, 10, 2221962. [Google Scholar] [CrossRef]
- Liu, R.; Wang, Y.; Xu, H.; Liu, B.; Sun, J.; Guo, Z.; Ma, W. Source Code Vulnerability Detection: Combining Code Language Models and Code Property Graphs. arXiv 2024, arXiv:2404.14719v1. [Google Scholar] [CrossRef]
- Tang, G.; Yang, L.; Ren, S.; Meng, L.; Yang, F.; Wang, H. An Automatic Source Code Vulnerability Detection Approach Based on KELM. Secur. Commun. Netw. 2021, 2021, 5566423. [Google Scholar] [CrossRef]
- Capuano, N.; Fenza, G.; Loia, V.; Stanzione, C. Explainable Artificial Intelligence in CyberSecurity: A Survey. IEEE Access 2022, 10, 93575–93600. [Google Scholar] [CrossRef]
- Docker. Available online: https://docs.docker.com/ (accessed on 25 August 2025).
- Kubernetes. Available online: https://kubernetes.io/ (accessed on 25 August 2025).
- Splunk: Splunk Inc. Splunk Enterprise Security. Available online: https://www.splunk.com (accessed on 25 August 2025).
- Huckelberry, J.; Zhang, Y.; Sansone, A.; Mickens, J.; Beerel, P. Vijay Janapa Reddi TinyML Security: Exploring Vulnerabilities in Resource-Constrained Machine Learning Systems. arXiv 2024, arXiv:2411.07114. [Google Scholar]
- Macas, M.; Wu, C.; Fuertes, W. A Survey on Deep Learning for Cybersecurity: Progress, Challenges, and Opportunities. Comput. Netw. 2022, 212, 109032. [Google Scholar] [CrossRef]
- Juliet Test Suite. Available online: https://samate.nist.gov/SARD/test-suites/112 (accessed on 25 August 2025).
- NVD (National Vulnerability Database). Available online: https://nvd.nist.gov/ (accessed on 25 August 2025).
- GitLab CI/CD. Available online: https://docs.gitlab.com/ee/ci/ (accessed on 25 August 2025).
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Chavan, S.; Pappas, G. From Detection to Decision: Transforming Cybersecurity with Deep Learning and Visual Analytics. AI 2025, 6, 214. https://doi.org/10.3390/ai6090214
Chavan S, Pappas G. From Detection to Decision: Transforming Cybersecurity with Deep Learning and Visual Analytics. AI. 2025; 6(9):214. https://doi.org/10.3390/ai6090214
Chicago/Turabian StyleChavan, Saurabh, and George Pappas. 2025. "From Detection to Decision: Transforming Cybersecurity with Deep Learning and Visual Analytics" AI 6, no. 9: 214. https://doi.org/10.3390/ai6090214
APA StyleChavan, S., & Pappas, G. (2025). From Detection to Decision: Transforming Cybersecurity with Deep Learning and Visual Analytics. AI, 6(9), 214. https://doi.org/10.3390/ai6090214