Application of Generative AI in Financial Risk Prediction: Enhancing Model Accuracy and Interpretability
Abstract
1. Introduction
2. Literature Review
2.1. Applications of Generative AI Models in Financial Risk Forecasting
2.2. Data Augmentation with Generative AI
3. Methodology
3.1. Data Preprocessing and Feature Engineering
3.2. Construction of a GAN-Based Financial Risk Prediction Model
3.2.1. Model Architecture
- 1.
- Generator:
- 2.
- Discriminator:
3.2.2. Loss Function
3.2.3. Training Procedure
3.3. Generative Adversarial Explanation (GAX) Framework
4. Results
4.1. Credit Risk Assessment Based on GANs
4.1.1. Experimental Design
- Dataset: The dataset consists of 10,000 small and medium enterprise (SME) loan records from a financial institution, with a 5% default rate (500 default samples), resulting in severe class imbalance.
- Data Preprocessing: Missing values were imputed using the median, outliers detected via Isolation Forest, and all features standardized using Z-score normalization.
- Model Architecture: The model employs an LSTM-based generator and a CNN-based discriminator as described in Section 3.2.
4.1.2. Experimental Results
4.2. Fraud Detection Using Large Language Models (LLMs)
4.2.1. Experimental Design
- Dataset: The dataset comprises 1,000,000 credit card transaction records from a commercial bank, including 0.1% fraudulent transactions (1000 samples) and associated transaction description texts.
- Model Selection: A pretrained BERT model (Devlin et al., 2018) was fine-tuned on the fraud detection dataset [18].
- Feature Extraction: Semantic features were extracted from transaction descriptions using BERT (BASE version), resulting in 768-dimensional embeddings. These were combined with structured features such as transaction amount and time.
- Model Architecture: BERT output was fed into a fully connected layer (256 units) followed by a sigmoid activation function for binary classification.
4.2.2. Experimental Results
- 1.
- Model Performance:
- 2.
- Interpretability Analysis: SHAP values were used to interpret model predictions. Key terms in the transaction descriptions—such as “suspicious” and “unusual”—were identified as significant contributors to fraud detection.
4.3. Market Risk Early Warning Using Generative AI
4.3.1. Experimental Design
- Dataset: A 10-year historical stock market dataset was used, including daily trading volume, price volatility, and news article texts.
- Model Selection: A hybrid approach combining GANs and LLMs was adopted—GANs were employed to simulate market scenarios, while LLMs were used to analyze market sentiment.
- Market Scenario Generation: GANs were trained to generate synthetic market conditions under extreme scenarios for stress testing purposes.
- Sentiment Analysis: LLMs performed sentiment analysis on news texts to quantify market sentiment indicators.
4.3.2. Experimental Results
- Scenario Generation: The GAN-generated market conditions closely resembled historical crash events, providing valuable input for portfolio stress testing.
- Sentiment scores extracted by LLMs showed a strong correlation with market volatility (Pearson r = 0.75), as reported in Figure 9.
4.4. Model Interpretability and Validation
5. Discussion
5.1. Data Quality and Security
5.2. Training Costs and Deployment Challenges
5.3. Regulatory Compliance and Ethical Concerns
6. Conclusions and Suggestions
6.1. Conclusions
6.2. Suggestions for Future Research
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Campbell, J.Y.; Lo, A.W.; MacKinlay, A.C.; Whitelaw, R.F. The econometrics of financial markets. Macroecon. Dyn. 1998, 2, 559–562. [Google Scholar] [CrossRef]
- Lo, A. Adaptive Markets: Financial Evolution at the Speed of Thought; Princeton University Press: Princeton, NJ, USA, 2017. [Google Scholar] [CrossRef]
- Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. Adv. Neural Inf. Process. Syst. 2014, 27. Available online: https://proceedings.neurips.cc/paper_files/paper/2014/file/f033ed80deb0234979a61f95710dbe25-Paper.pdf (accessed on 8 August 2025).
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30. Available online: https://proceedings.neurips.cc/paper_files/paper/2017/file/3f5ee243547dee91fbd053c1c4a845aa-Paper.pdf (accessed on 8 August 2025).
- Li, J.; Liu, H.; Yang, Z.; Han, L. A credit risk model with small sample data based on G-XGBoost. Appl. Artif. Intell. 2021, 35, 1550–1566. [Google Scholar] [CrossRef]
- Lei, Y.T.; Ma, C.Q.; Ren, Y.S.; Chen, X.Q.; Narayan, S.; Huynh, A.N.Q. A distributed deep neural network model for credit card fraud detection. Financ. Res. Lett. 2023, 58, 104547. [Google Scholar] [CrossRef]
- Hernandez Aros, L.; Bustamante Molano, L.X.; Gutierrez-Portela, F.; Moreno Hernandez, J.J.; Rodríguez Barrero, M.S. Financial fraud detection through the application of machine learning techniques: A literature review. Humanit. Soc. Sci. Commun. 2024, 11, 1130. [Google Scholar] [CrossRef]
- EBA. Discussion Paper on Machine Learning for IRB Models. 2021. Available online: https://reurl.cc/Rk7zRr (accessed on 8 August 2025).
- Bank of England & Financial Conduct Authority. FS2/23—Artificial Intelligence and Machine Learning: Feedback Statement. 2023. Available online: https://reurl.cc/oYb7vj (accessed on 8 August 2025).
- Financial Stability Institute (BIS). Regulating AI in the Financial Sector: Recent Developments and Supervisory Implications. FSI Insights No. 63. 2024. Available online: https://www.bis.org/fsi/publ/insights63.pdf?utm_source=chatgpt.com (accessed on 8 August 2025).
- Choi, E.; Biswal, S.; Malin, B.; Duke, J.; Stewart, W.F.; Sun, J. Generating multi-label discrete patient records using generative adversarial networks. In Proceedings of the Machine Learning for Healthcare Conference 2017, Boston, MA, USA, 18–19 August 2017; pp. 286–305. Available online: https://proceedings.mlr.press/v68/choi17a (accessed on 8 August 2025).
- Mirza, M.; Osindero, S. Conditional generative adversarial nets. arXiv 2014, arXiv:1411.1784. [Google Scholar] [CrossRef]
- Wang, C.; Xiao, Z. A deep learning approach for credit scoring using feature embedded transformer. Appl. Sci. 2022, 12, 10995. [Google Scholar] [CrossRef]
- Aljunaid, S.K.; Almheiri, S.J.; Dawood, H.; Khan, M.A. Secure and Transparent Banking: Explainable AI-Driven Federated Learning Model for Financial Fraud Detection. J. Risk Financ. Manag. 2025, 18, 179. [Google Scholar] [CrossRef]
- Zhu, S.; Ma, T.; Wu, H.; Ren, J.; He, D.; Li, Y.; Ge, R. Expanding and Interpreting Financial Statement Fraud Detection Using Supply Chain Knowledge Graphs. J. Theor. Appl. Electron. Commer. Res. 2025, 20, 26. [Google Scholar] [CrossRef]
- Adiputra, I.N.M.; Lin, P.-C.; Wanchai, P. The Effectiveness of Generative Adversarial Network-Based Oversampling Methods for Imbalanced Multi-Class Credit Score Classification. Electronics 2025, 14, 697. [Google Scholar] [CrossRef]
- Hinton, G.E.; Salakhutdinov, R.R. Reducing the dimensionality of data with neural networks. Science 2006, 313, 504507. [Google Scholar] [CrossRef]
- Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), Minneapolis, MN, USA, 2–7 June 2019; pp. 4171–4186. [Google Scholar] [CrossRef]
- Brown, T.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.D.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A. Language models are few-shot learners. Adv. Neural Inf. Process. Syst. 2020, 33, 1877–1901. [Google Scholar]
- Abdulla, Y.Y.; Al-Alawi, A.I. Advances in Machine Learning for Financial Risk Management: A Systematic Literature Review. In Proceedings of the 2024 ASU International Conference in Emerging Technologies for Sustainability and Intelligent Systems (ICETSIS 2024), Manama, Bahrain, 28–29 January 2024; IEEE: New York, NY, USA, 2024; pp. 531–535. [Google Scholar] [CrossRef]
- Gulrajani, I.; Ahmed, F.; Arjovsky, M.; Dumoulin, V.; Courville, A.C. Improved training of wasserstein gans. In Proceedings of the Advances in Neural Information Processing Systems 30, Long Beach, CA, USA, 4–9 December 2017; Available online: https://proceedings.neurips.cc/paper/2017/hash/892c3b1c6dccd52936e27cbd0ff683d6-Abstract.html (accessed on 8 August 2025).
- Szegedy, C.; Zaremba, W.; Sutskever, I.; Bruna, J.; Erhan, D.; Goodfellow, I.; Fergus, R. Intriguing properties of neural networks. arXiv 2013. [Google Scholar] [CrossRef]
- Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
- Little, R.J.; Rubin, D.B. Statistical Analysis with Missing Data; John Wiley & Sons: Hoboken, NJ, USA, 2019; Available online: https://reurl.cc/Y35eDn (accessed on 8 August 2025).
- Chandola, V.; Banerjee, A.; Kumar, V. Anomaly detection: A survey. ACM Comput. Surv. (CSUR) 2009, 41, 1–58. [Google Scholar] [CrossRef]
- Bishop, C.M. Pattern Recognition and Machine Learning; Springer: New York, NY, USA, 2006; Available online: https://link.springer.com/9780387310732 (accessed on 8 August 2025).
- James, G.; Witten, D.; Hastie, T.; Tibshirani, R. An Introduction to Statistical Learning: With Applications in R; Springer: New York, NY, USA, 2013. [Google Scholar] [CrossRef]
- Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
- Hastie, T.; Tibshirani, R.; Friedman, J.; Franklin, J. The elements of statistical learning: Data mining, inference and prediction. Math. Intell. 2005, 27, 83–85. Available online: https://cir.nii.ac.jp/crid/1370846644385113871 (accessed on 8 August 2025). [CrossRef]
- Kingma, D.P. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
- Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the International Conference on Machine Learning 2015, Lille, France, 6–11 July 2015; pp. 448–456. Available online: https://proceedings.mlr.press/v37/ioffe15.html (accessed on 8 August 2025).
- Goodfellow, I.; Bengio, Y.; Courville, A.; Bengio, Y. Deep Learning; MIT Press: Cambridge, MA, USA, 2016; Available online: https://synapse.koreamed.org/pdf/10.4258/hir.2016.22.4.351 (accessed on 8 August 2025).
- Chollet, F. Deep Learning with Python; Simon and Schuster: New York, NY, USA, 2021; Available online: https://reurl.cc/axbVXZ (accessed on 8 August 2025).
- Dwork, C.; Roth, A. The algorithmic foundations of differential privacy. Found. Trends® Theor. Comput. Sci. 2014, 9, 211–407. [Google Scholar] [CrossRef]
- Shostack, A. Threat Modeling: Designing for Security; John Wiley & Sons: Hoboken, NJ, USA, 2014; Available online: https://reurl.cc/9nA4WV (accessed on 8 August 2025).
- Hinton, G.; Vinyals, O.; Dean, J. Distilling the knowledge in a neural network. arXiv 2015. [Google Scholar] [CrossRef]
- Dean, J.; Corrado, G.; Monga, R.; Chen, K.; Devin, M.; Mao, M.; Ranzato, M.; Senior, A.; Tucker, P.; Yang, K.; et al. Large scale distributed deep networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1223–1231. Available online: https://reurl.cc/ekbDpx (accessed on 8 August 2025).
- Bengio, Y.; Courville, A.; Vincent, P. Representation learning: A review and new perspectives. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 1798–1828. [Google Scholar] [CrossRef]
- Engstrom, D.F.; Ho, D.E. Algorithmic accountability in the administrative state. Yale J. Reg. 2020, 37, 800. Available online: https://heinonline.org/HOL/LandingPage?handle=hein.journals/yjor37&div=22&id=&page= (accessed on 8 August 2025).
- Ribeiro, M.T.; Singh, S.; Guestrin, C. “ Why should i trust you?” Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 1135–1144. [Google Scholar] [CrossRef]
- Rudin, C. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nat. Mach. Intell. 2019, 1, 206–215. [Google Scholar] [CrossRef]
- Molnar, C. Interpretable Machine Learning-a Guide for Making Black Box Models Explainable. 2020. Available online: https://christophm.github.io/interpretable-ml-book/ (accessed on 8 August 2025).
- Mehrabi, N.; Morstatter, F.; Saxena, N.; Lerman, K.; Galstyan, A. A survey on bias and fairness in machine learning. ACM Comput. Surv. (CSUR) 2021, 54, 1–35. [Google Scholar] [CrossRef]
- Nakamoto, S. Bitcoin: A Peer-to-Peer Electronic Cash System. Available at SSRN 3440802. 2008. Available online: https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3440802 (accessed on 8 August 2025).
- McMahan, B.; Moore, E.; Ramage, D.; Hampson, S.; y Arcas, B.A. Communication-efficient learning of deep networks from decentralized data. In Proceedings of the 20th International Conference on Artificial Intelligence and Statistics, Fort Lauderdale, FL, USA, 20–22 April 2017; pp. 1273–1282. Available online: https://proceedings.mlr.press/v54/mcmahan17a?ref=https://githubhelp.com (accessed on 8 August 2025).
- Markowitz, H.M. Portfolio selection: Efficient Diversification of Investments; Yale University Press: New Haven, CT, USA, 2008; Available online: https://reurl.cc/Mzm8Mp (accessed on 8 August 2025).
- De Prado, M.L. Advances in Financial Machine Learning; John Wiley & Sons: Hoboken, NJ, USA, 2018; Available online: https://reurl.cc/GNqAGy (accessed on 8 August 2025).













| Method | AUC | F1 | Precision | Recall |
|---|---|---|---|---|
| Baseline (no augmentation) | 0.82 | 0.45 | 0.52 | 0.4 |
| SMOTE | 0.85 | 0.48 | 0.55 | 0.43 |
| ADASYN | 0.86 | 0.51 | 0.58 | 0.46 |
| Class-weighted XGBoost | 0.87 | 0.54 | 0.61 | 0.49 |
| Focal Loss (γ = 2) | 0.87 | 0.55 | 0.6 | 0.51 |
| Cost-sensitive Learning | 0.86 | 0.53 | 0.59 | 0.48 |
| Our GAN Approach | 0.88 | 0.58 | 0.63 | 0.54 |
| ε Value | δ | Model AUC | Data Utility | Privacy Budget |
|---|---|---|---|---|
| ∞ (no DP) | – | 0.88 | 100% | 0% |
| 10 | 10−5 | 0.875 | 98.20% | Low |
| 5 | 10−5 | 0.862 | 94.10% | Medium |
| 1 | 10−5 | 0.831 | 87.30% | High |
| 0.1 | 10−5 | 0.754 | 71.20% | Maximum |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Yao, K.-C.; Hung, H.-C.; Wang, C.-H.; Huang, W.-L.; Liang, H.-T.; Chu, T.-H.; Chen, B.-S.; Ho, W.-S. Application of Generative AI in Financial Risk Prediction: Enhancing Model Accuracy and Interpretability. Information 2025, 16, 857. https://doi.org/10.3390/info16100857
Yao K-C, Hung H-C, Wang C-H, Huang W-L, Liang H-T, Chu T-H, Chen B-S, Ho W-S. Application of Generative AI in Financial Risk Prediction: Enhancing Model Accuracy and Interpretability. Information. 2025; 16(10):857. https://doi.org/10.3390/info16100857
Chicago/Turabian StyleYao, Kai-Chao, Hsiu-Chu Hung, Ching-Hsin Wang, Wei-Lun Huang, Hui-Ting Liang, Tzu-Hsin Chu, Bo-Siang Chen, and Wei-Sho Ho. 2025. "Application of Generative AI in Financial Risk Prediction: Enhancing Model Accuracy and Interpretability" Information 16, no. 10: 857. https://doi.org/10.3390/info16100857
APA StyleYao, K.-C., Hung, H.-C., Wang, C.-H., Huang, W.-L., Liang, H.-T., Chu, T.-H., Chen, B.-S., & Ho, W.-S. (2025). Application of Generative AI in Financial Risk Prediction: Enhancing Model Accuracy and Interpretability. Information, 16(10), 857. https://doi.org/10.3390/info16100857

