Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (507)

Search Parameters:
Keywords = computational financial modeling

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
25 pages, 1321 KB  
Article
The Stingray Copula for Negative Dependence
by Alecos Papadopoulos
Stats 2026, 9(1), 13; https://doi.org/10.3390/stats9010013 - 4 Feb 2026
Abstract
We present a new single-parameter bivariate copula, called the Stingray, that is dedicated to representing negative dependence, and it nests the Independence copula. The Stingray copula is generated in a relatively novel way; it has a simple form and is always defined over [...] Read more.
We present a new single-parameter bivariate copula, called the Stingray, that is dedicated to representing negative dependence, and it nests the Independence copula. The Stingray copula is generated in a relatively novel way; it has a simple form and is always defined over the full support, unlike many copulas that model negative dependence. We provide visualizations of the copula, derive several dependence properties, and compute basic concordance measures. We compare it with other copulas and joint distributions with respect to the extent of dependence it can capture, and we find that the Stingray copula outperforms most of them while remaining competitive with well-known, widely used copulas such as the Gaussian and Frank copulas. Moreover, we show, through simulation, that the dependence structure it represents cannot be fully captured by these copulas, as it is asymmetric. We also show how the non-parametric Spearman’s rho measure of concordance can be used to formally test the hypothesis of statistical independence. As an illustration, we apply it to a financial data sample from the building construction sector in order to model the negative relationship between the level of capital employed and its gross rate of return. Full article
Show Figures

Figure 1

30 pages, 616 KB  
Article
Structural Preservation in Time Series Through Multiscale Topological Features Derived from Persistent Homology
by Luiz Carlos de Jesus, Francisco Fernández-Navarro and Mariano Carbonero-Ruz
Mathematics 2026, 14(3), 538; https://doi.org/10.3390/math14030538 - 2 Feb 2026
Abstract
A principled, model-agnostic framework for structural feature extraction in time series is presented, grounded in topological data analysis (TDA). The motivation stems from two gaps identified in the literature: First, compact and interpretable representations that summarise the global geometric organisation of trajectories across [...] Read more.
A principled, model-agnostic framework for structural feature extraction in time series is presented, grounded in topological data analysis (TDA). The motivation stems from two gaps identified in the literature: First, compact and interpretable representations that summarise the global geometric organisation of trajectories across scales remain scarce. Second, a unified, task-agnostic protocol for evaluating structure preservation against established non-topological families is still missing. To address these gaps, time-delay embeddings are employed to reconstruct phase space, sliding windows are used to generate local point clouds, and Vietoris–Rips persistent homology (up to dimension two) is computed. The resulting persistence diagrams are summarised with three transparent descriptors—persistence entropy, maximum persistence amplitude, and feature counts—and concatenated across delays and window sizes to yield a multiscale representation designed to complement temporal and spectral features while remaining computationally tractable. A unified experimental design is specified in which heterogeneous, regularly sampled financial series are preprocessed on native calendars and contrasted with competitive baselines spanning lagged, calendar-driven, difference/change, STL-based, delay-embedding PCA, price-based statistical, signature (FRUITS), and network-derived (NetF) features. Structure preservation is assessed through complementary criteria that probe spectral similarity, variance-scaled reconstruction fidelity, and the conservation of distributional shape (location, scale, asymmetry, tails). The study is positioned as an evaluation of representations, rather than a forecasting benchmark, emphasising interpretability, comparability, and methodological transparency while outlining avenues for adaptive hyperparameter selection and alternative filtrations. Full article
Show Figures

Figure 1

39 pages, 3699 KB  
Article
Enhancing Decision Intelligence Using Hybrid Machine Learning Framework with Linear Programming for Enterprise Project Selection and Portfolio Optimization
by Abdullah, Nida Hafeez, Carlos Guzmán Sánchez-Mejorada, Miguel Jesús Torres Ruiz, Rolando Quintero Téllez, Eponon Anvi Alex, Grigori Sidorov and Alexander Gelbukh
AI 2026, 7(2), 52; https://doi.org/10.3390/ai7020052 - 1 Feb 2026
Viewed by 190
Abstract
This study presents a hybrid analytical framework that enhances project selection by achieving reasonable predictive accuracy through the integration of expert judgment and modern artificial intelligence (AI) techniques. Using an enterprise-level dataset of 10,000 completed software projects with verified real-world statistical characteristics, we [...] Read more.
This study presents a hybrid analytical framework that enhances project selection by achieving reasonable predictive accuracy through the integration of expert judgment and modern artificial intelligence (AI) techniques. Using an enterprise-level dataset of 10,000 completed software projects with verified real-world statistical characteristics, we develop a three-step architecture for intelligent decision support. First, we introduce an extended Analytic Hierarchy Process (AHP) that incorporates organizational learning patterns to compute expert-validated criteria weights with a consistent level of reliability (CR=0.04), and Linear Programming is used for portfolio optimization. Second, we propose a machine learning architecture that integrates expert knowledge derived from AHP into models such as Transformers, TabNet, and Neural Oblivious Decision Ensembles through mechanisms including attention modulation, split criterion weighting, and differentiable tree regularization. Third, the hybrid AHP-Stacking classifier generates a meta-ensemble that adaptively balances expert-derived information with data-driven patterns. The analysis shows that the model achieves 97.5% accuracy, a 96.9% F1-score, and a 0.989 AUC-ROC, representing a 25% improvement compared to baseline methods. The framework also indicates a projected 68.2% improvement in portfolio value (estimated incremental value of USD 83.5 M) based on post factum financial results from the enterprise’s ventures.This study is evaluated retrospectively using data from a single enterprise, and while the results demonstrate strong robustness, generalizability to other organizational contexts requires further validation. This research contributes a structured approach to hybrid intelligent systems and demonstrates that combining expert knowledge with machine learning can provide reliable, transparent, and high-performing decision-support capabilities for project portfolio management. Full article
Show Figures

Figure 1

34 pages, 459 KB  
Article
Comparative Analysis and Optimisation of Machine Learning Models for Regression and Classification on Structured Tabular Datasets
by Siegfried Fredrich Stumpfe and Sandile Charles Shongwe
Mathematics 2026, 14(3), 473; https://doi.org/10.3390/math14030473 - 29 Jan 2026
Viewed by 144
Abstract
This research entails comparative analysis and optimisation of machine learning models for regression and classification tasks on structured tabular datasets. The primary target audience for this analysis comprises researchers and practitioners working with structured tabular data. Common fields include biostatistics, insurance, and financial [...] Read more.
This research entails comparative analysis and optimisation of machine learning models for regression and classification tasks on structured tabular datasets. The primary target audience for this analysis comprises researchers and practitioners working with structured tabular data. Common fields include biostatistics, insurance, and financial risk modelling, where computational efficiency and robust predictive performance are essential. Four machine learning techniques (i.e., linear/logistic regression, support vector machines (SVMs), Extreme Gradient Boosting (XGBoost), and Multi-Layered Perceptrons (MLPs)) were applied across 72 datasets sourced from OpenML and Kaggle. The datasets systematically varied by observation size, dimensionality, noise levels, linearity, and class balance. Based on extensive empirical analysis (72 datasets ×4 models ×2 configurations =576 experiments), it is observed that, understanding the dataset characteristics is more critical than extensive hyperparameter tuning for optimal model performance. Also, linear models are robust across various settings, while non-linear models, like XGBoost and MLP, perform better in complex and noisy environments. In general, this study provides valuable insights for model selection and benchmarking in machine learning applications that involve structured tabular datasets. Full article
(This article belongs to the Special Issue Computational Statistics: Analysis and Applications for Mathematics)
21 pages, 6506 KB  
Article
Strategic Energy Project Investment Decisions Using RoBERTa: A Framework for Efficient Infrastructure Evaluation
by Recep Özkan, Fatemeh Mostofi, Fethi Kadıoğlu, Vedat Toğan and Onur Behzat Tokdemir
Buildings 2026, 16(3), 547; https://doi.org/10.3390/buildings16030547 - 28 Jan 2026
Viewed by 243
Abstract
The task of identifying high-value projects from vast investment portfolios presents a major challenge in the construction industry, particularly within the energy sector, where decision-making carries high financial and operational stakes. This complexity is driven by both the volume and heterogeneity of project [...] Read more.
The task of identifying high-value projects from vast investment portfolios presents a major challenge in the construction industry, particularly within the energy sector, where decision-making carries high financial and operational stakes. This complexity is driven by both the volume and heterogeneity of project documentation, as well as the multidimensional criteria used to assess project value. Despite this, research gaps remain: large language models (LLMs) as pretrained transformer encoder models are underutilized in construction project selection, especially in domains where investment precision is paramount. Existing methodologies have largely focused on multi-criteria decision-making (MCDM) frameworks, often neglecting the potential of LLMs to automate and enhance early-phase project evaluation. However, deploying LLMs for such tasks introduces high computational demands, particularly in privacy-sensitive, enterprise-level environments. This study investigates the application of the robustly optimized BERT model (RoBERTa) for identifying high-value energy infrastructure projects. Our dual objective is to (1) leverage RoBERTa’s pre-trained language architecture to extract key information from unstructured investment texts and (2) evaluate its effectiveness in enhancing project selection accuracy. We benchmark RoBERTa against several leading LLMs: BERT, DistilBERT (a distilled variant), ALBERT (a lightweight version), and XLNet (a generalized autoregressive model). All models achieved over 98% accuracy, validating their utility in this domain. RoBERTa outperformed its counterparts with an accuracy of 99.6%. DistilBERT was fastest (1025.17 s), while RoBERTa took 2060.29 s. XLNet was slowest at 4145.49 s. In conclusion, RoBERTa can be the preferred option when maximum accuracy is required, while DistilBERT can be a viable alternative under computational or resource constraints. Full article
(This article belongs to the Section Construction Management, and Computers & Digitization)
Show Figures

Figure 1

19 pages, 376 KB  
Article
Multi-Platform Multivariate Regression with Group Sparsity for High-Dimensional Data Integration
by Shanshan Qin, Guanlin Zhang, Xin Gao and Yuehua Wu
Entropy 2026, 28(2), 135; https://doi.org/10.3390/e28020135 - 23 Jan 2026
Viewed by 185
Abstract
High-dimensional regression with multivariate responses poses significant challenges when data are collected across multiple platforms, each with potentially correlated outcomes. In this paper, we introduce a multi-platform multivariate high-dimensional linear regression (MM-HLR) model for simultaneously modeling within-platform correlation and cross-platform information fusion. Our [...] Read more.
High-dimensional regression with multivariate responses poses significant challenges when data are collected across multiple platforms, each with potentially correlated outcomes. In this paper, we introduce a multi-platform multivariate high-dimensional linear regression (MM-HLR) model for simultaneously modeling within-platform correlation and cross-platform information fusion. Our approach incorporates a mixture of Lasso and group Lasso penalties to promote both individual predictor sparsity and cross-platform group sparsity, thereby enhancing interpretability and estimation stability. We develop an efficient computational algorithm based on iteratively reweighted least squares and block coordinate descent to solve the resulting regularized optimization problem. We establish theoretical guarantees for our estimator, including oracle bounds on prediction error, estimation accuracy, and support recovery under mild conditions. Our simulation studies confirm the method’s strong empirical performance, demonstrating low bias, small variance, and robustness across various dimensions. The analysis of real financial data further validates the performance gains achieved by incorporating multivariate responses and integrating data across multiple platforms. Full article
35 pages, 7197 KB  
Article
Assessing the Sustainable Synergy Between Digitalization and Decarbonization in the Coal Power Industry: A Fuzzy DEMATEL-MultiMOORA-Borda Framework
by Yubao Wang and Zhenzhong Liu
Sustainability 2026, 18(3), 1160; https://doi.org/10.3390/su18031160 - 23 Jan 2026
Viewed by 117
Abstract
In the context of the “Dual Carbon” goals, achieving synergistic development between digitalization and green transformation in the coal power industry is essential for ensuring a just and sustainable energy transition. The core scientific problem addressed is the lack of a robust quantitative [...] Read more.
In the context of the “Dual Carbon” goals, achieving synergistic development between digitalization and green transformation in the coal power industry is essential for ensuring a just and sustainable energy transition. The core scientific problem addressed is the lack of a robust quantitative tool to evaluate the comprehensive performance of diverse transition scenarios in a complex environment characterized by multi-objective trade-offs and high uncertainty. This study establishes a sustainability-oriented four-dimensional performance evaluation system encompassing 22 indicators, covering Synergistic Economic Performance, Green-Digital Strategy, Synergistic Governance, and Technology Performance. Based on this framework, a Fuzzy DEMATEL–MultiMOORA–Borda integrated decision model is proposed to evaluate seven transition scenarios. The computational framework utilizes the Interval Type-2 Fuzzy DEMATEL (IT2FS-DEMATEL) method for robust causal analysis and weight determination, addressing the inherent subjectivity and vagueness in expert judgments. The model integrates MultiMOORA with Borda Count aggregation for enhanced ranking stability. All model calculations were implemented using Matlab R2022a. Results reveal that Carbon Price and Digital Hedging Capability (C13) and Digital-Driven Operational Efficiency (C43) are the primary drivers of synergistic performance. Among the scenarios, P3 (Digital Twin Empowerment and New Energy Co-integration) achieves the best overall performance (score: 0.5641), representing the most viable pathway for balancing industrial efficiency and environmental stewardship. Robustness tests demonstrate that the proposed model significantly outperforms conventional approaches such as Fuzzy AHP (Analytic Hierarchy Process) and TOPSIS under weight perturbations. Sensitivity analysis further identifies Financial Return (C44) and Green Transformation Marginal Economy (C11) as critical factors for long-term policy effectiveness. This study provides a data-driven framework and a robust decision-support tool for advancing the coal power industry’s low-carbon, intelligent, and resilient transition in alignment with global sustainability targets. Full article
Show Figures

Figure 1

29 pages, 6210 KB  
Article
Assessing Economic Vulnerability from Urban Flooding: A Case Study of Catu, a Commerce-Based City in Brazil
by Lais Das Neves Santana, Alarcon Matos de Oliveira, Lusanira Nogueira Aragão de Oliveira and Fabricio Ribeiro Garcia
Water 2026, 18(2), 282; https://doi.org/10.3390/w18020282 - 22 Jan 2026
Viewed by 207
Abstract
Flooding is a recurrent problem in many Brazilian cities, resulting in significant losses that affect health, assets, finance, and the environment. The uncertainty regarding extreme rainfall events due to climate change makes this challenge even more severe, compounded by inadequate urban planning and [...] Read more.
Flooding is a recurrent problem in many Brazilian cities, resulting in significant losses that affect health, assets, finance, and the environment. The uncertainty regarding extreme rainfall events due to climate change makes this challenge even more severe, compounded by inadequate urban planning and the occupation of risk areas, particularly for the municipality of Catu, in the state of Bahia, which also suffers from recurrent floods. Critical hotspots include the Santa Rita neighborhood and its surroundings, the main supply center, and the city center—the municipality’s commercial hub. The focus of this research is the unprecedented quantification of the socioeconomic impact of these floods on the low-income population and the region’s informal sector (street vendors). This research focused on analyzing and modeling the destructive potential of intense rainfall in the Santa Rita region (Supply Center) of Catu, Bahia, and its effects on the local economy across different recurrence intervals. A hydrological simulation software suite based on computational and geoprocessing technologies—specifically HEC-RAS 6.4, HEC-HMS 4.11, and QGIS— 3.16 was utilized. Two-dimensional (2D) modeling was applied to assess the flood-prone areas. For the socioeconomic impact assessment, a loss procedure based on linear regression was developed, which correlated the different return periods of extreme events with the potential losses. This methodology, which utilizes validated, indirect data, establishes a replicable framework adaptable to other regions facing similar socioeconomic and drainage challenges. The results revealed that the area becomes impassable during flood events, preventing commercial activities and causing significant economic losses, particularly for local market vendors. The total financial damage for the 100-year extreme event is approximately US $30,000, with the loss model achieving an R2 of 0.98. The research concludes that urgent measures are necessary to mitigate flood impacts, particularly as climate change reduces the return period of extreme events. The implementation of adequate infrastructure, informed by the presented risk modeling, and public awareness are essential for reducing vulnerability. Full article
(This article belongs to the Special Issue Water-Soil-Vegetation Interactions in Changing Climate)
Show Figures

Figure 1

15 pages, 323 KB  
Article
Assessing the Link Between the Misery Index and Dollarization: Regional Evidence from Türkiye
by Gökhan Özkul and İbrahim Yaşar Gök
J. Risk Financial Manag. 2026, 19(1), 93; https://doi.org/10.3390/jrfm19010093 - 22 Jan 2026
Viewed by 109
Abstract
This study analyzes the relationship between macroeconomic distress and financial dollarization in Türkiye using annual regional panel data for 26 Nomenclature of Territorial Units for Statistics 2 regions over the period 2005–2021. Macroeconomic distress is captured using the misery index, computed as the [...] Read more.
This study analyzes the relationship between macroeconomic distress and financial dollarization in Türkiye using annual regional panel data for 26 Nomenclature of Territorial Units for Statistics 2 regions over the period 2005–2021. Macroeconomic distress is captured using the misery index, computed as the compound of inflation and unemployment rates, while the share of foreign-currency-denominated deposits in total deposits measures financial dollarization. Applying second-generation panel econometric models that account for regional heterogeneity, we investigate both long-run equilibrium relationships and short-run interactions. Panel cointegration tests show a long-run connection between macroeconomic distress and dollarization. Short-run effects estimated using a Panel Vector Error Correction Model and a Cross-Sectionally Augmented ARDL framework point to bidirectional causality. Long-run coefficient estimates obtained via Dynamic Ordinary Least Squares indicate an apparent asymmetry. Increases in dollarization exert a substantial and economically significant effect on macroeconomic distress, whereas the long-run impact of distress on dollarization is comparatively modest. The findings suggest that dollarization functions not only as a response to macroeconomic instability but also as a structural element that intensifies inflationary pressures and labor market distortions over time. Focusing on regional patterns rather than national aggregates, the paper provides new evidence on the spatial dimension of the dollarization–instability link. Full article
(This article belongs to the Section Financial Markets)
26 pages, 2118 KB  
Article
A Hybrid HAR-LSTM-GARCH Model for Forecasting Volatility in Energy Markets
by Wiem Ben Romdhane and Heni Boubaker
J. Risk Financial Manag. 2026, 19(1), 77; https://doi.org/10.3390/jrfm19010077 - 17 Jan 2026
Viewed by 533
Abstract
Accurate volatility forecasting in energy markets is paramount for risk management, derivative pricing, and strategic policy planning. Traditional econometric models like the Heterogeneous Auto-regressive (HAR) model effectively capture the long-memory and multi-component nature of volatility but often fail to account for non-linearities and [...] Read more.
Accurate volatility forecasting in energy markets is paramount for risk management, derivative pricing, and strategic policy planning. Traditional econometric models like the Heterogeneous Auto-regressive (HAR) model effectively capture the long-memory and multi-component nature of volatility but often fail to account for non-linearities and complex, unseen dependencies. Deep learning models, particularly Long Short-Term Memory (LSTM) networks, excel at capturing these non-linear patterns but can be data-hungry and prone to overfitting, especially in noisy financial datasets. This paper proposes a novel hybrid model, HAR-LSTM-GARCH, which synergistically combines the strengths of the HAR model, an LSTM network, and a GARCH model to forecast the realized volatility of crude oil futures. The HAR component captures the persistent, multi-scale volatility dynamics, the LSTM network learns the non-linear residual patterns, and the GARCH component models the time-varying volatility of the residuals themselves. Using high-frequency data on Brent Crude futures, we compute daily Realized Volatility (RV). Our empirical results demonstrate that the proposed HAR-LSTM-GARCH model significantly outperforms the benchmark HAR, GARCH(1,1), and standalone LSTM models in both statistical accuracy and economic significance, offering a robust framework for volatility forecasting in the complex energy sector. Full article
(This article belongs to the Special Issue Mathematical Modelling in Economics and Finance)
Show Figures

Figure 1

22 pages, 312 KB  
Article
Machine Learning-Enhanced Database Cache Management: A Comprehensive Performance Analysis and Comparison of Predictive Replacement Policies
by Maryam Abbasi, Paulo Váz, José Silva, Filipe Cardoso, Filipe Sá and Pedro Martins
Appl. Sci. 2026, 16(2), 666; https://doi.org/10.3390/app16020666 - 8 Jan 2026
Viewed by 319
Abstract
The exponential growth of data-driven applications has intensified performance demands on database systems, where cache management represents a critical bottleneck. Traditional cache replacement policies such as Least Recently Used (LRU) and Least Frequently Used (LFU) rely on simple heuristics that fail to capture [...] Read more.
The exponential growth of data-driven applications has intensified performance demands on database systems, where cache management represents a critical bottleneck. Traditional cache replacement policies such as Least Recently Used (LRU) and Least Frequently Used (LFU) rely on simple heuristics that fail to capture complex temporal and frequency patterns in modern workloads. This research presents a modular machine learning-enhanced cache management framework that leverages pattern recognition to optimize database performance through intelligent replacement decisions. Our approach integrates multiple machine learning models—Random Forest classifiers, Long Short-Term Memory (LSTM) networks, Support Vector Machines (SVM), and Gradient Boosting methods—within a modular architecture enabling seamless integration with existing database systems. The framework incorporates sophisticated feature engineering pipelines extracting temporal, frequency, and contextual characteristics from query access patterns. Comprehensive experimental evaluation across synthetic workloads, real-world production datasets, and standard benchmarks (TPC-C, TPC-H, YCSB, and LinkBench) demonstrates consistent performance improvements. Machine learning-enhanced approaches achieve 8.4% to 19.2% improvement in cache hit rates, 15.3% to 28.7% reduction in query latency, and 18.9% to 31.4% increase in system throughput compared to traditional policies and advanced adaptive methods including ARC, LIRS, Clock-Pro, TinyLFU, and LECAR. Random Forest emerges as the most practical solution, providing 18.7% performance improvement with only 3.1% computational overhead. Case study analysis across e-commerce, financial services, and content management applications demonstrates measurable business impact, including 8.3% conversion rate improvements and USD 127,000 annual revenue increases. Statistical validation (p<0.001, Cohen’s d>0.8) confirms both statistical and practical significance. Full article
Show Figures

Figure 1

22 pages, 2930 KB  
Article
Developing and Assessing the Performance of a Machine Learning Model for Analyzing Drinking Behaviors in Minipigs for Experimental Research
by Frederik Deutch, Lars Schmidt Hansen, Firas Omar Saleh, Marc Gjern Weiss, Constanca Figueiredo, Cyril Moers, Anna Krarup Keller and Stefan Rahr Wagner
Sensors 2026, 26(2), 402; https://doi.org/10.3390/s26020402 - 8 Jan 2026
Viewed by 294
Abstract
Monitoring experimental animals is essential for ethical, scientific, and financial reasons. Conventional observation methods are limited by subjectivity and time constraints. Camera-based monitoring combined with machine learning offers a promising solution for automating the monitoring process. This study aimed to validate and assess [...] Read more.
Monitoring experimental animals is essential for ethical, scientific, and financial reasons. Conventional observation methods are limited by subjectivity and time constraints. Camera-based monitoring combined with machine learning offers a promising solution for automating the monitoring process. This study aimed to validate and assess the performance of a machine learning model for analyzing drinking behavior in minipigs. A novel, vision-based monitoring system was developed and tested to detect drinking behavior in minipigs. The system, based on low-cost Raspberry Pi units, enabled on-site video analysis. A dataset of 5297 images was used to train a YOLOv11n object detection model to identify key features such as pig heads and water faucets. Drinking events were defined by the spatial proximity of these features within video frames. The multi-class object detection model achieved an accuracy of above 97%. Manual validation using human-annotated ground truth on 72 h of video yielded an overall accuracy of 99.7%, with a precision of 99.7%, recall of 99.2%, and F1-score of 99.5%. Drinking patterns for three pigs were analyzed using 216 h of video. The results revealed a bimodal drinking pattern and substantial inter-pig variability. A limitation to the study was chosen methods missing distinguishment between multiple pigs and the absence of quantification of water intake. This study demonstrates the feasibility of a low-cost, computer vision-based system for monitoring drinking behavior in individually housed experimental pigs, supporting earlier detection of illness. Full article
Show Figures

Figure 1

44 pages, 4883 KB  
Article
Mapping the Role of Artificial Intelligence and Machine Learning in Advancing Sustainable Banking
by Alina Georgiana Manta, Claudia Gherțescu, Roxana Maria Bădîrcea, Liviu Florin Manta, Jenica Popescu and Mihail Olaru
Sustainability 2026, 18(2), 618; https://doi.org/10.3390/su18020618 - 7 Jan 2026
Viewed by 361
Abstract
The convergence of artificial intelligence (AI), machine learning (ML), blockchain, and big data analytics is transforming the governance, sustainability, and resilience of modern banking ecosystems. This study provides a multivariate bibliometric analysis using Principal Component Analysis (PCA) of research indexed in Scopus and [...] Read more.
The convergence of artificial intelligence (AI), machine learning (ML), blockchain, and big data analytics is transforming the governance, sustainability, and resilience of modern banking ecosystems. This study provides a multivariate bibliometric analysis using Principal Component Analysis (PCA) of research indexed in Scopus and Web of Science to explore how decentralized digital infrastructures and AI-driven analytical capabilities contribute to sustainable financial development, transparent governance, and climate-resilient digital societies. Findings indicate a rapid increase in interdisciplinary work integrating Distributed Ledger Technology (DLT) with large-scale data processing, federated learning, privacy-preserving computation, and intelligent automation—tools that can enhance financial inclusion, regulatory integrity, and environmental risk management. Keyword network analyses reveal blockchain’s growing role in improving data provenance, security, and trust—key governance dimensions for sustainable and resilient financial systems—while AI/ML and big data analytics dominate research on predictive intelligence, ESG-related risk modeling, customer well-being analytics, and real-time decision support for sustainable finance. Comparative analyses show distinct emphases: Web of Science highlights decentralized architectures, consensus mechanisms, and smart contracts relevant to transparent financial governance, whereas Scopus emphasizes customer-centered analytics, natural language processing, and high-throughput data environments supporting inclusive and equitable financial services. Patterns of global collaboration demonstrate strong internationalization, with Europe, China, and the United States emerging as key hubs in shaping sustainable and digitally resilient banking infrastructures. By mapping intellectual, technological, and collaborative structures, this study clarifies how decentralized intelligence—enabled by the fusion of AI/ML, blockchain, and big data—supports secure, scalable, and sustainability-driven financial ecosystems. The results identify critical research pathways for strengthening financial governance, enhancing climate and social resilience, and advancing digital transformation, which contributes to more inclusive, equitable, and sustainable societies. Full article
Show Figures

Figure 1

19 pages, 857 KB  
Article
Data-Driven Insights: Leveraging Sentiment Analysis and Latent Profile Analysis for Financial Market Forecasting
by Eyal Eckhaus
Big Data Cogn. Comput. 2026, 10(1), 24; https://doi.org/10.3390/bdcc10010024 - 7 Jan 2026
Viewed by 497
Abstract
Background: This study explores an innovative integration of big data analytics techniques aimed at enhancing predictive modeling in financial markets. It investigates how combining sentiment analysis with latent profile analysis (LPA) can accurately forecast stock prices. This research aligns with big data [...] Read more.
Background: This study explores an innovative integration of big data analytics techniques aimed at enhancing predictive modeling in financial markets. It investigates how combining sentiment analysis with latent profile analysis (LPA) can accurately forecast stock prices. This research aligns with big data methodologies by leveraging automated content analysis and segmentation algorithms to address real-world challenges in data-driven decision-making. This study leverages advanced computational methods to process and segment large-scale unstructured data, demonstrating scalability in data-rich environments. Methods: We compiled a corpus of 3843 financial news articles on Teva Pharmaceuticals from Bloomberg and Reuters. Sentiment scores were generated using the VADER tool, and LPA was applied to identify eight distinct sentiment profiles. These profiles were then used in segmented regression models and Structural Equation Modeling (SEM) to assess their predictive value for stock price fluctuations. Results: Six of the eight latent profiles demonstrated significantly higher predictive accuracy compared to traditional sentiment-based models. The combined profile-based regression model explained 47% of the stock price variance (R2 = 0.47), compared to 10% (R2 = 0.10) in the baseline model using sentiment analysis alone. Conclusion: This study pioneers the use of latent profile analysis (LPA) in sentiment analysis for stock price prediction, offering a novel integration of clustering and financial forecasting. By uncovering complex, non-linear links between market sentiment and stock movements, it addresses a key gap in the literature and establishes a powerful foundation for advancing sentiment-based financial models. Full article
Show Figures

Figure 1

22 pages, 5584 KB  
Article
Design and Evaluation of Machine Learning-Based Investment Strategies in Equity Funds
by Danillo Guimarães Cassiano da Silva, Estaner Claro Romão and Fabiano Fernandes Bargos
Int. J. Financial Stud. 2026, 14(1), 16; https://doi.org/10.3390/ijfs14010016 - 7 Jan 2026
Viewed by 415
Abstract
This study examines quantitative investment strategies for Brazilian equity funds, integrating traditional financial performance indicators with machine learning techniques to enhance fund selection. The main objective was to construct and validate predictive models for fund selection. The methodology involved collecting daily data from [...] Read more.
This study examines quantitative investment strategies for Brazilian equity funds, integrating traditional financial performance indicators with machine learning techniques to enhance fund selection. The main objective was to construct and validate predictive models for fund selection. The methodology involved collecting daily data from 2019 to 2025, computing a range of return and risk measures, and trained models to classify 1- and 3-month shifted windows. The 3-month models achieved the strongest predictive accuracy, exceeding 91%, with the Sharpe Ratio emerging as the most influential feature. A 12-month backtest (October/2024–September/2025) showed that ML-constructed portfolios delivered cumulative returns between 14.65% and 91.86%, depending on the selection criterion, substantially outperforming Brazil’s CDI risk-free benchmark (12.70%) and the Ibovespa (11.46%). These findings highlight the practical potential of ML-based fund selection, though successful implementation requires careful risk management and ongoing model validation. Full article
Show Figures

Figure 1

Back to TopTop