Journal Description
AI
AI
is an international, peer-reviewed, open access journal on artificial intelligence (AI), including broad aspects of cognition and reasoning, perception and planning, machine learning, intelligent robotics, and applications of AI, published monthly online by MDPI.
- Open Access— free for readers, with article processing charges (APC) paid by authors or their institutions.
- High Visibility: indexed within ESCI (Web of Science), Scopus, EBSCO, and other databases.
- Journal Rank: JCR - Q2 (Computer Science, Artificial Intelligence) / CiteScore - Q2 (Artificial Intelligence)
- Rapid Publication: manuscripts are peer-reviewed and a first decision is provided to authors approximately 18.9 days after submission; acceptance to publication is undertaken in 4.9 days (median values for papers published in this journal in the second half of 2024).
- Recognition of Reviewers: APC discount vouchers, optional signed peer review, and reviewer names published annually in the journal.
Impact Factor:
3.1 (2023);
5-Year Impact Factor:
3.3 (2023)
Latest Articles
Automated Pruning Framework for Large Language Models Using Combinatorial Optimization
AI 2025, 6(5), 96; https://doi.org/10.3390/ai6050096 (registering DOI) - 5 May 2025
Abstract
Currently, large language models (LLMs) have been utilized in many aspects of natural language processing. However, due to their significant size and high computational demands, large computational resources are required for deployment. In this research, we focus on the automated approach for size
[...] Read more.
Currently, large language models (LLMs) have been utilized in many aspects of natural language processing. However, due to their significant size and high computational demands, large computational resources are required for deployment. In this research, we focus on the automated approach for size reduction of such a model. We propose the framework to perform the automated pruning based on combinatorial optimization. Two techniques were particularly studied, i.e., particle swarm optimization (PSO) and whale optimization algorithm (WOA). The model pruning problem was modeled as a combinatorial optimization task whose the goal is to minimize model size while maintaining model accuracy. The framework systematically explores the search space to identify the most optimal pruning configurations, removing redundant or non-contributory parameters. The two optimizations, PSO and WOA, were evaluated for their ability to efficiently navigate the search space. As a result, with PSO, the proposed framework can reduce the model size of Llama-3.1-70B by 13.44% while keeping the loss of model accuracy at 19.25%; with WOA, the model size reduction is 12.07% with 22.81% loss of model accuracy. Since accuracy degradation may occur during pruning process, the framework integrates the post-process to recover the model accuracy. After this process, the pruned model loss can reduce to 12.72% and 14.83% using PSO and WOA, respectively.
Full article
(This article belongs to the Section AI Systems: Theory and Applications)
►
Show Figures
Open AccessArticle
Cybersecure XAI Algorithm for Generating Recommendations Based on Financial Fundamentals Using DeepSeek
by
Iván García-Magariño, Javier Bravo-Agapito and Raquel Lacuesta
AI 2025, 6(5), 95; https://doi.org/10.3390/ai6050095 - 2 May 2025
Abstract
Background: Investment decisions in stocks are one of the most complex tasks due to the uncertainty of which stocks will increase or decrease in their values. A diversified portfolio statistically reduces the risk; however, stock choice still substantially influences the profitability. Methods: This
[...] Read more.
Background: Investment decisions in stocks are one of the most complex tasks due to the uncertainty of which stocks will increase or decrease in their values. A diversified portfolio statistically reduces the risk; however, stock choice still substantially influences the profitability. Methods: This work proposes a methodology to automate investment decision recommendations with clear explanations. It utilizes generative AI, guided by prompt engineering, to interpret price predictions derived from neural networks. The methodology also includes the Artificial Intelligence Trust, Risk, and Security Management (AI TRiSM) model to provide robust security recommendations for the system. The proposed system provides long-term investment recommendations based on the financial fundamentals of companies, such as the price-to-earnings ratio (PER) and the net margin of profits over the total revenue. The proposed explainable artificial intelligence (XAI) system uses DeepSeek for describing recommendations and suggested companies, as well as several charts based on Shapley additive explanation (SHAP) values and local-interpretable model-agnostic explanations (LIMEs) for showing feature importance. Results: In the experiments, we compared the profitability of the proposed portfolios, ranging from 8 to 28 stock values, with the maximum expected price increases for 4 years in the NASDAQ-100 and S&P-500, where both bull and bear markets were, respectively, considered before and after the custom duties increases in international trade by the USA in April 2025. The proposed system achieved an average profitability of 56.62% while considering 120 different portfolio recommendations. Conclusions: A t-Student test confirmed that the difference in profitability compared to the index was statistically significant. A user study revealed that the participants agreed that the portfolio explanations were useful for trusting the system, with an average score of 6.14 in a 7-point Likert scale.
Full article
(This article belongs to the Special Issue AI in Finance: Leveraging AI to Transform Financial Services)
►▼
Show Figures

Figure 1
Open AccessArticle
Robust Single-Cell RNA-Seq Analysis Using Hyperdimensional Computing: Enhanced Clustering and Classification Methods
by
Hossein Mohammadi, Maziyar Baranpouyan, Krishnaprasad Thirunarayan and Lingwei Chen
AI 2025, 6(5), 94; https://doi.org/10.3390/ai6050094 - 1 May 2025
Abstract
Background. Single-cell RNA sequencing (scRNA-seq) has transformed genomics by enabling the study of cellular heterogeneity. However, its high dimensionality, noise, and sparsity pose significant challenges for data analysis. Methods. We investigate the use of Hyperdimensional Computing (HDC), a brain-inspired computational framework recognized for
[...] Read more.
Background. Single-cell RNA sequencing (scRNA-seq) has transformed genomics by enabling the study of cellular heterogeneity. However, its high dimensionality, noise, and sparsity pose significant challenges for data analysis. Methods. We investigate the use of Hyperdimensional Computing (HDC), a brain-inspired computational framework recognized for its noise robustness and hardware efficiency, to tackle the challenges in scRNA-seq data analysis. We apply HDC to both supervised classification and unsupervised clustering tasks. Results. Our experiments demonstrate that HDC consistently outperforms established methods such as XGBoost, Seurat reference mapping, and scANVI in terms of noise tolerance and scalability. HDC achieves superior accuracy in classification tasks and maintains robust clustering performance across varying noise levels. Conclusions. These results highlight HDC as a promising framework for accurate and efficient single-cell data analysis. Its potential extends to other high-dimensional biological datasets including proteomics, epigenomics, and transcriptomics, with implications for advancing bioinformatics and personalized medicine.
Full article
(This article belongs to the Special Issue Artificial Intelligence in Biomedical Engineering: Challenges and Developments)
►▼
Show Figures

Figure 1
Open AccessArticle
Personalized Non-Player Characters: A Framework for Character-Consistent Dialogue Generation
by
Xiao Liu, Zhenping Xie and Senlin Jiang
AI 2025, 6(5), 93; https://doi.org/10.3390/ai6050093 - 1 May 2025
Abstract
Generating character-consistent and personalized dialogue for Non-Player Characters (NPCs) in Role-Playing Games (RPGs) poses significant challenges, especially due to limited memory retention and inconsistent character representation. This paper proposes a framework for generating personalized dialogues based on character-specific knowledge. By combining static knowledge
[...] Read more.
Generating character-consistent and personalized dialogue for Non-Player Characters (NPCs) in Role-Playing Games (RPGs) poses significant challenges, especially due to limited memory retention and inconsistent character representation. This paper proposes a framework for generating personalized dialogues based on character-specific knowledge. By combining static knowledge fine-tuning and dynamic knowledge graph technology, the framework generates dialogue content that is more aligned with character settings and is highly personalized. Specifically, the paper introduces a protective static knowledge fine-tuning approach to ensure that the language model does not generate content beyond the character’s cognitive scope during conversations. Additionally, dynamic knowledge graphs are employed to store and update the interaction history between NPCs and players, forming unique “experience-response” patterns. During dialogue generation, the paper first parses player input into an Abstract Meaning Representation (AMR) graph, retrieves relevant memory nodes from the knowledge graph, and constructs a fused graph structure. This integrated graph is encoded via a graph neural network to generate high-dimensional semantic vectors, which are then used to retrieve and supplement knowledge from the vector database. Ultimately, the model generates personalized responses consistent with the NPC’s identity. Experimental results demonstrate that the framework significantly enhances the authenticity of NPC dialogues and player immersion and performs well on multiple large-scale language models.
Full article
(This article belongs to the Section AI Systems: Theory and Applications)
►▼
Show Figures

Figure 1
Open AccessArticle
Who Is to Blame for the Bias in Visualizations, ChatGPT or DALL-E?
by
Dirk H. R. Spennemann
AI 2025, 6(5), 92; https://doi.org/10.3390/ai6050092 - 29 Apr 2025
Abstract
Due to range of factors in the development stage, generative artificial intelligence (AI) models cannot be completely free from bias. Some biases are introduced by the quality of training data, and developer influence during both design and training of the large language models
[...] Read more.
Due to range of factors in the development stage, generative artificial intelligence (AI) models cannot be completely free from bias. Some biases are introduced by the quality of training data, and developer influence during both design and training of the large language models (LLMs), while others are introduced in the text-to-image (T2I) visualization programs. The bias and initialization at the interface between LLMs and T2I applications has not been examined to date. This study analyzes 770 images of librarians and curators generated by DALL-E from ChatGPT-4o prompts to investigate the source of gender, ethnicity, and age biases in these visualizations. Comparing prompts generated by ChatGPT-4o with DALL-E’s visual interpretations, the research demonstrates that DALL-E primarily introduces biases when ChatGPT-4o provides non-specific prompts. This highlights the potential for generative AI to perpetuate and amplify harmful stereotypes related to gender, age, and ethnicity in professional roles.
Full article
(This article belongs to the Special Issue AI Bias in the Media and Beyond)
Open AccessArticle
A Hybrid and Modular Integration Concept for Anomaly Detection in Industrial Control Systems
by
Christian Goetz and Bernhard G. Humm
AI 2025, 6(5), 91; https://doi.org/10.3390/ai6050091 - 27 Apr 2025
Abstract
Effective anomaly detection is essential for realizing modern and secure industrial control systems. However, the direct integration of anomaly detection within such a system is complex due to the wide variety of hardware used, different communication protocols, and given industrial requirements. Many components
[...] Read more.
Effective anomaly detection is essential for realizing modern and secure industrial control systems. However, the direct integration of anomaly detection within such a system is complex due to the wide variety of hardware used, different communication protocols, and given industrial requirements. Many components of an industrial control system allow direct integration, while others are designed as closed systems or do not have the required performance. At the same time, the effective usage of available resources and the sustainable use of energy are more important than ever for modern industry. Therefore, in this paper, we present a modular and hybrid concept that enables the integration of efficient and effective anomaly detection while optimising the use of available resources under consideration of industrial requirements. Because of the modular and hybrid properties, many functionalities can be outsourced to the respective devices, and at the same time, additional hardware can be integrated where required. The resulting flexibility allows the seamless integration of complete anomaly detection into existing and legacy systems without the need for expensive centralised or cloud-based solutions. Through a detailed evaluation within an industrial unit, we demonstrate the performance and versatility of our concept.
Full article
(This article belongs to the Special Issue Artificial Intelligence Challenges to the Industrial Internet of Things and Industrial Control Systems Applications)
►▼
Show Figures

Figure 1
Open AccessArticle
Should We Reconsider RNNs for Time-Series Forecasting?
by
Vahid Naghashi, Mounir Boukadoum and Abdoulaye Banire Diallo
AI 2025, 6(5), 90; https://doi.org/10.3390/ai6050090 - 25 Apr 2025
Abstract
(1) Background: In recent years, Transformer-based models have dominated the time-series forecasting domain, overshadowing recurrent neural networks (RNNs) such as Long Short-Term Memory (LSTM) and Gated Recurrent Unit (GRU). While Transformers demonstrate superior performance, their high computational cost limits their practical application in
[...] Read more.
(1) Background: In recent years, Transformer-based models have dominated the time-series forecasting domain, overshadowing recurrent neural networks (RNNs) such as Long Short-Term Memory (LSTM) and Gated Recurrent Unit (GRU). While Transformers demonstrate superior performance, their high computational cost limits their practical application in resource-constrained settings. (2) Methods: In this paper, we reconsider RNNs—specifically the GRU architecture—as an efficient alternative to time-series forecasting by leveraging this architecture’s sequential representation capability to capture cross-channel dependencies effectively. Our model also utilizes a feed-forward layer right after the GRU module to represent temporal dependencies, and aggregates it with the GRU layers to predict future values of a given time-series. (3) Results and conclusions: Our extensive experiments conducted on different real-world datasets show that our inverted GRU (iGRU) model achieves promising results in terms of error metrics and memory efficiency, challenging or surpassing state-of-the-art models on various benchmarks.
Full article
(This article belongs to the Section AI Systems: Theory and Applications)
►▼
Show Figures

Figure 1
Open AccessArticle
Non-Linear Synthetic Time Series Generation for Electroencephalogram Data Using Long Short-Term Memory Models
by
Bakr Rashid Alqaysi, Manuel Rosa-Zurera and Ali Abdulameer Aldujaili
AI 2025, 6(5), 89; https://doi.org/10.3390/ai6050089 - 25 Apr 2025
Abstract
Background/Objectives: The implementation of artificial intelligence-based systems for disease detection using biomedical signals is challenging due to the limited availability of training data. This paper deals with the generation of synthetic EEG signals using deep learning-based models, to be used in future research
[...] Read more.
Background/Objectives: The implementation of artificial intelligence-based systems for disease detection using biomedical signals is challenging due to the limited availability of training data. This paper deals with the generation of synthetic EEG signals using deep learning-based models, to be used in future research for training Parkinson’s disease detection systems. Methods: Linear models, such as AR, MA, and ARMA, are often inadequate due to the inherent non-linearity of time series. To overcome this drawback, long short-term memory (LSTM) networks are proposed to learn long-term dependencies in non-linear EEG time series and subsequently generate synthetic signals to enhance the training of detection systems. To learn the forward and backward time dependencies in the EEG signals, a Bidirectional LSTM model has been implemented. The LSTM model was trained on the UC San Diego Resting State EEG Dataset, which includes samples from two groups: individuals with Parkinson’s disease and a healthy control group. Results: To determine the optimal number of cells in the model, we evaluated the mean squared error (MSE) and cross-correlation between the original and synthetic signals. This method was also applied to select the length of the hidden state vector. The number of hidden cells was set to 14, and the length of the hidden state vector for each cell was fixed at 4. Increasing these values did not improve MSE or cross-correlation and unnecessarily increased computational complexity. The proposed model’s performance was evaluated using the mean-squared error (MSE), Pearson’s correlation coefficient, and the power spectra of the synthetic and original signals, demonstrating the suitability of the proposed method for this application. Conclusions: The proposed model was compared to Autoregressive Moving Average (ARMA) models, demonstrating superior performance. This confirms that deep learning-based models, such as LSTM, are strong alternatives to statistical models like ARMA for handling non-linear, multifrequency, and non-stationary signals.
Full article
(This article belongs to the Topic Mathematical Applications and Computational Intelligence in Medicine and Biology)
►▼
Show Figures

Figure 1
Open AccessArticle
Evaluating the Efficacy of Deep Learning Models for Identifying Manipulated Medical Fundus Images
by
Ho-Jung Song, Ju-Hyuck Han, You-Sang Cho and Yong-Suk Kim
AI 2025, 6(5), 88; https://doi.org/10.3390/ai6050088 - 24 Apr 2025
Abstract
(1) Background: The misuse of transformation technology using medical images is a critical problem that can endanger patients’ lives, and detecting manipulation via a deep learning model is essential to address issues of manipulated medical images that may arise in the healthcare field.
[...] Read more.
(1) Background: The misuse of transformation technology using medical images is a critical problem that can endanger patients’ lives, and detecting manipulation via a deep learning model is essential to address issues of manipulated medical images that may arise in the healthcare field. (2) Methods: The dataset was divided into a real fundus dataset and a manipulated dataset. The fundus image manipulation detection model uses a deep learning model based on a Convolution Neural Network (CNN) structure that applies a concatenate operation for fast computation speed and reduced loss of input image weights. (3) Results: For real data, the model achieved an average sensitivity of 0.98, precision of 1.00, F1-score of 0.99, and AUC of 0.988. For manipulated data, the model recorded sensitivity of 1.00, precision of 0.84, F1-score of 0.92, and AUC of 0.988. Comparatively, five ophthalmologists achieved lower average scores on manipulated data: sensitivity of 0.71, precision of 0.61, F1-score of 0.65, and AUC of 0.822. (4) Conclusions: This study presents the possibility of addressing and preventing problems caused by manipulated medical images in the healthcare field. The proposed approach for detecting manipulated fundus images through a deep learning model demonstrates higher performance than that of ophthalmologists, making it an effective method.
Full article
(This article belongs to the Section Medical & Healthcare AI)
►▼
Show Figures

Figure 1
Open AccessReview
Deep Reinforcement Learning for Automated Insulin Delivery Systems: Algorithms, Applications, and Prospects
by
Xia Yu, Zi Yang, Xiaoyu Sun, Hao Liu, Hongru Li, Jingyi Lu, Jian Zhou and Ali Cinar
AI 2025, 6(5), 87; https://doi.org/10.3390/ai6050087 - 23 Apr 2025
Abstract
Advances in continuous glucose monitoring (CGM) technologies and wearable devices are enabling the enhancement of automated insulin delivery systems (AIDs) towards fully automated closed-loop systems, aiming to achieve secure, personalized, and optimal blood glucose concentration (BGC) management for individuals with diabetes. While model
[...] Read more.
Advances in continuous glucose monitoring (CGM) technologies and wearable devices are enabling the enhancement of automated insulin delivery systems (AIDs) towards fully automated closed-loop systems, aiming to achieve secure, personalized, and optimal blood glucose concentration (BGC) management for individuals with diabetes. While model predictive control provides a flexible framework for developing AIDs control algorithms, models that capture inter- and intra-patient variability and perturbation uncertainty are needed for accurate and effective regulation of BGC. Advances in artificial intelligence present new opportunities for developing data-driven, fully closed-loop AIDs. Among them, deep reinforcement learning (DRL) has attracted much attention due to its potential resistance to perturbations. To this end, this paper conducts a literature review on DRL-based BGC control algorithms for AIDs. First, this paper systematically analyzes the benefits of utilizing DRL algorithms in AIDs. Then, a comprehensive review of various DRL techniques and extensions that have been proposed to address challenges arising from their integration with AIDs, including considerations related to low sample availability, personalization, and security are discussed. Additionally, the paper provides an application-oriented investigation of DRL-based AIDs control algorithms, emphasizing significant challenges in practical implementations. Finally, the paper discusses solutions to relevant BGC control problems, outlines prospects for practical applications, and suggests future research directions.
Full article
(This article belongs to the Special Issue Artificial Intelligence for Future Healthcare: Advancement, Impact, and Prospect in the Field of Cancer)
►▼
Show Figures

Figure 1
Open AccessArticle
Evaluating the Societal Impact of AI: A Comparative Analysis of Human and AI Platforms Using the Analytic Hierarchy Process
by
Bojan Srđević
AI 2025, 6(4), 86; https://doi.org/10.3390/ai6040086 - 20 Apr 2025
Abstract
►▼
Show Figures
A central focus of this study was the methodology used to evaluate both humans and AI platforms, particularly in terms of their competitiveness and the implications of six key challenges to society resulting from the development and increasing use of artificial intelligence (AI)
[...] Read more.
A central focus of this study was the methodology used to evaluate both humans and AI platforms, particularly in terms of their competitiveness and the implications of six key challenges to society resulting from the development and increasing use of artificial intelligence (AI) technologies. The list of challenges was compiled by consulting various online sources and cross-referencing with academics from 15 countries across Europe and the USA. Professors, scientific researchers, and PhD students were invited to independently and remotely evaluate the challenges. Rather than contributing another discussion based solely on social arguments, this paper seeks to provide a logical evaluation framework, moving beyond qualitative discourse by incorporating numerical values. The pairwise comparison of AI challenges was conducted by two groups of participants using the multicriteria decision-making model known as the analytic hierarchy process (AHP). Thirty-eight humans performed pairwise comparisons of the six challenges after they were listed in a distributed questionnaire. The same procedure was carried out by four AI platforms—ChatGPT, Gemini (BardAI), Perplexity, and DedaAI—who responded to the same requests as the human participants. The results from both groups were grouped and compared, revealing interesting differences in the prioritization of AI challenges’ impact on society. Both groups agreed on the highest importance of data privacy and security, as well as the lowest importance of social and cultural resistance, specifically the clash of AI with existing cultural norms and societal values.
Full article

Figure 1
Open AccessArticle
CacheFormer: High-Attention-Based Segment Caching
by
Sushant Singh and Ausif Mahmood
AI 2025, 6(4), 85; https://doi.org/10.3390/ai6040085 - 18 Apr 2025
Abstract
►▼
Show Figures
Efficiently handling long contexts in transformer-based language models with low perplexity is an active area of research. Numerous recent approaches like Linformer, Longformer, Performer, and Structured state space models (SSMs), have not fully resolved this problem. All these models strive to reduce the
[...] Read more.
Efficiently handling long contexts in transformer-based language models with low perplexity is an active area of research. Numerous recent approaches like Linformer, Longformer, Performer, and Structured state space models (SSMs), have not fully resolved this problem. All these models strive to reduce the quadratic time complexity of the attention mechanism while minimizing the loss in quality due to the effective compression of the long context. Inspired by the cache and virtual memory principle in computers, where in case of a cache miss, not only the needed data are retrieved from the memory, but the adjacent data are also obtained, we apply this concept to handling long contexts by dividing it into small segments. In our design, we retrieve the nearby segments in an uncompressed form when high segment-level attention occurs at the compressed level. Our enhancements for handling long context include aggregating four attention mechanisms consisting of short sliding window attention, long compressed segmented attention, dynamically retrieving top-k high-attention uncompressed segments, and overlapping segments in long segment attention to avoid segment fragmentation. These enhancements result in an architecture that outperforms existing SOTA architectures with an average perplexity improvement of 8.5% over similar model sizes.
Full article

Figure 1
Open AccessSystematic Review
Artificial Intelligence in Ovarian Cancer: A Systematic Review and Meta-Analysis of Predictive AI Models in Genomics, Radiomics, and Immunotherapy
by
Mauro Francesco Pio Maiorano, Gennaro Cormio, Vera Loizzi and Brigida Anna Maiorano
AI 2025, 6(4), 84; https://doi.org/10.3390/ai6040084 - 18 Apr 2025
Abstract
Background/Objectives: Artificial intelligence (AI) is increasingly influencing oncological research by enabling precision medicine in ovarian cancer through enhanced prediction of therapy response and patient stratification. This systematic review and meta-analysis was conducted to assess the performance of AI-driven models across three key
[...] Read more.
Background/Objectives: Artificial intelligence (AI) is increasingly influencing oncological research by enabling precision medicine in ovarian cancer through enhanced prediction of therapy response and patient stratification. This systematic review and meta-analysis was conducted to assess the performance of AI-driven models across three key domains: genomics and molecular profiling, radiomics-based imaging analysis, and prediction of immunotherapy response. Methods: Relevant studies were identified through a systematic search across multiple databases (2020–2025), adhering to PRISMA guidelines. Results: Thirteen studies met the inclusion criteria, involving over 10,000 ovarian cancer patients and encompassing diverse AI models such as machine learning classifiers and deep learning architectures. Pooled AUCs indicated strong predictive performance for genomics-based (0.78), radiomics-based (0.88), and immunotherapy-based (0.77) models. Notably, radiogenomics-based AI integrating imaging and molecular data yielded the highest accuracy (AUC = 0.975), highlighting the potential of multi-modal approaches. Heterogeneity and risk of bias were assessed, and evidence certainty was graded. Conclusions: Overall, AI demonstrated promise in predicting therapeutic outcomes in ovarian cancer, with radiomics and integrated radiogenomics emerging as leading strategies. Future efforts should prioritize explainability, prospective multi-center validation, and integration of immune and spatial transcriptomic data to support clinical implementation and individualized treatment strategies. Unlike earlier reviews, this study synthesizes a broader range of AI applications in ovarian cancer and provides pooled performance metrics across diverse models. It examines the methodological soundness of the selected studies and highlights current gaps and opportunities for clinical translation, offering a comprehensive and forward-looking perspective in the field.
Full article
(This article belongs to the Special Issue Artificial Intelligence for Future Healthcare: Advancement, Impact, and Prospect in the Field of Cancer)
►▼
Show Figures

Figure 1
Open AccessArticle
Efficient Detection of Mind Wandering During Reading Aloud Using Blinks, Pitch Frequency, and Reading Rate
by
Amir Rabinovitch, Eden Ben Baruch, Maor Siton, Nuphar Avital, Menahem Yeari and Dror Malka
AI 2025, 6(4), 83; https://doi.org/10.3390/ai6040083 - 18 Apr 2025
Abstract
►▼
Show Figures
Mind wandering is a common issue among schoolchildren and academic students, often undermining the quality of learning and teaching effectiveness. Current detection methods mainly rely on eye trackers and electrodermal activity (EDA) sensors, focusing on external indicators such as facial movements but neglecting
[...] Read more.
Mind wandering is a common issue among schoolchildren and academic students, often undermining the quality of learning and teaching effectiveness. Current detection methods mainly rely on eye trackers and electrodermal activity (EDA) sensors, focusing on external indicators such as facial movements but neglecting voice detection. These methods are often cumbersome, uncomfortable for participants, and invasive, requiring specialized, expensive equipment that disrupts the natural learning environment. To overcome these challenges, a new algorithm has been developed to detect mind wandering during reading aloud. Based on external indicators like the blink rate, pitch frequency, and reading rate, the algorithm integrates these three criteria to ensure the accurate detection of mind wandering using only a standard computer camera and microphone, making it easy to implement and widely accessible. An experiment with ten participants validated this approach. Participants read aloud a text of 1304 words while the algorithm, incorporating the Viola–Jones model for face and eye detection and pitch-frequency analysis, monitored for signs of mind wandering. A voice activity detection (VAD) technique was also used to recognize human speech. The algorithm achieved 76% accuracy in predicting mind wandering during specific text segments, demonstrating the feasibility of using noninvasive physiological indicators. This method offers a practical, non-intrusive solution for detecting mind wandering through video and audio data, making it suitable for educational settings. Its ability to integrate seamlessly into classrooms holds promise for enhancing student concentration, improving the teacher–student dynamic, and boosting overall teaching effectiveness. By leveraging standard, accessible technology, this approach could pave the way for more personalized, technology-enhanced education systems.
Full article

Figure 1
Open AccessArticle
BEV-CAM3D: A Unified Bird’s-Eye View Architecture for Autonomous Driving with Monocular Cameras and 3D Point Clouds
by
Daniel Ayo Oladele, Elisha Didam Markus and Adnan M. Abu-Mahfouz
AI 2025, 6(4), 82; https://doi.org/10.3390/ai6040082 - 18 Apr 2025
Abstract
Three-dimensional (3D) visual perception is pivotal for understanding surrounding environments in applications such as autonomous driving and mobile robotics. While LiDAR-based models dominate due to accurate depth sensing, their cost and sparse outputs have driven interest in camera-based systems. However, challenges like cross-domain
[...] Read more.
Three-dimensional (3D) visual perception is pivotal for understanding surrounding environments in applications such as autonomous driving and mobile robotics. While LiDAR-based models dominate due to accurate depth sensing, their cost and sparse outputs have driven interest in camera-based systems. However, challenges like cross-domain degradation and depth estimation inaccuracies persist. This paper introduces BEVCAM3D, a unified bird’s-eye view (BEV) architecture that fuses monocular cameras and LiDAR point clouds to overcome single-sensor limitations. BEVCAM3D integrates a deformable cross-modality attention module for feature alignment and a fast ground segmentation algorithm to reduce computational overhead by 40%. Evaluated on the nuScenes dataset, BEVCAM3D achieves state-of-the-art performance, with a 73.9% mAP and a 76.2% NDS, outperforming existing LiDAR-camera fusion methods like SparseFusion (72.0% mAP) and IS-Fusion (73.0% mAP). Notably, it excels in detecting pedestrians (91.0% AP) and traffic cones (89.9% AP), addressing the class imbalance in autonomous driving scenarios. The framework supports real-time inference at 11.2 FPS with an EfficientDet-B3 backbone and demonstrates robustness under low-light conditions (62.3% nighttime mAP).
Full article
(This article belongs to the Section AI in Autonomous Systems)
►▼
Show Figures

Figure 1
Open AccessArticle
The Impact of Ancient Greek Prompts on Artificial Intelligence Image Generation: A New Educational Paradigm
by
Anna Kalargirou, Dimitrios Kotsifakos and Christos Douligeris
AI 2025, 6(4), 81; https://doi.org/10.3390/ai6040081 - 18 Apr 2025
Abstract
►▼
Show Figures
Background/Objectives: This article explores the use of Ancient Greek as a prompt language in DALL·E 3, an Artificial Intelligence software for image generation. The research investigates three dimensions of Artificial Intelligence’s ability: (a) the sense and visualization of the concept of distance, (b)
[...] Read more.
Background/Objectives: This article explores the use of Ancient Greek as a prompt language in DALL·E 3, an Artificial Intelligence software for image generation. The research investigates three dimensions of Artificial Intelligence’s ability: (a) the sense and visualization of the concept of distance, (b) the mixing of representational as well as mythic contents, and (c) the visualization of emotions. More specifically, the research not only investigates AI’s potentialities in processing and representing Ancient Greek texts but also attempts to assess its interpretative boundaries. The key question is whether AI can faithfully represent the underlying conceptual and narrative structures of ancient literature or whether its representations are superficial and constrained by algorithmic procedures. Methods: This is a mixed-methods experimental research design examining whether a specified Artificial Intelligence software can sense, understand, and graphically represent linguistic and conceptual structures in the Ancient Greek language. Results: The study highlights Artificial Intelligence’s possibility in classical language education as well as digital humanities regarding linguistic complexity versus AI’s power in interpretation. More specifically, the research not only investigates AI’s potentialities in processing and representing Ancient Greek texts but also attempts to assess its interpretative boundaries. The key question is whether AI can faithfully represent the underlying conceptual and narrative structures of ancient literature or whether its representations are superficial and constrained by algorithmic procedures. The study highlights Artificial Intelligence’s possibility in classical language education as well as digital humanities regarding linguistic complexity versus AI’s power in interpretation. Conclusions: The research is a step toward a more extensive discussion on Artificial Intelligence in historical linguistics, digital pedagogy, as well as aesthetic representation by Artificial Intelligence environments.
Full article

Figure 1
Open AccessArticle
Radiomics-Based Machine Learning Models Improve Acute Pancreatitis Severity Prediction
by
Ahmet Yasin Karkas, Gorkem Durak, Onder Babacan, Timurhan Cebeci, Emre Uysal, Halil Ertugrul Aktas, Mehmet Ilhan, Alpay Medetalibeyoglu, Ulas Bagci, Mehmet Semih Cakir and Sukru Mehmet Erturk
AI 2025, 6(4), 80; https://doi.org/10.3390/ai6040080 - 18 Apr 2025
Abstract
(1) Acute pancreatitis (AP) is a medical emergency associated with high mortality rates. Early and accurate prognosis assessment during admission is crucial for optimizing patient management and outcomes. This study seeks to develop robust radiomics-based machine learning (ML) models to classify the severity
[...] Read more.
(1) Acute pancreatitis (AP) is a medical emergency associated with high mortality rates. Early and accurate prognosis assessment during admission is crucial for optimizing patient management and outcomes. This study seeks to develop robust radiomics-based machine learning (ML) models to classify the severity of AP using contrast-enhanced computed tomography (CECT) scans. (2) Methods: A retrospective cohort of 287 AP patients with CECT scans was analyzed, and clinical data were collected within 72 h of admission. Patients were classified as mild or moderate/severe based on the Revised Atlanta classification. Two radiologists manually segmented the pancreas and peripancreatic regions on CECT scans, and 234 radiomic features were extracted. The performance of the ML algorithms was compared with that of traditional scoring systems, including Ranson and Glasgow-Imrie scores. (3) Results: Traditional severity scoring systems produced AUC values of 0.593 (Ranson, Admission), 0.696 (Ranson, 48 h), 0.677 (Ranson, Cumulative), and 0.663 (Glasgow-Imrie). Using LASSO regression, 12 radiomic features were selected for the ML classifiers. Among these, the best-performing ML classifier achieved an AUC of 0.876 in the training set and 0.777 in the test set. (4) Conclusions: Radiomics-based ML classifiers significantly enhanced the prediction of AP severity in patients undergoing CECT scans within 72 h of admission, outperforming traditional severity scoring systems. This research is the first to successfully predict prognosis by analyzing radiomic features from both pancreatic and peripancreatic tissues using multiple ML algorithms applied to early CECT images.
Full article
(This article belongs to the Section Medical & Healthcare AI)
►▼
Show Figures

Figure 1
Open AccessArticle
Cross-Context Stress Detection: Evaluating Machine Learning Models on Heterogeneous Stress Scenarios Using EEG Signals
by
Omneya Attallah, Mona Mamdouh and Ahmad Al-Kabbany
AI 2025, 6(4), 79; https://doi.org/10.3390/ai6040079 - 14 Apr 2025
Abstract
Background/Objectives: This article addresses the challenge of stress detection across diverse contexts. Mental stress is a worldwide concern that substantially affects human health and productivity, rendering it a critical research challenge. Although numerous studies have investigated stress detection through machine learning (ML) techniques,
[...] Read more.
Background/Objectives: This article addresses the challenge of stress detection across diverse contexts. Mental stress is a worldwide concern that substantially affects human health and productivity, rendering it a critical research challenge. Although numerous studies have investigated stress detection through machine learning (ML) techniques, there has been limited research on assessing ML models trained in one context and utilized in another. The objective of ML-based stress detection systems is to create models that generalize across various contexts. Methods: This study examines the generalizability of ML models employing EEG recordings from two stress-inducing contexts: mental arithmetic evaluation (MAE) and virtual reality (VR) gaming. We present a data collection workflow and publicly release a portion of the dataset. Furthermore, we evaluate classical ML models and their generalizability, offering insights into the influence of training data on model performance, data efficiency, and related expenses. EEG data were acquired leveraging MUSE-STM hardware during stressful MAE and VR gaming scenarios. The methodology entailed preprocessing EEG signals using wavelet denoising mother wavelets, assessing individual and aggregated sensor data, and employing three ML models—linear discriminant analysis (LDA), support vector machine (SVM), and K-nearest neighbors (KNN)—for classification purposes. Results: In Scenario 1, where MAE was employed for training and VR for testing, the TP10 electrode attained an average accuracy of 91.42% across all classifiers and participants, whereas the SVM classifier achieved the highest average accuracy of 95.76% across all participants. In Scenario 2, adopting VR data as the training data and MAE data as the testing data, the maximum average accuracy achieved was 88.05% with the combination of TP10, AF8, and TP9 electrodes across all classifiers and participants, whereas the LDA model attained the peak average accuracy of 90.27% among all participants. The optimal performance was achieved with Symlets 4 and Daubechies-2 for Scenarios 1 and 2, respectively. Conclusions: The results demonstrate that although ML models exhibit generalization capabilities across stressors, their performance is significantly influenced by the alignment between training and testing contexts, as evidenced by systematic cross-context evaluations using an 80/20 train–test split per participant and quantitative metrics (accuracy, precision, recall, and F1-score) averaged across participants. The observed variations in performance across stress scenarios, classifiers, and EEG sensors provide empirical support for this claim.
Full article
(This article belongs to the Special Issue Artificial Intelligence in Biomedical Engineering: Challenges and Developments)
►▼
Show Figures

Figure 1
Open AccessArticle
On the Integration of Social Context for Enhanced Fake News Detection Using Multimodal Fusion Attention Mechanism
by
Hachemi Nabil Dellys, Halima Mokeddem and Layth Sliman
AI 2025, 6(4), 78; https://doi.org/10.3390/ai6040078 - 11 Apr 2025
Abstract
►▼
Show Figures
Detecting fake news has become a critical challenge in today’s information-dense society. Existing research on fake news detection predominantly emphasizes multi-modal approaches, focusing primarily on textual and visual features. However, despite its clear importance, the integration of social context has received limited attention
[...] Read more.
Detecting fake news has become a critical challenge in today’s information-dense society. Existing research on fake news detection predominantly emphasizes multi-modal approaches, focusing primarily on textual and visual features. However, despite its clear importance, the integration of social context has received limited attention in the literature. To address this gap, this study proposes a novel three-dimensional multimodal fusion framework that integrates textual, visual, and social context features for effective fake news detection on social media platforms. The proposed methodology leverages an advanced Vision-and-Language Bidirectional Encoder Representations from Transformers multi-task model to extract fused attention features from text and images concurrently, capturing intricate inter-modal correlations. Comprehensive experiments validate the efficacy of the proposed approach. The results demonstrate that the proposed solution achieves the highest balanced accuracy of 77%, surpassing other baseline models. Furthermore, the incorporation of social context features significantly enhances model performance. The proposed multimodal architecture also outperforms state-of-the-art approaches, providing a robust and scalable framework for fake news detection using artificial intelligence. This study contributes to advancing the field by offering a comprehensive and practical engineering solution for combating fake news.
Full article

Figure 1
Open AccessArticle
Enhancing the Classification of Imbalanced Arabic Medical Questions Using DeepSMOTE
by
Bushra Al-Smadi, Bassam Hammo, Hossam Faris and Pedro A. Castillo
AI 2025, 6(4), 77; https://doi.org/10.3390/ai6040077 - 11 Apr 2025
Abstract
The growing demand for telemedicine has highlighted the need for automated healthcare services, particularly in medical question classification. This study presents a deep learning model designed to address key challenges in telemedicine, including class imbalance and accurate routing of Arabic medical questions to
[...] Read more.
The growing demand for telemedicine has highlighted the need for automated healthcare services, particularly in medical question classification. This study presents a deep learning model designed to address key challenges in telemedicine, including class imbalance and accurate routing of Arabic medical questions to the correct specialties. The model combines AraBERTv0.2-Twitter, fine-tuned for informal Arabic, with Bidirectional Long Short-Term Memory (BiLSTM) networks to capture deep semantic relationships in medical text. We used a labeled dataset of 5000 Arabic consultation records from Altibbi, covering five key medical specialties selected for their clinical relevance and frequency. The data underwent preprocessing to remove noise and normalize text. We employed stratified sampling to ensure representative distribution across the selected medical specialties. We evaluate multiple models using macro precision, macro recall, macro F1-score, weighted F1-score, and G-Mean. Our results demonstrate that DeepSMOTE combined with cross-entropy loss achieves the best performance. The findings offer statistically significant improvements and have practical implications for improving screening and patient routing in telemedicine platforms.
Full article
(This article belongs to the Section Medical & Healthcare AI)
►▼
Show Figures

Figure 1
Highly Accessed Articles
Latest Books
E-Mail Alert
News
Topics
Topic in
AI, Applied Sciences, Education Sciences, Electronics, Information
Explainable AI in Education
Topic Editors: Guanfeng Liu, Karina Luzia, Luke Bozzetto, Tommy Yuan, Pengpeng ZhaoDeadline: 30 June 2025
Topic in
Applied Sciences, Energies, Buildings, Smart Cities, AI
Smart Electric Energy in Buildings
Topic Editors: Daniel Villanueva Torres, Ali Hainoun, Sergio Gómez MelgarDeadline: 15 July 2025
Topic in
AI, BDCC, Fire, GeoHazards, Remote Sensing
AI for Natural Disasters Detection, Prediction and Modeling
Topic Editors: Moulay A. Akhloufi, Mozhdeh ShahbaziDeadline: 25 July 2025
Topic in
Algorithms, Applied Sciences, Electronics, MAKE, AI, Software
Applications of NLP, AI, and ML in Software Engineering
Topic Editors: Affan Yasin, Javed Ali Khan, Lijie WenDeadline: 31 August 2025

Conferences
Special Issues
Special Issue in
AI
Machine Learning in Bioinformatics: Current Research and Development
Guest Editors: Yinghao Wu, Kalyani Dhusia, Zhaoqian SuDeadline: 30 June 2025
Special Issue in
AI
Exploring the Use of Artificial Intelligence in Education
Guest Editor: Hyeon JoDeadline: 30 June 2025
Special Issue in
AI
Trustworthy AI and Distributed Intelligence in Smart Cities
Guest Editors: Kashif Ahmad, Jebran KhanDeadline: 30 June 2025
Special Issue in
AI
Artificial Intelligence-Based Object Detection and Tracking: Theory and Applications
Guest Editors: Di Yuan, Xiu ShuDeadline: 30 June 2025