Next Issue
Volume 9, June
Previous Issue
Volume 9, April
 
 

Big Data Cogn. Comput., Volume 9, Issue 5 (May 2025) – 32 articles

Cover Story (view full-size image): In the context of breast tumor detection and classification using ultrasound images, recent years have seen growing research interest, driven by the intersection of deep learning algorithms and medical image analysis. This article focuses on comparing a traditional, flat, three-class model with a hierarchical, two-tier classification approach, which first distinguishes normal from tumorous tissue and then classifies tumors as benign or malignant. Aimed at rethinking current methodologies, the study evaluates a novel architecture, providing insights for future algorithm development, broader clinical applicability, and the seamless integration of the proposed model into an existing web application for deployment. View this paper
  • Issues are regarded as officially published after their release is announced to the table of contents alert mailing list.
  • You may sign up for e-mail alerts to receive table of contents of newly released issues.
  • PDF is the official format for papers published in both, html and pdf forms. To view the papers in pdf format, click on the "PDF Full-text" link, and use the free Adobe Reader to open them.
Order results
Result details
Select all
Export citation of selected articles as:
42 pages, 1639 KiB  
Article
A Comprehensive Evaluation of Embedding Models and LLMs for IR and QA Across English and Italian
by Ermelinda Oro, Francesco Maria Granata and Massimo Ruffolo
Big Data Cogn. Comput. 2025, 9(5), 141; https://doi.org/10.3390/bdcc9050141 - 21 May 2025
Viewed by 14
Abstract
This study presents a comprehensive evaluation of embedding techniques and large language models (LLMs) for Information Retrieval (IR) and question answering (QA) across languages, focusing on English and Italian. We address a significant research gap by providing empirical evidence of model performance across [...] Read more.
This study presents a comprehensive evaluation of embedding techniques and large language models (LLMs) for Information Retrieval (IR) and question answering (QA) across languages, focusing on English and Italian. We address a significant research gap by providing empirical evidence of model performance across linguistic boundaries. We evaluate 12 embedding models on diverse IR datasets, including Italian SQuAD and DICE, English SciFact, ArguAna, and NFCorpus. We assess four LLMs (GPT4o, LLama-3.1 8B, Mistral-Nemo, and Gemma-2b) for QA tasks within a retrieval-augmented generation (RAG) pipeline. We evaluate them on SQuAD, CovidQA, and NarrativeQA datasets, including cross-lingual scenarios. The results show multilingual models perform more competitively than language-specific ones. The embed-multilingual-v3.0 model achieves top nDCG@10 scores of 0.90 for English and 0.86 for Italian. In QA evaluation, Mistral-Nemo demonstrates superior answer relevance (0.91–1.0) while maintaining strong groundedness (0.64–0.78). Our analysis reveals three key findings: (1) multilingual embedding models effectively bridge performance gaps between English and Italian, though performance consistency decreases in specialized domains, (2) model size does not consistently predict performance, and (3) all evaluated QA systems exhibit a critical trade-off between answer relevance and factual groundedness. Our evaluation framework combines traditional metrics with innovative LLM-based assessment techniques. It establishes new benchmarks for multilingual language technologies while providing actionable insights for real-world IR and QA system deployment. Full article
(This article belongs to the Special Issue Advances in Natural Language Processing and Text Mining)
Show Figures

Figure 1

25 pages, 2733 KiB  
Article
Polarity of Yelp Reviews: A BERT–LSTM Comparative Study
by Rachid Belaroussi, Sié Cyriac Noufe, Francis Dupin and Pierre-Olivier Vandanjon
Big Data Cogn. Comput. 2025, 9(5), 140; https://doi.org/10.3390/bdcc9050140 - 21 May 2025
Viewed by 9
Abstract
With the rapid growth in social network comments, the need for more effective methods to classify their polarity—negative, neutral, or positive—has become essential. Sentiment analysis, powered by natural language processing, has evolved significantly with the adoption of advanced deep learning techniques. Long Short-Term [...] Read more.
With the rapid growth in social network comments, the need for more effective methods to classify their polarity—negative, neutral, or positive—has become essential. Sentiment analysis, powered by natural language processing, has evolved significantly with the adoption of advanced deep learning techniques. Long Short-Term Memory networks capture long-range dependencies in text, while transformers, with their attention mechanisms, excel at preserving contextual meaning and handling high-dimensional, semantically complex data. This study compares the performance of sentiment analysis models based on LSTM and BERT architectures using key evaluation metrics. The dataset consists of business reviews from the Yelp Open Dataset. We tested LSTM-based methods against BERT and its variants—RoBERTa, BERTweet, and DistilBERT—leveraging popular pipelines from the Hugging Face Hub. A class-by-class performance analysis is presented, revealing that more complex BERT-based models do not always guarantee superior results in the classification of Yelp reviews. Additionally, the use of bidirectionality in LSTMs does not necessarily lead to better performance. However, across a diversity of test sets, transformer models outperform traditional RNN-based models, as their generalization capability is greater than that of a simple LSTM model. Full article
(This article belongs to the Special Issue Advances in Natural Language Processing and Text Mining)
Show Figures

Figure 1

17 pages, 1137 KiB  
Article
Tri-Collab: A Machine Learning Project to Leverage Innovation Ecosystems in Portugal
by Ângelo Marujo, Bruno Afonso, Inês Martins, Lisandro Pires and Sílvia Fernandes
Big Data Cogn. Comput. 2025, 9(5), 139; https://doi.org/10.3390/bdcc9050139 - 20 May 2025
Viewed by 148
Abstract
This project consists of a digital platform named Tri-Collab, where investors, entrepreneurs, and other agents (mainly talents) can cooperate on their ideas and eventually co-create. It is a digital means for this triad of actors (among other potential ones) to better adjust their [...] Read more.
This project consists of a digital platform named Tri-Collab, where investors, entrepreneurs, and other agents (mainly talents) can cooperate on their ideas and eventually co-create. It is a digital means for this triad of actors (among other potential ones) to better adjust their requirements. It includes an app that easily communicates with a database of projects, innovation agents and their profiles, and the originality lies in the matching algorithm. Thus, co-creation can have better support through this assertive interconnection of players and their resources. This work also highlights the usefulness of the Canvas Business Model in structuring the idea and its dashboard, allowing a comprehensive view of channels, challenges and gains. Also, the potential of machine learning in improving matchmaking platforms is discussed, especially when technological advancements allow for forecasts and match people at scale. Full article
Show Figures

Figure 1

29 pages, 4204 KiB  
Article
A Comparative Study of Ensemble Machine Learning and Explainable AI for Predicting Harmful Algal Blooms
by Omer Mermer, Eddie Zhang and Ibrahim Demir
Big Data Cogn. Comput. 2025, 9(5), 138; https://doi.org/10.3390/bdcc9050138 - 20 May 2025
Viewed by 63
Abstract
Harmful algal blooms (HABs), driven by environmental pollution, pose significant threats to water quality, public health, and aquatic ecosystems. This study enhances the prediction of HABs in Lake Erie, part of the Great Lakes system, by utilizing ensemble machine learning (ML) models coupled [...] Read more.
Harmful algal blooms (HABs), driven by environmental pollution, pose significant threats to water quality, public health, and aquatic ecosystems. This study enhances the prediction of HABs in Lake Erie, part of the Great Lakes system, by utilizing ensemble machine learning (ML) models coupled with explainable artificial intelligence (XAI) for interpretability. Using water quality data from 2013 to 2020, various physical, chemical, and biological parameters were analyzed to predict chlorophyll-a (Chl-a) concentrations, which are a commonly used indicator of phytoplankton biomass and a proxy for algal blooms. This study employed multiple ensemble ML models, including random forest (RF), deep forest (DF), gradient boosting (GB), and XGBoost, and compared their performance against individual models, such as support vector machine (SVM), decision tree (DT), and multi-layer perceptron (MLP). The findings revealed that the ensemble models, particularly XGBoost and deep forest (DF), achieved superior predictive accuracy, with R2 values of 0.8517 and 0.8544, respectively. The application of SHapley Additive exPlanations (SHAPs) provided insights into the relative importance of the input features, identifying the particulate organic nitrogen (PON), particulate organic carbon (POC), and total phosphorus (TP) as the critical factors influencing the Chl-a concentrations. This research demonstrates the effectiveness of ensemble ML models for achieving high predictive accuracy, while the integration of XAI enhances model interpretability. The results support the development of proactive water quality management strategies and highlight the potential of advanced ML techniques for environmental monitoring. Full article
(This article belongs to the Special Issue Machine Learning Applications and Big Data Challenges)
Show Figures

Figure 1

21 pages, 1559 KiB  
Article
The Development of Small-Scale Language Models for Low-Resource Languages, with a Focus on Kazakh and Direct Preference Optimization
by Nurgali Kadyrbek, Zhanseit Tuimebayev, Madina Mansurova and Vítor Viegas
Big Data Cogn. Comput. 2025, 9(5), 137; https://doi.org/10.3390/bdcc9050137 - 20 May 2025
Viewed by 109
Abstract
Low-resource languages remain underserved by contemporary large language models (LLMs) because they lack sizable corpora, bespoke preprocessing tools, and the computing budgets assumed by mainstream alignment pipelines. Focusing on Kazakh, we present a 1.94B parameter LLaMA-based model that demonstrates how strong, culturally aligned [...] Read more.
Low-resource languages remain underserved by contemporary large language models (LLMs) because they lack sizable corpora, bespoke preprocessing tools, and the computing budgets assumed by mainstream alignment pipelines. Focusing on Kazakh, we present a 1.94B parameter LLaMA-based model that demonstrates how strong, culturally aligned performance can be achieved without massive infrastructure. The contribution is threefold. (i) Data and tokenization—we compile a rigorously cleaned, mixed-domain Kazakh corpus and design a tokenizer that respects the language’s agglutinative morphology, mixed-script usage, and diacritics. (ii) Training recipe—the model is built in two stages: causal language modeling from scratch followed by instruction tuning. Alignment is further refined with Direct Preference Optimization (DPO), extended by contrastive and entropy-based regularization to stabilize training under sparse, noisy preference signals. Two complementary resources support this step: ChatTune-DPO, a crowd-sourced set of human preference pairs, and Pseudo-DPO, an automatically generated alternative that repurposes instruction data to reduce annotation cost. (iii) Evaluation and impact—qualitative and task-specific assessments show that targeted monolingual training and the proposed DPO variant markedly improve factuality, coherence, and cultural fidelity over baseline instruction-only and multilingual counterparts. The model and datasets are released under open licenses, offering a reproducible blueprint for extending state-of-the-art language modeling to other under-represented languages and domains. Full article
Show Figures

Figure 1

15 pages, 1909 KiB  
Article
Helium Speech Recognition Method Based on Spectrogram with Deep Learning
by Yonghong Chen, Shibing Zhang and Dongmei Li
Big Data Cogn. Comput. 2025, 9(5), 136; https://doi.org/10.3390/bdcc9050136 - 20 May 2025
Viewed by 122
Abstract
With the development of the marine economy and the increase in marine activities, deep saturation diving has gained significant attention. Helium speech communication is indispensable for saturation diving operations and is a critical technology for deep saturation diving, serving as the sole communication [...] Read more.
With the development of the marine economy and the increase in marine activities, deep saturation diving has gained significant attention. Helium speech communication is indispensable for saturation diving operations and is a critical technology for deep saturation diving, serving as the sole communication method to ensure the smooth execution of such operations. This study introduces deep learning into helium speech recognition and proposes a spectrogram-based dual-model helium speech recognition method. First, we extract the spectrogram features from the helium speech. Then, we combine a deep fully convolutional neural network with connectionist temporal classification (CTC) to form an acoustic model, in which the spectrogram features of helium speech are used as an input to convert speech signals into phonetic sequences. Finally, a maximum entropy hidden Markov model (MEMM) is employed as the language model to convert the phonetic sequences to word outputs, which is regarded as a dynamic programming problem. We use a Viterbi algorithm to find the optimal path to decode the phonetic sequences to word sequences. The simulation results show that the method can effectively recognize helium speech with a recognition rate of 97.89% for isolated words and 95.99% for continuous helium speech. Full article
Show Figures

Figure 1

35 pages, 3235 KiB  
Article
Applying Big Data for Maritime Accident Risk Assessment: Insights, Predictive Insights and Challenges
by Vicky Zampeta, Gregory Chondrokoukis and Dimosthenis Kyriazis
Big Data Cogn. Comput. 2025, 9(5), 135; https://doi.org/10.3390/bdcc9050135 - 19 May 2025
Viewed by 202
Abstract
Maritime safety is a critical concern for the transport sector and remains a key challenge for the international shipping industry. Recognizing that maritime accidents pose significant risks to both safety and operational efficiency, this study explores the application of big data analysis techniques [...] Read more.
Maritime safety is a critical concern for the transport sector and remains a key challenge for the international shipping industry. Recognizing that maritime accidents pose significant risks to both safety and operational efficiency, this study explores the application of big data analysis techniques to understand the factors influencing maritime transport accidents (MTA). Specifically, using extensive datasets derived from vessel performance measurements, environmental conditions, and accident reports, it seeks to identify the key intrinsic and extrinsic factors contributing to maritime accidents. The research examines more than 90 thousand incidents for the period 2014–2022. Leveraging big data analytics and advanced statistical techniques, the findings reveal significant correlations between vessel size, speed, and specific environmental factors. Furthermore, the study highlights the potential of big data analytics in enhancing predictive modeling, real-time risk assessment, and decision-making processes for maritime traffic management. The integration of big data with intelligent transportation systems (ITSs) can optimize safety strategies, improve accident prevention mechanisms, and enhance the resilience of ocean-going transportation systems. By bridging the gap between big data applications and maritime safety research, this work contributes to the literature by emphasizing the importance of examining both intrinsic and extrinsic factors in predicting maritime accident risks. Additionally, it underscores the transformative role of big data in shaping safer and more efficient waterway transportation systems. Full article
Show Figures

Figure 1

20 pages, 1064 KiB  
Article
Predicting Early Employability of Vietnamese Graduates: Insights from Data-Driven Analysis Through Machine Learning Methods
by Long-Sheng Chen, Thao-Trang Huynh-Cam, Van-Canh Nguyen, Tzu-Chuen Lu and Dang-Khoa Le-Huynh
Big Data Cogn. Comput. 2025, 9(5), 134; https://doi.org/10.3390/bdcc9050134 - 19 May 2025
Viewed by 185
Abstract
Graduate employability remains a crucial challenge for higher education institutions, especially in developing economies. This study investigates the key academic and vocational factors influencing early employment outcomes among recent graduates at a public university in Vietnam’s Mekong Delta region. By leveraging predictive analytics, [...] Read more.
Graduate employability remains a crucial challenge for higher education institutions, especially in developing economies. This study investigates the key academic and vocational factors influencing early employment outcomes among recent graduates at a public university in Vietnam’s Mekong Delta region. By leveraging predictive analytics, the research explores how data-driven approaches can enhance career readiness strategies. The analysis employed AI-driven models, particularly classification and regression trees (CARTs), using a dataset of 610 recent graduates from a public university in the Mekong Delta to predict early employability. The input factors included gender, field of study, university entrance scores, and grade point average (GPA) scores for four university years. The output factor was recent graduates’ (un)employment within six months after graduation. Among all input factors, third-year GPA, university entrance scores, and final-year academic performance are the most significant predictors of early employment. Among the tested models, CARTs achieved the highest accuracy (93.6%), offering interpretable decision rules that can inform curriculum design and career support services. This study contributes to the intersection of artificial intelligence and vocational education by providing actionable insights for universities, policymakers, and employers, supporting the alignment of education with labor market demands and improving graduate employability outcomes. Full article
Show Figures

Figure 1

26 pages, 2125 KiB  
Article
Adaptive Augmented Reality Architecture for Optimising Assistance and Safety in Industry 4.0
by Ginés Morales Méndez and Francisco del Cerro Velázquez
Big Data Cogn. Comput. 2025, 9(5), 133; https://doi.org/10.3390/bdcc9050133 - 19 May 2025
Viewed by 198
Abstract
The present study proposes adaptive augmented reality (AR) architecture, specifically designed to enhance real-time operator assistance and occupational safety in industrial environments, which is representative of Industry 4.0. The proposed system addresses key challenges in AR adoption, such as the need for dynamic [...] Read more.
The present study proposes adaptive augmented reality (AR) architecture, specifically designed to enhance real-time operator assistance and occupational safety in industrial environments, which is representative of Industry 4.0. The proposed system addresses key challenges in AR adoption, such as the need for dynamic personalisation of instructions based on operator profiles and the mitigation of technical and cognitive barriers. Architecture integrates theoretical modelling, modular design, and real-time adaptability to match instruction complexity with user expertise and environmental conditions. A working prototype was implemented using Microsoft HoloLens 2, Unity 3D, and Vuforia and validated in a controlled industrial scenario involving predictive maintenance and assembly tasks. The experimental results demonstrated statistically significant enhancements in task completion time, error rates, perceived cognitive load, operational efficiency, and safety indicators in comparison with conventional methods. The findings underscore the system’s capacity to enhance both performance and consistency while concomitantly bolstering risk mitigation in intricate operational settings. This study proposes a scalable and modular AR framework with built-in safety and adaptability mechanisms, demonstrating practical benefits for human–machine interaction in Industry 4.0. The present study is subject to certain limitations, including validation in a simulated environment, which limits the direct extrapolation of the results to real industrial scenarios; further evaluation in various operational contexts is required to verify the overall scalability and applicability of the proposed system. It is recommended that future research studies explore the long-term ergonomics, scalability, and integration of emerging technologies in decision support within adaptive AR systems. Full article
Show Figures

Figure 1

25 pages, 11184 KiB  
Article
Comparative Evaluation of Multimodal Large Language Models for No-Reference Image Quality Assessment with Authentic Distortions: A Study of OpenAI and Claude.AI Models
by Domonkos Varga
Big Data Cogn. Comput. 2025, 9(5), 132; https://doi.org/10.3390/bdcc9050132 - 16 May 2025
Viewed by 193
Abstract
This study presents a comparative analysis of several multimodal large language models (LLMs) for no-reference image quality assessment, with a particular focus on images containing authentic distortions. We evaluate three models developed by OpenAI and three models from Claude.AI, comparing their performance in [...] Read more.
This study presents a comparative analysis of several multimodal large language models (LLMs) for no-reference image quality assessment, with a particular focus on images containing authentic distortions. We evaluate three models developed by OpenAI and three models from Claude.AI, comparing their performance in estimating image quality without reference images. Our results demonstrate that these LLMs outperform traditional methods based on hand-crafted features. However, more advanced deep learning models, especially those based on deep convolutional networks, surpass LLMs in performance. Notably, we make a unique contribution by publishing the processed outputs of the LLMs, providing a transparent and direct comparison of their quality assessments based solely on the predicted quality scores. This work underscores the potential of multimodal LLMs in image quality evaluation, while also highlighting the continuing advantages of specialized deep learning approaches. Full article
(This article belongs to the Special Issue Advances in Natural Language Processing and Text Mining)
Show Figures

Figure 1

17 pages, 959 KiB  
Article
Evaluating the Impact of Artificial Intelligence Tools on Enhancing Student Academic Performance: Efficacy Amidst Security and Privacy Concerns
by Jwern Tick Kiet Phua, Han-Foon Neo and Chuan-Chin Teo
Big Data Cogn. Comput. 2025, 9(5), 131; https://doi.org/10.3390/bdcc9050131 - 15 May 2025
Viewed by 253
Abstract
The rapid advancements in artificial intelligence (AI) have significantly transformed various domains, including education, by introducing innovative tools that reshape teaching and learning processes. This research investigates the perceptions and attitudes of students towards the use of AI tools in their academic activities, [...] Read more.
The rapid advancements in artificial intelligence (AI) have significantly transformed various domains, including education, by introducing innovative tools that reshape teaching and learning processes. This research investigates the perceptions and attitudes of students towards the use of AI tools in their academic activities, focusing on constructs such as perceived usefulness, the perceived ease of use, security and privacy concerns, and both positive and negative attitudes towards AI. On the basis of Technology Acceptance Model (TAM) and the General Attitudes towards Artificial Intelligence Scale (GAAIS), this research seeks to identify the factors influencing students’ behavioral intentions and actual adoption of AI tools in educational settings. A structured survey was administered to students at Multimedia University, Malaysia, capturing their experiences and opinions on widely used AI tools such as ChatGPT, Quillbot, Grammarly, and Perplexity. Hypothesis testing was used to evaluate the statistical significance of relationships between the constructs and behavioral intention and actual use of the AI tools. The findings reveal a high level of engagement with AI tools among University students, primarily driven by their perceived benefits in enhancing academic performance, improving efficiency, and facilitating personalized learning experiences. The findings also uncover significant concerns related to data security, privacy, and the potential over-reliance on AI tools, which may hinder the development of critical thinking and problem-solving skills. Full article
(This article belongs to the Special Issue Security, Privacy, and Trust in Artificial Intelligence Applications)
Show Figures

Figure 1

20 pages, 5358 KiB  
Article
Machine Learning-Based Classification of Sulfide Mineral Spectral Emission in High Temperature Processes
by Carlos Toro, Walter Díaz, Gonzalo Reyes, Miguel Peña, Nicolás Caselli, Carla Taramasco, Pablo Ormeño-Arriagada and Eduardo Balladares
Big Data Cogn. Comput. 2025, 9(5), 130; https://doi.org/10.3390/bdcc9050130 - 14 May 2025
Viewed by 218
Abstract
Accurate classification of sulfide minerals during combustion is essential for optimizing pyrometallurgical processes such as flash smelting, where efficient combustion impacts resource utilization, energy efficiency, and emission control. This study presents a deep learning-based approach for classifying visible and near-infrared (VIS-NIR) emission spectra [...] Read more.
Accurate classification of sulfide minerals during combustion is essential for optimizing pyrometallurgical processes such as flash smelting, where efficient combustion impacts resource utilization, energy efficiency, and emission control. This study presents a deep learning-based approach for classifying visible and near-infrared (VIS-NIR) emission spectra from the combustion of high-grade sulfide minerals. A one-dimensional convolutional neural network (1D-CNN) was developed and trained on experimentally acquired spectral data, achieving a balanced accuracy score of 99.0% in a test set. The optimized deep learning model outperformed conventional machine learning methods, highlighting the effectiveness of deep learning for spectral analysis in high-temperature environments. In addition, Gradient-weighted Class Activation Mapping (Grad-CAM) was applied to enhance model interpretability and identify key spectral regions contributing to classification decisions. The results demonstrated that the model successfully distinguished spectral features associated with different mineral species, offering insights into combustion dynamics. These findings support the potential integration of deep learning for real-time spectral monitoring in industrial flash smelting operations, thereby enabling more precise process control and decision-making. Full article
(This article belongs to the Special Issue Machine Learning and AI Technology for Sustainable Development)
Show Figures

Figure 1

21 pages, 9562 KiB  
Article
Identifying Influential Nodes in Complex Networks via Transformer with Multi-Scale Feature Fusion
by Tingshuai Jiang, Yirun Ruan, Tianyuan Yu, Liang Bai and Yifei Yuan
Big Data Cogn. Comput. 2025, 9(5), 129; https://doi.org/10.3390/bdcc9050129 - 14 May 2025
Viewed by 193
Abstract
In complex networks, the identification of critical nodes is vital for optimizing information dissemination. Given the significant role of these nodes in network structures, researchers have proposed various identification methods. In recent years, deep learning has emerged as a promising approach for identifying [...] Read more.
In complex networks, the identification of critical nodes is vital for optimizing information dissemination. Given the significant role of these nodes in network structures, researchers have proposed various identification methods. In recent years, deep learning has emerged as a promising approach for identifying key nodes in networks. However, existing algorithms fail to effectively integrate local and global structural information, leading to incomplete and limited network understanding. To overcome this limitation, we introduce a transformer framework with multi-scale feature fusion (MSF-Former). In this framework, we construct local and global feature maps for nodes and use them as input. Through the transformer module, node information is effectively aggregated, thereby improving the model’s ability to recognize key nodes. We perform evaluations using six real-world and three synthetic network datasets, comparing our method against multiple baselines using the SIR model to validate its effectiveness. Experimental analysis confirms that MSF-Former achieves consistently high accuracy in the identification of influential nodes across real-world and synthetic networks. Full article
(This article belongs to the Special Issue Advances in Complex Networks)
Show Figures

Figure 1

21 pages, 12662 KiB  
Review
Benchmarking of Anomaly Detection Methods for Industry 4.0: Evaluation, Ranking, and Practical Recommendations
by Aurélie Cools, Mohammed Amin Belarbi and Sidi Ahmed Mahmoudi
Big Data Cogn. Comput. 2025, 9(5), 128; https://doi.org/10.3390/bdcc9050128 - 13 May 2025
Viewed by 307
Abstract
Quality control and predictive maintenance are two essential pillars of Industry 4.0, aiming to optimize production, reduce operational costs, and enhance system reliability. Real-time visual inspection ensures early detection of manufacturing defects, assembly errors, or texture inconsistencies, preventing defective products from reaching customers. [...] Read more.
Quality control and predictive maintenance are two essential pillars of Industry 4.0, aiming to optimize production, reduce operational costs, and enhance system reliability. Real-time visual inspection ensures early detection of manufacturing defects, assembly errors, or texture inconsistencies, preventing defective products from reaching customers. Predictive maintenance leverages sensor data by analyzing vibrations, temperature, and pressure signals to anticipate failures and avoid production downtime. Image-based quality control has become critical in industries such as automotive, electronics, aerospace, and food processing, where visual appearance is a key quality indicator. Although advances in deep learning and computer vision have significantly improved anomaly detection, industrial deployments remain challenged by the scarcity of labeled anomalies and the variability of defects. These issues increasingly lead to the adoption of unsupervised methods and generative approaches, which, despite their effectiveness, introduce substantial computational complexity. We conduct a unified comparison of ten anomaly detection methods, categorizing them according to their reliance on synthetic anomaly generation and their detection strategy, either reconstruction-based or feature-based. All models are trained exclusively on normal data to mirror realistic industrial conditions. Our evaluation framework combines performance metrics such as recall, precision, and their harmonic mean, emphasizing the need to minimize false negatives that could lead to critical production failures. In addition, we assess environmental impact and hardware complexity to better guide method selection. Practical recommendations are provided to balance robustness, operational feasibility, and sustainability in industrial applications. Full article
(This article belongs to the Special Issue Fault Diagnosis and Detection Based on Deep Learning)
Show Figures

Figure 1

18 pages, 11024 KiB  
Article
Rail Surface Defect Diagnosis Based on Image–Vibration Multimodal Data Fusion
by Zhongmei Wang, Shenao Peng, Wenxiu Ao, Jianhua Liu and Changfan Zhang
Big Data Cogn. Comput. 2025, 9(5), 127; https://doi.org/10.3390/bdcc9050127 - 12 May 2025
Viewed by 282
Abstract
To address the challenges in existing multi-sensor data fusion methods for rail surface defect diagnosis, particularly their limitations in fully exploiting potential synergistic information among multimodal data and effectively bridging the semantic gap between heterogeneous multi-source data, this paper proposes a diagnostic approach [...] Read more.
To address the challenges in existing multi-sensor data fusion methods for rail surface defect diagnosis, particularly their limitations in fully exploiting potential synergistic information among multimodal data and effectively bridging the semantic gap between heterogeneous multi-source data, this paper proposes a diagnostic approach based on a Progressive Joint Representation Graph Attention Fusion Network (PJR-GAFN). The methodology comprises five principal phases: Firstly, shared and specific autoencoders are used to extract joint representations of multimodal features through shared and modality-specific representations. Secondly, a squeeze-and-excitation module is implemented to amplify defect-related features while suppressing non-essential characteristics. Thirdly, a progressive fusion module is introduced to comprehensively utilize cross-modal synergistic information during feature extraction. Fourthly, a source domain classifier and domain discriminator are employed to capture modality-invariant features across different modalities. Finally, the spatial attention aggregation properties of graph attention networks are leveraged to fuse multimodal features, thereby fully exploiting contextual semantic information. Experimental results on real-world rail surface defect datasets from domestic railway lines demonstrate that the proposed method achieves 95% diagnostic accuracy, confirming its effectiveness in rail surface defect detection. Full article
Show Figures

Figure 1

28 pages, 11946 KiB  
Article
Introducing a Novel Fast Neighbourhood Component Analysis–Deep Neural Network Model for Enhanced Driver Drowsiness Detection
by Sama Hussein Al-Gburi, Kanar Alaa Al-Sammak, Ion Marghescu, Claudia Cristina Oprea, Ana-Maria Claudia Drăgulinescu, George Suciu, Khattab M. Ali Alheeti, Nayef A. M. Alduais and Nawar Alaa Hussein Al-Sammak
Big Data Cogn. Comput. 2025, 9(5), 126; https://doi.org/10.3390/bdcc9050126 - 8 May 2025
Viewed by 293
Abstract
Driver fatigue is a key factor in road accidents worldwide, requiring effective real-time detection mechanisms. Traditional deep neural network (DNN)-based solutions have shown promising results in detecting drowsiness; however, they are often less suitable for real-time applications due to their high computational complexity, [...] Read more.
Driver fatigue is a key factor in road accidents worldwide, requiring effective real-time detection mechanisms. Traditional deep neural network (DNN)-based solutions have shown promising results in detecting drowsiness; however, they are often less suitable for real-time applications due to their high computational complexity, risk of overfitting, and reliance on large datasets. Hence, this paper introduces an innovative approach that integrates fast neighbourhood component analysis (FNCA) with a deep neural network (DNN) to enhance the detection of driver drowsiness using electroencephalogram (EEG) data. FNCA is employed to optimize feature representation, effectively highlighting critical features for drowsiness detection, which are then analysed using a DNN to achieve high accuracy in recognizing signs of driver fatigue. Our model has been evaluated on the SEED-VIG dataset and achieves state-of-the-art accuracy: 94.29% when trained on 12 subjects and 90.386% with 21 subjects, surpassing existing methods such as TSception, ConvNeXt LMDA-Net, and CNN + LSTM. Full article
Show Figures

Figure 1

23 pages, 3804 KiB  
Article
Quantifying Post-Purchase Service Satisfaction: A Topic–Emotion Fusion Approach with Smartphone Data
by Peijun Guo, Huan Li and Xinyue Mo
Big Data Cogn. Comput. 2025, 9(5), 125; https://doi.org/10.3390/bdcc9050125 - 8 May 2025
Viewed by 306
Abstract
Effectively identifying factors related to user satisfaction is crucial for evaluating customer experience. This study proposes a two-phase analytical framework that combines natural language processing techniques with hierarchical decision-making methods. In Phase 1, an ERNIE-LSTM-based emotion model (ELEM) is used to detect fake [...] Read more.
Effectively identifying factors related to user satisfaction is crucial for evaluating customer experience. This study proposes a two-phase analytical framework that combines natural language processing techniques with hierarchical decision-making methods. In Phase 1, an ERNIE-LSTM-based emotion model (ELEM) is used to detect fake reviews from 4016 smartphone evaluations collected from JD.com (accuracy: 84.77%, recall: 84.86%, F1 score: 84.81%). The filtered genuine reviews are then analyzed using Biterm Topic Modeling (BTM) to extract key satisfaction-related topics, which are weighted based on sentiment scores and organized into a multi-criteria evaluation matrix through the Analytic Hierarchy Process (AHP). These topics are further clustered into five major factors: user-centered design (70.8%), core performance (10.0%), imaging features (8.6%), promotional incentives (7.8%), and industrial design (2.8%). This framework is applied to a comparative analysis of two smartphone stores, revealing that Huawei Mate 60 Pro emphasizes performance, while Redmi Note 11 5G focuses on imaging capabilities. Further clustering of user reviews identifies six distinct user groups, all prioritizing user-centered design and core performance, but showing differences in other preferences. In Phase 2, a comparison of word frequencies between product reviews and community Q and A content highlights hidden user concerns often missed by traditional single-source sentiment analysis, such as screen calibration and pixel density. These findings provide insights into how product design influences satisfaction and offer practical guidance for improving product development and marketing strategies. Full article
Show Figures

Figure 1

20 pages, 1750 KiB  
Article
Enhancing Recommendation Systems with Real-Time Adaptive Learning and Multi-Domain Knowledge Graphs
by Zeinab Shahbazi, Rezvan Jalali and Zahra Shahbazi
Big Data Cogn. Comput. 2025, 9(5), 124; https://doi.org/10.3390/bdcc9050124 - 8 May 2025
Viewed by 312
Abstract
In the era of information explosion, recommendation systems play a crucial role in filtering vast amounts of content for users. Traditional recommendation models leverage knowledge graphs, sentiment analysis, social capital, and generative AI to enhance personalization. However, existing models still struggle to adapt [...] Read more.
In the era of information explosion, recommendation systems play a crucial role in filtering vast amounts of content for users. Traditional recommendation models leverage knowledge graphs, sentiment analysis, social capital, and generative AI to enhance personalization. However, existing models still struggle to adapt dynamically to users’ evolving interests across multiple content domains in real-time. To address this gap, the cross-domain adaptive recommendation system (CDARS) is proposed, which integrates real-time behavioral tracking with multi-domain knowledge graphs to refine user preference modeling continuously. Unlike conventional methods that rely on static or historical data, CDARS dynamically adjusts its recommendation strategies based on contextual factors such as real-time engagement, sentiment fluctuations, and implicit preference drifts. Furthermore, a novel explainable adaptive learning (EAL) module was introduced, providing transparent insights into recommendations’ evolving nature, thereby improving user trust and system interpretability. To enable such real-time adaptability, CDARS incorporates multimodal sentiment analysis of user-generated content, behavioral pattern mining (e.g., click timing, revisit frequency), and learning trajectory modeling through time-aware embeddings and incremental updates of user representations. These dynamic signals are mapped into evolving knowledge graphs, forming continuously updated learning charts that drive more context-aware and emotionally intelligent recommendations. Our experimental results on datasets spanning social media, e-commerce, and entertainment domains demonstrate that CDARS significantly enhances recommendation relevance, achieving an average improvement of 7.8% in click-through rate (CTR) and 8.3% in user engagement compared to state-of-the-art models. This research presents a paradigm shift toward truly dynamic and explainable recommendation systems, creating a way for more personalized and user-centric experiences in the digital landscape. Full article
Show Figures

Figure 1

20 pages, 3241 KiB  
Article
Assessing the Transformation of Armed Conflict Types: A Dynamic Approach
by Dong Jiang, Jun Zhuo, Peiwei Fan, Fangyu Ding, Mengmeng Hao, Shuai Chen, Jiping Dong and Jiajie Wu
Big Data Cogn. Comput. 2025, 9(5), 123; https://doi.org/10.3390/bdcc9050123 - 8 May 2025
Viewed by 328
Abstract
Armed conflict is a dynamic social phenomenon, yet existing research often overlooks its evolving nature. We propose a method to simulate the dynamic transformations of armed conflicts. First, we enhanced the Spatial Conflict Dynamic Indicator (SCDi) by integrating conflict intensity and clustering, which [...] Read more.
Armed conflict is a dynamic social phenomenon, yet existing research often overlooks its evolving nature. We propose a method to simulate the dynamic transformations of armed conflicts. First, we enhanced the Spatial Conflict Dynamic Indicator (SCDi) by integrating conflict intensity and clustering, which allowed for the distinction of various conflict types. Second, we established transformation rules for the SCDi, quantifying five types of transformations: outbreak, stabilization, escalation, de-escalation, and maintaining peace. Using the random forest algorithm with multiple covariates, we simulated these transformations and analyzed the driving factors. The results reveal a global trend of polarization in armed conflicts over the past 20 years, with an increase in clustered/high-intensity (CH) and dispersed/low-intensity (DL) conflicts. Stable regions of ongoing conflict have emerged, notably in areas like Syria, the border of Afghanistan, and Nepal’s border region. New conflicts are more likely to arise near these zones. Various driving forces shape conflict transformations, with neighboring conflict scenarios acting as key catalysts. The capacity of a region to maintain peace largely depends on neighboring conflict dynamics, while local factors are more influential in other types of transformations. This study quantifies the dynamic process of conflict transformations and reveals detailed changes. Full article
Show Figures

Figure 1

19 pages, 5047 KiB  
Article
Robust Anomaly Detection of Multivariate Time Series Data via Adversarial Graph Attention BiGRU
by Yajing Xing, Jinbiao Tan, Rui Zhang and Jiafu Wan
Big Data Cogn. Comput. 2025, 9(5), 122; https://doi.org/10.3390/bdcc9050122 - 8 May 2025
Viewed by 274
Abstract
Multivariate time series data (MTSD) anomaly detection due to complex spatio-temporal dependencies among sensors and pervasive environmental noise. The existing methods struggle to balance anomaly detection accuracy with robustness against data contamination. Hence, this paper proposes a robust multivariate temporal data anomaly detection [...] Read more.
Multivariate time series data (MTSD) anomaly detection due to complex spatio-temporal dependencies among sensors and pervasive environmental noise. The existing methods struggle to balance anomaly detection accuracy with robustness against data contamination. Hence, this paper proposes a robust multivariate temporal data anomaly detection method based on graph attention for training convolutional neural networks (PGAT-BiGRU-NRA). Firstly, the parallel graph attention (PGAT) mechanism extracts the time-dependent and spatially related features of MTSD to realize the MTSD fusion. Then, a bidirectional gate recurrent unit (BiGRU) is utilized to extract the contextual information of the data to avoid information loss. In addition, reconstructing the noise for adversarial training aims to achieve a more robust anomaly detection of MTSD. The experiments conducted on real industrial equipment datasets evaluate the effectiveness of the method in the task of MTSD, and the comparative experiments verify that the proposed method outperforms the mainstream baseline model. The proposed method achieves anomaly detection and robust performance in noise interference, which provides feasible technical support for the stable operation of industrial equipment in complex environments. Full article
Show Figures

Figure 1

21 pages, 2085 KiB  
Article
Edge vs. Cloud: Empirical Insights into Data-Driven Condition Monitoring
by Chikumbutso Christopher Walani and Wesley Doorsamy
Big Data Cogn. Comput. 2025, 9(5), 121; https://doi.org/10.3390/bdcc9050121 - 8 May 2025
Viewed by 349
Abstract
This study evaluates edge and cloud computing paradigms in the context of data-driven condition monitoring of rotating electrical machines. Two well-known platforms, the Raspberry Pi and Amazon Web Services Elastic Compute Cloud, are used to compare and contrast these two computing paradigms in [...] Read more.
This study evaluates edge and cloud computing paradigms in the context of data-driven condition monitoring of rotating electrical machines. Two well-known platforms, the Raspberry Pi and Amazon Web Services Elastic Compute Cloud, are used to compare and contrast these two computing paradigms in terms of different metrics associated with their application suitability. The tested induction machine fault diagnosis models are developed using popular algorithms, namely support vector machines, k-nearest neighbours, and decision trees. The findings reveal that while the cloud platform offers superior computational and memory resources, making it more suitable for complex machine learning tasks, it also incurs higher costs and latency. On the other hand, the edge platform excels in real-time processing and reduces network data burden, but its computational and memory resources are found to be a limitation with certain tasks. The study provides both quantitative and qualitative insights into the trade-offs involved in selecting the most suitable computing approach for condition monitoring applications. Although the scope of the empirical study is primarily limited to factors such as computational efficiency, scalability, and resource utilisation, particularly in the context of specific machine learning models, this paper offers broader discussion and future research directions of other key issues, including latency, network variability, and energy consumption. Full article
(This article belongs to the Special Issue Application of Cloud Computing in Industrial Internet of Things)
Show Figures

Figure 1

22 pages, 12121 KiB  
Article
A Computational–Cognitive Model of Audio-Visual Attention in Dynamic Environments
by Hamideh Yazdani, Alireza Bosaghzadeh, Reza Ebrahimpour and Fadi Dornaika
Big Data Cogn. Comput. 2025, 9(5), 120; https://doi.org/10.3390/bdcc9050120 - 6 May 2025
Viewed by 179
Abstract
Human visual attention is influenced by multiple factors, including visual, auditory, and facial cues. While integrating auditory and visual information enhances prediction accuracy, many existing models rely solely on visual-temporal data. Inspired by cognitive studies, we propose a computational model that combines spatial, [...] Read more.
Human visual attention is influenced by multiple factors, including visual, auditory, and facial cues. While integrating auditory and visual information enhances prediction accuracy, many existing models rely solely on visual-temporal data. Inspired by cognitive studies, we propose a computational model that combines spatial, temporal, face (low-level and high-level visual cues), and auditory saliency to predict visual attention more effectively. Our approach processes video frames to generate spatial, temporal, and face saliency maps, while an audio branch localizes sound-producing objects. These maps are then integrated to form the final audio-visual saliency map. Experimental results on the audio-visual dataset demonstrate that our model outperforms state-of-the-art image and video saliency models and the basic model and aligns more closely with behavioral and eye-tracking data. Additionally, ablation studies highlight the contribution of each information source to the final prediction. Full article
Show Figures

Figure 1

26 pages, 2193 KiB  
Article
Discovering Key Successful Factors of Mobile Advertisements Using Feature Selection Approaches
by Kai-Fu Yang, Venkateswarlu Nalluri, Chun-Cheng Liu and Long-Sheng Chen
Big Data Cogn. Comput. 2025, 9(5), 119; https://doi.org/10.3390/bdcc9050119 - 5 May 2025
Viewed by 329
Abstract
Programmatic buying has attracted growing interest from manufacturers and has become a driving force behind the growth of digital advertising. Among various formats, mobile advertisements (ads) have emerged as a preferred choice over traditional ones due to their advanced automation, adaptability, and cost-effectiveness. [...] Read more.
Programmatic buying has attracted growing interest from manufacturers and has become a driving force behind the growth of digital advertising. Among various formats, mobile advertisements (ads) have emerged as a preferred choice over traditional ones due to their advanced automation, adaptability, and cost-effectiveness. Despite their increasing adoption, academic research on mobile ads remains relatively limited. Unlike conventional statistical analysis techniques, the proposed feature selection methods eliminate the need for assumptions related to data properties such as independence, normal distribution, and constant variance in regression. Additionally, feature selection techniques have recently gained traction in big data analysis, addressing the limitations inherent in traditional statistical approaches. Consequently, this study aims to determine the key success factors of mobile ads in fostering customer loyalty, offering advertisers valuable insights for optimizing mobile ad design. This study begins by identifying potential factors influencing mobile advertising effectiveness. Then, it applies Support Vector Machine Recursive Feature Elimination (SVM-RFE), correlation-based selection, and consistency-based selection methods to determine the key drivers of customer retention. The findings reveal that “Price” and “Preference” are the most significant contributors to enhancing repurchase intention. Moreover, factors such as “Language”, “Perceived Usefulness”, “Interest”, “Mobile Device”, and “Informativeness” are also essential in maximizing the effectiveness of mobile advertising. Full article
Show Figures

Figure 1

47 pages, 29654 KiB  
Review
A Survey on Object-Oriented Assembly and Disassembly Operations in Nuclear Applications
by Wenxing Liu, Ipek Caliskanelli, Hanlin Niu, Kaiqiang Zhang and Robert Skilton
Big Data Cogn. Comput. 2025, 9(5), 118; https://doi.org/10.3390/bdcc9050118 - 28 Apr 2025
Viewed by 267
Abstract
Nuclear environments demand exceptional precision, reliability, and safety, given the high stakes involved in handling radioactive materials and maintaining reactor systems. Object-oriented assembly and disassembly operations in nuclear applications represent a cutting-edge approach to managing complex, high-stakes operations with enhanced precision and safety. [...] Read more.
Nuclear environments demand exceptional precision, reliability, and safety, given the high stakes involved in handling radioactive materials and maintaining reactor systems. Object-oriented assembly and disassembly operations in nuclear applications represent a cutting-edge approach to managing complex, high-stakes operations with enhanced precision and safety. This paper discusses the challenges associated with nuclear robotic remote operations, summarizes current methods for handling object-oriented assembly and disassembly operations, and explores potential future research directions in this field. Object-oriented assembly and disassembly operations are vital in nuclear applications due to their ability to manage complexity, ensure precision, and enhance safety and reliability, all of which are paramount in the demanding and high-risk environment of nuclear technology. Full article
(This article belongs to the Special Issue Field Robotics and Artificial Intelligence (AI))
Show Figures

Figure 1

29 pages, 1391 KiB  
Review
A Review of the State-of-the-Art Techniques and Analysis of Transformers for Bengali Text Summarization
by MD Iftekharul Mobin, Mahamodul Hasan Mahadi, Al-Sakib Khan Pathan and A. F. M. Suaib Akhter
Big Data Cogn. Comput. 2025, 9(5), 117; https://doi.org/10.3390/bdcc9050117 - 28 Apr 2025
Viewed by 528
Abstract
Text summarization is a complex and essential task in natural language processing (NLP) research, focused on extracting the most important information from a document. This study focuses on the Extractive and Abstractive approaches of Bengali Text Summarization (BTS). With the breakthrough advancements in [...] Read more.
Text summarization is a complex and essential task in natural language processing (NLP) research, focused on extracting the most important information from a document. This study focuses on the Extractive and Abstractive approaches of Bengali Text Summarization (BTS). With the breakthrough advancements in deep learning, summarization is no longer a major challenge for English, given the availability of extensive resources dedicated to this global language. However, the Bengali language remains underexplored. Hence, in this work, a comprehensive review has been conducted on BTS research from 2007 to 2023, analyzing trends, datasets, preprocessing techniques, methodologies, evaluations, and challenges. Leveraging 106 journal and conference papers, this review offers insights into emerging topics and trends in Bengali Abstractive summarization. The review has been augmented with experiments using transformer models from Hugging Face and publicly available datasets to assess the Rouge score accuracy for Abstractive summarization. The extensive literature review conducted in this study reveals that before the advent of transformers, LSTM (Long Short-Term Memory) models were the dominant deep learning approach for text summarization across various languages. For transformers, one of the key datasets utilized was XL-SUM with the MT5 model emerging as the best performer among various contemporary multilingual models. These findings contribute to understanding the contemporary techniques and challenges in BTS. Furthermore, recommendations are made to guide future research endeavors, aiming to provide valuable insights and directions for researchers in this field. Full article
Show Figures

Figure 1

16 pages, 1091 KiB  
Article
Transferring Natural Language Datasets Between Languages Using Large Language Models for Modern Decision Support and Sci-Tech Analytical Systems
by Dmitrii Popov, Egor Terentev, Danil Serenko, Ilya Sochenkov and Igor Buyanov
Big Data Cogn. Comput. 2025, 9(5), 116; https://doi.org/10.3390/bdcc9050116 - 28 Apr 2025
Viewed by 311
Abstract
The decision-making process to rule R&D relies on information related to current trends in particular research areas. In this work, we investigated how one can use large language models (LLMs) to transfer the dataset and its annotation from one language to another. This [...] Read more.
The decision-making process to rule R&D relies on information related to current trends in particular research areas. In this work, we investigated how one can use large language models (LLMs) to transfer the dataset and its annotation from one language to another. This is crucial since sharing knowledge between different languages could boost certain underresourced directions in the target language, saving lots of effort in data annotation or quick prototyping. We experiment with English and Russian pairs, translating the DEFT (Definition Extraction from Texts) corpus. This corpus contains three layers of annotation dedicated to term-definition pair mining, which is a rare annotation type for Russian. The presence of such a dataset is beneficial for the natural language processing methods of trend analysis in science since the terms and definitions are the basic blocks of any scientific field. We provide a pipeline for the annotation transfer using LLMs. In the end, we train the BERT-based models on the translated dataset to establish a baseline. Full article
Show Figures

Figure 1

25 pages, 3585 KiB  
Article
Enhancing Deep Learning Sustainability by Synchronized Multi Augmentation with Rotations and Multi-Backbone Architectures
by Nikita Gordienko, Yuri Gordienko and Sergii Stirenko
Big Data Cogn. Comput. 2025, 9(5), 115; https://doi.org/10.3390/bdcc9050115 - 27 Apr 2025
Viewed by 190
Abstract
Deep learning applications for Edge Intelligence (EI) face challenges in achieving high model performance while maintaining computational efficiency, particularly under varying image orientations and perspectives. This study investigates the synergy of multi-backbone (MB) configurations and Synchronized Multi Augmentation (SMA) to address these challenges [...] Read more.
Deep learning applications for Edge Intelligence (EI) face challenges in achieving high model performance while maintaining computational efficiency, particularly under varying image orientations and perspectives. This study investigates the synergy of multi-backbone (MB) configurations and Synchronized Multi Augmentation (SMA) to address these challenges by leveraging diverse input representations and spatial transformations. SMA employs synchronously augmented input data across MBs during training, thereby improving feature extraction across diverse representations. The outputs provided by these MBs are merged through different fusion strategies: Averaging Fusion with aggregation of predictions and Dense Fusion with integration of features via a fully connected neural network. It aims to increase model accuracy on previously unseen input data and to reduce computational requirements by minimizing neural network size, particularly advantageous for EI systems characterized by the limited computing resources. This study employed MBs with the MobileNetV3 architecture and the CIFAR-10 dataset to investigate the impact of SMA techniques and different fusion strategies on model robustness and performance. SMA techniques were applied to simulate diverse image orientations, and MB architectures were tested with Averaging and Dense fusion strategies to assess their ability to learn diverse feature representations and improve robustness. The experiments revealed that models augmented with SMA outperformed the baseline MobileNetV3 on modified datasets, achieving higher robustness to orientation variations. Models with Averaging fusion exhibited the most stable performance across datasets, while Dense fusion achieved the highest metrics under specific conditions. Results indicate that SMAs incorporating image transformation adjustments, such as rotation, significantly enhance generalization across varying orientation conditions. This approach enables the production of more stable results using the same pretrained weights in real-world applications by configuring Image Signal Processing (ISP) to effectively use SMA. The findings encourage further exploration of SMA techniques in conjunction with diverse camera sensor configurations and ISP settings to optimize real-world deployments. Full article
(This article belongs to the Special Issue Machine Learning and AI Technology for Sustainable Development)
Show Figures

Figure 1

18 pages, 6190 KiB  
Article
From Accuracy to Vulnerability: Quantifying the Impact of Adversarial Perturbations on Healthcare AI Models
by Sarfraz Brohi and Qurat-ul-ain Mastoi
Big Data Cogn. Comput. 2025, 9(5), 114; https://doi.org/10.3390/bdcc9050114 - 27 Apr 2025
Viewed by 307
Abstract
As AI becomes indispensable in healthcare, its vulnerability to adversarial attacks demands serious attention. Even minimal changes to the input data can mislead Deep Learning (DL) models, leading to critical errors in diagnosis and endangering patient safety. In this study, we developed an [...] Read more.
As AI becomes indispensable in healthcare, its vulnerability to adversarial attacks demands serious attention. Even minimal changes to the input data can mislead Deep Learning (DL) models, leading to critical errors in diagnosis and endangering patient safety. In this study, we developed an optimized Multi-layer Perceptron (MLP) model for breast cancer classification and exposed its cybersecurity vulnerabilities through a real-world-inspired adversarial attack. Unlike prior studies, we conducted a quantitative evaluation on the impact of a Fast Gradient Sign Method (FGSM) attack on an optimized DL model designed for breast cancer detection to demonstrate how minor perturbations reduced the model’s accuracy from 98% to 53%, and led to a substantial increase in the classification errors, as revealed by the confusion matrix. Our findings demonstrate how an adversarial attack can significantly compromise the performance of a healthcare AI model, underscoring the importance of aligning AI development with cybersecurity readiness. This research highlights the demand for designing resilient AI by integrating rigorous cybersecurity practices at every stage of the AI development lifecycle, i.e., before, during, and after the model engineering to prioritize the effectiveness, accuracy, and safety of AI in real-world healthcare environments. Full article
(This article belongs to the Special Issue Big Data Analytics with Machine Learning for Cyber Security)
Show Figures

Figure 1

30 pages, 37454 KiB  
Article
Cognitive Computing for Understanding and Restoring Color in Renaissance Art
by Artyom M. Grigoryan, Sos S. Agaian and Shao Liu
Big Data Cogn. Comput. 2025, 9(5), 113; https://doi.org/10.3390/bdcc9050113 - 23 Apr 2025
Viewed by 334
Abstract
In this article, for the first time on this topic, we analyze the historical color palettes of Renaissance oil paintings by using machine-learning methods and digital images. Our work has two main parts: we collect data on their historical color palettes and then [...] Read more.
In this article, for the first time on this topic, we analyze the historical color palettes of Renaissance oil paintings by using machine-learning methods and digital images. Our work has two main parts: we collect data on their historical color palettes and then use machine learning to predict the original colors of paintings. This model studies color ratios, enhancement levels, symbolic meanings, and historical records. It looks at key colors, measures their relationships, and learns how they have changed. The main contributions of this work are as follows: (i) we develop a model that predicts a painting’s original color palette based on multiple factors, such as the color ratios and symbolic meanings, and (ii) we propose a framework for using cognitive computing tools to recover the original colors of historical artworks. This helps us to rediscover lost emotional and cultural details. Full article
Show Figures

Figure 1

21 pages, 541 KiB  
Article
Cognitive Computing with Large Language Models for Student Assessment Feedback
by Noorhan Abbas and Eric Atwell
Big Data Cogn. Comput. 2025, 9(5), 112; https://doi.org/10.3390/bdcc9050112 - 23 Apr 2025
Viewed by 436
Abstract
Effective student feedback is fundamental to enhancing learning outcomes in higher education. While traditional assessment methods emphasise both achievements and development areas, the process remains time-intensive for educators. This research explores the application of cognitive computing, specifically open-source Large Language Models (LLMs) Mistral-7B [...] Read more.
Effective student feedback is fundamental to enhancing learning outcomes in higher education. While traditional assessment methods emphasise both achievements and development areas, the process remains time-intensive for educators. This research explores the application of cognitive computing, specifically open-source Large Language Models (LLMs) Mistral-7B and CodeLlama-7B, to streamline feedback generation for student reports containing both Python programming elements and English narrative content. The findings indicate that these models can provide contextually appropriate feedback on both technical Python coding and English specification and documentation. They effectively identified coding weaknesses and provided constructive suggestions for improvement, as well as insightful feedback on English language quality, structure, and clarity in report writing. These results contribute to the growing body of knowledge on automated assessment feedback in higher education, offering practical insights for institutions considering the implementation of open-source LLMs in their workflows. There are around 22 thousand assessment submissions per year in the School of Computer Science, which is one of eight schools in the Faculty of Engineering and Physical Sciences, which is one of seven faculties in the University of Leeds, which is one of one hundred and sixty-six universities in the UK, so there is clear potential for our methods to scale up to millions of assessment submissions. This study also examines the limitations of current approaches and proposes potential enhancements. The findings support a hybrid system where cognitive computing manages routine tasks and educators focus on complex, personalised evaluations, enhancing feedback quality, consistency, and efficiency in educational settings. Full article
(This article belongs to the Special Issue Advances in Natural Language Processing and Text Mining)
Previous Issue
Next Issue
Back to TopTop