Journal Description
Big Data and Cognitive Computing
Big Data and Cognitive Computing
is an international, peer-reviewed, open access journal on big data and cognitive computing published monthly online by MDPI.
- Open Access— free for readers, with article processing charges (APC) paid by authors or their institutions.
- High Visibility: indexed within Scopus, ESCI (Web of Science), dblp, Inspec, Ei Compendex, and other databases.
- Journal Rank: JCR - Q1 (Computer Science, Theory and Methods) / CiteScore - Q1 (Computer Science Applications)
- Rapid Publication: manuscripts are peer-reviewed and a first decision is provided to authors approximately 24.5 days after submission; acceptance to publication is undertaken in 4.6 days (median values for papers published in this journal in the first half of 2025).
- Recognition of Reviewers: reviewers who provide timely, thorough peer-review reports receive vouchers entitling them to a discount on the APC of their next publication in any MDPI journal, in appreciation of the work done.
- Journal Cluster of Artificial Intelligence: AI, AI in Medicine, Algorithms, BDCC, MAKE, MTI, Stats, Virtual Worlds and Computers.
Impact Factor:
4.4 (2024);
5-Year Impact Factor:
4.2 (2024)
Latest Articles
Large Language Models in Mechanical Engineering: A Scoping Review of Applications, Challenges, and Future Directions
Big Data Cogn. Comput. 2025, 9(12), 305; https://doi.org/10.3390/bdcc9120305 (registering DOI) - 30 Nov 2025
Abstract
Following PRISMA-ScR guidelines, this scoping review systematically maps the landscape of Large Language Models (LLMs) in mechanical engineering. A search of four major databases (Scopus, IEEE Xplore, ACM Digital Library, Web of Science) and a rigorous screening process yielded 66 studies for final
[...] Read more.
Following PRISMA-ScR guidelines, this scoping review systematically maps the landscape of Large Language Models (LLMs) in mechanical engineering. A search of four major databases (Scopus, IEEE Xplore, ACM Digital Library, Web of Science) and a rigorous screening process yielded 66 studies for final analysis. The findings reveal a nascent, rapidly accelerating field, with over 68% of publications from 2024 (representing a year-on-year growth of 150% from 2023 to 2024), and applications concentrated on front-end design processes like conceptual design and Computer-Aided Design (CAD) generation. The technological landscape is dominated by OpenAI’s GPT-4 variants. A persistent challenge identified is weak spatial and geometric reasoning, shifting the primary research bottleneck from traditional data scarcity to inherent model limitations. This, alongside reliability concerns, forms the main barrier to deeper integration into engineering workflows. A consensus on future directions points to the need for specialized datasets, multimodal inputs to ground models in engineering realities, and robust, engineering-specific benchmarks. This review concludes that LLMs are currently best positioned as powerful ‘co-pilots’ for engineers rather than autonomous designers, providing an evidence-based roadmap for researchers, practitioners, and educators.
Full article
(This article belongs to the Special Issue Artificial Intelligence (AI) and Natural Language Processing (NLP))
►
Show Figures
Open AccessArticle
Development of Traffic Rules Training Platform Using LLMs and Cloud Video Streaming
by
Artem Kazarian, Vasyl Teslyuk, Oleh Berezsky and Oleh Pitsun
Big Data Cogn. Comput. 2025, 9(12), 304; https://doi.org/10.3390/bdcc9120304 (registering DOI) - 30 Nov 2025
Abstract
►▼
Show Figures
Driving safety education remains a critical societal priority, and understanding traffic rules is essential for reducing road accidents and improving driver awareness. This study presents the development and evaluation of a virtual simulator for learning traffic rules, incorporating spherical video technology and interactive
[...] Read more.
Driving safety education remains a critical societal priority, and understanding traffic rules is essential for reducing road accidents and improving driver awareness. This study presents the development and evaluation of a virtual simulator for learning traffic rules, incorporating spherical video technology and interactive training scenarios. The primary objective was to enhance the accessibility and effectiveness of traffic rule education by utilizing modern virtual reality approaches without the need for specialized equipment. A key research component is using Petri net-based models to study the simulator’s dynamic states, enabling the analysis and optimization of system behavior. The developed simulator employs large language models for the automated generation of educational content and test questions, supporting personalized learning experiences. Additionally, a model for determining the camera rotation angle was proposed, ensuring a realistic and immersive presentation of training scenarios within the simulator. The system’s cloud-based, modular software architecture and cross-platform algorithms ensure flexibility, scalability, and compatibility across devices. The simulator allows users to practice traffic rules in realistic road environments with the aid of spherical videos and receive immediate feedback through contextual prompts. The developed system stands out from existing traffic rule learning platforms by combining spherical video technology, large language model-based content generation, and cloud architecture to create a more interactive, adaptive, and realistic learning experience. The experimental results confirm the simulator’s high efficiency in improving users’ knowledge of traffic rules and practical decision-making skills.
Full article

Figure 1
Open AccessArticle
Optimization of Machine Learning Algorithms with Distillation and Quantization for Early Detection of Attacks in Resource-Constrained Systems
by
Mikhail Rusanov, Mikhail Babenko and Maria Lapina
Big Data Cogn. Comput. 2025, 9(12), 303; https://doi.org/10.3390/bdcc9120303 (registering DOI) - 28 Nov 2025
Abstract
►▼
Show Figures
This study addresses the problem of automatic attack detection targeting Linux-based machines and web applications through the analysis of system logs, with a particular focus on reducing the computational requirements of existing solutions. The aim of the research is to develop and evaluate
[...] Read more.
This study addresses the problem of automatic attack detection targeting Linux-based machines and web applications through the analysis of system logs, with a particular focus on reducing the computational requirements of existing solutions. The aim of the research is to develop and evaluate the effectiveness of machine learning models capable of classifying system events as benign or malicious, while also identifying the type of attack under resource-constrained conditions. The Linux-APT-Dataset-2024 was employed as the primary source of data. To mitigate the challenge of high computational complexity, model optimization techniques such as parameter quantization, knowledge distillation, and architectural simplifications were applied. Experimental results demonstrate that the proposed approaches significantly reduce computational overhead and hardware requirements while maintaining high classification accuracy. The findings highlight the potential of optimized machine learning algorithms for the development of practical early threat detection systems in Linux environments with limited resources, which is particularly relevant for deployment in IoT devices and edge computing systems.
Full article

Figure 1
Open AccessArticle
DotA 2 Match Outcome Prediction System Using Decision Tree Ensemble Algorithms
by
Sukhrob Yangibaev, Jamolbek Mattiev and Sello Mokwena
Big Data Cogn. Comput. 2025, 9(12), 302; https://doi.org/10.3390/bdcc9120302 - 27 Nov 2025
Abstract
►▼
Show Figures
This paper explores the replication of the DotA Plus prediction system using decision tree algorithms. The study implements and evaluates Extra Trees Classifier, Random Forest Classifier, and Hist Gradient Boosting Classifier, along with their combined average, for predicting the outcome of Defense of
[...] Read more.
This paper explores the replication of the DotA Plus prediction system using decision tree algorithms. The study implements and evaluates Extra Trees Classifier, Random Forest Classifier, and Hist Gradient Boosting Classifier, along with their combined average, for predicting the outcome of Defense of the Ancients (DotA) 2 matches. Data was collected using the OpenDotA API and the Steam API, and various features such as game duration, tower and barracks states, net-worth, assists, last hits, gold, level, gold per minute, and experience per minute were extracted for model training. Additionally, hero and item win rate features, derived from Dotabuff data, were incorporated to enhance the models’ predictive accuracy. The models were trained on datasets with varying match durations, including segments for matches under 10 min, between 10 and 20 min, and over 20 min. The experimental results show that the Extra Trees Classifier consistently outperformed other individual models and performed comparably to the averaged models, achieving a peak performance of 98.6% test accuracy on matches longer than 20 min when using match duration segmentation and hero/item embeddings. The study highlights the effectiveness of decision tree-based methods for real-time match outcome prediction in DotA 2 and offers insights into feature importance. The combined average of Extra Trees Classifier, Random Forest Classifier, and Hist Gradient Boosting Classifier models provides a robust and reliable prediction of DotA 2 match outcomes, thus showing potential as a real-time prediction system.
Full article

Figure 1
Open AccessArticle
Crime Spatiotemporal Prediction Through Urban Region Representation by Using Building Footprints
by
Tao Wang, Peng Chen and Miaoxuan Shan
Big Data Cogn. Comput. 2025, 9(12), 301; https://doi.org/10.3390/bdcc9120301 - 27 Nov 2025
Abstract
►▼
Show Figures
Current crime spatiotemporal prediction models are limited by the insufficient ability of POI data to represent the continuity and mixed-use nature of urban spatial functions. To address this, our study applies an urban region representation method based on building footprints and validates its
[...] Read more.
Current crime spatiotemporal prediction models are limited by the insufficient ability of POI data to represent the continuity and mixed-use nature of urban spatial functions. To address this, our study applies an urban region representation method based on building footprints and validates its effectiveness in improving the accuracy of crime spatiotemporal prediction. Specially, we first use the Region Dual Contrastive Learning algorithm to generate region representations as a region graph by integrating building footprints and POI data. Then, the region graph combined with crime data is input into crime prediction models to predict four crime types, including Burglary, Robbery, Felony Assault, and Grand Larceny. Finally, ablation experiments are conducted to quantify the contribution of building footprints to prediction improvement. The experimental results on New York City crime data indicate that (1) the region representations significantly improve deep learning model performance, with the most improved LSTM achieving average increases of 5.66% in Macro-F1 and 18.57% in Micro-F1, particularly benefiting baseline models with lower accuracy, and (2) the region representations yield more significant improvements for low-frequency crime categories and mitigates temporal memory decay in long-term predictions. These findings confirm that incorporating urban region representation based on building footprints effectively enhances crime spatiotemporal prediction performance, providing a more precise and efficient tool for urban security management to optimize police resource allocation and crime prevention strategies.
Full article

Figure 1
Open AccessArticle
Unordered Stacked Pillbox Detection Algorithm Based on Improved YOLOv8
by
Jiahang Pan, Rui Zhou, Jie Feng, Mincheng Wu, Xiang Wu and Hui Dong
Big Data Cogn. Comput. 2025, 9(12), 300; https://doi.org/10.3390/bdcc9120300 - 26 Nov 2025
Abstract
To enable fully automated medicine warehousing in intelligent pharmacy systems, accurately detecting disordered, stacked pillboxes is essential. This paper proposes a high-precision detection algorithm for such scenarios based on an improved YOLOv8 framework. The proposed method integrates a novel convolutional module that replaces
[...] Read more.
To enable fully automated medicine warehousing in intelligent pharmacy systems, accurately detecting disordered, stacked pillboxes is essential. This paper proposes a high-precision detection algorithm for such scenarios based on an improved YOLOv8 framework. The proposed method integrates a novel convolutional module that replaces traditional stride convolutions and pooling layers, enhancing the detection of small, low-resolution targets in computer vision tasks. To further enhance detection accuracy, the Bi-Level Routing Attention (BiFormer) Vision Transformer is incorporated as a Cognitive Computing module. Additionally, the circular Smooth Label (CSL) technique is employed to mitigate boundary discontinuities and periodic anomalies in angle prediction, which often arise in the detection of rotated objects. The experimental results demonstrate that the proposed method achieves a precision of 94.24%, a recall of 90.39%, and a mean average precision (mAP) of 94.16%—improvements of 3.34%, 2.53%, and 3.35%, respectively, over the baseline YOLOv8 model. Moreover, the enhanced detection model outperforms existing rotated-object detection methods while maintaining real-time inference speed. To facilitate reproducibility and future benchmarking, the full dataset and source code used in this study have been released publicly. Although no standardized benchmark currently exists for pillbox detection, our self-constructed dataset reflects key industrial variations in pillbox size, orientation, and stacking, thereby providing a foundation for future cross-domain validation.
Full article
(This article belongs to the Special Issue AI, Computer Vision and Human–Robot Interaction)
►▼
Show Figures

Figure 1
Open AccessArticle
CURE: Confidence-Driven Unified Reasoning Ensemble Framework for Medical Question Answering
by
Ziad Elshaer and Essam A. Rashed
Big Data Cogn. Comput. 2025, 9(12), 299; https://doi.org/10.3390/bdcc9120299 - 23 Nov 2025
Abstract
High-performing medical Large Language Models (LLMs) typically require extensive fine-tuning with substantial computational resources, limiting accessibility for resource-constrained healthcare institutions. This study introduces a confidence-driven multi-model framework that leverages model diversity to enhance medical question answering without fine-tuning. Our framework employs a two-stage
[...] Read more.
High-performing medical Large Language Models (LLMs) typically require extensive fine-tuning with substantial computational resources, limiting accessibility for resource-constrained healthcare institutions. This study introduces a confidence-driven multi-model framework that leverages model diversity to enhance medical question answering without fine-tuning. Our framework employs a two-stage architecture: a confidence detection module assesses the primary model’s certainty, and an adaptive routing mechanism directs low-confidence queries to Helper models with complementary knowledge for collaborative reasoning. We evaluate our approach using Qwen3-30B-A3B-Instruct, Phi-4 14B, and Gemma 2 12B across three medical benchmarks; MedQA, MedMCQA, and PubMedQA. Results demonstrate that our framework achieves competitive performance, with particularly strong results in PubMedQA (0.95) and MedMCQA (0.78). Ablation studies confirm that confidence-aware routing combined with multi-model collaboration substantially outperforms single-model approaches and uniform reasoning strategies. This work establishes that strategic model collaboration offers a practical, computationally efficient pathway to improve medical AI systems, with significant implications for democratizing access to advanced medical AI in resource-limited settings.
Full article
(This article belongs to the Special Issue Advances in Large Language Models for Biological and Medical Applications)
►▼
Show Figures

Figure 1
Open AccessArticle
Metadata Suffices: Optimizer-Aware Fake Account Detection with Minimal Multimodal Input
by
Ziad Elgammal, Khaled Elgammal and Reda Alhajj
Big Data Cogn. Comput. 2025, 9(12), 298; https://doi.org/10.3390/bdcc9120298 - 21 Nov 2025
Abstract
►▼
Show Figures
Social media platforms are currently confronted with a substantial problem concerning the presence of fake accounts, which pose a threat by spreading harmful content, spam, and misinformation. This study aims to address the problem by differentiating between fake and real X accounts (formerly
[...] Read more.
Social media platforms are currently confronted with a substantial problem concerning the presence of fake accounts, which pose a threat by spreading harmful content, spam, and misinformation. This study aims to address the problem by differentiating between fake and real X accounts (formerly Twitter). The need to mitigate the negative impact of fake accounts on online communities serves as the driving force for this work, with the goal of developing an effective method for identifying fake accounts and their fraudulent activities, such as posting harmful links, engaging in spamming behaviors, and disrupting online communities. The scope of this work focuses specifically on fake Twitter account detection. A comprehensive approach is taken, leveraging user information and tweets to discern between genuine and fake accounts. Various deep learning architectures are proposed and implemented, utilizing different optimizers and evaluating performance metrics. The models are trained and tested using a collected dataset, augmented to cover diverse real-life scenarios. The results show promising progress in distinguishing between fake and real accounts, revealing that the inclusion of tweet content along with user metadata does not significantly improve the classification of fake accounts. It also highlights the importance of selecting appropriate optimizers. The implications of this study are relevant to social media platforms, users, and researchers. The findings provide insights into combating fake accounts and their fraudulent activities, contributing to the enhancement of online community safety. While the research is specific to Twitter, the methodology and insights gained may be potentially generalizable to other social media platforms.
Full article

Figure 1
Open AccessArticle
Interpretable Predictive Modeling for Educational Equity: A Workload-Aware Decision Support System for Early Identification of At-Risk Students
by
Aigul Shaikhanova, Oleksandr Kuznetsov, Kainizhamal Iklassova, Aizhan Tokkuliyeva and Laura Sugurova
Big Data Cogn. Comput. 2025, 9(11), 297; https://doi.org/10.3390/bdcc9110297 - 20 Nov 2025
Abstract
Educational equity and access to quality learning opportunities represent fundamental pillars of sustainable societal development, directly aligned with the United Nations Sustainable Development Goal 4 (Quality Education). Student retention remains a critical challenge in higher education, with early disengagement strongly predicting eventual failure
[...] Read more.
Educational equity and access to quality learning opportunities represent fundamental pillars of sustainable societal development, directly aligned with the United Nations Sustainable Development Goal 4 (Quality Education). Student retention remains a critical challenge in higher education, with early disengagement strongly predicting eventual failure and limiting opportunities for social mobility. While machine learning models have demonstrated impressive predictive accuracy for identifying at-risk students, most systems prioritize performance metrics over practical deployment constraints, creating a gap between research demonstrations and real-world impact for social good. We present an accountable and interpretable decision support system that balances three competing objectives essential for responsible AI deployment: ultra-early prediction timing (day 14 of semester), manageable instructor workload (flagging 15% of students), and model transparency (multiple explanation mechanisms). Using the Open University Learning Analytics Dataset (OULAD) containing 22,437 students across seven modules, we develop predictive models from activity patterns, assessment performance, and demographics observable within two weeks. We compare threshold-based rules, logistic regression (interpretable linear modeling), and gradient boosting (ensemble modeling) using temporal validation where early course presentations train models tested on later cohorts. Results show gradient boosting achieves AUC (Area Under the ROC Curve, measuring discrimination ability) of 0.789 and average precision of 0.722, with logistic regression performing nearly identically (AUC 0.783, AP 0.713), revealing that linear modeling captures most predictive signal and makes interpretability essentially free. At our recommended threshold of 0.607, the predictive model flags 15% of students with 84% precision and 35% recall, creating actionable alert lists instructors can manage within normal teaching duties while maintaining accountability for false positives. Calibration analysis confirms that predicted probabilities match observed failure rates, ensuring trustworthy risk estimates. Feature importance modeling reveals that assessment completion and activity patterns dominate demographic factors, providing transparent evidence that behavioral engagement matters more than student background. We implement a complete decision support system generating instructor reports, explainable natural language justifications for each alert, and personalized intervention templates. Our contribution advances responsible AI for social good by demonstrating that interpretable predictive modeling can support equitable educational outcomes when designed with explicit attention to timing, workload, and transparency—core principles of accountable artificial intelligence.
Full article
(This article belongs to the Special Issue Applied Data Science for Social Good: 2nd Edition)
►▼
Show Figures

Figure 1
Open AccessArticle
Music Genre Classification Using Prosodic, Stylistic, Syntactic and Sentiment-Based Features
by
Erik-Robert Kovacs and Stefan Baghiu
Big Data Cogn. Comput. 2025, 9(11), 296; https://doi.org/10.3390/bdcc9110296 - 19 Nov 2025
Abstract
Romanian popular music has had a storied history across the last century and a half. Incorporating different influences at different times, today it boasts a wide range of both autochthonous and imported genres, such as traditional folk music, rock, rap, pop, and manele,
[...] Read more.
Romanian popular music has had a storied history across the last century and a half. Incorporating different influences at different times, today it boasts a wide range of both autochthonous and imported genres, such as traditional folk music, rock, rap, pop, and manele, to name a few. We aim to trace the linguistic differences between the lyrics of these genres using natural language processing and a computational linguistics approach by studying the prosodic, stylistic, syntactic, and sentiment-based features of each genre. For this purpose, we have crawled a dataset of ~14,000 Romanian songs from publicly available websites along with the user-provided genre labels, and characterized each song and each genre, respectively, with regard to these features, discussing similarities and differences. We improve on existing tools for Romanian language natural language processing by building a lexical analysis library well suited to song lyrics or poetry which encodes a set of 17 linguistic features. In addition, we build lexical analysis tools for profanity-based features and improve the SentiLex sentiment analysis library by manually rebalancing its lexemes to overcome the limitations introduced by it having been machine translated into Romanian. We estimate the accuracy gain using a benchmark Romanian sentiment analysis dataset and register a 25% increase in accuracy over the SentiLex baseline. The contribution is meant to describe the characteristics of the Romanian expression of autochthonous as well as international genres and provide technical support to researchers in natural language processing, musicology or the digital humanities in studying the lyrical content of Romanian music. We have released our data and code for research use.
Full article
(This article belongs to the Special Issue Artificial Intelligence (AI) and Natural Language Processing (NLP))
►▼
Show Figures

Figure 1
Open AccessArticle
FL-Swarm MRCM: A Novel Federated Learning Framework for Cross-Site Medical Image Reconstruction
by
Ailya Izhar and Syed Muhammad Anwar
Big Data Cogn. Comput. 2025, 9(11), 295; https://doi.org/10.3390/bdcc9110295 - 19 Nov 2025
Abstract
Magnetic Resonance Imaging (MRI) reconstruction is computationally heavy from under sampled data, and centralized data sharing within deep learning models was met with privacy concerns. We therefore propose FL-Swarm MRCM, a novel federated learning framework that integrates FedDyn dynamic regularization, a swarm-optimized generative
[...] Read more.
Magnetic Resonance Imaging (MRI) reconstruction is computationally heavy from under sampled data, and centralized data sharing within deep learning models was met with privacy concerns. We therefore propose FL-Swarm MRCM, a novel federated learning framework that integrates FedDyn dynamic regularization, a swarm-optimized generative adversarial network (SwarmGAN), and a structure-aware cross-entropy loss to enhance cross-site MRI reconstruction without sharing raw data. The framework avoids client drift, locally adapts hyper-parameters using Particle Swarm Optimization, and preserves anatomic fidelity. Evaluations on fastMRI, BraTS-2020, and OASIS datasets under non-IID partitions show that FL-Swarm MRCM improves reconstruction quality, achieving PSNR = 29.78 dB and SSIM = 0.984, outscoring FL-MR and FL-MRCM baselines. The federated framework for adversarial training proposed here enables reproducible, privacy-preserving, and strongly multi-institutional MRI reconstruction with better cross-site generalization for clinical use.
Full article
(This article belongs to the Topic Internet of Things Architectures, Applications, and Strategies: Emerging Paradigms, Technologies, and Advancing AI Integration)
►▼
Show Figures

Figure 1
Open AccessArticle
Pervasive Auto-Scaling Method for Improving the Quality of Resource Allocation in Cloud Platforms
by
Vimal Raja Rajasekar and G. Santhi
Big Data Cogn. Comput. 2025, 9(11), 294; https://doi.org/10.3390/bdcc9110294 - 18 Nov 2025
Abstract
►▼
Show Figures
Cloud resource provider deployment at random locations increases operational costs regardless of the application demand intervals. To provide adaptable load balancing under varying application traffic intervals, the auto-scaling concept has been introduced. This article introduces a Pervasive Auto-Scaling Method (PASM) for Computing Resource
[...] Read more.
Cloud resource provider deployment at random locations increases operational costs regardless of the application demand intervals. To provide adaptable load balancing under varying application traffic intervals, the auto-scaling concept has been introduced. This article introduces a Pervasive Auto-Scaling Method (PASM) for Computing Resource Allocation (CRA) to improve the application quality of service. In this auto-scaling method, deep reinforcement learning is employed to verify shared instances of up-scaling and down-scaling pervasively. The overflowing application demands are computed for their service failures and are used to train the learning network. In this process, the scaling is decided based on the maximum computing resource allocation to the demand ratio. Therefore, the learning network is also trained using scaling rates from the previous (completed) allocation intervals. This process is thus recurrent until maximum resource allocation with high sharing is achieved. The resource provider migrates to reduce the wait time based on the high-to-low demand ratio between successive computing intervals. This enhances the resource allocation rate without high wait times. The proposed method’s performance is validated using the metrics resource allocation rate, service delay, allocated wait time, allocation failures, and resource utilization.
Full article

Figure 1
Open AccessArticle
Enhancing POI Recognition with Micro-Level Tagging and Deep Learning
by
Paraskevas Messios, Ioanna Dionysiou and Harald Gjermundrød
Big Data Cogn. Comput. 2025, 9(11), 293; https://doi.org/10.3390/bdcc9110293 - 15 Nov 2025
Abstract
►▼
Show Figures
Background: Understanding visual context in images is essential for enhanced Point-of-Interest (POI) recommender systems. Traditional models often rely on global features, overlooking object-level information, which can limit contextual accuracy. Methods: This study introduces micro-level contextual tagging, a method for extracting metadata from individual
[...] Read more.
Background: Understanding visual context in images is essential for enhanced Point-of-Interest (POI) recommender systems. Traditional models often rely on global features, overlooking object-level information, which can limit contextual accuracy. Methods: This study introduces micro-level contextual tagging, a method for extracting metadata from individual objects in images, including object type, frequency, and color. This enriched information is used to train WORLDO, a Vision Transformer model designed for multi-task learning. The model performs scene classification, contextual tag prediction, and object presence detection. It is then integrated into a content-based recommender system that supports feature configurations. Results: The model was evaluated on its ability to classify scenes, predict tags, and detect objects within images. Ablation analysis confirmed the complementary role of tag, object, and scene features in representation learning, while benchmarking against CNN architectures showed the superior performance of the transformer-based model. Additionally, its integration with a POI recommender system demonstrated consistent performance across different feature settings. The recommender system produced relevant suggestions and maintained robustness even when specific components were disabled. Conclusions: Micro-level contextual tagging enhances the representation of scene context and supports more informative recommendations. WORLDO provides a practical framework for incorporating object-level semantics into POI applications through efficient visual modeling.
Full article

Figure 1
Open AccessArticle
Lightweight Multimodal Adapter for Visual Object Tracking
by
Vasyl Borsuk, Vitaliy Yakovyna and Nataliya Shakhovska
Big Data Cogn. Comput. 2025, 9(11), 292; https://doi.org/10.3390/bdcc9110292 - 15 Nov 2025
Abstract
Visual object tracking is a fundamental computer vision task recently extended to multimodal settings, where natural language descriptions complement visual information. Existing multimodal trackers typically rely on large-scale transformer architectures that jointly train visual and textual encoders, resulting in hundreds of millions of
[...] Read more.
Visual object tracking is a fundamental computer vision task recently extended to multimodal settings, where natural language descriptions complement visual information. Existing multimodal trackers typically rely on large-scale transformer architectures that jointly train visual and textual encoders, resulting in hundreds of millions of trainable parameters and substantial computational overhead. We propose a lightweight multimodal adapter that integrates textual descriptions into a state-of-the-art visual-only framework with minimal overhead. The pretrained visual and text encoders are frozen, and only a small projection network is trained to align text embeddings with visual features. The adapter is modular, can be toggled at inference, and has negligible impact on speed. Extensive experiments demonstrate that textual cues improve tracking robustness and enable efficient multimodal integration with over 100× fewer trainable parameters than heavy multimodal trackers, allowing training and deployment on resource-limited devices.
Full article
(This article belongs to the Special Issue AI, Computer Vision and Human–Robot Interaction)
►▼
Show Figures

Figure 1
Open AccessArticle
Attention-Driven Deep Learning for News-Based Prediction of Disease Outbreaks
by
Avneet Singh Gautam, Zahid Raza, Maria Lapina and Mikhail Babenko
Big Data Cogn. Comput. 2025, 9(11), 291; https://doi.org/10.3390/bdcc9110291 - 14 Nov 2025
Abstract
►▼
Show Figures
Natural Language Processing is being used for Disease Outbreak Prediction using news data. However, the available research focuses on predicting outbreaks for only specific diseases using disease-specific data such as COVID-19, Zika, SARS, MERS, and Ebola, etc. To address the challenge of disease
[...] Read more.
Natural Language Processing is being used for Disease Outbreak Prediction using news data. However, the available research focuses on predicting outbreaks for only specific diseases using disease-specific data such as COVID-19, Zika, SARS, MERS, and Ebola, etc. To address the challenge of disease outbreak prediction without relying on prior knowledge or introducing bias, this research proposes a model that leverages a news dataset devoid of specific disease names. This approach ensures generalizability and domain independence in identifying potential outbreaks. To facilitate supervised learning, spaCy was employed to annotate the dataset, enabling the classification of articles as either related or unrelated to disease outbreaks. LSTM, Bi-LSTM, and Bi-LSTM with a Multi-Head Attention mechanism, and transformer have been used and compared for the purpose of classification. Experimental results exhibit good prediction accuracy with Bi-LSTM with Multi-Head Attention and transformer on the test dataset. The work serves as a pro-active and unbiased approach to predict any disease outbreak without being specific to any disease.
Full article

Figure 1
Open AccessArticle
Wildfire Prediction in British Columbia Using Machine Learning and Deep Learning Models: A Data-Driven Framework
by
Maryam Nasourinia and Kalpdrum Passi
Big Data Cogn. Comput. 2025, 9(11), 290; https://doi.org/10.3390/bdcc9110290 - 14 Nov 2025
Abstract
Wildfires pose a growing threat to ecosystems, infrastructure, and public safety, particularly in the province of British Columbia (BC), Canada. In recent years, the frequency, severity, and scale of wildfires in BC have increased significantly, largely due to climate change, human activity, and
[...] Read more.
Wildfires pose a growing threat to ecosystems, infrastructure, and public safety, particularly in the province of British Columbia (BC), Canada. In recent years, the frequency, severity, and scale of wildfires in BC have increased significantly, largely due to climate change, human activity, and changing land use patterns. This study presents a comprehensive, data-driven approach to wildfire prediction, leveraging advanced machine learning (ML) and deep learning (DL) techniques. A high-resolution dataset was constructed by integrating five years of wildfire incident records from the Canadian Wildland Fire Information System (CWFIS) with ERA5 reanalysis climate data. The final dataset comprises more than 3.6 million spatiotemporal records and 148 environmental, meteorological, and geospatial features. Six feature selection techniques were evaluated, and five predictive models—Random Forest, XGBoost, LightGBM, CatBoost, and an RNN + LSTM—were trained and compared. The CatBoost model achieved the highest predictive performance with an accuracy of 93.4%, F1-score of 92.1%, and ROC-AUC of 0.94, while Random Forest achieved an accuracy of 92.6%. The study identifies key environmental variables, including surface temperature, humidity, wind speed, and soil moisture, as the most influential predictors of wildfire occurrence. These findings highlight the potential of data-driven AI frameworks to support early warning systems and enhance operational wildfire management in British Columbia.
Full article
(This article belongs to the Topic AI for Natural Disasters Detection, Prediction and Modeling)
►▼
Show Figures

Figure 1
Open AccessReview
Speech Separation Using Advanced Deep Neural Network Methods: A Recent Survey
by
Zeng Wang and Zhongqiang Luo
Big Data Cogn. Comput. 2025, 9(11), 289; https://doi.org/10.3390/bdcc9110289 - 14 Nov 2025
Abstract
►▼
Show Figures
Speech separation, as an important research direction in audio signal processing, has been widely studied by the academic community since its emergence in the mid-1990s. In recent years, with the rapid development of deep neural network technology, speech processing based on deep neural
[...] Read more.
Speech separation, as an important research direction in audio signal processing, has been widely studied by the academic community since its emergence in the mid-1990s. In recent years, with the rapid development of deep neural network technology, speech processing based on deep neural networks has shown outstanding performance in speech separation. While existing studies have surveyed the application of deep neural networks in speech separation from multiple dimensions including learning paradigms, model architectures, loss functions, and training strategies, current achievements still lack systematic comprehension of the field’s developmental trajectory. To address this, this paper focuses on single-channel supervised speech separation tasks, proposing a technological evolution path “U-Net–TasNet–Transformer–Mamba” as the main thread to systematically analyze the impact mechanisms of core architectural designs on separation performance across different stages. By reviewing the transition process from traditional methods to deep learning paradigms and delving into the improvements and integration of deep learning architectures at various stages, this paper summarizes milestone achievements, mainstream evaluation frameworks, and typical datasets in the field, while also providing prospects for future research directions. Through this detailed-focused review perspective, we aim to provide researchers in the speech separation field with a clearly articulated technical evolution map and practical reference.
Full article

Figure 1
Open AccessArticle
Your Eyes Under Pressure: Real-Time Estimation of Cognitive Load with Smooth Pursuit Tracking
by
Pierluigi Dell’Acqua, Marco Garofalo, Francesco La Rosa and Massimo Villari
Big Data Cogn. Comput. 2025, 9(11), 288; https://doi.org/10.3390/bdcc9110288 - 13 Nov 2025
Abstract
Understanding and accurately estimating cognitive workload is crucial for the development of adaptive, user-centered interactive systems across a variety of domains including augmented reality, automotive driving assistance, and intelligent tutoring systems. Cognitive workload assessment enables dynamic system adaptation to improve user experience and
[...] Read more.
Understanding and accurately estimating cognitive workload is crucial for the development of adaptive, user-centered interactive systems across a variety of domains including augmented reality, automotive driving assistance, and intelligent tutoring systems. Cognitive workload assessment enables dynamic system adaptation to improve user experience and safety. In this work, we introduce a novel framework that leverages smooth pursuit eye movements as a non-invasive and temporally precise indicator of mental effort. A key innovation of our approach is the development of trajectory-independent algorithms that address a significant limitation of existing methods, which generally rely on a predefined or known stimulus trajectory. Our framework leverages two solutions to provide accurate cognitive load estimation, without requiring knowledge of the exact target path, based on Kalman filter and B-spline heuristic classifiers. This enables the application of our methods in more naturalistic and unconstrained environments where stimulus trajectories may be unknown. We evaluated these algorithms against classical supervised machine learning models on a publicly available benchmark dataset featuring diverse pursuit trajectories and varying cognitive workload conditions. The results demonstrate competitive performance along with robustness across different task complexities and trajectory types. Moreover, our framework supports real-time inference, making it viable for continuous cognitive workload monitoring. To further enhance deployment feasibility, we propose a federated learning architecture, allowing privacy-preserving adaptation of models across heterogeneous devices without the need to share raw gaze data. This scalable approach mitigates privacy concerns and facilitates collaborative model improvement in distributed real-world scenarios. Experimental findings confirm that metrics derived from smooth pursuit eye movements reliably reflect fluctuations in cognitive states induced by working memory load tasks, substantiating their use for real-time, continuous workload estimation. By integrating trajectory independence, robust classification techniques, and federated privacy-aware learning, our work advances the state of the art in adaptive human–computer interaction. This framework offers a scientifically grounded, privacy-conscious, and practically deployable solution for cognitive workload estimation that can be adapted to diverse application contexts.
Full article
(This article belongs to the Special Issue Advances in Artificial Intelligence for Computer Vision, Augmented Reality Virtual Reality and Metaverse)
►▼
Show Figures

Figure 1
Open AccessArticle
Construction of a Person–Job Temporal Knowledge Graph Using Large Language Models
by
Zhongshan Zhang, Junzhi Wang, Bo Li, Xiang Lin and Mingyu Liu
Big Data Cogn. Comput. 2025, 9(11), 287; https://doi.org/10.3390/bdcc9110287 - 12 Nov 2025
Abstract
►▼
Show Figures
Person–job data are multi-source, heterogeneous, and strongly temporal, making knowledge modeling and analysis challenging. We present an automated approach for constructing a Human-Resources Temporal Knowledge Graph. We first formalize a schema in which temporal relations are represented as sets of time intervals. On
[...] Read more.
Person–job data are multi-source, heterogeneous, and strongly temporal, making knowledge modeling and analysis challenging. We present an automated approach for constructing a Human-Resources Temporal Knowledge Graph. We first formalize a schema in which temporal relations are represented as sets of time intervals. On top of this schema, a large language model (LLM) pipeline extracts entities, relations, and temporal expressions, augmented by self-verification and external knowledge injection to enforce schema compliance, resolve ambiguities, and automatically repair outputs. Context-aware prompting and confidence-based escalation further improve robustness. Evaluated on a corpus of 2000 Chinese resumes, our method outperforms strong baselines, and ablations confirm the necessity and synergy of each component; notably, temporal extraction attains an F1 of 0.9876. The proposed framework provides a reusable path and engineering foundation for downstream HR tasks—such as profiling, relational reasoning, and position matching—supporting more reliable, time-aware decision-making in complex organizations.
Full article

Figure 1
Open AccessArticle
Overcoming Domain Shift in Violence Detection with Contrastive Consistency Learning
by
Zhenche Xia, Zhenhua Tan and Bin Zhang
Big Data Cogn. Comput. 2025, 9(11), 286; https://doi.org/10.3390/bdcc9110286 - 12 Nov 2025
Abstract
►▼
Show Figures
Automated violence detection in video surveillance is critical for public safety; however, existing methods frequently suffer notable performance degradation across diverse real-world scenarios due to domain shift. Substantial distributional discrepancies between source training data and target environments severely hinder model generalization, limiting practical
[...] Read more.
Automated violence detection in video surveillance is critical for public safety; however, existing methods frequently suffer notable performance degradation across diverse real-world scenarios due to domain shift. Substantial distributional discrepancies between source training data and target environments severely hinder model generalization, limiting practical deployment. To overcome this, we propose CoMT-VD, a new contrastive Mean Teacher-based violence detection model, engineered for enhanced adaptability in unseen target domains. CoMT-VD innovatively integrates a Mean Teacher architecture to adequately leverage unlabeled target domain data, fostering stable, domain-invariant feature representations by enforcing consistency regularization between student and teacher networks, crucial for bridging the domain gap. Furthermore, to mitigate supervisory noise from pseudo-labels and refine the feature space, CoMT-VD incorporates a dual-strategy contrastive learning module. DCL systematically refines features through intra-sample consistency, minimizing latent space distances for compact representations, and inter-sample consistency, maximizing feature dissimilarity across distinct categories to sharpen decision boundaries. This dual regularization purifies the learned feature space, boosting discriminativeness while mitigating noisy pseudo-labels. Broad evaluations on five benchmark datasets unequivocally demonstrate that CoMT-VD achieves the superior generalization performance (in the four integrated scenarios from five benchmark datasets, the improvements were 5.0∼12.0%, 6.0∼12.5%, 5.0∼11.2%, 5.0∼11.2%, and 6.3∼12.3%, respectively), marking a notable advancement towards robust and reliable real-world violence detection systems.
Full article

Figure 1
Journal Menu
► ▼ Journal Menu-
- BDCC Home
- Aims & Scope
- Editorial Board
- Reviewer Board
- Topical Advisory Panel
- Instructions for Authors
- Special Issues
- Topics
- Topical Collections
- Article Processing Charge
- Indexing & Archiving
- Editor’s Choice Articles
- Most Cited & Viewed
- Journal Statistics
- Journal History
- Journal Awards
- Conferences
- Editorial Office
Journal Browser
► ▼ Journal BrowserHighly Accessed Articles
Latest Books
E-Mail Alert
News
Topics
Topic in
Actuators, Algorithms, BDCC, Future Internet, JMMP, Machines, Robotics, Systems
Smart Product Design and Manufacturing on Industrial Internet
Topic Editors: Pingyu Jiang, Jihong Liu, Ying Liu, Jihong YanDeadline: 31 December 2025
Topic in
Computers, Information, AI, Electronics, Technologies, BDCC
Graph Neural Networks and Learning Systems
Topic Editors: Huijia Li, Jun Hu, Weichen Zhao, Jie CaoDeadline: 31 January 2026
Topic in
AI, BDCC, Fire, GeoHazards, Remote Sensing
AI for Natural Disasters Detection, Prediction and Modeling
Topic Editors: Moulay A. Akhloufi, Mozhdeh ShahbaziDeadline: 31 March 2026
Topic in
Applied Sciences, Electronics, J. Imaging, MAKE, Information, BDCC, Signals
Applications of Image and Video Processing in Medical Imaging
Topic Editors: Jyh-Cheng Chen, Kuangyu ShiDeadline: 30 April 2026
Conferences
Special Issues
Special Issue in
BDCC
Applied Data Science for Social Good: 2nd Edition
Guest Editors: Vishnu S. Pendyala, Celestine IwendiDeadline: 15 December 2025
Special Issue in
BDCC
Deep Learning-Based Pose Estimation: Applications in Vision, Robotics, and Beyond
Guest Editors: Jyotindra Narayan, Chaiyawan AuepanwiriyakulDeadline: 31 December 2025
Special Issue in
BDCC
Transforming Cyber Security Provision Through Utilizing Artificial Intelligence
Guest Editors: Peter R. J. Trim, Yang-Im LeeDeadline: 31 December 2025
Special Issue in
BDCC
Bioinformatics Dry Laboratories for Research and Education: Computational Systems and Infrastructures
Guest Editor: Emanuel MaldonadoDeadline: 31 December 2025





