Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (170)

Search Parameters:
Keywords = adaptive deep ensemble learning

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
25 pages, 4100 KB  
Article
An Adaptive Unsupervised Learning Approach for Credit Card Fraud Detection
by John Adejoh, Nsikak Owoh, Moses Ashawa, Salaheddin Hosseinzadeh, Alireza Shahrabi and Salma Mohamed
Big Data Cogn. Comput. 2025, 9(9), 217; https://doi.org/10.3390/bdcc9090217 - 25 Aug 2025
Viewed by 219
Abstract
Credit card fraud remains a major cause of financial loss around the world. Traditional fraud detection methods that rely on supervised learning often struggle because fraudulent transactions are rare compared to legitimate ones, leading to imbalanced datasets. Additionally, the models must be retrained [...] Read more.
Credit card fraud remains a major cause of financial loss around the world. Traditional fraud detection methods that rely on supervised learning often struggle because fraudulent transactions are rare compared to legitimate ones, leading to imbalanced datasets. Additionally, the models must be retrained frequently, as fraud patterns change over time and require new labeled data for retraining. To address these challenges, this paper proposes an ensemble unsupervised learning approach for credit card fraud detection that combines Autoencoders (AEs), Self-Organizing Maps (SOMs), and Restricted Boltzmann Machines (RBMs), integrated with an Adaptive Reconstruction Threshold (ART) mechanism. The ART dynamically adjusts anomaly detection thresholds by leveraging the clustering properties of SOMs, effectively overcoming the limitations of static threshold approaches in machine learning and deep learning models. The proposed models, AE-ASOMs (Autoencoder—Adaptive Self-Organizing Maps) and RBM-ASOMs (Restricted Boltzmann Machines—Adaptive Self-Organizing Maps), were evaluated on the Kaggle Credit Card Fraud Detection and IEEE-CIS datasets. Our AE-ASOM model achieved an accuracy of 0.980 and an F1-score of 0.967, while the RBM-ASOM model achieved an accuracy of 0.975 and an F1-score of 0.955. Compared to models such as One-Class SVM and Isolation Forest, our approach demonstrates higher detection accuracy and significantly reduces false positive rates. In addition to its performance, the model offers considerable computational efficiency with a training time of 200.52 s and memory usage of 3.02 megabytes. Full article
Show Figures

Figure 1

27 pages, 1604 KB  
Review
A Review of State-of-the-Art AI and Data-Driven Techniques for Load Forecasting
by Jian Liu, Xiaotian He, Kangji Li and Wenping Xue
Energies 2025, 18(16), 4408; https://doi.org/10.3390/en18164408 - 19 Aug 2025
Viewed by 470
Abstract
With the gradual penetration of new energy generation/storage, accurate and reliable load forecasting (LF) plays an increasingly important role in different energy management applications (e.g., power resource allocation, peak demand response, energy supply and demand optimization). In recent years, data-driven and artificial intelligence [...] Read more.
With the gradual penetration of new energy generation/storage, accurate and reliable load forecasting (LF) plays an increasingly important role in different energy management applications (e.g., power resource allocation, peak demand response, energy supply and demand optimization). In recent years, data-driven and artificial intelligence (AI) technologies have received considerable attention in the field of LF. This study provides a comprehensive review on the existing advanced AI and data-driven techniques used for LF tasks. First, the reviewed studies are classified from the load’s spatial scale and forecasting time scale, and the research gap that this study aims to fill in the existing reviews is revealed. It was found that short-term forecasting dominates in the time scale (accounting for about 83.1%). Second, based on the summary of basic preprocessing methods, some advanced preprocessing methods are presented and analyzed. These advanced methods have greatly increased complexity compared with basic methods, while they can bring significant performance improvements such as adaptability and accuracy. Then, various LF models using the latest AI techniques, including deep learning, reinforcement learning, transfer learning, and ensemble learning, are reviewed and analyzed. These models are also summarized from several aspects, such as computational cost, interpretability, application scenarios, and so on. Finally, from the perspectives of data, techniques, and operations, a detailed discussion is given on some challenges and opportunities for LF. Full article
(This article belongs to the Section G: Energy and Buildings)
Show Figures

Figure 1

22 pages, 4350 KB  
Review
A Review of Artificial Intelligence Techniques in Fault Diagnosis of Electric Machines
by Christos Zachariades and Vigila Xavier
Sensors 2025, 25(16), 5128; https://doi.org/10.3390/s25165128 - 18 Aug 2025
Viewed by 514
Abstract
Rotating electrical machines are critical assets in industrial systems, where unexpected failures can lead to costly downtime and safety risks. This review presents a comprehensive and up-to-date analysis of artificial intelligence (AI) techniques for fault diagnosis in electric machines. It categorizes and evaluates [...] Read more.
Rotating electrical machines are critical assets in industrial systems, where unexpected failures can lead to costly downtime and safety risks. This review presents a comprehensive and up-to-date analysis of artificial intelligence (AI) techniques for fault diagnosis in electric machines. It categorizes and evaluates supervised, unsupervised, deep learning, and hybrid/ensemble approaches in terms of diagnostic accuracy, adaptability, and implementation complexity. A comparative analysis highlights the strengths and limitations of each method, while emerging trends such as explainable AI, self-supervised learning, and digital twin integration are discussed as enablers of next-generation diagnostic systems. To support practical deployment, the article proposes a modular implementation framework and offers actionable recommendations for practitioners. This work serves as both a reference and a guide for researchers and engineers aiming to develop scalable, interpretable, and robust AI-driven fault diagnosis solutions for rotating electrical machines. Full article
(This article belongs to the Special Issue Sensors for Fault Diagnosis of Electric Machines)
Show Figures

Figure 1

23 pages, 1304 KB  
Article
A GIS-Driven, Machine Learning-Enhanced Framework for Adaptive Land Bonitation
by Bogdan Văduva, Anca Avram, Oliviu Matei, Laura Andreica and Teodor Rusu
Agriculture 2025, 15(16), 1735; https://doi.org/10.3390/agriculture15161735 - 12 Aug 2025
Viewed by 268
Abstract
Land bonitation, or land rating, is a core instrument in agricultural policy used to evaluate land productivity based on environmental and climatic indicators. However, conventional Bonitation Coefficient (BC) methods are often rigid, require complete indicator sets, and lack mechanisms for handling missing or [...] Read more.
Land bonitation, or land rating, is a core instrument in agricultural policy used to evaluate land productivity based on environmental and climatic indicators. However, conventional Bonitation Coefficient (BC) methods are often rigid, require complete indicator sets, and lack mechanisms for handling missing or forecasted data—limiting their applicability under data scarcity and climate variability. This paper proposes a GIS-integrated, modular framework that couples classical BC computation with machine learning-based temporal forecasting and spatial generalization. Specifically, we apply deep learning models (LSTM, GRU, and CNN) to predict monthly precipitation—one of the 17 indicators in the Romanian BC formula—using over 61 years of data. The forecasts are spatially interpolated using Voronoi tessellation and then incorporated into the bonitation process via an adaptive logic that accommodates both complete and incomplete datasets. Results show that the ensemble forecast model outperforms individual predictors, achieving an R2 of up to 0.648 and an RMSE of 18.8 mm, compared to LSTM (R2=0.59), GRU (R2=0.61), and CNN (R2=0.57). While the case study focuses on precipitation, the framework is generalizable to other BC indicators and regions. This integration of forecasting, spatial generalization, and classical land evaluation addresses key limitations of existing bonitation methods and lays the groundwork for scalable, AI-enhanced land assessment systems. The forecasting module supports BC computation by supplying missing climate indicators, reinforcing that the primary aim remains adaptive land bonitation. Full article
(This article belongs to the Section Artificial Intelligence and Digital Agriculture)
Show Figures

Figure 1

23 pages, 8286 KB  
Article
Context-Guided SAR Ship Detection with Prototype-Based Model Pretraining and Check–Balance-Based Decision Fusion
by Haowen Zhou, Zhe Geng, Minjie Sun, Linyi Wu and He Yan
Sensors 2025, 25(16), 4938; https://doi.org/10.3390/s25164938 - 10 Aug 2025
Viewed by 380
Abstract
To address the challenging problem of multi-scale inshore–offshore ship detection in synthetic aperture radar (SAR) remote sensing images, we propose a novel deep learning-based automatic ship detection method within the framework of compositional learning. The proposed method is supported by three pillars: context-guided [...] Read more.
To address the challenging problem of multi-scale inshore–offshore ship detection in synthetic aperture radar (SAR) remote sensing images, we propose a novel deep learning-based automatic ship detection method within the framework of compositional learning. The proposed method is supported by three pillars: context-guided region proposal, prototype-based model-pretraining, and multi-model ensemble learning. To reduce the false alarms induced by the discrete ground clutters, the prior knowledge of the harbour’s layout is exploited to generate land masks for terrain delimitation. To prepare the model for the diverse ship targets of different sizes and orientations it might encounter in the test environment, a novel cross-dataset model pretraining strategy is devised, where the SAR images of several key ship target prototypes from the auxiliary dataset are used to support class-incremental learning. To combine the advantages of diverse model architectures, an adaptive decision-level fusion framework is proposed, which consists of three components: a dynamic confidence threshold assignment strategy based on the sizes of targets, a weighted fusion mechanism based on president-senate check–balance, and Soft-NMS-based Dense Group Target Bounding Box Fusion (Soft-NMS-DGT-BBF). The performance enhancement brought by contextual knowledge-aided terrain delimitation, cross-dataset prototype-based model pretraining and check–balance-based adaptive decision-level fusion are validated with a series of ingeniously devised experiments based on the FAIR-CSAR-Ship dataset. Full article
(This article belongs to the Special Issue SAR Imaging Technologies and Applications)
Show Figures

Figure 1

24 pages, 2791 KB  
Article
Short-Term Wind Power Forecasting Based on Improved Modal Decomposition and Deep Learning
by Bin Cheng, Wenwu Li and Jie Fang
Processes 2025, 13(8), 2516; https://doi.org/10.3390/pr13082516 - 9 Aug 2025
Viewed by 397
Abstract
With the continued growth in wind power installed capacity and electricity generation, accurate wind power forecasting has become increasingly critical for power system stability and economic operations. Currently, short-term wind power forecasting often employs deep learning models following modal decomposition of wind power [...] Read more.
With the continued growth in wind power installed capacity and electricity generation, accurate wind power forecasting has become increasingly critical for power system stability and economic operations. Currently, short-term wind power forecasting often employs deep learning models following modal decomposition of wind power time series. However, the optimal length of the time series used for decomposition remains unclear. To address this issue, this paper proposes a short-term wind power forecasting method that integrates improved modal decomposition with deep learning techniques. First, the historical wind power series is segmented using the Pruned Exact Linear Time (PELT) method. Next, the segmented series is decomposed using an enhanced Complete Ensemble Empirical Mode Decomposition with Adaptive Noise (ICEEMDAN) to extract multiple modal components. High-frequency oscillatory components are then further decomposed using Variational Mode Decomposition (VMD), and the resulting modes are clustered using the K-means algorithm. The reconstructed components are subsequently input into a Long Short-Term Memory (LSTM) network for prediction, and the final forecast is obtained by aggregating the outputs of the individual modes. The proposed method is validated using historical wind power data from a wind farm. Experimental results demonstrate that this approach enhances forecasting accuracy, supports grid power balance, and increases the economic benefits for wind farm operators in electricity markets. Full article
(This article belongs to the Section Energy Systems)
Show Figures

Figure 1

24 pages, 12489 KB  
Article
Hyperspectral Lithological Classification of 81 Rock Types Using Deep Ensemble Learning Algorithms
by Shanjuan Xie, Yichun Qiu, Shixian Cao and Wenyuan Wu
Minerals 2025, 15(8), 844; https://doi.org/10.3390/min15080844 - 8 Aug 2025
Viewed by 294
Abstract
To address overfitting due to limited sample size, and the challenges posed by “Spectral Homogeneity with Material Heterogeneity (SHMH)” and “Material Consistency with Spectral Divergence (MCSD)”—which arise from subtle spectral differences and limited classification accuracy—this study proposes a deep integration model that combines [...] Read more.
To address overfitting due to limited sample size, and the challenges posed by “Spectral Homogeneity with Material Heterogeneity (SHMH)” and “Material Consistency with Spectral Divergence (MCSD)”—which arise from subtle spectral differences and limited classification accuracy—this study proposes a deep integration model that combines the Adaptive Boosting (AdaBoost) algorithm with a convolutional recurrent neural network (CRNN). The model adopts a dual-branch architecture integrating a 2D-CNN and gated recurrent unit to effectively fuse spatial and spectral features of rock samples, while the integration of the AdaBoost algorithm optimizes performance by enhancing system stability and generalization capability. The experiment used a hyperspectral dataset containing 81 rock samples (46 igneous rocks and 35 metamorphic rocks) and evaluated model performance through five-fold cross-validation. The results showed that the proposed 2D-CRNN-AdaBoost model achieved 92.55% overall accuracy, which was significantly better than that of other comparative models, demonstrating the effectiveness of multimodal feature fusion and ensemble learning strategy. Full article
(This article belongs to the Section Mineral Exploration Methods and Applications)
Show Figures

Figure 1

25 pages, 2915 KB  
Article
Multi-Model Identification of Rice Leaf Diseases Based on CEL-DL-Bagging
by Zhenghua Zhang, Rufeng Wang and Siqi Huang
AgriEngineering 2025, 7(8), 255; https://doi.org/10.3390/agriengineering7080255 - 7 Aug 2025
Viewed by 359
Abstract
This study proposes CEL-DL-Bagging (Cross-Entropy Loss-optimized Deep Learning Bagging), a multi-model fusion framework that integrates cross-entropy loss-weighted voting with Bootstrap Aggregating (Bagging). First, we develop a lightweight recognition architecture by embedding a salient position attention (SPA) mechanism into four base networks (YOLOv5s-cls, EfficientNet-B0, [...] Read more.
This study proposes CEL-DL-Bagging (Cross-Entropy Loss-optimized Deep Learning Bagging), a multi-model fusion framework that integrates cross-entropy loss-weighted voting with Bootstrap Aggregating (Bagging). First, we develop a lightweight recognition architecture by embedding a salient position attention (SPA) mechanism into four base networks (YOLOv5s-cls, EfficientNet-B0, MobileNetV3, and ShuffleNetV2), significantly enhancing discriminative feature extraction for disease patterns. Our experiments show that these SPA-enhanced models achieve consistent accuracy gains of 0.8–1.7 percentage points, peaking at 97.86%. Building on this, we introduce DB-CEWSV—an ensemble framework combining Deep Bootstrap Aggregating (DB) with adaptive Cross-Entropy Weighted Soft Voting (CEWSV). The system dynamically optimizes model weights based on their cross-entropy performance, using SPA-augmented networks as base learners. The final integrated model attains 98.33% accuracy, outperforming the strongest individual base learner by 0.48 percentage points. Compared with single models, the ensemble learning algorithm proposed in this study led to better generalization and robustness of the ensemble learning model and better identification of rice diseases in the natural background. It provides a technical reference for applying rice disease identification in practical engineering. Full article
(This article belongs to the Topic Digital Agriculture, Smart Farming and Crop Monitoring)
Show Figures

Figure 1

29 pages, 3173 KB  
Article
Graph Neural Networks for Sustainable Energy: Predicting Adsorption in Aromatic Molecules
by Hasan Imani Parashkooh and Cuiying Jian
ChemEngineering 2025, 9(4), 85; https://doi.org/10.3390/chemengineering9040085 - 6 Aug 2025
Viewed by 441
Abstract
The growing need for rapid screening of adsorption energies in organic materials has driven substantial progress in developing various architectures of equivariant graph neural networks (eGNNs). This advancement has largely been enabled by the availability of extensive Density Functional Theory (DFT)-generated datasets, sufficiently [...] Read more.
The growing need for rapid screening of adsorption energies in organic materials has driven substantial progress in developing various architectures of equivariant graph neural networks (eGNNs). This advancement has largely been enabled by the availability of extensive Density Functional Theory (DFT)-generated datasets, sufficiently large to train complex eGNN models effectively. However, certain material groups with significant industrial relevance, such as aromatic compounds, remain underrepresented in these large datasets. In this work, we aim to bridge the gap between limited, domain-specific DFT datasets and large-scale pretrained eGNNs. Our methodology involves creating a specialized dataset by segregating aromatic compounds after a targeted ensemble extraction process, then fine-tuning a pretrained model via approaches that include full retraining and systematically freezing specific network sections. We demonstrate that these approaches can yield accurate energy and force predictions with minimal domain-specific training data and computation. Additionally, we investigate the effects of augmenting training datasets with chemically related but out-of-domain groups. Our findings indicate that incorporating supplementary data that closely resembles the target domain, even if approximate, would enhance model performance on domain-specific tasks. Furthermore, we systematically freeze different sections of the pretrained models to elucidate the role each component plays during adaptation to new domains, revealing that relearning low-level representations is critical for effective domain transfer. Overall, this study contributes valuable insights and practical guidelines for efficiently adapting deep learning models for accurate adsorption energy predictions, significantly reducing reliance on extensive training datasets. Full article
Show Figures

Figure 1

43 pages, 2466 KB  
Article
Adaptive Ensemble Learning for Financial Time-Series Forecasting: A Hypernetwork-Enhanced Reservoir Computing Framework with Multi-Scale Temporal Modeling
by Yinuo Sun, Zhaoen Qu, Tingwei Zhang and Xiangyu Li
Axioms 2025, 14(8), 597; https://doi.org/10.3390/axioms14080597 - 1 Aug 2025
Viewed by 713
Abstract
Financial market forecasting remains challenging due to complex nonlinear dynamics and regime-dependent behaviors that traditional models struggle to capture effectively. This research introduces the Adaptive Financial Reservoir Network with Hypernetwork Flow (AFRN–HyperFlow) framework, a novel ensemble architecture integrating Echo State Networks, temporal convolutional [...] Read more.
Financial market forecasting remains challenging due to complex nonlinear dynamics and regime-dependent behaviors that traditional models struggle to capture effectively. This research introduces the Adaptive Financial Reservoir Network with Hypernetwork Flow (AFRN–HyperFlow) framework, a novel ensemble architecture integrating Echo State Networks, temporal convolutional networks, mixture density networks, adaptive Hypernetworks, and deep state-space models for enhanced financial time-series prediction. Through comprehensive feature engineering incorporating technical indicators, spectral decomposition, reservoir-based representations, and flow dynamics characteristics, the framework achieves superior forecasting performance across diverse market conditions. Experimental validation on 26,817 balanced samples demonstrates exceptional results with an F1-score of 0.8947, representing a 12.3% improvement over State-of-the-Art baseline methods, while maintaining robust performance across asset classes from equities to cryptocurrencies. The adaptive Hypernetwork mechanism enables real-time regime-change detection with 2.3 days average lag and 95% accuracy, while systematic SHAP analysis provides comprehensive interpretability essential for regulatory compliance. Ablation studies reveal Echo State Networks contribute 9.47% performance improvement, validating the architectural design. The AFRN–HyperFlow framework addresses critical limitations in uncertainty quantification, regime adaptability, and interpretability, offering promising directions for next-generation financial forecasting systems incorporating quantum computing and federated learning approaches. Full article
(This article belongs to the Special Issue Financial Mathematics and Econophysics)
Show Figures

Figure 1

27 pages, 4682 KB  
Article
DERIENet: A Deep Ensemble Learning Approach for High-Performance Detection of Jute Leaf Diseases
by Mst. Tanbin Yasmin Tanny, Tangina Sultana, Md. Emran Biswas, Chanchol Kumar Modok, Arjina Akter, Mohammad Shorif Uddin and Md. Delowar Hossain
Information 2025, 16(8), 638; https://doi.org/10.3390/info16080638 - 27 Jul 2025
Viewed by 338
Abstract
Jute, a vital lignocellulosic fiber crop with substantial industrial and ecological relevance, continues to suffer considerable yield and quality degradation due to pervasive foliar pathologies. Traditional diagnostic modalities reliant on manual field inspections are inherently constrained by subjectivity, diagnostic latency, and inadequate scalability [...] Read more.
Jute, a vital lignocellulosic fiber crop with substantial industrial and ecological relevance, continues to suffer considerable yield and quality degradation due to pervasive foliar pathologies. Traditional diagnostic modalities reliant on manual field inspections are inherently constrained by subjectivity, diagnostic latency, and inadequate scalability across geographically distributed agrarian systems. To transcend these limitations, we propose DERIENet, a robust and scalable classification approach within a deep ensemble learning framework. It is meticulously engineered by integrating three high-performing convolutional neural networks—ResNet50, InceptionV3, and EfficientNetB0—along with regularization, batch normalization, and dropout strategies, to accurately classify jute leaf diseases such as Cercospora Leaf Spot, Golden Mosaic Virus, and healthy leaves. A key methodological contribution is the design of a novel augmentation pipeline, termed Geometric Localized Occlusion and Adaptive Rescaling (GLOAR), which dynamically modulates photometric and geometric distortions based on image entropy and luminance to synthetically upscale a limited dataset (920 images) into a significantly enriched and diverse dataset of 7800 samples, thereby mitigating overfitting and enhancing domain generalizability. Empirical evaluation, utilizing a comprehensive set of performance metrics—accuracy, precision, recall, F1-score, confusion matrices, and ROC curves—demonstrates that DERIENet achieves a state-of-the-art classification accuracy of 99.89%, with macro-averaged and weighted average precision, recall, and F1-score uniformly at 99.89%, and an AUC of 1.0 across all disease categories. The reliability of the model is validated by the confusion matrix, which shows that 899 out of 900 test images were correctly identified and that there was only one misclassification. Comparative evaluations of the various ensemble baselines, such as DenseNet201, MobileNetV2, and VGG16, and individual base learners demonstrate that DERIENet performs noticeably superior to all baseline models. It provides a highly interpretable, deployment-ready, and computationally efficient architecture that is ideal for integrating into edge or mobile platforms to facilitate in situ, real-time disease diagnostics in precision agriculture. Full article
Show Figures

Figure 1

23 pages, 4256 KB  
Article
A GAN-Based Framework with Dynamic Adaptive Attention for Multi-Class Image Segmentation in Autonomous Driving
by Bashir Sheikh Abdullahi Jama and Mehmet Hacibeyoglu
Appl. Sci. 2025, 15(15), 8162; https://doi.org/10.3390/app15158162 - 22 Jul 2025
Viewed by 361
Abstract
Image segmentation is a foundation for autonomous driving frameworks that empower vehicles to explore and navigate their surrounding environment. It gives a fundamental setting to the dynamic cycles by dividing the image into significant parts like streets, vehicles, walkers, and traffic signs. Precise [...] Read more.
Image segmentation is a foundation for autonomous driving frameworks that empower vehicles to explore and navigate their surrounding environment. It gives a fundamental setting to the dynamic cycles by dividing the image into significant parts like streets, vehicles, walkers, and traffic signs. Precise segmentation ensures safe navigation and the avoidance of collisions, while following the rules of traffic is very critical for seamless operation in self-driving cars. The most recent deep learning-based image segmentation models have demonstrated impressive performance in structured environments, yet they often fall short when applied to the complex and unpredictable conditions encountered in autonomous driving. This study proposes an Adaptive Ensemble Attention (AEA) mechanism within a Generative Adversarial Network architecture to deal with dynamic and complex driving conditions. The AEA integrates the features of self, spatial, and channel attention adaptively and powerfully changes the amount of each contribution as per input and context-oriented relevance. It does this by allowing the discriminator network in GAN to evaluate the segmentation mask created by the generator. This explains the difference between real and fake masks by considering a concatenated pair of an original image and its mask. The adversarial training will prompt the generator, via the discriminator, to mask out the image in such a way that the output aligns with the expected ground truth and is also very realistic. The exchange of information between the generator and discriminator improves the quality of the segmentation. In order to check the accuracy of the proposed method, the three widely used datasets BDD100K, Cityscapes, and KITTI were selected to calculate average IoU, where the value obtained was 89.46%, 89.02%, and 88.13% respectively. These outcomes emphasize the model’s effectiveness and consistency. Overall, it achieved a remarkable accuracy of 98.94% and AUC of 98.4%, indicating strong enhancements compared to the State-of-the-art (SOTA) models. Full article
Show Figures

Figure 1

24 pages, 2173 KB  
Article
A Novel Ensemble of Deep Learning Approach for Cybersecurity Intrusion Detection with Explainable Artificial Intelligence
by Abdullah Alabdulatif
Appl. Sci. 2025, 15(14), 7984; https://doi.org/10.3390/app15147984 - 17 Jul 2025
Viewed by 971
Abstract
In today’s increasingly interconnected digital world, cyber threats have grown in frequency and sophistication, making intrusion detection systems a critical component of modern cybersecurity frameworks. Traditional IDS methods, often based on static signatures and rule-based systems, are no longer sufficient to detect and [...] Read more.
In today’s increasingly interconnected digital world, cyber threats have grown in frequency and sophistication, making intrusion detection systems a critical component of modern cybersecurity frameworks. Traditional IDS methods, often based on static signatures and rule-based systems, are no longer sufficient to detect and respond to complex and evolving attacks. To address these challenges, Artificial Intelligence and machine learning have emerged as powerful tools for enhancing the accuracy, adaptability, and automation of IDS solutions. This study presents a novel, hybrid ensemble learning-based intrusion detection framework that integrates deep learning and traditional ML algorithms with explainable artificial intelligence for real-time cybersecurity applications. The proposed model combines an Artificial Neural Network and Support Vector Machine as base classifiers and employs a Random Forest as a meta-classifier to fuse predictions, improving detection performance. Recursive Feature Elimination is utilized for optimal feature selection, while SHapley Additive exPlanations (SHAP) provide both global and local interpretability of the model’s decisions. The framework is deployed using a Flask-based web interface in the Amazon Elastic Compute Cloud environment, capturing live network traffic and offering sub-second inference with visual alerts. Experimental evaluations using the NSL-KDD dataset demonstrate that the ensemble model outperforms individual classifiers, achieving a high accuracy of 99.40%, along with excellent precision, recall, and F1-score metrics. This research not only enhances detection capabilities but also bridges the trust gap in AI-powered security systems through transparency. The solution shows strong potential for application in critical domains such as finance, healthcare, industrial IoT, and government networks, where real-time and interpretable threat detection is vital. Full article
Show Figures

Figure 1

16 pages, 2355 KB  
Article
Generalising Stock Detection in Retail Cabinets with Minimal Data Using a DenseNet and Vision Transformer Ensemble
by Babak Rahi, Deniz Sagmanli, Felix Oppong, Direnc Pekaslan and Isaac Triguero
Mach. Learn. Knowl. Extr. 2025, 7(3), 66; https://doi.org/10.3390/make7030066 - 16 Jul 2025
Viewed by 471
Abstract
Generalising deep-learning models to perform well on unseen data domains with minimal retraining remains a significant challenge in computer vision. Even when the target task—such as quantifying the number of elements in an image—stays the same, data quality, shape, or form variations can [...] Read more.
Generalising deep-learning models to perform well on unseen data domains with minimal retraining remains a significant challenge in computer vision. Even when the target task—such as quantifying the number of elements in an image—stays the same, data quality, shape, or form variations can deviate from the training conditions, often necessitating manual intervention. As a real-world industry problem, we aim to automate stock level estimation in retail cabinets. As technology advances, new cabinet models with varying shapes emerge alongside new camera types. This evolving scenario poses a substantial obstacle to deploying long-term, scalable solutions. To surmount the challenge of generalising to new cabinet models and cameras with minimal amounts of sample images, this research introduces a new solution. This paper proposes a novel ensemble model that combines DenseNet-201 and Vision Transformer (ViT-B/8) architectures to achieve generalisation in stock-level classification. The novelty aspect of our solution comes from the fact that we combine a transformer with a DenseNet model in order to capture both the local, hierarchical details and the long-range dependencies within the images, improving generalisation accuracy with less data. Key contributions include (i) a novel DenseNet-201 + ViT-B/8 feature-level fusion, (ii) an adaptation workflow that needs only two images per class, (iii) a balanced layer-unfreezing schedule, (iv) a publicly described domain-shift benchmark, and (v) a 47 pp accuracy gain over four standard few-shot baselines. Our approach leverages fine-tuning techniques to adapt two pre-trained models to the new retail cabinets (i.e., standing or horizontal) and camera types using only two images per class. Experimental results demonstrate that our method achieves high accuracy rates of 91% on new cabinets with the same camera and 89% on new cabinets with different cameras, significantly outperforming standard few-shot learning methods. Full article
(This article belongs to the Section Data)
Show Figures

Figure 1

30 pages, 4273 KB  
Article
Hybrid Attention-Enhanced Xception and Dynamic Chaotic Whale Optimization for Brain Tumor Diagnosis
by Aliyu Tetengi Ibrahim, Ibrahim Hayatu Hassan, Mohammed Abdullahi, Armand Florentin Donfack Kana, Amina Hassan Abubakar, Mohammed Tukur Mohammed, Lubna A. Gabralla, Mohamad Khoiru Rusydi and Haruna Chiroma
Bioengineering 2025, 12(7), 747; https://doi.org/10.3390/bioengineering12070747 - 9 Jul 2025
Viewed by 545
Abstract
In medical diagnostics, brain tumor classification remains essential, as accurate and efficient models aid medical professionals in early detection and treatment planning. Deep learning methodologies for brain tumor classification have gained popularity due to their potential to deliver prompt and precise diagnostic results. [...] Read more.
In medical diagnostics, brain tumor classification remains essential, as accurate and efficient models aid medical professionals in early detection and treatment planning. Deep learning methodologies for brain tumor classification have gained popularity due to their potential to deliver prompt and precise diagnostic results. This article proposes a novel classification technique that integrates the Xception model with a hybrid attention mechanism and progressive image resizing to enhance performance. The methodology is built on a combination of preprocessing techniques, transfer learning architecture reconstruction, and dynamic fine-tuning strategies. To optimize key hyper-parameters, this study employed the Dynamic Chaotic Whale Optimization Algorithm. Additionally, we developed a novel learning rate scheduler that dynamically adjusts the learning rate based on image size at each training phase, improving training efficiency and model adaptability. Batch sizes and layer freezing methods were also adjusted according to image size. We constructed an ensemble approach by preserving models trained on different image sizes and merging their results using weighted averaging, bagging, boosting, stacking, blending, and voting techniques. Our proposed method was evaluated on benchmark datasets achieving remarkable accuracies of 99.67%, 99.09%, and 99.67% compared to the classical algorithms. Full article
Show Figures

Figure 1

Back to TopTop