Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (4,251)

Search Parameters:
Keywords = ensemble learning model

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
29 pages, 20184 KB  
Article
Estimation of Canopy Traits and Yield in Maize–Soybean Intercropping Systems Using UAV Multispectral Imagery and Machine Learning
by Li Wang, Shujie Jia, Jinguang Zhao, Canru Liang and Wuping Zhang
Agriculture 2026, 16(4), 487; https://doi.org/10.3390/agriculture16040487 (registering DOI) - 22 Feb 2026
Abstract
Strip intercropping of maize and soybean is a key practice for improving land productivity and ensuring food and oil security in the hilly regions of the Loess Plateau. However, complex interspecific interactions generate highly heterogeneous canopy structures, making it difficult for traditional linear [...] Read more.
Strip intercropping of maize and soybean is a key practice for improving land productivity and ensuring food and oil security in the hilly regions of the Loess Plateau. However, complex interspecific interactions generate highly heterogeneous canopy structures, making it difficult for traditional linear models to capture yield variability within mixed pixels. Based on a single-season (2025) field experiment, this study developed a UAV multispectral imagery-based yield estimation framework integrating multiple machine-learning algorithms. Shapley additive explanations (SHAP) and partial dependence plots (PDP) were used to interpret the spectral–yield relationships under different spatial configurations. The predictive performance of linear regression and eight nonlinear algorithms was compared using 20 spectral features. Ensemble learning outperformed linear approaches in all intercropping scenarios. In the maize–soybean 3:2 pattern, the GBDT model delivered the highest accuracy (R2 = 0.849; NRMSE = 9.28%), whereas in the 4:2 pattern with stronger shading stress on soybean, the random forest model showed the greatest robustness (R2 = 0.724). Interpretation results indicated that yield in monoculture systems was mainly driven by physiological traits characterized by visible-band indices, while yield in intercropping systems was dominated by structural and stress-response traits represented by near-infrared and soil-adjusted vegetation indices. The generated centimeter-scale yield maps revealed clear strip-like spatial variability driven by interspecific competition. Overall, explainable machine learning combined with UAV multispectral data shows promise for within-season yield estimation in intercropping systems and can support spatially differentiated precision management under the sampled conditions. Full article
(This article belongs to the Section Artificial Intelligence and Digital Agriculture)
Show Figures

Figure 1

23 pages, 2608 KB  
Article
Designing Predictive Models: A Comparative Evaluation of Machine Learning Algorithms for Predicting Body Carcass Fat in Ewes at Weaning
by Ahmad Shalaldeh, Mosleh Abualhaj, Ahmad Adel Abu-Shareha, Ayman Elshenawy, Yassen Saoudi, Muzammil Hussain, Ahmad Shubita, Majeed Safa and Chris Logan
Agriculture 2026, 16(4), 488; https://doi.org/10.3390/agriculture16040488 (registering DOI) - 22 Feb 2026
Abstract
Accurate estimation of Body Carcass Fat (BCF) is essential for evaluating the physiological condition of ewes. Traditional assessment via Body Condition Score (BCS) through palpation is inaccurate and subjective. BCF can now be predicted more precisely using objective measurements. This study presents a [...] Read more.
Accurate estimation of Body Carcass Fat (BCF) is essential for evaluating the physiological condition of ewes. Traditional assessment via Body Condition Score (BCS) through palpation is inaccurate and subjective. BCF can now be predicted more precisely using objective measurements. This study presents a comparative analysis of eight machine learning (ML) models for predicting BCF in Coopworth ewes, using weight and RGB-image-based body measurements. Four non-linear regression methods and four neural network architectures were evaluated using a dataset of 74 ewes with 13 independent variables. The dataset was partitioned into training (52 ewes), validation (11 ewes), and testing (11 ewes) sets. The Gradient Boosting Regression achieved the highest predictive accuracy with an R2 value of 0.9434 using body weight and width, followed by Ensemble Neural Network (R2 = 0.9371) using body weight. The findings demonstrate the effectiveness of the Gradient Boosting Regression, Ensemble Neural Network and Random Forest tree-based approaches for morphometric prediction tasks in biological applications. BCF values obtained from image analysis were validated against those derived from computerized tomography (CT), considered the gold standard. These findings highlight the potential of image-guided, ML-driven models for objective, non-invasive, cost-effective assessment of ewe body composition in modern livestock systems. Full article
Show Figures

Figure 1

24 pages, 3518 KB  
Article
A Diffusion Weighted Ensemble Framework for Robust Short-Horizon Global SST Forecasting from Multivariate GODAS Data
by Gwangun Yu, GilHan Choi, Moonseung Choi, Sun-hong Min and Yonggang Kim
Mathematics 2026, 14(4), 740; https://doi.org/10.3390/math14040740 (registering DOI) - 22 Feb 2026
Abstract
Accurate time series forecasting of sea surface temperature (SST) is essential for understanding the ocean climate system and large-scale ocean circulation, yet it remains challenging due to regime-dependent variability and correlated errors across heterogeneous prediction models. This study addresses these challenges by formulating [...] Read more.
Accurate time series forecasting of sea surface temperature (SST) is essential for understanding the ocean climate system and large-scale ocean circulation, yet it remains challenging due to regime-dependent variability and correlated errors across heterogeneous prediction models. This study addresses these challenges by formulating SST ensemble time series forecasting aggregation as a stochastic, sample-adaptive weighting problem. We propose a diffusion-conditioned ensemble framework in which heterogeneous base forecasters generate out-of-sample SST predictions that are combined through a noise-conditioned weighting network. The proposed framework produces convex, sample-specific mixture weights without requiring iterative reverse-time sampling. The approach is evaluated on short-horizon global SST forecasting using the Global Ocean Data Assimilation System (GODAS) reanalysis as a representative multivariate dataset. Under a controlled experimental protocol with fixed input windows and one-step-ahead prediction, the proposed method is compared against individual deep learning forecasters and conventional global pooling strategies, including uniform averaging and validation-optimized convex weighting. The results show that adaptive, diffusion-weighted aggregation yields consistent improvements in error metrics over the best single-model baseline and static pooling rules, with more pronounced gains in several mid- to high-latitude regimes. These findings indicate that stochastic, condition-dependent weighting provides an effective and computationally practical framework for enhancing the robustness of multivariate time series forecasting, with direct applicability to global SST prediction from large-scale geophysical reanalysis data. Full article
20 pages, 4349 KB  
Article
Agricultural Carbon Flux Estimation Using Multi-Source Remote Sensing and Ensemble Models
by Jiang Qiu, Qinrong Li, Weiyu Yu and Jinping Chen
Appl. Sci. 2026, 16(4), 2118; https://doi.org/10.3390/app16042118 (registering DOI) - 22 Feb 2026
Abstract
To accurately understand and investigate carbon fluxes in cropland ecosystems, this study adopted a machine learning ensemble model for estimation. Focusing on the Jinzhou station of the ChinaFLUX, we integrated eddy covariance carbon flux observations with multi-source satellite remote sensing data to construct [...] Read more.
To accurately understand and investigate carbon fluxes in cropland ecosystems, this study adopted a machine learning ensemble model for estimation. Focusing on the Jinzhou station of the ChinaFLUX, we integrated eddy covariance carbon flux observations with multi-source satellite remote sensing data to construct a machine learning-based cropland carbon flux estimation model. For environmental driver selection, a strategy combining correlation analysis with ecological mechanism understanding was employed to screen LST, NDVI, and NDMI as model input variables, effectively avoiding multicollinearity issues. Using footprint-weighted integrated data from 2005 to 2014 for model training and validation, a Stacking ensemble model was constructed with the RF model serving as the meta-learner to stack the predictions of RF, CART, and GBM. The ensemble model further reduced the prediction error (RMSE = 39.82), maintaining an R2 > 0.9 in most years and effectively improving predictive performance during anomalous years where single models underperformed. Based on these findings, the model was applied to analyze the spatiotemporal evolution of NEE in Jinzhou croplands from 2005 to 2014. The analysis revealed that while the region functioned overall as a carbon sink, it exhibited significant spatiotemporal heterogeneity. Spatially, the distribution followed a pattern of “strong intensity in the northeast and center, and weak intensity in the northwest and southwest.” Temporally, the sink intensity underwent significant interannual oscillations characterized by a “strengthening–weakening–re-strengthening–declining” trajectory. The high-precision prediction method proposed in this study is of great significance for revealing spatiotemporal variations in carbon sources/sinks, guiding green agricultural development, and supporting relevant policy formulation. Full article
Show Figures

Figure 1

15 pages, 1547 KB  
Article
Development and Evaluation of a Urinary Na/K Ratio Prediction Model: A Systematic Comparison from Attention-Based Deep Learning to Classical Ensemble Approaches
by Emi Yuda, Itaru Kaneko and Daisuke Hirahara
Bioengineering 2026, 13(2), 252; https://doi.org/10.3390/bioengineering13020252 (registering DOI) - 21 Feb 2026
Abstract
The urinary sodium-to-potassium (Na/K) ratio is a clinically established predictor of blood pressure and cardiovascular risk. This study aimed to develop and rigorously evaluate machine learning models for estimating the urinary Na/K ratio using four easily obtainable physiological variables: body weight, systolic blood [...] Read more.
The urinary sodium-to-potassium (Na/K) ratio is a clinically established predictor of blood pressure and cardiovascular risk. This study aimed to develop and rigorously evaluate machine learning models for estimating the urinary Na/K ratio using four easily obtainable physiological variables: body weight, systolic blood pressure, diastolic blood pressure, and pulse rate. A dataset of 82 participants was analyzed under a nested cross-validation framework to ensure strict generalization assessment. We first designed an attention-based deep learning model (MIDIP: Multi-Integrated Deep Ion Prediction). Although MIDIP showed reduced training error, nested validation revealed performance instability, indicating overfitting in this small-sample setting. We then compared classical machine learning models and ensemble strategies. Among all configurations, simple averaging of Random Forest, Gradient Boosting, and Linear Regression (Group A) achieved the best performance (MAE = 1.756, RMSE = 2.349, R2 = 0.390). In contrast, incorporating a Transformer model (Group B) degraded performance (MAE = 1.855, R2 = 0.294). Similarly, adaptive weighting (AWE) did not improve accuracy (Group A: MAE = 1.836, R2 = 0.266; Group B: MAE = 2.133, R2 = 0.035). These results demonstrate that, under limited sample conditions (N = 82), model simplicity and equal-weight ensemble integration provide superior generalization compared to attention-based or adaptively weighted deep architectures. The findings underscore the importance of strict validation and controlled model complexity when developing clinically applicable prediction models from small datasets. Full article
(This article belongs to the Section Biosignal Processing)
Show Figures

Figure 1

16 pages, 1212 KB  
Article
GenReP: An Ensemble Model for Predicting TP53 in Response to Pharmaceutical Compounds
by Austin Spadaro, Alok Sharma and Iman Dehzangi
Molecules 2026, 31(4), 739; https://doi.org/10.3390/molecules31040739 (registering DOI) - 21 Feb 2026
Abstract
TP53 is a tumor-suppressor gene involved in regulating apoptosis, DNA repair, and genomic stability. Mutations in TP53 are implicated in approximately half of all detected cancers, including breast, lung, colorectal, and ovarian cancers, making it a significant target for therapeutic interventions. Many pharmaceutical [...] Read more.
TP53 is a tumor-suppressor gene involved in regulating apoptosis, DNA repair, and genomic stability. Mutations in TP53 are implicated in approximately half of all detected cancers, including breast, lung, colorectal, and ovarian cancers, making it a significant target for therapeutic interventions. Many pharmaceutical drugs aim to restore TP53 function, and there is a need for predictive tools to assess how compounds may affect TP53 expression. In this study, we propose a new ensemble machine-learning model to predict the direction of TP53 relative gene expression in response to pharmaceutical compounds. Our model utilizes molecular fingerprints, descriptors, and scaffold-based features extracted from SMILES representations of compounds concatenated into a single feature vector. Trained using our newly generated benchmark dataset based on the Connectivity Map (CMap) database and addressing class imbalance with the Synthetic Minority Over-sampling Technique (SMOTE), our model achieves 62.9%, 93.9%, 40.3%, and 0.39 in terms of accuracy, sensitivity, specificity, and Matthews Correlation Coefficient (MCC), respectively. As the first-of-its-kind TP53 gene regulation prediction, our study serves as a convincing proof-of-concept that paves the way for future investigation. GenReP as a stand-alone predictor, its source code, and our newly generated benchmark dataset are publicly available. Full article
(This article belongs to the Special Issue Computational Insights into Protein Engineering and Molecular Design)
25 pages, 1245 KB  
Article
Machine Learning-Driven Intrusion Detection for Securing IoT-Based Wireless Sensor Networks
by Yirga Yayeh Munaye, Abebaw Demelash Gebeyehu, Li-Chia Tai, Zemenu Alem Abebe, Aeneas Bekele Workneh, Robel Berie Tarekegn, Yenework Belayneh Chekol and Getaneh Berie Tarekegn
Future Internet 2026, 18(2), 113; https://doi.org/10.3390/fi18020113 (registering DOI) - 21 Feb 2026
Abstract
Wireless sensor networks (WSNs) have become a critical component of modern Internet of Things (IoT) infrastructures; however, their constrained resources and distributed deployment expose them to various cyber threats. In this work, we present a machine learning-driven intrusion detection framework optimized for WSN-based [...] Read more.
Wireless sensor networks (WSNs) have become a critical component of modern Internet of Things (IoT) infrastructures; however, their constrained resources and distributed deployment expose them to various cyber threats. In this work, we present a machine learning-driven intrusion detection framework optimized for WSN-based IoT environments. The proposed approach employs the WSN-DS benchmark dataset and integrates adaptive synthetic sampling (ADASYN) to address class imbalance, followed by a hybrid feature selection strategy combining Feature Importance Selection (FIS) and Recursive Feature Elimination (RFE) to reduce dimensionality and improve learning efficiency. An XGBoost classifier is then trained using five-fold cross-validation to ensure robust generalization. The experimental results demonstrate that the proposed framework significantly outperforms baseline methods, achieving an overall accuracy of 99.87%, with substantial gains in terms of F1-score, precision, and recall. Comparative analysis against recent WSN-DS studies confirms the effectiveness of combining imbalance correction, optimized feature selection, and ensemble learning. These findings highlight the potential of the proposed model as a lightweight and highly accurate intrusion detection solution for emerging WSN-IoT deployments. Full article
(This article belongs to the Special Issue Machine Learning and Internet of Things in Industry 4.0)
Show Figures

Figure 1

33 pages, 12030 KB  
Article
An Interpretable Ensemble Transformer Framework for Breast Cancer Detection in Ultrasound Images
by Riyadh M. Al-Tam, Aymen M. Al-Hejri, Fatma A. Hashim, Sachin M. Narangale, Mugahed A. Al-Antari and Sarah A. Alzakari
Diagnostics 2026, 16(4), 622; https://doi.org/10.3390/diagnostics16040622 - 20 Feb 2026
Viewed by 20
Abstract
Background/Objectives: Early and accurate detection of breast cancer is essential for reducing mortality and improving patient outcomes. However, the manual interpretation of breast ultrasound images is challenging due to image variability, noise, and inter-observer subjectivity. This study aims to address these limitations [...] Read more.
Background/Objectives: Early and accurate detection of breast cancer is essential for reducing mortality and improving patient outcomes. However, the manual interpretation of breast ultrasound images is challenging due to image variability, noise, and inter-observer subjectivity. This study aims to address these limitations by developing an automated and interpretable computer-aided diagnosis (CAD) system. Methods: We propose an automated and interpretable computer-aided diagnosis (CAD) system that integrates ensemble transfer learning with Vision Transformer architectures. The system combines the Data-Efficient Image Transformer (Deit) and Vision Transformer (ViT) through concatenation-based feature fusion to exploit their complementary representations. Preprocessing, normalization, and targeted data augmentation enhance robustness, while Gradient-weighted Class Activation Mapping (Grad-CAM) provides visual explanations to support clinical interpretability. The proposed model is benchmarked against state-of-the-art CNNs (VGG16, ResNet50, DenseNet201) and Transformer models (ViT, DeiT, Swin, Beit) using the Breast Ultrasound Images (BUSI) dataset. Results: The ensemble achieved 96.92% accuracy and 97.10% AUC for binary classification, and 94.27% accuracy with 94.81% AUC for three-class classification. External validation on independent datasets demonstrated strong generalizability, with 87.76%/88.07% accuracy/AUC on BrEaST, 86.77%/85.90% on BUS-BRA, and 86.99%/86.99% on BUSI_WHU. Performance decreased for fine-grained BI-RADS classification—76.68%/84.59% accuracy/AUC on BUS-BRA and 68.75%/81.10% on BrEaST—reflecting the inherent complexity and subjectivity of clinical subclassification. Conclusions: The proposed Vision Transformer-based ensemble demonstrates high diagnostic accuracy, strong cross-dataset generalization, and clinically meaningful explainability. These findings highlight its potential as a reliable second-opinion CAD tool for breast cancer diagnosis, particularly in resource-limited clinical environments. Full article
(This article belongs to the Special Issue Artificial Intelligence in Biomedical Imaging and Signal Processing)
Show Figures

Figure 1

25 pages, 6562 KB  
Article
An Adaptive Transfer Learning Approach for Dynamic Demand Response Potential Prediction of Load Aggregators
by Dongli Jia, Huiyu Zhan, Keyan Liu, Kunhang Xie and Bin Gou
Energies 2026, 19(4), 1083; https://doi.org/10.3390/en19041083 - 20 Feb 2026
Viewed by 28
Abstract
Accurate forecasting of aggregated demand response (DR) potential is critical for load aggregators, yet remains challenging under severe data scarcity and domain shift conditions. This paper proposes a domain-adaptive transfer learning framework based on an ensemble of Random Vector Functional-Link (RVFL) neural networks [...] Read more.
Accurate forecasting of aggregated demand response (DR) potential is critical for load aggregators, yet remains challenging under severe data scarcity and domain shift conditions. This paper proposes a domain-adaptive transfer learning framework based on an ensemble of Random Vector Functional-Link (RVFL) neural networks for DR potential prediction without requiring any labeled target-domain data. By integrating domain adaptation layers and Maximum Mean Discrepancy (MMD) regularization, the proposed method explicitly reduces marginal feature distribution discrepancies between source and target domains, enabling effective knowledge transfer across heterogeneous operating scenarios. Compared with deep learning architectures, the RVFL-based framework offers favorable theoretical and practical properties for this application, including closed-form least-squares training, reduced risk of overfitting under limited data, and stable generalization under distribution shifts due to its direct-link structure and randomized hidden representations. These characteristics lead to significantly lower computational complexity and training cost than gradient-based deep models, while maintaining strong predictive capability. Case studies using real-world residential consumption data from the Pecan Street dataset demonstrate that the proposed approach consistently outperforms benchmark methods, including SVR, RF, and LSTM, across both intra-year and cross-year transfer scenarios. Reliable prediction accuracy is achieved even when only 10% of source-domain data are available, indicating strong data efficiency and scalability for practical aggregator deployment in day-ahead DR planning. Full article
43 pages, 1927 KB  
Article
A Large-Scale Empirical Study of LLM Orchestration and Ensemble Strategies for Sentiment Analysis in Recommender Systems
by Konstantinos I. Roumeliotis, Dionisis Margaris, Dimitris Spiliotopoulos and Costas Vassilakis
Future Internet 2026, 18(2), 112; https://doi.org/10.3390/fi18020112 - 20 Feb 2026
Viewed by 41
Abstract
This paper presents a comprehensive empirical evaluation comparing meta-model aggregation strategies with traditional ensemble methods and standalone models for sentiment analysis in recommender systems beyond standalone large language model (LLM) performance. We investigate whether aggregating multiple LLMs through a reasoning-based meta-model provides measurable [...] Read more.
This paper presents a comprehensive empirical evaluation comparing meta-model aggregation strategies with traditional ensemble methods and standalone models for sentiment analysis in recommender systems beyond standalone large language model (LLM) performance. We investigate whether aggregating multiple LLMs through a reasoning-based meta-model provides measurable performance advantages over individual models and standard statistical aggregation approaches in zero-shot sentiment classification. Using a balanced dataset of 5000 verified Amazon purchase reviews (1000 reviews per rating category from 1 to 5 stars, sampled via two-stage stratified sampling across five product categories), we evaluate 12 different leading pre-trained LLMs from four major providers (OpenAI, Anthropic, Google, and DeepSeek) in both standalone and meta-model configurations. Our experimental design systematically compares individual model performance against GPT-based meta-model aggregation and traditional ensemble baselines (majority voting, mean aggregation). Results show statistically significant improvements (McNemar’s test, p < 0.001): the GPT-5 meta-model achieves 71.40% accuracy (10.15 percentage point improvement over the 61.25% individual model average), while the GPT-5 mini meta-model reaches 70.32% (9.07 percentage point improvement). These observed improvements surpass traditional ensemble methods (majority voting: 62.64%; mean aggregation: 62.96%), suggesting potential value in meta-model aggregation for sentiment analysis tasks. Our analysis reveals empirical patterns including neutral sentiment classification challenges (3-star ratings show 64.83% failure rates across models), model influence hierarchies, and cost-accuracy trade-offs ($130.45 aggregation cost vs. $0.24–$43.97 for individual models per 5000 predictions). This work provides evidence-based insights into the comparative effectiveness of LLM aggregation strategies in recommender systems, demonstrating that meta-model aggregation with natural language reasoning capabilities achieves measurable performance gains beyond statistical aggregation alone. Full article
(This article belongs to the Special Issue Intelligent Agents and Their Application)
Show Figures

Figure 1

26 pages, 2804 KB  
Article
From Experiment to Prediction: Machine Learning Solutions for Concrete Strength Assessment with Steel Clamps
by Panumas Saingam, Burachat Chatveera, Gritsada Sua-Iam, Preeda Chaimahawan, Chisanuphong Suthumma, Panuwat Joyklad, Qudeer Hussain and Afaq Ahmad
Buildings 2026, 16(4), 851; https://doi.org/10.3390/buildings16040851 - 20 Feb 2026
Viewed by 45
Abstract
This study examines the confined compressive strength (Fcc) of circular, square, and rectangular column geometries under varying confinement conditions. Results indicate that circular columns have the highest Fcc values, exceeding those of square and rectangular shapes. Increased confinement through clamps significantly enhances compressive [...] Read more.
This study examines the confined compressive strength (Fcc) of circular, square, and rectangular column geometries under varying confinement conditions. Results indicate that circular columns have the highest Fcc values, exceeding those of square and rectangular shapes. Increased confinement through clamps significantly enhances compressive strength. Five machine learning models, Linear Regression, Decision Tree, Random Forest, AdaBoost, and Gradient Boosting, were used to predict Fcc based on geometric and confinement parameters. Linear Regression and Decision Tree models achieved moderate predictive performance, with R2 values of 0.84 and 0.83, respectively, and relatively higher error measures (RMSE, MAE, and MAPE), indicating limited ability to capture complex nonlinear relationships in the data. In contrast, ensemble-based methods demonstrated superior performance. The Random Forest model improved the coefficient of determination to 0.90 while substantially reducing all error metrics, reflecting enhanced generalization through bagging. The boosting-based approaches yielded the best results, with AdaBoost achieving the highest R2 value of 0.99 and the lowest RMSE, MAE, and MAPE among all models, followed closely by Gradient Boosting with an R2 of 0.98. These results confirm that ensemble learning techniques, particularly boosting algorithms, yield more accurate and robust predictions than single learners for the problem studied. Data visualization techniques, including Regression Error Characteristic curves (REC) and SHapley Additive exPlanations (SHAP) value analysis, highlighted model performance and feature importance, emphasizing the roles of confinement and geometry in compressive strength. This research demonstrates the potential of machine learning to optimize structural engineering design and suggests further exploration of alternative shapes and confinement strategies to enhance structural integrity. Full article
21 pages, 2437 KB  
Article
Evaluating SWIR Spectral Data and Random Forest Models for Copper Mineralization Discrimination in the Zhunuo Porphyry Deposit
by Jiale Cao, Lifang Wang, Xiaofeng Liu and Song Wu
Minerals 2026, 16(2), 213; https://doi.org/10.3390/min16020213 - 19 Feb 2026
Viewed by 76
Abstract
In recent years, with the widespread application of shortwave infrared (SWIR) spectroscopy in mineral identification and hydrothermal alteration studies, an increasing number of studies have attempted to integrate SWIR spectral data with machine learning approaches to fully exploit mineralization-related discriminative information embedded in [...] Read more.
In recent years, with the widespread application of shortwave infrared (SWIR) spectroscopy in mineral identification and hydrothermal alteration studies, an increasing number of studies have attempted to integrate SWIR spectral data with machine learning approaches to fully exploit mineralization-related discriminative information embedded in high-dimensional spectral datasets. In this study, the Zhunuo porphyry copper deposit in Tibet was selected as the research target. SWIR drill core spectral data were systematically acquired, and a random forest (RF) machine learning model was applied to full-band SWIR spectra (1300–2500 nm) to conduct integrated analyses of copper grade regression and mineralization discrimination. A total of 2140 drill core samples were measured, with three replicate measurements per sample, yielding 6420 spectra. After standardized preprocessing and interpolation resampling, a unified spectral feature dataset was constructed for regression and classification analyses. SWIR spectral data are characterized by a large number of bands, strong inter-band correlations, and relatively limited sample sizes; under such conditions, model generalization ability and stability become critical factors in method selection. Based on ensemble learning, the random forest model constructs multiple decision trees and aggregates their predictions through voting or averaging, effectively reducing model variance and mitigating overfitting, and is therefore well suited for high-dimensional, small-sample, and highly correlated geological spectral datasets. In porphyry copper systems, the spectral characteristics of hydrothermal alteration minerals and mineralization intensity commonly exhibit complex nonlinear relationships, which can be effectively captured by random forest models without requiring predefined functional forms. The regression results indicate that accurate quantitative prediction of copper grade based solely on SWIR spectral data remains limited. In contrast, when a threshold-based binary classification was introduced using an industrial cutoff grade of 0.2% Cu, the model achieved an overall accuracy of 75%, an F1 score of 0.69, and an area under the ROC curve (AUC) of 0.80, demonstrating strong mineralization discrimination capability and stability. Overall, the integration of SWIR spectroscopy with machine learning methods provides an efficient, reliable, and geologically interpretable technical approach for early-stage exploration and detailed drill core interpretation in porphyry copper deposits. Full article
29 pages, 3439 KB  
Article
HCHS-Net: A Multimodal Handcrafted Feature and Metadata Framework for Interpretable Skin Lesion Classification
by Ahmet Solak
Biomimetics 2026, 11(2), 154; https://doi.org/10.3390/biomimetics11020154 - 19 Feb 2026
Viewed by 154
Abstract
Accurate and timely classification of skin lesions is critical for early cancer detection, yet current deep learning approaches suffer from high computational costs, limited interpretability, and poor transparency for clinical deployment. This study presents HCHS-Net, a lightweight and interpretable multimodal framework for six-class [...] Read more.
Accurate and timely classification of skin lesions is critical for early cancer detection, yet current deep learning approaches suffer from high computational costs, limited interpretability, and poor transparency for clinical deployment. This study presents HCHS-Net, a lightweight and interpretable multimodal framework for six-class skin lesion classification on the PAD-UFES-20 dataset. The proposed framework extracts a 116-dimensional visual feature vector through three complementary handcrafted modules: a Color Module employing multi-channel histogram analysis to capture chromatic diagnostic patterns, a Haralick Module deriving texture descriptors from the gray-level co-occurrence matrix (GLCM) that quantify surface characteristics correlated with malignancy, and a Shape Module encoding morphological properties via Hu moment invariants aligned with the clinical ABCD rule. The architectural design of HCHS-Net adopts a biomimetic approach by emulating the hierarchical information processing of the human visual system and the cognitive diagnostic workflows of expert dermatologists. Unlike conventional black-box deep learning models, this framework employs parallel processing branches that simulate the selective attention mechanisms of the human eye by focusing on biologically significant visual cues such as chromatic variance, textural entropy, and morphological asymmetry. These visual features are concatenated with a 12-dimensional clinical metadata vector encompassing patient demographics and lesion characteristics, yielding a compact 128-dimensional multimodal representation. Classification is performed through an ensemble of three gradient boosting algorithms (XGBoost, LightGBM, CatBoost) with majority voting. HCHS-Net achieves 97.76% classification accuracy with only 0.25 M parameters, outperforming deep learning baselines, including VGG-16 (94.60%), ResNet-50 (94.80%), and EfficientNet-B2 (95.16%), which require 60–97× more parameters. The framework delivers an inference time of 0.11 ms per image, enabling real-time classification on standard CPUs without GPU acceleration. Ablation analysis confirms the complementary contribution of each feature module, with metadata integration providing a 2.53% accuracy gain. The model achieves perfect melanoma and nevus recall (100%) with 99.55% specificity, maintaining reliable discrimination at safety-critical diagnostic boundaries. Comprehensive benchmarking against 13 published methods demonstrates that domain-informed handcrafted features combined with clinical metadata can match or exceed deep learning fusion approaches while offering superior interpretability and computational efficiency for point-of-care deployment. Full article
(This article belongs to the Section Bioinspired Sensorics, Information Processing and Control)
Show Figures

Figure 1

28 pages, 1384 KB  
Article
Prediction of Blaine Fineness of Final Product in Cement Production Using Industrial Quality Control Data Based on Chemical and Granulometric Inputs Using Machine Learning
by Mustafa Taha Topaloğlu, Cevher Kürşat Macit, Ukbe Usame Uçar and Burak Tanyeri
Appl. Sci. 2026, 16(4), 2046; https://doi.org/10.3390/app16042046 - 19 Feb 2026
Viewed by 94
Abstract
The cement industry is central to sustainable manufacturing due to its high energy demand and associated CO2 emissions. In cement production, a substantial share of electrical energy is consumed in the clinker grinding circuit, where Blaine fineness (specific surface area, cm2 [...] Read more.
The cement industry is central to sustainable manufacturing due to its high energy demand and associated CO2 emissions. In cement production, a substantial share of electrical energy is consumed in the clinker grinding circuit, where Blaine fineness (specific surface area, cm2/g), a key quality output, affects both cement performance and specific energy consumption. However, laboratory Blaine measurements are typically available with a 30–60 min delay, which limits timely process interventions and may promote conservative operating practices (e.g., precautionary over-grinding) to secure quality. This study develops machine-learning models to predict the finished-product Blaine fineness (Blaine-F) from routinely recorded industrial quality-control inputs, including XRF-based oxide composition, derived chemical moduli (lime saturation factor, LSF; silica modulus, SM; alumina modulus, AM), laser-diffraction particle-size distribution descriptors (Q10/Q50/Q90 corresponding to D10/D50/D90 percentile diameters; and R3 residual fractions at selected cut sizes), and intermediate in-process fineness (Blaine-P). The models were trained on over 200 finished-product samples obtained from the quality-control laboratory information management system (LIMS) of Seza Cement Factory (SYCS Group, Turkey). Ridge regression, Random Forest, XGBoost, LightGBM, and CatBoost were tuned using RandomizedSearchCV with five-fold cross-validation and evaluated on a held-out test set using MAE, RMSE, and R2. The results show that the linear baseline provides limited explanatory power (Ridge: R2 ≈ 0.50), consistent with the strongly non-linear behavior of the grinding–separation system, whereas tree-based ensemble methods achieve higher predictive accuracy. XGBoost yields the best overall performance (R2 = 0.754; RMSE = 76.9 cm2/g), while Random Forest attains R2 = 0.744 with the lowest MAE (61.7 cm2/g). Explainability analyses indicate that Blaine-F is primarily influenced by the fine-tail PSD descriptor Q10 (D10 particle size) and the intermediate fineness Blaine-P, whereas chemistry-related variables (e.g., LSF and SiO2, and particularly SM) provide secondary yet meaningful contributions. These findings support the use of the proposed model as a virtual sensor to reduce decision latency associated with delayed laboratory Blaine measurements and to enable tighter fineness targeting. Potential energy and CO2 implications should be quantified using site-specific, plant-calibrated relationships between kWh/t and Blaine fineness, rather than inferred as measured outcomes within the present study. Full article
(This article belongs to the Special Issue Applications of Artificial Intelligence in Industrial Engineering)
21 pages, 1504 KB  
Article
A Data-Driven Reduced-Order Model for Rotary Kiln Temperature Field Prediction Using Autoencoder and TabPFN
by Ya Mao, Yuhang Li, Yanhui Lai and Fangshuo Fan
Appl. Sci. 2026, 16(4), 2029; https://doi.org/10.3390/app16042029 - 18 Feb 2026
Viewed by 95
Abstract
The accurate reconstruction of the internal temperature field in rotary kilns is critical for optimizing the clinker calcination process and ensuring energy efficiency. In this study, a rapid and high-fidelity surrogate modeling framework is proposed, utilizing snapshot ensembles generated by full-order Computational Fluid [...] Read more.
The accurate reconstruction of the internal temperature field in rotary kilns is critical for optimizing the clinker calcination process and ensuring energy efficiency. In this study, a rapid and high-fidelity surrogate modeling framework is proposed, utilizing snapshot ensembles generated by full-order Computational Fluid Dynamics (CFD) simulations to reconstruct the temperature field of the axial center section. The framework incorporates a symmetric Autoencoder (AE) coupled with a TabPFN network as its core components. Capitalizing on the kiln’s strong axial symmetry, this reduction–regression system efficiently maps the high-dimensional nonlinear thermodynamic topology of the central section into a compact low-dimensional latent manifold via AE, while utilizing TabPFN to establish a robust mapping between operating boundary conditions and these latent features. By leveraging the In-Context Learning (ICL) mechanism for prior-data fitting, TabPFN effectively overcomes the data scarcity inherent in high-cost CFD sampling. Predictive results demonstrate that the model achieves a coefficient of determination (R2) of 0.897 for latent feature regression, outperforming traditional algorithms by 6.53%. In terms of field reconstruction on the test set, the model yields an average temperature error of 15.31 K. Notably, 93.83% of the nodal errors are confined within a narrow range of 0–50 K, and the reconstructed distributions exhibit high consistency with the CFD benchmarks. Furthermore, compared to the hours required for full-scale simulations, the inference time is reduced to 0.45 s, representing a speedup of four orders of magnitude. Consequently, the predictive system demonstrates excellent accuracy and efficiency, serving as an effective substitute for traditional models to realize online monitoring and intelligent optimization. Full article
(This article belongs to the Special Issue Fuel Cell Technologies in Power Generation and Energy Recovery)
Back to TopTop