Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (481)

Search Parameters:
Keywords = ridge estimation

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
23 pages, 1180 KB  
Article
Carbon Emission Prediction Model for Railway Passenger Stations on the Qinghai–Tibet Plateau
by Guanguan Jia and Qingqin Wang
Sustainability 2026, 18(8), 3881; https://doi.org/10.3390/su18083881 - 14 Apr 2026
Abstract
Controlling operation-stage carbon emissions (CE) from transport buildings is crucial for China’s dual-carbon goals and the ecological security of the Qinghai–Tibet Plateau (QTP), and the sustainable development of plateau transport infrastructure. For plateau railway passenger stations (RPS), limited monitoring and distinctive high-altitude, cold-climate [...] Read more.
Controlling operation-stage carbon emissions (CE) from transport buildings is crucial for China’s dual-carbon goals and the ecological security of the Qinghai–Tibet Plateau (QTP), and the sustainable development of plateau transport infrastructure. For plateau railway passenger stations (RPS), limited monitoring and distinctive high-altitude, cold-climate operations make daily CE prediction difficult with conventional measurement- or simulation-based methods. This study develops a machine-learning approach based on a Monte Carlo synthetic database and derives engineering-standard formulas for direct use. Building scale, meteorology and passenger flow volume (PFV) were compiled for 12 representative RPS, and a large synthetic database of daily carbon emission was generated under multiple distribution constraints. With daily mean temperature, heating degree days, altitude, station floor area and PFV as inputs, four models were trained and assessed using mean absolute error, root mean square error, mean absolute percentage error (MAPE) and R2. The results show that random forest (RF) performed best, achieving ~6% MAPE and R2 > 0.99 on the test set, and markedly lower errors than multivariable linear regression. Interpretation of RF via feature importance and partial dependence shows that floor area, altitude and PFV dominate emissions and exhibit nonlinear response patterns. To improve transparency and transferability, ridge regression was used to fit a linear surrogate to RF predictions, producing engineering-standard formulas for daily and annual operation-stage CE. The formulas retain most predictive accuracy while requiring only readily obtainable variables, enabling rapid estimation and scenario analysis for cold, high-altitude RPS. The proposed workflow provides a replicable pathway for operational CE assessment in data-scarce regions and supports low-carbon planning, design and operation of RPS on the QTP, thereby contributing to more sustainable infrastructure development in high-altitude regions. Full article
(This article belongs to the Section Green Building)
17 pages, 834 KB  
Article
Improved Data-Driven Shrinkage Estimators for Regression Models Under Severe Multicollinearity
by Ali Rashash R. Alzahrani and Asma Ahmad Alzahrani
Mathematics 2026, 14(8), 1245; https://doi.org/10.3390/math14081245 - 9 Apr 2026
Viewed by 152
Abstract
Multicollinearity is a critical issue in regression analysis, often resulting in inflated variances and unstable parameter estimates. Ridge regression is a widely adopted solution to address this challenge; however, existing ridge estimators are typically tailored to specific scenarios, limiting their universal applicability. Akhtar [...] Read more.
Multicollinearity is a critical issue in regression analysis, often resulting in inflated variances and unstable parameter estimates. Ridge regression is a widely adopted solution to address this challenge; however, existing ridge estimators are typically tailored to specific scenarios, limiting their universal applicability. Akhtar and Alharthi developed ridge estimators based on condition-adjusted ridge estimators (CAREs) to handle severe multicollinearity issues. However, their approach did not account for the error variances in the estimation process. In this study, we propose improvements to these CAREs by incorporating error variances, resulting in the development of multiscale ridge estimators (MSRE1, MSRE2, MSRE3 and MSRE4) that more effectively address the challenges posed by severe multicollinearity. We compare the performance of our newly proposed estimators with ordinary least square (OLS) and other existing ridge estimators using both simulation studies and real-life datasets. The evaluation, based on estimated mean squared error (MSE), demonstrates that the proposed estimators consistently outperform existing methods, particularly in scenarios with significant multicollinearity, larger sample sizes, and higher predictor dimensions. Results from three real-life datasets further validate the proposed estimators’ ability to reduce estimation error and improve predictive accuracy across diverse practical applications. Full article
(This article belongs to the Special Issue Statistical Machine Learning: Models and Its Applications)
Show Figures

Figure 1

27 pages, 23751 KB  
Article
A Mathematical Framework for Retinal Vessel Segmentation: Fractional Hessian-Based Curvature Analysis
by Priyanka Harjule, Mukesh Delu, Rajesh Kumar and Pilani Nkomozepi
Fractal Fract. 2026, 10(4), 246; https://doi.org/10.3390/fractalfract10040246 - 8 Apr 2026
Viewed by 165
Abstract
This study proposes an improved retinal blood vessel segmentation method to enhance the diagnosis of microvascular retinal complications. The proposed method extracts local shape features from retinal images utilizing a fractional Hessian matrix, which models blood vessels as surface structures characterized by ridges [...] Read more.
This study proposes an improved retinal blood vessel segmentation method to enhance the diagnosis of microvascular retinal complications. The proposed method extracts local shape features from retinal images utilizing a fractional Hessian matrix, which models blood vessels as surface structures characterized by ridges and valleys resulting from variations in curvature. The methodology integrates adaptive principal curvature estimation with a new framework leveraging the fractional Hessian matrix with nonsingular and nonlocal kernels. The effectiveness of the suggested method is assessed using publicly accessible datasets, including DRIVE, HRF, STARE, and some real images obtained from a local hospital. The proposed segmentation achieves 96.77% accuracy and 98.82% specificity on the DRIVE database, 96.91% accuracy and 98.69% specificity on STARE, and 95.90% accuracy and 98.36% specificity on the HRF database. Optimal parameters for the fractional order and Gaussian standard deviation were empirically determined by maximizing segmentation accuracy. Our findings show that the proposed approach achieves competitive performance compared to the listed methods, including several deep learning approaches, while maintaining significant computational efficiency. The output of the suggested method can be further utilized with deep learning techniques, which will be applied in the clinical context of diabetic retinopathy and glaucoma to identify abnormalities likely related to disease progression and different stages. Full article
Show Figures

Figure 1

15 pages, 1474 KB  
Article
Prognostic Power of Ensemble Learning in Colorectal Cancer with Peritoneal Metastasis: A Multi-Institutional Analysis
by Yoshiko Bamba, Michio Itabashi, Hirotoshi Kobayashi, Kenjiro Kotake, Masayasu Kawasaki, Yukihide Kanemitsu, Yusuke Kinugasa, Hideki Ueno, Kotaro Maeda, Takeshi Suto, Kimihiko Funahashi, Heita Ozawa, Fumikazu Koyama, Shingo Noura, Hideyuki Ishida, Masayuki Ohue, Tomomichi Kiyomatsu, Soichiro Ishihara, Keiji Koda, Hideo Baba, Kenji Kawada, Yojiro Hashiguchi, Takanori Goi, Yuji Toiyama, Naohiro Tomita, Eiji Sunami, Yoshito Akagi, Jun Watanabe, Kenichi Hakamada, Goro Nakayama, Kenichi Sugihara and Yoichi Ajiokaadd Show full author list remove Hide full author list
Bioengineering 2026, 13(4), 434; https://doi.org/10.3390/bioengineering13040434 - 8 Apr 2026
Viewed by 205
Abstract
Background: Owing to significant clinical heterogeneity, the achievement of accurate survival forecasting for individuals with colorectal cancer and peritoneal metastasis continues to be a complex undertaking. We aimed to transcend traditional prognostic limitations by evaluating machine learning boosting models against standard regression-based methods [...] Read more.
Background: Owing to significant clinical heterogeneity, the achievement of accurate survival forecasting for individuals with colorectal cancer and peritoneal metastasis continues to be a complex undertaking. We aimed to transcend traditional prognostic limitations by evaluating machine learning boosting models against standard regression-based methods in terms of estimating overall survival (OS). Methods: Utilizing a multi-institutional registry of 150 patients diagnosed with synchronous peritoneal metastasis of colorectal cancer, we integrated 124 clinicopathological variables to refine our predictive models. Beyond standard preprocessing—including standardization and median imputation—we rigorously compared XGBoost and LightGBM against Ridge, Lasso, and linear regression via five-fold cross-validation. To specifically address right-censoring, an XGBoost Cox model was implemented and validated using Harrell’s C-index, with SHAP and LIME providing essential model interpretability. Results: Boosting models consistently outperformed linear alternatives, which struggled with high error rates and negative R2 values. Specifically, XGBoost achieved an MAE of 475 ± 60 and an RMSE of 585 ± 88. The XGBoost Cox model reached a C-index of 0.64 ± 0.06. SHAP analysis highlighted inflammatory markers and peritoneal disease extent as the most influential prognostic drivers. Conclusions: While boosting models offer a clear accuracy advantage over linear methods, their prognostic power remains moderate. These findings underscore the potential of ensemble learning in oncology, yet mandate external validation before these tools can be integrated into clinical decision-making. Full article
(This article belongs to the Section Biosignal Processing)
Show Figures

Figure 1

30 pages, 507 KB  
Article
Beyond MSE in Poisson Ridge Regression: New Ridge Parameter Estimators with Additional Distributional Performance Criteria
by Selman Mermi
Mathematics 2026, 14(7), 1190; https://doi.org/10.3390/math14071190 - 2 Apr 2026
Viewed by 203
Abstract
Despite its widespread use for mitigating multicollinearity in count data models, Poisson ridge regression (PRR) remains methodologically constrained by the choice of the ridge parameter k. Existing studies predominantly evaluate ridge parameter estimators using only the mean squared error (MSE) criterion, largely [...] Read more.
Despite its widespread use for mitigating multicollinearity in count data models, Poisson ridge regression (PRR) remains methodologically constrained by the choice of the ridge parameter k. Existing studies predominantly evaluate ridge parameter estimators using only the mean squared error (MSE) criterion, largely neglecting their distributional properties and estimation stability. Such a narrow evaluation framework may yield unreliable inference, particularly under high correlation and small sample sizes. This study makes two original contributions to the PRR literature. First, we conduct a comprehensive comparison of 13 commonly used ridge parameter estimators and introduce two new estimators that exhibit superior empirical performance. Second, we extend performance evaluation beyond MSE by incorporating outlier ratios and conformity to normality, thereby establishing a multidimensional framework that explicitly addresses distributional robustness and estimator stability. Monte Carlo simulations across 180 scenarios—varying the number of predictors, sample size, correlation level, and intercept value—show that several estimators deemed optimal under MSE perform poorly in terms of outlier prevalence and normality. In contrast, the proposed estimators consistently achieve a balanced performance between error minimization and distributional stability. Two real-data applications further support these findings. Full article
(This article belongs to the Special Issue Statistical Models and Their Applications)
Show Figures

Figure 1

23 pages, 2936 KB  
Article
A Global Multi-Hazard Framework for Projecting Climate Migration Flows to 2100 Along Shared Socioeconomic Pathways (SSPs)
by Zachary M. Hirsch, Danielle N. Medgyesi, Jasmina M. Buresch and Jeremy R. Porter
Climate 2026, 14(4), 81; https://doi.org/10.3390/cli14040081 - 2 Apr 2026
Viewed by 503
Abstract
Climate-induced migration is increasingly recognized as a major demographic consequence of environmental change, yet projections vary widely due to differences in spatial scale, hazard coverage, and modeling approaches. This study introduces the First Street Global Climate Migration Model (FS-GCMM), a globally consistent, multi-hazard [...] Read more.
Climate-induced migration is increasingly recognized as a major demographic consequence of environmental change, yet projections vary widely due to differences in spatial scale, hazard coverage, and modeling approaches. This study introduces the First Street Global Climate Migration Model (FS-GCMM), a globally consistent, multi-hazard framework that estimates climate-driven population redistribution at a 12.5 km resolution across all countries through 2100. The model integrates high-resolution global climate hazard datasets, including flood (GloFAS), wind (IBTrACS and ERA5), drought (ERA5), wildfire (Global Fire Atlas), and extreme heat and cold (ERA5-LAND) datasets, with gridded population data from NASA SEDAC’s Gridded Population of the World (GPWv4) and Shared Socioeconomic Pathway (SSP) projections. To identify climate-related migration effects, we applied within-country propensity score matching to construct balanced samples of exposed and unexposed grid cells with similar socioeconomic, demographic, geographic, and governance characteristics. Hazard-specific impacts on annualized population change from 2000 to 2020 were then estimated using mixed-effects ridge regression with country-level random effects to account for cross-national heterogeneity and multicollinearity. These empirically derived coefficients were applied to SSP1-2.6, SSP2-4.5, and SSP5-8.5 scenarios to project future climate-driven outmigration, which was subsequently redistributed using a spatial attractiveness framework incorporating economic opportunity, population density, climate safety, and geographic proximity. Results indicate statistically significant negative effects of all modeled hazards on population retention globally, with approximately 199.5 million people projected to experience climate-driven displacement by 2055 under SSP2-4.5. Full article
Show Figures

Figure 1

16 pages, 3658 KB  
Article
Runoff and Sediment Flux on the North Coast of KwaZulu-Natal: Counter-Acting Beach Erosion from Rising Seas?
by Mark R. Jury
Coasts 2026, 6(2), 13; https://doi.org/10.3390/coasts6020013 - 1 Apr 2026
Viewed by 305
Abstract
A remote analysis of coastal sedimentation in northern KwaZulu-Natal (KZN), South Africa, describes how summer runoff and winter wave-action operate within a highly variable climate. Despite rising sea levels, the sediment flux can sustain beaches under certain conditions. Daily satellite red-band reflectivity and [...] Read more.
A remote analysis of coastal sedimentation in northern KwaZulu-Natal (KZN), South Africa, describes how summer runoff and winter wave-action operate within a highly variable climate. Despite rising sea levels, the sediment flux can sustain beaches under certain conditions. Daily satellite red-band reflectivity and ocean–atmosphere reanalysis datasets were studied over the period of 2018–2025. Statistical results indicate that streamflow discharges are spread northward by oblique wave-driven currents. Sediment concentrations peak during late winter (>1 mg/L, May–October) when deep turbulent mixing (>40 m) mobilizes sand from the seabed. A case study from September 2021 revealed that ridging high-pressure/cut-off low weather patterns can simultaneously increase streamflow, wave energy, and wind power, creating a surf-zone sediment conveyor along the coast of northern KZN. Long-term climate diagnostics from 1981 to 2025 reveal upward trends in coastal runoff, vegetation, and turbidity (0.29 σ/yr) that point to an increasingly vigorous water cycle. The warming of the southeast Atlantic intensifies the sub-tropical upper-level westerlies and late winter storms over southeast Africa. These processes occur in 5–8 year cycles and drive shoreline advance and retreat, from accretion ~1 T/m and storm surge inundations up to 5.5 m. Using Digital Earth, it was noted that ~1/4 of beaches around Africa are gaining sediment while ~1/3 are eroding. Although remote information could not close the sediment budget, realistic estimates of long-shore transport in the surf-zone (>104 kg/yr/m) and on the beach (>103 kg/yr/m) were calculated. These provide an emerging explanation for the resilience of northern KZN beaches, as sea levels rise at a rate of 0.6 cm/yr. Full article
Show Figures

Figure 1

13 pages, 44672 KB  
Article
ARMANI: Dictionary-Learning-Inspired Data-Free Deep Generative Modeling with Meta-Attention and Implicit Preconditioning for Compressively Sampled Magnetic Resonance Imaging
by Ming Wu, Jing Cheng, Qingyong Zhu and Dong Liang
Electronics 2026, 15(7), 1402; https://doi.org/10.3390/electronics15071402 - 27 Mar 2026
Viewed by 242
Abstract
Magnetic resonance imaging (MRI) reconstruction from undersampled k-space data enables accelerated acquisition but leads to a severely ill-posed inverse problem. Although supervised deep learning methods have achieved strong performance, they typically rely on large paired datasets that are difficult to obtain in clinical [...] Read more.
Magnetic resonance imaging (MRI) reconstruction from undersampled k-space data enables accelerated acquisition but leads to a severely ill-posed inverse problem. Although supervised deep learning methods have achieved strong performance, they typically rely on large paired datasets that are difficult to obtain in clinical practice. To address these limitations, we propose a dictionary-learning-inspired dAta-fRee deep generative modeling with Meta-Attention and implicit precoNditIoning for compressively sampled MRI (CS-MRI), termed ARMANI. Specifically, a meta-attention-augmented deep image prior (MA-DIP) generator performs a joint optimization over the latent input η and the network parameter θ, where η is regularized via gradient-domain sparsity and θ is constrained by a ridge penalty, mirroring the adaptive estimation of sparse coefficients and an empirical sparsifying dictionary. Furthermore, we integrate a single-step pseudo-orthogonal projection to achieve implicit preconditioning, which modulates the loss landscape and mitigates ill-conditioning of the forward operator. Experimental results demonstrate that ARMANI consistently outperforms existing SOTA data-free and self-supervised methods, and, with limited training data, achieves performance comparable to or slightly better than the supervised benchmark MoDL, with effective artifact suppression and faithful recovery of fine structural details. Overall, ARMANI shows strong scalability and potential for practical deployment in fully data-free CS-MRI reconstruction scenarios. Full article
Show Figures

Figure 1

27 pages, 8176 KB  
Article
Climate and Vegetation Dominate Lake Eutrophication in the Inner Mongolia–Xinjiang Plateau (2000–2024)
by Yuzheng Zhang, Feifei Cao, Yuping Rong, Linglong Wen, Wei Su, Jianjun Wu, Yaling Yin, Zhilin Zi, Shasha Liu and Leizhen Liu
Remote Sens. 2026, 18(7), 988; https://doi.org/10.3390/rs18070988 - 25 Mar 2026
Viewed by 492
Abstract
Lakes on the Inner Mongolia–Xinjiang Plateau (IMXP) are increasingly vulnerable to eutrophication under climate change and human pressure, yet long-term monitoring remains limited by sparse field sampling. Here, we reconstruct multi-decadal trophic dynamics across the IMXP using Landsat time series and temporally transferable [...] Read more.
Lakes on the Inner Mongolia–Xinjiang Plateau (IMXP) are increasingly vulnerable to eutrophication under climate change and human pressure, yet long-term monitoring remains limited by sparse field sampling. Here, we reconstruct multi-decadal trophic dynamics across the IMXP using Landsat time series and temporally transferable machine-learning models and further quantify the underlying natural and anthropogenic drivers. We compiled monthly in situ water-quality observations (chlorophyll-a, Chl-a; total phosphorus, TP; total nitrogen, TN; Secchi depth, SD; and permanganate index, CODMn;) and calculated the trophic level index (TLI). After rigorous quality control and monthly aggregation, we compiled a dataset of 1345 matched lake–month samples spanning 2000–2024, and divided it into a training set (n = 1076; ≤2019) and an independent test set (n = 269; 2020–2024) to evaluate temporal transferability. We utilized Google Earth Engine to generate monthly surface reflectance composites from Landsat 7 ETM+, Landsat 8 OLI, and Landsat 9 OLI-2. Four supervised regression algorithms—ridge regression (RR), support vector regression (SVR), random forest (RF), and eXtreme Gradient Boosting (XGBoost)—were trained to estimate TLI. On the independent test period, XGBoost performed best (R2 = 0.780, RMSE = 3.290, MAE = 1.779), followed by RF (R2 = 0.770, RMSE = 3.364), SVR (R2 = 0.700, RMSE = 3.842), and RR (R2 = 0.630, RMSE = 4.267); we then used XGBoost to reconstruct monthly and yearly TLI for 610 perennial grassland lakes from 2000 to 2024. From 2000 to 2024, the annual mean TLI (48–49) across the IMXP exhibited a statistically significant upward trend (slope = 0.0158 TLI yr−1; 95% confidence interval (CI) = 0.0050–0.0267; p = 0.006). Meanwhile, spatial heterogeneity was distinct (TLI: 41.51–59.70). High values concentrated in endorheic and desert–oasis basins (e.g., Eastern Inner Mongolia Plateau, >51), whereas lower values characterized high-altitude regions (e.g., Yarkant River, <45). Overall, trends ranged from −0.49 to 0.51 yr−1, increasing in 54% of lakes (15.6% significantly) and decreasing in 46% (15.4% significantly). Attribution analyses identified NDVI (33.92%) and temperature (21.67%) as dominant drivers (55.59% combined), followed by precipitation (13.99%) and human proxies (30.42% combined: population 10.66%, grazing 10.31%, built-up 9.45%). Across 53 sub-basins, NDVI was the primary driver in 28, followed by temperature (11), population (7), precipitation (3), grazing (3), and built-up land (1); notably, the top two drivers explained 56.6–87.1% of variations. TWFE estimates revealed bidirectional NDVI effects (significant in 31/53): positive associations in 22 basins were linked to nutrient retention, contrasting with negative effects in nine basins associated with agricultural return flows. Temperature effects were significant in 15 basins and predominantly negative (14/15), except for the Qiangtang Plateau. Overall, eutrophication risk across the IMXP lake region reflects the combined influences of climatic conditions, vegetation conditions, and human activities, with their relative contributions varying among basins. Full article
Show Figures

Figure 1

34 pages, 1788 KB  
Article
A Two-Stage Comparative Framework for Predicting Photovoltaic Cleaning Schedules: Modeling and Comparisons Based on Real and Simulated Data
by Ali Al-Humairi, Enmar Khalis, Zuhair A. Al Hemyari and Peter Jung
Appl. Sci. 2026, 16(6), 2976; https://doi.org/10.3390/app16062976 - 19 Mar 2026
Viewed by 287
Abstract
This study develops and validates a two-stage comparative framework for predicting Photovoltaic (PV) cleaning schedules by integrating high-resolution operational data with regression-based simulated datasets generated from statistical models trained on real measurements. The work directly addresses the growing need to assess whether model-based [...] Read more.
This study develops and validates a two-stage comparative framework for predicting Photovoltaic (PV) cleaning schedules by integrating high-resolution operational data with regression-based simulated datasets generated from statistical models trained on real measurements. The work directly addresses the growing need to assess whether model-based regression-based simulated data can reliably substitute real measurements in predictive PV maintenance. These models are employed to generate clean-condition power baselines and to estimate daily energy losses attributable to soiling under two distinct paradigms: (i) using real historical PV performance and environmental measurements, and (ii) using regression-derived, regression-based simulated data representing idealized clean operating conditions. Model performance is rigorously quantified using correlation coefficients (R), coefficients of determination (R2), mean absolute deviations, and binary classification metrics including accuracy, precision, recall, and F1-score. The comprehensive results demonstrate that regression-based simulated datasets exhibit high fidelity with real measurements across key electrical variables. This is evident for datasets generated using PLSR, Ridge Regression, and Robust Regression. Strong correlations are observed for DC power (R2 = 0.9545) and DC current (R2 = 0.9520), with mean deviations consistently below 2.2%. When a threshold-based binary decision rule (“clean” versus “do not clean”) is applied, cleaning decisions derived from simulated and real datasets show near-perfect concordance, achieving a mean F1-score of 0.9792. These results indicate that for a fixed performance-loss threshold, models using regression-based simulated data reproduce real-data-based cleaning triggers with an accuracy exceeding 97%. Furthermore, the findings confirm that regression-based simulation frameworks constitute a reliable and scalable foundation for data-driven PV maintenance optimization. By enabling efficient cleaning scheduling, these frameworks can significantly reduce operational expenditure and maximize energy yield, particularly in regions where continuous, high-quality PV monitoring data are limited or difficult to obtain. Full article
Show Figures

Figure 1

26 pages, 1011 KB  
Article
A Study on Machine Learning-Based Cost Estimation Models for AI Training Data Construction
by Yoon-Seok Ko and Bong Gyou Lee
Appl. Sci. 2026, 16(6), 2891; https://doi.org/10.3390/app16062891 - 17 Mar 2026
Viewed by 547
Abstract
This study proposes an explainable machine learning framework for estimating the total project cost (TPC) of AI training-data construction, where cost information is difficult to structure due to heterogeneous workflows and quality requirements. Using 386 public AI training-data projects conducted between 2020 and [...] Read more.
This study proposes an explainable machine learning framework for estimating the total project cost (TPC) of AI training-data construction, where cost information is difficult to structure due to heterogeneous workflows and quality requirements. Using 386 public AI training-data projects conducted between 2020 and 2022, we derive 24 numerical predictors from standardized final reports and construct three input tracks: a baseline feature set, a principal component analysis (PCA)-enhanced set, and a factor analysis (FA)–enhanced set capturing latent cost structures. Four regression models (Ridge, Random Forest, XGBoost, and LightGBM) are evaluated using nested cross-validation. XGBoost achieves the best overall performance across all three tracks (Baseline, PCA-enhanced, and FA-enhanced). Among them, PCA-enhanced XGBoost attains the highest predictive accuracy (R2 = 0.868; RMSE = 1084.9; MAE = 746.9; MAPE = 0.358; pooled out-of-fold), while Baseline XGBoost yields the lowest MAE (731.4; R2 = 0.863). To support transparent decision-making, Shapley Additive exPlanations (SHAP)-based attribution and scenario-based sensitivity analyses are conducted. Results show that project scale and process-level unit costs are dominant cost-drivers, while cloud usage, expert participation, and de-identification requirements exhibit secondary effects. The proposed framework provides an interpretable, data-driven approach to cost information management and decision support for data-intensive AI projects. Full article
Show Figures

Figure 1

30 pages, 1713 KB  
Article
Safe-Calibrated TCN–Transformer Transfer Learning for Reliable Battery SoH Estimation Under Lab-to-Field Domain Shift
by Kumbirayi Nyachionjeka and Ehab H. E. Bayoumi
World Electr. Veh. J. 2026, 17(3), 149; https://doi.org/10.3390/wevj17030149 - 17 Mar 2026
Viewed by 526
Abstract
Battery state-of-health (SoH) estimation is central to transportation electrification because it conditions safety limits, warranty accounting, power capability management, and long-horizon fleet optimization. Although deep temporal architectures can achieve high laboratory accuracy, field deployment is frequently limited by laboratory (Lab)-to-field (L2F) domain shift [...] Read more.
Battery state-of-health (SoH) estimation is central to transportation electrification because it conditions safety limits, warranty accounting, power capability management, and long-horizon fleet optimization. Although deep temporal architectures can achieve high laboratory accuracy, field deployment is frequently limited by laboratory (Lab)-to-field (L2F) domain shift that alters input statistics, feature definitions, and noise regimes. Under such a shift, predictors may remain strongly monotonic, preserving degradation ordering and become operationally unreliable due to systematic output distortion (e.g., compression/warping of the SoH scale). A deployment-complete L2F transfer learning pipeline is presented, built around a gated Temporal Convolutional Network (TCN)–Transformer fusion backbone, domain-specific adapters and heads, alignment-regularized fine-tuning, and row-level inference via sliding-window overlap averaging. To address the dominant deployment failure mode, a Safe Calibration stage robustly filters calibration pairs and selects among candidate calibrators under a strict do-no-harm criterion. On an unseen deployment stream (2154 labeled rows), overlap-averaged raw inference achieves MAE = 0.0439, RMSE = 0.0501, and R2 = 0.7451, consistent with mid-to-high SoH range compression, while Safe Calibration (Isotonic-Balanced selected) corrects nonlinear scaling without violating monotonic structure, improving to MAE = 0.0188, RMSE = 0.0252, and R2 = 0.9357 to obtain a complete understanding of the challenges due to domain shifts, evaluation is extended to include other architecture baselines such as TCN-only, Transformer-only, Gated Recurrent Unit (GRU), and Long Short-Term Memory (LSTM), and a Ridge regression baseline. Also added is explicit alignment and calibration ablations that include CORAL off/on, that is, none vs. Safe-Global vs. Context-Aware under identical leakage-safe splits and the same overlap-averaged deployment inference operator. This work goes beyond peak-score reporting and looks at the robustness of a pipeline under domain shift, which is quantified across four random seeds and multiple deployment streams, with uncertainty summarized via mean ± std and bootstrap confidence intervals for Mean of Absolute value of Errors (MAE)/Root of the Mean of the Square of Errors (RMSE) computed from per-example absolute errors. Full article
(This article belongs to the Section Storage Systems)
Show Figures

Figure 1

22 pages, 10555 KB  
Article
Deep Learning-Based Recognition of Arch-Back Direction in Bare-Root Strawberry Seedlings for Mechanized Transplanting
by Jinhao Zhou, Pengcheng Zhang, Menglei Wei, Wei Liu, Jiawei Shi, Youheng Tan and Jianping Hu
Agriculture 2026, 16(6), 657; https://doi.org/10.3390/agriculture16060657 - 13 Mar 2026
Viewed by 271
Abstract
Correct arch-back orientation is essential in ridge-based strawberry transplanting. Improper orientation can increase soil contact and soil-borne disease risk, leading to yield loss and reduced harvest efficiency. In current practice, arch-back orientation of bare-root seedlings is still mainly judged and corrected manually, which [...] Read more.
Correct arch-back orientation is essential in ridge-based strawberry transplanting. Improper orientation can increase soil contact and soil-borne disease risk, leading to yield loss and reduced harvest efficiency. In current practice, arch-back orientation of bare-root seedlings is still mainly judged and corrected manually, which is labor-intensive and not always accurate under field conditions. Although plug seedlings are easier for mechanized transplanting, they are about three times more expensive than bare-root seedlings. Therefore, bare-root seedlings remain widely used for cost-effective production. However, accurate real-time orientation perception for bare-root seedlings is still challenging because stems are thin, morphology varies widely, and leaves often occlude key curvature cues. To address this gap, we propose a lightweight machine-vision method for bare-root strawberry seedlings that detects three characteristic keypoints on the new stem. The three-keypoint design is inspired by farmers’ practical judgement: farmers often determine arch-back direction by observing the stem and using manual touch to sense curvature changes. Similarly, three keypoints provide a simple geometric representation of curvature trend, enabling real-time estimation of both arch-back direction and bending angle. Physical tests on 100 bare-root seedlings achieved a 93% agronomically compliant orientation rate, with an MAE of 5.74° and an RMSE of 7.44° for bending-angle estimation. For edge deployment, the optimized model achieved real-time performance on an embedded GPU platform, reaching 152.51 FPS (FP16) and 154.26 FPS (INT8). Overall, the proposed method provides a practical perception module that can be integrated into strawberry transplanting machines to support cost-effective, orientation-aware mechanized transplanting. Full article
(This article belongs to the Section Agricultural Technology)
Show Figures

Figure 1

17 pages, 566 KB  
Article
Analyst-of-Record: A Proof-of-Concept for Influence-Based Analyst Credit Assignment in Human-Feedback Decision Support
by Devon L. Brown and Danda B. Rawat
Electronics 2026, 15(6), 1210; https://doi.org/10.3390/electronics15061210 - 13 Mar 2026
Viewed by 341
Abstract
The purpose of this study is to examine whether analyst-level credit can be assigned quantitatively in a lightweight human-feedback decision-support pipeline. In intelligence and national security workflows, analysts often provide edits, comments, and evaluative feedback during the production of analytic products, yet these [...] Read more.
The purpose of this study is to examine whether analyst-level credit can be assigned quantitatively in a lightweight human-feedback decision-support pipeline. In intelligence and national security workflows, analysts often provide edits, comments, and evaluative feedback during the production of analytic products, yet these intermediate contributions are usually discarded, leaving no auditable record of how individual feedback shaped the final output. To address this problem, this study proposes a proof-of-concept Analyst-of-Record framework that combines synthetic analyst feedback, a linear ridge reward model, first-order influence functions, and additive Shapley aggregation to estimate both feedback-item and analyst-level contribution scores. The research design uses the Fact Extraction and VERification (FEVER) fact-verification dataset under controlled experimental settings. The pipeline retrieves evidence with Best Matching 25 (BM25), generates a grounded template-based response, derives three synthetic analyst feedback channels from FEVER annotations, trains a reward model on simple claim–answer and analyst-identity features, and aggregates per-feedback influence scores into an Analyst Contribution Index (ACI). The main experiments are conducted on a 500-claim subset across five random seeds, with additional ablation and bootstrap analyses used to assess sensitivity and stability. The findings show that the reward model achieves a mean validation R2 of 0.801±0.037, indicating that the synthetic feedback signals are learnable under the selected featureization. The analyst-level contribution scores remain stable across random seeds, with approximately half of the total influence magnitude attributed to the explanation-quality channel and the remainder split across the other two channels. Ablation results further show that removing the explanation-quality channel collapses validation fit, while bootstrap resampling demonstrates tight concentration of absolute ACI magnitudes. Theoretically, this study extends attribution research beyond document-only grounding by showing how analyst feedback itself can be modeled as an object of contribution analysis. It also demonstrates that influence functions and Shapley-style aggregation can be adapted into a tractable framework for estimating interpretable analyst-level credit in a reproducible experimental setting. Practically, the proposed framework offers an initial foundation for more traceable and accountable decision-support workflows in which intermediate analyst contributions can be preserved rather than lost. The results also provide a feasible implementation path for future systems that incorporate stronger generators, richer evidence representations, and real analyst annotations. Full article
(This article belongs to the Section Computer Science & Engineering)
Show Figures

Figure 1

22 pages, 2804 KB  
Article
A Comprehensive Evaluation Method for Greenhouse-Grown Lettuce Based on RGB Images and Hyperspectral Data
by Duoer Ma, Hong Ren, Qi Zeng, Yidi Liu, Lulu Ma, Qiang Zhang, Ze Zhang and Jiangli Wang
Agronomy 2026, 16(6), 600; https://doi.org/10.3390/agronomy16060600 - 11 Mar 2026
Viewed by 369
Abstract
Quality grading of greenhouse lettuce requires rapid external appearance screening and nondestructive internal quality assessment. However, existing detection methods struggle to simultaneously evaluate both external and internal quality while maintaining efficiency, resulting in a lack of scientific and comprehensive integrated evaluation standards for [...] Read more.
Quality grading of greenhouse lettuce requires rapid external appearance screening and nondestructive internal quality assessment. However, existing detection methods struggle to simultaneously evaluate both external and internal quality while maintaining efficiency, resulting in a lack of scientific and comprehensive integrated evaluation standards for current crop grading. To address this issue, this study leveraged the technical strengths of different sensors to construct separate models: an RGB image-based monitoring model for external quality and a hyperspectral-based estimation model for internal quality. Using a combined objective–subjective weighting method, this approach scientifically integrated external and internal quality monitoring indicators to establish a comprehensive evaluation method for greenhouse lettuce quality. The results demonstrate that features such as canopy projection area, compactness, and color components can be extracted from RGB images. Combined with Ridge regression, this approach achieves high-accuracy estimation of lettuce fresh weight and leaf area (R2 ≥ 0.880). For intrinsic quality, by combining hyperspectral data with the CARS and SPA band selection algorithms, a Random Forest (RF)-based inversion model for chlorophyll, soluble sugar, protein, and vitamin C content was developed. The AHP-CRITIC method effectively resolved the weight imbalance caused by an excessive coefficient of variation in appearance indicators, thereby achieving the scientific integration of appearance and internal quality data. The grading outcomes of this integrated evaluation method were highly consistent with industry standards (kappa coefficient: 0.788). This approach establishes an effective link between the rapid monitoring of external and internal quality for comprehensive evaluation, providing a novel technical pathway and scientific basis for nondestructive post-harvest detection and automated grading of greenhouse vegetables. Full article
(This article belongs to the Section Precision and Digital Agriculture)
Show Figures

Figure 1

Back to TopTop