Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (964)

Search Parameters:
Keywords = learning-based design automation

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
15 pages, 392 KB  
Article
Random Forest Predicts Human Ratings of Creative Stories Using Very Small Training Samples
by Baptiste Barbot and Thomas Calogero Kiekens
Behav. Sci. 2026, 16(4), 576; https://doi.org/10.3390/bs16040576 (registering DOI) - 11 Apr 2026
Abstract
The Consensual Assessment Technique (CAT) is a gold standard of creativity assessment which provides valid product-based creativity scores that are contextually grounded (stemming from raters with unique expertise, culturally and historically situated). However, its implementation is often demanding (raters’ burden, complex rating designs). [...] Read more.
The Consensual Assessment Technique (CAT) is a gold standard of creativity assessment which provides valid product-based creativity scores that are contextually grounded (stemming from raters with unique expertise, culturally and historically situated). However, its implementation is often demanding (raters’ burden, complex rating designs). This study investigates whether machine learning can effectively simulate expert-panel judgments of creativity using minimal training data. Using a dataset of 411 short stories, we compared the performance of Random Forest (RF), Gradient Boosted Trees, and Decision Tree models, based on story length and Divergent Semantic Integration, to predict expert CAT ratings by (1) identifying the optimal algorithm and (2) the minimum training sample size required for reliable prediction. Results indicate that RF consistently outperformed other algorithms, achieving high correlations with CAT scores (r = 0.80) using as few as 25 training stories. Furthermore, RF demonstrated superior accuracy and lower reliance on story length compared to LLM-based scoring models. These findings provide a robust proof-of-concept for using simulated expert panels as a scalable alternative to (decontextualized) automated assessment methods, while reducing human raters’ burden and the logistical constraints of complex rating designs. Extension of this work to different contexts, creativity tasks and domains are necessary to gauge its generalizability. Full article
(This article belongs to the Section Cognition)
Show Figures

Figure 1

20 pages, 4549 KB  
Article
Online Track Anomaly Detection: Comparison of Different Machine Learning Techniques Through Injection of Synthetic Defects on Experimental Datasets
by Giovanni Bellacci, Luca Di Carlo, Marco Fiaschi, Luca Bocciolini, Carmine Zappacosta and Luca Pugi
Machines 2026, 14(4), 424; https://doi.org/10.3390/machines14040424 - 10 Apr 2026
Abstract
The adoption of instrumented wheelsets on diagnostic trains offers the possibility of continuous monitoring of wheel–rail contact forces. The collection of large datasets can be exploited for diagnostic purposes, aiming to localize specific track defects, allowing significant improvements in terms of safety and [...] Read more.
The adoption of instrumented wheelsets on diagnostic trains offers the possibility of continuous monitoring of wheel–rail contact forces. The collection of large datasets can be exploited for diagnostic purposes, aiming to localize specific track defects, allowing significant improvements in terms of safety and maintenance costs. Machine learning (ML) techniques can be used to automate anomaly detection. In this work, the authors compare the application of various ML algorithms based on the identification of different frequency or time-based features of analyzed signals. To perform the activity, a significant number and variety of local defects have been included in the recorded data. From a practical point of view, the insertion of real known defects into an existing line is extremely time-consuming, expensive, and not immune to safety issues. On the other hand, the design of anomaly detection algorithms involves the usage of relatively extended datasets with different faulty conditions. The authors propose deliberately adding real contact force profiles of healthy lines to a mix of synthetic signals, which substantially reproduce the behavior and the variability of foreseen faulty conditions. The results of this work, although preliminary and still to be completed, offer a contribution to the scientific community both in terms of obtained results and adopted methodologies. Full article
(This article belongs to the Special Issue AI-Driven Reliability Analysis and Predictive Maintenance)
Show Figures

Figure 1

27 pages, 1324 KB  
Review
Artificial Intelligence Architectures in Oral Rehabilitation: A Focused Review of Deep Learning Models for Implant Planning, Prosthodontic Design, and Peri-Implant Diagnosis
by Hossam Dawa, Carlos Aroso, Ana Sofia Vinhas, José Manuel Mendes and Arthur Rodriguez Gonzalez Cortes
Appl. Sci. 2026, 16(8), 3739; https://doi.org/10.3390/app16083739 - 10 Apr 2026
Abstract
Deep learning is increasingly integrated into oral rehabilitation workflows, particularly in implant planning, prosthodontic design automation, and peri-implant diagnosis. However, reported performance is heterogeneous and difficult to compare across tasks, modalities, and validation designs. The goal of this study was to critically analyze [...] Read more.
Deep learning is increasingly integrated into oral rehabilitation workflows, particularly in implant planning, prosthodontic design automation, and peri-implant diagnosis. However, reported performance is heterogeneous and difficult to compare across tasks, modalities, and validation designs. The goal of this study was to critically analyze deep learning architecture families applied to oral rehabilitation and to provide task-driven selection guidance supported by an evidence table reporting dataset characteristics, validation strategy, and performance metrics. A focused narrative review was conducted using transparent, database-specific search criteria (final n = 10 included studies), emphasizing implant planning (cone–beam computed tomography [CBCT]-based segmentation), prosthodontic design (intraoral scan [IOS]/mesh inputs), and peri-implant diagnosis (periapical/panoramic radiographs). Evidence certainty for each clinical task was assessed using GRADE-informed ratings (High/Moderate/Low/Very Low). Extracted variables included clinical task, imaging modality, dataset size, architecture, validation strategy (internal vs. internal + external), split level, ground truth protocol, and performance metrics. A structured computational and hardware feasibility analysis was conducted for each architecture family to support real-world deployment planning. Encoder–decoder networks (U-Net/nnU-Net) dominate CBCT segmentation for implant planning, while detection architectures (Faster R-CNN, YOLO) support implant localization and peri-implant assessment on radiographs. Generative models (3D GANs, transformer-based point-to-mesh networks) enable crown design from three-dimensional scans. Hybrid CNN–Transformer architectures show promise for multimodal CBCT–IOS fusion, though direct evidence from the included studies remains limited to a single study. External validation remains uncommon yet essential given the risk of domain shift. In conclusion, architecture selection should be anchored to task geometry (2D vs. 3D), artifact burden, and required clinical output type. Reporting standards should prioritize dataset transparency, validation rigor, multi-center external testing, and uncertainty-aware outputs. Full article
Show Figures

Figure 1

26 pages, 3869 KB  
Article
Conceptual AI-Informed Institutional Learning Analytics: Extending the TAM to Strengthen Inclusive Digital Justice
by Soledad Zabala, José Javier Galán Hernández, Alberto Garcés Jiménez, José Manuel Gómez Pulido, Susana Ester Medina and María Belén Morales Cevallos
Appl. Sci. 2026, 16(8), 3737; https://doi.org/10.3390/app16083737 - 10 Apr 2026
Abstract
This study examines institutional processes in digital justice through a mixed conceptual approach that integrates bibliometric analysis and technology-adoption modeling, incorporating artificial intelligence (AI) as a projected component rather than an implemented system. A corpus of approximately 200 Scopus-indexed documents (2003–2024) was analyzed, [...] Read more.
This study examines institutional processes in digital justice through a mixed conceptual approach that integrates bibliometric analysis and technology-adoption modeling, incorporating artificial intelligence (AI) as a projected component rather than an implemented system. A corpus of approximately 200 Scopus-indexed documents (2003–2024) was analyzed, identifying five dominant thematic clusters: advanced technologies, institutional justice, digital government, judicial information management, and digital criminal justice. The results reveal persistent gaps in the literature, particularly in rural and underserved communities, where connectivity barriers and the limited application of adoption models hinder inclusive digital transformation. As an institutional contribution, the study presents the conceptual design of the digital solution “Travel Permits—Accessible Justice”, developed under a Service-Oriented Architecture (SOA) and projected for future integration with AI-supported components to automate judicial authorizations through biometric validation, electronic signatures, and digital delivery. To evaluate its potential acceptance, the Technology Acceptance Model (TAM) is analytically adapted and extended to the community-based judicial context, framing institutional learning processes as a prospective form of learning analytics focused on user interaction, perceived usefulness, perceived ease of use, and behavioral intention. Taken together, the integration of bibliometric evidence with an extended TAM, along with the projected incorporation of AI-supported institutional learning processes, offers a coherent foundation for future studies on inclusive digital innovation in justice environments. Full article
Show Figures

Figure 1

35 pages, 3452 KB  
Article
LUMINA-Net: Acute Lymphocytic Leukemia Subtype Classification via Interpretable Convolution Neural Network Based on Wavelet and Attention Mechanisms
by Omneya Attallah
Algorithms 2026, 19(4), 298; https://doi.org/10.3390/a19040298 - 10 Apr 2026
Abstract
Acute Lymphoblastic Leukemia (ALL) is a highly prevalent hematological malignancy, especially in children, for whom precise and prompt subtype identification is essential to establish suitable treatment protocols. Current deep learning-based computer-aided diagnosis (CAD) methods for identifying ALL are hindered by numerous drawbacks, such [...] Read more.
Acute Lymphoblastic Leukemia (ALL) is a highly prevalent hematological malignancy, especially in children, for whom precise and prompt subtype identification is essential to establish suitable treatment protocols. Current deep learning-based computer-aided diagnosis (CAD) methods for identifying ALL are hindered by numerous drawbacks, such as a dependence on solely spatial feature depictions, elevated feature dimensions, computationally extensive deep learning architectures, inadequate multi-layer feature utilization, and poor interpretability. This paper introduces LUMINA-Net, a custom, lightweight, and interpretable deep learning CAD for the automated identification and subtype diagnosis of ALL using microscopic blood smear pictures. LUMINA-Net makes four principal contributions: first, it integrates a self-attention module within a lightweight custom Convolution Neural Network (CNN) to effectively capture long-range spatial relationships across clinically pertinent cytological patterns while preserving a compact design. Second, it employs a Discrete Wavelet Transform (DWT)-based wavelet pooling layer that decreases feature dimensions by up to 96.875% while enhancing the obtained depictions with spatial-spectral information. Third, it utilizes a multi-layer feature fusion strategy that combines wavelet-pooled features from two deep layers with a third fully connected layer to create a discriminating multi-scale feature vector. Fourth, it incorporates Gradient-weighted Class Activation Mapping as a dedicated explainability process to furnish clinicians with apparent visual explanations for each classification decision. Withoit the need for image enhancement or segmentation preprocessing, LUMINA-Net outperforms the competing state-of-the-art methods on the same dataset, achieving a peak accuracy of 99.51%, specificity of 99.84%, and sensitivity of 99.51% on the publicly available Kaggle ALL dataset. This demonstrates that LUMINA-Net has the potential to be a dependable, effective, and clinically interpretable CAD tool for ALL diagnosis. Full article
19 pages, 1237 KB  
Article
Reinforcement Learning-Based Inverse Design of Multilayer Particles
by Zhaohui Li, Fang Gao and Delian Liu
Computation 2026, 14(4), 91; https://doi.org/10.3390/computation14040091 - 10 Apr 2026
Viewed by 129
Abstract
Multilayered particles possess exceptional optical properties and hold significant potential for applications in chemical analysis, life sciences, optical sensing, and photonic integration. In practical applications, however, it is often necessary to perform inverse design of multilayered particles with given optical characteristics to meet [...] Read more.
Multilayered particles possess exceptional optical properties and hold significant potential for applications in chemical analysis, life sciences, optical sensing, and photonic integration. In practical applications, however, it is often necessary to perform inverse design of multilayered particles with given optical characteristics to meet specific requirements, a process that remains time-consuming. To overcome this challenge, we propose a reinforcement learning-based method for the automated design of multilayered particles. Leveraging the self-learning capacity of reinforcement learning models in combination with an optical characteristics calculation model, the method iteratively determines particle parameters that fulfill the desired optical responses. This method effectively addresses the many-to-one parameter mapping problem in inverse design, eliminates the need for extensive pre-computations, and provides an innovative approach to the automated design of complex nanostructures. Full article
(This article belongs to the Section Computational Engineering)
Show Figures

Figure 1

36 pages, 7325 KB  
Article
Intelligent Scheduling of Rail-Guided Shuttle Cars via Deep Reinforcement Learning Integrating Dynamic Graph Neural Networks and Transformer Model
by Fang Zhu and Shanshan Peng
Algorithms 2026, 19(4), 289; https://doi.org/10.3390/a19040289 - 8 Apr 2026
Viewed by 119
Abstract
With the rapid development of e-commerce and smart manufacturing, automated warehouse systems have become critical infrastructure for modern logistics. In China’s vast market, the dynamic scheduling of Rail-Guided Vehicles (RGVs) faces significant challenges due to complex task uncertainties, hierarchical supply chain structures, and [...] Read more.
With the rapid development of e-commerce and smart manufacturing, automated warehouse systems have become critical infrastructure for modern logistics. In China’s vast market, the dynamic scheduling of Rail-Guided Vehicles (RGVs) faces significant challenges due to complex task uncertainties, hierarchical supply chain structures, and real-time collision avoidance requirements. Traditional rule-based methods and static optimization models often fail to adapt to such dynamic environments. To address these issues, this paper proposes a novel hybrid deep reinforcement learning framework integrating a Dynamic Graph Neural Network (DGNN) and a Transformer model. The DGNN captures the spatiotemporal dependencies of the warehouse network topology, while the Transformer mechanism enhances long-range feature extraction for task prioritization. Furthermore, we design a centralized Deep Q-network (DQN) framework with parameterized action spaces to coordinate multiple RGVs collaboratively. While the system manages multiple physical vehicles, the learning architecture employs a single-agent global scheduler to avoid the non-stationarity issues inherent in multi-agent reinforcement learning. Experimental results based on real-world data from a large-scale electronics manufacturing warehouse demonstrate that our method reduces average task completion time by 18.5% and improves system throughput by 22.3% compared to state-of-the-art baselines. The proposed approach demonstrates potential for intelligent warehouse management in dynamic industrial scenarios. Full article
Show Figures

Figure 1

19 pages, 1748 KB  
Article
Evaluating Embedding Representations for Multiclass Code Smell Detection: A Comparative Study of CodeBERT and General-Purpose Embeddings
by Marcela Mosquera and Rodolfo Bojorque
Appl. Sci. 2026, 16(8), 3622; https://doi.org/10.3390/app16083622 - 8 Apr 2026
Viewed by 119
Abstract
Code smells are indicators of potential design problems in software systems and are commonly used to guide refactoring activities. Recent advances in representation learning have enabled the use of embedding-based models for analyzing source code, offering an alternative to traditional approaches based on [...] Read more.
Code smells are indicators of potential design problems in software systems and are commonly used to guide refactoring activities. Recent advances in representation learning have enabled the use of embedding-based models for analyzing source code, offering an alternative to traditional approaches based on manually engineered metrics. However, the effectiveness of different embedding representations for multiclass code smell detection remains insufficiently explored. This study presents an empirical comparison of embedding models for the automatic detection of three widely studied code smells: Long Method, God Class, and Feature Envy. Using the Crowdsmelling dataset as an empirical basis, source code fragments were extracted from the original projects and transformed into vector representations using two embedding approaches: a general-purpose embedding model and the code-specialized CodeBERT model. The resulting representations were evaluated using several machine learning classifiers under a stratified group-based validation protocol. The results show that CodeBERT consistently outperforms the general-purpose embeddings across multiple evaluation metrics, including balanced accuracy, macro F1-score, and Matthews correlation coefficient. Dimensionality reduction analyses using PCA and t-SNE further indicate that CodeBERT organizes code smell instances in a more structured latent representation space, which facilitates the separation of smell categories. In particular, CodeBERT achieved a macro F1-score of 0.8619, outperforming general-purpose embeddings (0.7622) and substantially surpassing a classical TF-IDF baseline (0.4555). These findings highlight the value of this study as a controlled multiclass evaluation of embedding representations and demonstrate the practical value of domain-specific representations for improving automated code smell detection and class separability in real-world software engineering scenarios. Full article
Show Figures

Graphical abstract

30 pages, 28721 KB  
Article
Dual-Arm Robotic Textile Unfolding with Depth-Corrected Perception and Fold Resolution
by Tilla Egerhei Båserud, Joakim Johansen, Ajit Jha and Ilya Tyapin
Robotics 2026, 15(4), 78; https://doi.org/10.3390/robotics15040078 - 8 Apr 2026
Viewed by 227
Abstract
Reliable textile recycling requires automated unfolding to expose hidden hard components such as zippers, buttons, and metal fasteners, which otherwise risk damaging machinery and compromising downstream processes. This paper presents the design and implementation of an automated textile unfolding system based on a [...] Read more.
Reliable textile recycling requires automated unfolding to expose hidden hard components such as zippers, buttons, and metal fasteners, which otherwise risk damaging machinery and compromising downstream processes. This paper presents the design and implementation of an automated textile unfolding system based on a dual-arm robotic manipulation framework. The system uses two Interbotix WidowX 250s 6-DoF robotic arms and an Intel RealSense L515 LiDAR camera for visual perception. The unfolding process consists of three stages: initial dual-arm stretching to reduce major folds, refinement through a second stretch targeting the lower region, and a machine-learning stage that employs a YOLOv11 framework trained on depth-encoded textile images, followed by a depth-gradient-based estimator for fold direction. The system applies an extremity-based grasping strategy that selects leftmost and rightmost textile points from a custom error-corrected depth map, enabling robust grasp point selection, and a fold direction estimation method based on depth gradients around the detected fold. The most confident fold region is selected, an unfolding direction is determined using depth ranking, and the textile is manipulated until a flat state is confirmed through depth uniformity. Experiments show that depth correction significantly reduces spatial error in the robot frame, while segmentation and extremity detection achieve high accuracy across varied fold configurations, and the YOLOv11n-based model reaches 98.8% classification accuracy, while fold direction is estimated correctly in 87% of test cases. By enabling robust, largely autonomous textile unfolding, the system demonstrates a practical approach that could support safer and more efficient automated textile recycling workflows. Full article
(This article belongs to the Section Sensors and Control in Robotics)
Show Figures

Figure 1

25 pages, 5864 KB  
Article
Climate-Generalizable Energy Prediction in PCM-Integrated Building Envelopes: A Physics-Informed Machine Learning Framework for Sustainable Envelope Design
by Sadia Jahan Noor, Hyosoo Moon, Raymond C. Tesiero and Seyedali Mirmotalebi
Sustainability 2026, 18(7), 3609; https://doi.org/10.3390/su18073609 - 7 Apr 2026
Viewed by 158
Abstract
Phase change materials (PCMs) offer potential for passive thermal regulation in building envelopes through latent heat storage; however, their effectiveness remains strongly climate-dependent, and configurations optimized for one region often underperform in others. Existing evaluation approaches rely largely on location-specific simulations or surrogate [...] Read more.
Phase change materials (PCMs) offer potential for passive thermal regulation in building envelopes through latent heat storage; however, their effectiveness remains strongly climate-dependent, and configurations optimized for one region often underperform in others. Existing evaluation approaches rely largely on location-specific simulations or surrogate models with limited climate transferability. This study develops a physics-informed, climate-aware machine-learning framework to assess PCM-integrated wall assemblies across diverse climates. A structured dataset of 720 EnergyPlus simulations was generated across nine PCM materials, ten ASHRAE climate zones, two placement configurations, and four thickness levels using automated model generation and batch simulation through Eppy-based workflows. Ensemble-based models (XGBoost, LightGBM, CatBoost, Random Forest) were trained under climate-grouped validation to predict total annual energy consumption, peak cooling demand, and peak heating demand. The models achieved high predictive accuracy for total annual energy use (R2 ≈ 0.98–0.99) and peak cooling demand (R2 ≈ 0.93–0.96), outperforming statistical, climate-only, and PCM-agnostic baselines. In contrast, peak heating demand showed low predictability (R2 ≤ 0.26), indicating limited sensitivity to PCM parameters under the studied configuration. These results demonstrate that climate-aware validation enables defensible cross-climate PCM assessment, supporting energy demand reduction and sustainable envelope design decisions aligned with global building decarbonization goals. Full article
Show Figures

Figure 1

20 pages, 4228 KB  
Article
Design and Application of an Automated Microinjection System Combining Deep Learning Vision Positioning and Neural Network Sliding Mode Motion Control
by Zhihao Deng, Yifan Xu and Shengzheng Kang
Actuators 2026, 15(4), 208; https://doi.org/10.3390/act15040208 - 5 Apr 2026
Viewed by 192
Abstract
Microinjection is one of the most established and effective techniques for introducing foreign substances into cells. However, issues such as cumbersome procedures, low success rates, and poor repeatability in manual cell microinjection have seriously restricted its practical applications in biomedical research and engineering. [...] Read more.
Microinjection is one of the most established and effective techniques for introducing foreign substances into cells. However, issues such as cumbersome procedures, low success rates, and poor repeatability in manual cell microinjection have seriously restricted its practical applications in biomedical research and engineering. Responding to such problems, this paper designs an automated microinjection system that combines deep learning visual positioning and adaptive neural network sliding-mode motion control. The machine vision solution based on the deep learning YOLOv8 target detection algorithm is utilized by the system to provide positional prerequisites for automated microinjection. Then, stable and fast puncture is completed by controlling the end effector (composed of a piezoelectric actuator and a displacement amplification mechanism). Since the piezoelectric actuator has strong nonlinearity, the motion control of the end effector adopts the control strategy combining sliding mode variable structure and adaptive neural networks to meet the requirements of precise displacement output of microinjection. At the same time, a host computer control system is developed to integrate hardware equipment, visual positioning algorithms and motion control algorithms to achieve corresponding automated microinjection tasks. Finally, the effectiveness of the designed automated microinjection system is successfully verified on zebrafish embryos. Full article
Show Figures

Figure 1

23 pages, 1629 KB  
Article
AI-Based Automated Scoring Layer Using Large Language Models and Semantic Analysis
by Anastasia Vangelova and Veska Gancheva
Appl. Sci. 2026, 16(7), 3537; https://doi.org/10.3390/app16073537 - 4 Apr 2026
Viewed by 645
Abstract
Automated scoring of open-ended questions is an important research direction in educational technology and artificial intelligence, as manual grading is time-consuming and often subject to inter-rater variation. This paper proposes an AI-based framework for automated scoring that combines large language models (LLMs), Retrieval-Augmented [...] Read more.
Automated scoring of open-ended questions is an important research direction in educational technology and artificial intelligence, as manual grading is time-consuming and often subject to inter-rater variation. This paper proposes an AI-based framework for automated scoring that combines large language models (LLMs), Retrieval-Augmented Generation (RAG), analytical rubrics, and structured machine-readable output within a Moodle-supported e-learning environment. The framework is designed to support context-grounded and criterion-based evaluation by combining the student response, retrieved instructional context, and rubric-defined scoring criteria within a controlled assessment workflow. The proposed approach aims to improve the consistency, traceability, and practical applicability of automated scoring for open-ended responses. To examine its performance, an experimental study was conducted in a real university setting involving a five-task open-ended examination. AI-generated scores were compared with independent human scores using agreement, reliability, correlation, and error metrics. The results indicate a strong level of agreement between automated and expert scoring within the tested setting, together with relatively low average deviation. These findings suggest that the proposed framework has practical potential for supporting automated assessment in digital learning environments, while also highlighting the importance of careful interpretation within the scope of the experimental design. Full article
(This article belongs to the Special Issue Application of Semantic Web Technologies for E-Learning)
Show Figures

Figure 1

33 pages, 16801 KB  
Article
A GNSS–Vision Integrated Autonomous Navigation System for Trellis Orchard Transportation Robots
by Huaiyang Liu, Haiyang Gu, Yong Wang, Tianjiao Zhong, Tong Tian and Changxing Geng
AI 2026, 7(4), 125; https://doi.org/10.3390/ai7040125 - 1 Apr 2026
Viewed by 331
Abstract
Autonomous navigation is essential for orchard transportation robots to support automated operations and precision orchard management. However, in trellis orchards, dense vegetation and complex canopy structures often degrade the stability of GNSS-based navigation in in-row environments. To address this issue, this study proposes [...] Read more.
Autonomous navigation is essential for orchard transportation robots to support automated operations and precision orchard management. However, in trellis orchards, dense vegetation and complex canopy structures often degrade the stability of GNSS-based navigation in in-row environments. To address this issue, this study proposes a GNSS–vision integrated navigation framework for orchard transportation robots. The performance of GNSS-based navigation in out-of-row environments and vision-based navigation in in-row environments was experimentally evaluated under representative orchard operating conditions. In out-of-row areas, the robot employs GNSS-based path planning and trajectory tracking to achieve reliable navigation in relatively open, lightly occluded environments. During in-row navigation, a deep learning-based real-time object detection approach is used to detect tree trunks and trellis supporting structures. By integrating corner-point selection with temporal RANSAC-based line fitting, a stable orchard row structure is constructed to generate robust navigation references. The visual perception module serves as the front-end sensing component of the navigation system and is designed to be independent of specific object detection architectures, allowing flexible integration with different real-time detection models. Field experiments were conducted under various orchard layouts and growth stages. The average lateral deviation of GNSS-based navigation in out-of-row scenarios ranged from 0.093 to 0.221 m, while the average heading deviation of in-row visual navigation was approximately 5.23° at a robot speed of 0.6 m/s. These results indicate that the proposed perception and navigation methods can maintain stable navigation performance within their respective applicable scenarios in trellis orchard environments. The experimental findings provide a practical and engineering-oriented basis for future research on automatic navigation mode switching and system-level integration of orchard transportation robots. Full article
Show Figures

Figure 1

23 pages, 6673 KB  
Article
ERZA-DETR: A Deep Learning-Based Detection Transformer with Enhanced Relational-Zone Aggregation for WCE Lesion Detection
by Shiren Ye, Haipeng Ma, Zetong Zhang and Liangjing Li
Algorithms 2026, 19(4), 268; https://doi.org/10.3390/a19040268 - 1 Apr 2026
Viewed by 218
Abstract
Wireless capsule endoscopy (WCE) plays a vital role in non-invasive screening of small intestinal lesions. However, the automated detection of lesions remains challenging due to low contrast, uneven illumination, and severe visual variability across images. Existing convolutional detectors rely heavily on manually designed [...] Read more.
Wireless capsule endoscopy (WCE) plays a vital role in non-invasive screening of small intestinal lesions. However, the automated detection of lesions remains challenging due to low contrast, uneven illumination, and severe visual variability across images. Existing convolutional detectors rely heavily on manually designed anchors and post-processing, while end-to-end detection transformers developed for natural images exhibit limited adaptability to the complex texture and spectral characteristics of WCE data. To overcome these limitations, this study proposes a deep learning-based detection transformer with enhanced relational-zone aggregation for WCE lesion detection, termed ERZA-DETR, specifically tailored for WCE lesion detection. The framework integrates three complementary modules: a Dual-Band Adaptive Fourier Spectral module (DBFS) that recalibrates frequency responses to suppress illumination artifacts and highlight lesion boundaries; a Fused Dual-scale Gated Convolutional module (FD-gConv) that selectively fuses multi-scale texture features; and a Graph-Linked Embedding at Semantic Scales module (GLES) that preserves local topological relationships through coordinate-gated aggregation. Experimental evaluations on the SEE-AI small intestine dataset demonstrate that ERZA-DETR achieves a 3.2% improvement in mAP@50 and a 12.4% reduction in parameters compared with RT-DETRv2, achieving a superior balance between detection accuracy, computational efficiency, and clinical applicability. Full article
Show Figures

Figure 1

42 pages, 899 KB  
Review
Bridging the Semantic Gap: A Review of Data Interoperability Challenges and Advanced Methodologies from BIM to LCA
by Yilong Jia, Peng Zhang and Qinjun Liu
Sustainability 2026, 18(7), 3352; https://doi.org/10.3390/su18073352 - 30 Mar 2026
Viewed by 725
Abstract
Building Information Modelling (BIM) offers a pivotal opportunity to automate Life Cycle Assessment (LCA) within the Architecture, Engineering, and Construction (AEC) industry. However, seamless integration is persistently hindered by a semantic gap, a critical misalignment between the object-oriented, geometric definitions of BIM and [...] Read more.
Building Information Modelling (BIM) offers a pivotal opportunity to automate Life Cycle Assessment (LCA) within the Architecture, Engineering, and Construction (AEC) industry. However, seamless integration is persistently hindered by a semantic gap, a critical misalignment between the object-oriented, geometric definitions of BIM and the process-based material data required by Life Cycle Inventory (LCI) databases. This paper presents a comprehensive review of data interoperability challenges and evaluates advanced methodologies designed to bridge this divide, moving beyond simple tool comparison to analyse structural integration barriers. Through a systematic review of 124 primary studies published between 2010 and 2025, this research inductively derives the BIM-LCA Interoperability Triad. This framework analyses causal dependencies across three dimensions, including Semantic and Ontological Structures, Workflow and Temporal Integration, and System Architecture and Interoperability. Furthermore, by establishing a comparative challenge–solution matrix, the analysis reveals a maturity paradox in current methodologies. While semi-automated commercial plugins dominate practice due to accessibility, they frequently function as opaque black boxes with limited transparency. Conversely, advanced approaches utilising Semantic Web technologies and Machine Learning demonstrate superior capability in resolving terminological mismatches but currently face significant barriers regarding infrastructure and expertise. This study contributes a novel theoretical model for understanding integration failures. It concludes that future research must pivot from static schema mapping towards AI-driven semantic healing, dynamic Digital Twins, and explicit system boundary harmonisation to achieve truly automated, context-aware environmental assessments and support whole-life circularity. Full article
Show Figures

Figure 1

Back to TopTop