Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (209)

Search Parameters:
Keywords = fidelity metrics

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
14 pages, 1728 KiB  
Article
Accelerating High-Frequency Circuit Optimization Using Machine Learning-Generated Inverse Maps for Enhanced Space Mapping
by Jorge Davalos-Guzman, Jose L. Chavez-Hurtado and Zabdiel Brito-Brito
Electronics 2025, 14(15), 3097; https://doi.org/10.3390/electronics14153097 (registering DOI) - 3 Aug 2025
Abstract
The optimization of high-frequency circuits remains a computationally intensive task due to the need for repeated high-fidelity electromagnetic (EM) simulations. To address this challenge, we propose a novel integration of machine learning-generated inverse maps within the space mapping (SM) optimization framework to significantly [...] Read more.
The optimization of high-frequency circuits remains a computationally intensive task due to the need for repeated high-fidelity electromagnetic (EM) simulations. To address this challenge, we propose a novel integration of machine learning-generated inverse maps within the space mapping (SM) optimization framework to significantly accelerate circuit optimization while maintaining high accuracy. The proposed approach leverages Bayesian Neural Networks (BNNs) and surrogate modeling techniques to construct an inverse mapping function that directly predicts design parameters from target performance metrics, bypassing iterative forward simulations. The methodology was validated using a low-pass filter optimization scenario, where the inverse surrogate model was trained using electromagnetic simulations from COMSOL Multiphysics 2024 r6.3 and optimized using MATLAB R2024b r24.2 trust region algorithm. Experimental results demonstrate that our approach reduces the number of high-fidelity simulations by over 80% compared to conventional SM techniques while achieving high accuracy with a mean absolute error (MAE) of 0.0262 (0.47%). Additionally, convergence efficiency was significantly improved, with the inverse surrogate model requiring only 31 coarse model simulations, compared to 580 in traditional SM. These findings demonstrate that machine learning-driven inverse surrogate modeling significantly reduces computational overhead, accelerates optimization, and enhances the accuracy of high-frequency circuit design. This approach offers a promising alternative to traditional SM methods, paving the way for more efficient RF and microwave circuit design workflows. Full article
(This article belongs to the Special Issue Advances in Algorithm Optimization and Computational Intelligence)
Show Figures

Figure 1

20 pages, 586 KiB  
Article
Implementing High-Intensity Gait Training in Stroke Rehabilitation: A Real-World Pragmatic Approach
by Jennifer L. Moore, Pia Krøll, Håvard Hansen Berg, Merethe B. Sinnes, Roger Arntsen, Chris E. Henderson, T. George Hornby, Stein Arne Rimehaug, Ingvild Lilleheie and Anders Orpana
J. Clin. Med. 2025, 14(15), 5409; https://doi.org/10.3390/jcm14155409 (registering DOI) - 31 Jul 2025
Viewed by 191
Abstract
Background: High-intensity gait training (HIT) is an evidence-based intervention recommended for stroke rehabilitation; however, its implementation in routine practice is inconsistent. This study examined the real-world implementation of HIT in an inpatient rehabilitation setting in Norway, focusing on fidelity, barriers, and knowledge [...] Read more.
Background: High-intensity gait training (HIT) is an evidence-based intervention recommended for stroke rehabilitation; however, its implementation in routine practice is inconsistent. This study examined the real-world implementation of HIT in an inpatient rehabilitation setting in Norway, focusing on fidelity, barriers, and knowledge translation (KT) strategies. Methods: Using the Knowledge-to-Action (KTA) framework, HIT was implemented in three phases: pre-implementation, implementation, and competency. Fidelity metrics and coverage were assessed in 99 participants post-stroke. Barriers and facilitators were documented and categorized using the Consolidated Framework for Implementation Research. Results: HIT was delivered with improved fidelity during the implementation and competency phases, reflected by increased stepping and heart rate metrics. A coverage rate of 52% was achieved. Barriers evolved over time, beginning with logistical and knowledge challenges and shifting toward decision-making complexity. The KT interventions, developed collaboratively by clinicians and external facilitators, supported implementation. Conclusions: Structured pre-implementation planning, clinician engagement, and external facilitation enabled high-fidelity HIT implementation in a real-world setting. Pragmatic, context-sensitive strategies were critical to overcoming evolving barriers. Future research should examine scalable, adaptive KT strategies that balance theoretical guidance with clinical feasibility to sustain evidence-based practice in rehabilitation. Full article
Show Figures

Figure 1

19 pages, 2913 KiB  
Article
Radiation Mapping: A Gaussian Multi-Kernel Weighting Method for Source Investigation in Disaster Scenarios
by Songbai Zhang, Qi Liu, Jie Chen, Yujin Cao and Guoqing Wang
Sensors 2025, 25(15), 4736; https://doi.org/10.3390/s25154736 (registering DOI) - 31 Jul 2025
Viewed by 98
Abstract
Structural collapses caused by accidents or disasters could create unexpected radiation shielding, resulting in sharp gradients within the radiation field. Traditional radiation mapping methods often fail to accurately capture these complex variations, making the rapid and precise localization of radiation sources a significant [...] Read more.
Structural collapses caused by accidents or disasters could create unexpected radiation shielding, resulting in sharp gradients within the radiation field. Traditional radiation mapping methods often fail to accurately capture these complex variations, making the rapid and precise localization of radiation sources a significant challenge in emergency response scenarios. To address this issue, based on standard Gaussian process regression (GPR) models that primarily utilize a single Gaussian kernel to reflect the inverse-square law in free space, a novel multi-kernel Gaussian process regression (MK-GPR) model is proposed for high-fidelity radiation mapping in environments with physical obstructions. MK-GPR integrates two additional kernel functions with adaptive weighting: one models the attenuation characteristics of intervening materials, and the other captures the energy-dependent penetration behavior of radiation. To validate the model, gamma-ray distributions in complex, shielded environments were simulated using GEometry ANd Tracking 4 (Geant4). Compared with conventional methods, including linear interpolation, nearest-neighbor interpolation, and standard GPR, MK-GPR demonstrated substantial improvements in key evaluation metrics, such as MSE, RMSE, and MAE. Notably, the coefficient of determination (R2) increased to 0.937. For practical deployment, the optimized MK-GPR model was deployed to an RK-3588 edge computing platform and integrated into a mobile robot equipped with a NaI(Tl) detector. Field experiments confirmed the system’s ability to accurately map radiation fields and localize gamma sources. When combined with SLAM, the system achieved localization errors of 10 cm for single sources and 15 cm for dual sources. These results highlight the potential of the proposed approach as an effective and deployable solution for radiation source investigation in post-disaster environments. Full article
(This article belongs to the Section Navigation and Positioning)
Show Figures

Figure 1

17 pages, 920 KiB  
Article
Enhancing Early GI Disease Detection with Spectral Visualization and Deep Learning
by Tsung-Jung Tsai, Kun-Hua Lee, Chu-Kuang Chou, Riya Karmakar, Arvind Mukundan, Tsung-Hsien Chen, Devansh Gupta, Gargi Ghosh, Tao-Yuan Liu and Hsiang-Chen Wang
Bioengineering 2025, 12(8), 828; https://doi.org/10.3390/bioengineering12080828 - 30 Jul 2025
Viewed by 272
Abstract
Timely and accurate diagnosis of gastrointestinal diseases (GIDs) remains a critical bottleneck in clinical endoscopy, particularly due to the limited contrast and sensitivity of conventional white light imaging (WLI) in detecting early-stage mucosal abnormalities. To overcome this, this research presents Spectrum Aided Vision [...] Read more.
Timely and accurate diagnosis of gastrointestinal diseases (GIDs) remains a critical bottleneck in clinical endoscopy, particularly due to the limited contrast and sensitivity of conventional white light imaging (WLI) in detecting early-stage mucosal abnormalities. To overcome this, this research presents Spectrum Aided Vision Enhancer (SAVE), an innovative, software-driven framework that transforms standard WLI into high-fidelity hyperspectral imaging (HSI) and simulated narrow-band imaging (NBI) without any hardware modification. SAVE leverages advanced spectral reconstruction techniques, including Macbeth Color Checker-based calibration, principal component analysis (PCA), and multivariate polynomial regression, achieving a root mean square error (RMSE) of 0.056 and structural similarity index (SSIM) exceeding 90%. Trained and validated on the Kvasir v2 dataset (n = 6490) using deep learning models like ResNet-50, ResNet-101, EfficientNet-B2, both EfficientNet-B5 and EfficientNetV2-B0 were used to assess diagnostic performance across six key GI conditions. Results demonstrated that SAVE enhanced imagery and consistently outperformed raw WLI across precision, recall, and F1-score metrics, with EfficientNet-B2 and EfficientNetV2-B0 achieving the highest classification accuracy. Notably, this performance gain was achieved without the need for specialized imaging hardware. These findings highlight SAVE as a transformative solution for augmenting GI diagnostics, with the potential to significantly improve early detection, streamline clinical workflows, and broaden access to advanced imaging especially in resource constrained settings. Full article
Show Figures

Figure 1

34 pages, 1156 KiB  
Systematic Review
Mathematical Modelling and Optimization Methods in Geomechanically Informed Blast Design: A Systematic Literature Review
by Fabian Leon, Luis Rojas, Alvaro Peña, Paola Moraga, Pedro Robles, Blanca Gana and Jose García
Mathematics 2025, 13(15), 2456; https://doi.org/10.3390/math13152456 - 30 Jul 2025
Viewed by 210
Abstract
Background: Rock–blast design is a canonical inverse problem that joins elastodynamic partial differential equations (PDEs), fracture mechanics, and stochastic heterogeneity. Objective: Guided by the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) protocol, a systematic review of mathematical methods for geomechanically informed [...] Read more.
Background: Rock–blast design is a canonical inverse problem that joins elastodynamic partial differential equations (PDEs), fracture mechanics, and stochastic heterogeneity. Objective: Guided by the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) protocol, a systematic review of mathematical methods for geomechanically informed blast modelling and optimisation is provided. Methods: A Scopus–Web of Science search (2000–2025) retrieved 2415 records; semantic filtering and expert screening reduced the corpus to 97 studies. Topic modelling with Bidirectional Encoder Representations from Transformers Topic (BERTOPIC) and bibliometrics organised them into (i) finite-element and finite–discrete element simulations, including arbitrary Lagrangian–Eulerian (ALE) formulations; (ii) geomechanics-enhanced empirical laws; and (iii) machine-learning surrogates and multi-objective optimisers. Results: High-fidelity simulations delimit blast-induced damage with ≤0.2 m mean absolute error; extensions of the Kuznetsov–Ram equation cut median-size mean absolute percentage error (MAPE) from 27% to 15%; Gaussian-process and ensemble learners reach a coefficient of determination (R2>0.95) while providing closed-form uncertainty; Pareto optimisers lower peak particle velocity (PPV) by up to 48% without productivity loss. Synthesis: Four themes emerge—surrogate-assisted PDE-constrained optimisation, probabilistic domain adaptation, Bayesian model fusion for digital-twin updating, and entropy-based energy metrics. Conclusions: Persisting challenges in scalable uncertainty quantification, coupled discrete–continuous fracture solvers, and rigorous fusion of physics-informed and data-driven models position blast design as a fertile test bed for advances in applied mathematics, numerical analysis, and machine-learning theory. Full article
Show Figures

Figure 1

17 pages, 1603 KiB  
Perspective
A Perspective on Quality Evaluation for AI-Generated Videos
by Zhichao Zhang, Wei Sun and Guangtao Zhai
Sensors 2025, 25(15), 4668; https://doi.org/10.3390/s25154668 - 28 Jul 2025
Viewed by 238
Abstract
Recent breakthroughs in AI-generated content (AIGC) have transformed video creation, empowering systems to translate text, images, or audio into visually compelling stories. Yet reliable evaluation of these machine-crafted videos remains elusive because quality is governed not only by spatial fidelity within individual frames [...] Read more.
Recent breakthroughs in AI-generated content (AIGC) have transformed video creation, empowering systems to translate text, images, or audio into visually compelling stories. Yet reliable evaluation of these machine-crafted videos remains elusive because quality is governed not only by spatial fidelity within individual frames but also by temporal coherence across frames and precise semantic alignment with the intended message. The foundational role of sensor technologies is critical, as they determine the physical plausibility of AIGC outputs. In this perspective, we argue that multimodal large language models (MLLMs) are poised to become the cornerstone of next-generation video quality assessment (VQA). By jointly encoding cues from multiple modalities such as vision, language, sound, and even depth, the MLLM can leverage its powerful language understanding capabilities to assess the quality of scene composition, motion dynamics, and narrative consistency, overcoming the fragmentation of hand-engineered metrics and the poor generalization ability of CNN-based methods. Furthermore, we provide a comprehensive analysis of current methodologies for assessing AIGC video quality, including the evolution of generation models, dataset design, quality dimensions, and evaluation frameworks. We argue that advances in sensor fusion enable MLLMs to combine low-level physical constraints with high-level semantic interpretations, further enhancing the accuracy of visual quality assessment. Full article
(This article belongs to the Special Issue Perspectives in Intelligent Sensors and Sensing Systems)
Show Figures

Figure 1

20 pages, 3386 KiB  
Article
Design of Realistic and Artistically Expressive 3D Facial Models for Film AIGC: A Cross-Modal Framework Integrating Audience Perception Evaluation
by Yihuan Tian, Xinyang Li, Zuling Cheng, Yang Huang and Tao Yu
Sensors 2025, 25(15), 4646; https://doi.org/10.3390/s25154646 - 26 Jul 2025
Viewed by 369
Abstract
The rise of virtual production has created an urgent need for both efficient and high-fidelity 3D face generation schemes for cinema and immersive media, but existing methods are often limited by lighting–geometry coupling, multi-view dependency, and insufficient artistic quality. To address this, this [...] Read more.
The rise of virtual production has created an urgent need for both efficient and high-fidelity 3D face generation schemes for cinema and immersive media, but existing methods are often limited by lighting–geometry coupling, multi-view dependency, and insufficient artistic quality. To address this, this study proposes a cross-modal 3D face generation framework based on single-view semantic masks. It utilizes Swin Transformer for multi-level feature extraction and combines with NeRF for illumination decoupled rendering. We utilize physical rendering equations to explicitly separate surface reflectance from ambient lighting to achieve robust adaptation to complex lighting variations. In addition, to address geometric errors across illumination scenes, we construct geometric a priori constraint networks by mapping 2D facial features to 3D parameter space as regular terms with the help of semantic masks. On the CelebAMask-HQ dataset, this method achieves a leading score of SSIM = 0.892 (37.6% improvement from baseline) with FID = 40.6. The generated faces excel in symmetry and detail fidelity with realism and aesthetic scores of 8/10 and 7/10, respectively, in a perceptual evaluation with 1000 viewers. By combining physical-level illumination decoupling with semantic geometry a priori, this paper establishes a quantifiable feedback mechanism between objective metrics and human aesthetic evaluation, providing a new paradigm for aesthetic quality assessment of AI-generated content. Full article
(This article belongs to the Special Issue Convolutional Neural Network Technology for 3D Imaging and Sensing)
Show Figures

Figure 1

47 pages, 18189 KiB  
Article
Synthetic Scientific Image Generation with VAE, GAN, and Diffusion Model Architectures
by Zineb Sordo, Eric Chagnon, Zixi Hu, Jeffrey J. Donatelli, Peter Andeer, Peter S. Nico, Trent Northen and Daniela Ushizima
J. Imaging 2025, 11(8), 252; https://doi.org/10.3390/jimaging11080252 - 26 Jul 2025
Viewed by 469
Abstract
Generative AI (genAI) has emerged as a powerful tool for synthesizing diverse and complex image data, offering new possibilities for scientific imaging applications. This review presents a comprehensive comparative analysis of leading generative architectures, ranging from Variational Autoencoders (VAEs) to Generative Adversarial Networks [...] Read more.
Generative AI (genAI) has emerged as a powerful tool for synthesizing diverse and complex image data, offering new possibilities for scientific imaging applications. This review presents a comprehensive comparative analysis of leading generative architectures, ranging from Variational Autoencoders (VAEs) to Generative Adversarial Networks (GANs) on through to Diffusion Models, in the context of scientific image synthesis. We examine each model’s foundational principles, recent architectural advancements, and practical trade-offs. Our evaluation, conducted on domain-specific datasets including microCT scans of rocks and composite fibers, as well as high-resolution images of plant roots, integrates both quantitative metrics (SSIM, LPIPS, FID, CLIPScore) and expert-driven qualitative assessments. Results show that GANs, particularly StyleGAN, produce images with high perceptual quality and structural coherence. Diffusion-based models for inpainting and image variation, such as DALL-E 2, delivered high realism and semantic alignment but generally struggled in balancing visual fidelity with scientific accuracy. Importantly, our findings reveal limitations of standard quantitative metrics in capturing scientific relevance, underscoring the need for domain-expert validation. We conclude by discussing key challenges such as model interpretability, computational cost, and verification protocols, and discuss future directions where generative AI can drive innovation in data augmentation, simulation, and hypothesis generation in scientific research. Full article
(This article belongs to the Special Issue Celebrating the 10th Anniversary of the Journal of Imaging)
Show Figures

Graphical abstract

27 pages, 30210 KiB  
Article
Research on a Rapid Three-Dimensional Compressor Flow Field Prediction Method Integrating U-Net and Physics-Informed Neural Networks
by Chen Wang and Hongbing Ma
Mathematics 2025, 13(15), 2396; https://doi.org/10.3390/math13152396 - 25 Jul 2025
Viewed by 130
Abstract
This paper presents a neural network model, PINN-AeroFlow-U, for reconstructing full-field aerodynamic quantities around three-dimensional compressor blades, including regions near the wall. This model is based on structured CFD training data and physics-informed loss functions and is proposed for direct 3D compressor flow [...] Read more.
This paper presents a neural network model, PINN-AeroFlow-U, for reconstructing full-field aerodynamic quantities around three-dimensional compressor blades, including regions near the wall. This model is based on structured CFD training data and physics-informed loss functions and is proposed for direct 3D compressor flow prediction. It maps flow data from the physical domain to a uniform computational domain and employs a U-Net-based neural network capable of capturing the sharp local transitions induced by fluid acceleration near the blade leading edge, as well as learning flow features associated with internal boundaries (e.g., the wall boundary). The inputs to PINN-AeroFlow-U are the flow-field coordinate data from high-fidelity multi-geometry blade solutions, the 3D blade geometry, and the first-order metric coefficients obtained via mesh transformation. Its outputs include the pressure field, temperature field, and velocity vector field within the blade passage. To enhance physical interpretability, the network’s loss function incorporates both the Euler equations and gradient constraints. PINN-AeroFlow-U achieves prediction errors of 1.063% for the pressure field and 2.02% for the velocity field, demonstrating high accuracy. Full article
Show Figures

Figure 1

35 pages, 4256 KiB  
Article
Automated Segmentation and Morphometric Analysis of Thioflavin-S-Stained Amyloid Deposits in Alzheimer’s Disease Brains and Age-Matched Controls Using Weakly Supervised Deep Learning
by Gábor Barczánfalvi, Tibor Nyári, József Tolnai, László Tiszlavicz, Balázs Gulyás and Karoly Gulya
Int. J. Mol. Sci. 2025, 26(15), 7134; https://doi.org/10.3390/ijms26157134 - 24 Jul 2025
Viewed by 377
Abstract
Alzheimer’s disease (AD) involves the accumulation of amyloid-β (Aβ) plaques, whose quantification plays a central role in understanding disease progression. Automated segmentation of Aβ deposits in histopathological micrographs enables large-scale analyses but is hindered by the high cost of detailed pixel-level annotations. Weakly [...] Read more.
Alzheimer’s disease (AD) involves the accumulation of amyloid-β (Aβ) plaques, whose quantification plays a central role in understanding disease progression. Automated segmentation of Aβ deposits in histopathological micrographs enables large-scale analyses but is hindered by the high cost of detailed pixel-level annotations. Weakly supervised learning offers a promising alternative by leveraging coarse or indirect labels to reduce the annotation burden. We evaluated a weakly supervised approach to segment and analyze thioflavin-S-positive parenchymal amyloid pathology in AD and age-matched brains. Our pipeline integrates three key components, each designed to operate under weak supervision. First, robust preprocessing (including retrospective multi-image illumination correction and gradient-based background estimation) was applied to enhance image fidelity and support training, as models rely more on image features. Second, class activation maps (CAMs), generated by a compact deep classifier SqueezeNet, were used to identify, and coarsely localize amyloid-rich parenchymal regions from patch-wise image labels, serving as spatial priors for subsequent refinement without requiring dense pixel-level annotations. Third, a patch-based convolutional neural network, U-Net, was trained on synthetic data generated from micrographs based on CAM-derived pseudo-labels via an extensive object-level augmentation strategy, enabling refined whole-image semantic segmentation and generalization across diverse spatial configurations. To ensure robustness and unbiased evaluation, we assessed the segmentation performance of the entire framework using patient-wise group k-fold cross-validation, explicitly modeling generalization across unseen individuals, critical in clinical scenarios. Despite relying on weak labels, the integrated pipeline achieved strong segmentation performance with an average Dice similarity coefficient (≈0.763) and Jaccard index (≈0.639), widely accepted metrics for assessing segmentation quality in medical image analysis. The resulting segmentations were also visually coherent, demonstrating that weakly supervised segmentation is a viable alternative in histopathology, where acquiring dense annotations is prohibitively labor-intensive and time-consuming. Subsequent morphometric analyses on automatically segmented Aβ deposits revealed size-, structural complexity-, and global geometry-related differences across brain regions and cognitive status. These findings confirm that deposit architecture exhibits region-specific patterns and reflects underlying neurodegenerative processes, thereby highlighting the biological relevance and practical applicability of the proposed image-processing pipeline for morphometric analysis. Full article
Show Figures

Figure 1

25 pages, 6911 KiB  
Article
Image Inpainting Algorithm Based on Structure-Guided Generative Adversarial Network
by Li Zhao, Tongyang Zhu, Chuang Wang, Feng Tian and Hongge Yao
Mathematics 2025, 13(15), 2370; https://doi.org/10.3390/math13152370 - 24 Jul 2025
Viewed by 287
Abstract
To address the challenges of image inpainting in scenarios with extensive or irregular missing regions—particularly detail oversmoothing, structural ambiguity, and textural incoherence—this paper proposes an Image Structure-Guided (ISG) framework that hierarchically integrates structural priors with semantic-aware texture synthesis. The proposed methodology advances a [...] Read more.
To address the challenges of image inpainting in scenarios with extensive or irregular missing regions—particularly detail oversmoothing, structural ambiguity, and textural incoherence—this paper proposes an Image Structure-Guided (ISG) framework that hierarchically integrates structural priors with semantic-aware texture synthesis. The proposed methodology advances a two-stage restoration paradigm: (1) Structural Prior Extraction, where adaptive edge detection algorithms identify residual contours in corrupted regions, and a transformer-enhanced network reconstructs globally consistent structural maps through contextual feature propagation; (2) Structure-Constrained Texture Synthesis, wherein a multi-scale generator with hybrid dilated convolutions and channel attention mechanisms iteratively refines high-fidelity textures under explicit structural guidance. The framework introduces three innovations: (1) a hierarchical feature fusion architecture that synergizes multi-scale receptive fields with spatial-channel attention to preserve long-range dependencies and local details simultaneously; (2) spectral-normalized Markovian discriminator with gradient-penalty regularization, enabling adversarial training stability while enforcing patch-level structural consistency; and (3) dual-branch loss formulation combining perceptual similarity metrics with edge-aware constraints to align synthesized content with both semantic coherence and geometric fidelity. Our experiments on the two benchmark datasets (Places2 and CelebA) have demonstrated that our framework achieves more unified textures and structures, bringing the restored images closer to their original semantic content. Full article
Show Figures

Figure 1

25 pages, 2129 KiB  
Article
Zero-Shot 3D Reconstruction of Industrial Assets: A Completion-to-Reconstruction Framework Trained on Synthetic Data
by Yongjie Xu, Haihua Zhu and Barmak Honarvar Shakibaei Asli
Electronics 2025, 14(15), 2949; https://doi.org/10.3390/electronics14152949 - 24 Jul 2025
Viewed by 221
Abstract
Creating high-fidelity digital twins (DTs) for Industry 4.0 applications, it is fundamentally reliant on the accurate 3D modeling of physical assets, a task complicated by the inherent imperfections of real-world point cloud data. This paper addresses the challenge of reconstructing accurate, watertight, and [...] Read more.
Creating high-fidelity digital twins (DTs) for Industry 4.0 applications, it is fundamentally reliant on the accurate 3D modeling of physical assets, a task complicated by the inherent imperfections of real-world point cloud data. This paper addresses the challenge of reconstructing accurate, watertight, and topologically sound 3D meshes from sparse, noisy, and incomplete point clouds acquired in complex industrial environments. We introduce a robust two-stage completion-to-reconstruction framework, C2R3D-Net, that systematically tackles this problem. The methodology first employs a pretrained, self-supervised point cloud completion network to infer a dense and structurally coherent geometric representation from degraded inputs. Subsequently, a novel adaptive surface reconstruction network generates the final high-fidelity mesh. This network features a hybrid encoder (FKAConv-LSA-DC), which integrates fixed-kernel and deformable convolutions with local self-attention to robustly capture both coarse geometry and fine details, and a boundary-aware multi-head interpolation decoder, which explicitly models sharp edges and thin structures to preserve geometric fidelity. Comprehensive experiments on the large-scale synthetic ShapeNet benchmark demonstrate state-of-the-art performance across all standard metrics. Crucially, we validate the framework’s strong zero-shot generalization capability by deploying the model—trained exclusively on synthetic data—to reconstruct complex assets from a custom-collected industrial dataset without any additional fine-tuning. The results confirm the method’s suitability as a robust and scalable approach for 3D asset modeling, a critical enabling step for creating high-fidelity DTs in demanding, unseen industrial settings. Full article
Show Figures

Figure 1

18 pages, 2545 KiB  
Article
Reliable Indoor Fire Detection Using Attention-Based 3D CNNs: A Fire Safety Engineering Perspective
by Mostafa M. E. H. Ali and Maryam Ghodrat
Fire 2025, 8(7), 285; https://doi.org/10.3390/fire8070285 - 21 Jul 2025
Viewed by 489
Abstract
Despite recent advances in deep learning for fire detection, much of the current research prioritizes model-centric metrics over dataset fidelity, particularly from a fire safety engineering perspective. Commonly used datasets are often dominated by fully developed flames, mislabel smoke-only frames as non-fire, or [...] Read more.
Despite recent advances in deep learning for fire detection, much of the current research prioritizes model-centric metrics over dataset fidelity, particularly from a fire safety engineering perspective. Commonly used datasets are often dominated by fully developed flames, mislabel smoke-only frames as non-fire, or lack intra-video diversity due to redundant frames from limited sources. Some works treat smoke detection alone as early-stage detection, even though many fires (e.g., electrical or chemical) begin with visible flames and no smoke. Additionally, attempts to improve model applicability through mixed-context datasets—combining indoor, outdoor, and wildland scenes—often overlook the unique false alarm sources and detection challenges specific to each environment. To address these limitations, we curated a new video dataset comprising 1108 annotated fire and non-fire clips captured via indoor surveillance cameras. Unlike existing datasets, ours emphasizes early-stage fire dynamics (pre-flashover) and includes varied fire sources (e.g., sofa, cupboard, and attic fires), realistic false alarm triggers (e.g., flame-colored objects, artificial lighting), and a wide range of spatial layouts and illumination conditions. This collection enables robust training and benchmarking for early indoor fire detection. Using this dataset, we developed a spatiotemporal fire detection model based on the mixed convolutions ResNets (MC3_18) architecture, augmented with Convolutional Block Attention Modules (CBAM). The proposed model achieved 86.11% accuracy, 88.76% precision, and 84.04% recall, along with low false positive (11.63%) and false negative (15.96%) rates. Compared to its CBAM-free baseline, the model exhibits notable improvements in F1-score and interpretability, as confirmed by Grad-CAM++ visualizations highlighting attention to semantically meaningful fire features. These results demonstrate that effective early fire detection is inseparable from high-quality, context-specific datasets. Our work introduces a scalable, safety-driven approach that advances the development of reliable, interpretable, and deployment-ready fire detection systems for residential environments. Full article
Show Figures

Figure 1

39 pages, 2628 KiB  
Article
A Decentralized Multi-Venue Real-Time Video Broadcasting System Integrating Chain Topology and Intelligent Self-Healing Mechanisms
by Tianpei Guo, Ziwen Song, Haotian Xin and Guoyang Liu
Appl. Sci. 2025, 15(14), 8043; https://doi.org/10.3390/app15148043 - 19 Jul 2025
Viewed by 450
Abstract
The rapid growth in large-scale distributed video conferencing, remote education, and real-time broadcasting poses significant challenges to traditional centralized streaming systems, particularly regarding scalability, cost, and reliability under high concurrency. Centralized approaches often encounter bottlenecks, increased bandwidth expenses, and diminished fault tolerance. This [...] Read more.
The rapid growth in large-scale distributed video conferencing, remote education, and real-time broadcasting poses significant challenges to traditional centralized streaming systems, particularly regarding scalability, cost, and reliability under high concurrency. Centralized approaches often encounter bottlenecks, increased bandwidth expenses, and diminished fault tolerance. This paper proposes a novel decentralized real-time broadcasting system employing a peer-to-peer (P2P) chain topology based on IPv6 networking and the Secure Reliable Transport (SRT) protocol. By exploiting the global addressing capability of IPv6, our solution simplifies direct node interconnections, effectively eliminating complexities associated with Network Address Translation (NAT). Furthermore, we introduce an innovative chain-relay transmission method combined with distributed node management strategies, substantially reducing reliance on central servers and minimizing deployment complexity. Leveraging SRT’s low-latency UDP transmission, packet retransmission, congestion control, and AES-128/256 encryption, the proposed system ensures robust security and high video stream quality across wide-area networks. Additionally, a WebSocket-based real-time fault detection algorithm coupled with a rapid fallback self-healing mechanism is developed, enabling millisecond-level fault detection and swift restoration of disrupted links. Extensive performance evaluations using Video Multi-Resolution Fidelity (VMRF) metrics across geographically diverse and heterogeneous environments confirm significant performance gains. Specifically, our approach achieves substantial improvements in latency, video quality stability, and fault tolerance over existing P2P methods, along with over tenfold enhancements in frame rates compared with conventional RTMP-based solutions, thereby demonstrating its efficacy, scalability, and cost-effectiveness for real-time video streaming applications. Full article
Show Figures

Figure 1

25 pages, 6248 KiB  
Article
Low-Cost Strain-Gauge Force-Sensing Sidestick for 6-DoF Flight Simulation: Design and Human-in-the-Loop Evaluation
by Patrik Rožić, Milan Vrdoljak, Karolina Krajček Nikolić and Jurica Ivošević
Sensors 2025, 25(14), 4476; https://doi.org/10.3390/s25144476 - 18 Jul 2025
Viewed by 333
Abstract
Modern fly-by-wire (FBW) aircraft demand high-fidelity simulation systems for research and training, yet existing force-sensing solutions are often prohibitively expensive. This study presents the design, development, and validation of a low-cost, reconfigurable force-sensing sidestick. The system utilizes four strain-gauge load cells to capture [...] Read more.
Modern fly-by-wire (FBW) aircraft demand high-fidelity simulation systems for research and training, yet existing force-sensing solutions are often prohibitively expensive. This study presents the design, development, and validation of a low-cost, reconfigurable force-sensing sidestick. The system utilizes four strain-gauge load cells to capture pure pilot force inputs, integrated with a 6-DoF non-linear flight model. To evaluate its performance, a pitch-angle tracking task was conducted with 16 participants (pilots and non-pilots). Objective metrics revealed that the control strategy was a primary determinant of performance. Participants employing a proactive feedforward control strategy exhibited roughly an order of magnitude lower tracking-error variance than those relying on reactive corrections. Subjective assessments using the Cooper-Harper scale and NASA-TLX corroborated the objective data, confirming the sidestick’s ability to differentiate control techniques. This work demonstrates an open-source platform that makes high-fidelity FBW simulation accessible for academic research, pilot training, and human factors analysis at a fraction of the cost of commercial systems. Full article
Show Figures

Figure 1

Back to TopTop