Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (1,079)

Search Parameters:
Keywords = lightweight deep learning model

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
28 pages, 3531 KiB  
Review
Review of Acoustic Emission Detection Technology for Valve Internal Leakage: Mechanisms, Methods, Challenges, and Application Prospects
by Dongjie Zheng, Xing Wang, Lingling Yang, Yunqi Li, Hui Xia, Haochuan Zhang and Xiaomei Xiang
Sensors 2025, 25(14), 4487; https://doi.org/10.3390/s25144487 - 18 Jul 2025
Abstract
Internal leakage within the valve body constitutes a severe potential safety hazard in industrial fluid control systems, attributable to its high concealment and the resultant difficulty in detection via conventional methodologies. Acoustic emission (AE) technology, functioning as an efficient non-destructive testing approach, is [...] Read more.
Internal leakage within the valve body constitutes a severe potential safety hazard in industrial fluid control systems, attributable to its high concealment and the resultant difficulty in detection via conventional methodologies. Acoustic emission (AE) technology, functioning as an efficient non-destructive testing approach, is capable of capturing the transient stress waves induced by leakage, thereby furnishing an effective means for the real-time monitoring and quantitative assessment of internal leakage within the valve body. This paper conducts a systematic review of the theoretical foundations, signal-processing methodologies, and the latest research advancements related to the technology for detecting internal leakage in the valve body based on acoustic emission. Firstly, grounded in Lechlier’s acoustic analogy theory, the generation mechanism of acoustic emission signals arising from valve body leakage is elucidated. Secondly, a detailed analysis is conducted on diverse signal processing techniques and their corresponding optimization strategies, encompassing parameter analysis, time–frequency analysis, nonlinear dynamics methods, and intelligent algorithms. Moreover, this paper recapitulates the current challenges encountered by this technology and delineates future research orientations, such as the fusion of multi-modal sensors, the deployment of lightweight deep learning models, and integration with the Internet of Things. This study provides a systematic reference for the engineering application and theoretical development of the acoustic emission-based technology for detecting internal leakage in valves. Full article
(This article belongs to the Topic Advances in Non-Destructive Testing Methods, 3rd Edition)
Show Figures

Figure 1

20 pages, 1811 KiB  
Article
Sentinel-2 Satellite-Derived Bathymetry with Data-Efficient Domain Adaptation
by Christos G. E. Anagnostopoulos, Vassilios Papaioannou, Konstantinos Vlachos, Anastasia Moumtzidou, Ilias Gialampoukidis, Stefanos Vrochidis and Ioannis Kompatsiaris
J. Mar. Sci. Eng. 2025, 13(7), 1374; https://doi.org/10.3390/jmse13071374 - 18 Jul 2025
Abstract
Satellite-derived bathymetry (SDB) enables the efficient mapping of shallow waters such as coastal zones but typically requires extensive local ground truth data to achieve high accuracy. This study evaluates the effectiveness of transfer learning in reducing this requirement while keeping estimation accuracy at [...] Read more.
Satellite-derived bathymetry (SDB) enables the efficient mapping of shallow waters such as coastal zones but typically requires extensive local ground truth data to achieve high accuracy. This study evaluates the effectiveness of transfer learning in reducing this requirement while keeping estimation accuracy at acceptable levels by adapting a deep learning model pretrained on data from Puck Lagoon (Poland) to a new coastal site in Agia Napa (Cyprus). Leveraging the open MagicBathyNet benchmark dataset and a lightweight U-Net architecture, three scenarios were studied and compared: direct inference to Cyprus, site-specific training in Cyprus, and fine-tuning from Poland to Cyprus with incrementally larger subsets of training data. Results demonstrate that fine-tuning with 15 samples reduces RMSE by over 50% relative to the direct inference baseline. In addition, the domain adaptation approach using 15 samples shows comparable performance to the site-specific model trained on all available data in Cyprus. Depth-stratified error analysis and paired statistical tests confirm that around 15 samples represent a practical lower bound for stable SDB, according to the MagicBathyNet benchmark. The findings of this work provide quantitative evidence on the effectiveness of deploying data-efficient SDB pipelines in settings of limited in situ surveys, as well as a practical lower bound for clear and shallow coastal waters. Full article
(This article belongs to the Section Physical Oceanography)
36 pages, 7426 KiB  
Article
PowerLine-MTYOLO: A Multitask YOLO Model for Simultaneous Cable Segmentation and Broken Strand Detection
by Badr-Eddine Benelmostafa and Hicham Medromi
Drones 2025, 9(7), 505; https://doi.org/10.3390/drones9070505 - 18 Jul 2025
Abstract
Power transmission infrastructure requires continuous inspection to prevent failures and ensure grid stability. UAV-based systems, enhanced with deep learning, have emerged as an efficient alternative to traditional, labor-intensive inspection methods. However, most existing approaches rely on separate models for cable segmentation and anomaly [...] Read more.
Power transmission infrastructure requires continuous inspection to prevent failures and ensure grid stability. UAV-based systems, enhanced with deep learning, have emerged as an efficient alternative to traditional, labor-intensive inspection methods. However, most existing approaches rely on separate models for cable segmentation and anomaly detection, leading to increased computational overhead and reduced reliability in real-time applications. To address these limitations, we propose PowerLine-MTYOLO, a lightweight, one-stage, multitask model designed for simultaneous power cable segmentation and broken strand detection from UAV imagery. Built upon the A-YOLOM architecture, and leveraging the YOLOv8 foundation, our model introduces four novel specialized modules—SDPM, HAD, EFR, and the Shape-Aware Wise IoU loss—that improve geometric understanding, structural consistency, and bounding-box precision. We also present the Merged Public Power Cable Dataset (MPCD), a diverse, open-source dataset tailored for multitask training and evaluation. The experimental results show that our model achieves up to +10.68% mAP@50 and +1.7% IoU compared to A-YOLOM, while also outperforming recent YOLO-based detectors in both accuracy and efficiency. These gains are achieved with a smaller model memory footprint and a similar inference speed compared to A-YOLOM. By unifying detection and segmentation into a single framework, PowerLine-MTYOLO offers a promising solution for autonomous aerial inspection and lays the groundwork for future advances in fine-structure monitoring tasks. Full article
Show Figures

Figure 1

21 pages, 5917 KiB  
Article
VML-UNet: Fusing Vision Mamba and Lightweight Attention Mechanism for Skin Lesion Segmentation
by Tang Tang, Haihui Wang, Qiang Rao, Ke Zuo and Wen Gan
Electronics 2025, 14(14), 2866; https://doi.org/10.3390/electronics14142866 - 17 Jul 2025
Abstract
Deep learning has advanced medical image segmentation, yet existing methods struggle with complex anatomical structures. Mainstream models, such as CNN, Transformer, and hybrid architectures, face challenges including insufficient information representation and redundant complexity, which limit their clinical deployment. Developing efficient and lightweight networks [...] Read more.
Deep learning has advanced medical image segmentation, yet existing methods struggle with complex anatomical structures. Mainstream models, such as CNN, Transformer, and hybrid architectures, face challenges including insufficient information representation and redundant complexity, which limit their clinical deployment. Developing efficient and lightweight networks is crucial for accurate lesion localization and optimized clinical workflows. We propose the VML-UNet, a lightweight segmentation network with core innovations including the CPMamba module and the multi-scale local supervision module (MLSM). The CPMamba module integrates the visual state space (VSS) block and a channel prior attention mechanism to enable efficient modeling of spatial relationships with linear computational complexity through dynamic channel-space weight allocation, while preserving channel feature integrity. The MLSM enhances local feature perception and reduces the inference burden. Comparative experiments were conducted on three public datasets, including ISIC2017, ISIC2018, and PH2, with ablation experiments performed on ISIC2017. VML-UNet achieves 0.53 M parameters, 2.18 MB memory usage, and 1.24 GFLOPs time complexity, with its performance on the datasets outperforming comparative networks, validating its effectiveness. This study provides valuable references for developing lightweight, high-performance skin lesion segmentation networks, advancing the field of skin lesion segmentation. Full article
(This article belongs to the Section Bioelectronics)
Show Figures

Figure 1

29 pages, 5825 KiB  
Article
BBSNet: An Intelligent Grading Method for Pork Freshness Based on Few-Shot Learning
by Chao Liu, Jiayu Zhang, Kunjie Chen and Jichao Huang
Foods 2025, 14(14), 2480; https://doi.org/10.3390/foods14142480 - 15 Jul 2025
Viewed by 186
Abstract
Deep learning approaches for pork freshness grading typically require large datasets, which limits their practical application due to the high costs associated with data collection. To address this challenge, we propose BBSNet, a lightweight few-shot learning model designed for accurate freshness classification with [...] Read more.
Deep learning approaches for pork freshness grading typically require large datasets, which limits their practical application due to the high costs associated with data collection. To address this challenge, we propose BBSNet, a lightweight few-shot learning model designed for accurate freshness classification with a limited number of images. BBSNet incorporates a batch channel normalization (BCN) layer to enhance feature distinguishability and employs BiFormer for optimized fine-grained feature extraction. Trained on a dataset of 600 pork images graded by microbial cell concentration, BBSNet achieved an average accuracy of 96.36% in a challenging 5-way 80-shot task. This approach significantly reduces data dependency while maintaining high accuracy, presenting a viable solution for cost-effective real-time pork quality monitoring. This work introduces a novel framework that connects laboratory freshness indicators to industrial applications in data-scarce conditions. Future research will investigate its extension to various food types and optimization for deployment on portable devices. Full article
Show Figures

Figure 1

20 pages, 5700 KiB  
Article
Multimodal Personality Recognition Using Self-Attention-Based Fusion of Audio, Visual, and Text Features
by Hyeonuk Bhin and Jongsuk Choi
Electronics 2025, 14(14), 2837; https://doi.org/10.3390/electronics14142837 - 15 Jul 2025
Viewed by 186
Abstract
Personality is a fundamental psychological trait that exerts a long-term influence on human behavior patterns and social interactions. Automatic personality recognition (APR) has exhibited increasing importance across various domains, including Human–Robot Interaction (HRI), personalized services, and psychological assessments. In this study, we propose [...] Read more.
Personality is a fundamental psychological trait that exerts a long-term influence on human behavior patterns and social interactions. Automatic personality recognition (APR) has exhibited increasing importance across various domains, including Human–Robot Interaction (HRI), personalized services, and psychological assessments. In this study, we propose a multimodal personality recognition model that classifies the Big Five personality traits by extracting features from three heterogeneous sources: audio processed using Wav2Vec2, video represented as Skeleton Landmark time series, and text encoded through Bidirectional Encoder Representations from Transformers (BERT) and Doc2Vec embeddings. Each modality is handled through an independent Self-Attention block that highlights salient temporal information, and these representations are then summarized and integrated using a late fusion approach to effectively reflect both the inter-modal complementarity and cross-modal interactions. Compared to traditional recurrent neural network (RNN)-based multimodal models and unimodal classifiers, the proposed model achieves an improvement of up to 12 percent in the F1-score. It also maintains a high prediction accuracy and robustness under limited input conditions. Furthermore, a visualization based on t-distributed Stochastic Neighbor Embedding (t-SNE) demonstrates clear distributional separation across the personality classes, enhancing the interpretability of the model and providing insights into the structural characteristics of its latent representations. To support real-time deployment, a lightweight thread-based processing architecture is implemented, ensuring computational efficiency. By leveraging deep learning-based feature extraction and the Self-Attention mechanism, we present a novel personality recognition framework that balances performance with interpretability. The proposed approach establishes a strong foundation for practical applications in HRI, counseling, education, and other interactive systems that require personalized adaptation. Full article
(This article belongs to the Special Issue Explainable Machine Learning and Data Mining)
Show Figures

Figure 1

21 pages, 4147 KiB  
Article
AgriFusionNet: A Lightweight Deep Learning Model for Multisource Plant Disease Diagnosis
by Saleh Albahli
Agriculture 2025, 15(14), 1523; https://doi.org/10.3390/agriculture15141523 - 15 Jul 2025
Viewed by 163
Abstract
Timely and accurate identification of plant diseases is critical to mitigating crop losses and enhancing yield in precision agriculture. This paper proposes AgriFusionNet, a lightweight and efficient deep learning model designed to diagnose plant diseases using multimodal data sources. The framework integrates RGB [...] Read more.
Timely and accurate identification of plant diseases is critical to mitigating crop losses and enhancing yield in precision agriculture. This paper proposes AgriFusionNet, a lightweight and efficient deep learning model designed to diagnose plant diseases using multimodal data sources. The framework integrates RGB and multispectral drone imagery with IoT-based environmental sensor data (e.g., temperature, humidity, soil moisture), recorded over six months across multiple agricultural zones. Built on the EfficientNetV2-B4 backbone, AgriFusionNet incorporates Fused-MBConv blocks and Swish activation to improve gradient flow, capture fine-grained disease patterns, and reduce inference latency. The model was evaluated using a comprehensive dataset composed of real-world and benchmarked samples, showing superior performance with 94.3% classification accuracy, 28.5 ms inference time, and a 30% reduction in model parameters compared to state-of-the-art models such as Vision Transformers and InceptionV4. Extensive comparisons with both traditional machine learning and advanced deep learning methods underscore its robustness, generalization, and suitability for deployment on edge devices. Ablation studies and confusion matrix analyses further confirm its diagnostic precision, even in visually ambiguous cases. The proposed framework offers a scalable, practical solution for real-time crop health monitoring, contributing toward smart and sustainable agricultural ecosystems. Full article
(This article belongs to the Special Issue Computational, AI and IT Solutions Helping Agriculture)
Show Figures

Figure 1

23 pages, 3492 KiB  
Article
A Multimodal Deep Learning Framework for Accurate Biomass and Carbon Sequestration Estimation from UAV Imagery
by Furkat Safarov, Ugiloy Khojamuratova, Misirov Komoliddin, Xusinov Ibragim Ismailovich and Young Im Cho
Drones 2025, 9(7), 496; https://doi.org/10.3390/drones9070496 - 14 Jul 2025
Viewed by 126
Abstract
Accurate quantification of above-ground biomass (AGB) and carbon sequestration is vital for monitoring terrestrial ecosystem dynamics, informing climate policy, and supporting carbon neutrality initiatives. However, conventional methods—ranging from manual field surveys to remote sensing techniques based solely on 2D vegetation indices—often fail to [...] Read more.
Accurate quantification of above-ground biomass (AGB) and carbon sequestration is vital for monitoring terrestrial ecosystem dynamics, informing climate policy, and supporting carbon neutrality initiatives. However, conventional methods—ranging from manual field surveys to remote sensing techniques based solely on 2D vegetation indices—often fail to capture the intricate spectral and structural heterogeneity of forest canopies, particularly at fine spatial resolutions. To address these limitations, we introduce ForestIQNet, a novel end-to-end multimodal deep learning framework designed to estimate AGB and associated carbon stocks from UAV-acquired imagery with high spatial fidelity. ForestIQNet combines dual-stream encoders for processing multispectral UAV imagery and a voxelized Canopy Height Model (CHM), fused via a Cross-Attentional Feature Fusion (CAFF) module, enabling fine-grained interaction between spectral reflectance and 3D structure. A lightweight Transformer-based regression head then performs multitask prediction of AGB and CO2e, capturing long-range spatial dependencies and enhancing generalization. Proposed method achieves an R2 of 0.93 and RMSE of 6.1 kg for AGB prediction, compared to 0.78 R2 and 11.7 kg RMSE for XGBoost and 0.73 R2 and 13.2 kg RMSE for Random Forest. Despite its architectural complexity, ForestIQNet maintains a low inference cost (27 ms per patch) and generalizes well across species, terrain, and canopy structures. These results establish a new benchmark for UAV-enabled biomass estimation and provide scalable, interpretable tools for climate monitoring and forest management. Full article
(This article belongs to the Special Issue UAVs for Nature Conservation Tasks in Complex Environments)
Show Figures

Figure 1

22 pages, 2492 KiB  
Article
VJDNet: A Simple Variational Joint Discrimination Network for Cross-Image Hyperspectral Anomaly Detection
by Shiqi Wu, Xiangrong Zhang, Guanchun Wang, Puhua Chen, Jing Gu, Xina Cheng and Licheng Jiao
Remote Sens. 2025, 17(14), 2438; https://doi.org/10.3390/rs17142438 - 14 Jul 2025
Viewed by 79
Abstract
To enhance the generalization of networks and avoid redundant training efforts, cross-image hyperspectral anomaly detection (HAD) based on deep learning has been gradually studied in recent years. Cross-image HAD aims to perform anomaly detection on unknown hyperspectral images after a single training process [...] Read more.
To enhance the generalization of networks and avoid redundant training efforts, cross-image hyperspectral anomaly detection (HAD) based on deep learning has been gradually studied in recent years. Cross-image HAD aims to perform anomaly detection on unknown hyperspectral images after a single training process on the network, thereby improving detection efficiency in practical applications. However, the existing approaches may require additional supervised information or stacking of networks to improve model performance, which may impose high demands on data or hardware in practical applications. In this paper, a simple and lightweight unsupervised cross-image HAD method called Variational Joint Discrimination Network (VJDNet) is proposed. We leverage the reconstruction and distribution representation ability of the variational autoencoder (VAE), learning the global and local discriminability of anomalies jointly. To integrate these representations from the VAE, a probability distribution joint discrimination (PDJD) module is proposed. Through the PDJD module, the VJDNet can directly output the anomaly score mask of pixels. To further facilitate the unsupervised paradigm, a sample pair generation module is proposed, which is able to generate anomaly samples and background representation samples tailored for the cross-image HAD task. The experimental results show that the proposed method is able to maintain the detection accuracy with only a small number of parameters. Full article
Show Figures

Figure 1

13 pages, 2828 KiB  
Article
Efficient Single-Exposure Holographic Imaging via a Lightweight Distilled Strategy
by Jiaosheng Li, Haoran Liu, Zeyu Lai, Yifei Chen, Chun Shan, Shuting Zhang, Youyou Liu, Tude Huang, Qilin Ma and Qinnan Zhang
Photonics 2025, 12(7), 708; https://doi.org/10.3390/photonics12070708 - 14 Jul 2025
Viewed by 73
Abstract
Digital holography can capture and reconstruct 3D object information, making it valuable for biomedical imaging and materials science. However, traditional holographic reconstruction methods require the use of phase shift operation in the time or space domain combined with complex computational processes, which, to [...] Read more.
Digital holography can capture and reconstruct 3D object information, making it valuable for biomedical imaging and materials science. However, traditional holographic reconstruction methods require the use of phase shift operation in the time or space domain combined with complex computational processes, which, to some extent, limits the range of application areas. The integration of deep learning (DL) advancements with physics-informed methodologies has opened new avenues for tackling this challenge. However, most of the existing DL-based holographic reconstruction methods have high model complexity. In this study, we first design a lightweight model with fewer parameters through the synergy of deep separable convolution and Swish activation function and then employ it as a teacher to distill a smaller student model. By reducing the number of network layers and utilizing knowledge distillation to improve the performance of a simple model, high-quality holographic reconstruction is achieved with only one hologram, greatly reducing the number of parameters in the network model. This distilled lightweight method cuts computational expenses dramatically, with its parameter count representing just 5.4% of the conventional Unet-based method, thereby facilitating efficient holographic reconstruction in settings with limited resources. Full article
(This article belongs to the Special Issue Advancements in Optical Metrology and Imaging)
Show Figures

Figure 1

15 pages, 6090 KiB  
Article
Automated Detection of Tailing Impoundments in Multi-Sensor High-Resolution Satellite Images Through Advanced Deep Learning Architectures
by Lin Qin and Wenyue Song
Sensors 2025, 25(14), 4387; https://doi.org/10.3390/s25144387 - 14 Jul 2025
Viewed by 176
Abstract
Accurate spatial mapping of Tailing Impoundments (TIs) is vital for environmental sustainability in mining ecosystems. While remote sensing enables large-scale monitoring, conventional methods relying on single-sensor data and traditional machine learning-based algorithm suffer from reduced accuracy in cluttered environments. This research proposes a [...] Read more.
Accurate spatial mapping of Tailing Impoundments (TIs) is vital for environmental sustainability in mining ecosystems. While remote sensing enables large-scale monitoring, conventional methods relying on single-sensor data and traditional machine learning-based algorithm suffer from reduced accuracy in cluttered environments. This research proposes a deep learning framework leveraging multi-source high-resolution imagery to address these limitations. An upgraded You Only Look Once (YOLO) model is introduced, integrating three key innovations: a multi-scale feature aggregation layer, a lightweight hierarchical fusion mechanism, and a modified loss metric. These components enhance the model’s ability to capture spatial dependencies, optimize inference speed, and ensure stable training dynamics. A comprehensive dataset of TIs across varied terrains was constructed, expanded through affine transformations, spectral perturbations, and adversarial sample synthesis. Evaluations confirm the framework’s superior performance in complex scenarios, achieving higher precision and computational efficiency than state-of-the-art detectors. Full article
(This article belongs to the Section Remote Sensors)
Show Figures

Figure 1

25 pages, 9517 KiB  
Article
YOLOv8n-SSDW: A Lightweight and Accurate Model for Barnyard Grass Detection in Fields
by Yan Sun, Hanrui Guo, Xiaoan Chen, Mengqi Li, Bing Fang and Yingli Cao
Agriculture 2025, 15(14), 1510; https://doi.org/10.3390/agriculture15141510 - 13 Jul 2025
Viewed by 162
Abstract
Barnyard grass is a major noxious weed in paddy fields. Accurate and efficient identification of barnyard grass is crucial for precision field management. However, existing deep learning models generally suffer from high parameter counts and computational complexity, limiting their practical application in field [...] Read more.
Barnyard grass is a major noxious weed in paddy fields. Accurate and efficient identification of barnyard grass is crucial for precision field management. However, existing deep learning models generally suffer from high parameter counts and computational complexity, limiting their practical application in field scenarios. Moreover, the morphological similarity, overlapping, and occlusion between barnyard grass and rice pose challenges for reliable detection in complex environments. To address these issues, this study constructed a barnyard grass detection dataset using high-resolution images captured by a drone equipped with a high-definition camera in rice experimental fields in Haicheng City, Liaoning Province. A lightweight field barnyard grass detection model, YOLOv8n-SSDW, was proposed to enhance detection precision and speed. Based on the baseline YOLOv8n model, a novel Separable Residual Coord Conv (SRCConv) was designed to replace the original convolution module, significantly reducing parameters while maintaining detection accuracy. The Spatio-Channel Enhanced Attention Module (SEAM) was introduced and optimized to improve sensitivity to barnyard grass edge features. Additionally, the lightweight and efficient Dysample upsampling module was incorporated to enhance feature map resolution. A new WIoU loss function was developed to improve bounding box classification and regression accuracy. Comprehensive performance analysis demonstrated that YOLOv8n-SSDW outperformed state-of-the-art models. Ablation studies confirmed the effectiveness of each improvement module. The final fused model achieved lightweight performance while improving detection accuracy, with a 2.2% increase in mAP_50, 3.8% higher precision, 0.6% higher recall, 10.6% fewer parameters, 9.8% lower FLOPs, and an 11.1% reduction in model size compared to the baseline. Field tests using drones combined with ground-based computers further validated the model’s robustness in real-world complex paddy environments. The results indicate that YOLOv8n-SSDW exhibits excellent accuracy and efficiency. This study provides valuable insights for barnyard grass detection in rice fields. Full article
Show Figures

Figure 1

25 pages, 3827 KiB  
Article
Source-Free Domain Adaptation Framework for Rotary Machine Fault Diagnosis
by Hoejun Jeong, Seungha Kim, Donghyun Seo and Jangwoo Kwon
Sensors 2025, 25(14), 4383; https://doi.org/10.3390/s25144383 - 13 Jul 2025
Viewed by 348
Abstract
Intelligent fault diagnosis for rotary machinery often suffers performance degradation under domain shifts between training and deployment environments. To address this, we propose a robust fault diagnosis framework incorporating three key components: (1) an order-frequency-based preprocessing method to normalize rotational variations, (2) a [...] Read more.
Intelligent fault diagnosis for rotary machinery often suffers performance degradation under domain shifts between training and deployment environments. To address this, we propose a robust fault diagnosis framework incorporating three key components: (1) an order-frequency-based preprocessing method to normalize rotational variations, (2) a U-Net variational autoencoder (U-NetVAE) to enhance adaptation through reconstruction learning, and (3) a test-time training (TTT) strategy enabling unsupervised target domain adaptation without access to source data. Since existing works rarely evaluate under true domain shift conditions, we first construct a unified cross-domain benchmark by integrating four public datasets with consistent class and sensor settings. The experimental results show that our method outperforms conventional machine learning and deep learning models in both F1-score and recall across domains. Notably, our approach maintains an F1-score of 0.47 and recall of 0.51 in the target domain, outperforming others under identical conditions. Ablation studies further confirm the contribution of each component to adaptation performance. This study highlights the effectiveness of combining mechanical priors, self-supervised learning, and lightweight adaptation strategies for robust fault diagnosis in the practical domain. Full article
(This article belongs to the Special Issue Sensor Data-Driven Fault Diagnosis Techniques)
Show Figures

Figure 1

22 pages, 6645 KiB  
Article
Visual Detection on Aircraft Wing Icing Process Using a Lightweight Deep Learning Model
by Yang Yan, Chao Tang, Jirong Huang, Zhixiong Cen and Zonghong Xie
Aerospace 2025, 12(7), 627; https://doi.org/10.3390/aerospace12070627 - 12 Jul 2025
Viewed by 103
Abstract
Aircraft wing icing significantly threatens aviation safety, causing substantial losses to the aviation industry each year. High transparency and blurred edges of icing areas in wing images pose challenges to wing icing detection by machine vision. To address these challenges, this study proposes [...] Read more.
Aircraft wing icing significantly threatens aviation safety, causing substantial losses to the aviation industry each year. High transparency and blurred edges of icing areas in wing images pose challenges to wing icing detection by machine vision. To address these challenges, this study proposes a detection model, Wing Icing Detection DeeplabV3+ (WID-DeeplabV3+), for efficient and precise aircraft wing leading edge icing detection under natural lighting conditions. WID-DeeplabV3+ adopts the lightweight MobileNetV3 as its backbone network to enhance the extraction of edge features in icing areas. Ghost Convolution and Atrous Spatial Pyramid Pooling modules are incorporated to reduce model parameters and computational complexity. The model is optimized using the transfer learning method, where pre-trained weights are utilized to accelerate convergence and enhance performance. Experimental results show WID-DeepLabV3+ segments the icing edge at 1920 × 1080 within 0.03 s. The model achieves the accuracy of 97.15%, an IOU of 94.16%, a precision of 97%, and a recall of 96.96%, representing respective improvements of 1.83%, 3.55%, 1.79%, and 2.04% over DeeplabV3+. The number of parameters and computational complexity are reduced by 92% and 76%, respectively. With high accuracy, superior IOU, and fast inference speed, WID-DeeplabV3+ provides an effective solution for wing-icing detection. Full article
(This article belongs to the Section Aeronautics)
Show Figures

Figure 1

19 pages, 2641 KiB  
Article
MSFF-Net: Multi-Sensor Frequency-Domain Feature Fusion Network with Lightweight 1D CNN for Bearing Fault Diagnosis
by Miao Dai, Hangyeol Jo, Moonsuk Kim and Sang-Woo Ban
Sensors 2025, 25(14), 4348; https://doi.org/10.3390/s25144348 - 11 Jul 2025
Viewed by 296
Abstract
This study proposes MSFF-Net, a lightweight deep learning framework for bearing fault diagnosis based on frequency-domain multi-sensor fusion. The vibration and acoustic signals are initially converted into the frequency domain using the fast Fourier transform (FFT), enabling the extraction of temporally invariant spectral [...] Read more.
This study proposes MSFF-Net, a lightweight deep learning framework for bearing fault diagnosis based on frequency-domain multi-sensor fusion. The vibration and acoustic signals are initially converted into the frequency domain using the fast Fourier transform (FFT), enabling the extraction of temporally invariant spectral features. These features are processed by a compact one-dimensional convolutional neural network, where modality-specific representations are fused at the feature level to capture complementary fault-related information. The proposed method demonstrates robust and superior performance under both full and scarce data conditions, as verified through experiments on a publicly available dataset. Experimental results on a publicly available dataset indicate that the proposed model attains an average accuracy of 99.73%, outperforming state-of-the-art (SOTA) methods in both accuracy and stability. With only about 70.3% of the parameters of the SOTA model, it offers faster inference and reduced computational cost. Ablation studies confirm that multi-sensor fusion improves all classification metrics over single-sensor setups. Under few-shot conditions with 20 samples per class, the model retains 94.69% accuracy, highlighting its strong generalization in data-limited scenarios. The results validate the effectiveness, computational efficiency, and practical applicability of the model for deployment in data-constrained industrial environments. Full article
(This article belongs to the Special Issue Condition Monitoring in Manufacturing with Advanced Sensors)
Show Figures

Figure 1

Back to TopTop