Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (901)

Search Parameters:
Keywords = convolution dimension

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
31 pages, 7049 KB  
Article
Objective Emotion Assessment Using a Triple Attention Network for an EEG-Based Brain–Computer Interface
by Lihua Zhang, Xin Zhang, Xiu Zhang, Changyi Yu and Xuguang Liu
Brain Sci. 2025, 15(11), 1167; https://doi.org/10.3390/brainsci15111167 - 29 Oct 2025
Abstract
Background: The assessment of emotion recognition holds growing significance in research on the brain–computer interface and human–computer interaction. Among diverse physiological signals, electroencephalography (EEG) occupies a pivotal position in affective computing due to its exceptional temporal resolution and non-invasive acquisition. However, EEG signals [...] Read more.
Background: The assessment of emotion recognition holds growing significance in research on the brain–computer interface and human–computer interaction. Among diverse physiological signals, electroencephalography (EEG) occupies a pivotal position in affective computing due to its exceptional temporal resolution and non-invasive acquisition. However, EEG signals are inherently complex, characterized by substantial noise contamination and high variability, posing considerable challenges to accurate assessment. Methods: To tackle these challenges, we propose a Triple Attention Network (TANet), a triple-attention EEG emotion recognition framework that integrates Conformer, Convolutional Block Attention Module (CBAM), and Mutual Cross-Modal Attention (MCA). The Conformer component captures temporal feature dependencies, CBAM refines spatial channel representations, and MCA performs cross-modal fusion of differential entropy and power spectral density features. Results: We evaluated TANet on two benchmark EEG emotion datasets, DEAP and SEED. On SEED, using a subject-specific cross-validation protocol, the model reached an average accuracy of 98.51 ± 1.40%. On DEAP, we deliberately adopted a segment-level splitting paradigm—in line with influential state-of-the-art methods—to ensure a direct and fair comparison of model architecture under an identical evaluation protocol. This approach, designed specifically to assess fine-grained within-trial pattern discrimination rather than cross-subject generalization, yielded accuracies of 99.69 ± 0.15% and 99.67 ± 0.13% for the valence and arousal dimensions, respectively. Compared with existing benchmark approaches under similar evaluation protocols, TANet delivers substantially better results, underscoring the strong complementary effects of its attention mechanisms in improving EEG-based emotion recognition performance. Conclusions: This work provides both theoretical insights into multi-dimensional attention for physiological signal processing and practical guidance for developing high-performance, robust EEG emotion assessment systems. Full article
(This article belongs to the Section Neurotechnology and Neuroimaging)
Show Figures

Figure 1

22 pages, 6546 KB  
Article
Birds-YOLO: A Bird Detection Model for Dongting Lake Based on Modified YOLOv11
by Shuai Fang, Yue Shen, Haojie Zou, Yerong Yin, Wei Jin and Haoyu Zhou
Biology 2025, 14(11), 1515; https://doi.org/10.3390/biology14111515 - 29 Oct 2025
Abstract
To address the challenges posed by complex background interference, varying target sizes, and high species diversity in bird detection tasks in the Dongting Lake region, this paper proposes an enhanced bird detection model named Birds-YOLO, based on the YOLOv11 framework. First, the EMA [...] Read more.
To address the challenges posed by complex background interference, varying target sizes, and high species diversity in bird detection tasks in the Dongting Lake region, this paper proposes an enhanced bird detection model named Birds-YOLO, based on the YOLOv11 framework. First, the EMA mechanism is introduced to replace the original C2PSA module. This mechanism synchronously captures global dependencies in the channel dimension and local detailed features in the spatial dimension, thereby enhancing the model’s robustness in cluttered environments. Second, the model incorporates an improved RepNCSPELAN4-ECO module, by reasonably integrating depthwise separable convolution modules and combining them with an adaptive channel compression mechanism, to strengthen feature extraction and multi-scale feature fusion, effectively enhances the detection capability for bird targets at different scales. Finally, the neck component of the network is redesigned using lightweight GSConv convolution, which integrates the principles of grouped and spatial convolutions. This design preserves the feature modeling capacity of standard convolution while incorporating the computational efficiency of depthwise separable convolution, thereby reducing model complexity without sacrificing accuracy. Experimental results show that, compared to the baseline YOLOv11n, Birds-YOLO achieves a 5.0% improvement in recall and a 3.5% increase in mAP@0.5 on the CUB200-2011 dataset. On the in-house DTH-Birds dataset, it gains 3.7% in precision, 3.7% in recall, and 2.6% in mAP@0.5, demonstrating consistent performance enhancement across both public and private benchmarks. The model’s generalization ability and robustness are further validated through extensive ablation studies and comparative experiments, indicating its strong potential for practical deployment in bird detection tasks in complex natural environments such as Dongting Lake. Full article
(This article belongs to the Special Issue Bird Biology and Conservation)
Show Figures

Figure 1

33 pages, 4302 KB  
Article
Artificial Intelligence-Based Plant Disease Classification in Low-Light Environments
by Hafiz Ali Hamza Gondal, Seong In Jeong, Won Ho Jang, Jun Seo Kim, Rehan Akram, Muhammad Irfan, Muhammad Hamza Tariq and Kang Ryoung Park
Fractal Fract. 2025, 9(11), 691; https://doi.org/10.3390/fractalfract9110691 - 27 Oct 2025
Abstract
The accurate classification of plant diseases is vital for global food security, as diseases can cause major yield losses and threaten sustainable and precision agriculture. The classification of plant diseases in low-light noisy environments is crucial because crops can be continuously monitored even [...] Read more.
The accurate classification of plant diseases is vital for global food security, as diseases can cause major yield losses and threaten sustainable and precision agriculture. The classification of plant diseases in low-light noisy environments is crucial because crops can be continuously monitored even at night. Important visual cues of disease symptoms can be lost due to the degraded quality of images captured under low-illumination, resulting in poor performance of conventional plant disease classifiers. However, researchers have proposed various techniques for classifying plant diseases in daylight, and no studies have been conducted for low-light noisy environments. Therefore, we propose a novel model for classifying plant diseases from low-light noisy images called dilated pixel attention network (DPA-Net). DPA-Net uses a pixel attention mechanism and multi-layer dilated convolution with a high receptive field, which obtains essential features while highlighting the most relevant information under this challenging condition, allowing more accurate classification results. Additionally, we performed fractal dimension estimation on diseased and healthy leaves to analyze the structural irregularities and complexities. For the performance evaluation, experiments were conducted on two public datasets: the PlantVillage and Potato Leaf Disease datasets. In both datasets, the image resolution is 256 × 256 pixels in joint photographic experts group (JPG) format. For the first dataset, DPA-Net achieved an average accuracy of 92.11% and harmonic mean of precision and recall (F1-score) of 89.11%. For the second dataset, it achieved an average accuracy of 88.92% and an F1-score of 88.60%. These results revealed that the proposed method outperforms state-of-the-art methods. On the first dataset, our method achieved an improvement of 2.27% in average accuracy and 2.86% in F1-score compared to the baseline. Similarly, on the second dataset, it attained an improvement of 6.32% in average accuracy and 6.37% in F1-score over the baseline. In addition, we confirm that our method is effective with the real low-illumination dataset self-constructed by capturing images at 0 lux using a smartphone at night. This approach provides farmers with an affordable practical tool for early disease detection, which can support crop protection worldwide. Full article
Show Figures

Figure 1

27 pages, 2176 KB  
Article
Intelligent Fault Diagnosis of Rolling Bearings Based on Digital Twin and Multi-Scale CNN-AT-BiGRU Model
by Jiayu Shi, Liang Qi, Shuxia Ye, Changjiang Li, Chunhui Jiang, Zhengshun Ni, Zheng Zhao, Zhe Tong, Siyu Fei, Runkang Tang, Danfeng Zuo and Jiajun Gong
Symmetry 2025, 17(11), 1803; https://doi.org/10.3390/sym17111803 - 26 Oct 2025
Viewed by 262
Abstract
Rolling bearings constitute critical rotating components within rolling mill equipment. Production efficiency and the operational safety of the whole mechanical system are directly governed by their operational health state. To address the dual challenges of the over-reliance of conventional diagnostic methods on expert [...] Read more.
Rolling bearings constitute critical rotating components within rolling mill equipment. Production efficiency and the operational safety of the whole mechanical system are directly governed by their operational health state. To address the dual challenges of the over-reliance of conventional diagnostic methods on expert experience and the scarcity of fault samples in industrial scenarios, we propose a virtual–physical data fusion-optimized intelligent fault diagnosis framework. Initially, a dynamics-based digital twin model for rolling bearings is developed by leveraging their geometric symmetry. It is capable of generating comprehensive fault datasets through parametric adjustments of bearing dimensions and operational environments in virtual space. Subsequently, a symmetry-informed architecture is constructed, which integrates multi-scale convolutional neural networks with attention mechanisms and bidirectional gated recurrent units (MCNN-AT-BiGRU). This architecture enables spatiotemporal feature extraction and enhances critical fault characteristics. The experimental results demonstrate 99.5% fault identification accuracy under single operating conditions. It maintains stable performance under low SNR conditions. Furthermore, the framework exhibits superior generalization capability and transferability across the different bearing types. Full article
(This article belongs to the Section Computer)
Show Figures

Figure 1

19 pages, 1781 KB  
Article
HiSeq-TCN: High-Dimensional Feature Sequence Modeling and Few-Shot Reinforcement Learning for Intrusion Detection
by Yadong Pei, Yanfei Tan, Wei Gao, Fangwei Li and Mingyue Wang
Electronics 2025, 14(21), 4168; https://doi.org/10.3390/electronics14214168 - 25 Oct 2025
Viewed by 215
Abstract
Intrusion detection is essential to cybersecurity. However, the curse of dimensionality and class imbalance limit detection accuracy and impede the identification of rare attacks. To address these challenges, this paper proposes the high-dimensional feature sequence temporal convolutional network (HiSeq-TCN) for intrusion detection. The [...] Read more.
Intrusion detection is essential to cybersecurity. However, the curse of dimensionality and class imbalance limit detection accuracy and impede the identification of rare attacks. To address these challenges, this paper proposes the high-dimensional feature sequence temporal convolutional network (HiSeq-TCN) for intrusion detection. The proposed HiSeq-TCN transforms high-dimensional feature vectors into pseudo-temporal sequences, enabling the network to capture contextual dependencies across feature dimensions. This enhances feature representation and detection robustness. In addition, a few-shot reinforcement strategy adaptively assigns larger loss weights to minority classes, mitigating class imbalance and improving the recognition of rare attacks. Experiments on the NSL-KDD dataset show that HiSeq-TCN achieves an overall accuracy of 99.44%, outperforming support vector machines, deep neural networks, and long short-term memory models. More importantly, it significantly improves the detection of rare attack types such as remote-to-local and user-to-root attacks. These results highlight the potential of HiSeq-TCN for robust and reliable intrusion detection in practical cybersecurity environments. Full article
(This article belongs to the Special Issue Applications of Deep Learning in Cyber Threat Detection)
Show Figures

Graphical abstract

25 pages, 5292 KB  
Article
Deep Learning-Based Non-Parametric System Identification and Interpretability Analysis for Improving Ship Motion Prediction
by Shaojie Guo, Siqing Zhuang, Junyi Wang, Xi Peng and Yihua Liu
J. Mar. Sci. Eng. 2025, 13(10), 2017; https://doi.org/10.3390/jmse13102017 - 21 Oct 2025
Viewed by 259
Abstract
The proposed hybrid model integrates a convolutional neural network, bidirectional long short-term memory network, and attention mechanism. This model is applied to the nonparametric system identification of ship motion, incorporating wind factors. The model processes input data with different historical dimensions after preprocessing, [...] Read more.
The proposed hybrid model integrates a convolutional neural network, bidirectional long short-term memory network, and attention mechanism. This model is applied to the nonparametric system identification of ship motion, incorporating wind factors. The model processes input data with different historical dimensions after preprocessing, extracts local features using a CNN layer, captures bidirectional temporal dependencies via a BiLSTM layer to provide comprehensive bidirectional information, and finally introduces a multi-head attention mechanism to enhance the model’s expressive and learning capabilities. However, the use of deep neural networks introduces difficulties in explaining internal mechanisms. The coupled CNN-BiLSTM-Attention model with SHapley Additive exPlanations was adopted for the prediction of ship motion processes and the identification of key input feature factors. The effectiveness of the proposed model was validated through experiments using a ship free-running motion dataset with wind interference. The findings indicate that, in comparison to conventional single-architecture models and composite architecture models, the proposed model attains smaller prediction errors and demonstrates augmented generalizability and robustness. Full article
(This article belongs to the Section Ocean Engineering)
Show Figures

Figure 1

24 pages, 1993 KB  
Article
The Downstream Supply Chain for Electricity Generated from Renewables in Egypt: A Dynamic Analysis
by Islam Hassanin, Tariq Muneer and Matjaz Knez
Logistics 2025, 9(4), 150; https://doi.org/10.3390/logistics9040150 - 21 Oct 2025
Viewed by 289
Abstract
Background: Generating electricity from renewable sources continues to receive significant attention from both scholars and professional communities. This is mainly because traditional energy use harms public health, threatens biodiversity, and increases pollution, particularly in developing countries. Meanwhile, renewable technologies are considered one of [...] Read more.
Background: Generating electricity from renewable sources continues to receive significant attention from both scholars and professional communities. This is mainly because traditional energy use harms public health, threatens biodiversity, and increases pollution, particularly in developing countries. Meanwhile, renewable technologies are considered one of the most effective solutions to enrich energy security for future usage with clean practices and affordable prices. However, planning such applications may become complex due to the convolution of many technical, economic, environmental, and social dimensions, particularly from a supply chain management viewpoint. Methods: The paper identifies the dimensions affecting the supply chain variables of downstream processes in renewable energy supply systems, especially for generating electricity in Egypt. Also, this paper investigates the relationships between the dimensions of renewable energy supply systems and the downstream supply chain variables that are closely related to the Egyptian energy sector. Results: The different relationships between these indicators and downstream supply chain variables are revealed. Conclusions: This study employed conceptual causality diagramming to organize these relationships harmoniously, which helps to analyze the behavior of the supply chain during the transitions to renewable energy applications and its implications, whether at the managerial or policy and procedural levels. Full article
Show Figures

Figure 1

25 pages, 8228 KB  
Article
Soybean Seed Classification and Identification Based on Corner Point Multi-Feature Segmentation and Improved MobileViT
by Yu Xia, Rui Zhu, Fan Ji, Junlan Zhang, Kunjie Chen and Jichao Huang
AgriEngineering 2025, 7(10), 354; https://doi.org/10.3390/agriengineering7100354 - 21 Oct 2025
Viewed by 243
Abstract
To address the challenges of high model complexity, substantial computational resource consumption, and insufficient classification accuracy in existing soybean seed identification research, we first perform soybean seed segmentation based on polygon features, constructing a dataset comprising five categories: whole seeds, broken seeds, seeds [...] Read more.
To address the challenges of high model complexity, substantial computational resource consumption, and insufficient classification accuracy in existing soybean seed identification research, we first perform soybean seed segmentation based on polygon features, constructing a dataset comprising five categories: whole seeds, broken seeds, seeds with epidermal damage, immature seeds, and spotted seeds. The MobileViT module is then optimized by employing Depthwise Separable Convolution (DSC) in place of standard convolutions, applying Transformer Half-Dimension (THD) for dimensional reconstruction, and integrating Dynamic Channel Recalibration (DCR) to reduce model parameters and enhance inter-channel interactions. Furthermore, by incorporating the CBAM attention mechanism into the MV2 module and replacing the ReLU6 activation function with the Mish activation function, the model’s feature extraction capability and generalization performance are further improved. These enhancements culminate in a novel soybean seed detection model, MobileViT-SD (MobileViT for Soybean Detection). Experimental results demonstrate that the proposed MobileViT-SD model contains only 2.09 million parameters while achieving a classification accuracy of 98.39% and an F1 score of 98.38%, representing improvements of 2.86% and 2.88%, respectively, over the original MobileViT model. Comparative experiments further show that MobileViT-SD not only outperforms several representative lightweight models in both detection accuracy and efficiency but also surpasses a number of mainstream heavyweight models. Its highly optimized, lightweight architecture combines efficient inference performance with low resource consumption, making it well-suited for deployment in computing-constrained environments, such as edge devices. Full article
Show Figures

Figure 1

23 pages, 1986 KB  
Article
GMHCA-MCBILSTM: A Gated Multi-Head Cross-Modal Attention-Based Network for Emotion Recognition Using Multi-Physiological Signals
by Xueping Li, Yanbo Li, Yuhang Li and Yuan Yang
Algorithms 2025, 18(10), 664; https://doi.org/10.3390/a18100664 - 20 Oct 2025
Viewed by 444
Abstract
To address the limitations of the single-modal electroencephalogram (EEG), such as its single physiological dimension, weak anti-interference ability, and inability to fully reflect emotional states, this paper proposes a gated multi-head cross-attention module (GMHCA) for multimodal fusion of EEG, electrooculography (EOG),and electrodermal activity [...] Read more.
To address the limitations of the single-modal electroencephalogram (EEG), such as its single physiological dimension, weak anti-interference ability, and inability to fully reflect emotional states, this paper proposes a gated multi-head cross-attention module (GMHCA) for multimodal fusion of EEG, electrooculography (EOG),and electrodermal activity (EDA). This attention module employs three independent and parallel attention computation units to assign independent attention weights to different feature subsets across modalities. Combined with a modality complementarity metric, the gating mechanism suppresses redundant heads and enhances the information transmission of key heads. Through multi-head concatenation, cross-modal interaction results from different perspectives are fused. For the backbone network, a multi-scale convolution and bidirectional long short-term memory network (MC-BiLSTM) is designed for feature extraction, tailored to the characteristics of each modality. Experiments show that this method, which primarily fuses eight-channel EEG with peripheral physiological signals, achieves an emotion recognition accuracy of 89.45%, a 7.68% improvement over single-modal EEG. In addition, in cross-subject experiments conducted on the SEED-IV dataset, the EEG+EOG modality achieved a classification accuracy of 92.73%. All were significantly better than the baseline method. This fully demonstrates the effectiveness of the innovative GMHCA module architecture and MC-BiLSTM feature extraction network proposed in this paper for multimodal fusion methods. Through the novel attention gating mechanism, higher recognition accuracy is achieved while significantly reducing the number of EEG channels, providing new ideas and approaches based on attention mechanisms and gated fusion for multimodal emotion recognition in resource-constrained environments. Full article
(This article belongs to the Special Issue Machine Learning in Medical Signal and Image Processing (4th Edition))
Show Figures

Graphical abstract

21 pages, 1453 KB  
Review
Current Trends and Future Opportunities of AI-Based Analysis in Mesenchymal Stem Cell Imaging: A Scoping Review
by Maksim Solopov, Elizaveta Chechekhina, Viktor Turchin, Andrey Popandopulo, Dmitry Filimonov, Anzhelika Burtseva and Roman Ishchenko
J. Imaging 2025, 11(10), 371; https://doi.org/10.3390/jimaging11100371 - 18 Oct 2025
Viewed by 370
Abstract
This scoping review explores the application of artificial intelligence (AI) methods for analyzing mesenchymal stem cells (MSCs) images. The aim of this study was to identify key areas where AI-based image processing techniques are utilized for MSCs analysis, assess their effectiveness, and highlight [...] Read more.
This scoping review explores the application of artificial intelligence (AI) methods for analyzing mesenchymal stem cells (MSCs) images. The aim of this study was to identify key areas where AI-based image processing techniques are utilized for MSCs analysis, assess their effectiveness, and highlight existing challenges. A total of 25 studies published between 2014 and 2024 were selected from six databases (PubMed, Dimensions, Scopus, Google Scholar, eLibrary, and Cochrane) for this review. The findings demonstrate that machine learning algorithms outperform traditional methods in terms of accuracy (up to 97.5%), processing speed and noninvasive capabilities. Among AI methods, convolutional neural networks (CNNs) are the most widely employed, accounting for 64% of the studies reviewed. The primary applications of AI in MSCs image analysis include cell classification (20%), segmentation and counting (20%), differentiation assessment (32%), senescence analysis (12%), and other tasks (16%). The advantages of AI methods include automation of image analysis, elimination of subjective biases, and dynamic monitoring of live cells without the need for fixation and staining. However, significant challenges persist, such as the high heterogeneity of the MSCs population, the absence of standardized protocols for AI implementation, and limited availability of annotated datasets. To advance this field, future efforts should focus on developing interpretable and multimodal AI models, creating standardized validation frameworks and open-access datasets, and establishing clear regulatory pathways for clinical translation. Addressing these challenges is crucial for accelerating the adoption of AI in MSCs biomanufacturing and enhancing the efficacy of cell therapies. Full article
Show Figures

Figure 1

13 pages, 1352 KB  
Article
“Speed”: A Dataset for Human Speed Estimation
by Zainab R. Bachir and Usman Tariq
Sensors 2025, 25(20), 6335; https://doi.org/10.3390/s25206335 - 14 Oct 2025
Viewed by 379
Abstract
Over the years, researchers have developed several speed estimation techniques using wearable inertial measurement units (IMUs). In this paper, we introduce a medium-scale dataset, containing measurements of walking/running at speeds ranging from 4.0 km/h (1.11 m/s) to 9.5 km/h (2.64 m/s) in increments [...] Read more.
Over the years, researchers have developed several speed estimation techniques using wearable inertial measurement units (IMUs). In this paper, we introduce a medium-scale dataset, containing measurements of walking/running at speeds ranging from 4.0 km/h (1.11 m/s) to 9.5 km/h (2.64 m/s) in increments of 0.5 km/h (0.14 m/s) from 33 healthy subjects wearing IMUs. We name it the “Speed” dataset. In summary, we present accelerometer and gyroscope data from 12 speeds and 22 subject-independent sets with the full range of 12 speeds. The data in each set consists of overlapping sections of 250 time samples (corresponding to 2.5 s, sampled at 100 Hz), and six dimensions (corresponding to the three axes of the accelerometer and three axes of the gyroscope). Each speed set contains 1775 examples. We benchmark the existing approaches used in the literature for the purpose of speed estimation on this dataset. These include support vector regression, Gaussian Process Regression, and shallow neural networks. We then design a deep Convolutional Neural Network (CNN), SpeedNet, for baseline results. The proposed SpeedNet yields an average Root Mean Square Error (RMSE) of 0.4819 km/h (0.13 m/s), following a subject-independent approach. Then, the SpeedNet obtained from the subject-independent approach are adapted using a portion of subject-specific data. The average RMSE for the remainder of the data for all subjects then drops down to 0.1747 km/h (0.05 m/s). The suggested SpeedNet yields a lower RMSE in comparison to the other approaches. In addition, we also compare the proposed method to others in terms of the average testing time, to give an idea of computational complexity. The proposed SpeedNet, despite being more accurate, yields real-time performance. Full article
(This article belongs to the Section Wearables)
Show Figures

Figure 1

31 pages, 3416 KB  
Article
Accurate Estimation of Forest Canopy Height Based on GEDI Transmitted Deconvolution Waveforms
by Longtao Cai, Jun Wu, Inthasone Somsack, Xuemei Zhao and Jiasheng He
Remote Sens. 2025, 17(20), 3412; https://doi.org/10.3390/rs17203412 - 11 Oct 2025
Viewed by 457
Abstract
Accurate estimation of the forest canopy height is crucial in monitoring the global carbon cycle and evaluating progress toward carbon neutrality goals. The Global Ecosystem Dynamics Investigation (GEDI) mission provides an important data source for canopy height estimation at a global scale. However, [...] Read more.
Accurate estimation of the forest canopy height is crucial in monitoring the global carbon cycle and evaluating progress toward carbon neutrality goals. The Global Ecosystem Dynamics Investigation (GEDI) mission provides an important data source for canopy height estimation at a global scale. However, the non-zero half-width of the transmitted laser pulses (NHWTLP) and the influence of terrain slope can cause waveform broadening and overlap between canopy returns and ground returns in GEDI waveforms, thereby reducing the estimation accuracy. To address these limitations, we propose a canopy height retrieval method that combines the deconvolution of GEDI’s transmitted waveforms with terrain slope constraints on the ground response function. The method consists of two main components. The first is performing deconvolution on GEDI’s effective return waveforms using their corresponding transmitted waveforms to obtain the true ground response function within each GEDI footprint, thereby mitigating waveform broadening and overlap induced by NHWTLP. This process includes constructing a convolution convergence function for GEDI waveforms, denoising GEDI waveform data, transforming one-dimensional ground response functions into two dimensions, and applying amplitude difference regularization between the convolved and observed waveforms. The second is incorporating terrain slope parameters derived from a digital terrain model (DTM) as constraints in the canopy height estimation model to alleviate waveform broadening and overlap in ground response functions caused by topographic effects. The proposed approach enhances the precision of forest canopy height estimation from GEDI data, particularly in areas with complex terrain. The results demonstrate that, under various conditions—including GEDI full-power beams and coverage beams, different terrain slopes, varying canopy closures, and multiple study areas—the retrieved height (rh) model constructed from ground response functions derived via the inverse deconvolution of the transmitted waveforms (IDTW) outperforms the RH (the official height from GEDI L2A) model constructed using RH parameters from GEDI L2A data files in forest canopy height estimation. Specifically, without incorporating terrain slope, the rh model for canopy height estimation using full-power beams achieved a coefficient of determination (R2) of 0.58 and a root mean square error (RMSE) of 5.23 m, compared to the RH model, which had an R2 of 0.58 and an RMSE of 5.54 m. After incorporating terrain slope, the rh_g model for full-power beams in canopy height estimation yielded an R2 of 0.61 and an RMSE of 5.21 m, while the RH_g model attained an R2 of 0.60 and an RMSE of 5.45 m. These findings indicate that the proposed method effectively mitigates waveform broadening and overlap in GEDI waveforms, thereby enhancing the precision of forest canopy height estimation, particularly in areas with complex terrain. This approach provides robust technical support for global-scale forest resource assessment and contributes to the accurate monitoring of carbon dynamics. Full article
(This article belongs to the Collection Feature Paper Special Issue on Forest Remote Sensing)
Show Figures

Figure 1

19 pages, 8850 KB  
Article
Intelligent Defect Recognition of Glazed Components in Ancient Buildings Based on Binocular Vision
by Youshan Zhao, Xiaolan Zhang, Ming Guo, Haoyu Han, Jiayi Wang, Yaofeng Wang, Xiaoxu Li and Ming Huang
Buildings 2025, 15(20), 3641; https://doi.org/10.3390/buildings15203641 - 10 Oct 2025
Viewed by 201
Abstract
Glazed components in ancient Chinese architecture hold profound historical and cultural value. However, over time, environmental erosion, physical impacts, and human disturbances gradually lead to various forms of damage, severely impacting the durability and stability of the buildings. Therefore, preventive protection of glazed [...] Read more.
Glazed components in ancient Chinese architecture hold profound historical and cultural value. However, over time, environmental erosion, physical impacts, and human disturbances gradually lead to various forms of damage, severely impacting the durability and stability of the buildings. Therefore, preventive protection of glazed components is crucial. The key to preventive protection lies in the early detection and repair of damage, thereby extending the component’s service life and preventing significant structural damage. To address this challenge, this study proposes a Restoration-Scale Identification (RSI) method that integrates depth information. By combining RGB-D images acquired from a depth camera with intrinsic camera parameters, and embedding a Convolutional Block Attention Module (CBAM) into the backbone network, the method dynamically enhances critical feature regions. It then employs a scale restoration strategy to accurately identify damage areas and recover the physical dimensions of glazed components from a global perspective. In addition, we constructed a dedicated semantic segmentation dataset for glazed tile damage, focusing on cracks and spalling. Both qualitative and quantitative evaluation results demonstrate that, compared with various high-performance semantic segmentation methods, our approach significantly improves the accuracy and robustness of damage detection in glazed components. The achieved accuracy deviates by only ±10 mm from high-precision laser scanning, a level of precision that is essential for reliably identifying and assessing subtle damages in complex glazed architectural elements. By integrating depth information, real scale information can be effectively obtained during the intelligent recognition process, thereby efficiently and accurately identifying the type of damage and size information of glazed components, and realizing the conversion from two-dimensional (2D) pixel coordinates to local three-dimensional (3D) coordinates, providing a scientific basis for the protection and restoration of ancient buildings, and ensuring the long-term stability of cultural heritage and the inheritance of historical value. Full article
(This article belongs to the Section Building Materials, and Repair & Renovation)
Show Figures

Figure 1

12 pages, 768 KB  
Article
ECG Waveform Segmentation via Dual-Stream Network with Selective Context Fusion
by Yongpeng Niu, Nan Lin, Yuchen Tian, Kaipeng Tang and Baoxiang Liu
Electronics 2025, 14(19), 3925; https://doi.org/10.3390/electronics14193925 - 2 Oct 2025
Viewed by 336
Abstract
Electrocardiogram (ECG) waveform delineation is fundamental to cardiac disease diagnosis. This task requires precise localization of key fiducial points, specifically the onset, peak, and offset positions of P-waves, QRS complexes, and T-waves. Current methods exhibit significant performance degradation in noisy clinical environments (baseline [...] Read more.
Electrocardiogram (ECG) waveform delineation is fundamental to cardiac disease diagnosis. This task requires precise localization of key fiducial points, specifically the onset, peak, and offset positions of P-waves, QRS complexes, and T-waves. Current methods exhibit significant performance degradation in noisy clinical environments (baseline drift, electromyographic interference, powerline interference, etc.), compromising diagnostic reliability. To address this limitation, we introduce ECG-SCFNet: a novel dual-stream architecture employing selective context fusion. Our framework is further enhanced by a consistency training paradigm, enabling it to maintain robust waveform delineation accuracy under challenging noise conditions.The network employs a dual-stream architecture: (1) A temporal stream captures dynamic rhythmic features through sequential multi-branch convolution and temporal attention mechanisms; (2) A morphology stream combines parallel multi-scale convolution with feature pyramid integration to extract multi-scale waveform structural features through morphological attention; (3) The Selective Context Fusion (SCF) module adaptively integrates features from the temporal and morphology streams using a dual attention mechanism, which operates across both channel and spatial dimensions to selectively emphasize informative features from each stream, thereby enhancing the representation learning for accurate ECG segmentation. On the LUDB and QT datasets, ECG-SCFNet achieves high performance, with F1-scores of 97.83% and 97.80%, respectively. Crucially, it maintains robust performance under challenging noise conditions on these datasets, with 88.49% and 86.25% F1-scores, showing significantly improved noise robustness compared to other methods and demonstrating exceptional robustness and precise boundary localization for clinical ECG analysis. Full article
Show Figures

Figure 1

25 pages, 9710 KB  
Article
SCS-YOLO: A Lightweight Cross-Scale Detection Network for Sugarcane Surface Cracks with Dynamic Perception
by Meng Li, Xue Ding, Jinliang Wang and Rongxiang Luo
AgriEngineering 2025, 7(10), 321; https://doi.org/10.3390/agriengineering7100321 - 1 Oct 2025
Viewed by 418
Abstract
Detecting surface cracks on sugarcane is a critical step in ensuring product quality control, with detection precision directly impacting raw material screening efficiency and economic benefits in the sugar industry. Traditional methods face three core challenges: (1) complex background interference complicates texture feature [...] Read more.
Detecting surface cracks on sugarcane is a critical step in ensuring product quality control, with detection precision directly impacting raw material screening efficiency and economic benefits in the sugar industry. Traditional methods face three core challenges: (1) complex background interference complicates texture feature extraction; (2) variable crack scales limit models’ cross-scale feature generalization capabilities; and (3) high computational complexity hinders deployment on edge devices. To address these issues, this study proposes a lightweight sugarcane surface crack detection model, SCS-YOLO (Surface Cracks on Sugarcane-YOLO), based on the YOLOv10 architecture. This model incorporates three key technical innovations. First, the designed RFAC2f module (Receptive-Field Attentive CSP Bottleneck with Dual Convolution) significantly enhances feature representation capabilities in complex backgrounds through dynamic receptive field modeling and multi-branch feature processing/fusion mechanisms. Second, the proposed DSA module (Dynamic SimAM Attention) achieves adaptive spatial optimization of cross-layer crack features by integrating dynamic weight allocation strategies with parameter-free spatial attention mechanisms. Finally, the DyHead detection head employs a dynamic feature optimization mechanism to reduce parameter count and computational complexity. Experiments demonstrate that on the Sugarcane Crack Dataset v3.1, compared to the baseline model YOLOv10, our model achieves mAP50:95 to 71.8% (up 2.1%). Simultaneously, it achieves significant reductions in parameter count (down 19.67%) and computational load (down 11.76%), while boosting FPS to 122 to meet real-time detection requirements. Considering the multiple dimensions of precision indicators, complexity indicators, and FPS comprehensively, the SCS—YOLO detection framework proposed in this study provides a feasible technical reference for the intelligent detection of sugarcane quality in the raw materials of the sugar industry. Full article
Show Figures

Figure 1

Back to TopTop