Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (518)

Search Parameters:
Keywords = cross-entropy loss

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
25 pages, 7449 KB  
Article
Influence of Volumetric Geometry on Meteorological Time Series Measurements: Fractality and Thermal Flows
by Patricio Pacheco Hernández, Gustavo Navarro Ahumada, Eduardo Mera Garrido and Diego Zemelman de la Cerda
Fractal Fract. 2025, 9(10), 639; https://doi.org/10.3390/fractalfract9100639 - 30 Sep 2025
Abstract
This work analyzes the behavior of the boundary layer subjected to stresses by obstacles using hourly measurements, in the form of time series, of meteorological variables (temperature (T), relative humidity (RH), and magnitude of the wind speed (WS)) in a given period. The [...] Read more.
This work analyzes the behavior of the boundary layer subjected to stresses by obstacles using hourly measurements, in the form of time series, of meteorological variables (temperature (T), relative humidity (RH), and magnitude of the wind speed (WS)) in a given period. The study region is Santiago, the capital of Chile. The measurement location is in a rugged basin geography with a nearly pristine atmospheric environment. The time series are analyzed through chaos theory, demonstrating that they are chaotic through the calculation of the parameters Lyapunov exponent (λ > 0), correlation dimension (DC < 5), Kolmogorov entropy (SK > 0), Hurst exponent (0.5 < H < 1), and Lempel–Ziv complexity (LZ > 0). These series are simultaneous measurements of the variables of interest, before and after, of three different volumetric geometries arranged as obstacles: a parallelepiped, a cylinder, and a miniature mountain. The three geometries are subject to the influence of the wind and present the same cross-sectional area facing the measuring instruments oriented in the same way. The entropies calculated for each variable in each geometry are compared. It is demonstrated, in a first approximation, that volumetric geometry impacts the magnitude of the entropic fluxes associated with the measured variables, which can affect micrometeorology and, by extension, the climate in general. Furthermore, the study examines which geometry favors greater information loss or greater fractality in the measured variables. Full article
(This article belongs to the Special Issue Fractals in Earthquake and Atmospheric Science)
Show Figures

Figure 1

25 pages, 9804 KB  
Article
GLFNet: Attention Mechanism-Based Global–Local Feature Fusion Network for Micro-Expression Recognition
by Meng Zhang, Long Yao, Wenzhong Yang and Yabo Yin
Entropy 2025, 27(10), 1023; https://doi.org/10.3390/e27101023 - 28 Sep 2025
Abstract
Micro-expressions are extremely subtle and short-lived facial muscle movements that often reveal an individual’s genuine emotions. However, micro-expression recognition (MER) remains highly challenging due to its short duration, low motion intensity, and the imbalanced distribution of training samples. To address these issues, this [...] Read more.
Micro-expressions are extremely subtle and short-lived facial muscle movements that often reveal an individual’s genuine emotions. However, micro-expression recognition (MER) remains highly challenging due to its short duration, low motion intensity, and the imbalanced distribution of training samples. To address these issues, this paper proposes a Global–Local Feature Fusion Network (GLFNet) to effectively extract discriminative features for MER. Specifically, GLFNet consists of three core modules: the Global Attention (LA) module, which captures subtle variations across the entire facial region; the Local Block (GB) module, which partitions the feature map into four non-overlapping regions to emphasize salient local movements while suppressing irrelevant information; and the Adaptive Feature Fusion (AFF) module, which employs an attention mechanism to dynamically adjust channel-wise weights for efficient global–local feature integration. In addition, a class-balanced loss function is introduced to replace the conventional cross-entropy loss, mitigating the common issue of class imbalance in micro-expression datasets. Extensive experiments are conducted on three benchmark databases, SMIC, CASME II, and SAMM, under two evaluation protocols. The experimental results demonstrate that under the Composite Database Evaluation protocol, GLFNet consistently outperforms existing state-of-the-art methods in overall performance. Specifically, the unweighted F1-scores on the Combined, SAMM, CASME II, and SMIC datasets are improved by 2.49%, 2.02%, 0.49%, and 4.67%, respectively, compared to the current best methods. These results strongly validate the effectiveness and superiority of the proposed global–local feature fusion strategy in micro-expression recognition tasks. Full article
Show Figures

Figure 1

18 pages, 3172 KB  
Article
Enhancing Confidence and Interpretability of a CNN-Based Wafer Defect Classification Model Using Temperature Scaling and LIME
by Jieun Lee, Yeonwoo Ju, Junho Lim, Sungmin Hong, Soo-Whang Baek and Jonghwan Lee
Micromachines 2025, 16(9), 1057; https://doi.org/10.3390/mi16091057 - 17 Sep 2025
Viewed by 360
Abstract
Accurate classification of defects in the semiconductor manufacturing process is critical for improving yield and ensuring quality. While previous works have mainly focused on improving classification accuracy, we propose a model that can simultaneously assess accuracy, prediction confidence, and interpretability in wafer defect [...] Read more.
Accurate classification of defects in the semiconductor manufacturing process is critical for improving yield and ensuring quality. While previous works have mainly focused on improving classification accuracy, we propose a model that can simultaneously assess accuracy, prediction confidence, and interpretability in wafer defect classification. To solve the class imbalance problem, we used a weighted cross-entropy loss function and convolutional neural network–based model to achieve a high accuracy of 97.8% on the test dataset and applied a temperature-scaling technique to enhance confidence. Furthermore, by simultaneously employing local interpretable model-agnostic explanations and gradient-weighted class activation mapping, the rationale for the predictions of the model was visualized, allowing users to understand the decision-making process of the model from various perspectives. This research can provide a direction for the next generation of intelligent quality management systems by enhancing the applicability of the proposed model in actual semiconductor production sites through explainable predictions. Full article
(This article belongs to the Special Issue Semiconductor and Energy Materials and Processing Technology)
Show Figures

Figure 1

21 pages, 7692 KB  
Article
Deployable Deep Learning Models for Crack Detection: Efficiency, Interpretability, and Severity Estimation
by Amna Altaf, Adeel Mehmood, Massimo Leonardo Filograno, Soltan Alharbi and Jamshed Iqbal
Buildings 2025, 15(18), 3362; https://doi.org/10.3390/buildings15183362 - 17 Sep 2025
Viewed by 462
Abstract
Concrete infrastructure inspection is essential for maintaining the safety and longevity of urban environments. Traditional manual crack detection methods are labor-intensive, inconsistent, and difficult to scale. Recent advancements in deep learning and computer vision offer automated alternatives, particularly when deployed via unmanned aerial [...] Read more.
Concrete infrastructure inspection is essential for maintaining the safety and longevity of urban environments. Traditional manual crack detection methods are labor-intensive, inconsistent, and difficult to scale. Recent advancements in deep learning and computer vision offer automated alternatives, particularly when deployed via unmanned aerial vehicles (UAVs) for enhanced coverage and flexibility. However, achieving real-time performance on embedded systems requires models that are not only accurate but also lightweight and computationally efficient. This study presents CrackDetect-Lite, a comparative analysis of three deep learning architectures for binary crack detection using the SDNET2018 benchmark dataset: CNNSimple (a custom lightweight model), RSNet (a shallow residual network), and MobileVNet (a fine-tuned MobileNetV2). Class imbalance was addressed using a weighted cross-entropy loss function, and models were evaluated across multiple criteria including classification accuracy, crack-class F1-score, inference latency, and model size. Among the models, MobileVNet achieved the best balance between detection performance and deployability, with an accuracy of 90.5% and a crack F1-score of 0.73, while maintaining a low computational footprint suitable for UAV-based deployment. These findings demonstrate that carefully selected lightweight CNN architectures can deliver reliable, real-time crack detection, supporting scalable and autonomous infrastructure monitoring in smart city systems. Full article
(This article belongs to the Special Issue AI in Construction: Automation, Optimization, and Safety)
Show Figures

Figure 1

21 pages, 3788 KB  
Article
Research on BBHL Model Based on Hybrid Loss Optimization for Fake News Detection
by Minghu Tang, Jiayi Zhang, Xuan Bu, Junjie Wang and Peng Luo
Appl. Sci. 2025, 15(18), 10028; https://doi.org/10.3390/app151810028 - 13 Sep 2025
Viewed by 612
Abstract
With the rapid development of social media, the spread of fake news has become a significant issue affecting social stability. To address the problems of incomplete feature extraction and simplistic loss function design in traditional fake news detection, this paper proposes a BBHL [...] Read more.
With the rapid development of social media, the spread of fake news has become a significant issue affecting social stability. To address the problems of incomplete feature extraction and simplistic loss function design in traditional fake news detection, this paper proposes a BBHL model based on hybrid loss optimization. The model achieves deep extraction of text features by integrating BERT, Bi-LSTM, and attention mechanisms, and innovatively fuses binary cross-entropy (BCE) loss with contrastive loss to enhance feature discriminability and the model’s generalization ability. Experiments on the Weibo, Twitter, and Pheme datasets demonstrate that the BBHL model significantly outperforms baseline models such as EANN and MCNN in metrics including accuracy and F1-score. Ablation experiments verify the effectiveness of contrastive loss, providing a robust and generalizable solution for fake news detection. Full article
(This article belongs to the Special Issue Natural Language Processing in the Era of Artificial Intelligence)
Show Figures

Figure 1

21 pages, 4721 KB  
Article
Automated Brain Tumor MRI Segmentation Using ARU-Net with Residual-Attention Modules
by Erdal Özbay and Feyza Altunbey Özbay
Diagnostics 2025, 15(18), 2326; https://doi.org/10.3390/diagnostics15182326 - 13 Sep 2025
Viewed by 438
Abstract
Background/Objectives: Accurate segmentation of brain tumors in Magnetic Resonance Imaging (MRI) scans is critical for diagnosis and treatment planning due to their life-threatening nature. This study aims to develop a robust and automated method capable of precisely delineating heterogeneous tumor regions while improving [...] Read more.
Background/Objectives: Accurate segmentation of brain tumors in Magnetic Resonance Imaging (MRI) scans is critical for diagnosis and treatment planning due to their life-threatening nature. This study aims to develop a robust and automated method capable of precisely delineating heterogeneous tumor regions while improving segmentation accuracy and generalization. Methods: We propose Attention Res-UNet (ARU-Net), a novel Deep Learning (DL) architecture integrating residual connections, Adaptive Channel Attention (ACA), and Dimensional-space Triplet Attention (DTA) modules. The encoding module efficiently extracts and refines relevant feature information by applying ACA to the lower layers of convolutional and residual blocks. The DTA is fixed to the upper layers of the decoding module, decoupling channel weights to better extract and fuse multi-scale features, enhancing both performance and efficiency. Input MRI images are pre-processed using Contrast Limited Adaptive Histogram Equalization (CLAHE) for contrast enhancement, denoising filters, and Linear Kuwahara filtering to preserve edges while smoothing homogeneous regions. The network is trained using categorical cross-entropy loss with the Adam optimizer on the BTMRII dataset, and comparative experiments are conducted against baseline U-Net, DenseNet121, and Xception models. Performance is evaluated using accuracy, precision, recall, F1-score, Dice Similarity Coefficient (DSC), and Intersection over Union (IoU) metrics. Results: Baseline U-Net showed significant performance gains after adding residual connections and ACA modules, with DSC improving by approximately 3.3%, accuracy by 3.2%, IoU by 7.7%, and F1-score by 3.3%. ARU-Net further enhanced segmentation performance, achieving 98.3% accuracy, 98.1% DSC, 96.3% IoU, and a superior F1-score, representing additional improvements of 1.1–2.0% over the U-Net + Residual + ACA variant. Visualizations confirmed smoother boundaries and more precise tumor contours across all six tumor classes, highlighting ARU-Net’s ability to capture heterogeneous tumor structures and fine structural details more effectively than both baseline U-Net and other conventional DL models. Conclusions: ARU-Net, combined with an effective pre-processing strategy, provides a highly reliable and precise solution for automated brain tumor segmentation. Its improvements across multiple evaluation metrics over U-Net and other conventional models highlight its potential for clinical application and contribute novel insights to medical image analysis research. Full article
(This article belongs to the Special Issue Advances in Functional and Structural MR Image Analysis)
Show Figures

Figure 1

23 pages, 10375 KB  
Article
Extraction of Photosynthetic and Non-Photosynthetic Vegetation Cover in Typical Grasslands Using UAV Imagery and an Improved SegFormer Model
by Jie He, Xiaoping Zhang, Weibin Li, Du Lyu, Yi Ren and Wenlin Fu
Remote Sens. 2025, 17(18), 3162; https://doi.org/10.3390/rs17183162 - 12 Sep 2025
Viewed by 401
Abstract
Accurate monitoring of the coverage and distribution of photosynthetic (PV) and non-photosynthetic vegetation (NPV) in the grasslands of semi-arid regions is crucial for understanding the environment and addressing climate change. However, the extraction of PV and NPV information from Unmanned Aerial Vehicle (UAV) [...] Read more.
Accurate monitoring of the coverage and distribution of photosynthetic (PV) and non-photosynthetic vegetation (NPV) in the grasslands of semi-arid regions is crucial for understanding the environment and addressing climate change. However, the extraction of PV and NPV information from Unmanned Aerial Vehicle (UAV) remote sensing imagery is often hindered by challenges such as low extraction accuracy and blurred boundaries. To overcome these limitations, this study proposed an improved semantic segmentation model, designated SegFormer-CPED. The model was developed based on the SegFormer architecture, incorporating several synergistic optimizations. Specifically, a Convolutional Block Attention Module (CBAM) was integrated into the encoder to enhance early-stage feature perception, while a Polarized Self-Attention (PSA) module was embedded to strengthen contextual understanding and mitigate semantic loss. An Edge Contour Extraction Module (ECEM) was introduced to refine boundary details. Concurrently, the Dice Loss function was employed to replace the Cross-Entropy Loss, thereby more effectively addressing the class imbalance issue and significantly improving both the segmentation accuracy and boundary clarity of PV and NPV. To support model development, a high-quality PV and NPV segmentation dataset for Hengshan grassland was also constructed. Comprehensive experimental results demonstrated that the proposed SegFormer-CPED model achieved state-of-the-art performance, with a mIoU of 93.26% and an F1-score of 96.44%. It significantly outperformed classic architectures and surpassed all leading frameworks benchmarked here. Its high-fidelity maps can bridge field surveys and satellite remote sensing. Ablation studies verified the effectiveness of each improved module and its synergistic interplay. Moreover, this study successfully utilized SegFormer-CPED to perform fine-grained monitoring of the spatiotemporal dynamics of PV and NPV in the Hengshan grassland, confirming that the model-estimated fPV and fNPV were highly correlated with ground survey data. The proposed SegFormer-CPED model provides a robust and effective solution for the precise, semi-automated extraction of PV and NPV from high-resolution UAV imagery. Full article
Show Figures

Graphical abstract

15 pages, 2578 KB  
Article
Effects of Composite Cross-Entropy Loss on Adversarial Robustness
by Ning Ding and Knut Möller
Electronics 2025, 14(17), 3529; https://doi.org/10.3390/electronics14173529 - 4 Sep 2025
Viewed by 463
Abstract
Convolutional neural networks (CNNs) can efficiently extract image features and perform corresponding classification. Typically, the CNN architecture uses the softmax layer to map the extracted features to classification probabilities, and the cost function used for training is the cross-entropy loss. In this paper, [...] Read more.
Convolutional neural networks (CNNs) can efficiently extract image features and perform corresponding classification. Typically, the CNN architecture uses the softmax layer to map the extracted features to classification probabilities, and the cost function used for training is the cross-entropy loss. In this paper, we evaluate the influence of a number of representative composite cross-entropy loss functions on the learned feature space at the fully connected layer, when a target classification is introduced into a multi-class classification task. In addition, the accuracy and robustness of CNN models trained with different composite cross-entropy loss functions are investigated. Improved robustness is achieved by changing the loss between the input and the target classification. Preliminary experiments were conducted using ResNet-50 on the Cholec80 dataset for surgical tool recognition. Furthermore, the model trained with the proposed composite cross-entropy loss incorporating another target all-one classification demonstrates a 31% peak improvement in adversarial robustness. Adversarial training with target adversarial samples yields 80% robustness against PGD attack. This investigation shows that the careful choice of the loss function can improve the robustness of CNN models. Full article
(This article belongs to the Special Issue Convolutional Neural Networks and Vision Applications, 4th Edition)
Show Figures

Figure 1

18 pages, 568 KB  
Article
Beyond Cross-Entropy: Discounted Least Information Theory of Entropy (DLITE) Loss and the Impact of Loss Functions on AI-Driven Named Entity Recognition
by Sonia Pascua, Michael Pan and Weimao Ke
Information 2025, 16(9), 760; https://doi.org/10.3390/info16090760 - 2 Sep 2025
Viewed by 486
Abstract
Loss functions play a significant role in shaping model behavior in machine learning, yet their design implications remain underexplored in natural language processing tasks such as Named Entity Recognition (NER). This study investigates the performance and optimization behavior of five loss functions—L1, L2, [...] Read more.
Loss functions play a significant role in shaping model behavior in machine learning, yet their design implications remain underexplored in natural language processing tasks such as Named Entity Recognition (NER). This study investigates the performance and optimization behavior of five loss functions—L1, L2, Cross-Entropy (CE), KL Divergence (KL), and the proposed DLITE (Discounted Least Information Theory of Entropy) Loss—within transformer-based NER models. DLITE introduces a bounded, entropy-discounting approach to penalization, prioritizing recall and training stability, especially under noisy or imbalanced data conditions. We conducted empirical evaluations across three benchmark NER datasets: Basic NER, CoNLL-2003, and the Broad Twitter Corpus. While CE and KL achieved the highest weighted F1-scores in clean datasets, DLITE Loss demonstrated distinct advantages in macro recall, precision–recall balance, and convergence stability—particularly in noisy environments. Our findings suggest that the choice of loss function should align with application-specific priorities, such as minimizing false negatives or managing uncertainty. DLITE adds a new dimension to model design by enabling more measured predictions, making it a valuable alternative in high-stakes or real-world NLP deployments. Full article
Show Figures

Figure 1

17 pages, 1149 KB  
Article
IP Spoofing Detection Using Deep Learning
by İsmet Kaan Çekiş, Buğra Ayrancı, Fezayim Numan Salman and İlker Özçelik
Appl. Sci. 2025, 15(17), 9508; https://doi.org/10.3390/app15179508 - 29 Aug 2025
Viewed by 601
Abstract
IP spoofing is a critical component in many cyberattacks, enabling attackers to evade detection and conceal their identities. This study rigorously compares eight deep learning models—LSTM, GRU, CNN, MLP, DNN, RNN, ResNet1D, and xLSTM—for their efficacy in detecting IP spoofing attacks. Overfitting was [...] Read more.
IP spoofing is a critical component in many cyberattacks, enabling attackers to evade detection and conceal their identities. This study rigorously compares eight deep learning models—LSTM, GRU, CNN, MLP, DNN, RNN, ResNet1D, and xLSTM—for their efficacy in detecting IP spoofing attacks. Overfitting was mitigated through techniques such as dropout, early stopping, and normalization. Models were trained using binary cross-entropy loss and the Adam optimizer. Performance was assessed via accuracy, precision, recall, F1 score, and inference time, with each model executed a total of 15 times to account for stochastic variability. Results indicate a powerful performance across all models, with LSTM and GRU demonstrating superior detection efficacy. After ONNX conversion, the MLP and DNN models retained their performance while achieving significant reductions in inference time, miniaturized model sizes, and platform independence. These advancements facilitated the effective utilization of the developed systems in real-time network security applications. The comprehensive performance metrics presented are crucial for selecting optimal IP spoofing detection strategies tailored to diverse application requirements, serving as a valuable reference for network anomaly monitoring and targeted attack detection. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

16 pages, 1464 KB  
Article
Transient Stability Assessment of Power Systems Built upon a Deep Spatio-Temporal Feature Extraction Network
by Yu Nan, Meng Tong, Zhenzhen Kong, Huichao Zhao and Yadong Zhao
Energies 2025, 18(17), 4547; https://doi.org/10.3390/en18174547 - 27 Aug 2025
Viewed by 490
Abstract
The rapid and accurate identification of power system transient stability status is a fundamental prerequisite for ensuring the secure and reliable operation of large-scale power grids. With the increasing complexity and heterogeneity of modern power system components, system nonlinearity has grown significantly, rendering [...] Read more.
The rapid and accurate identification of power system transient stability status is a fundamental prerequisite for ensuring the secure and reliable operation of large-scale power grids. With the increasing complexity and heterogeneity of modern power system components, system nonlinearity has grown significantly, rendering traditional time-domain simulation and direct methods unable to meet accuracy and efficiency requirements simultaneously. To further improve the prediction accuracy of power system transient stability and provide more refined assessment results, this paper integrates deep learning with power system transient stability and proposes a transient stability assessment of power systems built upon a deep spatio-temporal feature extraction network method. First, a spatio-temporal feature extraction module is constructed by combining an improved graph attention network with a residual bidirectional temporal convolutional network, aiming to capture the spatial and bidirectional temporal characteristics of transient stability data. Second, a classification module is developed using the Kolmogorov–Arnold network to establish the mapping relationship between spatio-temporal features and transient stability states. This enables the accurate determination of the system’s transient stability status within a short time after fault occurrence. Finally, a weighted cross-entropy loss function is employed to address the issue of low prediction accuracy caused by the imbalanced sample distribution in the evaluation model. The feasibility, effectiveness, and superiority of the proposed method are validated through tests on the New England 10-machine 39-bus system and the NPCC 48-machine 140-bus system. Full article
Show Figures

Figure 1

25 pages, 3904 KB  
Article
Physics-Guided Multi-Representation Learning with Quadruple Consistency Constraints for Robust Cloud Detection in Multi-Platform Remote Sensing
by Qing Xu, Zichen Zhang, Guanfang Wang and Yunjie Chen
Remote Sens. 2025, 17(17), 2946; https://doi.org/10.3390/rs17172946 - 25 Aug 2025
Cited by 1 | Viewed by 733
Abstract
With the rapid expansion of multi-platform remote sensing applications, cloud contamination significantly impedes cross-platform data utilization. Current cloud detection methods face critical technical challenges in cross-platform settings, including neglect of atmospheric radiative transfer mechanisms, inadequate multi-scale structural decoupling, high intra-class variability coupled with [...] Read more.
With the rapid expansion of multi-platform remote sensing applications, cloud contamination significantly impedes cross-platform data utilization. Current cloud detection methods face critical technical challenges in cross-platform settings, including neglect of atmospheric radiative transfer mechanisms, inadequate multi-scale structural decoupling, high intra-class variability coupled with inter-class similarity, cloud boundary ambiguity, cross-modal feature inconsistency, and noise propagation in pseudo-labels within semi-supervised frameworks. To address these issues, we introduce a Physics-Guided Multi-Representation Network (PGMRN) that adopts a student–teacher architecture and fuses tri-modal representations—Pseudo-NDVI, structural, and textural features—via atmospheric priors and intrinsic image decomposition. Specifically, PGMRN first incorporates an InfoNCE contrastive loss to enhance intra-class compactness and inter-class discrimination while preserving physical consistency; subsequently, a boundary-aware regional adaptive weighted cross-entropy loss integrates PA-CAM confidence with distance transforms to refine edge accuracy; furthermore, an Uncertainty-Aware Quadruple Consistency Propagation (UAQCP) enforces alignment across structural, textural, RGB, and physical modalities; and finally, a dynamic confidence-screening mechanism that couples PA-CAM with information entropy and percentile-based thresholding robustly refines pseudo-labels. Extensive experiments on four benchmark datasets demonstrate that PGMRN achieves state-of-the-art performance, with Mean IoU values of 70.8% on TCDD, 79.0% on HRC_WHU, and 83.8% on SWIMSEG, outperforming existing methods. Full article
Show Figures

Figure 1

22 pages, 6265 KB  
Article
A Multi-Level Fusion Framework for Bearing Fault Diagnosis Using Multi-Source Information
by Xiaojun Deng, Yuanhao Sun, Lin Li and Xia Peng
Processes 2025, 13(8), 2657; https://doi.org/10.3390/pr13082657 - 21 Aug 2025
Cited by 1 | Viewed by 533
Abstract
Rotating machinery is essential to modern industrial systems, where rolling bearings play a critical role in ensuring mechanical stability and operational efficiency. Failures in bearings can result in serious safety risks and significant financial losses, which highlights the need for accurate and robust [...] Read more.
Rotating machinery is essential to modern industrial systems, where rolling bearings play a critical role in ensuring mechanical stability and operational efficiency. Failures in bearings can result in serious safety risks and significant financial losses, which highlights the need for accurate and robust methods for diagnosing bearing faults. Traditional diagnostic methods relying on single-source data often fail to fully leverage the rich information provided by multiple sensors and are more prone to performance degradation under noisy conditions. Therefore, this paper proposes a novel bearing fault diagnosis method based on a multi-level fusion framework. First, the Symmetrized Dot Pattern (SDP) method is applied to fuse multi-source signals into unified SDP images, enabling effective fusion at the data level. Then, a combination of RepLKNet and Bidirectional Gated Recurrent Unit (BiGRU) networks extracts multi-modal features, which are then fused through a cross-attention mechanism to enhance feature representation. Finally, information entropy is utilized to assess the reliability of each feature channel, enabling dynamic weighting to further strengthen model robustness. The experiments conducted on public datasets and noise-augmented datasets demonstrate that the proposed method significantly surpasses other single-source and multi-source data fusion models in terms of diagnostic accuracy and robustness to noise. Full article
(This article belongs to the Section Process Control and Monitoring)
Show Figures

Figure 1

29 pages, 12228 KB  
Article
Conditional Domain Adaptation with α-Rényi Entropy Regularization and Noise-Aware Label Weighting
by Diego Armando Pérez-Rosero, Andrés Marino Álvarez-Meza and German Castellanos-Dominguez
Mathematics 2025, 13(16), 2602; https://doi.org/10.3390/math13162602 - 14 Aug 2025
Viewed by 485
Abstract
Domain adaptation is a key approach to ensure that artificial intelligence models maintain reliable performance when facing distributional shifts between training (source) and testing (target) domains. However, existing methods often struggle to simultaneously preserve domain-invariant representations and discriminative class structures, particularly in the [...] Read more.
Domain adaptation is a key approach to ensure that artificial intelligence models maintain reliable performance when facing distributional shifts between training (source) and testing (target) domains. However, existing methods often struggle to simultaneously preserve domain-invariant representations and discriminative class structures, particularly in the presence of complex covariate shifts and noisy pseudo-labels in the target domain. In this work, we introduce Conditional Rényi α-Entropy Domain Adaptation, named CREDA, a novel deep learning framework for domain adaptation that integrates kernel-based conditional alignment with a differentiable, matrix-based formulation of Rényi’s quadratic entropy. The proposed method comprises three main components: (i) a deep feature extractor that learns domain-invariant representations from labeled source and unlabeled target data; (ii) an entropy-weighted approach that down-weights low-confidence pseudo-labels, enhancing stability in uncertain regions; and (iii) a class-conditional alignment loss, formulated as a Rényi-based entropy kernel estimator, that enforces semantic consistency in the latent space. We validate CREDA on standard benchmark datasets for image classification, including Digits, ImageCLEF-DA, and Office-31, showing competitive performance against both classical and deep learning-based approaches. Furthermore, we employ nonlinear dimensionality reduction and class activation maps visualizations to provide interpretability, revealing meaningful alignment in feature space and offering insights into the relevance of individual samples and attributes. Experimental results confirm that CREDA improves cross-domain generalization while promoting accuracy, robustness, and interpretability. Full article
Show Figures

Figure 1

18 pages, 9486 KB  
Article
MCCSAN: Automatic Modulation Classification via Multiscale Complex Convolution and Spatiotemporal Attention Network
by Songchen Xu, Duona Zhang, Yuanyao Lu, Zhe Xing and Weikai Ma
Electronics 2025, 14(16), 3192; https://doi.org/10.3390/electronics14163192 - 11 Aug 2025
Viewed by 480
Abstract
Automatic Modulation Classification (AMC) is vital for adaptive wireless communication, yet it faces challenges in complex environments, including insufficient feature extraction, feature redundancy, and high interclass similarity among modulation schemes. To address these limitations, this paper proposes the Multiscale Complex Convolution Spatiotemporal Attention [...] Read more.
Automatic Modulation Classification (AMC) is vital for adaptive wireless communication, yet it faces challenges in complex environments, including insufficient feature extraction, feature redundancy, and high interclass similarity among modulation schemes. To address these limitations, this paper proposes the Multiscale Complex Convolution Spatiotemporal Attention Network (MCCSAN). In this work, we propose three key innovations tailored for AMC tasks: a multiscale complex convolutional module that directly processes raw I/Q sequences, preserving critical phase and amplitude information while extracting diverse signal features. A spatiotemporal attention mechanism dynamically weights temporal steps and feature channels to suppress redundancy and enhance discriminative feature focus. A combined loss function integrating cross-entropy and center loss improves intraclass compactness and interclass separability. Evaluated on the RML2018.01A dataset and RML2016.10A across SNR levels from −6 dB to 12 dB, MCCSAN achieves a state-of-the-art classification accuracy of 97.03% (peak) and an average accuracy improvement of 3.98% over leading methods. The study confirms that integrating complex-domain processing with spatiotemporal attention significantly enhances AMC performance. Full article
Show Figures

Figure 1

Back to TopTop