Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (156)

Search Parameters:
Keywords = catastrophic forgetting

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
21 pages, 4182 KB  
Article
Incremental Pavement Distress Classification in UAV-Based Remote Sensing via Analytic Geometric Alignment
by Quanziang Wang, Xin Li, Jiangjun Peng, Xixi Jia and Renzhen Wang
Remote Sens. 2026, 18(8), 1141; https://doi.org/10.3390/rs18081141 - 12 Apr 2026
Viewed by 216
Abstract
Automated pavement distress classification using high-resolution Unmanned Aerial Vehicle (UAV) imagery is pivotal for intelligent transportation systems. However, long-term UAV monitoring faces a continuous stream of evolving distress types and changing remote sensing background textures, necessitating Class-Incremental Learning (CIL) capabilities. Existing methods struggle [...] Read more.
Automated pavement distress classification using high-resolution Unmanned Aerial Vehicle (UAV) imagery is pivotal for intelligent transportation systems. However, long-term UAV monitoring faces a continuous stream of evolving distress types and changing remote sensing background textures, necessitating Class-Incremental Learning (CIL) capabilities. Existing methods struggle to balance stability and plasticity, especially under the severe storage limitations typical of local edge stations in air–ground collaborative systems. This data scarcity leads to catastrophic forgetting and confusion among fine-grained distress categories. To address these challenges, we propose a data-efficient approach named Analytic Geometric Alignment (AGA). Our framework mainly consists of three key components. First, to overcome the optimization gap between the feature extractor and the fixed geometric target, we introduce a Subspace-Aware Analytic Initialization (SAI) that computes a closed-form projection to instantly align the feature subspace with the ETF manifold before each task training. Second, on this aligned basis, a Decoupled Geometric Adapter (DGA) is incorporated to facilitate continuous non-linear adaptation to complex aerial textures. Finally, for stable incremental training, we design a Memory-Prioritized Regression (MPR) loss to enforce tighter geometric constraints on replay samples, significantly enhancing model stability. Extensive experiments on the UAV-PDD2023 dataset demonstrate that AGA significantly outperforms state-of-the-art methods, showcasing excellent robustness and data efficiency. Full article
Show Figures

Figure 1

41 pages, 2153 KB  
Review
A Review of Domain-Adaptive Continual Deep Learning Remaining Useful Life Estimation for Bearing Fault Prognosis Under Evolving Data Distributions
by Stamatis Apeiranthitis, Christos Drosos, Avraam Chatzopoulos, Michail Papoutsidakis and Evangellos Pallis
Machines 2026, 14(4), 412; https://doi.org/10.3390/machines14040412 - 8 Apr 2026
Viewed by 282
Abstract
Estimating remaining useful life (RUL) and predicting bearing faults based on data-driven models have become central components of modern Prognostics and Health Management (PHM) systems. Although deep learning models have demonstrated strong performance under controlled and stationary operating conditions, their reliability in real-world [...] Read more.
Estimating remaining useful life (RUL) and predicting bearing faults based on data-driven models have become central components of modern Prognostics and Health Management (PHM) systems. Although deep learning models have demonstrated strong performance under controlled and stationary operating conditions, their reliability in real-world industrial and marine environments is limited. In practice, operating conditions, sensor properties, and degradation mechanisms evolve continuously over time, leading to non-stationary and shifting data distributions that violate the assumptions of conventional static learning approaches. To address these challenges, two research areas have gained increasing attention: Domain Adaptation (DA), which aims to mitigate distribution discrepancies across operating conditions or machines, and Continual Learning (CL), which enables models to learn sequentially while mitigating catastrophic forgetting. However, existing studies often examine these paradigms in isolation, limiting their effectiveness in long-term deployments, where domain shifts and temporal evolution coexist. This paper presents a comprehensive and systematic review of data-driven methods for bearing fault prognosis and remaining useful life (RUL) prediction under evolving data distributions, adopting the framework of Domain-Adaptive Continual Learning (DACL). By jointly examining the DA and CL methods, this review analyses how these approaches have been individually and implicitly combined to cope with non-stationarity, knowledge retention, and limited label availability in practical PHM scenarios. We categorised existing methods, highlighted their underlying assumptions and limitations, and critically assessed their applicability to long-term, real-world monitoring systems. Furthermore, key open challenges, including scalability, robustness under sequential domain shifts, uncertainty handling, and plasticity–stability trade-offs, are identified, and research directions are outlined based on the identified limitations and practical deployment requirements of the proposed method. This review aims to establish a structured and critical reference framework for understanding the role of domain-adaptive CL in data-driven prognostics, clarifying current research trends, limitations, and open challenges in evolving data distributions. Full article
Show Figures

Figure 1

22 pages, 7355 KB  
Article
IAE-Net: Incremental Learning-Based Attention-Enhanced DenseNet for Robust Facial Emotion Recognition
by Haseeb Ali Khan and Jong-Ha Lee
Mathematics 2026, 14(6), 1023; https://doi.org/10.3390/math14061023 - 18 Mar 2026
Viewed by 290
Abstract
Facial emotion recognition (FER) is an important component of human–computer interaction and healthcare-oriented affective computing. However, reliable deployment remains difficult in unconstrained settings due to appearance and geometric variability (e.g., pose, illumination, and occlusion), demographic imbalance, and dataset bias. In practice, two additional [...] Read more.
Facial emotion recognition (FER) is an important component of human–computer interaction and healthcare-oriented affective computing. However, reliable deployment remains difficult in unconstrained settings due to appearance and geometric variability (e.g., pose, illumination, and occlusion), demographic imbalance, and dataset bias. In practice, two additional constraints frequently limit real-world FER systems: the computational overhead of heavy architectures and limited adaptability when data evolve over time, where sequential updates can cause catastrophic forgetting. To address these challenges, we propose the Incremental Attention-Enhanced Network (IAE-Net), a compact single-branch framework built on a DenseNet121 backbone and a cascaded refinement pipeline. The model incorporates Channel Attention (CA) to emphasize expression-relevant feature channels and suppress less informative responses, followed by a deformable attention module (DA) that reduces feature misalignment caused by non-rigid facial motion and pose shifts, thereby improving robustness under geometric variability. For continual deployment, IAE-Net supports class-incremental updates via weight transfer, exemplar replay, and knowledge distillation to improve retention during sequential learning. We evaluate IAE-Net on four widely used benchmarks, FER2013, FERPlus, KDEF, and AffectNet, covering both controlled and in-the-wild conditions under a unified training protocol. The proposed approach achieves accuracies of 79.15%, 92.03%, 99.48%, and 74.20% on FER2013, FERPlus, KDEF, and AffectNet, respectively, with balanced precision, recall, and F1-score trends. These results indicate that IAE-Net provides an efficient and extensible FER framework with potential utility in dynamic real-world and longitudinal healthcare-oriented applications. Full article
(This article belongs to the Special Issue Recent Advances and Applications of Artificial Neural Networks)
Show Figures

Figure 1

27 pages, 3391 KB  
Article
A Hybrid Federated–Incremental Learning Framework for Continuous Authentication in Zero-Trust Networks
by Jie Ji, Shi Qiu, Shengpeng Ye and Xin Liu
Future Internet 2026, 18(3), 154; https://doi.org/10.3390/fi18030154 - 16 Mar 2026
Viewed by 301
Abstract
Zero-trust architecture (ZTA) requires continuous and adaptive identity authentication to maintain security in dynamic environments. However, current federated learning (FL)-based authentication models often struggle to incorporate evolving attack patterns without experiencing catastrophic forgetting. Moreover, non-independent and identically distributed (non-IID) client data and concept [...] Read more.
Zero-trust architecture (ZTA) requires continuous and adaptive identity authentication to maintain security in dynamic environments. However, current federated learning (FL)-based authentication models often struggle to incorporate evolving attack patterns without experiencing catastrophic forgetting. Moreover, non-independent and identically distributed (non-IID) client data and concept drift frequently lead to degraded model robustness and personalization. To address these issues, this paper presents a hybrid learning framework that integrates federated learning with incremental learning (IL) for sustainable authentication. A Dynamic Weighted Federated Aggregation (DWFA) algorithm is developed to mitigate concept drift by adjusting aggregation weights in real time, ensuring that the global model adapts to changing data distributions. This approach enables continuous learning from distributed threat data while maintaining privacy and eliminating the need for historical data retention. Experimental results on real-world traffic datasets indicate that the proposed framework outperforms conventional FL baselines, reducing the overall error rate by approximately 56% and improving the detection rate for novel attack types by over 17.8%. Furthermore, the framework remains stable against performance decay while maintaining efficient communication overhead. This study provides an adaptive, privacy-preserving solution for identity authentication in zero-trust systems. Full article
(This article belongs to the Special Issue Cybersecurity in the Age of AI, IoT, and Edge Computing)
Show Figures

Graphical abstract

31 pages, 23615 KB  
Article
A Memory-Efficient Class-Incremental Learning Framework for Remote Sensing Scene Classification via Feature Replay
by Yunze Wei, Yuhan Liu, Ben Niu, Xiantai Xiang, Jingdun Lin, Yuxin Hu and Yirong Wu
Remote Sens. 2026, 18(6), 896; https://doi.org/10.3390/rs18060896 - 15 Mar 2026
Viewed by 382
Abstract
Most existing deep learning models for remote sensing scene classification (RSSC) adopt an offline learning paradigm, where all classes are jointly optimized on fixed-class datasets. In dynamic real-world scenarios with streaming data and emerging classes, such paradigms are inherently prone to catastrophic forgetting [...] Read more.
Most existing deep learning models for remote sensing scene classification (RSSC) adopt an offline learning paradigm, where all classes are jointly optimized on fixed-class datasets. In dynamic real-world scenarios with streaming data and emerging classes, such paradigms are inherently prone to catastrophic forgetting when models are incrementally trained on new data. Recently, a growing number of class-incremental learning (CIL) methods have been proposed to tackle these issues, some of which achieve promising performance by rehearsing training data from previous tasks. However, implementing such strategy in real-world scenarios is often challenging, as the requirement to store historical data frequently conflicts with strict memory constraints and data privacy protocols. To address these challenges, we propose a novel memory-efficient feature-replay CIL framework (FR-CIL) for RSSC that retains compact feature embeddings, rather than raw images, as exemplars for previously learned classes. Specifically, a progressive multi-scale feature enhancement (PMFE) module is proposed to alleviate representation ambiguity. It adopts a progressive construction scheme to enable fine-grained and interactive feature enhancement, thereby improving the model’s representation capability for remote sensing scenes. Then, a specialized feature calibration network (FCN) is trained in a transductive learning paradigm with manifold consistency regularization to adapt stored feature descriptors to the updated feature space, thereby effectively compensating for feature space drift and enabling a unified classifier. Following feature calibration, a bias rectification (BR) strategy is employed to mitigate prediction bias by exclusively optimizing the classifier on a balanced exemplar set. As a result, this memory-efficient CIL framework not only addresses data privacy concerns but also mitigates representation drift and classifier bias. Extensive experiments on public datasets demonstrate the effectiveness and robustness of the proposed method. Notably, FR-CIL outperforms the leading state-of-the-art CIL methods in mean accuracy by margins of 3.75%, 3.09%, and 2.82% on the six-task AID, seven-task RSI-CB256, and nine-task NWPU-45 datasets, respectively. At the same time, it reduces memory storage requirements by over 94.7%, highlighting its strong potential for real-world RSSC applications under strict memory constraints. Full article
Show Figures

Figure 1

47 pages, 6456 KB  
Article
A Disentangled Prototype-Driven Continual Learning Framework for Fault Diagnosis of Cotton Harvester Picking-Head Drivetrains Under Gradually Expanding Operating Conditions
by Huachao Jiao, Wenlei Sun, Hongwei Wang and Xiaojing Wan
Agriculture 2026, 16(5), 566; https://doi.org/10.3390/agriculture16050566 - 2 Mar 2026
Viewed by 312
Abstract
The picking-head drivetrain is a critical transmission component of cotton harvesters, and its fault condition monitoring and diagnosis are essential for ensuring stable and reliable operation of the equipment. In practical engineering applications, diagnostic models for picking-head drivetrains are typically initialized using data [...] Read more.
The picking-head drivetrain is a critical transmission component of cotton harvesters, and its fault condition monitoring and diagnosis are essential for ensuring stable and reliable operation of the equipment. In practical engineering applications, diagnostic models for picking-head drivetrains are typically initialized using data collected under a limited number of representative operating conditions. Although sufficient fault samples can often be obtained during the initial training stage, the coverage of operating conditions is inherently restricted. As the model is deployed and used in the field, fault samples collected under new operating conditions are gradually acquired in a stage-wise manner. How to stably update the diagnostic model while the operating-condition coverage continuously expands, and how to avoid performance degradation and catastrophic forgetting, remain critical challenges. To address these issues, this paper proposes a continual learning method, termed DP-CL (Disentangled Prototype-Driven Continual Learning), for fault diagnosis of cotton harvester picking-head drivetrains under gradually expanding operating conditions. The proposed method is built upon an explicit disentanglement of condition-invariant features and condition-specific features. Within a unified framework, three types of structured prototypes, including class prototypes, condition prototypes, and condition-aware class prototypes, are constructed to form a multi-level representation hierarchy. A prototype-driven structured update mechanism is then employed to impose stable constraints on fault-discriminative semantics across different operating conditions. In addition, an operating-condition similarity measurement based on condition-specific features is introduced, based on which a proportion-adaptive sample selection strategy is designed. This strategy enables controlled knowledge transfer and preservation of discriminative structures during multi-stage model updates. Experimental results obtained under a laboratory-constructed cumulative operating-condition expansion scenario demonstrate that the proposed method achieves superior performance in terms of overall performance retention, cross-stage stability, and resistance to performance degradation. Moreover, as the number of operating conditions increases, the proposed method maintains a relatively smooth performance variation trend, while preserving clear class structures and a controllable level of confusion. These results validate the effectiveness of the proposed approach for stable fault diagnosis under expanding operating-condition coverage. Full article
(This article belongs to the Section Agricultural Technology)
Show Figures

Figure 1

24 pages, 4218 KB  
Article
SD-IDD: Selective Distillation for Incremental Defect Detection
by Jing Li, Chenggang Dai, Xiaobin Wang and Chengjun Chen
Sensors 2026, 26(5), 1413; https://doi.org/10.3390/s26051413 - 24 Feb 2026
Viewed by 294
Abstract
Surface defects in industrial production are complex and diverse. Therefore, deep learning-based defect detection models must consistently adapt to newly emerging defect categories. The trained models generally suffer from catastrophic forgetting as they learn new defect categories. To address this issue, we propose [...] Read more.
Surface defects in industrial production are complex and diverse. Therefore, deep learning-based defect detection models must consistently adapt to newly emerging defect categories. The trained models generally suffer from catastrophic forgetting as they learn new defect categories. To address this issue, we propose a selective distillation for incremental defect detection (SD-IDD) model based on GFLv1. Specifically, three selective distillation strategies are proposed, including high-confidence classification distillation, dual-stage cascaded regression distillation, and Intersection over Union (IoU)-driven difficulty-aware feature distillation. The high-confidence classification distillation aims to preserve critical discriminative knowledge of old categories within semantic confusion regions of the classification head, reducing interference from low-value regions. Dual-stage cascaded regression distillation focuses on high-quality anchors through geometric prior coarse filtering and statistical fine filtering, utilizing IoU-weighted KL divergence distillation loss to accurately transfer localization knowledge. IoU-driven difficulty-aware feature distillation adaptively allocates distillation resources, prioritizing features of high-difficulty targets. These selective distillation strategies significantly mitigate catastrophic forgetting while enhancing the detection accuracy of new classes, without requiring access to old training samples. Experimental results demonstrate that SD-IDD achieves superior performance, with mAP_old of 58.2% and 99.3%, mAP_new of 69.0% and 97.3%, and mAP_all of 63.6% and 98.3% on the NEU-DET and DeepPCB datasets, respectively, surpassing existing incremental detection methods. Full article
(This article belongs to the Section Fault Diagnosis & Sensors)
Show Figures

Figure 1

27 pages, 2216 KB  
Article
Exploring a New Architecture for Efficient Parameter Fine-Tuning in SLoRA Multitasking Scenarios
by Ce Shi and Jin-Woo Jung
Appl. Sci. 2026, 16(5), 2174; https://doi.org/10.3390/app16052174 - 24 Feb 2026
Viewed by 423
Abstract
Propose an enhanced LoRA (Low-Rank Adaptation) MoE (mixed expert) architecture, SLoRA (Enhanced LoRA MoE Architecture), aimed at addressing the key problem of efficient parameter fine-tuning in multitasking scenarios. Given the high cost of traditional full fine-tuning as the parameter size of visual language [...] Read more.
Propose an enhanced LoRA (Low-Rank Adaptation) MoE (mixed expert) architecture, SLoRA (Enhanced LoRA MoE Architecture), aimed at addressing the key problem of efficient parameter fine-tuning in multitasking scenarios. Given the high cost of traditional full fine-tuning as the parameter size of visual language models increases, and the limitations of LoRA as a popular PEFT (parameter-efficient fine-tuning) method in multitasking, such as inadequate adaptability and difficulty in capturing complex task patterns, as well as the catastrophic forgetting and knowledge fragmentation challenges faced by existing research on integrating mixed expert (MoE) mechanisms into LoRA, SLoRA utilizes orthogonal constraint optimization to reduce disturbance to existing knowledge through constraint solution space initialization, alleviating catastrophic forgetting (old task accuracy retention rate reaches 92.4%, 16.1% higher than LoRA), and an optimized MoE structure that includes general experts (retaining pre-trained knowledge) and task-specific experts (dynamic routing adaptation tasks) to enhance multitask adaptability. Experimental results show that in commonsense reasoning tasks, SLoRA’s accuracy is 9.0% higher than LoRA and 3.7% higher than AdaLoRA on the WSC dataset, and its F1 score is 7.7% higher than LoRA and 2.9% higher than AdaLoRA on the CommonsenseQA dataset; in multimodal tasks, its average score is up to 15.3% higher than LoRA, demonstrating significant advantages over existing methods. Full article
(This article belongs to the Special Issue Advancements in Deep Learning and Its Applications)
Show Figures

Figure 1

25 pages, 2131 KB  
Article
Symmetry-Aware Continual Learning for Dynamic Dimensional Multivariate Time Series Forecasting: Integrating Redundancy Clustering and Multi-LoRA Adapters
by Liyang Qin, Xiaoli Wang and Yulong Wang
Symmetry 2026, 18(2), 363; https://doi.org/10.3390/sym18020363 - 15 Feb 2026
Viewed by 425
Abstract
Continual learning of multivariate time series (MTS) forecasting is critical for process industries where working condition drift is frequent due to the variation in the feed properties and other factors. However, existing continual learning methods struggle with dynamic input dimension changes, and the [...] Read more.
Continual learning of multivariate time series (MTS) forecasting is critical for process industries where working condition drift is frequent due to the variation in the feed properties and other factors. However, existing continual learning methods struggle with dynamic input dimension changes, and the lack of symmetry-aware feature and dimension regulation further exacerbates the interference of irrelevant variables and dimensional inconsistency. To overcome this problem, G-MLoRA, a continual learning method based on dynamic redundancy clustering and multiple low-rank adapters, is proposed in this paper. This method can effectively enhance the network’s capability for prediction of multivariate time series under dynamic input dimensions. First, it groups MTS via Wasserstein distance-K-means clustering to reduce irrelevant variable interference. Second, each group is assigned to an exclusive LoRA adapter, with pre-trained backbone weights frozen during fine-tuning to lower complexity and mitigate catastrophic forgetting. Third, mini-batch gradient accumulation enables reuse of inconsistent-dimensional historical knowledge. Extensive experiments on two real grinding classification datasets show G-MLoRA outperforms baselines in new/historical knowledge compatibility, especially under dynamic dimensions. Full article
(This article belongs to the Section Computer)
Show Figures

Figure 1

19 pages, 3571 KB  
Article
Few-Shot Class-Incremental SAR Target Recognition Based on Dynamic Task-Adaptive Classifier
by Dan Li, Feng Zhao, Yong Li and Wei Cheng
Remote Sens. 2026, 18(3), 527; https://doi.org/10.3390/rs18030527 - 6 Feb 2026
Viewed by 530
Abstract
Current synthetic aperture radar automatic target recognition (SAR ATR) tasks face challenges including limited training samples and poor generalization capability to novel classes. To address these issues, few-shot class-incremental learning (FSCIL) has emerged as a promising research direction. Few-shot learning facilitates the expedited [...] Read more.
Current synthetic aperture radar automatic target recognition (SAR ATR) tasks face challenges including limited training samples and poor generalization capability to novel classes. To address these issues, few-shot class-incremental learning (FSCIL) has emerged as a promising research direction. Few-shot learning facilitates the expedited adaptation to novel tasks utilizing a limited number of labeled samples, whereas incremental learning concentrates on the continuous refinement of the model as new categories are incorporated without eradicating previously learned knowledge. Although both methodologies present potential resolutions to the challenges of sample scarcity and class evolution in SAR target recognition, they are not without their own set of difficulties. Fine-tuning with emerging classes can perturb the feature distribution of established classes, culminating in catastrophic forgetting, while training exclusively on a handful of new samples can induce bias towards older classes, leading to distribution collapse and overfitting. To surmount these limitations and satisfy practical application requirements, we propose a Few-Shot Class-Incremental SAR Target Recognition method based on a Dynamic Task-Adaptive Classifier (DTAC). This approach underscores task adaptability through a feature extraction module, a task information encoding module, and a classifier generation module. The feature extraction module discerns both target-specific and task-specific characteristics, while the task information encoding module modulates the network parameters of the classifier generation module based on pertinent task information, thereby improving adaptability. Our innovative classifier generation module, honed with task-specific insights, dynamically assembles classifiers tailored to the current task, effectively accommodating a variety of scenarios and novel class samples. Our extensive experiments on SAR datasets demonstrate that our proposed method generally outperforms the baselines in few-shot class incremental SAR target recognition. Full article
Show Figures

Figure 1

27 pages, 6439 KB  
Article
Contrastive–Transfer-Synergized Dual-Stream Transformer for Hyperspectral Anomaly Detection
by Lei Deng, Jiaju Ying, Qianghui Wang, Yue Cheng and Bing Zhou
Remote Sens. 2026, 18(3), 516; https://doi.org/10.3390/rs18030516 - 5 Feb 2026
Viewed by 589
Abstract
Hyperspectral anomaly detection (HAD) aims to identify pixels that significantly differ from the background without prior knowledge. While deep learning-based reconstruction methods have shown promise, they often suffer from limited feature representation, inefficient training cycles, and sensitivity to imbalanced data distributions. To address [...] Read more.
Hyperspectral anomaly detection (HAD) aims to identify pixels that significantly differ from the background without prior knowledge. While deep learning-based reconstruction methods have shown promise, they often suffer from limited feature representation, inefficient training cycles, and sensitivity to imbalanced data distributions. To address these challenges, this paper proposes a novel contrastive–transfer-synergized dual-stream transformer for hyperspectral anomaly detection (CTDST-HAD). The framework integrates contrastive learning and transfer learning within a dual-stream architecture, comprising a spatial stream and a spectral stream, which are pre-trained separately and synergistically fine-tuned. Specifically, the spatial stream leverages general visual and hyperspectral-view datasets with adaptive elastic weight consolidation (EWC) to mitigate catastrophic forgetting. The spectral stream employs a variational autoencoder (VAE) enhanced with the RossThick–LiSparseR (R-L) physical-kernel-driven model for spectrally realistic data augmentation. During fine-tuning, spatial and spectral features are fused for pixel-level anomaly detection, with focal loss addressing class imbalance. Extensive experiments on nine real hyperspectral datasets demonstrate that CTDST-HAD outperforms state-of-the-art methods in detection accuracy and efficiency, particularly in complex backgrounds, while maintaining competitive inference speed. Full article
Show Figures

Figure 1

20 pages, 6530 KB  
Article
Multi-Center Prototype Feature Distribution Reconstruction for Class-Incremental SAR Target Recognition
by Ke Zhang, Bin Wu, Peng Li, Zhi Kang and Lin Zhang
Sensors 2026, 26(3), 979; https://doi.org/10.3390/s26030979 - 3 Feb 2026
Viewed by 354
Abstract
In practical applications of deep learning-based Synthetic Aperture Radar (SAR) Automatic Target Recognition (ATR) systems, new target categories emerge continuously. This requires the systems to learn incrementally—acquiring new knowledge while retaining previously learned information. To mitigate catastrophic forgetting in Class-Incremental Learning (CIL), this [...] Read more.
In practical applications of deep learning-based Synthetic Aperture Radar (SAR) Automatic Target Recognition (ATR) systems, new target categories emerge continuously. This requires the systems to learn incrementally—acquiring new knowledge while retaining previously learned information. To mitigate catastrophic forgetting in Class-Incremental Learning (CIL), this paper proposes a CIL method for SAR ATR named Multi-center Prototype Feature Distribution Reconstruction (MPFR). It has two core components. First, a Multi-scale Hybrid Attention feature extractor is designed. Trained via a feature space optimization strategy, it fuses and extracts discriminative features from both SAR amplitude images and Attribute Scattering Center data, while preserving feature space capacity for new classes. Second, each class is represented by multiple prototypes to capture complex feature distributions. Old class knowledge is retained by modeling their feature distributions through parameterized Gaussian diffusion, alleviating feature confusion in incremental phases. Experiments on public SAR datasets show MPFR achieves superior performance compared to existing approaches, including recent SAR-specific CIL methods. Ablation studies validate each component’s contribution, confirming MPFR’s effectiveness in addressing CIL for SAR ATR without storing historical raw data. Full article
(This article belongs to the Section Radar Sensors)
Show Figures

Figure 1

27 pages, 9162 KB  
Article
Multi-Domain Incremental Learning for Semantic Segmentation via Visual Domain Prompt in Remote Sensing Data
by Junxi Li, Zhiyuan Yan, Wenhui Diao, Yidan Zhang, Zicong Zhu, Yichen Tian and Xian Sun
Remote Sens. 2026, 18(3), 464; https://doi.org/10.3390/rs18030464 - 1 Feb 2026
Viewed by 693
Abstract
Domain incremental learning for semantic segmentation has gained lots of attention due to its importance for many fields including urban planning and autonomous driving. The catastrophic forgetting problem caused by domain shift has been alleviated by structure expansion of the model or data [...] Read more.
Domain incremental learning for semantic segmentation has gained lots of attention due to its importance for many fields including urban planning and autonomous driving. The catastrophic forgetting problem caused by domain shift has been alleviated by structure expansion of the model or data rehearsal. However, these methods ignore similar contextual knowledge between the new and the old data domain and assume that new knowledge and old knowledge are completely mutually exclusive, which cause the model to be trained in a suboptimal direction. Motivated by the prompt learning, we proposed a new domain incremental learning framework named RS-VDP. The key innovation of RS-VDP is to utilize a visual domain prompt to change the optimization direction from input data space and feature space. First, we designed a domain prompt based on a dynamic location module, which applied a visual domain prompt according to a local entropy map to update the distribution of the input images. Second, in order to filter the feature vectors with high confidence, a representation feature alignment based on an entropy map module is proposed. This module ensures the accuracy and stability of the feature vectors involved in the regularization loss, alleviating the problem of semantic drift. Finally, we introduced a new evaluation metric to measure the overall performance of the incremental learning models, solving the problem that the traditional evaluation metric is affected by the single-task accuracy. Comprehensive experiments demonstrated the effectiveness of the proposed method by significantly reducing the degree of catastrophic forgetting. Full article
Show Figures

Figure 1

27 pages, 4422 KB  
Article
LaGu-RCL: Language-Guided Resolution-Continual Learning for Semantic Segmentation of Remote Sensing Images
by Penglong Li, Zezhong Ma, Haifeng Li and Zhenyang Huang
Remote Sens. 2026, 18(3), 452; https://doi.org/10.3390/rs18030452 - 1 Feb 2026
Viewed by 392
Abstract
Remote sensing image semantic segmentation faces substantial challenges in training and transferring models across images with varying resolutions. This issue can be effectively mitigated by continuously learning knowledge derived from new resolutions; however, this learning process is severely plagued by catastrophic forgetting. To [...] Read more.
Remote sensing image semantic segmentation faces substantial challenges in training and transferring models across images with varying resolutions. This issue can be effectively mitigated by continuously learning knowledge derived from new resolutions; however, this learning process is severely plagued by catastrophic forgetting. To address this problem, this paper proposes a novel continual learning framework termed Language-Guided Resolution-Continual Learning (i.e., LaGu-RCL), which alleviates catastrophic forgetting through two complementary strategies. On the one hand, a multi-resolution image augmentation pipeline is introduced to synthesize higher- and lower-resolution variants for each training batch, allowing the model to learn from images of diverse resolutions at every training step. On the other hand, a language-guided learning strategy is proposed to aggregate features of the same resolution while separating those of different resolutions. This ensures that the knowledge acquired from previously learned resolutions is not disrupted by that from unseen resolutions, thereby mitigating catastrophic forgetting. To validate the effectiveness of the proposed approach, we construct MR-ExcavSeg, a multi-resolution dataset covering several counties in Chongqing, and conduct comparative experiments between LaGu-RCL and several state-of-the-art continual learning baselines. Experimental results demonstrate that LaGu-RCL achieves significantly superior segmentation performance and continual learning capability, verifying its advantages. Full article
(This article belongs to the Section AI Remote Sensing)
Show Figures

Figure 1

17 pages, 797 KB  
Article
Continued Electromagnetic Signal Classification Based on Vector Space Separation
by Lu Jia, Yan Zhao, Shichuan Chen and Zhijin Zhao
Electronics 2026, 15(3), 613; https://doi.org/10.3390/electronics15030613 - 30 Jan 2026
Viewed by 315
Abstract
Incremental electromagnetic signal classification is crucial in realistic wireless environments where new signal types continuously emerge and historical training data are often unavailable. This paper proposes a model-based incremental learning method driven by vector space separation to mitigate catastrophic forgetting without accessing old-task [...] Read more.
Incremental electromagnetic signal classification is crucial in realistic wireless environments where new signal types continuously emerge and historical training data are often unavailable. This paper proposes a model-based incremental learning method driven by vector space separation to mitigate catastrophic forgetting without accessing old-task samples or requiring semantic information. We show that forgetting is largely caused by insufficient separation between old and new classes in the classifier weight space. To address this issue, we jointly introduce weight normalization, a cosine-similarity separation loss, and regularization, together with cross-entropy supervision for new classes. Based on these designs, we propose an incremental learning method based on vector space separation for electromagnetic signal classification, enabling the model to continually recognize modulation signals without requiring semantic information or access to raw data from previous tasks during incremental updates. Experiments on two simulated modulation datasets under multiple task sequences demonstrate that the proposed method consistently alleviates catastrophic forgetting and achieves stable incremental performance, outperforming baselines while avoiding data rehearsal. Full article
(This article belongs to the Section Circuit and Signal Processing)
Show Figures

Figure 1

Back to TopTop