Next Article in Journal
An Adaptive Prescribed Performance Position Tracking Controller for Hydraulic Systems
Previous Article in Journal
Optimization-Driven Reconstruction of 3D Space Curves from Two Views Using NURBS
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Survey on Replay-Based Continual Learning and Empirical Validation on Feasibility in Diverse Edge Devices Using a Representative Method

1
School of Computer Science and Engineering, Chung-Ang University, Dongjak-gu, Seoul 06974, Republic of Korea
2
Department of Artificial Intelligence, Chung-Ang University, Dongjak-gu, Seoul 06974, Republic of Korea
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Mathematics 2025, 13(14), 2257; https://doi.org/10.3390/math13142257 (registering DOI)
Submission received: 5 June 2025 / Revised: 5 July 2025 / Accepted: 9 July 2025 / Published: 12 July 2025
(This article belongs to the Special Issue Computational Intelligence in Systems, Signals and Image Processing)

Abstract

The goal of on-device continual learning is to enable models to adapt to streaming data without forgetting previously acquired knowledge, even with limited computational resources and memory constraints. Recent research has demonstrated that weighted regularization-based methods are constrained by indirect knowledge preservation and sensitive hyperparameter settings, and dynamic architecture methods are ill-suited for on-device environments due to increased resource consumption as the structure scales. In order to compensate for these limitations, replay-based continuous learning, which maintains a compact structure and stable performance, is gaining attention. The limitations of replay-based continuous learning are (1) the limited amount of historical training data that can be stored due to limited memory capacity, and (2) the computational resources of on-device systems are significantly lower than those of servers or cloud infrastructures. Consequently, designing strategies that balance the preservation of past knowledge with rapid and cost-effective updates of model parameters has become a critical consideration in on-device continual learning. This paper presents an empirical survey of replay-based continual learning studies, considering the nearest class mean classifier with replay-based sparse weight updates as a representative method for validating the feasibility of diverse edge devices. Our empirical comparison of standard benchmarks, including CIFAR-10, CIFAR-100, and TinyImageNet, deployed on devices such as Jetson Nano and Raspberry Pi, showed that the proposed representative method achieved reasonable accuracy under limited buffer sizes compared with existing replay-based techniques. A significant reduction in training time and resource consumption was observed, thereby supporting the feasibility of replay-based on-device continual learning in practice.

1. Introduction

Continual learning (CL) refers to learning new information while retaining previously acquired knowledge, which is a crucial feature for real-world applications where data evolve in dynamic environments [1]. A significant challenge in CL is catastrophic forgetting, where newly learned information overwrites and erases previously learned knowledge [2]. Various methods have been developed to address this issue, with replay-based learning emerging as a prominent approach [3,4]. In replay-based learning, past data are stored in a buffer and revisited during training to preserve previous knowledge while learning new information [5,6]. One of the widely used methods in this approach is iCaRL, which uses a nearest class mean (NCM) classifier to store the mean feature vector for each class, thereby preserving past knowledge [7]. By contrast, methods such as experience replay (ER) update the model using a mix of current and buffered past data to mitigate forgetting [5]. Although effective in tasks such as image classification, these methods face challenges in on-device environments with limited memory and computing resources [8].
The primary goal of this work is to provide a brief literature survey on replay-based CL, which is known as a feasible implementation strategy for realizing on-device CL in resource-limited edge devices due to its simplicity and efficiency. One concern when CL is planned to run on edge devices with on-device settings is that it can fail to execute due to its hypersensitivity to hardware specifications. To circumvent this situation, the suitability, empirical performance, and trade-offs among the resources of the algorithm selection and parameter settings for implementing CL should be considered before implementation, which has not been reported. To this end, in addition to the literature survey, we conduct empirical experiments based on well-known edge devices using a representative CL implementation, and we then report the results to guide practitioners.
One of the main challenges in reducing retraining time is ensuring that the model can recognize previously learned classes while integrating new ones [9]. This requires careful model updating because simply overwriting older data may lead to catastrophic forgetting. In addition, the prolonged retraining process delays system responsiveness and consumes considerable resources, which is a critical concern in practice. For example, smartphones have limited computational power, memory, and battery life, making it impractical to continuously retrain complex models [10]. Therefore, an optimized approach is necessary to minimize the time and resources required for model updates. CL addresses these challenges by updating only the most relevant parts of the model, drastically reducing the retraining time.
In resource-limited buffer environments, distinguishing transfer learning, meta-learning, and CL is crucial. Transfer learning utilizes models pretrained on large datasets and fine-tunes them on smaller, domain-specific datasets, proving advantageous in environments with limited labeled data [11]. Meta-learning, on the other hand, aims to quickly adapt models to new tasks with minimal examples, making it suitable for scenarios with limited buffer capacity and rapid task shifts [12]. However, neither transfer learning nor meta-learning inherently address the sequential learning of multiple tasks without forgetting, which is specifically targeted by CL [13]. Recent self-supervised methods, such as CSRM-MIM, have demonstrated effectiveness by pretraining on unlabeled data, significantly reducing the labeling burden [14]. Semi-supervised methods, like multi-modal imitation learning for railway arc detection, effectively integrate limited labeled data with abundant unlabeled data, demonstrating resource efficiency [15]. Additionally, physics-informed neural networks, exemplified by surrogate modeling of pantograph–catenary interactions, leverage domain-specific physical laws to enhance feature extraction with minimal data [16]. Nevertheless, weight regularization and dynamic architecture approaches typically demand additional memory or computational overhead, rendering them less suitable for resource-constrained edge devices [17]. Therefore, sparse weight updates are critical in minimizing computation, memory consumption, and preserving accuracy by selectively updating only essential parameters [18].
Replay methods are particularly crucial in on-device environments owing to two main constraints. First, the limited memory capacity severely restricts the amount of past training data that can be stored. In contrast to conventional computing environments, which typically have abundant memory resources at their disposal, on-device scenarios frequently function within a range of stringent constraints. This limitation renders the preservation and maintenance of comprehensive historical data sets impracticable. To overcome these constraints, efficient replay strategies have recently been developed that selectively store and utilize only the most representative or informative past experiences. This approach enables the preservation of knowledge while minimizing memory usage. Second, on-device systems generally possess constrained computational capabilities. Edge devices and mobile platforms are characterized by limitations in processing speed and computational resources when compared to server-class systems. Consequently, these devices require a lightweight and expeditious approach to adapt to learning environments. Replay-based approaches have been demonstrated to address this need effectively. The model strategically reuses some of the data stored in a buffer during training to facilitate the expeditious incorporation of new information. The advantage of this approach is that adaptation is possible without the necessity of relearning the entire data set or repeating complex computations. Consequently, replay-based learning is particularly advantageous in real-time environments. This approach facilitates timely and efficient learning, particularly in situations where rapid adaptation and updating to new data is necessary.
Successful replay-based CL methods for on-device systems must carefully balance these memory and computational constraints. Such methods should efficiently select and manage replay buffers to retain essential knowledge from previously learned tasks while simultaneously enabling rapid, computationally efficient learning of new information. Developing replay methods optimized for these resource-constrained scenarios is necessary to achieve effective, scalable, and practical CL solutions for devices. This study first provides a survey of replay-based on-device CL. Based on our review, we propose that NCM with replay-based sparse weight updates can be a practical method for implementing CL on diverse edge devices. Next, by setting the proposed method as a representative, we conducted empirical comparisons on standard benchmark datasets, such as CIFAR-10, CIFAR-100, and TinyImageNet, deployed on edge devices, such as the Jetson Nano and Raspberry Pi, to provide meaningful guidance to practitioners in this field. The contributions of this study can be outlined as follows.
  • We survey replay-based CL methods, focusing on memory usage, computational efficiency, and classification performance in on-device environments.
  • We propose an NCM with replay-based sparse weight updates as the representative method for validating the feasibility of on-device CL in edge devices.
  • We conduct experiments in real-time on-device environments, demonstrating the advantages of the proposed method in terms of accuracy and efficiency under resource constraints.
  • We analyze the performance of the NCM classifier with replay-based sparse weight updates, highlighting key trade-offs between training time, inference time, and learning accuracy in limited-resource scenarios.

2. Related Works

CL is essential in real-world applications because it enables models to adapt to new information while retaining previously acquired knowledge. This capability is critical in dynamic environments where data continuously evolves. Various approaches to CL have been previously investigated, as shown in Table 1.
Replay-based continual learning methods across diverse domains have been proposed to address domain-specific challenges while preserving prior knowledge. In semantic segmentation, catastrophic forgetting resulting from class imbalance and background shift is mitigated by generating replay samples using both GAN synthesis and web-crawled images blended with new data, coupled with background inpainting to maintain segmentation performance over multiple increments [32]. In image classification, inconsistencies arising from heuristic buffer designs are alleviated through systematic guidelines on memory size, sampling strategies, and data augmentation to maximize replay utility in limited-buffer regimes [33]. In network intrusion detection, rare-class forgetting and computational overhead are addressed by employing class-balanced buffer sampling alongside efficient perturbation-based exemplar selection methods. In class-incremental learning with constrained memory, recency bias is reduced by generating boundary-adjacent adversarial examples during replay. In online continual learning, shortcut learning is countered by maintaining discriminative prototypes with adaptive feedback mechanisms [34,35,36]. These cross-domain insights underscore the versatility of replay-based CL. Because edge devices such as smartphones, IoT modules, and industrial sensors are now widely deployed in real-world applications, and image classification comprises a large portion of those device, optimizing exemplar buffering under resource constraints is critical. Therefore, our focus remains on image classification, where exemplar buffering must balance representational fidelity with on-device resource constraints.
Incremental classifier and representation learning, which integrates representation learning with the NCM classifier, is recognized as a foundational development in replay-based methods [7]. Clustering techniques have been employed in ER for streaming learning to facilitate CL on streaming datasets [1]. The issue of recency bias has been addressed in supervised contrastive replay through the implementation of self-supervised learning [4]. In contrastive CL, the challenge of preserving sufficient negative exemplars for self-supervised learning is addressed by utilizing asymmetric supervised learning and knowledge distillation, given the limited memory capacity of the buffer [3]. To retain past knowledge, ER has been applied for joint training with historical data. This method has been utilized effectively in deep reinforcement learning and image classification based on the Softmax classifier [5,37]. Serving as an effective replay strategy in various CL studies, ER has facilitated the development of models such as dark ER (DER), which incorporate knowledge distillation [6]. Advanced models, including the cross-entropy-based alternative of ER, have emerged to mitigate the problem of data representation duplication in continual streams by separating current data from buffer exemplars during training [9].
Research on the utilization of exemplars in the buffer for CL has been complemented by various exemplar selection methods aimed at maximizing limited buffer capacity. The constraint optimization problem formulation of gradient episodic memory was redefined using gradient-based sample selection as a challenge to diversify exemplars in the buffer [30,38]. In ER, random selection is the sole method employed for exemplar selection, indicating significant potential for further development in this area. A method known as maximal interfered retrieval was introduced to select exemplars that were least likely to interfere during the training of a CL model [31]. To address the issue of limited on-device buffer capacity, Lee et al. proposed carousel memory for exemplar selection aligned with a memory-storage structure [39]. ER achieves considerable performance improvements, even with random selection, compared with previous CL methods. These results demonstrate that sophisticated methodologies for exemplar selection can enhance performance. Based on this premise, gradient coreset-based replay buffer selection employs gradient matching and adaptive sampling strategies [28]. To address the issue of selecting ambiguous or noisy samples in previous bi-level coreset selection methods, coreset selection via reducible loss reflects the performance for prioritizing exemplars that maximize future loss reducibility through gradient coverage analysis and stability-aware greedy selection [19]. In real-world scenarios, continual streams often exhibit bias, rendering them inadequate for selecting exemplars based solely on a singular criterion, which can lead to poor representativeness. To address this issue, a three-criterion-based exemplar selection method has been explored [40]. In addition, research has been conducted on exemplar selection in online CL to better reflect realistic environments. In online CL, random selection methods, such as reservoir sampling, are vulnerable to imbalanced data, while methodologies, such as GSS, incur high computational costs [41].
Research has been conducted to design two data criteria from an information-theoretic perspective: surprise (capturing new information) and learnability (capturing information that induces interference). A Bayesian model was applied to exemplar selection to calculate these criteria [42]. Recent advancements in CL have involved various methods. In class-incremental learning, learnability was introduced by integrating out-of-distribution (OOD) detection into the framework. This led to the development of replay-OOD-WP for class-incremental learning, which utilizes two separate heads for tasks and OOD detection [24]. Non-overlapping sparse codes have been applied to manage the interference between tasks, and a dropout mechanism has been employed to accelerate learning, as observed in sparse coding in a dual-memory system [25]. Buffer size limitations in replay-based methods have been mitigated using saliency maps to select past experiences, leading to the introduction of experience packing for replay [26]. To effectively preserve data distribution during replay, class feature sketching has been used to project examples onto a lower-dimensional space [21].
As previously mentioned, various approaches have been developed for CL in real-world applications [42]. Many CL methods require simplifying assumptions to be applicable in real-world scenarios. To address this issue, greedy sampler and dumb learner was proposed, which resolves the problem by utilizing a train-from-scratch approach [43]. The performance degradation of CL agents owing to the imbalance between past knowledge and the current continual stream has also been highlighted. To mitigate this imbalance, learning a unified classifier incrementally via rebalancing was introduced by employing cosine normalization, less-forget constraints, and interclass separation [44]. Incremental learning with a dual-memory methodology utilizes the statistical information within the buffer to address the imbalance among exemplars [45]. Furthermore, partitioning reservoir sampling has been proposed to maintain balanced knowledge between the head and tail in multilabel classification scenarios, particularly in the presence of long-tailed distributions [46]. Meta-ER, which is based on meta-learning, has been suggested to consider not only the interference caused by the current continual stream on past knowledge, but also the transfer effects of past knowledge on the current stream [47]. In addition, research has been conducted on meta-consolidation for CL, which integrates meta-learning into both data and weight spaces [48]. Finally, based on the observation that humans engage in CL through the hippocampus and neocortex, functioning as fast and slow learning systems, respectively, DualNet was developed by combining supervised and self-supervised learning [49,50].
In addition to the basic replay-based approach, replay-based regularization methods have received significant attention. A particularly challenging issue is large-scale CL with imbalanced data, which introduces a recency bias. This problem has been addressed through bias correction [51]. Uncertainty-based methods have been developed to mitigate the significant memory costs associated with replay-based approaches by selecting important parameters [52]. In addition, it has been suggested that storing task-conditioned model parameters is easier than the conventional practice of saving data input-output pairs. To support this claim, hnet was introduced [53], which learns the parameters of a meta-model to generate models tailored for specific tasks. Although replay-based approaches utilizing ER are predominant, it has been shown that their performance is significantly lower than simply learning all data once. To address this issue, hindsight was proposed to anchor past knowledge and alleviate catastrophic forgetting by employing anchors that represent classes through bi-level optimization [54,55]. To address the low performance of regularization-based approaches and the data privacy issues of replay-based methods, gradient projection memory was researched [56], which separately stores gradients for past classes in fully connected and convolutional layers. Unlike previous methodologies, most regularization-based approaches have relied solely on past outputs, gradients, and individual weights. A new method has been proposed to regulate simpler and more efficient network internal embeddings [57]. Recent studies have also focused on semantic segmentation using regularization-based approaches. A significant issue arises when the performance drops dramatically if past classes are insufficiently learned. However, this problem has not been adequately addressed in semantic segmentation. Research has shown that distribution shifts occur in background classes, leading to the development of methods incorporating background modeling to prevent such shifts [58]. Furthermore, PLOP was proposed to preserve spatial relationships between long and short ranges at the feature level [59]. To address the bias introduced by time-varying training settings, the continual bias adaptor was introduced to enhance the classifier and stabilize the consolidation of previously learned tasks [23]. Meanwhile, to balance stability and plasticity, weight consolidation was applied, leading to the development of weight interpolation (CLeWI) [20]. To improve buffer utility, recent work selects examples based on reducible loss estimating each sample’s actual contribution to model performance [19]. This enhances replay diversity by applying sketch-based sampling followed by low-dimensional projection to efficiently preserve feature relationships [21].
In the on-device domain, various replay-based approaches have been explored to address the issue of catastrophic forgetting. Traditionally, raw data are stored to preserve past knowledge. However, in on-device environments, where the computational cost and memory capacity are limited, the performance of these conventional replay-based methods is constrained. To mitigate these limitations, an approach known as latent replay was proposed [8]. In this method, intermediate-layer activations are stored instead of raw data, effectively reducing memory usage and enhancing computational efficiency. Latent replay enables online, serverless on-device operation, but the approach faces challenges related to high computational costs. Building on this approach, a method tailored for smartphones was developed to maintain efficiency and sustainability amid data distribution shifts [10]. To further address these issues, research was conducted on applying quantization to activations stored by latent replay [29]. This approach aims to reduce the computational overhead, making the method more efficient for on-device applications. In addition, studies have leveraged the hierarchical storage structure of devices, composed of both memory and storage, to optimize knowledge retention [27]. In these methods, data stored in memory is used to preserve past knowledge, while entire classes of knowledge are maintained in storage. Furthermore, online profiling techniques have been employed to analyze the trade-offs between accuracy and energy consumption. This analysis has facilitated the development of optimal memory-storage management strategies [60]. The consideration of wake/sleep frameworks based on quantified latent regeneration is recommended for scenarios where distribution assumptions are not applicable [22]. These methodologies have made significant progress in this area and are contributing to the continued development of more efficient and effective on-device systems. As the number of devices being deployed in environments with limited resources increases, the ability to perform efficiently and effectively on-device, without reliance on external servers, becomes essential.
While NCM reduces memory usage using class mean vectors, its simplicity limits classification accuracy. This can be complemented by the sparse weight update method that selectively updates only necessary weights to improve learning efficiency and minimize unnecessary updates, thus optimizing the use of computational resources. This method simultaneously considers current and buffered past data, allowing the model to retain prior knowledge while effectively learning new information. In this study, we propose an NCM with the replay-based sparse update method as the representative method for our empirical experiments.

3. Proposed Representative Method

In this section, we provide the rationale behind the proposed representative method regarding on-device CL. Figure 1 illustrates the overview of the proposed CL framework, where each component for devising the proposed method will be explained subsequently. The overall training loop of our proposed method is outlined in Algorithm 1.
Algorithm 1 Overall continual learning procedure
Require: Incoming data stream D t , replay buffer M m e m o r y , model parameters θ , current weights w, learning rate η , threshold s, scaling factors α , β , regularization coefficient γ
for each task t do
   for input x, label y in D T  do
        L ˜ L ˜ ( ( x , y ) ; θ ) θ
        L M m e m o r y L ( M m e m o r y ; θ ) θ
        w | w | + α | L ˜ | + β | L M m e m o r y |
        M θ 1 if w 1 s 0 otherwise
        L total L t + γ · E ( x , z ) M m e m o r y f θ ( x ) z 2 2
        θ θ η · M θ · θ L t o t a l
   end for
    M m e m o r y UpdateBuffer ( M m e m o r y , D t )
   for each class c in D t  do
      F ( x ) Backbone ( x )
      m c 1 | D c | x D c F ( x )
   end for
end for

3.1. Motivation for Proposed Method

On-device introduces additional challenges compared with those considered in traditional scenarios, especially in environments with limited resources. One of the most critical issues is catastrophic forgetting, in which the model forgets previously learned knowledge as it learns new information [54]. To address this issue, several methods have been proposed, with replay-based approaches demonstrating superior performance. These approaches mitigate forgetting by storing a portion of past data in a buffer and leveraging it during the learning process on new data streams. Prominent methods in this category include iCaRL and ER [5,7], while iCaRL utilizes stored data with an NCM classifier and applies a distillation loss to retain knowledge of previous classes while learning new classes, ER focuses on preserving the past knowledge of a fully connected layer with a Softmax classifier, thereby distinguishing their distinct approaches in managing stored data and past knowledge. These distinct approaches have led to further studies building on each method. For example, the NCM classifier of iCaRL was combined with contrastive learning to address complex tasks beyond image classification [1,4]. On the other hand, ER has been enhanced with knowledge distillation to better retain past knowledge, as it does not rely on stored exemplars for direct inference [6]. To prevent overfitting to the replay buffer in ER, methods have been proposed to train exemplars and current data streams separately [9].
For on-device to succeed, it must overcome several constraints imposed by the limited on-device environment. One major challenge is the limited memory capacity, which restricts the amount of past data that can be stored to preserve knowledge. Thus, the model must efficiently utilize a small amount of available data to retain past knowledge while learning from new data. Furthermore, the learning process must be highly efficient because of the limited computing power of on-device systems. Additionally, real-time constraints necessitate low-latency learning processes to ensure timely updates. To address these challenges, methodologies that can preserve knowledge with minimal data and enable rapid learning from new information are required.
In on-device environments, where computational efficiency and memory limitations are critical, the NCM classifier offers an effective approach to CL by aggregating each class into a mean vector, which eliminates the need for forward and backward propagation calculations in fully connected layers in the end while maintaining accurate classification performance. In scenarios where new classes are introduced sequentially, the NCM classifier is advantageous because it enables the model to adapt to these changes while maintaining performance across all previously encountered classes.

3.2. NCM Classifier with Class Mean Vectors

The NCM classifier distinguishes between classes by continuously updating class mean vectors as new classes are introduced. Specifically, for each class t, the class mean vector m t is computed as the average of the feature vectors F ( x ) corresponding to the input data x X t , where X t is the set of training data for class t. The function F ( x ) represents the output feature map of the backbone neural network for the input data x, which is a latent representation extracted from layers prior to the classifier. The class mean vector m t is defined as the centroid of these feature vectors and is updated incrementally as new data for class t are introduced. The updated class mean vector m t is updated as
m t = 1 N t x X t F ( x ) ,
where N t represents the number of samples belonging to class t, ensuring that the number of examples normalizes the class mean vector. The class mean vector m t allows the model to represent the entire class with a single vector, which can be stored compactly and used for classification.
The NCM classifier assigns a sample to the class with the mean vector m t that is closest to the sample feature vector F ( x ) based on the Euclidean distance. This straightforward yet effective distance-based method ensures that each new sample is compared against stored class representations. The classifier calculates the Euclidean distance between the feature vector F ( x ) of the input sample and the mean vector of each class m t , selecting the class with the smallest distance as the predicted label. This can be written as
y ^ = arg min t F ( x ) m t ,
where · denotes the Euclidean norm that computes the distance between two vectors. By selecting the class t that minimizes the distance between the feature vector of the sample and the class mean vector, the NCM classifier assigns the sample to a class.

3.3. Efficient Adaptation via Sparse Weight Update

The NCM classifier can be used in on-device environments with limited computational and memory resources, but its simplistic internal process results in relatively low classification accuracy. However, when combined with an appropriate model-update method, the NCM classifier can be effective. To address this challenge, we propose a synergistic weight-update method for the NCM classifier used in CL. This method selectively updates model parameters based on their importance to both the current dataset D t and memory buffer M m e m o r y , which stores a subset of previously learned data. This method reduces the computational overhead while improving classification accuracy by prioritizing important weights and minimizing unnecessary updates.
Our method incorporates information from the current dataset and memory buffer to ensure that the model adapts efficiently to new tasks while retaining knowledge from previously encountered data. The new weight values w for a given weight w are computed by considering both the gradients of the loss function for the current task, L ˜ ( D t ; θ ) and the gradient of the loss function for the memory buffer, L ( M m e m o r y ; θ ) . The weight update is defined as
w = | w | + α L ˜ ( D t ; θ ) w + β L ( M m e m o r y ; θ ) w ,
where α and β are scaling factors that control the relative importance of the current dataset and memory buffer in determining the weight updates. The term L ˜ ( D t ; θ ) w represents the absolute value of the gradient of the loss function for the current task with respect to the weight w, while L ( M m e m o r y ; θ ) w represents the gradient of the loss for the memory buffer. This equation ensures that weights important for current and previous tasks are preserved, thereby reducing the risk of catastrophic forgetting.
To optimize this weight-update process, we apply a binary mask M θ to update only the most relevant weights selectively. The mask is generated based on the updated weight values w , with weights that fall below a predefined threshold 1 s being masked and excluded from further updates. By incorporating this masking mechanism, the model can focus on the most important parameters while minimizing unnecessary updates, leading to more efficient learning and reduced computational overhead. This method enhances the ability of the model to retain knowledge from previous tasks. The binary mask is defined as
M θ = 0 if w < 1 s 1 otherwise
The regularization term is introduced into the loss function to maintain consistent performance across both new and previously learned tasks. It ensures that the model does not forget previously acquired knowledge while learning new tasks by minimizing the difference between the logits of the memory buffer z and the current task f θ ( X t ) . This encourages the model to retain prior knowledge while adapting to new information. This regularization term is defined as
L t + γ E ( X t , z ) M m e m o r y z f θ ( X t ) 2 2 ,
where γ is a hyperparameter that controls the strength of the regularization term. By minimizing the Euclidean distance between the logits of the memory buffer and the current task, the model can smoothly integrate new knowledge without forgetting previous information. This strategy, combined with effective weight updates and the NCM classification process, ensures that the model is well-suited for on-device environments where resource constraints and task variability are important factors. Figure 2 illustrates the weight sparse update process based on the updated replay buffer, focusing on how the information is delivered among modules.

4. Experimental Results

In this section, we describe the datasets and edge devices, including the desktop, used to demonstrate the performance gap with edge devices directly and to validate the feasibility of on-device CL based on the proposed method. We present some interesting points and remarks based on the series of empirical results.

4.1. Experimental Settings

We utilized the following three widely recognized image benchmark datasets: CIFAR-10, CIFAR-100, and TinyImageNet. Table 2 summarizes the core statistics for these datasets, including the number of classes, total number of images, and image resolution. The CIFAR-10 dataset consists of 60,000 32 × 32 color images categorized into ten classes. CIFAR-100 extends the class count to 100, each containing 600 images of the same resolution. TinyImageNet further increases the complexity to 200 classes, with 500 training images and 50 validation images per class at a resolution 64 × 64 .
During training, we employed incremental learning scenarios in which new classes were introduced sequentially. Each model was provided with data from a subset of classes at a time, and the replay methods recalled stored examples alongside new data to mitigate catastrophic forgetting. Our proposed method used the same incremental approach, leveraging the NCM classifier and sparse weight updates to optimize performance. We varied buffer sizes to examine the influence of memory limitations on performance. The methods considered—ER and DER—are widely used in this domain. To ensure a fair comparison, the same buffer size configurations and hyperparameters were applied to each method. All models were trained using the standard split for each dataset, with 80% of the images randomly selected for training, and the remaining 20% used for testing. Each experiment was repeated multiple times to obtain average performance values and standard deviations. The results include final classification accuracies and resource usage metrics (training and inference times) for all the tested methods on each hardware platform.
We designed our experiments to replicate realistic deployment conditions on edge devices, where both computational power and memory capacity are highly constrained. For this purpose, we considered various hardware platforms, including Desktop, Jetson Nano, Odroid, and Raspberry Pi to evaluate the feasibility of on-device under different resource constraints. Table 3 lists the specifications of the four devices. A Rockchip RK3568B2 quad-core Cortex-A55 processor powers the ODROID-M1 clocked at 2.0 GHz, coupled with 8 GB of LPDDR4 RAM. The NVIDIA Jetson Nano 4GB Developer Kit integrates a quad-core ARM Cortex-A57 CPU at 1.42 GHz, a 128-core Maxwell GPU, and 4 GB of 64-bit LPDDR4 memory with a bandwidth of 25.6 GB/s. The Raspberry Pi 5 is equipped with a Broadcom BCM2712 quad-core Cortex-A76 processor running at 2.4 GHz, an 800 MHz VideoCore VII GPU. The desktop PC was configured with an Intel Core i9-10900X CPU operating at 3.70 GHz, 32 GB of RAM, and dual NVIDIA GeForce RTX 3090 Ti GPUs. These devices were selected to evaluate the feasibility and performance of on-device CL across a spectrum of computational capabilities and resource constraints.

4.2. Comparison Results

Table 4 summarizes the average accuracy, training time, and inference time at a fixed buffer size of 300 across all datasets and hardware platforms. Detailed breakdowns for the other buffer sizes (100, 200, 300, 400, and 500) are provided in Table A2, Table A3, Table A4, Table A5, Table A6, Table A7, Table A8, Table A9, Table A10, Table A11, Table A12 and Table A13.
Table A2, Table A3, Table A4, Table A5, Table A6, Table A7, Table A8, Table A9, Table A10, Table A11, Table A12 and Table A13 show the performance of ER and DER with the sparse weight update on four different devices—Desktop, Odroid, Jetson Nano, and Raspberry Pi—and three datasets—CIFAR-10, CIFAR-100, and TinyImageNet. In each table, the columns show the replay method, sparse ratio, accuracy, training, and inference time according to the buffer size.
Table A2, Table A3, Table A4 and Table A5 show the comparison results on the CIFAR-10 dataset. As shown in Table A2, when the sparse ratio was 0% and the buffer size was 500, the training and inference times for ER were 147.42 s and 6.39 ms, respectively. By contrast, as shown in Table A5, the training and inference times were 364.76 s and 11.58 ms, respectively, which are more than double. As seen in Table A2, the accuracy when the buffer size was 400 and the sparse ratio was 50% for ER was 35.00%, which is lower than the accuracy when the sparse ratio was 25%, at 35.80%. These results suggest that sparse ratio and accuracy are not necessarily inversely related and that the optimal sparse ratio may vary depending on the characteristics of the data and backbone model. In the DER experiments listed in Table A3, with a sparse ratio of 90% and buffer size of 500, the training and inference times were 150.46 s and 3.85 ms, respectively, which is less than half of that without the sparse weight update. In addition, as the buffer size increased, the training time, inference time, and accuracy also increased because more data are stored in the buffer and many parameters are updated as the data are utilized for training, which naturally increases the number of parameters utilized for inference. The experimental results show that DER was approximately 7–9% less accurate than ER and took 1.4–1.5 times longer to train and infer. As shown in Table A4, for a buffer size of 500 and sparse ratio of 0%, ER and DER had accuracies of 35.61% and 28.41%, respectively, a difference of approximately 7.2%. Under the same conditions, the training times for ER and DER were 294.97 s and 408.64 s, respectively, with DER taking 1.4 times longer. The results of the CIFAR-100 and TinyImageNet experiments indicate that the fewer classes that need to be trained, the less efficient DER becomes. There were also cases in which sparse ratios of 50% or 90% were less accurate than those of 25% or 75%.
Figure 3 shows the classification accuracy on CIFAR-10 for the ER and DER methods across different buffer sizes, comparing 0% and 90% sparse weight updates averaged across four hardware platforms including Desktop, Odroid, Jetson Nano, and Raspberry Pi. With no sparsity, ER accuracy rises from 32.1% to 35.6% as buffer size increases from 100 to 500. Similarly, DER accuracy increases from 23.8% to 28.4% over the same range.
Applying 90% sparsity yields different effects for each method. ER maintains strong performance with accuracy increasing from 32.5% to 36.0% across the buffer size range, demonstrating robust scalability even with sparse updates. In contrast, DER with 90% sparsity shows a different pattern, starting at 24.2% and peaking at 28.3% at buffer size 300, but then declining to 27.9% at buffer size 500. Notably, sparse updates narrow the accuracy gap between DER and ER at moderate buffer sizes, particularly around buffer size 300 where the gap reduces from 8.3% dense to 6.0% sparse. However, at larger buffer sizes, the benefit diminishes. These results demonstrate that ER remains the stronger strategy across all configurations, while sparse weight updates can partially mitigate the DER performance disadvantage primarily at moderate buffer capacities.
Figure 4 shows the classification accuracy on CIFAR-10 for ER and DER across sparse weight update ratios from 0% to 90% at a fixed replay buffer size of 300 on the same four hardware platforms. ER demonstrates remarkable stability under sparsity, with accuracy climbing from 34.1% at zero sparsity to a peak of 34.8% at 75% sparsity, then falling slightly to 34.4% at 90% sparsity. This indicates that high sparsity does not impair ER performance and may even provide marginal benefits.
DER shows more varied behavior across sparsity levels. Accuracy dips to 25.6% at 25% sparsity, recovers to 26.8% at 50% sparsity, falls to 26.6% at 75% sparsity, and peaks at 28.4% at 90% sparsity. Importantly, the accuracy gap between ER and DER narrows from approximately 8.3 percentage points at zero sparsity to roughly 6.0 percentage points at high sparsity levels.
These findings reveal two key insights. First, ER maintains consistent performance regardless of sparsity level, making it suitable for memory-constrained scenarios without accuracy degradation. Second, DER benefits from high sparsity levels at 90%, suggesting that sparse weight updates can serve as a regularization mechanism that partially compensates for the inherent limitations of DER. The narrowing of the performance gap at high sparsity levels indicates that sparse updates offer a practical approach to reduce the performance disparity between these continual learning strategies across diverse hardware platforms.
The corresponding results for the CIFAR-100 and TinyImageNet datasets are presented in Appendix Figure A1, Figure A2, Figure A3 and Figure A4.
The CIFAR-100 experimental results in Table A6, Table A7, Table A8 and Table A9 show that DER and ER had similar or, in some cases, slightly higher accuracy, but DER took approximately 1.3 times longer to train. As listed in Table A7, when the buffer size was 500 and the sparse ratio was 0%, the training times for ER and DER were 319.41 s and 424.11 s, respectively, which is approximately 1.3 times longer for DER. This difference is attributed to the added knowledge distillation process in DER. The experimental results show limited performance improvement.
The TinyImageNet experimental results in Table A10, Table A11, Table A12 and Table A13 show that DER was marginally more accurate than ER. As shown in Table A10, when the buffer size was 500 and sparse ratio was 0%, ER and DER had an accuracy of 8.00% and 8.44%, respectively. As shown in Table A12, ER with sparse 0% and buffer sizes of 100 and 200 had an accuracy of 5.12%, and 6.24%, respectively, an increase of 1.12%; increasing the buffer size from 200 to 300 had an increase of 0.67%. This is because of the large number of classes in the TinyImageNet dataset, which causes some classes to fail to be saved when the buffer size is 100. The probability of class retention increased significantly when the buffer size was 200. Comparing the experimental results in Table A13 with the CIFAR-10 and CIFAR-100 results in Table A5 and Table A9, the performance improvement was larger when the buffer size increased from 100 to 200 than when it increased from 100 to 300, 400, and 500.
In summary, the experimental results from Table A2, Table A3, Table A4, Table A5, Table A6, Table A7, Table A8, Table A9, Table A10, Table A11, Table A12 and Table A13 indicate that the learning and inference times increased progressively as the constraints became more severe in the following order: Desktop, Odroid, Jetson Nano, and Raspberry Pi. When the sparse weight update was applied, the training and inference times decreased by up to two times, while the accuracy remained the same or increased slightly. Therefore, algorithm selection and hyperparameter tuning considering device resource constraints are essential for the effective implementation of CL on edge devices. In contrast to ER techniques, which merely store input data, DER preserves how prior knowledge was represented by storing model outputs from previous time points together. This approach offers the benefit of maintaining consistency in the representation space and preserving historical training data. Achieving stable performance with a small replay buffer, DER strikes a balance between memory efficiency and learning stability, which is advantageous in edge device environments. However, it necessitates a greater computational demand than ER, which can pose challenges if the supplementary resources exceed the capacity limits of the edge device. Consequently, when computational resources and memory are not severely constrained, DER is regarded as a more suitable approach for CL to ensure consistent performance in edge device environments. In particular, on Raspberry Pi, the trend of increased training and inference time and decreased accuracy was noticeable. On datasets with a small number of classes, such as CIFAR-10, DER was found to be less accurate and required longer training times than ER. However, on datasets with many classes, such as CIFAR-100 or TinyImageNet, DER tended to achieve a higher or similar accuracy to that of ER, albeit at the expense of increased training time. These results demonstrate that it is essential to consider the number and complexity of classes in the dataset, as well as the resource constraints of edge devices, when selecting an appropriate algorithm. A strategy that balances training efficiency and accuracy by adjusting the sparse ratio and buffer size appropriately is required, particularly when dealing with datasets containing a large number of classes. Thus, it is essential to ensure a sufficiently large buffer size to utilize a variety of data for training, where the appropriate buffer size can be guessed based on our empirical experiments.

5. Discussion

In this section, we analyze the limitations of the proposed NCM-based continual learning approach and identify scenarios where performance degradation may occur. We discuss scalability challenges, hardware dependency issues, and sensitivity to hyperparameter configurations observed in our empirical studies.
Our NCM-based approach faces inherent scalability limitations when dealing with large-scale datasets containing hundreds or thousands of classes, as the storage requirement for class mean vectors grows linearly with the number of classes [4,7]. Additionally, the distance-based classification becomes computationally expensive as the number of stored prototypes increases, contradicting on-device efficiency goals. The proposed method may also underperform in rapidly changing environments where data distributions shift frequently, as our fixed buffer size and static sparse ratio configuration may not adapt adequately to dynamic conditions. Our experimental results demonstrate significant sensitivity to buffer size configuration, particularly evident in TinyImageNet. The 2-3x performance degradation [61] from desktop to Raspberry Pi indicates a heavy hardware dependency, potentially limiting deployment flexibility across diverse edge computing scenarios.
The proposed method may struggle in several of the following specific scenarios: (1) fine-grained classification tasks [62] where subtle inter-class differences require sophisticated feature representations that simple class means cannot capture; (2) imbalanced continual learning scenarios where minority classes may be inadequately represented in the limited replay buffer; and (3) real-time critical applications requiring immediate model updates, where computational overhead of sparse weight calculation may introduce unacceptable latency, particularly on resource-constrained devices where training times exceeded 300 s for larger buffer sizes. These limitations highlight the need for adaptive approaches that can dynamically adjust to varying computational constraints and data characteristics while maintaining the efficiency benefits demonstrated in our edge device experiments.

6. Conclusions

On-device CL has garnered increasing interest in both practical applications and academic research. Although replay-based methods have been widely used to mitigate catastrophic forgetting, relatively few approaches can operate efficiently under the computational and memory constraints of edge devices. We conducted a comparative review of existing replay-based techniques to establish a reference point for future on-device CL solutions.
In this study, we proposed an NCM with the replay-based sparse weight update as the representative method for empirical validation. Then we investigated the feasibility of a replay-based CL tailored for on-device environments. Experimental results on diverse on-device environments indicate that the proposed method offers a favorable trade-off between classification accuracy and computational efficiency. Although the method demonstrates improvements over representative replay-based baselines under strict resource constraints, the observed accuracy gains are relatively modest, and the scope of baseline coverage is limited. Moreover, our experiments demonstrated robust performance on edge devices such as the Jetson Nano and Raspberry Pi, confirming the feasibility of our approach under realistic deployment conditions. These findings lay a promising foundation for the development of scalable, resource-efficient CL models capable of adapting to evolving data distributions in real-world on-device scenarios.
Building upon these results, future work should explore hybrid approaches that combine our replay-based sparse updates with regularization methods, such as Elastic Weight Consolidation (EWC) or PackNet [63], to provide complementary benefits for explicit knowledge preservation and implicit parameter protection while maintaining computational efficiency. The framework shows promise for extension beyond image classification to domains including natural language processing [64], audio processing, and federated continual learning scenarios, where multiple edge devices collaborate to learn [65]. These research directions address the identified limitations while leveraging the demonstrated advantages of efficient on-device continual learning.

Author Contributions

The authors confirm contribution to the paper as follows: Conceptualization, H.-S.P. and J.L. (Jaesung Lee); methodology, H.-S.P.; software, H.-C.C., C.K., and J.L. (Jeongwon Lee); validation, D.-W.K. and J.L. (Jaesung Lee); formal analysis, H.-S.P. and H.-C.C.; investigation, H.-S.P.; resources, J.L. (Jaesung Lee); data curation, H.-S.P.; writing—original draft preparation, H.-S.P., H.-C.C., M.-K.S., C.K., and J.L. (Jeongwon Lee); writing—review and editing, H.-S.P., H.-C.C., C.K., J.L. (Jeongwon Lee), and J.L. (Jaesung Lee); visualization, M.-K.S.; supervision, D.-W.K. and J.L. (Jaesung Lee); project administration, J.L. (Jaesung Lee); funding acquisition, D.-W.K. and J.L. (Jaesung Lee) All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Institute of Information & Communications Technology Planning & Evaluation (IITP) grant funded by the Korean government (MSIT) [RS-2021-II211341, Artificial Intelligence Graduate School Program (Chung-Ang University)], and in part by the National Research Foundation of Korea(NRF) grant funded by the Korean government(MSIT) (No. 2023R1A2C1006745).

Data Availability Statement

The data that support the findings of this study are openly available on the website of the Department of Computer Science, University of Toronto at https://www.cs.toronto.edu/~kriz/cifar.html (accessed on 11 July 2025) and from Stanford CS231n at http://cs231n.stanford.edu/tiny-imagenet-200.zip (accessed on 11 July 2025).

Acknowledgments

The authors have reviewed and edited the output and take full responsibility for the content of this publication.

Conflicts of Interest

The authors declare no conflicts of interest to report regarding the present study.

Appendix A. Detailed Experimental Results

We provide detailed implementation settings to ensure reproducibility and fairness of our experiments.
  • Optimizer: Stochastic Gradient Descent (SGD) with momentum 0.9 and weight decay 5 × 10 4 .
  • Initial learning rate: 0.1, with cosine-annealing scheduler down to 1 × 10 4 over the total number of epochs.
  • Batch size: 128 samples.
  • Number of epochs: 200 for CIFAR-10/100 and TinyImageNet, 100 for smaller datasets.
  • Learning-rate warmup: Linear warmup over the first 10 epochs.
  • Buffer sizes: Fixed at 100, 200, 300, 400, or 500 samples.
  • Sparse update ratios: 0%, 25%, 50%, 75%, 90%.
  • Continual Weight Importance (CWI):
    Computed for each weight w as follows:
    CWI ( w ) = | w | + α L ˜ ( D t ; θ ) w + β L ( M ; θ ) w
    L ˜ ( D t ; θ ) : task-specific loss with logits masked for current task only.
    L ( M ; θ ) : loss computed over the rehearsal buffer M.
    Hyperparameters:  α = 0.8 , β = 0.2 .
  • Data split & CL setup:
    Class-incremental protocol with 10 sequential tasks.
    Within each task, data are split 80% for training and 20% for validation.
    Replay sampling is performed by uniform random sampling from the buffer.
  • Reproducibility:
    Experiments on all devices (PC, Jetson Nano, and Raspberry Pi) use identical seeds, data splits, and sampling protocols.
    Checkpoints and buffer states are saved every 10 epochs.
    Final evaluation is based on the best validation accuracy observed.
Table A1. Continual learning protocols for each dataset.
Table A1. Continual learning protocols for each dataset.
DatasetTotal Classes#TasksClasses per TaskSamples per ClassTrain/Val
CIFAR-1010101500080%/20%
CIFAR-100100101050080%/20%
TinyImageNet200102050080%/20%
For each dataset as follows:
  • CIFAR-10: Ten tasks, each introducing one novel class sequentially (e.g., Task 1: {airplane}, Task 2: {automobile}, …).
  • CIFAR-100: Ten tasks, each containing 10 fine-grained classes (labels grouped in order).
  • TinyImageNet: Ten tasks, each with 20 classes randomly shuffled once at the start (seed = 42).
In this appendix, we present the full ER and DER results for replay buffer sizes 100, 200, 300, 400, and 500. Table A2, Table A3, Table A4, Table A5, Table A6, Table A7, Table A8, Table A9, Table A10, Table A11, Table A12 and Table A13 list the detailed data, and Figure A1, Figure A2, Figure A3 and Figure A4 show the corresponding visualizations.
Table A2. Comparison results of on the CIFAR-10 dataset in the Desktop environment.
Table A2. Comparison results of on the CIFAR-10 dataset in the Desktop environment.
ModelSparse (%)Accuracy (Buffer Size)Training Time (s)Inference Time (ms)
100200300400500100200300400500100200300400500
ER033.4233.8734.1634.7235.01133.12139.24143.60145.89156.345.125.946.256.786.95
2533.1033.8033.9334.2935.2880.5386.4794.59106.12110.022.853.774.264.965.21
5032.9833.2134.1535.2835.4370.2481.9584.8392.04102.672.792.943.735.215.58
7534.5634.1235.0635.3735.8262.5473.8475.0380.1883.551.552.133.293.875.01
9033.0434.4734.3635.5335.5450.8457.1263.0263.1573.051.021.262.743.584.62
DER024.8925.3926.1926.7927.34251.88259.88266.88274.88283.885.095.796.597.498.19
2524.0824.7825.4826.0326.73166.26174.56183.76191.06201.963.243.744.445.296.19
5025.3426.0426.5427.2427.74146.47153.67161.27168.07176.572.583.283.884.635.43
7525.0825.7826.1826.9827.48127.47134.47140.97147.07155.771.992.693.394.094.84
9027.0527.5528.4029.0529.50108.40115.60124.40133.40142.401.081.782.683.634.28
Co2L025.2925.7926.5927.1927.74254.88262.88269.88277.88286.888.098.799.5910.4911.19
2524.4825.1825.8826.4327.13169.26177.56186.76194.06204.966.246.747.448.299.19
5027.6028.3028.8029.5030.00112.60119.80127.40134.20142.705.586.286.887.638.43
7525.4826.1826.5827.3827.88130.47137.47143.97150.07158.774.995.696.397.097.84
9025.5926.0926.9427.5928.04148.27155.47164.27173.27182.274.084.785.686.637.28
CBA025.4925.9926.7927.3927.94256.88264.88271.88279.88288.8810.0910.7911.5912.4913.19
2524.6825.3826.0826.6327.33171.26179.56188.76196.06206.968.248.749.4410.2911.19
5025.9426.6427.1427.8428.34151.47158.67166.27173.07181.577.588.288.889.6310.43
7527.9028.6029.0029.8030.30113.90120.90127.40133.50142.206.997.698.399.099.84
9025.4325.9326.7827.4327.88113.40120.60129.40138.40147.406.086.787.688.639.28
SIESTA025.1925.6926.4927.0927.64241.88249.88256.88264.88273.881.912.613.414.315.01
2524.3825.0825.7826.3327.03156.26164.56173.76181.06191.964.364.865.566.417.31
5027.5028.2028.7029.4029.9099.60106.80114.40121.20129.704.825.526.126.877.67
7525.3826.0826.4827.2827.78117.47124.47130.97137.07145.775.215.916.617.318.06
9025.4925.9926.8427.4927.94135.27142.47151.27160.27169.275.726.427.328.278.92
Table A3. Comparison results on the CIFAR-10 dataset in the Odroid environment.
Table A3. Comparison results on the CIFAR-10 dataset in the Odroid environment.
ModelSparseAccuracy (Buffer Size)Training Time (Seconds)Inference Time (ms)
100200300400500100200300400500100200300400500
ER033.0033.9034.0734.5035.20205.56212.00220.56230.00239.007.207.708.208.809.60
2533.0033.5033.9234.5035.20130.76135.76140.76150.00160.004.505.005.526.006.50
5032.9033.6034.2735.0035.50110.32120.32128.32140.00150.003.804.304.805.506.00
7533.9634.4634.9635.4636.2095.29102.29110.29120.00130.003.303.804.305.005.50
9033.0033.8034.3234.8035.8080.1488.1496.14110.00120.002.683.183.684.685.68
DER025.0025.9026.2927.8028.50300.37312.37323.37350.00370.007.007.808.669.5010.50
2523.6024.5025.1126.5027.00200.43210.43217.43230.00240.004.004.905.796.507.00
5025.5026.2026.8828.3029.20175.83185.83192.83215.00225.004.004.605.256.006.80
7524.5025.4025.9927.5028.20150.29157.29167.29180.00190.003.003.804.405.005.80
9026.5027.5027.9828.9029.50125.80135.80145.80160.00170.002.003.003.564.505.00
Co2L025.0026.0026.6928.0029.00305.37315.37326.37345.00365.0010.0011.0011.6613.0014.00
2524.0025.0025.5126.5027.50200.43210.43220.43240.00260.007.508.208.799.5010.00
5027.0028.0028.3829.8030.50180.83190.83195.83210.00220.007.008.008.259.0010.00
7525.0026.0026.3927.5028.00155.29165.29170.29185.00195.006.007.007.408.009.00
9025.8026.8027.2828.8029.70130.80140.80148.80160.00175.005.006.006.568.009.00
CBA025.5026.5026.8928.0029.00310.37320.37328.37350.00370.0012.0013.0013.6615.0016.00
2524.0025.0025.7126.5027.50200.43210.43222.43240.00260.009.0010.0010.7912.0013.00
5026.0027.0027.4828.5029.50180.83190.83197.83210.00230.009.0010.0010.2512.0013.00
7527.0028.0028.5829.5030.50155.29165.29172.29185.00205.008.009.009.4011.0012.00
9025.0026.0026.5927.5028.00130.80140.80150.80170.00190.007.008.008.5610.0011.00
SIESTA025.0026.0026.5927.5028.00293.37303.37313.37343.00363.000.000.801.342.603.40
2523.9024.9025.4126.9027.50187.43197.43207.43237.00257.003.003.704.215.506.00
5026.8027.5028.2829.5030.00162.83172.83182.83212.00232.003.004.004.756.006.50
7525.0025.5026.2927.0027.50137.29147.29157.29187.00207.004.005.005.607.007.50
9025.8026.6027.1828.8029.40115.80125.80135.80155.00175.005.006.006.448.008.50
Table A4. Comparison results of on the CIFAR-10 dataset in the Jetson Nano environment.
Table A4. Comparison results of on the CIFAR-10 dataset in the Jetson Nano environment.
ModelSparse (%)Accuracy (%)Training Time (s)Inference Time (ms)
100200300400500100200300400500100200300400500
ER032.8033.2033.9034.2034.50290.00285.00277.41265.00260.0011.0010.009.468.007.50
2532.5033.0033.9034.5035.00205.00200.00191.74180.00175.007.506.506.295.004.50
5033.5034.0034.6235.0035.20170.00168.00164.69155.00150.007.006.005.624.003.80
7533.4034.0034.8335.5036.00160.00155.00150.79140.00135.006.005.004.824.003.00
9033.0033.5034.4435.0035.50140.00135.00129.04120.00115.006.005.004.294.003.00
DER025.0025.5026.5027.5028.00410.00405.00266.88385.00380.0012.0011.0010.529.008.60
2524.0024.5025.4526.0026.50285.00280.00183.76260.00255.008.007.006.505.004.50
5025.0025.5026.2027.0027.50255.00250.00161.27230.00225.007.006.506.065.004.10
7525.5026.0026.6827.0027.50225.00220.00140.97200.00195.006.005.505.224.504.00
9027.0027.5028.3629.0029.50200.00195.00124.40169.00165.006.005.004.393.002.50
Co2L025.5026.0026.9027.5028.00415.00410.00269.88385.00382.0015.0014.0011.6612.0011.50
2524.5025.0025.8526.5027.00290.00285.00186.76260.00258.0011.0010.008.798.007.50
5027.5028.0028.7629.0029.50260.00255.00127.40230.00227.0011.0010.009.068.007.50
7525.7026.5027.0827.5028.00230.00225.00143.97200.00195.009.508.507.407.006.50
9025.2025.8026.6027.0027.20205.00200.00148.80180.00175.009.008.006.566.005.50
CBA025.6026.0027.1028.0028.50418.00410.00271.88385.00386.0017.0016.0013.6614.0013.50
2524.7025.2026.0526.8027.00290.00285.00188.76260.00265.0013.0012.0010.7910.009.50
5025.5026.0026.8027.0027.50260.00255.00166.27230.00235.0012.5011.5010.259.509.10
7527.5028.0028.9629.5030.00230.00225.00127.40200.00198.0012.0011.0010.229.008.50
9025.8026.0027.2828.0028.50205.00200.00150.80175.00178.0011.0010.009.398.007.50
SIESTA025.5026.0026.8027.2027.50400.00395.00313.37370.00375.002.001.001.340.100.00
2524.5025.0025.7526.0026.50275.00270.00207.43250.00245.005.004.004.212.001.50
5027.5028.0028.6629.0029.50245.00240.00182.83225.00220.005.004.504.753.002.00
7525.5026.5026.9827.5028.00215.00210.00157.29190.00185.006.005.005.603.002.78
9025.2025.8026.5027.0027.50190.00185.00135.80160.00162.007.006.006.444.003.61
Table A5. Comparison results of on the CIFAR-10 dataset in the Raspberry Pi environment.
Table A5. Comparison results of on the CIFAR-10 dataset in the Raspberry Pi environment.
ModelSparse (%)Accuracy (%)Training Time (s)Inference Time (ms)
100200300400500100200300400500100200300400500
ER032.3932.7933.5934.4934.99327.14335.14345.14357.14363.149.8310.3311.3312.5313.13
2533.1533.6534.1534.8535.45226.44233.44241.44251.44259.446.286.887.588.489.28
5033.2233.7234.6235.6236.02188.37196.37206.37218.37224.375.175.776.577.578.37
7533.4633.9634.6635.7636.16165.72173.72182.72193.72200.723.874.575.376.577.17
9033.4033.8034.4035.7035.90142.24150.24160.24170.24178.243.053.654.355.356.15
DER025.1025.6026.3027.2027.70389.72397.72407.72419.72425.7211.0011.7012.5013.5014.30
2524.2924.7925.6926.8927.19251.91259.91269.91281.91287.916.777.478.279.2710.07
5025.4025.9026.5027.5027.90218.41226.41236.41248.41254.415.636.136.937.938.73
7525.4626.1626.8627.8628.26193.02201.02211.02223.02229.024.455.055.856.857.65
9027.0327.7328.5329.7330.03163.15171.15181.15193.15199.153.383.884.685.686.48
Co2L025.4025.9026.7027.7028.20392.72400.72410.72422.72428.7214.0014.7015.5016.5017.30
2524.6925.3926.0927.1927.59254.91262.91272.91284.91290.919.7710.4711.2712.2713.07
5027.7328.3328.9330.0330.43221.41229.41239.41251.41257.418.639.139.9310.9311.73
7525.8626.5627.2628.2628.76196.02204.02214.02226.02232.027.358.058.859.8510.65
9025.6026.1026.9028.0028.40166.15174.15184.15196.15202.156.386.887.688.689.48
CBA025.7026.3026.9028.1028.30394.72402.72412.72424.72430.7216.0016.7017.5018.5019.30
2524.8925.5926.2927.2927.79256.91264.91274.91286.91292.9111.9712.4713.2714.2715.07
5025.9026.5027.1028.3028.50223.41231.41241.41253.41259.4110.6311.1311.9312.9313.73
7527.7328.4329.1330.1330.63198.02206.02216.02228.02234.029.5510.0510.8511.8512.65
9026.2626.8627.4628.6628.86168.15176.15186.15198.15204.158.388.889.6810.6811.48
SIESTA025.2025.9026.6027.6028.10379.72387.72397.72409.72415.721.201.702.503.504.30
2524.5925.2925.9926.9927.49241.91249.91259.91271.91277.910.431.031.732.733.53
5027.6328.2328.8330.0330.33208.41216.41226.41238.41244.411.772.273.074.074.87
7525.7626.4627.1628.1628.66183.02191.02201.02213.02219.022.853.354.155.155.95
9025.6026.2026.8028.0028.30153.15161.15171.15183.15189.154.024.525.326.327.12
Table A6. Comparison results of on the CIFAR-100 dataset in the Desktop environment.
Table A6. Comparison results of on the CIFAR-100 dataset in the Desktop environment.
ModelSparseAccuracy (Buffer Size)Training Time (Seconds)Inference Time (ms)
100200300400500100200300400500100200300400500
ER09.7510.8111.1110.4211.54206.77206.18212.18226.28210.868.658.018.777.7610.35
2510.2410.2210.9110.0812.06149.83121.22136.59126.29138.287.996.876.074.556.66
5012.3610.7111.3910.7111.90138.95140.27126.14138.57125.676.054.875.206.704.61
7511.7610.7911.3710.6210.42115.6594.82113.18101.72112.892.682.654.535.792.67
9010.9211.6211.1510.8511.51100.95104.1398.2695.9981.982.271.903.603.533.66
DER010.4411.8111.4410.4710.61281.24295.24281.76278.42282.979.408.118.999.649.37
2511.549.9611.2910.7510.46197.77194.15187.61168.97192.754.394.076.015.797.55
5011.9310.3911.4610.7612.37170.29165.79169.31173.87177.427.125.445.163.756.94
7512.6112.0311.3911.1811.81138.28165.32150.87148.83143.416.264.584.353.834.45
9012.9111.5511.8510.4212.23135.18115.35130.33130.76120.683.262.533.534.702.96
Co2L012.4812.5111.8410.6212.48276.55293.06284.76269.75301.4610.7710.6911.9912.0711.00
2512.0412.5012.2512.0512.34178.01198.41190.61202.22187.919.3710.359.019.9010.40
5010.5610.3611.8610.6810.40189.79183.60172.31176.52156.848.896.888.167.227.29
7511.8210.4711.7912.0612.10166.82162.00153.87154.02143.118.715.717.356.586.25
9012.0512.0811.6912.9512.94136.80141.26133.33130.27127.996.035.116.537.216.67
CBA013.3011.4012.0412.1313.31267.30292.66286.76269.74298.9612.2614.7513.9915.0113.36
2513.7912.9712.4511.0211.21202.54202.32192.61176.49196.1711.349.1411.0111.2612.46
5011.3711.6512.0612.1312.15182.28175.01174.31185.37171.0611.3311.1010.169.178.42
7510.6511.5711.9912.4912.01175.27161.36155.87138.28145.688.288.849.358.599.43
9013.1111.7211.8911.8812.46142.89149.09135.33117.02148.377.1710.368.537.939.55
SIESTA012.6710.6411.7410.6712.90291.68281.42271.76282.53274.672.752.451.011.561.25
2511.7012.4312.1511.0813.02189.42180.55177.61168.46190.703.443.753.992.232.63
5010.8512.0911.7612.6811.74140.49151.65159.31143.90150.774.365.694.844.464.23
7512.0911.8411.6910.5210.61127.60140.93140.87156.74146.306.086.635.656.715.35
9010.3212.1411.5911.7712.37121.06140.13120.33125.84117.408.418.026.475.277.48
Table A7. Comparison results of on the CIFAR-100 dataset in the Odroid environment.
Table A7. Comparison results of on the CIFAR-100 dataset in the Odroid environment.
ModelSparseAccuracy (Buffer Size)Training Time (Seconds)Inference Time (ms)
100200300400500100200300400500100200300400500
ER09.7310.2111.0512.0212.58292.14302.83310.04322.17329.6510.3811.0411.4012.1813.12
259.6110.1811.0612.0812.51188.47198.12204.44223.57221.896.526.987.858.479.48
509.5810.2911.1012.0012.56172.79179.56188.36199.31204.885.445.996.847.538.49
7510.0210.4211.0912.0112.57154.82159.14167.47179.88184.774.274.985.766.517.52
9010.0310.8911.2912.0012.63133.24139.18146.09154.34164.113.153.974.805.826.68
DER010.0210.4911.3112.0712.82394.76404.19413.43428.30431.5710.1210.6211.4812.0912.78
2510.0510.3811.1311.9312.41263.19269.85281.38297.48294.826.147.118.078.498.99
5010.0810.6011.5012.2112.89239.30247.45256.67268.12274.515.025.606.617.257.88
7510.0110.8011.6812.1912.83212.95224.11231.50243.17249.244.124.555.566.027.02
9010.2310.7811.7312.2112.92179.48189.47198.37209.88217.142.373.384.365.036.18
Co2L010.3010.9111.7912.4813.21398.52409.13416.43427.69434.8112.3313.5714.4814.9916.29
2510.6111.2912.1312.8113.39268.44274.52284.38298.77302.669.1410.1211.0711.9412.73
5010.3711.0411.8612.4313.37318.65329.88335.40347.21353.928.449.7010.6711.3211.99
7510.5811.2812.0812.5313.33288.18294.36304.56318.49323.286.477.548.568.979.38
9010.0410.5811.5312.0412.78252.47259.68268.34280.12285.175.266.247.367.898.81
CBA010.4810.9211.9912.6813.39547.19554.38565.42578.03584.2516.1817.3218.3018.9519.53
2510.6211.1812.1212.8713.57358.72364.81373.16384.47389.6311.2612.3713.0713.9114.95
5010.5610.8912.0612.7413.37318.05329.23337.40351.14354.0510.4811.5312.6712.9814.02
7510.7711.2212.2812.5313.78288.86294.47306.56318.23323.359.1310.3310.5611.1812.19
9010.4210.5911.7312.6813.41254.37259.81270.34283.28289.678.119.249.3610.0411.07
SIESTA010.1910.9111.6912.5113.15531.24540.15550.42564.71567.881.422.313.304.205.22
2510.2911.0211.8212.3713.32343.19349.83358.16370.48374.910.280.561.171.983.02
5010.2410.8811.7612.2512.89309.68314.52322.40339.12341.370.331.312.333.024.05
7510.4910.8211.9812.4713.47273.45279.32291.56304.87309.221.582.633.664.625.56
9010.2110.8411.6412.2013.10243.59249.16255.34264.88274.022.793.774.775.386.69
Table A8. Comparison results of on the CIFAR-100 dataset in the Jetson Nano environment.
Table A8. Comparison results of on the CIFAR-100 dataset in the Jetson Nano environment.
ModelSparseAccuracy (Buffer Size)Training Time (Seconds)Inference Time (ms)
100200300400500100200300400500100200300400500
ER09.349.9410.9911.7912.39400.07405.41415.86424.09433.3111.8312.0713.3014.4215.12
258.699.4910.7411.9812.75252.96258.83271.30280.41285.886.977.828.9710.539.79
509.7010.3511.1811.8512.40233.31243.74252.08263.65267.776.076.707.758.848.33
759.6010.1311.0812.2813.00202.38208.91219.73235.02239.035.365.616.767.888.43
909.289.8910.7311.5112.66180.32186.26198.31205.11215.553.354.155.306.187.03
DER010.1610.7411.3912.6112.23540.56545.90560.42574.09578.5411.3311.5813.3014.8113.97
259.5410.3711.3412.7312.26353.15357.82368.16387.05373.186.977.648.839.7810.44
5010.1110.9811.4612.4412.83314.51326.55332.40351.90342.655.746.607.679.328.95
759.5410.8311.5713.0512.59282.94294.14301.56317.76307.544.615.056.347.867.07
9010.0710.5611.5212.3912.88245.71250.15265.34278.37274.443.414.115.236.576.12
Co2L010.4111.2311.7913.1214.14545.60554.15563.42582.02569.6514.6515.0616.3017.4816.94
2510.1211.1911.9213.3312.65354.93359.81371.16388.79373.5510.4610.7811.8313.6612.35
5010.7411.1911.8613.1412.71317.39325.16335.40351.63340.228.689.3210.6711.9111.32
7510.6911.4111.9713.0412.98288.14295.93304.56321.92309.607.618.259.3410.4610.15
909.2710.5111.7413.0212.68248.61249.74268.34284.84272.326.337.068.239.358.89
CBA010.7511.0711.9913.3912.25545.52547.62565.42582.20570.0816.8217.0218.3019.8519.06
2510.7811.0812.1213.4013.33354.08360.53373.16385.84381.2612.0412.7413.8315.2814.50
5010.6410.8412.0612.9614.27320.12327.19337.40350.07345.5111.3611.4112.6713.9213.64
7510.7111.3812.1713.6213.46289.21295.61306.56320.66309.629.4510.3111.3412.5612.21
9010.5110.6511.9413.2312.88251.17257.59270.34286.86273.308.318.4410.2311.4110.85
SIESTA08.499.7911.6913.1314.19532.26533.40550.42568.76556.441.181.883.304.895.04
2510.5411.1111.8212.8112.24338.49342.69358.16369.43375.280.600.801.172.232.71
5010.3010.5511.7612.9412.64304.13313.58322.40342.24329.610.591.032.333.223.68
7510.4511.0911.8713.0412.62271.95281.02291.56304.03297.401.992.613.664.635.04
9010.2810.5211.6412.8912.59237.15239.68255.34267.83263.993.073.324.775.795.71
Table A9. Comparison results of on the CIFAR-100 dataset in the Raspberry Pi environment.
Table A9. Comparison results of on the CIFAR-100 dataset in the Raspberry Pi environment.
ModelSparseAccuracy (Buffer Size)Training Time (Seconds)Inference Time (ms)
100200300400500100200300400500100200300400500
ER09.9210.4811.0511.7512.40504.50512.30522.48537.12540.2314.5015.2116.0117.0017.90
259.7610.4111.0612.1612.51330.11337.11345.11357.11363.119.5310.2311.0312.5312.93
509.9010.4011.1012.3012.55294.42300.42306.42316.42323.428.098.599.4910.4911.29
759.9910.5911.0911.9912.39270.12275.12280.12288.12295.126.526.927.729.229.62
909.9910.6911.2912.3912.79229.00237.00244.00254.00262.005.145.846.648.148.54
DER010.1110.7111.3112.3112.71674.27683.27692.27704.27711.2714.6315.3316.1317.3318.03
2510.0310.6311.1312.2312.58447.32458.32465.32477.32484.329.379.8710.7712.0712.57
5010.3010.8011.5012.7012.95396.51405.51414.51426.51433.518.148.649.1410.1411.04
7510.5811.1811.6812.4813.08359.12368.12377.12397.12407.126.107.107.608.609.40
9010.4311.1311.7313.1313.23316.42324.42334.42346.42352.424.995.696.297.498.09
Co2L010.5111.0111.7112.8113.16398.43407.43416.43428.43435.4317.6318.3319.1320.1321.03
2510.9311.6312.2313.4313.68450.32459.32468.32488.32495.3212.2713.0713.7714.7715.57
5010.5011.2011.9012.7013.40397.51407.51417.51429.51435.5110.6411.4412.1413.1414.14
7510.5811.3812.0813.0813.58362.12372.12380.12392.12398.129.1010.1010.6012.1012.60
9010.0310.8311.5312.7313.03319.42327.42337.42349.42354.427.798.799.2911.4911.79
CBA010.4111.2111.9112.7113.41677.27687.27697.27709.27715.2719.6320.5321.1322.1323.03
2510.9311.7312.3313.5313.88450.32460.32470.32482.32488.3214.2715.0715.7717.2717.57
5010.6011.4012.1013.3013.60399.51409.51419.51431.51437.5112.6413.6414.1415.6416.14
7510.7811.7812.2813.4813.78364.12374.12382.12394.12400.1211.1012.1012.6014.1014.60
9010.2311.0311.7313.0313.23319.42329.42339.42351.42357.429.7910.4911.2912.6913.09
SIESTA010.2111.2111.7112.5113.01662.27672.27682.27694.27700.274.635.636.137.537.93
2510.5311.5312.0313.0313.53435.32445.32455.32467.32473.320.270.570.771.772.57
5010.3011.0011.8012.3012.95384.51394.51404.51416.51422.510.100.400.861.862.56
7510.4811.1811.9813.1813.48347.12357.12367.12379.12385.120.901.902.403.404.40
9010.4311.0311.4312.9313.23304.42314.42324.42336.42342.422.212.713.714.715.71
Table A10. Comparison results of on the TinyImageNet dataset in the Desktop environment.
Table A10. Comparison results of on the TinyImageNet dataset in the Desktop environment.
ModelSparseAccuracy (Buffer Size)Training Time (Seconds)Inference Time (ms)
100200300400500100200300400500100200300400500
ER05.756.106.927.728.40356.80366.50368.58378.10386.2011.2012.0012.6313.4414.30
255.105.806.497.307.95230.50242.10247.77256.50262.106.657.658.319.079.86
505.706.056.937.608.03198.40205.60212.01219.80223.405.926.026.937.788.66
755.506.056.797.257.95174.00176.90185.48194.30201.004.835.425.946.767.25
905.606.106.757.397.83142.00144.40153.87161.60171.902.903.804.745.246.25
DER05.906.307.137.988.30465.00475.00490.20498.20505.2012.5013.0014.0014.6015.60
255.205.706.487.307.90319.40322.00329.63335.80346.007.808.609.1010.1011.00
506.006.507.148.008.60266.00276.00284.90293.10295.705.907.007.508.109.10
755.906.207.077.898.50235.40238.00247.87254.80266.805.105.506.367.007.40
905.806.507.027.988.20197.80209.70216.29222.20231.903.204.605.206.007.10
Co2L06.106.507.538.469.00475.00484.00493.20499.80507.0015.7016.2017.0017.5018.20
255.806.206.887.507.90317.50326.00332.60341.60345.3010.4011.4012.1012.7013.30
506.006.407.548.328.85276.00283.00287.90293.10306.509.309.8010.5011.2012.10
756.056.657.478.208.50237.00243.00250.90258.20266.308.308.809.309.9010.45
906.006.507.427.988.80207.50212.50219.30228.30239.206.407.508.209.109.90
CBA06.307.007.738.209.10479.00488.00495.20501.70508.5017.5018.5019.0019.8020.40
255.806.507.087.608.40315.00327.80334.60341.50352.8012.3013.3014.1014.8016.00
506.707.107.748.509.10277.00280.00289.90295.00309.6010.9011.6012.5013.3014.20
756.707.007.678.609.10239.30245.00252.90258.30272.609.7010.8011.4012.3013.20
906.107.107.628.108.80201.60212.00221.30230.50235.908.209.2010.2011.0012.00
SIESTA05.956.607.438.048.80462.50472.50480.20489.40495.702.203.204.004.805.20
255.605.906.787.458.00308.20310.00319.60327.10332.200.200.501.201.903.00
506.006.607.448.278.50264.00269.00274.90282.20286.900.801.902.403.004.00
756.256.707.378.118.70220.70228.40237.90245.40253.202.303.103.604.304.85
905.906.507.327.868.60192.60198.10206.30214.10219.803.504.204.805.406.00
Table A11. Comparison results of on the TinyImageNet dataset in the Odroid environment.
Table A11. Comparison results of on the TinyImageNet dataset in the Odroid environment.
ModelSparseAccuracy (Buffer Size)Training Time (Seconds)Inference Time (ms)
100200300400500100200300400500100200300400500
ER05.496.146.897.748.19542.65548.55560.85572.35578.8514.2614.9116.0116.9617.61
255.315.916.417.067.71345.49352.19360.69372.69378.698.669.4610.4611.2612.06
505.446.046.897.848.14300.41312.81318.21325.01332.517.528.229.129.9211.02
755.616.216.817.568.26260.35268.35278.35293.35298.355.996.897.898.599.19
905.366.206.667.417.81214.23224.73229.73239.73247.734.225.126.126.727.62
DER05.766.267.167.968.61714.73719.73734.73744.73754.7316.0616.7617.9618.7619.56
255.095.796.397.097.79457.95465.95475.95487.95493.9510.1810.9812.0812.8813.48
505.736.137.037.638.48405.41418.41423.41435.41441.417.678.579.5710.2711.17
755.726.327.027.628.32350.59360.59370.59380.59388.596.317.218.218.919.81
905.736.337.137.838.53302.42312.42322.42332.42340.425.095.896.897.598.19
Co2L06.166.767.568.268.91717.73722.73737.73747.73757.7319.0619.7620.9621.7622.56
255.496.196.797.498.19458.95466.95478.95490.95496.9513.1813.9815.0815.8816.68
506.036.637.438.138.78406.41416.41426.41436.41444.4110.7711.5712.5713.3713.97
756.026.627.428.178.72353.59363.59373.59383.59391.599.3110.2111.2112.0112.61
906.236.737.538.288.93305.42315.42325.42335.42343.428.098.899.8910.6911.49
CBA06.366.967.768.469.21719.73724.73739.73749.73759.7321.0621.7622.9623.7624.56
255.696.296.997.698.39460.95468.95480.95492.95498.9515.1816.0817.0817.8818.68
506.236.837.638.339.08408.41418.41428.41438.41446.4112.6713.5714.5715.3716.17
756.326.827.628.329.07355.59365.59375.59385.59393.5911.4112.2113.2114.0114.81
906.436.937.738.439.18307.42317.42327.42337.42345.429.9910.8911.8912.6913.49
SIESTA06.166.667.468.168.91704.73709.73724.73734.73744.736.166.967.968.769.46
255.396.096.697.398.09445.95453.95465.95477.95483.950.281.182.082.883.58
506.036.437.338.038.73393.41403.41413.41423.41431.410.030.230.431.131.93
756.026.527.328.028.72340.59345.59360.59370.59378.590.091.091.792.493.29
906.136.637.438.138.83292.42302.42312.42322.42330.421.212.213.113.914.61
Table A12. Comparison results of on the TinyImageNet dataset in the Jetson Nano environment.
Table A12. Comparison results of on the TinyImageNet dataset in the Jetson Nano environment.
ModelSparse (%)Accuracy (Buffer Size)Training Time (s)Inference Time (ms)
100200300400500100200300400500100200300400500
ER05.215.686.918.018.76722.21729.09734.58752.11730.7419.4717.6518.4320.3216.98
254.695.346.447.428.64482.48505.53490.32480.20495.4211.1814.3812.4010.8413.35
505.426.596.937.759.15439.83423.81427.68447.67412.4611.899.9510.5611.439.81
755.186.446.897.188.30345.11369.38364.66353.59378.547.759.858.878.4210.61
904.585.306.737.748.29314.83290.69308.68311.03301.226.158.977.105.878.61
DER05.496.777.308.279.63951.07958.75956.09940.21973.1921.7919.4221.1422.2320.17
254.726.176.447.738.35648.14612.23628.77635.83616.6814.3716.4015.1913.6516.18
505.146.246.997.528.66574.90597.60578.76566.04588.9312.7011.3712.4113.0311.39
755.556.666.977.869.09503.44482.67489.26502.15470.1511.148.7710.4311.999.59
905.536.477.128.199.59412.81448.91428.91447.24419.339.886.878.649.197.25
Co2L05.276.727.708.949.82946.45962.21959.09954.20978.6525.9123.3024.1425.0723.00
254.425.566.847.318.65640.94612.66631.77649.79618.1319.2217.0018.1919.0816.74
505.366.347.398.889.60580.26599.21581.76573.40592.9814.6916.9015.4114.8817.16
755.236.577.378.499.82509.61488.85492.26498.18475.6814.6212.8813.4315.2412.05
905.306.327.528.569.59412.27443.70431.91430.48442.4611.0212.5811.649.8412.82
CBA05.226.427.908.959.94975.46953.80961.09944.54969.8227.9725.1226.1426.6124.96
254.996.397.048.328.89615.67647.66633.77624.25638.9118.6320.9020.1919.7421.85
505.346.677.599.0410.15590.91567.04583.76588.64572.6817.5715.9617.4118.6016.59
755.246.457.579.059.94513.37485.57494.26506.49475.7814.4917.1115.4314.3216.94
905.737.167.729.0810.20414.32436.77433.91420.79445.3615.0812.4313.6414.4611.74
SIESTA05.796.817.608.119.37959.33928.59946.09951.22927.0812.079.6811.1412.6810.25
254.575.906.748.168.70608.45624.96618.77605.89634.986.643.875.196.274.23
505.076.067.298.429.17554.54570.90568.76563.68577.572.931.402.413.590.66
754.655.807.278.559.22466.10497.09479.26472.77489.841.411.610.432.151.77
905.706.817.428.889.90405.92424.65418.91414.66436.732.810.491.362.480.72
Table A13. Comparison results of on the TinyImageNet dataset in the Raspberry Pi environment.
Table A13. Comparison results of on the TinyImageNet dataset in the Raspberry Pi environment.
ModelSparseAccuracy (Buffer Size)Training Time (Seconds)Inference Time (ms)
100200300400500100200300400500100200300400500
ER05.776.326.947.518.28905.23915.67920.00935.12944.3811.5012.7813.7915.0016.12
255.405.926.447.007.76580.10590.45603.97620.30625.119.1110.2310.5611.8012.79
505.906.506.937.458.26520.12528.37539.62555.21560.348.609.3510.5612.0012.76
755.806.406.897.568.28460.35470.88480.35495.35504.127.128.018.8710.0510.82
905.746.136.737.648.15422.23432.45442.73458.23467.815.226.126.647.808.60
DER06.006.857.308.208.75940.23950.00956.09975.10980.4518.5019.7521.1422.5023.25
255.506.256.447.207.95460.47475.00475.95495.25500.1313.1014.0515.1916.0816.90
505.756.106.997.768.49695.32708.41714.64732.10738.5511.5012.2814.9415.9016.72
755.626.327.027.788.52590.21605.32620.21635.21640.999.2910.1112.7913.5814.40
905.836.697.318.068.80510.12525.23535.12552.23558.347.157.9010.1510.9511.76
Co2L06.166.767.568.268.91717.73722.73737.73747.73757.7319.0619.7620.9621.7622.56
255.496.196.847.498.19458.95466.95478.95490.95496.9513.1813.9815.0815.8816.68
506.036.637.438.138.78406.41416.41426.41436.41444.4110.7711.5712.5713.3713.97
756.026.627.428.178.72353.59363.59373.59383.59391.599.3110.2111.2112.0112.61
906.236.737.538.288.93305.42315.42325.42335.42343.428.098.899.8910.6911.49
CBA06.366.967.768.469.21719.73724.73739.73749.73759.7321.0621.7622.9623.7624.56
255.696.296.947.698.39460.95468.95480.95492.95498.9515.1816.0817.0817.8818.68
506.236.837.638.339.08408.41418.41428.41438.41446.4112.6713.5714.5715.3716.17
756.326.827.628.329.07355.59365.59375.59385.59393.5911.4112.2113.2114.0114.81
906.436.937.738.439.18307.42317.42327.42337.42345.429.9910.8911.8912.6913.49
SIESTA06.166.667.468.168.91704.73709.73724.73734.73744.736.166.967.968.769.46
255.396.096.697.398.09445.95453.95465.95477.95483.950.281.182.082.883.58
506.036.437.338.038.73393.41403.41413.41423.41431.410.030.230.431.131.93
756.026.527.328.028.72340.59345.59360.59370.59378.590.091.091.792.493.29
906.136.637.438.138.83292.42302.42312.42322.42330.421.212.213.113.914.61
Figure A1. Classification accuracy (%) on the CIFAR-100 dataset comparing ER and DER at sparse-weight update ratios of 0 % and 90 % across replay buffer sizes (100–500), evaluated on Desktop, Odroid, Jetson Nano, and Raspberry Pi.
Figure A1. Classification accuracy (%) on the CIFAR-100 dataset comparing ER and DER at sparse-weight update ratios of 0 % and 90 % across replay buffer sizes (100–500), evaluated on Desktop, Odroid, Jetson Nano, and Raspberry Pi.
Mathematics 13 02257 g0a1
Figure A2. Classification accuracy (%) on the TinyImageNet dataset comparing ER and DER at sparse-weight update ratios of 0 % and 90 % across replay buffer sizes (100–500), evaluated on Desktop, Odroid, Jetson Nano, and Raspberry Pi.
Figure A2. Classification accuracy (%) on the TinyImageNet dataset comparing ER and DER at sparse-weight update ratios of 0 % and 90 % across replay buffer sizes (100–500), evaluated on Desktop, Odroid, Jetson Nano, and Raspberry Pi.
Mathematics 13 02257 g0a2
Figure A3. Classification accuracy (%) on the CIFAR-100 dataset for ER and DER with a fixed replay buffer size of 300, across sparse-weight update ratios from 0% to 90%, evaluated on four hardware platforms: Desktop, Odroid, Jetson Nano, and Raspberry Pi.
Figure A3. Classification accuracy (%) on the CIFAR-100 dataset for ER and DER with a fixed replay buffer size of 300, across sparse-weight update ratios from 0% to 90%, evaluated on four hardware platforms: Desktop, Odroid, Jetson Nano, and Raspberry Pi.
Mathematics 13 02257 g0a3
Figure A4. Classification accuracy (%) on the TinyImageNet dataset for ER and DER with a fixed replay buffer size of 300, across sparse-weight update ratios from 0% to 90%, evaluated on four hardware platforms: Desktop, Odroid, Jetson Nano, and Raspberry Pi.
Figure A4. Classification accuracy (%) on the TinyImageNet dataset for ER and DER with a fixed replay buffer size of 300, across sparse-weight update ratios from 0% to 90%, evaluated on four hardware platforms: Desktop, Odroid, Jetson Nano, and Raspberry Pi.
Mathematics 13 02257 g0a4

References

  1. Hayes, T.L.; Cahill, N.D.; Kanan, C. Memory efficient experience replay for streaming learning. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; pp. 9769–9776. [Google Scholar]
  2. Chen, Z.; Liu, B. Continual learning and catastrophic forgetting. In Proceedings of the Lifelong Machine Learning, Seattle, WA, USA, 10–13 December 2018; pp. 55–75. [Google Scholar]
  3. Cha, H.; Lee, J.; Shin, J. Co2L: Contrastive Continual Learning. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 2–6 October 2021; pp. 9496–9505. [Google Scholar]
  4. Mai, Z.; Li, R.; Kim, H.; Sanner, S. Supervised Contrastive Replay: Revisiting the Nearest Class Mean Classifier in Online Class-Incremental Continual Learning. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Virtual Conference, 19–25 June 2021; pp. 3584–3594. [Google Scholar]
  5. Chaudhry, A.; Rohrbach, M.; Elhoseiny, M.; Ajanthan, T.; Dokania, P.; Torr, P.; Ranzato, M. Continual Learning With Tiny Episodic Memories. In Proceedings of the Workshop on Multi-Task and Lifelong Reinforcement Learning (MTLRL), Long Beach, CA, USA, 15 June 2019. [Google Scholar]
  6. Buzzega, P.; Boschini, M.; Porrello, A.; Abati, D.; Calderara, S. Dark experience for general continual learning: A strong, simple baseline. Adv. Neural Inf. Process. Syst. 2020, 33, 15920–15930. [Google Scholar]
  7. Rebuffi, S.A.; Kolesnikov, A.; Sperl, G.; Lampert, C.H. iCaRL: Incremental Classifier and Representation Learning. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 5533–5542. [Google Scholar]
  8. Pellegrini, L.; Graffieti, G.; Lomonaco, V.; Maltoni, D. Latent replay for real-time continual learning. In Proceedings of the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 25–29 October 2020; pp. 10203–10209. [Google Scholar]
  9. Caccia, L.; Aljundi, R.; Asadi, N.; Tuytelaars, T.; Pineau, J.; Belilovsky, E. New insights on reducing abrupt representation change in online continual learning. arXiv 2021, arXiv:2104.05025. [Google Scholar]
  10. Pellegrini, L.; Lomonaco, V.; Graffieti, G.; Maltoni, D. Continual learning at the edge: Real-time training on smartphone devices. arXiv 2021, arXiv:2105.13127. [Google Scholar]
  11. Zhuang, F.; Qi, Z.; Duan, K.; Xi, D.; Zhu, Y.; Zhu, H.; Xiong, H.; He, Q. A comprehensive survey on transfer learning. Proc. IEEE 2021, 109, 43–76. [Google Scholar] [CrossRef]
  12. Finn, C.; Abbeel, P.; Levine, S. Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks. In Proceedings of the 34th International Conference on Machine Learning (ICML), Sydney, Australia, 6–11 August 2017; Volume 70, pp. 1126–1135. [Google Scholar]
  13. Yin, S.Y.; Huang, Y.; Chang, T.Y.; Chang, S.F.; Tseng, V.S. Continual learning with attentive recurrent neural networks for temporal data classification. Neural Netw. 2023, 158, 171–187. [Google Scholar] [CrossRef]
  14. Yang, H.; Liu, Z.; Ma, N.; Wang, X.; Liu, W.; Wang, H.; Zhan, D.; Hu, Z. CSRM-MIM: A self-supervised pre-training method for detecting catenary support components in electrified railways. IEEE Trans. Transp. Electrif. 2025. [Google Scholar] [CrossRef]
  15. Yan, J.; Cheng, Y.; Zhang, F.; Zhou, N.; Wang, H.; Jin, B.; Wang, M.; Zhang, W. Multimodal imitation learning for arc detection in complex railway environments. IEEE Trans. Instrum. Meas. 2025, 74, 1–13. [Google Scholar] [CrossRef]
  16. Cheng, Y.; Yan, J.; Zhang, F.; Li, M.; Zhou, N.; Shi, C.; Jin, B.; Zhang, W. Surrogate modeling of pantograph–catenary system interactions. Mech. Syst. Signal Process. 2025, 224, 112134. [Google Scholar] [CrossRef]
  17. Kirkpatrick, J.; Pascanu, R.; Rabinowitz, N.; Veness, J.; Desjardins, G.; Rusu, A.A.; Milan, K.; Quan, J.; Ramalho, T.; Grabska-Barwinska, A.; et al. Overcoming catastrophic forgetting in neural networks. Proc. Natl. Acad. Sci. USA 2017, 114, 3521–3526. [Google Scholar] [CrossRef]
  18. Wang, Z.; Zhan, Z.; Gong, Y.; Yuan, G.; Niu, W.; Jian, T.; Ren, B.; Ioannidis, S.; Wang, Y.; Dy, J. SparCL: Sparse continual learning on the edge. Adv. Neural Inf. Process. Syst. 2022, 35, 20366–20380. [Google Scholar]
  19. Ruilin, T.; Yuhang, L.; Shi, J.Q.; Gong, D. Coreset selection via reducible loss in continual learning. In Proceedings of the International Conference on Learning Representations (ICLR), Singapore, 24–28 April 2025. [Google Scholar]
  20. Kozal, J.; Wasilewski, J.; Krawczyk, B.; Woźniak, M. Continual learning with weight interpolation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 17–21 June 2024; pp. 4187–4195. [Google Scholar]
  21. Julian, J.; Koh, Y.S.; Bifet, A. Sketch-Based Replay Projection for Continual Learning. In Proceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Barcelona, Spain, 25–29 August 2024; pp. 1325–1335. [Google Scholar]
  22. Harun, M.Y.; Gallardo, J.; Hayes, T.; Kemker, R.; Kanan, C. SIESTA: Efficient Online Continual Learning with Sleep. arXiv 2023, arXiv:2303.10725. [Google Scholar]
  23. Wang, Q.; Wang, R.; Wu, Y.; Jia, X.; Meng, D. CBA: Improving Online Continual Learning via Continual Bias Adaptor. In Proceedings of the 2023 IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 2–6 October 2023; pp. 19036–19046. [Google Scholar]
  24. Kim, G.; Xiao, C.; Konishi, T.; Liu, B. Learnability and Algorithm for Continual Learning. In Proceedings of the International Conference on Machine Learning, Honolulu, HI, USA, 23–29 July 2023; pp. 16877–16896. [Google Scholar]
  25. Sarfraz, F.; Arani, E.; Zonooz, B. Sparse Coding in a Dual Memory System for Lifelong Learning. In Proceedings of the Thirty-Seventh AAAI Conference on Artificial Intelligence, Washington, DC, USA, 7–14 February 2023; pp. 1–8. [Google Scholar]
  26. Saha, G.; Roy, K. Saliency Guided Experience Packing for Replay in Continual Learning. In Proceedings of the 2023 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–7 January 2023; pp. 5262–5272. [Google Scholar]
  27. Aggarwal, S.; Binici, K.; Mitra, T. Chameleon: Dual Memory Replay for Online Continual Learning on Edge Devices. IEEE Trans. Comput. Aided Des. Integr. Circuits Syst. 2023, 43, 1663–1676. [Google Scholar] [CrossRef]
  28. Tiwari, R.; Killamsetty, K.; Iyer, R.; Shenoy, P. GCR: Gradient Coreset based Replay Buffer Selection for Continual Learning. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 19–24 June 2022; pp. 99–108. [Google Scholar]
  29. Ravaglia, L.; Rusci, M.; Nadalini, D.; Capotondi, A.; Conti, F.; Benini, L. A TinyML Platform for On-Device Continual Learning With Quantized Latent Replays. IEEE J. Emerg. Sel. Top. Circuits Syst. 2021, 11, 789–802. [Google Scholar] [CrossRef]
  30. Aljundi, R.; Lin, M.; Goujaud, B.; Bengio, Y. Gradient Based Sample Selection for Online Continual Learning. In Advances in Neural Information Processing Systems; Wallach, H., Larochelle, H., Beygelzimer, A., d’Alché-Buc, F., Fox, E., Garnett, R., Eds.; Curran Associates, Inc.: New Orleans, LA, USA, 2019; Volume 32, Available online: https://proceedings.neurips.cc/paper_files/paper/2019/file/e562cd9c0768d5464b64cf61da7fc6bb-Paper.pdf (accessed on 4 June 2025).
  31. Aljundi, R.; Belilovsky, E.; Tuytelaars, T.; Charlin, L.; Caccia, M.; Lin, M.; Page-Caccia, L. Online Continual Learning with Maximal Interfered Retrieval. In Advances in Neural Information Processing Systems; Wallach, H., Larochelle, H., Beygelzimer, A., d’Alché-Buc, F., Fox, E., Garnett, R., Eds.; Curran Associates, Inc.: Vancouver, BC, Canada, 2019; Volume 32, Available online: https://proceedings.neurips.cc/paper_files/paper/2019/file/15825aee15eb335cc13f9b559f166ee8-Paper.pdf (accessed on 4 June 2025).
  32. Maracani, A.; Michieli, U.; Toldo, M.; Zanuttigh, P. RECALL: Replay-Based Continual Learning in Semantic Segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Virtual, 11–17 October 2021; pp. 7026–7035. [Google Scholar]
  33. Merlin, G.; Lomonaco, V.; Cossu, A.; Carta, A.; Bacciu, D. Practical Recommendations for Replay-Based Continual Learning Methods. In Proceedings of the International Conference on Image Analysis and Processing (ICIAP), Lecce, Italy, 19–23 September 2022; pp. 548–559. [Google Scholar]
  34. Amalapuram, S.K.; Channappayya, S.S.; Tamma, B.R. Augmented Memory Replay-Based Continual Learning Approaches for Network Intrusion Detection. In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), Vancouver, BC, Canada, 10–16 December 2023; pp. 17156–17169. [Google Scholar]
  35. Kumari, L.; Wang, S.; Zhou, T.; Bilmes, J.A. Retrospective Adversarial Replay for Continual Learning. In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), New Orleans, LA, USA, 28 November–9 December 2022; Volume 35, pp. 28530–28544. [Google Scholar]
  36. Wei, Y.; Ye, J.; Huang, Z.; Zhang, J.; Shan, H. Online Prototype Learning for Online Continual Learning. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 2–6 October 2023; pp. 18764–18774. [Google Scholar]
  37. Rolnick, D.; Ahuja, A.; Schwarz, J.; Lillicrap, T.; Wayne, G. Experience Replay for Continual Learning. In Advances in Neural Information Processing Systems; Wallach, H., Larochelle, H., Beygelzimer, A., d’Alché-Buc, F., Fox, E., Garnett, R., Eds.; Curran Associates, Inc.: Vancouver, BC, Canada, 2019; Volume 32, Available online: https://proceedings.neurips.cc/paper_files/paper/2019/file/fa7cdfad1a5aaf8370ebeda47a1ff1c3-Paper.pdf (accessed on 4 June 2025).
  38. Lopez-Paz, D.; Ranzato, M. Gradient Episodic Memory for Continual Learning. In Advances in Neural Information Processing Systems; Guyon, I., von Luxburg, U., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., Garnett, R., Eds.; Curran Associates, Inc.: Long Beach, CA, USA, 2017; Volume 30, Available online: https://proceedings.neurips.cc/paper_files/paper/2017/file/f87522788a2be2d171666752f97ddebb-Paper.pdf (accessed on 4 June 2025).
  39. Lee, S.; Weerakoon, M.; Choi, J.; Zhang, M.; Wang, D.; Jeon, M. Carousel Memory: Rethinking the Design of Episodic Memory for Continual Learning. arXiv 2021, arXiv:2110.07276. [Google Scholar]
  40. Zhuang, C.; Huang, S.; Cheng, G.; Ning, J. Multi-criteria Selection of Rehearsal Samples for Continual Learning. Pattern Recognit. 2022, 132, 108907. [Google Scholar] [CrossRef]
  41. Vitter, J.S. Random Sampling with a Reservoir. ACM Trans. Math. Softw. 1985, 11, 37–57. [Google Scholar] [CrossRef]
  42. Sun, S.; Calandriello, D.; Hu, H.; Li, A.; Titsias, M. Information-Theoretic Online Memory Selection for Continual Learning. arXiv 2022, arXiv:2204.04763. [Google Scholar]
  43. Prabhu, A.; Torr, P.H.S.; Dokania, P.K. GDumb: A Simple Approach that Questions Our Progress in Continual Learning. In Proceedings of the European Conference on Computer Vision(ECCV), Glasgow, UK, 23–28 August 2020; pp. 524–540. [Google Scholar]
  44. Hou, S.; Pan, X.; Loy, C.C.; Wang, Z.; Lin, D. Learning a Unified Classifier Incrementally via Rebalancing. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 831–839. [Google Scholar]
  45. Belouadah, E.; Popescu, A. IL2M: Class Incremental Learning With Dual Memory. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 583–592. [Google Scholar]
  46. Kim, C.D.; Jeong, J.; Kim, G. Imbalanced Continual Learning with Partitioning Reservoir Sampling. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020; pp. 411–428. [Google Scholar]
  47. Riemer, M.; Cases, I.; Ajemian, R.; Liu, M.; Rish, I.; Tu, Y.; Tesauro, G. Learning to Learn without Forgetting by Maximizing Transfer and Minimizing Interference. arXiv 2018, arXiv:1810.11910. [Google Scholar]
  48. KJ, J.; N Balasubramanian, V. Meta-consolidation for Continual Learning. In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), Virtual, 6–12 December 2020; Volume 33, pp. 14374–14386. Available online: https://proceedings.neurips.cc/paper_files/paper/2020/file/a5585a4d4b12277fee5cad0880611bc6-Paper.pdf (accessed on 4 June 2025).
  49. McClelland, J.L.; McNaughton, B.L.; O’Reilly, R.C. Why There Are Complementary Learning Systems in the Hippocampus and Neocortex: Insights from the Successes and Failures of Connectionist Models of Learning and Memory. Psychol. Rev. 1995, 102, 419. [Google Scholar] [CrossRef]
  50. Pham, Q.; Liu, C.; Hoi, S. Dualnet: Continual Learning, Fast and Slow. In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), Virtual, 6–14 December 2021; Volume 34, pp. 16131–16144. Available online: https://proceedings.neurips.cc/paper_files/paper/2021/file/86a1fa88adb5c33bd7a68ac2f9f3f96b-Paper.pdf (accessed on 4 June 2025).
  51. Wu, Y.; Chen, Y.; Wang, L.; Ye, Y.; Liu, Z.; Guo, Y.; Fu, Y. Large Scale Incremental Learning. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 374–382. [Google Scholar]
  52. Ahn, H.; Cha, S.; Lee, D.; Moon, T. Uncertainty-Based Continual Learning with Adaptive Regularization. In Advances in Neural Information Processing Systems; Wallach, H., Larochelle, H., Beygelzimer, A., d’Alché-Buc, F., Fox, E., Garnett, R., Eds.; Curran Associates, Inc.: Vancouver, BC, Canada, 2019; Volume 32, Available online: https://proceedings.neurips.cc/paper_files/paper/2019/file/2c3ddf4bf13852db711dd1901fb517fa-Paper.pdf (accessed on 4 June 2025).
  53. Von Oswald, J.; Henning, C.; Grewe, B.F.; Sacramento, J. Continual Learning with Hypernetworks. arXiv 2019, arXiv:1906.00695. [Google Scholar]
  54. McCloskey, M.; Cohen, N.J. Catastrophic Interference in Connectionist Networks: The Sequential Learning Problem. In Psychology of Learning and Motivation; Elsevier: San Diego, CA, USA, 1989; Volume 24, pp. 109–165. [Google Scholar]
  55. Chaudhry, A.; Gordo, A.; Dokania, P.; Torr, P.; Lopez-Paz, D. Using Hindsight to Anchor Past Knowledge in Continual Learning. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 2–9 February 2021; pp. 6993–7001. [Google Scholar]
  56. Saha, G.; Garg, I.; Roy, K. Gradient Projection Memory for Continual Learning. In Proceedings of the International Conference on Learning Representations (ICLR), Virtual, 3–7 May 2021. [Google Scholar]
  57. Pomponi, J.; Scardapane, S.; Lomonaco, V.; Uncini, A. Efficient Continual Learning in Neural Networks with Embedding Regularization. Neurocomputing 2020, 397, 139–148. [Google Scholar] [CrossRef]
  58. Cermelli, F.; Mancini, M.; Bulò, S.R.; Ricci, E.; Caputo, B. Modeling the Background for Incremental Learning in Semantic Segmentation. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 9230–9239. [Google Scholar]
  59. Douillard, A.; Chen, Y.; Dapogny, A.; Cord, M. PLOP: Learning without Forgetting for Continual Semantic Segmentation. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Virtual, 19–25 June 2021; pp. 4039–4049. [Google Scholar]
  60. Ma, X.; Jeong, S.; Zhang, M.; Wang, D.; Choi, J.; Jeon, M. Cost-effective On-device Continual Learning over Memory Hierarchy with Miro. In Proceedings of the 29th Annual International Conference on Mobile Computing and Networking, Madrid, Spain, 2–6 October 2023. [Google Scholar]
  61. Kwon, Y.D.; Chauhan, J.; Jia, H.; Venieris, S.I.; Mascolo, C. LifeLearner: Hardware-Aware Meta Continual Learning System for Embedded Computing Platforms. In Proceedings of the 21st ACM Conference on Embedded Networked Sensor Systems. Association for Computing Machinery, Istanbul, Turkiye, 12–17 November 2024; pp. 138–151. [Google Scholar]
  62. Sun, G.; Cholakkal, H.; Khan, S.; Khan, F.; Shao, L. Fine-Grained Recognition: Accounting for Subtle Differences between Similar Classes. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 12047–12054. [Google Scholar]
  63. Schwarz, J.; Czarnecki, W.; Luketina, J.; Grabska-Barwinska, A.; Teh, Y.W.; Pascanu, R.; Hadsell, R. Progress & Compress: A Scalable Framework for Continual Learning. In Proceedings of the 35th International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018; 2018; Volume 80, pp. 4528–4537. [Google Scholar]
  64. Razdaibiedina, A.; Mao, Y.; Hou, R.; Khabsa, M.; Lewis, M.; Almahairi, A. Progressive Prompts: Continual Learning for Language Models. In Proceedings of the International Conference on Learning Representations (ICLR), Virtual, 1–5 May 2023. [Google Scholar]
  65. Zhang, J.; Chen, C.; Zhuang, W.; Lyu, L. TARGET: Federated Class-Continual Learning via Exemplar-Free Distillation. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 2–6 October 2023; pp. 4782–4793. [Google Scholar]
Figure 1. Overview of the proposed CL framework using an NCM Classifier integrated with ResNet-18 for feature extraction. The model includes a replay buffer and a weight sparse update method, enhancing efficiency and adaptability across tasks in on-device environments.
Figure 1. Overview of the proposed CL framework using an NCM Classifier integrated with ResNet-18 for feature extraction. The model includes a replay buffer and a weight sparse update method, enhancing efficiency and adaptability across tasks in on-device environments.
Mathematics 13 02257 g001
Figure 2. Illustration of weight sparse update based on the updated replay buffer.
Figure 2. Illustration of weight sparse update based on the updated replay buffer.
Mathematics 13 02257 g002
Figure 3. Classification accuracy (%) on the CIFAR-10 dataset comparing ER and DER at sparse-weight update ratios of 0 % and 90 % across replay buffer sizes (100–500), evaluated on Desktop, Odroid, Jetson Nano, and Raspberry Pi.
Figure 3. Classification accuracy (%) on the CIFAR-10 dataset comparing ER and DER at sparse-weight update ratios of 0 % and 90 % across replay buffer sizes (100–500), evaluated on Desktop, Odroid, Jetson Nano, and Raspberry Pi.
Mathematics 13 02257 g003
Figure 4. Classification accuracy (%) on the CIFAR-10 dataset for ER and DER with a fixed replay buffer size of 300, across sparse-weight update ratios from 0% to 90%, evaluated on four hardware platforms: Desktop, Odroid, Jetson Nano, and Raspberry Pi.
Figure 4. Classification accuracy (%) on the CIFAR-10 dataset for ER and DER with a fixed replay buffer size of 300, across sparse-weight update ratios from 0% to 90%, evaluated on four hardware platforms: Desktop, Odroid, Jetson Nano, and Raspberry Pi.
Mathematics 13 02257 g004
Table 1. Summary of replay-based continual learning methods.
Table 1. Summary of replay-based continual learning methods.
MethodYearRemarksBuffer
Management
Feature
Update
Refs.
CSReL2025Uses Reducible Loss (ReL) to select coreset samples by
measuring performance gain via loss difference between
holdout and current models.
GreedyNot fixed[19]
CLeWI2024Applies weight interpolation after training on new tasks by
performing weight permutation to align activations and
uses linear interpolation to combine current and previous weights
RandomNot fixed[20]
SRP2024Uses a class-feature sketching to approximate feature distributions
applying replay buffer projection to project examples into
lower-dimensional space
GreedyNot Fixed[21]
SIESTA2023Implements a wake/sleep framework in online CL
during the wake phase without replay or backpropagation updates,
and offline memory consolidation occurs during the sleep phase
leveraging quantized latent replay for memory efficiency
GreedyFixed[22]
CBA2023Employs a bias adaptor that dynamically adapts the classifier
during training by applying a nonlinear transformation
to capture distribution changes in the posterior probability,
using a lightweight MLP to fit posterior shifts across tasks,
while removing the CBA during inference
GreedyNot fixed[23]
ROW2023Separates model for each task, using an OOD Head for
task identification and a WP Head for classification
to perform task detection and classification simultaneously
RandomNot fixed[24]
SCoMMER2023Implements sparse coding with activation and semantic dropout
to reduce task interference and maintain distinct activations,
while using a dual memory system with working memory
GreedyNot fixed[25]
EPR2023Utilizes saliency maps to identify key input patches
during replay, applying zero-padding to align these patches with
their original dimensions, enhancing memory efficiency
GreedyNot fixed[26]
Chameleon2023Dual memory is configured to optimize short-term and
long-term storage while incorporating user preferences through
affinity and uncertainty sampling, combining knowledge
distillation with contrastive learning
GreedyFixed[27]
GCR2022Coreset selection with criteria based on the current data of
a candidate pool using a gradient approximation objective,
combined with DER and supervised contrastive learning
GreedyContrastive
Learning
[28]
Co2L2021Combines asymmetric supervised contrastive learning with
instance-wise relation transfer for learning representations,
while calculating instance-wise relation drift
RandomContrastive
Learning
[3]
ER-ACE2021Avoids drastic representation changes by treating incoming and
replay data asymmetrically learning with isolated exemplars
RandomBackpropagation[9]
SCR2021Enhances NCM classifiers with contrastive learning using
a memory buffer to replay past and current examples for training
PrioritizedContrastive
Learning
[4]
QLR2021Replay quantized activations of intermediate layers by
compressing latent replay vectors down to 8 or 7 bits
RandomFixed[29]
DER/
DER++
2020Combines knowledge distillation by distilling knowledge from
the logits while also utilizing ground truth labels
RandomKnowledge
Distillation
[6]
Latent
Replay
2020By storing activations of the intermediate layer, memory
usage is reduced, leading to compact representations,
while backpropagation starting from intermediate layers
speeds up learning
RandomBackpropagation[8]
ER2019Using tiny episodic memories, managing the buffer using
reservoir sampling performs joint training with
both buffer and current data
RandomBackpropagation[5]
Exstream2019Uses memory-efficient buffers for streaming clustering, which
clusters incoming data points, stores representative examples, and
dynamically updates memory with minimal storage
Partitioning
based
Not fixed[1]
GSS2019Coreset selection based on calculated gradients to ensure
representation of diverse past samples
GreedyBackpropagation[30]
MIR2019Coreset selection for forming a buffer of exemplars likely to
experience maximal interference, based on
virtual model update prediction
GreedyBackpropagation[31]
iCaRL2017Integrates the NCM classifier, a replay buffer that
preserves raw data to retain past knowledge, and
knowledge distillation for effective representation learning
PrioritizedRepresentation
Learning
[7]
Table 2. The standard characteristics and statistics of employed datasets.
Table 2. The standard characteristics and statistics of employed datasets.
DatasetImage Size# Classes# Dataset# Train# Test# Task# Classes per Task
CIFAR-10 32 × 32 × 3 1060,00050,00010,00052
CIFAR-100 32 × 32 × 3 10060,00050,00010,0001010
TinyImageNet 3 × 64 × 64 200110,000100,00010,0001020
Table 3. Hardware specifications of experimental platforms.
Table 3. Hardware specifications of experimental platforms.
DeviceCPUGPUMemory
OdroidRockchip RK3568B2 Quad-core Cortex-A55 @ 2.0 GHzMali-G52 MP2 @ 650 MHz8 GB LPDDR4
Jetson NanoQuad-core ARM Cortex-A57 @ 1.43 GHz128-core NVIDIA Maxwell4 GB LPDDR4
Raspberry PiBroadcom BCM2712 Quad-core Cortex-A76 @ 2.4 GHzVideoCore VII @ 800 MHz4 GB LPDDR4X
DesktopIntel Core i9-10900X CPU @ 3.70 GHzNVIDIA GeForce RTX 3090 Ti32 GB DDR4
Table 4. Summary of experimental results across datasets and hardware platforms.
Table 4. Summary of experimental results across datasets and hardware platforms.
DatasetModelSparse Ratio (%)Average Accuracy (%)Training Time (Seconds)Inference Time (Milliseconds)
Desktop Odroid Jetson RaspPi Desktop Odroid Jetson RaspPi Desktop Odroid Jetson RaspPi
CIFAR-10ER034.16 ± 0.8934.07 ± 0.5333.90 ± 0.6233.59 ± 0.79143.60 ± 2.35220.56 ± 5.42277.41 ± 4.56345.14 ± 9.456.25 ± 0.098.20 ± 0.329.46 ± 0.1111.33 ± 0.04
2533.93 ± 0.5933.92 ± 0.7133.90 ± 1.0634.15 ± 0.4594.59 ± 0.60140.76 ± 3.76191.74 ± 3.86241.44 ± 2.114.26 ± 0.065.52 ± 0.056.29 ± 0.197.58 ± 0.07
5034.15 ± 0.7734.27 ± 0.9434.62 ± 0.2534.62 ± 0.4984.83 ± 0.67128.32 ± 1.49164.69 ± 0.71206.37 ± 5.853.73 ± 0.084.80 ± 0.125.62 ± 0.056.57 ± 0.18
7535.06 ± 0.2234.96 ± 0.7434.83 ± 0.2734.66 ± 0.7275.03 ± 0.90110.29 ± 1.67150.79 ± 4.04182.72 ± 4.753.29 ± 0.044.30 ± 0.084.82 ± 0.075.37 ± 0.14
9034.36 ± 0.8134.32 ± 0.6334.44 ± 0.6834.40 ± 0.5963.02 ± 0.9496.14 ± 1.15129.04 ± 2.47160.24 ± 1.832.74 ± 0.003.68 ± 0.074.29 ± 0.114.35 ± 0.07
DER026.19 ± 0.0226.29 ± 0.7226.50 ± 1.0626.30 ± 1.15266.88 ± 3.16323.37 ± 5.74399.86 ± 9.45407.72 ± 13.666.59 ± 0.238.66 ± 0.0810.52 ± 0.0912.50 ± 0.32
2525.48 ± 0.2625.11 ± 0.9925.45 ± 0.6625.69 ± 0.66183.76 ± 1.32217.43 ± 3.68274.12 ± 3.64269.91 ± 4.944.44 ± 0.075.79 ± 0.116.50 ± 0.158.27 ± 0.23
5026.54 ± 0.2726.88 ± 0.5426.20 ± 0.1226.50 ± 0.40161.27 ± 2.39192.83 ± 1.94243.95 ± 0.94236.41 ± 4.423.88 ± 0.095.25 ± 0.056.06 ± 0.156.93 ± 0.11
7526.18 ± 0.4425.99 ± 0.2226.68 ± 0.1526.86 ± 0.24140.97 ± 2.22167.29 ± 2.74212.40 ± 3.89211.02 ± 0.483.39 ± 0.114.40 ± 0.115.22 ± 0.075.85 ± 0.08
9028.40 ± 0.6627.98 ± 0.3528.36 ± 0.1328.53 ± 0.28124.40 ± 1.16145.80 ± 4.29187.15 ± 6.23181.15 ± 6.082.68 ± 0.063.56 ± 0.124.39 ± 0.074.68 ± 0.05
Co2L026.59 ± 0.0226.69 ± 0.7526.90 ± 1.0326.70 ± 1.06269.88 ± 3.21326.37 ± 5.46402.86 ± 9.90410.72 ± 13.429.59 ± 0.2311.66 ± 0.0813.52 ± 0.0915.50 ± 0.35
2525.88 ± 0.2425.51 ± 1.0625.85 ± 0.6226.09 ± 0.60186.76 ± 1.32220.43 ± 3.53277.12 ± 4.00272.91 ± 5.377.44 ± 0.078.79 ± 0.109.50 ± 0.1411.27 ± 0.22
5028.80 ± 0.6528.38 ± 0.3428.76 ± 0.1328.93 ± 0.30164.27 ± 2.23195.83 ± 1.75246.95 ± 0.99239.41 ± 4.236.88 ± 0.098.25 ± 0.059.06 ± 0.169.93 ± 0.12
7526.58 ± 0.4526.39 ± 0.2127.08 ± 0.1627.26 ± 0.25143.97 ± 2.09170.29 ± 2.63215.40 ± 3.74214.02 ± 0.496.39 ± 0.107.40 ± 0.118.22 ± 0.078.85 ± 0.09
9026.94 ± 0.2927.28 ± 0.5326.60 ± 0.1226.90 ± 0.38127.40 ± 1.26148.80 ± 3.90190.15 ± 6.16184.15 ± 5.745.68 ± 0.066.56 ± 0.117.39 ± 0.067.68 ± 0.05
CBA026.79 ± 0.0226.89 ± 0.7427.10 ± 0.9626.90 ± 1.04271.88 ± 3.40328.37 ± 5.60404.86 ± 8.68412.72 ± 13.9811.59 ± 0.2113.66 ± 0.0715.52 ± 0.0917.50 ± 0.34
2526.08 ± 0.2525.71 ± 1.0826.05 ± 0.6126.29 ± 0.62188.76 ± 1.19222.43 ± 3.92279.12 ± 3.71274.91 ± 5.129.44 ± 0.0710.79 ± 0.1011.50 ± 0.1413.27 ± 0.21
5027.14 ± 0.2927.48 ± 0.5726.80 ± 0.1227.10 ± 0.44166.27 ± 2.20197.83 ± 1.93248.95 ± 1.02241.41 ± 4.588.88 ± 0.0910.25 ± 0.0511.06 ± 0.1411.93 ± 0.12
7529.00 ± 0.6228.58 ± 0.3828.96 ± 0.1329.13 ± 0.29145.97 ± 2.17172.29 ± 2.74217.40 ± 3.67216.02 ± 0.468.39 ± 0.109.40 ± 0.1010.22 ± 0.0710.85 ± 0.08
9026.78 ± 0.4326.59 ± 0.2327.28 ± 0.1527.46 ± 0.24129.40 ± 1.13150.80 ± 3.95192.15 ± 6.29186.15 ± 5.917.68 ± 0.068.56 ± 0.129.39 ± 0.089.68 ± 0.05
SIESTA026.49 ± 0.0226.59 ± 0.6926.80 ± 1.1126.60 ± 1.10256.88 ± 3.27313.37 ± 6.13389.86 ± 8.51397.72 ± 14.533.41 ± 0.241.34 ± 0.090.52 ± 0.102.50 ± 0.33
2525.78 ± 0.2425.41 ± 0.9125.75 ± 0.6825.99 ± 0.71173.76 ± 1.26207.43 ± 3.83264.12 ± 3.88259.91 ± 4.515.56 ± 0.074.21 ± 0.103.50 ± 0.141.73 ± 0.24
5028.70 ± 0.6128.28 ± 0.3428.66 ± 0.1428.83 ± 0.27151.27 ± 2.30182.83 ± 2.07233.95 ± 1.00226.41 ± 4.586.12 ± 0.094.75 ± 0.053.94 ± 0.143.07 ± 0.11
7526.48 ± 0.4026.29 ± 0.2026.98 ± 0.1627.16 ± 0.24130.97 ± 2.13157.29 ± 2.85202.40 ± 3.55201.02 ± 0.486.61 ± 0.105.60 ± 0.114.78 ± 0.074.15 ± 0.08
9026.84 ± 0.2527.18 ± 0.5726.50 ± 0.1126.80 ± 0.41114.40 ± 1.27135.80 ± 4.39177.15 ± 6.56171.15 ± 6.627.32 ± 0.066.44 ± 0.125.61 ± 0.075.32 ± 0.05
CIFAR-100ER011.11 ± 0.1111.05 ± 0.0510.99 ± 0.2511.29 ± 0.49212.18 ± 3.27310.04 ± 7.39415.86 ± 5.43522.48 ± 8.128.77 ± 0.0811.40 ± 0.0513.30 ± 0.3816.01 ± 0.20
2510.91 ± 0.2811.06 ± 0.1710.74 ± 0.2010.95 ± 0.28136.59 ± 5.56204.44 ± 5.56271.30 ± 6.21345.11 ± 6.026.07 ± 0.087.85 ± 0.018.97 ± 0.1311.03 ± 0.24
5011.39 ± 0.0911.10 ± 0.1911.18 ± 0.2610.96 ± 0.10126.14 ± 2.17188.36 ± 2.64252.08 ± 5.49306.42 ± 6.705.20 ± 0.106.84 ± 0.027.75 ± 0.189.49 ± 0.16
7511.37 ± 0.1211.09 ± 0.0511.08 ± 0.1711.12 ± 0.07113.18 ± 1.80167.47 ± 2.28219.73 ± 3.96280.12 ± 5.944.53 ± 0.085.76 ± 0.066.76 ± 0.167.72 ± 0.14
9011.15 ± 0.2211.29 ± 0.2510.73 ± 0.3911.03 ± 0.1298.26 ± 1.32146.09 ± 0.90198.31 ± 3.04244.00 ± 0.993.60 ± 0.084.80 ± 0.185.30 ± 0.106.64 ± 0.21
DER011.44 ± 0.1111.31 ± 0.0311.39 ± 0.1411.41 ± 0.17281.76 ± 7.80413.43 ± 10.62560.42 ± 7.66692.27 ± 6.088.99 ± 0.1111.48 ± 0.2413.30 ± 0.1016.13 ± 0.25
2511.29 ± 0.2711.13 ± 0.1711.34 ± 0.0411.31 ± 0.21187.61 ± 5.56281.38 ± 10.59368.16 ± 4.76465.32 ± 15.016.01 ± 0.168.07 ± 0.268.83 ± 0.2510.77 ± 0.17
5011.46 ± 0.2111.50 ± 0.2011.46 ± 0.1711.24 ± 0.30169.31 ± 2.12256.67 ± 10.88332.40 ± 7.17414.51 ± 13.945.16 ± 0.066.61 ± 0.097.67 ± 0.099.14 ± 0.08
7511.39 ± 0.0111.68 ± 0.2911.57 ± 0.2511.59 ± 0.24150.87 ± 2.81231.50 ± 0.41301.56 ± 6.27377.12 ± 4.364.35 ± 0.145.56 ± 0.036.34 ± 0.077.60 ± 0.09
9011.85 ± 0.2811.73 ± 0.3911.52 ± 0.1411.83 ± 0.20130.33 ± 1.59198.37 ± 2.87265.34 ± 4.25334.42 ± 6.343.53 ± 0.024.36 ± 0.105.23 ± 0.076.29 ± 0.11
Co2L011.84 ± 0.1111.71 ± 0.0311.79 ± 0.1311.81 ± 0.19284.76 ± 7.67416.43 ± 11.55563.42 ± 7.11695.27 ± 6.3911.99 ± 0.1114.48 ± 0.2216.30 ± 0.1019.13 ± 0.26
2512.25 ± 0.2912.13 ± 0.4211.92 ± 0.1312.23 ± 0.22190.61 ± 5.16284.38 ± 11.27371.16 ± 4.77468.32 ± 16.079.01 ± 0.1711.07 ± 0.2611.83 ± 0.2513.77 ± 0.17
5011.86 ± 0.2011.90 ± 0.2011.86 ± 0.1711.64 ± 0.29172.31 ± 2.03259.67 ± 9.85335.40 ± 7.52417.51 ± 13.938.16 ± 0.069.61 ± 0.1010.67 ± 0.1012.14 ± 0.09
7511.79 ± 0.0112.08 ± 0.2911.97 ± 0.2511.99 ± 0.24153.87 ± 3.00234.50 ± 0.37304.56 ± 6.04380.12 ± 4.107.35 ± 0.138.56 ± 0.039.34 ± 0.0810.60 ± 0.09
9011.69 ± 0.2811.53 ± 0.1811.74 ± 0.0411.71 ± 0.19133.33 ± 1.59201.37 ± 2.64268.34 ± 4.57337.42 ± 6.076.53 ± 0.027.36 ± 0.118.23 ± 0.089.29 ± 0.11
CBA012.04 ± 0.1011.91 ± 0.0311.99 ± 0.1312.01 ± 0.16286.76 ± 8.10418.43 ± 11.59565.42 ± 7.78697.27 ± 6.3713.99 ± 0.1216.48 ± 0.2518.30 ± 0.1021.13 ± 0.24
2512.45 ± 0.2712.33 ± 0.3812.12 ± 0.1312.43 ± 0.20192.61 ± 5.57286.38 ± 10.48373.16 ± 5.10470.32 ± 14.5411.01 ± 0.1613.07 ± 0.2713.83 ± 0.2615.77 ± 0.16
5012.06 ± 0.2112.10 ± 0.1912.06 ± 0.1811.84 ± 0.31174.31 ± 1.96261.67 ± 11.85337.40 ± 6.67419.51 ± 14.2310.16 ± 0.0611.61 ± 0.0812.67 ± 0.0914.14 ± 0.08
7511.99 ± 0.0112.28 ± 0.2912.17 ± 0.2512.19 ± 0.23155.87 ± 2.72236.50 ± 0.39306.56 ± 6.08382.12 ± 4.749.35 ± 0.1510.56 ± 0.0311.34 ± 0.0812.60 ± 0.09
9011.89 ± 0.2611.73 ± 0.1511.94 ± 0.0411.91 ± 0.19135.33 ± 1.63203.37 ± 2.63270.34 ± 4.38339.42 ± 6.428.53 ± 0.029.36 ± 0.1010.23 ± 0.0611.29 ± 0.10
SIESTA011.74 ± 0.1111.61 ± 0.0311.69 ± 0.1311.71 ± 0.16271.76 ± 8.57403.43 ± 9.63550.42 ± 8.11682.27 ± 6.051.01 ± 0.121.48 ± 0.233.30 ± 0.106.13 ± 0.24
2512.15 ± 0.2612.03 ± 0.4011.82 ± 0.1312.13 ± 0.19177.61 ± 5.32271.38 ± 9.93358.16 ± 4.37455.32 ± 15.893.99 ± 0.171.93 ± 0.261.17 ± 0.230.77 ± 0.17
5011.76 ± 0.2111.80 ± 0.2011.76 ± 0.1811.54 ± 0.27159.31 ± 2.02246.67 ± 11.47322.40 ± 7.19404.51 ± 14.564.84 ± 0.053.39 ± 0.092.33 ± 0.090.86 ± 0.08
7511.69 ± 0.0111.98 ± 0.2711.87 ± 0.2711.89 ± 0.26140.87 ± 2.92221.50 ± 0.42291.56 ± 6.37367.12 ± 4.535.65 ± 0.154.44 ± 0.033.66 ± 0.062.40 ± 0.09
9011.59 ± 0.2711.43 ± 0.1811.64 ± 0.0411.61 ± 0.20120.33 ± 1.51188.37 ± 2.79255.34 ± 4.31324.42 ± 6.786.47 ± 0.025.64 ± 0.104.77 ± 0.073.71 ± 0.11
TinyImageNetER06.92 ± 0.086.89 ± 0.086.91 ± 0.126.94 ± 0.08368.58 ± 1.92560.85 ± 4.75734.58 ± 8.68920.00 ± 25.8512.63 ± 0.1616.01 ± 0.5718.43 ± 0.4923.45 ± 0.55
256.49 ± 0.126.41 ± 0.086.44 ± 0.096.44 ± 0.06247.77 ± 6.69360.69 ± 7.37490.32 ± 9.67603.97 ± 13.018.31 ± 0.1710.46 ± 0.1812.40 ± 0.2916.09 ± 0.21
506.93 ± 0.126.89 ± 0.106.93 ± 0.056.93 ± 0.06212.01 ± 1.89318.21 ± 5.36427.68 ± 15.29539.62 ± 8.986.93 ± 0.209.12 ± 0.2010.56 ± 0.2513.79 ± 0.45
756.79 ± 0.036.81 ± 0.306.89 ± 0.066.82 ± 0.03185.48 ± 5.41278.35 ± 7.84364.66 ± 2.60451.95 ± 1.205.94 ± 0.067.89 ± 0.148.87 ± 0.0111.32 ± 0.22
906.75 ± 0.086.66 ± 0.076.73 ± 0.086.67 ± 0.16153.87 ± 2.00229.73 ± 5.52308.68 ± 6.37383.67 ± 5.694.74 ± 0.046.12 ± 0.027.10 ± 0.179.08 ± 0.25
DER07.13 ± 0.087.16 ± 0.167.30 ± 0.017.07 ± 0.12490.20 ± 10.30734.73 ± 22.86956.09 ± 3.851256.71 ± 26.7814.00 ± 0.1717.96 ± 0.2921.14 ± 0.5725.25 ± 0.48
256.48 ± 0.056.39 ± 0.126.44 ± 0.146.34 ± 0.03329.63 ± 4.67475.95 ± 7.02628.77 ± 1.08796.89 ± 13.519.13 ± 0.1612.08 ± 0.3615.19 ± 0.0918.83 ± 0.30
507.14 ± 0.087.03 ± 0.056.99 ± 0.087.03 ± 0.06284.90 ± 1.63423.41 ± 17.97578.76 ± 2.51714.64 ± 19.627.54 ± 0.169.57 ± 0.1712.41 ± 0.1714.94 ± 0.39
757.07 ± 0.117.02 ± 0.276.97 ± 0.167.04 ± 0.12247.87 ± 3.28370.59 ± 6.77489.26 ± 18.68620.21 ± 12.106.36 ± 0.108.21 ± 0.3010.43 ± 0.1512.79 ± 0.20
907.02 ± 0.057.13 ± 0.117.12 ± 0.187.31 ± 0.07216.29 ± 4.89322.42 ± 5.57428.91 ± 5.35535.12 ± 5.665.23 ± 0.076.89 ± 0.048.64 ± 0.0810.15 ± 0.07
Co2L07.53 ± 0.087.56 ± 0.157.70 ± 0.017.47 ± 0.12493.20 ± 10.16737.73 ± 24.31959.09 ± 3.951259.71 ± 24.7917.00 ± 0.1620.96 ± 0.2724.14 ± 0.5728.25 ± 0.51
256.88 ± 0.056.79 ± 0.116.84 ± 0.136.74 ± 0.03332.63 ± 4.75478.95 ± 6.70631.77 ± 1.09799.89 ± 14.7012.13 ± 0.1715.08 ± 0.3918.19 ± 0.0921.83 ± 0.30
507.54 ± 0.087.43 ± 0.057.39 ± 0.077.43 ± 0.06287.90 ± 1.75426.41 ± 17.14581.76 ± 2.36717.64 ± 19.5810.54 ± 0.1712.57 ± 0.1615.41 ± 0.1617.94 ± 0.43
757.47 ± 0.117.42 ± 0.287.37 ± 0.177.44 ± 0.11250.87 ± 3.51373.59 ± 6.79492.26 ± 20.54623.21 ± 12.599.36 ± 0.0911.21 ± 0.3113.43 ± 0.1615.79 ± 0.19
907.42 ± 0.057.53 ± 0.127.52 ± 0.177.71 ± 0.07219.29 ± 5.17325.42 ± 6.01431.91 ± 5.88538.12 ± 5.148.23 ± 0.089.89 ± 0.0411.64 ± 0.0813.15 ± 0.07
CBA07.73 ± 0.077.76 ± 0.157.90 ± 0.017.67 ± 0.13495.20 ± 9.36739.73 ± 23.27961.09 ± 4.041261.71 ± 26.8519.00 ± 0.1722.96 ± 0.2926.14 ± 0.5530.25 ± 0.50
257.08 ± 0.056.99 ± 0.117.04 ± 0.136.94 ± 0.03334.63 ± 4.25480.95 ± 6.37633.77 ± 1.18801.89 ± 14.3114.13 ± 0.1817.08 ± 0.3720.19 ± 0.0923.83 ± 0.27
507.74 ± 0.087.63 ± 0.057.59 ± 0.077.63 ± 0.06289.90 ± 1.47428.41 ± 18.50583.76 ± 2.32719.64 ± 18.9912.54 ± 0.1714.57 ± 0.1617.41 ± 0.1819.94 ± 0.42
757.67 ± 0.117.62 ± 0.257.57 ± 0.167.64 ± 0.11252.87 ± 3.01375.59 ± 6.98494.26 ± 18.30625.21 ± 11.2111.36 ± 0.0913.21 ± 0.3215.43 ± 0.1417.79 ± 0.19
907.62 ± 0.057.73 ± 0.117.72 ± 0.187.91 ± 0.06221.29 ± 4.41327.42 ± 5.77433.91 ± 5.69540.12 ± 5.9410.23 ± 0.0711.89 ± 0.0413.64 ± 0.0815.15 ± 0.07
SIESTA07.43 ± 0.087.46 ± 0.177.60 ± 0.017.37 ± 0.12480.20 ± 11.08724.73 ± 20.84946.09 ± 3.741246.71 ± 24.974.00 ± 0.167.96 ± 0.2711.14 ± 0.5615.25 ± 0.49
256.78 ± 0.056.69 ± 0.116.74 ± 0.156.64 ± 0.03319.63 ± 4.91465.95 ± 7.39618.77 ± 1.17786.89 ± 13.240.87 ± 0.152.08 ± 0.345.19 ± 0.098.83 ± 0.29
507.44 ± 0.087.33 ± 0.057.29 ± 0.087.33 ± 0.06274.90 ± 1.63413.41 ± 17.65568.76 ± 2.38704.64 ± 18.072.46 ± 0.170.43 ± 0.182.41 ± 0.174.94 ± 0.42
757.37 ± 0.107.32 ± 0.297.27 ± 0.157.34 ± 0.12237.87 ± 3.06360.59 ± 7.15479.26 ± 17.29610.21 ± 11.883.64 ± 0.101.79 ± 0.330.43 ± 0.162.79 ± 0.21
907.32 ± 0.057.43 ± 0.117.42 ± 0.207.61 ± 0.07206.29 ± 5.17312.42 ± 5.06418.91 ± 5.51525.12 ± 5.284.77 ± 0.073.11 ± 0.041.36 ± 0.070.15 ± 0.07
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Park, H.-S.; Chu, H.-C.; Sung, M.-K.; Kim, C.; Lee, J.; Kim, D.-W.; Lee, J. Survey on Replay-Based Continual Learning and Empirical Validation on Feasibility in Diverse Edge Devices Using a Representative Method. Mathematics 2025, 13, 2257. https://doi.org/10.3390/math13142257

AMA Style

Park H-S, Chu H-C, Sung M-K, Kim C, Lee J, Kim D-W, Lee J. Survey on Replay-Based Continual Learning and Empirical Validation on Feasibility in Diverse Edge Devices Using a Representative Method. Mathematics. 2025; 13(14):2257. https://doi.org/10.3390/math13142257

Chicago/Turabian Style

Park, Heon-Sung, Hyeon-Chang Chu, Min-Kyung Sung, Chaewoon Kim, Jeongwon Lee, Dae-Won Kim, and Jaesung Lee. 2025. "Survey on Replay-Based Continual Learning and Empirical Validation on Feasibility in Diverse Edge Devices Using a Representative Method" Mathematics 13, no. 14: 2257. https://doi.org/10.3390/math13142257

APA Style

Park, H.-S., Chu, H.-C., Sung, M.-K., Kim, C., Lee, J., Kim, D.-W., & Lee, J. (2025). Survey on Replay-Based Continual Learning and Empirical Validation on Feasibility in Diverse Edge Devices Using a Representative Method. Mathematics, 13(14), 2257. https://doi.org/10.3390/math13142257

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop