Next Article in Journal
SC-CoSF: Self-Correcting Collaborative and Co-Training for Image Fusion and Semantic Segmentation
Next Article in Special Issue
Intelligent Sports Weights
Previous Article in Journal
Secure and Trusted Crowdsensing for Outdoor Air Quality Monitoring: State of the Art and Perspectives
Previous Article in Special Issue
Tiny Machine Learning and On-Device Inference: A Survey of Applications, Challenges, and Future Directions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

TDA-L: Reducing Latency and Memory Consumption of Test-Time Adaptation for Real-Time Intelligent Sensing

by
Rahim Hossain
,
Md Tawheedul Islam Bhuian
and
Kyoung-Don Kang
*
School of Computing, State University of New York at Binghamton, 4400 Vestal Parkway East, Vestal, NY 13850, USA
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(12), 3574; https://doi.org/10.3390/s25123574
Submission received: 17 April 2025 / Revised: 28 May 2025 / Accepted: 3 June 2025 / Published: 6 June 2025
(This article belongs to the Special Issue Edge AI for Wearables and IoT)

Abstract

Vision–language models learn visual concepts from the supervision of natural language. It can significantly enhance the generalizability of real-time intelligent sensing, such as analyzing camera-captured real-time images for visually impaired users. However, adapting vision–language models to distribution shifts at test time, caused by several factors such as lighting or weather changes, remains challenging. In particular, most existing test-time adaptation methods rely on gradient-based fine-tuning and backpropagation, making them computationally expensive and unsuitable for real-time applications. To address this challenge, the Training-Free Dynamic Adapter (TDA) has recently been introduced as a lightweight alternative that uses a dynamic key–value cache and pseudo-label refinement for test-time adaptation without backpropagation. Building on this, we propose TDA-L, a new framework that integrates Low-Rank Adaptation (LoRA) to reduce the size of feature representations and related computational overhead at test time using pre-learned low-rank matrices. TDA-L applies LoRA transformations to both query and cached features during inference, cost-efficiently improving robustness to distribution shifts while maintaining the training-free nature of TDA. Experimental results on seven benchmarks show that TDA-L maintains accuracy but achieves lower latency, less memory consumption, and higher throughput, making it well-suited for AI-based real-time sensing.

1. Introduction

Vision–language models (VLMs) have emerged as a cornerstone of multimodal AI, enabling systems to process and integrate information from both visual and textual domains [1,2]. These models have demonstrated remarkable capabilities in applications such as image captioning, visual question answering, and object recognition compared with traditional machine learning [3]. By leveraging pre-trained embeddings from large-scale datasets, VLMs bridge the gap between visual understanding and natural language processing, paving the way for advanced sensing applications in real-world scenarios [4].
One of the most impactful applications of VLMs is intelligent sensing, where these models analyze real-time visual data to extract meaningful semantic information. This capability is crucial for a range of applications, including smart surveillance, autonomous navigation, and assistive technology. However, deploying VLMs in dynamic environments poses significant challenges due to domain shifts [5], which occur when the distribution of test-time data differs from that of training data. Domain shifts can be caused by variations in lighting, object orientation, environmental factors, or unseen scenarios, all of which can degrade model performance. These shifts highlight the need for real-time adaptation to ensure robustness and maintain high performance in such unpredictable settings [6].
Test-time adaptation (TTA) has been introduced as a promising solution to address these challenges, enabling models to dynamically adapt to new data distributions at inference time without requiring access to labeled training data. However, existing TTA methods often involve complex optimization steps, including gradient-based fine-tuning and backpropagation, which introduce high computational overhead and latency [7,8,9]. An alternative approach where many devices offload the analysis of real-time images to a centralized cloud is not scalable due to the increased latency and bandwidth consumption.
To overcome these limitations, the TDA (Training-Free Dynamic Adapter) [10] has recently been proposed as a lightweight solution for test-time adaptation. Unlike conventional methods [7,8,9], TDA enables adaptation without any training or backpropagation by leveraging a dynamic key–value (KV) cache that refines pseudo-labels iteratively. This approach allows for fast and efficient adaptation while maintaining high classification accuracy. Additionally, TDA incorporates a negative pseudo-labeling mechanism that mitigates the adverse effects of noisy pseudo-labels by assigning them to negative classes in uncertain cases.
Building on TDA, we propose TDA-L—a test-time adaptation framework that injects Low-Rank Adaptation (LoRA) into the dynamic key–value cache. Although LoRA [11] is usually adopted to trim compute and memory during training, TDA-L exploits it at inference, slashing both cache-related memory usage and latency. Concretely, two low-rank matrices are frozen after offline tuning and then applied as a residual projection, enriching every feature vector without backpropagation. In this way, TDA-L aims to support cost-effective TTA on an edge server rather than relying on a centralized cloud.
Because all similarity searches for cache lookups now occur in this LoRA-adapted space, the system can operate with reduced cache sizes and half-precision (float16) representations while nearly retaining the accuracy of TDA. The result is a dynamic, more robust representation space that supports smaller, lower-precision matrices and cached features, yielding an efficient training-free TTA pipeline desirable for robust real-time intelligent sensing at the edge.
We evaluate TDA-L across five ImageNet variants [12,13,14,15,16] and two other datasets [17,18] against several state-of-the-art approaches—CLIP [1], Tip-Adapter [19], CoOp [9], and TDA [10]. Our results indicate that TDA-L achieves comparable accuracy to TDA. Compared with TDA, the top-1 accuracy is decreased by 1.37–2% (0.76–1.18 percentage points) only, depending on the specific LoRA rank. However, its accuracy is higher than those of the other advanced baselines by 2.77–3.29 percentage points. Also, TDA-L considerably reduces inference latency and memory usage by applying LoRA-based adaptation to feature vectors. Compared with TDA, which provides the highest accuracy, TDA-L reduces memory consumption and latency by approximately 18–20% and 6–10%, respectively. Furthermore, it enhances the inference throughput—the number of images analyzed per second—by 17–22%, depending on ranks.
In summary, this paper makes the following key contributions:
  • We introduce TDA-L, a test-time adaptation framework that integrates LoRA-based feature adaptation with a dynamic KV cache, operating in a training-free setting.
  • We curate a custom dataset by combining 10% of multiple datasets (described in Section 4.1) and learn generalized LoRA matrices offline for diverse domain adaptation. By using a small fraction of samples, we avoid our model memorizing and overfitting them while supporting efficient training of the LoRA module.
  • We demonstrate that TDA-L preserves accuracy while significantly reducing both latency and memory usage, especially with low-rank configurations.
  • We conduct a thorough comparison of TDA-L with state-of-the-art approaches to TTA [9,10,19]—across seven benchmarks, showcasing its robustness under distribution shifts and computational efficiency in terms of latency, memory usage, and throughput, validating the real-world viability of our method on a commodity machine that mimics a cost-effective edge server much less expensive than a cloud server with high-end GPUs and massive resources.
The rest of the paper is organized as follows. Section 2 reviews related literature on vision–language models, test-time adaptation, and lightweight adaptation techniques. Section 3 presents the TDA-L framework, including its LoRA-enhanced adaptation process. Section 4 describes the experimental setup and key evaluation results. Section 5 discusses our limitations and future work issues. Section 6 concludes the paper. In addition, Appendix A provides instructions to download, install, and execute our source code to reproduce the results.

2. Related Work

2.1. Vision–Language Models and Test-Time Adaptation

Vision–language models (VLMs), such as CLIP [1], have significantly advanced generality and usability in computer vision by aligning visual and textual representations based on large-scale natural language supervision available on the Internet. These models have been successfully applied in diverse areas, including image retrieval, scene understanding, human-computer interaction, and autonomous systems [20]. Their ability to generalize across various domains without requiring task-specific fine-tuning makes them particularly attractive for real-world applications. Despite their versatility, VLMs often struggle when deployed in dynamic environments where the distribution of test-time data differs from that of training data [21]. Such distribution shifts can degrade model performance, especially in real-time scenarios where retraining is impractical.
Early works on TTA, such as [22,23,24], perform batch normalization (BN) adaptation to update batch statistics at test time without backpropagation; however, this strategy is purely normalization-based and often fails under large distribution shifts or in high-dimensional regimes. Tent [6] improves upon them by minimizing the entropy of model predictions at test time while performing BN. However, domain adaptation supported by BN and entropy minimization is relatively limited, possibly due to their basic nature. Complementary to these approaches, meta-learning strategies aim to learn an adaptation rule during training so that only a few backpropagation steps are needed at test time. For example, Bartler et al. [25,26] propose a meta-learned update that rapidly aligns the model to unseen domains, treating test-time adaptation itself as a learnable optimization problem. However, meta-learning approaches are significantly more complex and not desirable for real-time adaptation.

2.2. Key–Value Caching for Efficient Test-Time Adaptation

More recently, TTA techniques have evolved to improve model generalization by adjusting predictions to match the characteristics of unseen test distributions. Notable approaches include [27,28,29] that leverage KV caching, retrieval-based learning, and memory-augmented neural networks for efficient TTA. Among them, KV caching is especially cost-effective. The fundamental idea behind KV caching is to store and retrieve useful information from past observations to improve model predictions.
In the context of TTA, KV caches serve as a memory buffer that retains information about test-time distributions, enabling efficient adaptation without modifying model weights. Existing KV caching approaches can be categorized as follows:
  • Static KV Caching: Methods such as Tip-Adapter [27] use a fixed cache of precomputed embeddings from a few-shot labeled dataset. While effective, these approaches cannot adapt to shifting data distributions.
  • Dynamic KV Caching: The Training-Free Dynamic Adapter (TDA) [10] introduces dynamic KV caching, where test-time predictions are stored and progressively updated. This enables continuous adaptation without backpropagation. In particular, TDA maintains two dynamic caches for effective TTA: (1) a positive cache that stores high-confidence pseudo-labels to improve classification and (2) a negative cache that mitigates noisy predictions by storing ambiguous or uncertain test samples.
While dynamic KV caching enables efficient TTA, TDA’s reliance on full-precision floating-point embeddings increases memory consumption, limiting its applicability in real-time edge scenarios. Addressing this limitation is crucial for enabling scalable and efficient adaptation on edge servers with relatively fewer resources compared with a cloud. Pruning [30] and quantization [31] have widely been explored. However, they often reduce accuracy. Neither are they designed to support TTA to enhance the robustness of DNN models in the presence of domain shifts. To address these issues, we apply LoRA in this paper.

2.3. Parameter-Efficient Adaptation with LoRA

LoRA (Low-Rank Adaptation) [11] has recently emerged as a lightweight alternative to full model fine-tuning. Instead of updating all parameters of a neural network, LoRA introduces trainable low-rank matrices into existing layers. These matrices can be trained efficiently with fewer resources and added to the original model outputs.
The LoRA paradigm has been shown to achieve competitive performance in natural language processing and vision–language tasks with minimal computational overhead. It enables parameter-efficient adaptation while preserving the pre-trained model’s weights, making it effective in memory-limited environments.
In TDA-L, we integrate LoRA into the pipeline of TDA [10] via offline fine-tuning of the LoRA matrices using a small portion of diverse datasets to model varying domain shifts. At test time, instead of modifying the full feature space or computing gradients, we apply the learned LoRA transformation to the features before a cache lookup. This allows our method, TDA-L, to maintain the benefits of dynamic cache-based adaptation while considerably reducing inference time and resource usage via half-precision caching and matrix operations.

2.4. Test-Time Adaptation in Resource-Constrained Environments

Real-time AI applications, such as autonomous navigation, robotics, and remote sensing, increasingly require adaptive, robust models that can operate efficiently in resource-constrained edge computing environments [32].
Most existing TTA methods, such as [7,8,9], require computationally intensive optimization steps (e.g., backpropagation). Thus, they are unsuitable for real-time edge deployment. In contrast, TDA-L seamlessly integrates dynamic KV caching and lightweight LoRA adaptation. Doing this enables deep learning models to efficiently adapt to changing test distributions for robust predictions while significantly reducing memory usage and computational overhead. In summary, TDA-L advances the state of the art in TTA as follows:
  • TDA-L does not perform gradient-based fine-tuning and backpropagation that are resource-demanding, different from other TTA methods such as Tent and Tip-Adapter.
  • TDA-L supports dynamic KV caching, which is more effective than the static alternative, similar to TDA. Unlike TDA, it applies LoRA to support efficient adaptation to test-time distributions using lightweight, low-rank matrices.
This makes TDA-L a more practical and scalable solution for TTA at the network edge.

3. Method

In this section, we describe our TDA-L method that extends TDA [10] with Low-Rank Adaptation (LoRA) [11] for efficient adaptation of a pre-trained model at test time. We first discuss how TDA works, followed by the introduction of LoRA, the fine-tuning process for obtaining the LoRA matrices, and the test-time adaptation process in TDA-L.

3.1. TDA Background

In TDA [10], it is assumed that a pre-trained base model with classifier weights W generates feature vectors z from an input image x . The process starts by extracting the feature vector:
z = f θ ( x ) where z 2 = 1 ,
where f θ ( · ) represents the CLIP model with parameters θ and the feature vector z is normalized to unit length. The logits for classification are computed as
logits = z W
where W is the classifier weights.
TDA employs a KV cache C to store feature vectors from previously encountered samples. The positive and negative cache previous feature vectors were classified with both high and low confidence, respectively. In TDA, the positive and negative caches help refine the final logits for predictions:
logits final = z W + α · Δ pos α · Δ neg ,
where α and α are the scalar coefficients for the positive adjustment, Δ pos , and the negative adjustment, Δ neg , respectively.

3.1.1. Positive Key–Value Cache

In TDA, the positive KV cache stores high-confidence pseudo-labels that the model has generated during test-time adaptation. These are features and their associated predictions that the model is confident about, i.e., predictions where the model has high certainty. The purpose of this cache is to strengthen the model’s predictions by reinforcing the associations between image features and their high-confidence classes. Specifically, using the positive cache, TDA enhances the final logits by calculating α · Δ pos in Equation (3) where Δ pos is the correction from the positive cache and α is a scalar, which determines how much weight the model should give to these high-confidence adjustments. The positive cache helps improve classification by progressively reinforcing the model’s confidence in its high-confidence predictions. This is especially important in test-time adaptation, where the model needs to refine its outputs using only the unlabeled data it encounters at runtime.

3.1.2. Negative Key–Value Cache

TDA adopts a negative cache in addition to the positive cache in order to effectively deal with uncertain or low-confidence predictions that can significantly affect final prediction quality. These low-certainty predictions are produced when the features are noisy or do not strongly match any known class. The purpose of this cache is to mitigate the adverse impact of uncertain predictions by assigning negative pseudo-labels or refining the model to avoid making wrong predictions in ambiguous cases.
By looking the negative cache up, TDA computes α · Δ neg , which is the adjustment from the negative cache in Equation (3). In the equation, α is a scalar that controls how much influence the negative cache has on the model’s predictions. The negative cache acts as a filter to prevent overfitting to noise or misclassified examples by adjusting the model’s predictions away from uncertain or noisy areas in the feature space. This correction helps the model avoid degrading its performance by negatively reinforcing uncertain predictions as shown in Equation (3).
Thus, TDA’s novelty lies in its ability to adapt to new data distributions by leveraging these two KV caches to enhance the prediction quality on the fly without requiring additional training at test time. This makes it particularly useful for real-world applications where test-time data may deviate significantly from the training distribution.

3.1.3. Cache Management

TDA’s cache management policy is described in Algorithm 1. Since this policy is applied to both the positive and negative cache, we describe the cache management algorithm for the positive cache in this subsection. When there are M classes, each class can have a maximum of K samples in the cache. Thus, the maximum cache size is M K .
When a new feature vector z with predicted class y ^ , loss , and class probability vector p is encountered, it is inserted into the cache C if it is one of the three following cases:
  • Its predicted class y ^ has previously had no sample in the cache (lines 1–2 in Algorithm 1).
  • The total number of the samples currently in that class is less than K (lines 3–4).
  • The class already has K samples; however, the new sample’s loss is smaller than the highest loss of the cached samples in that class (lines 5–6).
Otherwise, the new sample is dropped (not inserted into the cache). Thereby, the cache management scheme retains no more than the k most reliable samples with the lowest losses for each class.
Algorithm 1: Cache Management for Test-Time Adaptation
Sensors 25 03574 i001
At test time, TDA approximates the loss for each sample using its prediction entropy. This acts as a proxy for uncertainty—higher entropy indicates less reliable predictions. We use this estimate to retain the lowest-entropy samples in the cache.

3.2. TDA-L: TDA with Low-Rank Adaptation

Figure 1 depicts the overall architecture of TDA-L. Its components are described in this subsection. First, the fundamental LoRA concept necessary to understand TDA-L is briefly reviewed. Second, the offline fine-tuning process of LoRA matrices is described. Third, the test-time adaptation procedure of TDA-L using the LoRA matrices learned offline is described. In addition, we discuss how TDA-L is different from TDA in detail.

3.2.1. Low-Rank Adaptation (LoRA)

We apply LoRA to enhance the efficiency of TDA by adapting the feature vector z = f θ ( x ) for an input image x using low-rank matrices A and B without altering the model parameters θ or the classifier weights W .
Given A and B , the feature vector z is transformed as:
z LoRA = z + ( z A ) B ,
where
A R d × r , B R r × d , r d .
Here, d is the feature dimension and r is the rank of A , and B such that r d . By training only the matrices A and B , LoRA enables efficient adaptation with minimal computational cost:
A F + B F W F ,
where · F denotes the Frobenius norm. Thus, LoRA ensure that the adaptation is lightweight and does not require any modifications to the original model weights.

3.2.2. Offline Learning of the LoRA Matrices

In TDA-L, the weights in the low-rank matrices, A and B in Equation (4), are learned offline and remain fixed at test time. At test time, TDA-L only updates the positive and negative caches using A and B to enhance robustness in the presence of domain shifts.
Our offline procedure for learning LoRA matrices is summarized in Algorithm 2. Line 1 initializes the low-rank matrices A and B with a small Gaussian variance to keep the first forward pass close to the frozen backbone:
A N ( 0 , σ 2 ) , B N ( 0 , σ 2 ) where σ 0.01 .
This initialization ensures that training starts from a state close to the original model’s behavior.
For each sample x j in a mini-batch, line 5 extracts the feature z j using the backbone, line 6 derives the LoRA-applied feature z LoRA , j = z j + ( z j A ) B , and line 7 re-normalizes it to unit length. Line 8 multiplies it by the classifier weights to obtain final logits: p j = z LoRA , j W . Line 9 evaluates the cross-entropy loss:
j = CE ( p j , y j ) = log exp ( p y j ) k = 1 N C exp ( p k ) ,
where p j , previously calculated in line 8 of Algorithm 2, represents the vector of the logits derived by our LoRA module, p y j is the logit for the true class y j , N C is the number of classes, and p k p j (the k t h element of the vector p j ).
The batch loss in line 10 of Algorithm 2 averages the individual losses and adds a regularization term:
L ( A , B ) = 1 B j = 1 B j + λ ( A F 2 + B F 2 ) ,
where B is the batch size, λ is the regularization coefficient, and A F 2 + B F 2 is the regularization term.
Since no gradients are backpropagated into the classifier or backbone, only A and B , comprising 2 d r parameters in total, are trained. The optimization step for fine-tuning is given by
A , B arg min A , B L ( A , B ) .
This optimization uses the Adam optimizer as specified in line 11 with a small learning rate (e.g., η = 1 × 10 4 ) and a conservative initialization of the LoRA matrices previously described in line 1.
In line 12, the LoRA module is evaluated on the validation set after each epoch. The entire process is repeated for a specified number of epochs, N _ E p o c h s . The final LoRA matrices are used during test time without further optimization.
Algorithm 2: Algorithm for Fine-Tuning LoRA
Sensors 25 03574 i002

3.2.3. Test Time Adaptation During Inference

At test time, for an input image x , we first extract the normalized feature vector z from the image:
z = f θ ( x ) where z 2 = 1 .
Second, we apply the LoRA transformation to z :
z LoRA = z + ( z A ) B
and normalize it to maintain stability during further computations:
z LoRA z LoRA z LoRA 2 .
Third, we compute the adjustment term based on the similarity between z LoRA and z LoRA k that is the k t h feature vector in the cache:
Δ ( z LoRA , z LoRA k ) = exp β · ( z LoRA · z LoRA k )
where β is a scaling factor, and the similarity is measured by the dot product.
Finally, we compute the fully adjusted logits by incorporating the cache-based adaptations:
logits final ( z LoRA ) = z LoRA W + α k = 1 | C p | Δ ( z LoRA , z LoRA k ) α k = 1 | C n | Δ ( z LoRA , z LoRA k )
where | C p | and | C n | are the sizes of the positive and negative cache, respectively.

3.2.4. Efficient Design of TDA-L to Reduce Latency and Resource Consumption

The main difference between TDA and TDA-L lies in the feature space. In TDA-L, all similarity operations occur in a LoRA-adapted feature space, which is more compact yet more robust and class-separable than the raw CLIP space used in TDA. Thus, TDA-L enables accurate inference with lower-precision tensors and leaner caches compared with TDA. More specifically, TDA and TDA-L differ as follows:
  • TDA caches triples ( z , , p ) in Algorithm 1, where the feature vector z and the probability vector p R N C are the full soft-max vector with float32 precision. However, TDA-L only stores ( z LoRA , ) without storing p . This reduces the memory usage and lookup time of the caches.
  • z LoRA and are float16 in TDA-L. By using half-precision compared with TDA, TDA-L not only saves memory space for the caches but also decreases the latency and memory consumption for matrix operations required for cache lookups in Equation (14) and final logit calculations in Equation (15).
  • To look up the negative cache, TDA requires a masking operation that filters out uncertain cache entries. To do it, TDA concatenates the cached p vectors and thresholds them to build a class-specific mask at every lookup by doing a cache-wide multiplication. In contrast, TDA-L builds a compact one-hot vector from integer labels with no additional multiplications.
Due to this design, TDA-L can considerably reduce memory requirements and latency while providing similar accuracy to TDA.

4. Evaluation

4.1. Datasets

For evaluation, we use seven benchmarks comprising several variants of ImageNet and two other datasets—Caltech101 [17] and UCF101 [18]—that represent various data distributions and domains in the real world:
  • ImageNet [12]: A large-scale dataset with over 1.2 million labeled images spanning 1000 categories. It is commonly used for training and evaluating image classification models and serves as a standard benchmark in computer vision.
  • ImageNet-A [13]: A subset of ImageNet with real-world, unmodified, difficult samples misclassified by ResNet (Residual Neural Network) models. It contains images specifically chosen to challenge the robustness of models trained on ImageNet.
  • ImageNet-S [15]: A dataset designed to test robustness under distributional shifts, especially in terms of style variations. It consists of the same categories as ImageNet but includes significant style differences in images (e.g., black-and-white and sketches).
  • ImageNet-R [14]: A dataset that includes altered ImageNet data. In particular, they are altered with various types of real-world transformations, such as rotations or occlusions. The objective of this dataset is to challenge models to recognize objects under more complex, real-world conditions.
  • ImageNet-V2 [16]: A re-evaluated and re-labeled version of ImageNet, addressing discrepancies in the dataset’s original annotations. It contains images drawn from the same categories; however, they have updated labels and different distributions.
  • Caltech101 [17]: A dataset containing 9,146 images from 101 object categories, including animals, vehicles, and other objects. Unlike ImageNet, Caltech101 images are low resolution and not normalized. It is widely used to evaluate fine-grained recognition tasks and object classification in more constrained environments.
  • UCF101 [18]: A popular human action recognition dataset of 13,320 videos collected from YouTube. UCF101 has 101 categories, extending UCF50 with 50 action classes. It provides a large diversity in terms of actions and large variations in camera rotation, object appearance and pose, object scale, viewpoint, etc.
From each of these datasets, we randomly pick 10% of the samples to curate a relatively small dataset for efficient fine-tuning of the LoRA matrices. We do not use a large fraction of samples to avoid our LoRA module memorizing and overfitting them. The rest of the datasets are used to thoroughly evaluate TDA-L and several state-of-the-art baselines for test time adaptation.

4.2. Experiment Setup

In this paper, TDA-L is compared against several advanced baselines that support test-time adaptation [1,9,10,27] in terms of accuracy, inference time, memory usage, throughput, and GPU utilization.
For fair comparisons, in both TDA and TDA-L, we use three positive cache entries and two negative cache entries per class. For example, in the Imagenet dataset with 1000 classes, we use 3000 entries for the positive cache and 2000 entries for the negative cache. However, the cached feature vectors in TDA are full precision, while those in TDA-L are half-precision, as discussed before.
For performance evaluation, we use a desktop computer to mimic a cost-effective edge server. It consists of commodity hardware components: an Intel Core i7-7820X CPU, 64 GB of RAM, and an NVIDIA GeForce RTX 3080 Ti GPU with 12 GB of GDDR6X memory.

4.3. Accuracy and Latency

Table 1 reports top-1 accuracy achieved by CLIP, three test-time–adaptation baselines—Tip-Adapter, CoOp, and TDA—and TDA-L with rank 16, 8, 4, 2, and 1, respectively. TDA attains the highest mean accuracy of 55.38%, surpassing CLIP, Tip-Adapter, and CoOp. Compared with TDA, TDA-L’s top-1 accuracy is decreased by approximately 1.37–2% (0.76–1.18 percentage points) depending on the ranks. However, it consistently outperforms the other state-of-the-art baselines, as summarized in the table. Even the rank-1 case still delivers a 54.22% mean accuracy, outperforming them. These results confirm that reducing LoRA rank has only a marginal impact on classification quality.
In Table 2, the per-image inference latency of TDA and TDA-L that achieved the best and second-best accuracy is compared. The average latency of TDA is 41.39 ms per image. On the other hand, the average latency of TDA-L decreases as the rank decreases: it is 38.73 ms/image at rank 16 and 37.09 ms/image at rank 1, reducing the per-image latency by 6.4–10.4%. The improvement is visible for every dataset. In particular, the largest absolute gains, 5–9 ms/image, are achieved on the corruption-heavy ImageNet-A/R/S splits.

4.4. Memory Usage, Throughput, and GPU Utilization

Table 3 reports real-time statistics obtained on the validation run using 50,000 images randomly picked from ImageNet. On average, TDA consumes 3108 MB of GPU memory while processing 50.6 images/s. Compared with TDA, TDA-L decreases the GPU memory consumption and utilization by approximately 18–20% and 13.5–29%, respectively, as the rank decreases from 16 to 1. Furthermore, TDA-L enhances throughput compared with TDA by approximately 17–22%.
In summary, TDA-L considerably reduces GPU memory consumption, utilization, and latency while increasing throughput. Although its accuracy is slightly lower than that of TDA, it is higher than those of the other state-of-the-art baselines due to its effective offline fine-tuning of the LoRA module and efficient test-time adaptation using the half-precision LoRA features and caches for robust inference. Moreover, TDA-L provides a range of tradeoffs between inference accuracy and efficiency. Among TDA-L variants with different ranks, TDA-L with rank 16 is optimal when achieving high accuracy without significantly increasing resource consumption, and latency has the highest priority, while TDA-L with rank 1 is most desirable when optimizing latency, throughput, and resource efficiency is critical at the cost of a slight loss in accuracy.

5. Discussion

While TDA-L supports efficient test-time adaptation, it has room for improvement in the future:
  • More Efficient Cache Management: Exploring more efficient cache management techniques, such as adaptive pruning or memory-efficient storage, could help mitigate the memory constraints of the cache. This could also include strategies for dynamically adapting the cache size without sacrificing performance.
  • A Case Study: It will be interesting to evaluate TDA-L in a realistic case study. For example, our system can be extended to classify objects and speak the results to a user who could be visually impaired with low latency. Further optimizing our approach for augmented/virtual reality applications can be an interesting direction as well.
  • Continual Learning of LoRA Matrices: In this paper, the LoRA matrices are trained offline and remain fixed at test time. Robustness could be further enhanced if they are continually updated in real time, especially in case of significant unforeseen domain shifts. A new challenge, however, is the cost of continual updates (deriving gradients and performing backpropagation) at test time, which could disturb ongoing real-time inferences. A thorough investigation to address this challenge is reserved for future work.
  • Other Vision Language Models: CLIP (Contrastive Language–Image Pre-training) [1] takes an innovative approach that directly learns from raw text about images, using web-scale public images and textual annotations, which provide a wide variety of natural language supervision. As a result, it is effective in zero-shot classification and semantic search. Other advanced VLMs support more sophisticated tasks [20,33,34,35]; however, they are more complex and computationally intensive, imposing challenges for real-time/low-latency inference. On the other hand, there are quantized CLIP models [36,37,38] that are more efficient but subject to lower accuracy/robustness. Generally speaking, it is challenging to support robust performance even in the presence of domain shifts while supporting real-time inference at the same time. A rigorous exploration of this issue is reserved for future studies.

6. Conclusions

In this paper, we propose TDA-L, a test-time adaptation framework that integrates Low-Rank Adaptation (LoRA) with the Training-Free Dynamic Adapter (TDA). TDA-L offers a lightweight solution to dynamically adapt features at test time to enhance robustness in the presence of domain shifts without requiring any backpropagation. By leveraging LoRA-based feature transformation and half-precision caches, TDA-L improves the efficiency of test-time adaptation while nearly maintaining accuracy. In our evaluation using seven benchmark datasets, TDA-L substantially improves the inference speed, throughput, and memory efficiency, providing similar accuracy as TDA. Overall, TDA-L represents a promising step toward cost-effective test-time adaptation for robust real-time intelligent sensing at the edge.

Author Contributions

Conceptualization, R.H., K.-D.K. and M.T.I.B.; methodology, R.H., K.-D.K. and M.T.I.B.; software, R.H. and M.T.I.B.; validation, R.H., M.T.I.B. and K.-D.K.; formal analysis, R.H., M.T.I.B. and K.-D.K.; investigation, R.H., M.T.I.B. and K.-D.K.; resources, K.-D.K.; data curation, R.H. and M.T.I.B.; writing—original draft preparation, R.H. and M.T.I.B.; writing—review and editing, K.-D.K.; visualization, R.H. and M.T.I.B.; supervision, K.-D.K.; project administration, K.-D.K.; funding acquisition, K.-D.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded, in part, by National Science Foundation grant CNS-2326796.

Data Availability Statement

Our source code is publicly available at https://github.com/Real-Time-Lab/TDA-L (accessed on 1 April 2025).

Acknowledgments

We appreciate the anonymous reviewers for their constructive reviews and the Editorial Office of MDPI Sensors for processing this paper.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Appendix A. Software Installation, Datasets, and Execution

Appendix A.1. Requirements

Appendix A.1.1. Installation

Follow these steps to set up a Conda environment and install the required dependencies:
  • git clone git@github.com:Real-Time-Lab/TDA-L.git
  • cd TDA-L
  • conda create -n tdal python=3.8.20
  • conda activate tdal
  • # The results are produced with PyTorch 2.4.1 and CUDA 12.1
  • conda install pytorch==2.4.1 torchvision==0.15.1 torchaudio==2.0.1 cudatoolkit=12.1 -c pytorch
  • pip install -r requirements.txt

Appendix A.1.2. Datasets

To prepare the required datasets, please follow the step-by-step instructions provided in the docs/DATASETS.md file. This guide includes instructions for the benchmark datasets used in our experiments.

Appendix A.2. Running TDA-L

Appendix A.2.1. Configuration

The TDA-L hyperparameters are configured via YAML files located in configs/. For each dataset, the file configs/dataset.yaml can be customized to control both cache types:
  • Positive Cache Configuration: A user may tune parameters such as shot_capacity, alpha, and beta to strengthen the influence of reliable predictions.
  • Negative Cache Configuration: In addition to shot_capacity, alpha, and beta, this cache supports entropy_threshold and mask_threshold settings.
These parameters were pre-tuned to yield strong performance across the datasets described in the paper. However, users may explore further tuning to optimize results for other data distributions.

Appendix A.2.2. Execution

Navigate to the scripts/ directory, which contains four bash scripts for automating dataset preparation, LoRA training, and benchmarking. Each script processes datasets sequentially as indicated by slashes (“/”) in the dataset string.
1.
LoRA Dataset Creation
To prepare the data required for LoRA fine-tuning:
  • bash ./scripts/run_create_lora_data.sh
2.
LoRA Fine-Tuning
To train LoRA modules on the extracted features:
  • bash ./scripts/run_train_lora.sh
  • Note: You may need to modify this script to experiment with different ranks or training durations.
3.
Benchmarking
Ensure that the correct lora_weight file is referenced in tda_runner_lora.py at line 194. This file should match the model configuration used during LoRA training.
  • ResNet50 Backbone:
    •     bash ./scripts/run_benchmark_rn50.sh
  • ViT-B/16 Backbone:
    •     bash ./scripts/run_benchmark_vit.sh
This modular execution structure allows flexible experimentation with various architectures and cache configurations.

References

  1. Radford, A.; Kim, J.W.; Hallacy, C.; Ramesh, A.; Goh, G.; Agarwal, S.; Sastry, G.; Askell, A.; Mishkin, P.; Clark, J. Learning Transferable Visual Models From Natural Language Supervision. Int. Conf. Mach. Learn. 2021, 8748–8763. [Google Scholar]
  2. Li, J.; Li, D.; Xiong, C.; Hoi, S.C. BLIP: Bootstrapped Language-Image Pre-training for Unified Vision-Language Understanding and Generation. Proc. IEEE Conf. Comput. Vis. Pattern Recognit. 2022, 12888–12900. [Google Scholar]
  3. Hossain, R.; Samad, M.D. A Hybrid Clustering Pipeline for Mining Baseline Local Patterns in 3D Point Cloud. In Proceedings of the 2021 6th International Conference for Convergence in Technology (I2CT), Maharashtra, India, 2–4 April 2021; pp. 1–6. [Google Scholar] [CrossRef]
  4. Kamath, S.; Singh, D.; LeCun, Y.; Synnaeve, G.; Misra, I.; Carion, N. MDETR: Modulated Detection for End-to-End Multi-Modal Understanding. Proc. IEEE Int. Conf. Comput. Vis. 2021, 1780–1790. [Google Scholar]
  5. Ahmed, S.; Al Arafat, A.; Rizve, M.N.; Hossain, R.; Guo, Z.; Rakin, A.S. SSDA: Secure Source-Free Domain Adaptation. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 1–6 October 2023; pp. 19180–19190. [Google Scholar]
  6. Wang, D.; Shelhamer, E.; Liu, S.; Olshausen, B.; Darrell, T. Tent: Fully Test-Time Adaptation by Entropy Minimization. Proc. Int. Conf. Learn. Represent. 2021. [Google Scholar] [CrossRef]
  7. Shu, M.; Nie, W.; Huang, D.A.; Yu, Z.; Goldstein, T.; Anandkumar, A.; Xiao, C. Test-Time Prompt Tuning for Zero-Shot Generalization in Vision-Language Models. arXiv 2022, arXiv:cs.CV/2209.07511. [Google Scholar]
  8. Feng, C.M.; He, Y.; Zou, J.; Khan, S.; Xiong, H.; Li, Z.; Zuo, W.; Goh, R.S.M.; Liu, Y. Diffusion-Enhanced Test-time Adaptation with Text and Image Augmentation. Int. J. Comput. Vis. 2025, 1–16. [Google Scholar] [CrossRef]
  9. Zhou, K.; Yang, J.; Loy, C.C.; Liu, Z. Learning to Prompt for Vision-Language Models. In Proceedings of the International Conference on Learning Representations (ICLR), Virtual Event, 25–29 April 2022. [Google Scholar]
  10. Karmanov, A.; Guan, D.; Lu, S.; El Saddik, A.; Xing, E. Efficient Test-Time Adaptation of Vision-Language Models. arXiv 2024, arXiv:2403.18293. [Google Scholar]
  11. Hu, E.J.; Shen, Y.; Wallis, P.; Allen-Zhu, Z.; Li, Y.; Wang, L.; Chen, W. LoRA: Low-rank adaptation of large language models. arXiv 2021, arXiv:2106.09685. [Google Scholar]
  12. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. ImageNet: A Large-Scale Hierarchical Image Database. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  13. Hendrycks, D.; Zhao, K.; Basart, S.; Chen, P.; Sharma, Y.; Song, D. Natural adversarial examples. Nat. Mach. Intell. 2021, 3, 741–746. [Google Scholar]
  14. L., R.R.; Zhao, Y.; Werling, N.; Li, X.; Chen, C.H. ImageNet-R: A Dataset of Fine-Grained Visual Categories for Robustness Evaluation. arXiv 2021, arXiv:2103.01457. [Google Scholar]
  15. Oh, S.; Park, S.; Lee, J. Robustness of deep learning-based classifiers under various image corruptions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  16. Recht, B.; Rebecca, R.; Ludwig, S.; Vaishaal, S. Do ImageNet Classifiers Generalize to ImageNet? Int. Conf. Mach. Learn. 2019, 5389–5400. [Google Scholar]
  17. Fei-Fei, L.; Fergus, R.; Perona, P. Learning generative visual models from few training examples: An incremental bayesian approach tested on 101 object categories. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Washington, DC, USA, 27 June–2 July 2004. [Google Scholar]
  18. Soomro, K.; Zamir, A.R.; Shah, M. UCF101: A dataset of 101 human actions classes from videos in the wild. arXiv 2012, arXiv:1212.0402. [Google Scholar]
  19. Zhang, R.; Zhang, W.; Fang, R.; Gao, P.; Li, K.; Dai, J.; Qiao, Y.; Li, H. Tip-adapter: Training-free adaption of clip for few-shot classification. arXiv 2021, arXiv:2111.03930. [Google Scholar]
  20. Lu, J.; Batra, D.; Parikh, D.; Lee, S. ViLBERT: Pretraining Task-Agnostic Visiolinguistic Representations for Vision-and-Language Tasks. arXiv 2019, arXiv:1908.02265. [Google Scholar]
  21. Murshed, R.U.; Dhruba, S.K.; Bhuian, M.T.I.; Akter, M.R. Automated Level Crossing System: A Computer Vision Based Approach with Raspberry Pi Microcontroller. In Proceedings of the 2022 12th International Conference on Electrical and Computer Engineering (ICECE), Dhaka, Bangladesh, 21–23 December 2022; pp. 180–183. [Google Scholar]
  22. Nado, Z.; Mosbach, M.; Farnia, F.; Krishnan, S.; Izmailov, P.; Wilson, A.G. Evaluating test-time adaptation for covariate shift. In Proceedings of the International Conference on Machine Learning, Virtual Event, 13–18 July 2020; pp. 8037–8048. [Google Scholar]
  23. Li, Y.; Grandvalet, Y.; Davoine, F. Revisiting Batch Normalization for Practical Domain Adaptation. In Proceedings of the ICLR Workshop, San Juan, Puerto Rico, 2–4 May 2016. [Google Scholar]
  24. Schneider, S.; Rusak, E.; Tran, N.; Schölkopf, B.; Bethge, M.; Brendel, W. Improving Robustness Against Common Corruptions by Covariate Shift Adaptation. In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), Online, 6–12 December 2020. [Google Scholar]
  25. Bartler, A.; Bühler, A.; Wiewel, F.; Döbler, M.; Yang, B. MT3: Meta Test-Time Training for Self-Supervised Test-Time Adaption. In Proceedings of the 25th International Conference on Artificial Intelligence and Statistics, Valencia, Spain, 28–30 March 2022; pp. 3080–3090. [Google Scholar]
  26. Choi, M.; Choi, J.; Baik, S.; Kim, T.H.; Lee, K.M. Test-Time Adaptation for Video Frame Interpolation via Meta-Learning. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 9615–9628. [Google Scholar] [CrossRef] [PubMed]
  27. Zhang, R.; Zhang, W.; Fang, R.; Gao, P.; Li, K.; Dai, J.; Qiao, Y.; Li, H. Tip-Adapter: Training-free Adaption of CLIP for Few-shot Classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022. [Google Scholar]
  28. Santoro, A.; Bartunov, S.; Botvinick, M.; Wierstra, D.; Lillicrap, T.P. One-shot Learning with Memory-Augmented Neural Networks. In Proceedings of the 33rd International Conference on Machine Learning (ICML), New York, NY, USA, 19–24 June 2016; pp. 1842–1850. [Google Scholar]
  29. Liu, H.; Zhang, P.; Liao, H.; Luo, J.; Xu, Q. Leveraging Key-Value Memory to Mitigate Feature Forgetting in Online Continual Learning. arXiv 2021, arXiv:2110.09430. [Google Scholar]
  30. Samad, M.D.; Hossain, R.; Iftekharuddin, K.M. Dynamic Perturbation of Weights for Improved Data Reconstruction in Unsupervised Learning. In Proceedings of the 2021 International Joint Conference on Neural Networks (IJCNN), Shenzhen, China, 18–22 July 2021; pp. 1–7. [Google Scholar] [CrossRef]
  31. Lin, Z.; Courbariaux, M.; Memisevic, R.; Bengio, Y. Fixed Point Quantization of Deep Convolutional Networks. In Proceedings of the International Conference on Machine Learning, New York, NY, USA, 19–24 June 2016. [Google Scholar]
  32. Hossain, R. A Deep Neural Network for Detecting Spotted Lanternflies Using Energy Efficient Wide Area Networks. Master’s Thesis, State University of New York at Binghamton, Binghamton, NY, USA, 2024. [Google Scholar]
  33. Li, Z.; Wu, X.; Du, H.; Liu, F.; Nghiem, H.; Shi, G. A Survey of State of the Art Large Vision Language Models: Alignment, Benchmark, Evaluations and Challenges. arXiv 2025, arXiv:cs.CV/2501.02189. [Google Scholar]
  34. Zhang, Z.; Boykov, Y. Collision Cross-entropy for Soft Class Labels and Deep Clustering. arXiv 2023, arXiv:2303.07321. [Google Scholar]
  35. Gao, P.; Geng, S.; Zhang, R.; Ma, T.; Fang, R.; Zhang, Y.; Li, H.; Qiao, Y. CLIP-Adapter: Better Vision-Language Models with Feature Adapters. arXiv 2021, arXiv:2110.04544. [Google Scholar] [CrossRef]
  36. Han, S.; Joohee, K. CLIP-VQDiffusion: Langauge Free Training of Text To Image generation using CLIP and vector quantized diffusion model. arXiv 2024, arXiv:2403.14944. [Google Scholar]
  37. Team, O.A. Post-Training Quantization of OpenAI CLIP model with NNCF. 2023. Available online: https://docs.openvino.ai/2023.3/notebooks/228-clip-zero-shot-quantize-with-output.html (accessed on 1 May 2025).
  38. Cullan, M. Quantizing CLIP with ONNX Pt. 1: Smaller, Faster, Feasible? 2021. Available online: https://heartbeat.comet.ml/quantizing-clip-with-onnx-pt-1-smaller-faster-feasible-919966d44dbb (accessed on 7 December 2021).
Figure 1. Overall Architecture of TDA-L that (1) processes input images through an image encoder, (2) applies LoRA to the image features for real-time adaptation, (3) uses the CLIP model for predictions based on the similarity of the image and textual embeddings in the shared latent space, (4) stores adapted features in a KV cache that consists of a positive and negative cache to reinforce high-confidence predictions and filter uncertain predictions, respectively, and (5) refines the final prediction by combining CLIP outputs with information in the KV cache, enabling efficient, real-time adaptation without retraining at test time. The main difference between TDA and TDA-L is the LoRA module and resulting half-precision (float16) entries stored in the positive and negative caches (where the caches are used for efficient test-time adaptation without backpropagation). This reduces the memory requirements of the caches and the computational complexity of matrix operations for cache lookups and logit calculations.
Figure 1. Overall Architecture of TDA-L that (1) processes input images through an image encoder, (2) applies LoRA to the image features for real-time adaptation, (3) uses the CLIP model for predictions based on the similarity of the image and textual embeddings in the shared latent space, (4) stores adapted features in a KV cache that consists of a positive and negative cache to reinforce high-confidence predictions and filter uncertain predictions, respectively, and (5) refines the final prediction by combining CLIP outputs with information in the KV cache, enabling efficient, real-time adaptation without retraining at test time. The main difference between TDA and TDA-L is the LoRA module and resulting half-precision (float16) entries stored in the positive and negative caches (where the caches are used for efficient test-time adaptation without backpropagation). This reduces the memory requirements of the caches and the computational complexity of matrix operations for cache lookups and logit calculations.
Sensors 25 03574 g001
Table 1. Top-1 accuracy of CLIP, Tip-Adapter, CoOP, TDA, and TDA-L with rank 16, 8, 4, 2, and 1.
Table 1. Top-1 accuracy of CLIP, Tip-Adapter, CoOP, TDA, and TDA-L with rank 16, 8, 4, 2, and 1.
DatasetCLIPTip-A.CoOpTDATDA-L (16)TDA-L (8)TDA-L (4)TDA-L (2)TDA-L (1)
ImageNet59.81%62.03%63.33%61.32%61.32%60.32%60.31%60.33%60.33%
ImageNet-A23.06%23.13%23.06%30.85%30.04%30.04%30.02%30.28%30.28%
ImageNet-R60.72%60.35%56.60%62.30%62.19%62.19%62.18%62.04%62.04%
ImageNet-S21.48%21.74%20.67%24.45%23.10%23.01%23.00%23.17%23.17%
ImageNet-V252.91%53.97%55.40%55.32%55.32%54.77%54.75%54.58%54.58%
Caltech-10187.26%87.53%86.53%89.45%88.02%87.87%87.87%87.87%87.87%
UCF-10159.48%59.55%59.05%64.02%62.33%61.27%61.27%61.27%61.27%
Average52.10%52.61%52.09%55.38%54.62%54.21%54.20%54.22%54.22%
Table 2. Per-image inference latency of TDA and TDA-L with rank 16, 8, 4, 2, and 1.
Table 2. Per-image inference latency of TDA and TDA-L with rank 16, 8, 4, 2, and 1.
DatasetTDATDA-L (16)TDA-L (8)TDA-L (4)TDA-L (2)TDA-L (1)
ImageNet18.45 ms16.27 ms15.71 ms15.43 ms15.42 ms15.38 ms
ImageNet-A63.88 ms60.36 ms58.68 ms60.24 ms58.62 ms58.61 ms
ImageNet-R64.05 ms59.42 ms58.52 ms56.14 ms56.14 ms56.14 ms
ImageNet-S68.23 ms67.79 ms67.23 ms64.78 ms64.71 ms64.75 ms
ImageNet-V263.08 ms55.34 ms53.27 ms54.77 ms53.79 ms52.97 ms
Caltech-1016.02 ms5.98 ms6.00 ms5.92 ms5.90 ms5.90 ms
UCF-1016.00 ms5.94 ms5.94 ms5.86 ms5.90 ms5.90 ms
Average41.39 ms38.73 ms37.91 ms37.59 ms37.21 ms37.09 ms
Table 3. Comparison of Memory Consumption, Throughput, and GPU Utilization.
Table 3. Comparison of Memory Consumption, Throughput, and GPU Utilization.
MethodDuration (Minute)Avg. Memory Usage (MB)Avg. GPU Utilization (%)Throughput (Images/s)
TDA [10]53107.7130.8050.55
TDA-L (rank 16)52535.1726.6359.23
TDA-L (rank 8)5251224.1860
TDA-L (rank 4)52495.1723.2261.50
TDA-L (rank 2)52489.1223.1961.62
TDA-L (rank 1)52476.3723.2061.63
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hossain, R.; Islam Bhuian, M.T.; Kang, K.-D. TDA-L: Reducing Latency and Memory Consumption of Test-Time Adaptation for Real-Time Intelligent Sensing. Sensors 2025, 25, 3574. https://doi.org/10.3390/s25123574

AMA Style

Hossain R, Islam Bhuian MT, Kang K-D. TDA-L: Reducing Latency and Memory Consumption of Test-Time Adaptation for Real-Time Intelligent Sensing. Sensors. 2025; 25(12):3574. https://doi.org/10.3390/s25123574

Chicago/Turabian Style

Hossain, Rahim, Md Tawheedul Islam Bhuian, and Kyoung-Don Kang. 2025. "TDA-L: Reducing Latency and Memory Consumption of Test-Time Adaptation for Real-Time Intelligent Sensing" Sensors 25, no. 12: 3574. https://doi.org/10.3390/s25123574

APA Style

Hossain, R., Islam Bhuian, M. T., & Kang, K.-D. (2025). TDA-L: Reducing Latency and Memory Consumption of Test-Time Adaptation for Real-Time Intelligent Sensing. Sensors, 25(12), 3574. https://doi.org/10.3390/s25123574

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop