Abstract
Knowledge distillation (KD) enables the training of lightweight yet effective models, particularly in the visual domain. Meanwhile, reinforcement learning (RL) facilitates adaptive learning through environment-driven interactions, addressing the limitations of KD in handling dynamic and complex tasks. We propose a novel two-stage framework integrating Knowledge Distillation with Reinforcement Learning (KDRL) to enhance model adaptability to complex data distributions, such as remote sensing and medical imaging. In the first stage, supervised fine-tuning guides the student model using logit and feature-based distillation. The second stage refines the model via RL, leveraging confidence-based and cluster alignment rewards while dynamically reducing reliance on task loss. By combining the strengths of supervised knowledge distillation and reinforcement learning, KDRL provides a comprehensive approach to address the dual challenges of model efficiency and domain heterogeneity. A key innovation is the introduction of auxiliary layers within the student encoder to evaluate and reward the alignment of the characteristics with the teacher’s cluster centers, promoting robust feature learning. Our framework demonstrates superior performance and computational efficiency across diverse tasks, establishing a scalable design for efficient model training. Across remote sensing benchmarks, KDRL boosts the lightweight CLIP/ViT-B-32 student to 69.51% zero-shot accuracy on AID and 80.08% on RESISC45; achieves state-of-the-art cross-modal retrieval on RSITMD with 67.44% (I→T) and 74.76% (T→I) at R@10; and improves DIOR-RSVG visual-grounding precision to 64.21% at Pr@0.9. These gains matter in real deployments by reducing missed targets and speeding analyst search on resource-constrained platforms.
1. Introduction
Knowledge distillation (KD) [1,2,3] has emerged as a powerful technique for transferring knowledge from large, high-capacity teacher models to compact, efficient student models. By leveraging soft target distributions and feature-based guidance, KD enables smaller models to achieve competitive performance while reducing computational costs. This paradigm is particularly valuable for resource-constrained environments, where deploying large-scale models is impractical.
Deep learning models have been increasingly applied to complex domains, where high-dimensional, multimodal data pose significant challenges. Remote sensing tasks—such as land cover classification, object detection, and image retrieval—require models capable of handling diverse spatial and spectral features [4,5,6]. Similarly, medical imaging involves high-stakes decision-making, privacy constraints, and substantial variations in data quality across institutions [7,8,9]. In both domains, knowledge transfer from high-performing teacher models can improve generalization and robustness, making KD a promising solution.
However, traditional KD methods have notable limitations when applied to these complex domains. Conventional KD operates with fixed objectives, assuming that the knowledge transfer process is uniform across all instances. This assumption does not hold for remote sensing and medical imaging, where data heterogeneity, domain shifts, and task complexity significantly affect knowledge transfer effectiveness. For example, in remote sensing, satellite images vary based on geographical regions, acquisition conditions, and sensor types, making static KD objectives suboptimal. In medical imaging, factors such as inter-patient variability and data imbalance can degrade the effectiveness of standard KD strategies.
Scientific problem: In practice, remote sensing and medical imaging pipelines often require lightweight models that run under strict latency and memory budgets, yet must remain reliable under domain shift (e.g., seasonal changes, sensor differences, and heterogeneous scene layouts). The central problem we address is how to transfer the generalization and robust representations of a large vision–language teacher into a compact student without over-regularizing the student or amplifying teacher biases.
Literature gap: Most KD approaches for vision tasks rely on static objectives (logit/feature matching) with fixed weights applied uniformly across samples and tasks, while RL-based KD studies typically embed RL inside the KD loop to adjust distillation signals, rather than using RL to refine a trained student using task-aware rewards. This leaves a gap: an objective mechanism to adapt the student after KD toward downstream task criteria, while keeping inference efficient.
These challenges highlight the need for an adaptive knowledge transfer mechanism that dynamically refines student models based on task-specific requirements. Relying solely on KD can lead to suboptimal feature representations and degraded performance in cases where the student model struggles to generalize across complex data distributions.
Reinforcement learning (RL) offers an alternative approach by enabling models to learn adaptive policies through interaction with an environment [10,11,12]. Recent advances, such as REFT [13], have demonstrated the potential of RL in fine-tuning pre-trained models for specific tasks. While these approaches optimize model performance by adapting learning strategies dynamically, their application to KD remains underexplored. Integrating RL into KD offers a promising direction for overcoming the limitations of traditional knowledge transfer, allowing student models to refine their learning process based on reinforcement-driven feedback.
To address these challenges, we introduce a novel two-stage framework that integrates KD with RL to refine student models through reward-driven optimization. The first stage employs supervised learning with KD, ensuring alignment between the student and teacher models. The second stage introduces reinforcement learning to dynamically refine knowledge transfer, based on our carefully designed reward mechanisms, including cluster-based rewards and confidence-based rewards. During the fine-tuning process, dynamic KD-RL adjustment is employed to assign varying levels of importance to KD and RL.
By addressing the fundamental challenges of KD through reinforcement-driven optimization, our research establishes a more adaptive and scalable knowledge transfer paradigm, significantly improving performance in complex real-world applications. Our work makes the following key contributions:
- Integration of KD and RL: We propose the first comprehensive framework that unifies KD with RL for adaptive student model refinement.
- Dynamic and Reward-Driven Training: Our multi-reward RL mechanism enables task-specific optimization beyond static KD.
- Scalable and Generalizable Framework: The method is designed to handle diverse domains such as remote sensing and medical imaging.
2. Related Work
This section reviews knowledge distillation and reinforcement learning studies most relevant to lightweight vision models and highlights the literature gap addressed by KDRL. Most KD approaches for vision tasks optimize fixed combinations of logit and feature matching losses; their weights are typically static and applied uniformly across samples. Meanwhile, KD + RL hybrids mainly use RL to select teachers, choose distillation routes, or adjust distillation strength within the KD loop. In contrast, KDRL adopts a two-stage strategy: a strong KD initialization is followed by reward-driven refinement that directly optimizes task-aware criteria (cluster alignment and confidence), while keeping inference unchanged.
2.1. Knowledge Distillation
KD has been extensively studied as a model compression technique. Hinton [1] first introduced the foundational concept of distilling knowledge from a large teacher model to a smaller student model by matching logits. Recent advancements, such as feature-based KD and attention transfer, have expanded the applicability of KD to tasks like object detection [14,15] and semantic segmentation [16,17]. In addition, advanced specific KD strategies/techniques have also been proposed for specific domains [18,19] or solving specific problems [20]. For example, TAKD [21] proposes a target-aware KD method to create a lightweight remote sensing scene classification model, effectively distilling target-specific knowledge from RS images, which often exhibit large variations in object scales. RoS-KD [22] introduces a robust stochastic KD framework to incorporate additional smoothening in the distillation step for noisy medical imaging, where knowledge is distilled from multiple teacher models to handle noisy labels effectively. The impressive results highlight the promising potential of KD in specialized domains such as remote sensing and medical imaging.
2.2. Reinforcement Learning
RL has shown promise in tasks requiring sequential decision making, such as object tracking, robotic manipulation, and games [23]. With the rapid development of large foundation models, some research has focused on leveraging these pre-trained foundation models to generate dense rewards for RL tasks [24,25]. Recently, the application of RL has expanded into fine-tuning the pre-trained models. A prominent example is REFT [13], which utilizes RL to fine-tune pre-trained models for specific tasks, effectively guiding the model to adapt to task-specific challenges. Another work [26] proposes to use RL techniques that align models with a task reward to tune vision models, addressing the issue of misalignment between model predictions and the task risk. In contrast to the work that applies RL in the fine-tuning phase, our approach integrates RL with KD, where we provide dynamic rewards based on the alignment of student features with teacher cluster centers.
2.3. Integration of Knowledge Distillation and Reinforcement Learning
Several recent studies have explored combining KD with RL to mitigate the limitations of static distillation schedules and weights. For example, Lee et al. (2021) [27] explicitly introduce RL into the distillation process for vision tasks, using reinforcement-driven signals to stimulate a student’s ability to absorb teacher knowledge beyond fixed hyperparameter settings. RLKD [28] further extends this direction by leveraging RL signals to guide instance-level distillation decisions.
These approaches are conceptually close in that RL is used to control the KD procedure (e.g., how/when to distill, or which distillation signals to emphasize). In contrast, KDRL adopts a two-stage design: Stage 1 performs standard supervised KD to obtain a stable student initialization, while Stage 2 performs a separate reward-driven refinement with rewards derived from the teacher’s representation geometry (cluster centers) and prediction confidence. This decoupling avoids embedding an RL controller directly into the KD loop, and it is particularly practical for remote sensing settings where the teacher is fixed and reward computation can be implemented as a lightweight auxiliary objective.
Overall, prior KD + RL studies motivate the use of adaptive reinforcement signals during distillation; our contribution is to instantiate these signals as cluster- and confidence-based rewards and to integrate them into a two-stage pipeline tailored to lightweight student models in remote sensing and medical imaging.
3. Methodology
3.1. Problem Definition
Deep learning models have demonstrated remarkable performance across various domains. However, deploying these models in real-world applications is challenging due to their high computational demands, memory consumption, and limited scalability. Domains such as remote sensing and medical imaging exacerbate these challenges due to unique characteristics: remote sensing data involves diverse spatial resolutions and temporal variations, while medical imaging data exhibits significant heterogeneity from variations in modalities (e.g., CT and MRI) and patient demographics.
To address these issues, KD has emerged as a promising technique. In KD, a large pre-trained teacher model T transfers its knowledge to a smaller student model S, which is designed to be lightweight and efficient. This process typically involves minimizing the difference between the teacher and student outputs using the following objectives:
- Logit Matching: The student’s output logits are aligned with the teacher’s, enabling S to emulate T’s decision boundaries.
- Feature Alignment: Intermediate feature representations of S are aligned with T, facilitating better transfer of structural information.
- Task-Specific Supervision: A task loss (e.g., classification or detection) ensures that S learns task-relevant features.
Despite its success, traditional KD methods often struggle when the student capacity is very small or when the downstream data distribution deviates from the teacher’s pre-training domain. To better adapt the student beyond static KD, we introduce an RL-inspired formulation for the refinement stage. Concretely, we treat each mini-batch update as a one-step contextual bandit: the student produces intermediate representations (state), a lightweight refinement module (RLayer) produces an additive residual (action), and a composite reward guides the refinement.
- State: the intermediate representations extracted by the student model (e.g., transformer block outputs).
- Action: an additive refinement produced by RLayer, applied to the student representation (Section 3.4).
- Reward: a composite score that encourages (i) alignment to the teacher representation geometry (cluster-based reward) and (ii) confident predictions (confidence-based reward).
Accordingly, we maximize the expected reward and minimize its negation:
where and balance the reward components. Note that this refinement is deterministic and differentiable, so we optimize directly with standard backpropagation and do not require policy gradients or an explicit actor–critic.
The final training objective combines KD and reward-driven refinement:
where is a time-dependent weight (defined in Section 3.4) that decays from KD to refinement.
This problem setting emphasizes the need for a unified framework that integrates KD and RL to handle the challenges of heterogeneous data distributions in remote sensing and medical imaging.
3.2. Knowledge Distillation with Reinforcement Learning (KDRL)
In terms of a framework overview, KDRL consists of two sequential stages operating on a frozen teacher T and a lightweight student S: (i) KD initialization, where S learns from T using logit and feature matching plus task loss; and (ii) reward-driven refinement, where small refinement layers (RLayer) update intermediate student representations to better satisfy task-aware rewards. The framework depends on (a) a pre-trained vision–language teacher (e.g., RemoteCLIP) to provide high-quality embeddings and logits, and (b) teacher feature clusters computed once (offline) to define a stable cluster-alignment reward.
- Main components.
- Teacher (frozen): provides logits and embedding features, used to compute cluster centers.
- Student : the compact image encoder to be deployed at inference time.
- RLayer (train-time only): lightweight MLP/residual adapters inserted after selected student blocks to produce refinement actions.
- Reward modules: a cluster-alignment reward (matching teacher cluster assignments/centroids) and a confidence reward (encouraging confident, teacher-consistent predictions).
- Dynamic schedule: gradually decays KD weight and increases reward-driven emphasis to stabilize optimization.
General flow. Compute teacher cluster centers offline; train S with KD (Stage 1); then enable RLayer and optimize the Stage 2 objective with dynamic KD–RL weighting; finally, discard RLayer for deployment so the inference cost matches the student backbone.
To address the challenges identified in the previous section, we propose Knowledge Distillation with Reinforcement Learning (KDRL), a novel framework designed to enhance model efficiency and adaptability in domains with complex and heterogeneous data distributions. An overview of KDRL is illustrated in Figure 1. The KDRL framework is systematically divided into two distinct training stages: (1) supervised fine-tuning with knowledge distillation and (2) reinforcement-based refinement.
Figure 1.
Overview of our Knowledge Distillation with Reinforcement Learning (KDRL) framework. KDRL operates in two stages: Stage 1 (top) involves supervised fine-tuning, where the model is initialized using knowledge distillation (KD) and task-specific loss to establish a robust baseline. Stage 2 (bottom) incorporates reinforcement learning (RL) for further refinement, leveraging RL-based losses such as cluster loss and confidence loss to enhance the model’s performance and generalizability. This two-stage process ensures an effective balance between initial knowledge transfer and adaptive optimization. Teacher cluster centers are computed offline by clustering teacher embeddings and are used as fixed anchors during Stage 2 to compute the cluster-based reward (as indicated by the arrows).
The supervised fine-tuning with knowledge distillation stage focuses on training an initial student model under the guidance of a pre-trained teacher model. This stage leverages classical KD techniques to align the student model’s outputs and intermediate feature representations with those of the teacher. The objective is to enable the student model to capture essential task-relevant information efficiently, thereby building a solid foundation for subsequent refinement.
The reinforcement-based refinement stage aims to further optimize the student model by incorporating task-specific adaptability and domain-aware decision-making. In this stage, the student model is treated as an agent within a RL framework, where the state is defined by the model’s learned representations, actions correspond to parameter updates, and the reward function provides feedback based on task-specific criteria, such as domain alignment, or other custom metrics. This stage allows the model to dynamically refine its parameters, enhancing performance on diverse and unseen data distributions.
By combining the strengths of supervised knowledge distillation and reinforcement learning, KDRL provides a comprehensive approach to address the dual challenges of model efficiency and domain heterogeneity. The two stages are seamlessly integrated, with the KD phase laying a strong groundwork for the RL stage, and the refinement phase adding adaptive capabilities to the student model. This ensures that the final model not only achieves high task performance but also generalizes effectively across various data domains.
3.3. Stage 1: Supervised Fine-Tuning with Knowledge Distillation
The first stage of our framework focuses on initializing the student model S using KD from the teacher model T. This step ensures that S effectively captures the essential knowledge from T and establishes a robust foundation for downstream tasks.
Logit Matching: The logit outputs of the teacher and student are aligned using Kullback–Leibler (KL) divergence:
where denotes the softmax function, and is the temperature parameter controlling the smoothness of the output distribution.
Feature Alignment: The intermediate features extracted by the teacher () and student () are aligned using a mean squared error (MSE) loss:
Task-Specific Loss: The student is trained to optimize a supervised task loss (e.g., cross-entropy for classification):
where is the student’s prediction for input , and is the ground truth label.
The overall objective in the KD stage is a weighted combination of these losses:
where are hyperparameters. Unless otherwise stated, we tune on a held-out validation split using a small grid search and use the default setting , , and across datasets, which we found to be stable and competitive.
The overall training objective for this stage is a weighted combination of these losses, balancing knowledge transfer and task-specific learning.
3.4. Stage 2: Reinforcement-Based Refinement
The second stage refines the student model S beyond static KD by optimizing a reward-shaped objective. As described in the problem definition, we treat each mini-batch update as a one-step contextual bandit: the student produces intermediate representations (state), a lightweight refinement layer produces an additive residual (action), and a composite reward guides the refinement.
3.4.1. Refinement Layer (RLayer)
RLayer is a lightweight two-layer MLP (Linear–GELU–Linear) applied to the output of a student block. Let denote the representation at block ℓ. RLayer produces an action , and the refined representation is computed via a residual connection:
In practice, we attach RLayer to the last few student blocks (Figure 1) and fine-tune the student jointly with these refinement parameters during Stage 2.
3.4.2. Cluster-Based Reward
To encourage the student representation to preserve the teacher’s geometric structure, we first compute teacher cluster centers offline by applying k-means on teacher embeddings. During Stage 2, each refined student embedding is assigned to its nearest teacher center (cosine distance), and we define the cluster-based reward as cosine similarity:
We optimize this reward by minimizing .
3.4.3. Confidence-Based Reward
We further encourage confident predictions that remain consistent with the teacher. Let and be teacher and student probabilities (temperature ). We define the teacher pseudo-label and compute:
Similarly, we optimize this term by minimizing .
3.4.4. Stage-2 Objective and Dynamic KD–RL Schedule
The Stage 2 refinement loss combines task supervision with the two reward-shaped terms:
where and balance the cluster and confidence components. We then blend Stage 1 KD and Stage 2 refinement with a decaying weight
and optimize
Hyperparameters. We tune and on a validation split and report the best configuration. Unless otherwise stated, we use , , and as defaults, which were stable across the benchmarks. Stage 2 is run for a small number of refinement epochs (5 by default), so the additional training overhead is modest (Section 4.2).
4. Experimental Results
4.1. Implementation Details and Setup
To evaluate the performance of the proposed KDRL framework, we conduct two comprehensive experiments: (1) zero-shot image classification, and (2) image–text retrieval. These experiments are designed to systematically assess the effectiveness of KDRL across diverse tasks and scenarios.
For all tasks, we utilize RemoteCLIP/ViT L-14 as the teacher model due to its robust generalization capabilities in the remote sensing domain. The student model, CLIP/ViT B-32, serves as a lightweight counterpart to evaluate the efficiency of the proposed knowledge distillation process in transferring knowledge and capabilities from the larger teacher model. This setup follows the evaluation protocols of recent remote sensing CLIP baselines, enabling a fair comparison while isolating the contribution of our two-stage KD + RL refinement.
Reproducibility and variance: Unless otherwise stated, we report the primary metrics from a single run with a fixed random seed for each setting, consistent with common practice in recent CLIP-style remote sensing benchmarks and to keep the large-scale training budget tractable. Importantly, our “RL” stage is implemented as reward-shaped optimization on minibatches (no stochastic environment rollouts); the rewards are deterministic given the batch and current model parameters. Together with the decaying schedule , this greatly stabilizes training compared to policy-gradient RL in interactive environments. We provide additional discussion of variance sources and practical mitigation (seed fixing, reward normalization, and early stopping) in Section 5.
4.2. Efficiency Analysis
We quantify efficiency in terms of (i) inference-time compute of the deployed student model and (ii) the extra training-time overhead introduced by Stage 2. Since the text prompts are fixed at inference, their embeddings are pre-computed and cached; therefore, online cost is dominated by the image encoder. Table 1 summarizes the parameter count and multiply–accumulate operations (MACs) of the teacher (ViT-L/14) and student (ViT-B/32) image encoders at resolution, computed analytically from the backbone configuration.
Table 1.
Teacher vs. student model footprint (image encoder only, ). N is the number of tokens including the class token. MACs are reported per forward pass; smaller is better.
- Training-time overhead. Inference-time cost is unchanged because RLayer is only used during Stage 2 training and is removed at inference, while the student backbone remains ViT-B/32. The refinement stage adds a small training overhead: we run Stage 2 for 5 epochs by default, corresponding to roughly 10–20% of the total training steps in our experiments. Moreover, the per-step overhead of reward computation is negligible (<2%) since teacher cluster centers are pre-computed and the reward terms consist of lightweight cosine-similarity and probability lookups. Overall, KDRL incurs a modest additional training cost compared with standard KD/fine-tuning, while keeping inference efficiency identical to the student backbone.
The deployed KDRL model therefore retains the compact ViT-B/32 backbone at inference, yielding a reduction in parameters and an reduction in image-encoder MACs relative to the ViT-L/14 teacher. Stage 2 adds training cost but does not require expensive environment interaction: reward computation uses pre-computed teacher cluster centers and batch-level statistics, adding only lightweight operations on embeddings. As a result, KDRL targets edge-friendly inference while using a moderate, offline refinement budget during training.
4.3. Zero-Shot Classification on Remote Sensing Dataset
This subsection reports zero-shot remote sensing classification results to evaluate how well the distilled student generalizes without task-specific training. We first summarize the datasets and evaluation protocol (Section 4.3), then present the benchmark comparison and provide a critical interpretation of dataset-specific gains and degradations.
We consider widely used remote sensing benchmarks that vary in spatial resolution, geographic diversity, and scene semantics. This diversity is important for evaluating whether KDRL improves robustness under domain shift rather than overfitting to a single dataset style.
4.3.1. Public Datasets for Image Classification
For completeness, we summarize additional remote sensing classification datasets used in subsequent supervised or ablation experiments, including their class counts, splits, and evaluation conventions.
To assess the effectiveness of the proposed KDRL framework for zero-shot classification, we conduct experiments on a variety of benchmark datasets widely used in the remote sensing domain: AID [29], EuroSAT [30], fMoW [31], Million-AID [32], PatternNet [33], NWPU-RESISC45 [34], and RSI-CB256 [35]. These datasets encompass a diverse range of scenes and environments, providing a rigorous evaluation ground for the generalization capabilities of the proposed framework. The datasets were curated, prepared, and made available through SkyScript [36], a comprehensive platform for remote sensing data processing and analysis. To ensure robust zero-shot performance, our model is fine-tuned on the SkyScript training dataset, which is specifically designed to enhance generalization across diverse classes and domains. Classification performance is quantified using accuracy, a standard metric that measures the proportion of correctly classified instances across all categories. The use of multiple datasets ensures a thorough evaluation, capturing the adaptability and scalability of KDRL for zero-shot tasks.
4.3.2. Ablation Study
We conducted an ablation study to assess the individual and combined contributions of the key components of the KDRL framework. Specifically, we evaluated the impact of the supervised fine-tuning (SFT) stage, which includes logit matching, feature alignment, and task-specific loss, and the reinforcement-based refinement (ReFT) stage, which incorporates confidence-based rewards, cluster-based rewards, and dynamic KD loss decay. The results of this ablation study are summarized in Table 2.
Table 2.
Ablation study evaluating the contributions of individual components in the proposed framework. The table compares various configurations of KD and RL techniques. The model configurations are tested on CLIP variants (ViT L-14 and ViT B-32), showing the impact of different methods on overall accuracy. Here, BB denotes the backbone architecture.
As shown in Table 2, the baseline models demonstrate that CLIP with ViT L-14 achieves an accuracy of 61.30%, while ViT B-32 achieves 53.02%. These models serve as the baseline for comparison. When only logit matching is applied (KD logit), the student model’s accuracy improves to 55.98%, demonstrating the benefit of aligning logits. Similarly, applying feature alignment (KD feature) alone increases accuracy to 56.36%, showing that aligning intermediate features enhances performance. Combining both logit and feature alignment (KD full) results in a further accuracy improvement of 56.92%, highlighting the value of both components in refining the student model.
Introducing RL via confidence-based rewards (RL Confidence) results in a slight accuracy increase to 53.46%, suggesting that RL provides guidance for the student model. Adding cluster-based rewards (RL Cluster) yields a larger accuracy boost to 56.70%, indicating that aligning the features with the teacher’s cluster centers has a significant positive impact on performance. When both confidence-based and cluster-based rewards are combined (RL full), the accuracy reaches 56.93%, further refining the student model.
The use of dynamic KD with RL (Dynamic RL) leads to an accuracy of 57.34%, demonstrating the benefits of gradually transitioning from KD to RL to improve generalization. Finally, combining all components of the KDRL framework—logit matching, feature alignment, confidence-based reward, cluster-based reward, and dynamic KD loss decay—achieves the highest accuracy of 59.22%. This result underscores the effectiveness of the full KDRL framework in enhancing the student model’s performance through synergistic contributions from both KD and RL.
4.3.3. Zero-Shot Classification Benchmark
Table 3 presents the zero-shot classification results on public datasets, with models trained on the SkyScript dataset. The benchmarks include CLIP-based methods, KD approaches, and the proposed KDRL method, which is also a CLIP-based approach built on RemoteCLIP. Performance is evaluated across datasets such as AID, EuroSAT, fMoW, Million-AID, PatternNet, RESISC, and RSI-CB.
Table 3.
Benchmark evaluation on zero-shot classification on public datasets. The methods are trained on the SkyScript dataset.
Among the CLIP-based models, SkyCLIP-50 achieves high performance across most datasets, with notable results such as 80.88% on PatternNet and 67.45% on Million-AID. However, its performance is slightly lower on RESISC (70.94%) compared to RemoteCLIP (74.31%), showing variability depending on the dataset. RemoteCLIP, while strong on RSI-CB, struggles on more challenging datasets like fMoW (16.77%) and EuroSAT (27.81%).
KD-based methods consistently outperform standard CLIP-based approaches by leveraging efficient knowledge transfer. Among these, KD-TAKD achieves the best results, with scores of 48.01% on EuroSAT, 67.62% on Million-AID, and 78.41% on PatternNet. GDKD is a close second, excelling on PatternNet (78.43%) and Million-AID (67.77%). These methods demonstrate the power of KD in improving the generalization and efficiency of smaller models.
As a CLIP-based method derived from RemoteCLIP, KDRL combines KD with RL to optimize knowledge transfer and exploration. KDRL outperforms other methods on most datasets, achieving 69.51% on SkyScript, 69.24% on AID, 70.30% on RESISC, and 80.08% on PatternNet. On challenging datasets like fMoW, KDRL scores 20.31%, a significant improvement over RemoteCLIP’s 16.77%. These results highlight KDRL’s ability to overcome the limitations of RemoteCLIP and deliver robust performance across diverse datasets.
The results show that integrating reinforcement learning into knowledge distillation can significantly enhance zero-shot classification performance. KDRL not only addresses the weaknesses of RemoteCLIP but also surpasses both CLIP-based and KD-based methods across most datasets. Its dynamic optimization approach makes it highly effective in handling the spatial and temporal complexities of remote sensing data.
Compared with KD-TAKD, KDRL is slightly weaker on EuroSAT and fMoW. KD-TAKD introduces an intermediate teacher-assistant, which can reduce the capacity gap and preserve fine-grained teacher information in a more gradual manner. This can be advantageous when (i) the target dataset is relatively compact with high signal-to-noise (EuroSAT) or (ii) scenes exhibit strong context and temporal bias (fMoW). In contrast, our Stage 2 refinement emphasizes global representation geometry (cluster reward) and confidence shaping, which may slightly over-regularize the student toward teacher embedding clusters learned on the pre-training distribution. This trade-off favors generalization on most datasets but can be suboptimal on EuroSAT and fMoW, where dataset-specific cues are more prominent. We view this behavior as a limitation and discuss directions to mitigate it (e.g., task-adaptive cluster construction or reward weighting) in Section 5.
4.4. Image–Text Retrieval on Remote Sensing Dataset
This subsection evaluates whether the KDRL representation refinement transfers beyond classification to cross-modal image–text retrieval. Because retrieval depends on global embedding alignment between modalities, it provides a complementary view of how the student embedding geometry changes after Stage 2.
4.4.1. Public Datasets for Image–Text Retrieval
We compare our KDRL against other baseline retrieval models, including VSE++ [42], SCAN [43], MTFN [44], AMFMN [45], LW-MRC [46], GaLR [47], CMFM-Net [48], HyperMatch [49], HVSA [50], FBCLM [51], DOVE [52], and PIR [53]. The evaluation is conducted by performing image-to-text and text-to-image retrieval tasks on two public datasets, RSITMD [54] and RSICD [55].
4.4.2. Retrieval Benchmark
We report Recall@K (higher is better) for both image-to-text and text-to-image retrieval. Beyond absolute numbers, we analyze which gains are consistent across datasets and which failures suggest limitations of global reward shaping.
We conducted a comprehensive evaluation of our proposed KDRL method against state-of-the-art models on two public datasets, RSITMD and RSICD, for image-to-text and text-to-image retrieval tasks. Table 4 presents the comparative results in terms of Recall@1, Recall@5, and Recall@10.
Table 4.
Benchmark evaluation of cross-modal image and text retrieval on RSITMD and RSICD datasets.
For RSITMD, KDRL demonstrates substantial performance improvements, achieving 18.01% and 15.60% Recall@1 for image-to-text and text-to-image tasks, respectively, significantly outperforming most baseline methods. Notably, KDRL surpasses recent methods such as PIR and DOVE, particularly in Recall@10, with 67.44% and 74.76%.
On RSICD, KDRL continues to lead, with Recall@1 scores of 11.22% for text-to-image tasks. While FBCLM and PIR perform strongly in some metrics, KDRL establishes a robust advantage in Recall@10, reaching 52.61% and 57.04%, which are among the highest across all tested methods.
These results affirm the effectiveness of KDRL in bridging the gap between image and text modalities, achieving new state-of-the-art performance in key retrieval benchmarks.
4.5. Visual Grounding on Remote Sensing Dataset
This subsection evaluates KDRL on DIOR-RSVG visual grounding, which requires aligning a text query with a region (box) rather than only global embeddings. This task is therefore a stress test for whether Stage 2 rewards (defined on global embedding clusters and confidence) translate to spatially precise localization.
In this section we evaluate KDRL on a visual grounding task. We use Faster R-CNN as our RPN and replace the classification head with our CLIP model to perform evaluation on DIOR-RSVG dataset.
From Table 5, KDRL is less competitive under permissive thresholds (Pr@0.5–0.6) but becomes competitive under stricter thresholds (Pr@0.7–0.9), especially compared to other ViT-B/32 baselines.
Table 5.
Performance of our KDRL on visual grounding.
Evaluation note: following the DIOR-RSVG protocol, Pr@ reports the precision after filtering region–text matches by a similarity/confidence threshold . Lower is more permissive (more predicted matches are kept), while higher focuses on the most confident alignments. At lower precision thresholds, KDRL demonstrates a notable performance gap, with RemoteCLIP achieving a 5.55% higher Pr@0.5 and SkyCLIP outperforming by 2.66%. Although the gap narrows at higher thresholds, such as Pr@0.9, where KDRL achieves comparable performance to SkyCLIP (64.21% vs. 62.11%), it still trails RemoteCLIP (63.38%). This trend suggests that while KDRL may perform adequately in stricter matching conditions, its ability to generalize under less restrictive thresholds is limited. This behavior is closely tied to our training objective and the DIOR-RSVG evaluation protocol. Stage 2 rewards are computed on global image–text embeddings and confidence, which tends to sharpen the similarity distribution for the top-ranked matches but does not explicitly optimize region-level representations or spatial alignment. Under permissive thresholds ( or ), many medium-confidence region proposals are retained; without region-aware distillation, these proposals include more false positives, reducing precision. As increases, evaluation focuses on the most confident matches, and the improved global semantics from cluster-based refinement helps KDRL approach (or exceed) other ViT-B/32 baselines. The remaining gap to RemoteCLIP (L-14) is expected given the student’s smaller capacity and the absence of localization-aware rewards; improving grounding likely requires incorporating region-level cluster rewards or cross-attention/box supervision in Stage 2 rather than relying solely on global feature refinement.
The RL refinement stage explicitly optimizes (i) alignment to teacher cluster structure and (ii) prediction confidence. This tends to sharpen the score distribution: a small set of highly confident matches become more accurate, which improves Pr@0.7–0.9, while a larger set of moderately scored matches may remain poorly calibrated, which hurts Pr@0.5–0.6. In contrast, the larger RemoteCLIP/SkyCLIP teachers have higher capacity for region–text alignment under permissive filtering, leading to higher precision when many ambiguous proposals are retained. This analysis motivates future grounding-specific rewards (e.g., region localization cues) to better balance calibration across thresholds. While KDRL achieves competitive results at higher precision levels, further optimization is necessary to close the performance gap with other state-of-the-art CLIP-based models.
5. Discussion and Conclusions
In addition to the experiments, we also visualize the feature clustering results to further analyze the performance of the models. Figure 2 presents a comparison of feature clustering between three approaches: RemoteCLIP (teacher model), KD, and KDRL. This comparison provides insights into the effectiveness of the KDRL framework.
Figure 2.
The feature clustering comparison between three different approaches (a) RemoteCLIP (teacher model), (b) knowledge distillation, and (c) KDRL illustrates the effectiveness of the KDRL method in refining the student model’s feature representation.
In Figure 2a, the teacher model (RemoteCLIP) shows well-separated and compact clusters, indicating that the model produces highly discriminative feature representations for each class. These clusters represent the target feature space that the student model aims to replicate. On the other hand, Figure 2b, which shows the clustering results of the student model after applying standard KD, reveals that while the clusters remain distinct, there is noticeable misalignment and reduced compactness in the feature embeddings. This suggests that although KD aids in feature transfer, the alignment between the student and teacher features is less precise, and the compactness is not fully achieved.
In Figure 2c, we show the clustering results of our proposed KDRL approach. Compared to both the teacher model and the KD student, the clusters are more compact and better separated. This improvement in clustering quality indicates the contribution of reinforcement learning during the refinement phase. The additional rewards guide the student model to align its features more closely with the teacher’s cluster centers. The enhanced clustering results demonstrate that KDRL not only facilitates knowledge transfer through KD but also fine-tunes the feature space via RL technique, leading to better generalization across tasks.
5.1. Strengths and Weaknesses Across Tasks
Across tasks, KDRL is strongest when the downstream objective is dominated by global semantic alignment—zero-shot classification and image–text retrieval. In these settings, Stage 1 distillation transfers the teacher’s broad vision–language knowledge, while Stage 2 improves representation compactness (Figure 2) and reduces domain-specific feature drift, yielding consistent gains over the ViT-B/32 baselines (Table 3 and Table 4).
In particular, the proposed framework is most beneficial when strong global semantic transfer is required, while grounding-oriented extensions are needed for region-level tasks.
In contrast, the visual grounding task exposes a weakness: grounding requires fine-grained region–language correspondence and good score calibration over many ambiguous proposals. As shown in Table 5, KDRL improves precision in the high-confidence regime but remains weaker under permissive thresholds, suggesting that the current rewards are better aligned with global semantics than with dense localization.
5.2. Failure Modes
We qualitatively observe three typical failure patterns for the student model: (i) small or cluttered objects where the ViT-B/32 image encoder (larger patch size) lacks fine spatial detail; (ii) attribute confusion (e.g., visually similar land-use categories) where prompts contain subtle distinctions; and (iii) long-tail phrasing in text queries where limited domain-specific grounding supervision reduces robustness. These failure modes are most visible in grounding and, to a lesser extent, in retrieval under permissive filtering.
5.3. Why RL Helps Clustering but Can Hurt Grounding
The cluster reward explicitly encourages the student embedding to move toward teacher cluster centers, which directly improves intra-class compactness and inter-class separation—precisely what is visualized in Figure 2. However, grounding depends on ranking regions rather than whole-image embeddings. Because our rewards are computed on embedding alignment and confidence, they can over-emphasize global separability and “peakiness” of scores, which helps high-threshold precision but may reduce calibration when many candidate regions are retained (Table 5). This explains the apparent tension: RL improves the structure of the embedding space, yet does not directly optimize region-level localization.
5.4. Ablation Interpretation and Stability
Table 2 indicates that the cluster-based reward contributes more than the confidence reward. This is expected because the cluster term provides a structured, teacher-informed target for intermediate representations, while confidence alone can be satisfied by over-confident predictions without improving alignment. The dynamic schedule further mitigates instability by gradually shifting optimization from imitation (KD) to refinement (reward-shaped loss), reducing gradient conflict between task loss and reward terms. In practice, our “RL” stage is deterministic minibatch optimization rather than interactive rollouts, which helps control variance (see the reproducibility note in Section 4.2).
5.5. Computational Overhead and Deployment Implications
From a deployment perspective, KDRL is designed for edge inference: the final model keeps the compact ViT-B/32 student backbone, and text embeddings can be cached for fixed prompt sets. The additional Stage 2 refinement increases training time but does not add the heavy runtime cost of the teacher. This is particularly relevant for remote sensing platforms such as UAVs, satellites, and field-deployed systems where compute and memory budgets are limited, yet robust generalization is needed under domain shift (season, sensor, and geography).
5.6. Limitations and Future Work
KDRL has several limitations. First, although Stage 2 uses a deterministic and differentiable reward objective, a full multi-seed evaluation with standard deviations would strengthen the statistical claims since RL-style training objectives can still be sensitive to initialization and data shuffling. Second, our rewards are primarily designed for global semantic alignment; consequently, region-level localization is not explicitly optimized, which contributes to weaker visual grounding at permissive thresholds. Third, KDRL introduces additional hyperparameters (, , , and the number of refinement epochs). While we observed stable defaults across benchmarks, these parameters can affect performance (especially grounding) and require validation tuning. Fourth, the two-stage pipeline increases training complexity compared with a single-stage fine-tuning or KD baseline, even though its additional wall-clock cost is modest (Section 4.2).
Future work will explore (i) automating reward design and weighting (e.g., via meta-learning or adaptive schedules); (ii) incorporating localization-aware and calibration-aware rewards (e.g., region-level cluster rewards, cross-attention supervision, and temperature scaling); (iii) extending KDRL to online/continual learning settings where clusters and rewards update as data streams; and (iv) investigating alternative RL algorithms (e.g., actor–critic or off-policy methods) for the refinement stage.
This work demonstrates the potential of combining knowledge distillation and reinforcement learning to enhance model performance, particularly in domains with complex data distributions. The proposed KDRL framework achieves superior feature alignment, improved generalization, and task performance, setting a new standard for lightweight model training in image processing tasks.
Author Contributions
Conceptualization, T.K.; Methodology, T.K.; Software, T.K.; Validation, T.K.; Formal analysis, T.K.; Investigation, T.K.; Resources, Y.X.; Data curation, Y.X.; Writing—original draft, T.K., Y.X. and P.S.; Writing—Review & editing, Y.X. and P.S.; Supervision, T.K.; Project administration, P.S.; Funding acquisition, T.K. All authors have read and agreed to the published version of the manuscript.
Funding
This work was supported in part by the National Natural Science Foundation of China under Grant 62202435.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
The experimental data used in this study are derived entirely from publicly available remote sensing image–text datasets. Specifically, we use the RSICD dataset, available at 1 December 2025 https://github.com/201528014227051/RSICD_optimal, and the RSITMD dataset, available at 1 December 2025 https://github.com/AICyberTeam/AMFMN/tree/main/RSITMD.
Conflicts of Interest
The authors declare no conflicts of interest.
References
- Hinton, G. Distilling the Knowledge in a Neural Network. arXiv 2015, arXiv:1503.02531. [Google Scholar] [CrossRef]
- Gou, J.; Yu, B.; Maybank, S.J.; Tao, D. Knowledge distillation: A survey. Int. J. Comput. Vis. 2021, 129, 1789–1819. [Google Scholar] [CrossRef]
- Wang, H.; Wang, Q.; Liu, Y.; Du, W.; Ye, H. Knowledge distillation in deep learning and its applications. Nat. Mach. Intell. 2021, 3, 774–786. [Google Scholar]
- Zhu, X.X.; Tuia, D.; Mou, L.; Xia, G.S.; Zhang, L.; Xu, F.; Fraundorfer, F. Deep learning in remote sensing: A comprehensive review and list of resources. IEEE Geosci. Remote Sens. Mag. 2017, 5, 8–36. [Google Scholar] [CrossRef]
- Li, W.; Wan, L.; Wang, Q.; Lu, X. A comprehensive review of deep learning in remote sensing: Data, methods, and applications. Int. J. Appl. Earth Obs. Geoinf. 2022, 113, 102783. [Google Scholar]
- Yuan, Q.; Lin, L.; Zhang, H. Deep learning for urban remote sensing: A review. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 1925–1946. [Google Scholar]
- Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M.; van der Laak, J.A.; van Ginneken, B.; Sánchez, C.I. A survey on deep learning in medical image analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef]
- Ker, J.; Wang, L.; Rao, J.; Lim, T. Deep learning applications in medical image analysis. Comput. Biol. Med. 2018, 98, 205–210. [Google Scholar] [CrossRef]
- Guan, H.; Yap, P.-T.; Bozoki, A.; Liu, M. Federated learning for medical image analysis: A survey. Pattern Recognit. 2024, 151, 110424. [Google Scholar] [CrossRef]
- Sutton, R.S.; Barto, A.G. Reinforcement Learning: An Introduction; MIT Press: Cambridge, MA, USA, 1998. [Google Scholar]
- Mnih, V.; Kavukcuoglu, K.; Silver, D.; Rusu, A.A.; Veness, J.; Bellemare, M.G.; Graves, A.; Riedmiller, M.; Fidjeland, A.K.; Ostrovski, G.; et al. Human-level control through deep reinforcement learning. Nature 2015, 518, 529–533. [Google Scholar] [CrossRef]
- Silver, D.; Huang, A.; Maddison, C.J.; Guez, A.; Sifre, L.; van den Driessche, G.; Schrittwieser, J.; Antonoglou, I.; Panneershelvam, V.; Lanctot, M.; et al. Mastering the game of Go with deep neural networks and tree search. Nature 2016, 529, 484–489. [Google Scholar] [CrossRef]
- Trung, L.; Zhang, X.; Jie, Z.; Sun, P.; Jin, X.; Li, H. Reft: Reasoning with reinforced fine-tuning. In Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Bangkok, Thailand, 11–16 August 2024; pp. 7601–7614. [Google Scholar]
- Chawla, A.; Yin, H.; Molchanov, P.; Alvarez, J. Data-free knowledge distillation for object detection. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Virtual, 5–9 January 2021; pp. 3289–3298. [Google Scholar]
- De Rijk, P.; Schneider, L.; Cordts, M.; Gavrila, D. Structural knowledge distillation for object detection. Adv. Neural Inf. Process. Syst. 2022, 35, 3858–3870. [Google Scholar]
- He, T.; Shen, C.; Tian, Z.; Gong, D.; Sun, C.; Yan, Y. Knowledge adaptation for efficient semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 578–587. [Google Scholar]
- Li, M.; Halstead, M.; Mccool, C. Knowledge Distillation for Efficient Instance Semantic Segmentation with Transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 5432–5439. [Google Scholar]
- Zhou, W.; Li, Y.; Huang, J.; Liu, Y.; Jiang, Q. MSTNet-KD: Multilevel transfer networks using knowledge distillation for the dense prediction of remote-sensing images. IEEE Trans. Geosci. Remote Sens. 2024, 62, 4504612. [Google Scholar] [CrossRef]
- Qin, D.; Bu, J.J.; Liu, Z.; Shen, X.; Zhou, S.; Gu, J.J.; Wang, Z.H.; Wu, L.; Dai, H.F. Efficient medical image segmentation based on knowledge distillation. IEEE Trans. Med. Imaging 2021, 40, 3820–3831. [Google Scholar] [CrossRef] [PubMed]
- Gou, J.; Hu, Y.; Sun, L.; Wang, Z.; Ma, H. Collaborative knowledge distillation via filter knowledge transfer. Expert Syst. Appl. 2024, 238, 121884. [Google Scholar] [CrossRef]
- Wu, J.; Fang, L.; Yue, J. TAKD: Target-Aware Knowledge Distillation for Remote Sensing Scene Classification. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 8188–8200. [Google Scholar] [CrossRef]
- Jaiswal, A.; Ashutosh, K.; Rousseau, J.F.; Peng, Y.; Wang, Z.; Ding, Y. RoS-KD: A robust stochastic knowledge distillation approach for noisy medical imaging. In Proceedings of the 2022 IEEE International Conference on Data Mining (ICDM), Orlando, FL, USA, 28 November–1 December 2022; pp. 981–986. [Google Scholar]
- Agostinelli, F.; McAleer, S.; Shmakov, A.; Baldi, P. Solving the Rubik’s cube with deep reinforcement learning and search. Nat. Mach. Intell. 2019, 1, 356–363. [Google Scholar] [CrossRef]
- Mahmoudieh, P.; Pathak, D.; Darrell, T. Zero-shot reward specification via grounded natural language. In Proceedings of the International Conference on Machine Learning, PMLR, Baltimore, Ml, USA, 17–23 July 2022; pp. 14743–14752. [Google Scholar]
- Kalashnikov, D.; Irpan, A.; Pastor, P.; Ibarz, J.; Herzog, A.; Jang, E.; Quillen, D.; Holly, E.; Kalakrishnan, M.; Vanhoucke, V.; et al. Scalable deep reinforcement learning for vision-based robotic manipulation. In Proceedings of the Conference on Robot Learning, PMLR, Zürich, Switzerland, 29–31 October 2018; pp. 651–673. [Google Scholar]
- Pinto, A.S.; Kolesnikov, A.; Shi, Y.; Beyer, L.; Zhai, X. Tuning computer vision models with task rewards. In Proceedings of the International Conference on Machine Learning, PMLR, Honolulu, HI, USA, 23–29 July 2023; pp. 33229–33239. [Google Scholar]
- Lee, Y.; Yang, D.; Wu, W. Stimulate Knowledge Distillation Ability via Reinforcement Learning. SSRN 4775210. Available online: https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4775210 (accessed on 1 February 2025).
- Zhang, Z.; Zhou, Y.; Gong, J.; Liu, J.; Tu, Z. Instance temperature knowledge distillation. arXiv 2024, arXiv:2407.00115. [Google Scholar]
- Xia, G.S.; Hu, J.; Hu, F.; Shi, B.; Bai, X.; Zhong, Y.; Zhang, L.; Lu, X. AID: A Benchmark Data Set for Performance Evaluation of Aerial Scene Classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 3965–3981. [Google Scholar] [CrossRef]
- Helber, P.; Bischke, B.; Dengel, A.; Borth, D. EuroSAT: A Novel Dataset and Deep Learning Benchmark for Land Use and Land Cover Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 2217–2226. [Google Scholar] [CrossRef]
- Christie, G.; Fendley, N.; Wilson, J.; Mukherjee, R. Functional Map of the World. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6172–6180. [Google Scholar] [CrossRef]
- Long, Y.; Xia, G.S.; Li, S.; Yang, W.; Yang, M.Y.; Zhu, X.X.; Zhang, L.; Li, D. On Creating Benchmark Dataset for Aerial Image Interpretation: Reviews, Guidances, and Million-AID. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 4205–4230. [Google Scholar] [CrossRef]
- Zhou, W.; Newsam, S.; Li, C.; Shao, Z. PatternNet: A benchmark dataset for performance evaluation of remote sensing image retrieval. ISPRS J. Photogramm. Remote Sens. 2018, 145, 197–209. [Google Scholar] [CrossRef]
- Cheng, G.; Han, J.; Lu, X. Remote Sensing Image Scene Classification: Benchmark and State of the Art. Proc. IEEE 2017, 105, 1865–1883. [Google Scholar] [CrossRef]
- Li, H.; Dou, X.; Tao, C.; Wu, Z.; Chen, J.; Peng, J.; Deng, M.; Zhao, L. RSI-CB: A Large-Scale Remote Sensing Image Classification Benchmark Using Crowdsourced Data. Sensors 2020, 20, 1594. [Google Scholar] [CrossRef]
- Wang, Z.; Prabha, R.; Huang, T.; Wu, J.; Rajagopal, R. SkyScript: A Large and Semantically Diverse Vision-Language Dataset for Remote Sensing. Proc. AAAI Conf. Artif. Intell. 2024, 38, 5805–5813. [Google Scholar] [CrossRef]
- Radford, A.; Kim, J.W.; Hallacy, C.; Ramesh, A.; Goh, G.; Agarwal, S.; Sastry, G.; Askell, A.; Mishkin, P.; Clark, J.; et al. Learning Transferable Visual Models From Natural Language Supervision. arXiv 2021, arXiv:2103.00020. [Google Scholar] [CrossRef]
- Schuhmann, C.; Beaumont, R.; Vencu, R.; Gordon, C.; Wightman, R.; Cherti, M.; Coombes, T.; Katta, A.; Mullis, C.; Wortsman, M.; et al. LAION-5B: An open large-scale dataset for training next generation image-text models. arXiv 2022, arXiv:2210.08402. [Google Scholar]
- Romero, A.; Ballas, N.; Kahou, S.E.; Chassang, A.; Gatta, C.; Bengio, Y. FitNets: Hints for Thin Deep Nets. arXiv 2015, arXiv:1412.6550. [Google Scholar] [CrossRef]
- Mirzadeh, S.I.; Farajtabar, M.; Li, A.; Levine, N.; Matsukawa, A.; Ghasemzadeh, H. Improved Knowledge Distillation via Teacher Assistant. Proc. AAAI Conf. Artif. Intell. 2020, 34, 5191–5198. [Google Scholar] [CrossRef]
- Son, W.; Na, J.; Choi, J.; Hwang, W. Densely Guided Knowledge Distillation Using Multiple Teacher Assistants. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 9395–9404. [Google Scholar]
- Faghri, F.; Fleet, D.J.; Kiros, J.R.; Fidler, S. VSE++: Improving Visual-Semantic Embeddings with Hard Negatives. arXiv 2018, arXiv:1707.05612. [Google Scholar]
- Lee, K.H.; Chen, X.; Hua, G.; Hu, H.; He, X. Stacked Cross Attention for Image-Text Matching. In Proceedings of the Computer Vision—ECCV 2018, Munich, Germany, 8–14 September 2018; Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y., Eds.; Springer: Cham, Switzerland, 2018; pp. 212–228. [Google Scholar]
- Wang, T.; Xu, X.; Yang, Y.; Hanjalic, A.; Shen, H.T.; Song, J. Matching Images and Text with Multi-modal Tensor Fusion and Re-ranking. In Proceedings of the 27th ACM International Conference on Multimedia, MM ’19, Nice, France, 21–25 October 2019; pp. 12–20. [Google Scholar] [CrossRef]
- Hoxha, G.; Melgani, F.; Demir, B. Toward Remote Sensing Image Retrieval Under a Deep Image Captioning Perspective. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 4462–4475. [Google Scholar] [CrossRef]
- Rahhal, M.M.A.; Bazi, Y.; Abdullah, T.; Mekhalfi, M.L.; Zuair, M. Deep Unsupervised Embedding for Remote Sensing Image Retrieval Using Textual Cues. Appl. Sci. 2020, 10, 8931. [Google Scholar] [CrossRef]
- Yuan, Z.; Zhang, W.; Tian, C.; Rong, X.; Zhang, Z.; Wang, H.; Fu, K.; Sun, X. Remote Sensing Cross-Modal Text-Image Retrieval Based on Global and Local Information. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5620616. [Google Scholar] [CrossRef]
- Yu, H.; Yao, F.; Lu, W.; Liu, N.; Li, P.; You, H.; Sun, X. Text-Image Matching for Cross-Modal Remote Sensing Image Retrieval via Graph Neural Network. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 812–824. [Google Scholar] [CrossRef]
- Yao, F.; Sun, X.; Liu, N.; Tian, C.; Xu, L.; Hu, L.; Ding, C. Hypergraph-Enhanced Textual-Visual Matching Network for Cross-Modal Remote Sensing Image Retrieval via Dynamic Hypergraph Learning. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 688–701. [Google Scholar] [CrossRef]
- Zhang, W.; Li, J.; Li, S.; Chen, J.; Zhang, W.; Gao, X.; Sun, X. Hypersphere-Based Remote Sensing Cross-Modal Text–Image Retrieval via Curriculum Learning. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5621815. [Google Scholar] [CrossRef]
- Li, H.; Xiong, W.; Cui, Y.; Xiong, Z. A fusion-based contrastive learning model for cross-modal remote sensing retrieval. Int. J. Remote Sens. 2022, 43, 3359–3386. [Google Scholar] [CrossRef]
- Ma, Q.; Pan, J.; Bai, C. Direction-Oriented Visual–Semantic Embedding Model for Remote Sensing Image–Text Retrieval. IEEE Trans. Geosci. Remote Sens. 2024, 62, 4704014. [Google Scholar] [CrossRef]
- Pan, J.; Ma, Q.; Bai, C. A Prior Instruction Representation Framework for Remote Sensing Image-text Retrieval. In Proceedings of the 31st ACM International Conference on Multimedia, MM ’23, Ottawa, ON, Canada, 29 October–3 November 2023; pp. 611–620. [Google Scholar] [CrossRef]
- Yuan, Z.; Zhang, W.; Fu, K.; Li, X.; Deng, C.; Wang, H.; Sun, X. Exploring a Fine-Grained Multiscale Method for Cross-Modal Remote Sensing Image Retrieval. IEEE Trans. Geosci. Remote Sens. 2022, 60, 4404119. [Google Scholar] [CrossRef]
- Lu, X.; Wang, B.; Zheng, X.; Li, X. Exploring models and data for remote sensing image caption generation. IEEE Trans. Geosci. Remote Sens. 2017, 56, 2183–2195. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license.