You are currently viewing a new version of our website. To view the old version click .
Tomography
  • Article
  • Open Access

20 October 2025

ECSA: Mitigating Catastrophic Forgetting and Few-Shot Generalization in Medical Visual Question Answering

,
,
,
and
School of Computer Science and Technology, Xinjiang University, Urumqi 830017, China
*
Author to whom correspondence should be addressed.

Abstract

Objective: Medical Visual Question Answering (Med-VQA), a key technology that integrates computer vision and natural language processing to assist in clinical diagnosis, possesses significant potential for enhancing diagnostic efficiency and accuracy. However, its development is constrained by two major bottlenecks: weak few-shot generalization capability stemming from the scarcity of high-quality annotated data and the problem of catastrophic forgetting when continually learning new knowledge. Existing research has largely addressed these two challenges in isolation, lacking a unified framework. Methods: To bridge this gap, this paper proposes a novel Evolvable Clinical-Semantic Alignment (ECSA) framework, designed to synergistically solve these two challenges within a single architecture. ECSA is built upon powerful pre-trained vision (BiomedCLIP) and language (Flan-T5) models, with two innovative modules at its core. First, we design a Clinical-Semantic Disambiguation Module (CSDM), which employs a novel debiased hard negative mining strategy for contrastive learning. This enables the precise discrimination of “hard negatives” that are visually similar but clinically distinct, thereby significantly enhancing the model’s representation ability in few-shot and long-tail scenarios. Second, we introduce a Prompt-based Knowledge Consolidation Module (PKC), which acts as a rehearsal-free non-parametric knowledge store. It consolidates historical knowledge by dynamically accumulating and retrieving task-specific “soft prompts,” thus effectively circumventing catastrophic forgetting without relying on past data. Results: Extensive experimental results on four public benchmark datasets, VQA-RAD, SLAKE, PathVQA, and VQA-Med-2019, demonstrate ECSA’s state-of-the-art or highly competitive performance. Specifically, ECSA achieves excellent overall accuracies of 80.15% on VQA-RAD and 85.10% on SLAKE, while also showing strong generalization with 64.57% on PathVQA and 82.23% on VQA-Med-2019. More critically, in continual learning scenarios, the framework achieves a low forgetting rate of just 13.50%, showcasing its significant advantages in knowledge retention. Conclusions: These findings validate the framework’s substantial potential for building robust and evolvable clinical decision support systems.

1. Introduction

Medical Visual Question Answering is a key multimodal learning task that integrates advanced computer vision and natural language processing techniques, aiming to build intelligent systems capable of understanding details in medical images and generating precise answers to clinical questions, thereby playing an important auxiliary role in clinical diagnosis and decision-making [1,2]. Unlike general-domain Visual Question Answering (VQA), Med-VQA faces more severe challenges [3]. Its models not only require excellent visual feature extraction capabilities to identify subtle pathological markers but also must possess strong language understanding abilities to parse specialized medical terminology and achieve deep reasoning and fusion of information from both modalities in high-risk clinical environments [1,4]. As highlighted in the review by Dong et al., the ultimate goal of Med-VQA systems is to alleviate the burden on medical professionals and reduce the diagnostic error rates, and its inherent complexity requires the development of specialized architectures that surpass general-purpose models [1].
However, the development and deployment of Med-VQA systems are dually constrained by the inherent characteristics of medical data, giving rise to two core challenges: few-shot generalization and continual learning [5]. First, the cost of annotating high-quality medical images is extremely high, requiring significant time and effort from clinical experts, which leads to a scarcity of large-scale high-quality labeled datasets [6]. Chen et al. have pointed out that acquiring a small number of annotations is feasible in practical clinical settings, making this problem naturally suited to the paradigm of few-shot learning [7,8]. Meanwhile, the natural distribution of diseases exhibits a typical long-tail characteristic, where imaging data for rare diseases are inherently scarce [9,10]. These dual factors make it difficult for models to learn robust feature representations under few-shot conditions, resulting in poor generalization to unseen or rare cases [11]. Second, the clinical environment is dynamically evolving, with new diseases, diagnostic criteria, and imaging technologies constantly emerging. This requires AI models to possess the capability for continual or lifelong learning—that is, to continuously acquire knowledge from new tasks without accessing past data [12,13]. However, as demonstrated by Verma et al. in their research on medical image classification tasks, traditional models encounter catastrophic forgetting during the sequential learning process—where memory of previously learned knowledge catastrophically degrades while learning new information [12,14]. Therefore, any Med-VQA framework intended for clinical application must simultaneously overcome the dual challenges of weak few-shot generalization and catastrophic forgetting in continual learning.
To address the aforementioned challenges, existing methods have explored various avenues, but they still exhibit significant limitations. Early Med-VQA research primarily employed discriminative models, which constrain answers to a predefined closed set; this approach cannot handle open-ended questions and struggles to adapt to the introduction of new knowledge [1,15]. In recent years, the research focus has shifted towards generative models. For example, the Multimodal Prompt Retrieval (MPR) model proposed by Ossowski and Hu enhances the model’s domain adaptation capability in few-shot scenarios by retrieving similar image–text pairs to construct prompts [16]. However, such methods rely on a static pre-constructed retrieval library and lack a mechanism for dynamically updating the knowledge base during continual learning. Concurrently, the rise of large-scale pre-trained foundation models has introduced a new paradigm to Med-VQA. On one hand, Vision-Language Models (VLMs) like MaCo have acquired powerful zero-shot visual understanding capabilities through contrastive learning on massive radiology image–text pairs [17]. On the other hand, Large Medical Language Models (LLMs) such as GatorTron have mastered a rich body of medical knowledge through pre-training on vast clinical texts [2]. Although these models perform exceptionally well in their respective modalities, when directly applied to continual learning scenarios, the method of updating model parameters via fine-tuning is highly prone to disrupting the general knowledge embedded in the pre-trained weights, thereby triggering catastrophic forgetting [18]. In summary, current approaches either cannot generate free-text answers, lack a mechanism for dynamic knowledge updates, or perform poorly in continual learning, failing to provide a unified solution that simultaneously addresses both few-shot generalization and catastrophic forgetting.
To bridge this research gap, this paper proposes a novel Evolvable Clinical-Semantic Alignment (ECSA) framework, which skillfully integrates the core ideas of few-shot learning and continual learning, aiming to achieve efficient knowledge transfer, retention, and accumulation. The main contributions of this paper are as follows:
1.
We introduce the novel Evolvable Clinical-Semantic Alignment (ECSA) framework, designed to synergistically address the dual challenges of few-shot generalization and rehearsal-free continual learning within a unified architecture for the Med-VQA task. ECSA is built upon powerful pre-trained vision (BiomedCLIP) and language (Flan-T5) foundation models and achieves continual knowledge accumulation through a parameter-efficient prompt-based learning strategy.
2.
We design the Clinical-Semantic Disambiguation Module (CSDM), which achieves deep alignment of multi-scale visual features and text embeddings via a cross-attention mechanism. This hierarchical fusion strategy not only enables the model to simultaneously capture both the global context and local lesion details of an image, but more critically, it enhances the model’s ability to distinguish hard negative samples—that is, distractors that are visually highly similar to the target but belong to a different class. By precisely filtering these confusing negative samples, CSDM significantly improves the model’s representation and recognition capabilities on few-shot and long-tail rare cases.
3.
We introduce the Prompt-based Knowledge Consolidation Module (PKC), which is a dynamically expandable memory store for storing task-specific “soft prompts” acquired through learning. When learning a new task, this module only adds and trains new prompts relevant to the new task while freezing the old ones, thereby effectively preserving historical knowledge without storing any past patient data and fundamentally mitigating the problem of catastrophic forgetting.
We conducted extensive experiments on four widely-used medical VQA benchmark datasets: VQA-RAD [19], SLAKE [20], PathVQA [21], and VQA-Med-2019 [22]. The results show that our proposed ECSA framework achieves state-of-the-art or competitive performance across all datasets. This fully validates its substantial potential for building robust and evolvable clinical decision support systems.

3. Methods

This section details our innovative Evolvable Clinical-Semantic Alignment (ECSA) framework, designed to address the dual challenges of precise clinical discrimination and verifiable knowledge generation in Med-VQA. The framework’s architecture is organized into two core stages, each directly reflecting a key aspect of its name. The first stage, Clinical-Semantic Alignment, establishes the model’s foundational understanding by processing input via Multimodal Prompt Encoding and then leveraging our Clinical-Semantic Disambiguation Module (CSDM) with its novel debiased hard negative mining strategy to construct a highly discriminative embedding space. Building on this, the second stage, Evolvable Knowledge Consolidation, endows the framework with its crucial adaptive capability through the Prompt-based Knowledge Consolidation Module (PKC), which functions as a Retrieval-Augmented Generation mechanism to facilitate continual learning and ensure the knowledge base remains current without catastrophic forgetting. Following this two-stage exposition, we will describe the overall training and inference process. The complete framework is illustrated in Figure 1.
Figure 1. Overall schematic of the Evolvable Clinical-Semantic Alignment (ECSA) framework. The framework consists of two core stages. Stage 1: Clinical-Semantic Alignment, responsible for constructing high-quality cross-modal representations. The Multimodal Prompt Encoding (MPE) module first encodes the input, after which the Clinical-Semantic Disambiguation Module (CSDM), through a Feature Fusion Cross-Modal Attention (FF-CMA) mechanism and a Debiased Hard Negative Mining (DHNM) strategy, produces robustly aligned features for cases that are visually similar but clinically distinct. Stage 2: Evolvable Knowledge Consolidation, responsible for the model’s continual learning and knowledge generation. The Prompt-based Knowledge Consolidation (PKC) module utilizes a dynamic non-parametric memory store and employs a Retrieval-Augmented Generation mechanism during inference to provide traceable knowledge support for the final answer generation.

3.1. Multimodal Prompt Encoding (MPE)

To provide an information-rich and structurally uniform input for the subsequent alignment, retrieval, and generation tasks, this framework jointly encodes a complete VQA triplet (image V, question Q, answer A) as a holistic “multimodal prompt”. This approach ensures that the model, when processing the input, can fully leverage the visual information from the image as well as the complete semantic context embedded in the question–answer pair, thereby laying the foundation for deep semantic understanding and alignment.

3.1.1. Vision Feature Extraction

In medical image analysis, a model’s performance is highly dependent on its ability to extract features with clinical diagnostic value from images. To address this issue, we select the Vision Transformer (ViT) from BiomedCLIP, pre-trained on the large-scale biomedical image–text pair dataset PMC-15M, as the image encoder [51]. This model has demonstrated excellent performance in multiple biomedical vision-language tasks, and its domain-specific pre-training endows it with a stronger capability for medical concept representation.
To preserve local pathological details that are crucial for diagnosis, we do not use a single global feature vector. Instead, we process the input image V through the ViT backbone of the BiomedCLIP encoder. We extract the sequence of patch embeddings from its penultimate layer, a practice that strikes a good balance between spatial detail and semantic abstraction. This process can be formally defined as
F v = BiomedCLIP ( V ) ,
where F v R N v × d v represents the sequence of visual features, where N v is the length of tokens, and d v is the feature dimension of each token. Pseudocode for the architectural details of the BiomedCLIP encoder has been provided in the Appendix A.

3.1.2. Textual Context Encoding

The Med-VQA task requires the model not only to understand the question but also to learn by associating it with the corresponding ground-truth answer A. Therefore, we concatenate the question Q with the corresponding ground-truth answer A and prepend a task-specific instruction I t a s k to form a complete text input T = [ I t a s k ; Q ; A ] . This combined text is then tokenized and fed into the encoder of Flan-T5 [52] to generate a sequence of context-aware token embeddings. This process can be formulated as
F t = Flan-T 5 ( T ) ,
where F t R N t × d t is the embedding sequence of text tokens, where N t is the length of the token sequence, and d t is its feature dimension. Pseudocode for the architectural details of the Flan-T5 encoder has been provided in the Appendix A.

3.2. Clinical-Semantic Disambiguation Module (CSDM)

The CSDM module is central to achieving high-precision diagnostic reasoning within our framework. Its primary objective is to construct a shared highly discriminative multimodal embedding space that enables the model to precisely distinguish pathological states that are visually similar but clinically distinct. To achieve this goal, the CSDM module employs a refined training strategy. This strategy first confronts the unique “False Negative Trap” in medical contrastive learning through our proposed Debiased Hard Negative Mining strategy; subsequently, it leverages a feature fusion mechanism based on cross-modal attention to provide deeply aligned information-rich image–text representations for this mining strategy. Finally, the entire module is directly driven and optimized by a unified contrastive loss function.

3.2.1. Debiased Hard Negative Mining for Robust Representation (DHNM)

A fundamental challenge in medical image analysis lies in the high degree of visual similarity between samples of different pathological states—that is, samples that are very close to the anchor in the embedding space but have different labels. This makes it difficult for standard models to learn discriminative features. Contrastive Learning is a powerful paradigm for addressing such tasks, with the core objective of learning an embedding space where semantically similar samples are pulled closer and semantically dissimilar ones are pushed farther apart. However, in the medical domain, this contrastive learning strategy heavily relies on hard negative samples during training, resulting in a fatal flaw, which this section terms the “False Negative Trap”. The root cause is that many different diseases or pathological states can be highly similar visually. For example, in a batch processing lung cancer images, the anchor sample ( V i , Q i , A i ) is ‘lung adenocarcinoma,’ while another sample in the batch, ( V j , Q j , A j ) , is ‘lung squamous cell carcinoma.’ Due to the high visual similarity of these two types of lung cancer on imaging, a model that relies solely on cross-modal embedding similarity is highly likely to incorrectly identify the text pair ( Q j , A j ) as a ‘hard negative sample’ for the image V i . When the model is trained to push these two samples apart, it is actually learning to erase the subtle yet critical diagnostic features that distinguish these two similar cancers, which severely undermines its clinical diagnostic capability.
To address this issue, we propose a novel Debiased Hard Negative Mining for Robust Representation (DHNM). The overall framework of DHNM is shown in Figure 2. First, we map the high-dimensional feature sequences to the contrastive learning embedding space via a pooling operation (denoted as Pool ( · ) ) and a projection head (a small MLP, denoted as ϕ ) to obtain normalized global representations z v = ϕ v ( Pool ( F v ) ) and z q a = ϕ q a ( Pool ( F t ) ) . For an anchor i in a batch of size B, we compute the cosine similarity between its image representation z v i and all in-batch text representations z q a j (where j i ), and select the indices of the Top-K most similar samples to form a candidate set:
J cand , i = arg top-k j { 1 , , B } , j i ( sim ( z v i , z q a j ) ) .
Figure 2. Flowchart of the Debiased Hard Negative Mining for Robust Representation (DHNM) module. For an anchor image feature F v within a batch, visual similarity is first computed against all textual features F q a in the batch, and the Top-K most similar texts are selected as a candidate set. Subsequently, an independent semantic encoder (Semantic Referee) is used to calculate the semantic similarity between these candidate texts and the anchor text. Only those candidate samples with high visual similarity but low semantic similarity are ultimately confirmed as ‘true’ hard negative samples for the subsequent contrastive learning loss calculation.
The corresponding hard negative sample set is N cand , i = { ( Q j , A j ) j J cand , i } . Next, we perform semantic filtering on the candidate set N cand , i . We utilize the PubMedBERT text encoder integrated within the BiomedCLIP model as a “semantic referee” E sem . For each candidate negative sample ( Q j , A j ) in N cand , i , we compute its semantic similarity with the anchor text ( Q i , A i ) . To avoid the “False Negative Trap”, we discard candidate samples that are semantically too similar to the anchor, retaining only the semantically irrelevant ones as the final hard negative sample set N hard , i . The filtering condition is controlled by a preset threshold θ , a hyperparameter (set to 0.85 in this study):
N hard , i = ( Q j , A j ) j J cand , i   and   sim E sem ( Q i , A i ) , E sem ( Q j , A j ) θ ,
where sim ( · , · ) represents the cosine similarity. This process ensures that only those samples that are visually confusing, but are indeed clinically-semantically irrelevant, are used for the subsequent contrastive learning.

3.2.2. Unified Debiased Contrastive Loss

To directly utilize the high-quality negative samples mined in the previous step to optimize the CSDM module, we adopt a unified optimized InfoNCE loss function L DCL . This loss frames the contrastive learning task as a multi-class classification problem. For a given anchor image z v i , its objective is to distinguish the corresponding positive text sample z q a i from all the true hard negative samples (from N hard , i ) within a set. Its formal definition is as follows:
L D C L = log exp ( sim ( z v i , z q a i ) / τ ) exp ( sim ( z v i , z q a i ) / τ ) + ( Q j , A j ) N hard , i exp ( sim ( z v i , z q a j ) / τ ) ,
where z q a j is the embedding representation corresponding to the hard negative sample ( Q k , A k ) , and τ is the temperature hyperparameter. The InfoNCE loss is itself a hardness-aware loss function: the temperature parameter τ naturally controls the degree of focus on hard negative samples.
The objective of this loss function is to increase exp ( sim ( z v i , z q a i ) to pull z v i and z q a i closer in similarity, while simultaneously decreasing exp ( sim ( z v i , z q a j ) to increase the distance between z v i and the hard negative sample z q a j . Therefore, the model can effectively avoid the False Negative Trap caused by hard negative pairs, thereby enhancing the robustness of the inputs F v and F t in scenarios where they are visually similar but clinically distinct.

3.2.3. Feature Fusion via Cross-Modal Attention (FF-CMA)

Based on the enhanced F v and F t from the previous section, which are robust to hard negative samples, we employ a Feature Fusion via Cross-Modal Attention(FF-CMA) to establish the relationship between F v and F t . The overall framework of FF-CMA is shown in Figure 3.
Figure 3. Architecture diagram of the Feature-Fused Cross-Modal Attention (FF-CMA) module. This module is designed to generate image–text fused representations H f u s e d .
This relationship accurately binds an image to its corresponding text and enhances the model’s accuracy in aligning image–text pairs in scenarios that are visually similar yet clinically distinct. Specifically, we use the textual feature sequence F t as the Query, and the visual feature sequence F v as both the Key and the Value. These inputs are first transformed through learnable linear projections:
Q = F t W Q , K = F v W K , V = F v W V ,
where W Q , W K , W V R d × d k are the corresponding projection matrices. The calculation of the attention weights and their output follows the scaled dot-product attention paradigm:
Attention ( Q , K , V ) = softmax Q K T d k V .
To stabilize training and integrate context, we adopt a standard Transformer block structure that includes residual connections and Layer Normalization, followed by a feed-forward network (FFN), to ultimately generate a “text-aware” fused feature representation  H fused :
H fused = LayerNorm ( Attention ( Q , K , V ) + F t )
H fused = LayerNorm ( FFN ( H fused ) + H fused ) .
The representation H fused captures the fine-grained correspondence between the image and text and is fed into the subsequent modules for processing.

3.3. Prompt-Based Knowledge Consolidation Module (PKC)

To address the “stale knowledge” problem of traditional models when encountering new information and to provide a traceable basis for the model’s answers, we introduce the Prompt-based Knowledge Consolidation Module (PKC). It is a framework based on Retrieval-Augmented Generation, whose core advantage lies in decoupling stable, consolidated reasoning capabilities from a dynamic, updatable knowledge base. The overall framework of PKC is shown in Figure 4.
Figure 4. Operational workflow of the Prompt-based Knowledge Consolidation (PKC) module. (Left) During training, new knowledge is encoded and incrementally added to a non-parametric memory store in a rehearsal-free manner. (Right) During inference, a retrieval-augmented generation process retrieves relevant historical exemplars to provide context for generating fact-based answers.

3.3.1. Memory as a Non-Parametric Knowledge Store

The core of the PKC is an external, persistent, and non-parametric memory store M , which stores a vast number of medical knowledge exemplars in the form of key–value pairs. After the CSDM module has completed its training, its parameters θ CSDM are permanently frozen. We use this frozen CSDM module (denoted as CSDM ( · ; θ CSDM * ) ) to process the ROCO-v2 [53] dataset offline and build the initial version of the memory store. For each V Q A triplet ( V i , Q i , A i ) in the dataset, we encode it into a single-vector key p i R d for efficient retrieval:
p i = Pool ( CSDM ( V i , Q i ; θ CSDM * ) ) ,
where Pool ( · ) can be an average pooling operation. This vector key p i , along with the corresponding raw answer string A i , forms a key–value pair ( p i , A i ) , and is stored in the memory store:
M [ 0 ] = { ( p i , A i ) | i = 1 , , N treatmem } ,
where N treatmem is the total number of exemplars in the initial memory store. The framework’s knowledge update capability is entirely transferred to the non-parametric memory store M . When a new knowledge dataset D new becomes available, we use the same frozen CSDM encoder to generate key–value pairs for the new data and add them incrementally to the memory store; this process never modifies any model parameters:
M [ t + 1 ] = M [ t ] { ( Pool ( CSDM ( V j , Q j ; θ CSDM * ) ) , A j ) ( V j , Q j , A j ) D new } .
This design fundamentally solves the “stale knowledge” problem of traditional models, as it allows the knowledge base to stay synchronized with the latest medical advancements without the need for costly and risky full model retraining.

3.3.2. Retrieval-Augmented Knowledge Replay

During the inference phase, when a new query ( V new , Q new ) arrives, the system follows a standard retrieval-augmented-generation workflow. First, the model uses its fixed CSDM module to encode the new query into a query vector q new :
q new = Pool ( C S D M ( V new , Q new ; θ C S D M * ) ) .
Subsequently, the system uses q new as a query to perform an efficient similarity search in the memory store M to retrieve the k most relevant historical exemplars. This retrieval process aims to find a set of indices J k whose corresponding keys have the highest cosine similarity to the query vector:
J k = arg TopK J M , | J | = k j J sim ( q new , p j ) .
The retrieved k answers A j j J k are concatenated into a single contextual prompt string C retrieved = concat ( A j 1 , , A j k ) . Finally, this retrieved context, along with the original question Q new , forms an augmented prompt that is fed into the generative decoder to generate the final answer, guided by the fused features of the current query, H fused , new .

3.4. Training and Inference

This section will integrate the aforementioned modules and introduce the model’s overall training objective and inference process.

3.4.1. Training Objective

In the initial training phase (i.e., before the CSDM module weights are frozen), the entire model is guided by a composite loss function. This function is designed to simultaneously shape the model’s generative capabilities and its core debiased discriminative ability. The total loss L total is a weighted sum of the autoregressive generation loss L LM and the unified debiased contrastive loss L DCL defined in Section 3.2.3:
L total = λ L LM + L DCL ,
where λ is a hyperparameter that balances the contributions of the generation task and the contrastive learning task. L LM is the core of the model’s generative capability and adopts the standard autoregressive cross-entropy loss. It aims to maximize the probability of generating the ground-truth answer sequence A = ( a 1 , a 2 , , a T ) , given the fused feature representation H fused :
L LM = t = 1 T log P ( a t a < t , A i ) ,
to enhance the model’s image–question–answer matching accuracy, while, through L DCL , improving the model’s robustness to ’confusing samples’, thereby improving the overall performance of multimodal alignment.

3.4.2. Autoregressive Answer Generation

During the inference phase, the answer generation for a novel image–question pair ( V new , Q new ) follows a carefully designed multi-step deterministic process to ensure the accuracy and trustworthiness of the results. The process begins by using the CSDM module with its frozen parameters to encode the input, producing a unified multimodal fused feature representation H fused . This high-dimensional representation encapsulates the fine-grained semantic alignment information between the relevant visual evidence in the image and the question text, serving as a critical bridge connecting perception with subsequent cognitive reasoning. Subsequently, H fused is used by the Prompt-based Knowledge Consolidation module to perform retrieval-based knowledge enhancement.
Specifically, a query vector is generated by pooling H fused and is used to perform efficient similarity retrieval in the non-parametric memory store M to locate the K most relevant historical medical exemplars. The answer texts corresponding to these exemplars are concatenated into a single contextual string C retrieved , providing external knowledge support for the generation process. Ultimately, the answer generation task is undertaken by the autoregressive decoder of Flan-T5. For its pseudocode, refer to the Appendix A. As the decoder generates each token, its state is not only guided by the augmented text prompt composed of the original question Q new and the retrieved context C retrieved , but more importantly, it is continuously conditioned on H fused to integrate visual information via the cross-modal attention mechanism. Regarding the choice of decoding strategy, considering that the determinism and reproducibility of the output are crucial in the high-risk medical domain, this study explicitly and exclusively adopts the Beam Search strategy. This strategy, compared to greedy search, which is prone to getting trapped in local optima, and various sampling methods (such as Top-p sampling) that introduce uncontrollable randomness, can more stably explore and output the answer sequence with the highest overall probability, thereby mechanistically ensuring the reliability and safety of the model’s output results.

4. Experiment

This section aims to provide a comprehensive and in-depth empirical evaluation of the Evolvable Clinical-Semantic Alignment (ECSA) framework proposed in this paper through a series of rigorous experiments. Our experimental design revolves around three core objectives: first, to compare ECSA with SOTA models across multiple dimensions on standard in-domain evaluation benchmarks; second, to quantify the generalization capability of ECSA in few-shot scenarios through specially designed dataset adaptation experiments and directly compare it with key related works; and finally, to dissect the independent contributions and synergistic effects of each key component within the framework through exhaustive ablation studies.

4.1. Datasets

To comprehensively evaluate the model performance and construct the knowledge enhancement module, this study utilized a total of five public datasets. VQA-RAD, SLAKE, PathVQA, and VQA-Med-2019 were used for model training and evaluation, while ROCO-v2 [53] was specifically used as an external knowledge base.
VQA-RAD: This is a radiology VQA dataset manually constructed by clinicians. The dataset contains 315 unique radiology images (such as CT, MRI, X-rays), covering multiple body parts including the head, chest, and abdomen, and is accompanied by 3515 question–answer pairs.
SLAKE: This is a larger and more diverse bilingual (English and Chinese) Med-VQA dataset. Following common practice, we only use its English portion, which contains 642 images and approximately 7000 question–answer pairs. Compared to VQA-RAD, SLAKE covers a broader range of imaging modalities and anatomical regions. One of its notable features is that, in addition to questions based purely on visual content, it also includes knowledge-based questions that require external medical knowledge to answer.
PathVQA: This is a VQA dataset focused on pathology images, containing 4998 images and nearly 33,000 question–answer pairs extracted from pathology textbooks and digital libraries. A key feature of this dataset is the high proportion of open-ended questions (approximately 50.2%), which presents a greater challenge for the models.
VQA-Med-2019: As part of the ImageCLEF challenge, this dataset focuses on radiology images, with a training set of 3200 images and 12,792 question–answer pairs. Its questions are primarily centered around four categories: Modality, Plane, Organ System, and Abnormality, designed to test the model’s ability to answer based solely on image content.
ROCO-v2 (Radiology Objects in COntext version 2): This is a large-scale radiology multimodal dataset containing over 81,000 images extracted from public PubMed Central articles and their corresponding captions. The dataset covers multiple imaging modalities such as CT, MRI, and X-ray.

4.2. Evaluation Metrics

To comprehensively evaluate the model performance from different dimensions, we adopted two core evaluation metrics. Accuracy is the primary performance basis in all experimental scenarios. Additionally, to specifically quantify the model’s knowledge retention capability in the ablation studies (Section 4.6), we also introduced the Forgetting rate (F).
Accuracy: In line with the prevailing evaluation practices in the Med-VQA community, we perform exact-string matching to determine correctness. This metric serves as the primary yardstick for model performance, and we formulate it via the confusion-matrix-based definition:
Accuracy = TP + TN TP + TN + FP + FN ,
where T P , T N , F P , and F N denote true positives, true negatives, false positives, and false negatives, respectively, obtained by treating each unique textual answer as a distinct class. We report accuracy on the overall, open-ended, and closed-ended question subsets.
Forgetting (F): This metric is specifically used in the ablation studies to quantify the model’s knowledge retention capability during the continual learning process. It measures the degree to which the model’s performance on old tasks degrades after learning a new task. Its calculation formula is as follows:
F K = 1 K 1 j = 1 K 1 ( a j , j a K , j ) ,
where a K , j represents the accuracy of the model on the test set of the j-th task after having learned the K-th task, and a j , j represents the best accuracy achieved by the model right after learning the j-th task. The lower the value of the forgetting rate F, the milder the catastrophic forgetting phenomenon.

4.3. Experimental Setup

4.3.1. Implementation Details

Our ECSA framework is based on two powerful pre-trained foundation models. The visual encoder uses the ViT-B/16 model from BiomedCLIP (approx. 86 million parameters), and the language model is Flan-T5-Base, which contains approximately 250 million parameters. Crucially, our parameter-efficient approach freezes the backbones of these large models. The trainable parameters are limited only to the Clinical-Semantic Disambiguation module, as the Prompt-based Knowledge Consolidation module is a non-parametric memory store with no trainable weights. This trainable module totals approximately 8.4 million parameters, representing less than 3% of the entire model’s parameter count. All models are trained with a unified hyperparameter setup: we use the AdamW optimizer with an initial learning rate set to 1 × 10 4 and employ a linear decay learning rate scheduling strategy. The batch size is set to 16. In our composite loss function, the weight hyperparameter λ that balances the generation loss and the contrastive loss is set to 1.0 . In the CSDM module, the temperature hyperparameter τ for the InfoNCE loss is set to 0.07 , and the similarity threshold θ for debiased hard negative mining is set to 0.85 . During the inference phase, the Prompt-based Knowledge Consolidation module retrieves the k = 3 most relevant exemplars. All experiments are conducted on a single NVIDIA GeForce RTX 3090 GPU with 24 GB of VRAM.

4.3.2. Hyperparameter Sensitivity Analysis

We conducted a sensitivity analysis on the selection of key hyperparameters to provide an experimental basis for their values, with the experimental data shown in Figure 5. To determine the optimal configuration, we adopted a controlled experiment approach: when validating a specific hyperparameter, all other parameters were kept at the default values described in Section 4.3.1. The results in Figure 5 provide a solid empirical basis for our hyperparameter choices. The experiments were conducted on the validation sets of VQA-RAD and SLAKE, respectively. The results show that when the loss function weight λ = 1.0 , the model achieves the best balance between the generation task and the contrastive learning task. For the contrastive loss temperature τ , a setting of 0.07 best balances the focus on hard negative samples. For the hard negative mining threshold θ , a threshold of 0.85 strikes the optimal balance between effectively filtering semantically similar “false negative” samples and retaining a sufficient number of hard negative samples, thereby maximizing the model’s discriminative ability.
Figure 5. Sensitivity analysis for key hyperparameters on the VQA-RAD and SLAKE validation sets. (Left) Model accuracy across different values of the loss weight λ . (Center) Model accuracy as a function of the contrastive loss temperature τ . (Right) Model accuracy under varying similarity thresholds θ for the DHNM strategy. Performance peaks indicate the empirically determined optimal values.

4.4. Comparison with SOTA Models

To place the performance of ECSA in the broader context of Med-VQA research, this section compares its performance under the standard in-domain evaluation paradigm with several non-continual learning SOTA models that have reported the best performance on the VQA-RAD, SLAKE, PathVQA, and VQA-Med-2019 test sets. All comparisons use publicly verifiable data reported in the original papers of the baseline models, ensuring the fairness and transparency of the comparison.
The data in Table 1 contextualize the performance of ECSA across four distinct datasets, benchmarking it against current SOTA models. On the two primary radiology datasets, VQA-RAD and SLAKE, ECSA achieves highly competitive overall accuracies of 80.15% and 85.10% respectively, outperforming several recent models like LaPA and performing on par with AMiF. Notably, on VQA-RAD’s more challenging open-ended questions, ECSA’s accuracy of 71.25% is particularly outstanding.
Table 1. Accuracy of models on the VQA-RAD, SLAKE, PathVQA, and VQA-Med-2019 test sets. In the table, bold indicates the best result, and underlined values denote the second-best.
To further assess the model’s robustness, its performance was evaluated on two additional datasets. On VQA-Med-2019, another radiology dataset, the ECSA achieves a strong accuracy of 82.23%, demonstrating excellent in-domain generalization. On the more challenging cross-domain PathVQA dataset, which focuses on pathology and features a higher proportion of difficult open-ended questions, the ECSA attains an overall accuracy of 64.57%. While the demanding nature of its open-ended questions (39.64%) impacts the overall score, the model maintains a high accuracy of 89.50% on closed-ended questions, confirming the solidity of its foundational capabilities.  
This comprehensive performance profile strongly supports our framework’s design philosophy. Models like AMiF typically adopt an end-to-end fine-tuning strategy to achieve peak performance on a single fixed dataset. In contrast, ECSA’s parameter-efficient approach—freezing most of the backbone network—achieves comparable or superior performance on multiple radiology benchmarks while also demonstrating reasonable generalization to a new medical domain. This validates that our design achieves an excellent balance between high performance and an evolvable efficient architecture.

4.5. Generalization Ability Evaluation

To evaluate the model’s generalization capability in few-shot scenarios, we follow the dataset adaptation (DA) paradigm from the paper “Multimodal Prompt Retrieval for Generative Visual Question Answering” (MPR) and directly compare with its reported results. The experimental setup is a k = 1 few-shot domain adaptation, where after the model is trained on a source dataset, during inference on the target dataset, it is allowed to retrieve one (k = 1) most similar question–answer pair from the target dataset as a contextual prompt.
The results in Table 2 clearly indicate that ECSA demonstrates a significant performance advantage in the few-shot domain adaptation task, comprehensively outperforming the MPR model, which serves as the basis for our work. In the adaptation task from SLAKE to VQA-RAD, ECSA achieves an overall accuracy of 67.5%, a significant improvement of 4.5 percentage points over MPR’s 63.0%. In the reverse adaptation task from VQA-RAD to SLAKE, ECSA also substantially leads MPR with an overall accuracy of 58.2%. This result is significant as it directly validates the effectiveness of our framework’s design. The performance improvement of ECSA is primarily attributed to its superior retrieval capability, which is driven by the Prompt-based Knowledge Consolidation Module (PKC). As an efficient Retrieval-Augmented Generation (RAG) mechanism, PKC can accurately find the most relevant historical exemplars for the current query in few-shot scenarios. This superior retrieval capability is made possible because the query vectors and memory store keys it relies on for retrieval are both generated by the Clinical-Semantic Disambiguation Module (CSDM). Through its hierarchical alignment and debiased hard negative mining strategy, CSDM can produce more discriminative image–text joint embeddings. Therefore, it is the synergistic effect between the high-quality representations provided by CSDM and the efficient retrieval mechanism of PKC that enables ECSA to achieve superior performance in few-shot generalization tasks.
Table 2. Cross-dataset evaluation results.

4.6. Ablation Study

To thoroughly validate the actual contributions of each innovative component in the ECSA framework, we conducted a series of rigorous ablation studies. By systematically removing or replacing key modules within the framework, we can quantify the impact of each part on the model’s final performance.
Our study begins with the validation of the base model selection. Before constructing a complex framework, it is essential to first establish a solid foundation. Therefore, starting with the generic models used in MPR (Generic CLIP + Generic T5), we progressively replace them with domain-specific models through ablation studies to demonstrate the rationale for our choices.
The results in Table 3 clearly demonstrate the importance of domain-specific models. Starting from the generic model combination, whether individually replacing the visual encoder with the biomedically pre-trained BiomedCLIP or individually replacing the language model with the instruction-finetuned Flan-T5, both lead to performance improvements on both datasets. This, respectively, highlights the importance of a domain-specific visual encoder for understanding subtle pathological features in medical images and the advantage of an instruction-finetuned language model for accurately parsing and generating complex medical question–answers. When we use both BiomedCLIP and Flan-T5 simultaneously, the performance is optimal, proving that their combination provides the most powerful foundation for our framework.
Table 3. Results of ablation study on base model selection.
After establishing the optimal base model combination, we further analyzed the necessity of several core innovative modules in ECSA. To rigorously evaluate the performance of each component in a continual learning setting, we constructed a challenging sequence of five tasks. This sequence was formed by strategically partitioning the VQA-RAD and SLAKE datasets, aiming to simulate a real-world scenario of Domain-Incremental Learning. The task sequence is as follows: Task 1: radiological abnormality recognition (the ‘abnormality’ category of VQA-RAD), Task 2: multimodal organ localization (the ‘organ system’ category of SLAKE), Task 3: anatomical plane and modality classification (the ‘plane’ and ‘modality’ categories of VQA-RAD), Task 4: general visual attribute question answering (the ‘color’, ‘shape’, etc., categories of SLAKE), and Task 5: comprehensive mixed-domain evaluation (reserved samples from both datasets). This design ensures significant distribution shifts between tasks, providing a stringent testbed for evaluating catastrophic forgetting.
The comprehensive forgetting rate reported in the table is the final value calculated according to the formula F 5 = 1 4 j = 1 4 ( a j , j a 5 , j ) after the model has completed training on all five tasks. This value explicitly quantifies the average performance degradation after the model has learned the entire sequence, relative to the best performance it achieved on each of tasks 1 to 4.
The quantitative results in Table 4 systematically deconstruct the success factors of ECSA. All experiments were conducted on the aforementioned five-task sequence, and the table reports the model’s average accuracy on the two datasets after the completion of sequential learning, as well as the comprehensive forgetting rate over the entire sequence. The results clearly demonstrate the contribution of each component in the framework to maintaining knowledge stability. The full ECSA model achieved a comprehensive forgetting rate of 18.5%. Although this design is not intended to compete with the current state-of-the-art continual learning methods that rely on data replay, the effectiveness of its internal components has been fully validated. When the Prompt-based Knowledge Consolidation module (w/o PKC), which serves as the core memory mechanism, was removed, the model’s performance experienced a catastrophic collapse, with the forgetting rate soaring to 33.2%. This irrefutably establishes that the PKC is the cornerstone for mitigating catastrophic forgetting within the framework. Similarly, when the Clinical-Semantic Disambiguation module (w/o CSDM) was removed, the forgetting rate also rose significantly to 26.1%, which indicates that the high-quality aligned representations produced by the CSDM module are crucial for the effective accumulation and consolidation of subsequent knowledge. Finally, after replacing our proposed debiased hard negative mining strategy with standard negative sampling (w/o Debiased Mining), the forgetting rate also increased to 21.3%. This clearly shows that our proposed mining strategy not only enhances the model’s discriminative ability, but the more robust and refined medical representations it learns are also more conducive to resisting knowledge forgetting.
Table 4. Results of ablation study on key components of ECSA.

4.7. Qualitative Analysis

Figure 6 provides a qualitative analysis that highlights the practical advantages of our ECSA framework. Our model demonstrates superior diagnostic accuracy, correctly identifying abnormalities in the liver and appendix, where a baseline approach might fail (top row). This enhanced discriminative capability is a direct result of the Clinical-Semantic Disambiguation Module (CSDM). By employing a debiased hard negative mining strategy, CSDM learns highly robust representations that can distinguish subtle pathological markers from normal tissue, which is crucial for accurate initial assessments.
Figure 6. Qualitative comparison between ECSA and a baseline model on representative clinical cases. (Top Row) ECSA demonstrates superior diagnostic accuracy, correctly identifying abnormalities in the liver and appendix that were missed by the baseline. (Bottom Row) ECSA generates more comprehensive and clinically precise answers, such as identifying bilateral pathology and specific imaging modalities.
Furthermore, the ECSA exhibits a capacity for more comprehensive and clinically nuanced reasoning. For instance, it correctly identifies pathology as affecting ‘bilateral lungs’ rather than a single lung (bottom row, left) and provides a more complete description of findings related to the appendix, noting both ‘extraluminal air and small fluid collection’ (bottom row, right). This ability to generate detailed and complete answers stems from the rich context-aware features produced by CSDM’s cross-modal attention mechanism. The model’s proficiency in knowledge-intensive tasks is also evident, as it correctly identifies the specific imaging modality as ‘MRI Diffusion Weighted’ (bottom row, middle). This level of specificity is enabled by the Prompt-based Knowledge Consolidation Module (PKC), which functions as a retrieval-augmented knowledge base to supply relevant factual information, grounding the model’s response in a vast repository of clinical examples. Together, these cases illustrate how ECSA’s components synergistically contribute to generating responses that are not only more accurate but also more clinically precise and reliable.

5. Conclusions

This paper addresses the two long-standing core challenges in the field of Medical Visual Question Answering—weak few-shot generalization and catastrophic forgetting in continual learning—by proposing an innovative unified framework named the Evolvable Clinical-Semantic Alignment framework. This framework achieves efficient knowledge alignment, transfer, and accumulation in a parameter-efficient manner, built upon frozen pre-trained large vision and language models.
Our main contributions are embodied in two meticulously designed modules:
1.
Clinical-Semantic Disambiguation Module: By introducing a novel debiased hard negative mining strategy, it effectively addresses the “False Negative Trap” of traditional contrastive learning in medical scenarios, significantly enhancing the model’s ability to discriminate difficult cases that are visually similar but clinically distinct;
2.
Prompt-based Knowledge Consolidation Module: Acting as a rehearsal-free dynamic knowledge base, it successfully retains knowledge learned from old tasks without storing any past patient data, thereby effectively overcoming catastrophic forgetting;
Comprehensive experimental evaluations quantitatively validate the effectiveness of our framework. ECSA achieves state-of-the-art or highly competitive performance across four public benchmarks, attaining overall accuracies of 80.15% on VQA-RAD and 85.10% on SLAKE, while maintaining this performance by fine-tuning less than 3% of the total parameters. In a challenging one-shot cross-dataset generalization task, the framework demonstrates a significant 4.5 percentage point improvement over strong baselines, directly validating the synergy between CSDM’s discriminative representations and PKC’s retrieval capabilities. Critically, in a five-task continual learning scenario, ECSA achieves a low forgetting rate of just 13.50%. Ablation studies confirm that this robust knowledge retention is primarily driven by the PKC module, the removal of which causes the forgetting rate to increase to 32.20%.
Limitations: While the ECSA has demonstrated promising results, we acknowledge several limitations that open avenues for future exploration. First, the scalability and generalizability of our framework were tested on standard public benchmarks, which, while representative, are limited in size and diversity compared to real-world clinical data. The model’s performance on larger institution-specific datasets remains to be validated. Second, while the framework was tested across radiology and pathology, its adaptability to other imaging modalities requires further investigation.
Future Work: Based on these limitations, future work could proceed in two main directions: (1) Conducting large-scale clinical validation. Applying the framework to broader institution-specific datasets and more diverse imaging modalities is key to verifying its ultimate value in real-world clinical workflows. (2) Integrating structured knowledge. Future research could explore integrating medical knowledge graphs into the PKC module to enhance the model’s complex reasoning capabilities. In conclusion, this research provides a promising path for building more robust, adaptive, and trustworthy next-generation clinical intelligent assistant systems.

Author Contributions

Conceptualization, Q.J. and S.L.; Methodology, Q.J.; Validation, Q.J., M.C., T.L., and J.Y.; Formal Analysis, Q.J.; Investigation, Q.J. and M.C.; Resources, Q.J., S.L., T.L., and J.Y.; Data Curation, Q.J.; Writing—Original Draft Preparation, Q.J.; Writing—Review and Editing, Q.J. and S.L.; Visualization, Q.J.; Supervision, S.L. and M.C.; Project Administration, S.L. and M.C. All authors have read and agreed to the published version of the manuscript.

Funding

National Natural Science Foundation of China: 61762085.

Data Availability Statement

The datasets used and analyzed in this study are all publicly available.

Acknowledgments

We sincerely thank all the authors cited in this paper for their valuable contributions.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Algorithm A1 BiomedCLIP Vision Transformer Encoder
   Input: Image I R H × W × 3
   Output: Sequence of visual features F v R N v × d v
1:
  function ViT_Encoder(I)
2:
         X p ImageToPatches ( I , P )
3:
         X f l a t Flatten ( X p )
4:
         E LinearProjection ( X f l a t )
5:
         x c l a s s LearnableClassToken ( )
6:
         Z 0 Concat ( [ x c l a s s , E ] )
7:
         Z 0 Z 0 + PositionalEncoding
8:
        for  l = 1 to L do
9:
               A l LayerNorm ( Z l 1 )
10:
             A l MultiHeadSelfAttention ( Q = A l , K = A l , V = A l )
11:
             Z l Z l 1 + A l
12:
             F l LayerNorm ( Z l )
13:
             F l FeedForwardNetwork ( F l )
14:
             Z l Z l + F l
15:
      end for
16:
       F v LayerNorm ( Z L )
17:
      return  F v
18:
end function
Algorithm A2 Flan-T5 Textual Context Encoder
   Input: Instruction I t a s k , Question Q, Ground-truth Answer A
   Output: Textual features F t R N t × d t
1:
  function TextualContextEncoder( I t a s k , Q , A )
2:
         T c o m b i n e d Concatenate ( I t a s k , Q , A )
3:
         E 0 TokenEmbedding ( T c o m b i n e d ) + PositionalEncoding
4:
        for  l = 1 to L e n c  do
5:
               E l TransformerEncoderLayer ( E l 1 )
6:
        end for
7:
         F t LayerNorm ( E l )
8:
        return  F t
9:
  end function
Algorithm A3 Flan-T5 Autoregressive Answer Generation (Inference)
  •    Input: Fused multimodal features H f u s e d , New question Q n e w , Memory store M
  •    Output: Generated answer sequence A g e n
  •    function AnswerGeneration( H f u s e d , Q n e w , M )
  •           q r e t r i e v a l Pool ( H f u s e d )
  •           C r e t r i e v e d RetrieveExemplars ( q r e t r i e v a l , M , K )
  •           P p r o m p t Concatenate ( Q n e w , C r e t r i e v e d )
  •           D 0 TokenEmbedding ( P p r o m p t ) + PositionalEncoding
  •           A g e n BeamSearchInitialize ( D 0 )
  •          for  i = 1  to MaxLength do
  •                 D i n GetLastHiddenStates ( A g e n )
  •                for  l = 1  to  L d e c  do
  •                       D l MaskedMultiHeadAttention ( Q = D i n , K = D i n , V = D i n )
  •                       D l LayerNorm ( D i n + D l )
  •                       D l MultiHeadAttention ( Q = D l , K = H f u s e d , V = H f u s e d )
  •                       D l LayerNorm ( D l + D l )
  •                       D o u t FeedForwardNetwork ( D l )
  •                       D i n LayerNorm ( D l + D o u t )
  •                end for
  •                 A g e n BeamSearchStep ( A g e n , D i n )
  •          end for
  •          return  BestHypothesis ( A g e n )
  •    end function

References

  1. Dong, W.; Shen, S.; Han, Y.; Tan, T.; Wu, J.; Xu, H. Generative models in medical visual question answering: A survey. Appl. Sci. 2025, 15, 2983. [Google Scholar] [CrossRef]
  2. Maity, S.; Saikia, M.J. Large Language Models in Healthcare and Medical Applications: A Review. Bioengineering 2025, 12, 631. [Google Scholar] [CrossRef] [PubMed]
  3. Song, X.; Shi, T.; Feng, Z.; Song, M.; Lin, J.; Lin, C.; Fan, C.; Yuan, Y. Unsupervised learning facial parameter regressor for action unit intensity estimation via differentiable renderer. In Proceedings of the 28th ACM International Conference on Multimedia, Seattle, WA, USA, 12–16 October 2020; pp. 2842–2851. [Google Scholar]
  4. Zhang, X.; Wu, C.; Zhao, Z.; Lin, W.; Zhang, Y.; Wang, Y.; Xie, W. Development of a large-scale medical visual question-answering dataset. Commun. Med. 2024, 4, 277. [Google Scholar] [CrossRef]
  5. Wu, Q.; Teney, D.; Wang, P.; Shen, C.; Dick, A.; Van Den Hengel, A. Visual question answering: A survey of methods and datasets. Comput. Vis. Image Underst. 2017, 163, 21–40. [Google Scholar] [CrossRef]
  6. Bian, D.; Wang, X.; Li, M. Diffusion-based visual representation learning for medical question answering. In Proceedings of the 15th Asian Conference on Machine Learning, PMLR, Istanbul, Turkey, 11–14 November 2023; Yanıkoğlu, B., Buntine, W., Eds.; pp. 169–184. [Google Scholar]
  7. Chen, P.; Wang, J.; Luo, L.; Lin, H.; Yang, Z. Learning to explain is a good biomedical few-shot learner. Bioinformatics 2024, 40, btae589. [Google Scholar] [CrossRef]
  8. Liu, J.; Fan, K.; Cai, X.; Niranjan, M. Few-shot learning for inference in medical imaging with subspace feature representations. PLoS ONE 2024, 19, e0309368. [Google Scholar] [CrossRef]
  9. Wu, Z.; Guo, K.; Luo, E.; Wang, T.; Wang, S.; Yang, Y.; Zhu, X.; Ding, R. Medical long-tailed learning for imbalanced data: Bibliometric analysis. Comput. Methods Programs Biomed. 2024, 247, 108106. [Google Scholar] [CrossRef]
  10. Zhou, S.K.; Greenspan, H.; Davatzikos, C.; Duncan, J.S.; Van Ginneken, B.; Madabhushi, A.; Prince, J.L.; Rueckert, D.; Summers, R.M. A review of deep learning in medical imaging: Imaging traits, technology trends, case studies with progress highlights, and future promises. Proc. IEEE 2021, 109, 820–838. [Google Scholar] [CrossRef]
  11. Wang, W.; Wang, J.; Chen, C.; Jiao, J.; Cai, Y.; Song, S.; Li, J. Fremim: Fourier transform meets masked image modeling for medical image segmentation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 2–7 January 2023; pp. 7860–7870. [Google Scholar]
  12. Verma, T.; Jin, L.; Zhou, J.; Huang, J.; Tan, M.; Choong, B.C.M.; Tan, T.F.; Gao, F.; Xu, X.; Ting, D.S.; et al. Privacy-preserving continual learning methods for medical image classification: A comparative analysis. Front. Med. 2023, 10, 1227515. [Google Scholar] [CrossRef]
  13. Zheng, G.; Braverman, V.; Leal, J.; Rowe, S.; Leung, D.; Jacobs, M.A.; Parekh, V.S. Towards a Collective Medical Imaging AI: Enabling Continual Learning from Peers. Proc. Med. Imaging Deep. Learn. 2024, 250, 1862–1877. [Google Scholar]
  14. Rose, E.D. Continual Learning: An Approach via Feature Maps Extrapolation (Short Paper). In Proceedings of the AIxIA Doctoral Consortium 2023 Co-Located with the 22nd International Conference of the Italian Association for Artificial Intelligence (AIxIA 2023), Rome, Italy, 6–7 November 2023; Poggioni, V., Rossi, S., Eds.; CEUR-WS.org. 2023; Volume 3670. [Google Scholar]
  15. Yu, T.; Tong, Z.; Yu, J.; Zhang, K. Fine-grained Adaptive Visual Prompt for Generative Medical Visual Question Answering. In Proceedings of the AAAI Conference on Artificial Intelligence, Philadelphia, PA, USA, 25 February–4 March 2025; Volume 39, pp. 9662–9670. [Google Scholar]
  16. Ossowski, T.; Hu, J. Retrieving Multimodal Prompts for Generative Visual Question Answering. In Proceedings of the Findings of the Association for Computational Linguistics: ACL 2023, Toronto, ON, Canada, 9–14 July 2023; Rogers, A., Boyd-Graber, J., Okazaki, N., Eds.; pp. 2518–2535. [Google Scholar] [CrossRef]
  17. Huang, W.; Li, C.; Zhou, H.Y.; Yang, H.; Liu, J.; Liang, Y.; Zheng, H.; Zhang, S.; Wang, S. Enhancing representation in radiography-reports foundation model: A granular alignment algorithm using masked contrastive learning. Nat. Commun. 2024, 15, 7620. [Google Scholar] [CrossRef] [PubMed]
  18. Dutt, R.; Ericsson, L.; Sanchez, P.; Tsaftaris, S.A.; Hospedales, T. Parameter-Efficient Fine-Tuning for Medical Image Analysis: The Missed Opportunity. In Proceedings of the 7nd International Conference on Medical Imaging with Deep Learning, Paris, France, 3–5 July 2024; Burgos, N., Petitjean, C., Vakalopoulou, M., Christodoulidis, S., Coupe, P., Delingette, H., Lartizien, C., Mateus, D., Eds.; Volume 250, pp. 406–425. [Google Scholar]
  19. Lau, J.J.; Gayen, S.; Ben Abacha, A.; Demner-Fushman, D. A dataset of clinically generated visual questions and answers about radiology images. Sci. Data 2018, 5, 1–10. [Google Scholar] [CrossRef] [PubMed]
  20. Liu, B.; Zhan, L.M.; Xu, L.; Ma, L.; Yang, Y.; Wu, X.M. Slake: A semantically-labeled knowledge-enhanced dataset for medical visual question answering. In Proceedings of the 2021 IEEE 18th International Symposium on Biomedical Imaging (ISBI), Nice, France, 13–16 April 2021; pp. 1650–1654. [Google Scholar]
  21. He, X. Towards Visual Question Answering on Pathology Images. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (ACL-IJCNLP 2021), Virtual, 1–6 August 2021; Volume 2. [Google Scholar]
  22. Ben Abacha, A.; Hasan, S.A.; Datla, V.V.; Demner-Fushman, D.; Müller, H. Vqa-med: Overview of the medical visual question answering task at imageclef 2019. In Proceedings of the CLEF (Conference and Labs of the Evaluation Forum) 2019 Working Notes, Lugano, Switzerland, 9–12 September 2019. [Google Scholar]
  23. Xia, J.; Tan, L.; Dai, P.; Zhao, M.; Wu, Y.; Cao, L. Attention disturbance and dual-path constraint network for occluded person re-identification. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 20–27 February 2024; Volume 38, pp. 6198–6206. [Google Scholar]
  24. Birkun, A.A.; Gautam, A. Large language model (LLM)-powered chatbots fail to generate guideline-consistent content on resuscitation and may provide potentially harmful advice. Prehospital Disaster Med. 2023, 38, 757–763. [Google Scholar] [CrossRef] [PubMed]
  25. McCloskey, M.; Cohen, N.J. Catastrophic interference in connectionist networks: The sequential learning problem. In Psychology of Learning and Motivation; Elsevier: Amsterdam, The Netherlands, 1989; Volume 24, pp. 109–165. [Google Scholar]
  26. French, R.M. Catastrophic forgetting in connectionist networks. Trends Cogn. Sci. 1999, 3, 128–135. [Google Scholar] [CrossRef]
  27. Cao, Y.H.; Yu, H.; Wu, J. Training vision transformers with only 2040 images. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 220–237. [Google Scholar]
  28. Zhang, X.; Zhang, F.; Xu, C. Vqacl: A novel visual question answering continual learning setting. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 19102–19112. [Google Scholar]
  29. Kirkpatrick, J.; Pascanu, R.; Rabinowitz, N.; Veness, J.; Desjardins, G.; Rusu, A.A.; Milan, K.; Quan, J.; Ramalho, T.; Grabska-Barwinska, A.; et al. Overcoming catastrophic forgetting in neural networks. Proc. Natl. Acad. Sci. USA 2017, 114, 3521–3526. [Google Scholar] [CrossRef]
  30. Zenke, F.; Poole, B.; Ganguli, S. Continual learning through synaptic intelligence. In Proceedings of the International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; pp. 3987–3995. [Google Scholar]
  31. Lester, B.; Al-Rfou, R.; Constant, N. The Power of Scale for Parameter-Efficient Prompt Tuning. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing (EMNLP 2021), Virtual, 7– 11 November 2021; Moens, M.F., Huang, X., Specia, L., Yih, S.W.t., Eds.; pp. 3045–3059. [Google Scholar] [CrossRef]
  32. Lei, S.W.; Gao, D.; Wu, J.Z.; Wang, Y.; Liu, W.; Zhang, M.; Shou, M.Z. Symbolic replay: Scene graph as prompt for continual learning on vqa task. In Proceedings of the AAAI Conference on Artificial Intelligence, Washington, DC, USA, 7–14 February 2023; Volume 37, pp. 1250–1259. [Google Scholar]
  33. Wang, Z.; Zhang, Z.; Lee, C.Y.; Zhang, H.; Sun, R.; Ren, X.; Su, G.; Perot, V.; Dy, J.; Pfister, T. Learning to prompt for continual learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 139–149. [Google Scholar]
  34. Wang, Z.; Zhang, Z.; Ebrahimi, S.; Sun, R.; Zhang, H.; Lee, C.Y.; Ren, X.; Su, G.; Perot, V.; Dy, J.; et al. Dualprompt: Complementary prompting for rehearsal-free continual learning. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 631–648. [Google Scholar]
  35. Gao, Z.; Cen, J.; Chang, X. Consistent prompting for rehearsal-free continual learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 17–21 June 2024; pp. 28463–28473. [Google Scholar]
  36. Qian, Z.; Wang, X.; Duan, X.; Qin, P.; Li, Y.; Zhu, W. Decouple before interact: Multi-modal prompt learning for continual visual question answering. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 1–6 October 2023; pp. 2953–2962. [Google Scholar]
  37. Liu, W.; Zhu, F.; Wei, L.; Tian, Q. C-CLIP: Multimodal continual learning for vision-language model. In Proceedings of the Thirteenth International Conference on Learning Representations, Singapore, 24–28 April 2025. [Google Scholar]
  38. Lewis, P.; Perez, E.; Piktus, A.; Petroni, F.; Karpukhin, V.; Goyal, N.; Küttler, H.; Lewis, M.; Yih, W.t.; Rocktäschel, T.; et al. Retrieval-augmented generation for knowledge-intensive nlp tasks. Adv. Neural Inf. Process. Syst. 2020, 33, 9459–9474. [Google Scholar]
  39. Amugongo, L.M.; Mascheroni, P.; Brooks, S.; Doering, S.; Seidel, J. Retrieval augmented generation for large language models in healthcare: A systematic review. PLoS Digit. Health 2025, 4, e0000877. [Google Scholar] [CrossRef]
  40. Lin, W.; Chen, J.; Mei, J.; Coca, A.; Byrne, B. Fine-grained late-interaction multi-modal retrieval for retrieval augmented visual question answering. Adv. Neural Inf. Process. Syst. 2023, 36, 22820–22840. [Google Scholar]
  41. Yang, Z.; Qin, J.; Chen, J.; Lin, L.; Liang, X. LogicSolver: Towards Interpretable Math Word Problem Solving with Logical Prompt-enhanced Learning. In Proceedings of the Findings of the Association for Computational Linguistics: EMNLP, Abu Dhabi, United Arab Emirates, 7–11 December 2022; Goldberg, Y., Kozareva, Z., Zhang, Y., Eds.; Association for Computational Linguistics: Abu Dhabi, United Arab Emirates, 2022; pp. 1–13. [Google Scholar] [CrossRef]
  42. Xia, P.; Zhu, K.; Li, H.; Wang, T.; Shi, W.; Wang, S.; Zhang, L.; Zou, J.; Yao, H. MMed-RAG: Versatile Multimodal RAG System for Medical Vision Language Models. In Proceedings of the Thirteenth International Conference on Learning Representations, Singapore, 24–28 April 2025. [Google Scholar]
  43. Chu, Y.W.; Zhang, K.; Malon, C.; Min, M.R. Reducing hallucinations of medical multimodal large language models with visual retrieval-augmented generation. arXiv 2025, arXiv:2502.15040. [Google Scholar] [CrossRef]
  44. Li, Y.; Li, Y.; Wang, X.; Jiang, Y.; Zhang, Z.; Zheng, X.; Wang, H.; Zheng, H.T.; Yu, P.S.; Huang, F.; et al. Benchmarking multimodal retrieval augmented generation with dynamic vqa dataset and self-adaptive planning agent. arXiv 2024, arXiv:2411.02937. [Google Scholar] [CrossRef]
  45. Yu, X.; Wu, Z.; Zhang, L.; Zhang, J.; Lyu, Y.; Zhu, D. Cp-clip: Core-periphery feature alignment clip for zero-shot medical image analysis. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Marrakesh, Morocco, 6–10 October 2024; Springer: Berlin/Heidelberg, Germany, 2024; pp. 88–97. [Google Scholar]
  46. Yan, Q.; Duan, J.; Wang, J. Multi-modal Concept Alignment Pre-training for Generative Medical Visual Question Answering. In Proceedings of the Findings of the Association for Computational Linguistics: ACL 2024, Bangkok, Thailand, 11–16 August 2024; Ku, L.W., Martins, A., Srikumar, V., Eds.; Association for Computational Linguistics: Bangkok, Thailand, 2024; pp. 5378–5389. [Google Scholar] [CrossRef]
  47. Zou, Y.; Yin, Z. Alignment, Mining and Fusion: Representation Alignment with Hard Negative Mining and Selective Knowledge Fusion for Medical Visual Question Answering. In Proceedings of the Computer Vision and Pattern Recognition Conference, Chongqing, China, 25–26 October 2025; pp. 29623–29633. [Google Scholar]
  48. Hou, Q.; Cheng, S.; Cao, P.; Yang, J.; Liu, X.; Tham, Y.C.; Zaiane, O.R. A clinical-oriented multi-level contrastive learning method for disease diagnosis in low-quality medical images. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Marrakesh, Morocco, 6–10 October 2024; Springer: Berlin/Heidelberg, Germany, 2024; pp. 13–23. [Google Scholar]
  49. Bie, Y.; Luo, L.; Chen, Z.; Chen, H. Xcoop: Explainable prompt learning for computer-aided diagnosis via concept-guided context optimization. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Marrakesh, Morocco, 6–10 October 2024; Springer: Berlin/Heidelberg, Germany, 2024; pp. 773–783. [Google Scholar]
  50. Venkataramanan, S.; Kijak, E.; Amsaleg, L.; Avrithis, Y. Alignmixup: Improving representations by interpolating aligned features. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 19174–19183. [Google Scholar]
  51. Zhang, S.; Xu, Y.; Usuyama, N.; Xu, H.; Bagga, J.; Tinn, R.; Preston, S.; Rao, R.; Wei, M.; Valluri, N.; et al. Biomedclip: A multimodal biomedical foundation model pretrained from fifteen million scientific image-text pairs. arXiv 2023, arXiv:2303.00915. [Google Scholar]
  52. Chung, H.W.; Hou, L.; Longpre, S.; Zoph, B.; Tay, Y.; Fedus, W.; Li, Y.; Wang, X.; Dehghani, M.; Brahma, S.; et al. Scaling instruction-finetuned language models. J. Mach. Learn. Res. 2024, 25, 1–53. [Google Scholar]
  53. Rückert, J.; Bloch, L.; Brüngel, R.; Idrissi-Yaghir, A.; Schäfer, H.; Schmidt, C.S.; Koitka, S.; Pelka, O.; Abacha, A.B.; Seco de Herrera, A.G.; et al. Rocov2: Radiology objects in context version 2, an updated multimodal image dataset. Sci. Data 2024, 11, 688. [Google Scholar] [CrossRef]
  54. Eslami, S.; Meinel, C.; De Melo, G. Pubmedclip: How much does clip benefit visual question answering in the medical domain? In Proceedings of the Findings of the Association for Computational Linguistics: EACL 2023, Dubrovnik, Croatia, 2–6 May 2023; pp. 1181–1193. [Google Scholar]
  55. Li, P.; Liu, G.; Tan, L.; Liao, J.; Zhong, S. Self-supervised vision-language pretraining for medial visual question answering. In Proceedings of the 2023 IEEE 20th International Symposium on Biomedical Imaging (ISBI), Cartagena de Indias, Colombia, 18–21 April 2023; pp. 1–5. [Google Scholar]
  56. Zhang, X.; Wu, C.; Zhao, Z.; Lin, W.; Zhang, Y.; Wang, Y.; Xie, W. Pmc-vqa: Visual instruction tuning for medical visual question answering. arXiv 2023, arXiv:2305.10415. [Google Scholar] [CrossRef]
  57. Li, P.; Liu, G.; He, J.; Zhao, Z.; Zhong, S. Masked vision and language pre-training with unimodal and multimodal contrastive losses for medical visual question answering. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Vancouver, BC, Canada, 8–12 October 2023; Springer: Berlin/Heidelberg, Germany, 2023; pp. 374–383. [Google Scholar]
  58. He, J.; Li, P.; Liu, G.; Zhao, Z.; Zhong, S. Pefomed: Parameter efficient fine-tuning on multimodal large language models for medical visual question answering. arXiv 2024, arXiv:2401.02797. [Google Scholar]
  59. Gu, T.; Yang, K.; Liu, D.; Cai, W. Lapa: Latent prompt assist model for medical visual question answering. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 17–21 June 2024; pp. 4971–4980. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.