1. Introduction
Healthcare is one of the critical sectors that significantly influences an individual’s well-being and quality of life, necessitating novel research, therapies, and guidelines [
1]. Artificial Intelligence (AI)-driven applications play a crucial role in transforming healthcare services across the globe [
2]. Natural language processing (NLP) is the foundation of the medical-question-answering system (MQAS) [
2,
3,
4]. It employs various techniques, including tokenization, part-of-speech tagging, named entity recognition, and syntactic parsing, to provide reliable medical information in response to a user query. MQAS supports patients and healthcare professionals in retrieving personalized responses [
5]. Medical queries are complex, requiring a deep understanding of medical terminologies, disease relationships, and treatment methods [
5]. The MQAS may face challenges in adequately addressing complex queries. The limited domain knowledge in existing MQASs limits their performance, reliability, and effectiveness in real-time settings. Inadequate or insufficient domain knowledge may influence an MQAS to provide inaccurate or misleading user responses [
5]. Consequently, the MQAS may generate inappropriate recommendations on a medical condition, compromising patient safety and worsening health outcomes.
The introduction of transformers has revolutionized the development of MQASs [
6]. Transformers leverage self-attention mechanisms to handle complex data. The parallelized processing feature enables the transformers to process long text sequences simultaneously [
7]. It reduces the training time of transformers in learning complex patterns from extensive datasets, leading to improved performance. There is significant attention on large language models (LLMs) for analyzing and generating medical responses to a user query [
8]. Transformers determine the significance of each term in a user query, enabling the development of practical NLP applications [
9]. Transformers, including the Generative Pre-trained Transformer (GPT), Bidirectional Encoder Representations from Transformers (BERT), and T5, have demonstrated remarkable capabilities in addressing the limitations of traditional QASs [
9]. For instance, ChatGPT (GPT-4) generates human-like text responses for a user query. However, these models lack the specialized domain knowledge to develop an appropriate response to a medical query [
10,
11]. An MQAS with static knowledge may not respond rapidly to these changes, preventing users from accessing the latest medical advances. The lack of adaptability may affect the timely and accurate responses, which are crucial for effective patient care [
12]. A large number of existing QASs rely on statistical models or superficial text matching to offer responses. These systems may overlook significant information or context in medical queries. Domain-specific knowledge, including medical ontologies and structured knowledge graphs, is crucial for a QAS to comprehend the complex relationships between disease, symptoms, and treatment [
13,
14,
15]. The lack of interpretability in offering insights into the QAS responses may reduce trust and confidence, which are essential in the healthcare environment [
15].
Advanced techniques, including knowledge-aware networks (KANs), synthetic datasets, and linked open data (LOD), can overcome the limitations of existing transformers and improve the MQAS response accuracy [
16]. The integration of LOD is essential. With the assistance of LOD, an MQAS can access a frequently updated source of medical information, enabling it to produce reliable responses to user queries [
16]. It can guarantee that its responses reflect the latest advances in medical knowledge by establishing access to several databases, scholarly articles, and medical guidelines. By integrating context and background information, LOD improves the system’s ability to respond to complicated queries [
17].
LLMs transform NLP techniques, leading to the development of highly effective MQASs [
18]. These models utilize massive amounts of textual data to provide contextually appropriate and syntactically reliable responses. Using subtle language patterns, LLMs can translate complicated medical queries and generate context-aware responses [
18]. Retrieval-augmented generation (RAG) enhances LLM-based systems with an additional layer of functionality [
18]. RAG supplements these models with external information to guarantee syntactically and factually correct responses. Real-time access to the recent medical literature improves the trustworthiness of MQAS responses [
18,
19]. RAG-enhanced LLMs can generate linguistically accurate and medically sound responses in bilingual QAS systems, providing solutions for low-resource languages, such as Arabic [
19]. The integration of LLMs and RAG allows bilingual MQASs to generate medical information [
19]. User satisfaction and trust can be improved by providing accurate, relevant, and personalized responses [
20]. Recent studies highlighted the performance of LLMs in answering medical queries across various domains and cultural contexts. Chen et al. [
21] evaluate ChatGPT and Bard on medical licensing examinations across diverse regions, indicating significant variations in accuracy and reasoning depth. Their findings indicate linguistic nuance and contextual bias as persistent challenges in multilingual medical examination. Chau et al. [
22] examine the performance of generative AI models in dental licensing examinations. The generative AI models struggle with clinically applied reasoning and terminology specific to the regional dental curriculum. These studies exhibit the importance of culturally adapted, domain-specific datasets and evaluation protocols for developing MQASs.
The development of bilingual MQASs holds great potential to improve global healthcare accessibility [
23]. In a bilingual setting, English-based medical QASs may not elicit meaningful responses from non-English-speaking users [
24]. In Middle Eastern countries, English-based MQASs are ineffective in meeting the demands of healthcare professionals and patients who primarily communicate in Arabic [
25]. The limited accessibility and understanding of medical responses may affect Arabic-speaking individuals. English-based medical QASs face challenges in handling cultural and linguistic differences, resulting in the generation of contextually inaccurate responses [
26]. The critical medical insights are associated with regional practices and health concerns [
25,
26]. As a result, English-based MQASs may not provide adequate services to Arabic-speaking individuals. These limitations underscore the need to develop a robust, contextually aware, and interpretable bilingual medical QAS, providing trustworthy, accurate, and culturally appropriate responses. Thus, we develop an interpretable bilingual MQAS to accurately and effectively respond to user queries. The study’s novelty lies in its synergistic integration of generative, discriminative, and knowledge-aware paradigms. The proposed model establishes a retrieval-conditioned and knowledge-grounded reasoning through the harmonization of GPT-Neo’s linguistic fluency and RoBERTa’s factual discrimination. KANs-mediated semantic alignment and RAG-based evidence retrieval improve the model’s ability to deliver reliable bilingual medical responses. The proposed approach enables intermediate and trustworthy cross-lingual medical information access, reducing hallucination and promoting culturally aligned, evidence-based responsegned, evidence-based repsoting culturally aligned, evidence-based repsoion access, ccess to medical informations. Furthermore, the SHAP-based explainability supports clinical auditability. The study contributions are listed as follows:
We employ query optimization techniques, including semantic parsing, language detection, and query reformation, to retrieve the most relevant and contextually appropriate information.
- 2.
Domain-specific data enrichment using RAG.
We integrate the GPT-Neo and RoBERTa models into a single framework to generate responses in Arabic and English. RAG enhances the model’s understanding of complex medical terminology and contextual relationships. It enhances the bilingual performance of the GPT-Neo and RoBERTs models.
- 3.
Ensemble-learning-based bilingual MQAS.
We use the weighted majority voting approach to ensemble the responses of the GPT-Neo and RoBERTa models. They enhance the model’s performance by integrating KANs and SHAP values. This improvement enables the proposed model to present a foundation for interpretable and trustworthy AI in healthcare.
The study is organized as follows:
Section 2 presents the features and limitations of the existing MQAS. The proposed methodology for developing the bilingual MQAS is outlined in
Section 3.
Section 4 highlights the findings of the experimental validation. The study’s implications and limitations are presented in
Section 5.
Section 6 concludes the study with its future directions.
2. Related Works
Many technological advancements have been established in MQASs, transforming the healthcare domain. Traditional MQASs, including rule-based systems, employ rules and patterns to respond to user inquiries [
27]. Medical ontologies are utilized to retrieve appropriate information. These systems are effective for straightforward and well-structured queries. However, they struggle with the complexities of natural language, ambiguity, and medical queries. In addition, rule-based systems require continuous manual updates to incorporate current healthcare data, which can lower their performance in real-time settings [
27]. Recently, machine learning techniques have been widely applied in healthcare to provide reliable services to individuals. These techniques generate responses using a supervised learning approach based on labeled medical datasets. Machine learning models outperformed rule-based systems in handling complex and diverse queries due to their larger datasets and access to high computational resources. Despite these advancements, recent MQASs encounter challenges in comprehending context and addressing queries that comprise multiple levels of interpretation.
Yasunaga et al. (2021) [
28] propose a GNN-based QAS. They address the limitations of traditional GNN models with relevance scoring and joint reasoning strategies. By leveraging RoBERTa and structured graph nodes, the model encodes text. An iterative message passing mechanism supports it to learn linguistic and relational representations. The findings of the performance evaluation (accuracy: 82.14%) reveal the importance of graph-based reasoning in generating meaningful responses. Zhang et al. (2022) [
29] employ a graphical neural network (GNN) for QAS development, achieving an accuracy of 84.7%. A cross-modal fusion component provides information from the knowledge graph to the language representation. The transformer is employed to update tokens and node embedding. An integrated GNN and transformer improve the model’s contextual understanding. The concept of co-training is applied to enable graph context to inform token-level embedding. The integration of joint gradient updates assists the model to outperform graph–text fusion systems. This architecture motivates the introduction of the proposed KANs module, bridging LOD with RAG-retrieved textual evidence. Alzubi et al. (2023) [
30] develop a COVID-19 QAS using BERT. This model obtained an overall accuracy of 58.91%. A retriever–reader dual algorithmic system is used to answer the complex queries. A TF-IDF vectorizer is used to capture the top 15 documents with optimal scores. A preprocessing technique is applied to fine-tune the model’s performance. Attention visualization is used for the model’s interpretability. The application of contextual embedding supports the model in achieving an F1-score of 0.89. However, the training and inference were limited to the English language.
Singhal et al. (2025) [
31] proposed an LLMs-based MQAS. Instruction tuning and reinforcement learning from human feedback are used to align outputs with professional medical standards. The findings demonstrate that the MultiMedQA benchmark shows near-clinician accuracy and improved factual grounding. The model is generalized on multiple clinical topic datasets. Wang et al. (2023) [
32] introduce augmenting black-box LLMs with medical textbooks in order to generate a response. The retrieval-augmented pipeline constructs a dense retriever of textbook passages using user queries. Using this strategy, the factual grounding of this model is improved, achieving a significant improvement in responding to a query. The model was evaluated on three open-domain datasets. However, the model is monolingual and non-interpretable. Van Sonsbeek et al. (2023) [
33] propose a medical visual QAS using pre-trained language models, achieving an accuracy of 84.3%. They map the visual features to a set of learnable tokens. Using a frozen GPT-2 decoder, image embeddings are fused with text prompts. The parameter-efficient fine-tuning strategy optimized the model’s parameters to achieve an optimal outcome. The study’s findings emphasize multimodal reasoning that leads us to introduce bilingual reasoning through pre-trained models. Lin et al. (2023) [
34] build a visual MQAS using a pre-trained vision encoder. They conduct internal and external validations using public repositories. They categorize MQAS into image–text attention, graph-fusion transformers, and multimodal LLMs. The challenges in interpretability and dataset imbalance are highlighted, underscoring the importance of enriching the dataset with structured medical information. Similarly, Kamel et al. (2023) [
35] develop a visual Arabic QAS using LLM techniques, achieving a lower accuracy of 57.62%. They use a question-tokenization approach and three-word embedding algorithms. Transformer backbones are integrated with the Arabic-BERT encoder. However, it lacks cross-lingual grounding or retrieval augmentation.
Sengupta et al. (2023) [
36] build an MQAS using the GPT-2 model. They enhance the model’s performance using a joint artificial intelligence system (jais) tokenizer with Adam optimization. Morphological richness and right-to-left tokenization support the model to produce reliable responses. However, the model obtains an accuracy of 28.8%. Compared to the existing multilingual baseline models, the Arabic-centric model demonstrates significant improvements. However, the model remains general-purpose. Jiang et al. (2023) [
37] introduce a model, Mistral 7B, using sliding windows and grouped-query attention. The grouped-query attention and sliding window mechanisms are used to reduce inference latency. This model lacks domain adaptation and explainability, restricting its application to specific domains (accuracy: 40.6%). Zeming et al. (2023) [
38] propose an MQAS using NVIDIA’s Megatron-LM distributed trainer. Multi-stage instruction tuning is employed to acquire clinical reasoning. A set of 4.6 trillion tokens from PubMed and clinical guidelines is used to train the model, achieving an accuracy of 60.7%. Yu et al. (2024) [
39] proposed a QAS model for the healthcare domain. They generated a dataset [
40] to train and validate the QAS model. Multiple transformers were used to generate an answer. The models, including Sentence-T5 and Mistral 7B, are integrated to enhance the model’s generative capabilities (precision of 76.2%). The synthetic minority over-sampling technique is employed to handle class imbalance. Preri et al. (2024) [
41] introduce a bilingual medical QAS. A bilingual medical mixture of expert LLMs is used to respond to user queries with an accuracy of 47.3%. Bilingual corpora and domain alignment are used to train the model. However, there is a lack of approaches to address limited interpretability and hallucination. The parameter-efficient fine-tuning (PEFT) technique is used to optimize the hyperparameters. Wu et al. (2024) [
42] develop an MQAS using a general-purpose foundation language model, achieving an accuracy of 47.2%. The factual grounding and domain understanding are improved through the fine-tuning process using the PubMed corpora.
Table 1 outlines the features and limitations of the existing models.
Visual QAS encounters challenges in managing the complex interdependencies between visual and textual medical information. It relies on image–text correlation. The lack of a comprehensive understanding of text-based context may lead to inaccurate responses. The limited Arabic visual datasets reduce the performance of visual QAS. GNN faces challenges in capturing long-range contextual information. Due to the limitations in handling language-specific requirements, GNN-based QASs may not produce better performance. Open-domain QASs retrieve and synthesize data from multiple sources in order to provide a proper response. They may misinterpret user intent, leading to irrelevant outcomes. The lack of reliability in the numerous sources may result in biased or incorrect responses. Multiple-choice QASs depend on multiple-choice patterns. They retrieve long or comprehensive responses rather than understanding the context. A significant limitation of multiple-choice QASs is the existence of distractor alternatives, misleading users with inappropriate responses. The existing models have difficulties in interpreting complex queries with limited resources. These limitations motivate us to develop a contextually grounded knowledge architecture to address the limitations of existing MQASs, handling the complexities in bilingual queries. Employing ensemble learning techniques with advanced evaluation metrics supports the model’s reliability and interpretability. The combination of high-performance NLP with domain-aware reasoning overcomes the shortcomings of existing MQASs. This strategy promotes future scalability across languages and specialties, leading to advancements in patient-centered healthcare technologies.
3. Materials and Methods
GPT-Neo can generate fluent and contextually rich responses for open-ended medical queries [
43]. RoBERTa excels in handling the structure and context of a query [
44]. It is commonly applied for query classification, context extraction, and ranking. The development of the proposed MQAS is motivated by the complementary strengths of GPT-Neo and RoBERTa in processing user queries and generating optimal medical responses in English and Arabic. Fine-tuning and facilitating external knowledge through RAG can improve the ability of these transformers to generate contextually accurate responses for bilingual queries. To improve the proposed model’s robustness and generalization, we employ an ensemble learning approach. The approach ensembles the outcomes of GPT-Neo and RoBERTa using a voting approach. In addition, it applies SHAP values to facilitate the model interpretability. In addition, value-aware quantization is used to minimize the computational resources [
45].
Figure 1 shows the proposed approach for developing bilingual MQAS. It highlights the process of integrating GPT-Neo and RoBERTa within a unified hybrid framework, leveraging the complementary strengths of generative and discriminative modeling. GPT-Neo functions as the generative reasoning engine, while RoBERTa serves as the discriminative verifier. Using this approach, factuality and linguistic coherence are evaluated. A user query is optimized and enriched using KANs and LOD, yielding a semantically expanded query representation. The introduction of a synthetic enrichment layer augments the training corpus with diverse, medically valid QA pairs in Arabic and English. The enrichment addresses data sparsity in underrepresented categories, improving the model’s generalization. RAG + T5 generates Arabic and English QA pairs from biomedical corpora and verified knowledge sources, producing linguistically natural and semantically diverse QA pairs. The enriched corpus is used to train the model. During inference, RAG retrieves top-k evidence documents corresponding to the user query, guiding GPT-Neo’s generation of candidate responses. These outputs are re-ranked by the RoBERTa model. The final response is selected using a weighted ensemble approach. SHAP-based interpretability reveals the linguistic and biomedical features that influence the model’s decision.
3.1. Data Acquisition
We utilize four public datasets in order to train and test the proposed model. The datasets used in this study are carefully selected and curated from reputable public repositories and trusted medical sources. Dataset 1 is an Arabic healthcare dataset containing 808,000 questions and answers across 90 categories. It offers a substantial sample size to capture the linguistic and conceptual features required for the model to understand the complex terms in Arabic. It is available in a public repository [
46]. Dataset 2 [
47] contains 87,930 Arabic medical questions and answers. It is divided into training, validation, and test sets. It enables structured performance assessment. Dataset 3 [
40] includes 247,000 English QA pairs collected from medical experts. The high clinical reliability and coverage of domain-specific terminology enhance the model’s interpretability. Dataset 4 [
48] is a specialized dataset covering 47,000 English questions and answers across 12 categories. It is collected from trusted medical websites. It provides high-quality and reliable medical questions and answers. It supports the proposed model to deliver reliable responses. We utilize datasets 1 and 3 for synthetic data generation. We split the datasets into a training set (80%) and a test set (20%). The large dataset sizes provide a robust foundation for training GPT-Neo and RoBERTa in enriching datasets 1 and 3. The balanced coverage across medical categories improves the proposed model’s stability, bilingual capability, and medical relevance. The dataset providers have indicated that the data were acquired through appropriate methods, including informed consent from the participants. They have anonymized the datasets, protecting sensitive user data privacy. As a result, Institutional Review Board approval or additional ethical clearance are not essential for this study. The use of public repositories is in line with the ethical and legal frameworks defined by the original dataset providers.
For the purpose of assessing the model’s generalization and preventing overfitting, we utilize 20% of datasets 2 and 4. This strategy determines the model’s generalization ability on unseen data. We limit the evaluation set to 20% to ensure statistically meaningful generalization testing. The 20% test ratio provides sufficient statistical power (95% confidence) for F1-score, bilingual evaluation understudy (BLEU), and recall-oriented understudy for gisting evaluation (ROUGE), reflecting a trade-off between representativeness and model’s generalization efficiency. In addition, it aligns with established NLP evaluation standards and ensures that Arabic and English models are tested on unseen and balanced subsets. A stratified random sampling technique is used for the selection process. Using this technique, the original class and category distribution are preserved. In order to guarantee extensive and comprehensive coverage of a wide range of medical conditions, each subset is designed to reflect a proportionate representation of all question categories. This strategy enables a fair comparison of Arabic-only, English-only, and combined bilingual MQAS models. It provides a dependable and representative foundation for quantitative and qualitative performance evaluation.
3.2. Synthetic Dataset
To improve the performance of GPT-Neo and RoBERTa, data enrichment is essential. RAG enriches the train set (Datasets 1 and 3) to enhance the performance of the proposed model. It involves a structured approach that integrates the retrieval and generating processes. The knowledge bases, including PubMed [
49] for English and e-Marefa [
50] for Arabic, are used in this study. PubMed is a comprehensive repository of the biomedical literature. It provides rich contextual information for English queries. e-Marefa is a specialized resource for Arabic terms. A dense passage retrieval system is used to index the two knowledge bases. Algorithm 1 outlines the procedure for generating the synthetic QA (enriched datasets 1 and 3) dataset. For instance, the query “What is the treatment for hypertension?”, retrieves relevant documents from PubMed, identifies the entity “hypertension” through KANs, and maps it to DBPedia and MeSH. The T5 model [
51] generates a response such as “Antihypertensive therapy with inhibitors or lifestyle modification”. Furthermore, the bilingual experts validate the generated QA pairs before including them in the final dataset.
| Algorithm 1: Synthetic dataset generation |
Input: Query Q (Arabic or English) Output: Synthetic QA pair (Q, A_syn) 1. Language Detection: if Q ∈ Arabic → L ← “AR” else L ← “EN” 2. Template Generation: Retrieve relevant question templates T_L based on query intent (diagnosis, treatment, prevention) 3. Document Retrieval (RAG step): if L = "EN" → Search PubMed/Medline if L = “AR” → Search e-Marefa/AraBase Retrieve top-k documents D = {d1, d2, …, dk} 4. Entity and Context Enrichment: Use KAN to extract medical entities E from D Map E ↔ Linked Open Data (DBpedia, MeSH, AraWiki) Append semantically aligned terms to query context 5. Synthetic Response Generation: Feed (Q + enriched context) into fine-tuned T5 model A_syn ← T5(Q, context) 6. Expert Validation: Two bilingual medical experts review A_syn if disagreement → resolve via consensus (third expert) 7. Finalization: Store (Q, A_syn) into bilingual QA dataset |
We developed question templates covering medical queries in Arabic and English. The templates, including “What is [X]?”, “How does [Y] work?”, “
ما هو [X]?” (What is [X]?) or “
كيف يعمل [Y]?” (How does [Y] work?) are employed. These templates offer a fundamental structure for developing synthetic queries associated with the medical domain. By generating a diverse question–answer pair across multiple medical domains, the proposed synthetic dataset generation improves the MQAS’s robustness. We follow the common structure of question templates in existing templates, frequently asked questions in health organizations, and user forums. They categorize the templates based on medical specialties, enabling the model to reflect realistic and domain-relevant inquiries. Using publicly available ontologies, each template is instantiated with medical terms, conditions, treatments, and symptoms. In order to maintain linguistic balance, we pair each question with an equivalent Arabic version using the combination of expert-crafted translations and back-translation techniques. Grammatical accuracy and cultural relevance are maintained for each template to ensure that the synthetically generated questions mimic natural linguistic patterns. This strategy enhances the model’s generalizability and guarantees fair performance across diverse user queries. User queries are standardized using these templates, facilitating management and categorization of the generated data. By leveraging these approaches, we employ RAG to retrieve relevant documents for each template using PubMed and e-Marefa repositories. In addition, a retrieval model is used to analyze the templates and rank documents based on semantic similarity to the query. Equation (1) shows the mathematical expression of the retrieval process.
where
is the top K documents retrieved from the knowledge base (KB) for a query (q).
In order to generate responses from the retrieved top k documents, we employ the T5 [
51] transformer. The potential of the T5 generator synthesizes coherent and contextually appropriate responses. Equation (2) presents the response generation for the Arabic and English queries.
where R is a response, and
represents that the generation of the response (r) is conditioned on the query (q) and the retrieved documents (
).
A language detection module is used to identify English and Arabic queries. We use a lightweight model, FastTest [
52], to detect the query’s language. The model uses character n-grams and text embeddings for language identification. Based on the query, the system utilizes either the PubMed or e-Marefa knowledge bases. In the context of bilingual MQAS development, the quality, relevance, and authenticity of training data are crucial.
To ensure consistency and reliability in the expert validation phase, a structured calibration and an inter-annotator agreement assessment protocol is implemented. The calibration phase comprises two bilingual experts and a medical language specialist (mediator). Two medical experts are actively engaged to ensure the clinical and linguistic validity of the synthetic dataset. They guide the process of answer pair generation, reflecting common real-world patient queries, terminology, and phrasing that align with clinical practice. Their bilingual medical expertise supports us in verifying cultural sensitivities and local healthcare practices, strengthening the dataset’s applicability to Arabic-speaking populations. Using the calibration phase, an initial random sample of 500 Arabic and 500 English QA pairs is reviewed. Calibration outcomes are consolidated into a shared annotation guide, ensuring clinical appropriateness and linguistic clarity. The inter-annotator agreement is quantified through Cohen’s kappa (k) for pairwise evaluations. As per Landis and Koch’s interpretation scale, a kappa coefficient of 0.86 is considered perfect agreement. A structured consensus-building approach is employed in order to address the disagreements or differences between the two experts. We isolate the items related to disagreements for discussion. Each expert is requested to independently annotate the rationale behind his or her recommendations. We facilitate a moderated session with a third expert to review the recommendations, enabling transparent discussion based on clinical standards and mutual understanding. In addition, a resolution protocol is applied to cross-validate the disputed items against medical guidelines. Using this multi-layered review protocol, the credibility and trustworthiness of the final synthetic dataset is strengthened.
Using authoritative medical repositories and consensus discussion, the resolution protocol ensures the clinical fidelity and linguistic equivalence of the bilingual QA pairs. To guarantee cross-lingual consistency between Arabic and English QA pairs, we employ a back-translation validation protocol using the MarianMT and mBART50 models. The original and translated versions are compared using cosine similarity over sentence embedding and reviewed by bilingual experts. Using this process, we ensure the preservation of linguistic nuances, idiomatic expressions, and medical terminologies, preventing potential misalignment between Arabic and English QA pairs during model training and evaluation. With the systematic validation, the credibility of the synthetic dataset is strengthened, and the reproducibility of the proposed MQAS pipeline is supported.
Normalization and control strategies are used to ensure the reliability of the cross-dataset comparisons. Datasets are standardized to comparable topic distributions, covering overlapping medical domains and preventive medicine. We exclude the queries with domain mismatches or ambiguous semantic categories through expert validation, maintaining a consistent conceptual scope. The sampling strategy of datasets 1 and 3 for evaluation guarantees the representation of similar semantic and complexity levels within its linguistic domains. The quantitative comparisons are interpreted within datasets after normalizing scoring metrics, reflecting genuine model-level generalization rather than dataset-driven variability. The semantic parallelism and controlled sampling of the dataset indicate that the datasets are sufficiently comparable for bilingual evaluation.
In
Table 2, the findings of the expert validation are outlined. Although the primary training set (datasets 1 and 3) collectively contains over one million QA pairs, only 49,000 English and 48,000 Arabic QA pairs are utilized for training, ensuring data quality, bilingual balance, and medical relevance. This approach prevents redundancy and linguistic noise present in the raw corpora, prioritizing data quality over quantity. The low modification ratio of 14.8% and high Cohen’s k of 0.86 indicate the quality of the synthetically generated QA pairs and reliability between independent annotators.
Table 3 outlines the importance of features of the synthetic datasets, contributing to the model’s outcomes. The SHAP value represents a numeric value indicating the role of each feature in generated responses. It provides a positive or negative numeric value. A positive values indicate the positive influence of a feature, whereas a negative value highlights the feature’s negative contribution.
3.3. Query Optimization
The query optimization process improves the accuracy and relevance of the proposed bilingual MQAS through multiple preprocessing techniques. The techniques, including tokenization, stemming, stop-word removal, and lemmatization, are used to optimize the queries. Equation (3) shows the mathematical form of query optimization.
where
is the actual query, and
is the optimized query.
Furthermore, we employ KANs and LOD to enhance the query. The query enhancement allows users to use prompts and complex queries to retrieve a response. KANs can interpret the user query by identifying their relationship without explicit named entity recognition. It can map query terms to entities in the knowledge bases. LOD is a repository of linked datasets. It facilitates data interoperability by integrating diverse datasets. We apply synonym enrichment using KANs to expand the medical terms in the query. The enrichment allows the proposed model to capture broader information. Equation (4) highlights the query enrichment process.
where
is the enriched query, and
is the set of synonyms for term (x).
The integration of LOD and KANs enhances the proposed MQAS’s ability to interpret and respond to user queries. For instance, a query related to “diabetes” can be enriched with related medical terms, “insulin resistance” or “glucose monitoring”. Equation (5) outlines the query enrichment process using KANs and LOD. We employ DBPedia for English and AraWiki for Arabic queries, respectively.
where
is the set of related terms from the LOD sources, and KAN is the KANs model for the comprehensive interpretation of the query.
To illustrate the query optimization process, consider a query, “What are the treatment options for controlling high blood sugar?”. The preprocessing step transforms the query into a standardized query representation: treat, option, control, blood, and sugar. Through semantic mapping, KANs identify contextual equivalents such as “therapy”, “management”, and “medication” for the term “treat” and associate “glucose” with “insulin” and “blood sugar”. The enriched query terms, i.e., treat, therapy, management, glucose, insulin, diabetes, and medication, are extended through LOD, resulting in the final optimized query terms: treat, therapy, diabetes, insulin, metformin, HbA1C, and insulin resistance. This richer semantic structure enhances the retrieval process within the RAG framework.
3.4. GPT-Neo and RoBERTa-Based Response Generation
In this section, we present the methodologies for response generation using the enhanced GPT-Neo and RoBERTa models. They highlight the recommended approach to enhance the overall effectiveness and reliability of the proposed MQAS in generating appropriate responses for English and Arabic queries.
GPT-Neo is an open-source alternative to OpenAI’s GPT-3. It has been trained on diverse datasets to comprehend and compose cohesive descriptions, conversations, and generative human-like narratives on a variety of themes. It is widely used for developing chatbots and producing content. It guarantees smooth interactions by handling multi-turn discussions. However, it may generate factually incorrect or nonsensical information and lacks fine-grained control over the generated results. The unique grammatical structure may limit GPT-Neo’s capability in tokenizing Arabic text. These limitations restrict the potential of GPT-Neo to produce precise outcomes for context-sensitive information.
RoBERTa is an enhanced version of the BERT model. It is used to understand and analyze the specific text. It demonstrates reliable performance on a wide range of NLP tasks, including text classification, sentiment analysis, and named entity recognition. A masked language modeling approach is used to capture intricate word–context relationships. RoBERTa demands high computational costs during training and inference. The effectiveness of the RoBERTa model depends on the selection of hyperparameters.
To overcome the limitations of GPT-Neo and RoBERTa, we employ domain-specific data (synthetic datasets (enriched datasets 1 and 3)), confidence-based dynamic weighting, and fine-tuning using BOHB optimization. Using these strategies, we improve the generative capability of GPT-Neo and the context understanding of RoBERTa. The objective of training the GPT-Neo and RoBERTa models is to reduce the loss function for the synthetic dataset.
Let
be the synthetic dataset, where s is a sequence of tokens. Let
be the corresponding target tokens for the individual sequence. GPT-Neo and RoBERTa predict the subsequent token in a sequence. These models compute the probability distribution, P
, for each token, (
, where
is the model’s parameters. Equation (6) presents the computation of a SoftMax function for each token over the logits (
where
is the vocabulary, and
is the output of the model prior to the SoftMax for the token (
).
A cross-entropy loss function L for GPT-Neo and RoBERTa is used to measure the difference between the predicted and the true classes. Equation (7) shows the mathematical form of the loss function.
where N is the number of input–output pairs, and
is the predicted probability for the true class (
Equation (8) outlines the back propagation and optimization processes through the computation of gradients of the loss function.
where
is the model’s parameter at iteration t,
is the learning rate, and
is the gradients of the loss function.
3.5. BOHB-Based Hyperparameter Optimization
In order to fine-tune the parameters of the transformers, we employ the BOHB algorithm. Compared to conventional techniques, such as grid search, random search, and standalone Bayesian optimization, BOHB offers superior efficiency, scalability, and adaptability in large, high-dimensional search spaces. Bayesian models guide the search operation, and HyperBand discards underperforming configurations without exhaustively exploring every possibility. During the fine-tuning process using the synthetic bilingual datasets, BOHB optimizes the hyperparameters. The parameters, including learning rate, batch size, number of epochs, and weight decay, are optimized to achieve exceptional accuracy. Additionally, BOHB manages the knowledge-injection parameters used in the KANs module, including the degree of graph attention and knowledge integration weights. These benefits render BOHB a practical and strategically superior choice for complex, bilingual, and knowledge-enriched NLP applications.
GPT-Neo and RoBERTa cover a traditional gradient-based optimization. However, the recommended BOHB guides the fine-tuning process, leading to obtaining accurate outcomes with limited computational resources. Equation (9) outlines the fine-tuning process using the synthetic bilingual dataset. The optimizations are initialized by balancing exploration and exploitation.
where
is the hyperparameter space,
represents the expectation over the loss function (
), and
is the optimal set of hyperparameters. A random selection of hyperparameters with multiple trials is used to identify the key parameters. We apply quantization to reduce computational costs. This maintains the trade-off between memory usage and acceptable accuracy levels, making the model suitable for deployment in low-resource environments. Equation (10) outlines the quantization process.
where
is the precision level,
is the model weights, and
is the quantized model weights.
3.6. Ensemble-Learning-Based Response Generation
To facilitate the ensemble-learning-based outcome generation and model interpretability, we compute the confidence score using the SoftMax probabilities. Equations (11) and (12) present the confidence scores for predicting outputs.
where
is the logit for class C, c is the possible answer classes,
and
are the model weights for query (q), and
and
are the confidence scores for the query (q).
We employ the weighted majority voting mechanism to combine the GPT-Neo and RoBERTa models’ outcomes based on the weights and confidence scores. This approach uses the transformers’ unique strengths, contributing to producing reliable results for each query. Each model’s vote is scaled by a weight based on its performance and reliability. The confidence-based dynamic weighting mechanisms maintain the individual model’s contribution based on its performance. This enables the proposed model to ensure that models operating with lower confidence are assigned with lower weights. Equation (13) computes the outcome using the ensemble-learning approach.
where
is the final response for class C,
is the confidence score, and
is the weight associated with ith model with GPT-Neo and RoBERTa.
3.7. Performance Evaluation
We employ the key evaluation metrics to determine the proposed model’s effectiveness and reliability. This uses a top-k document retrieval strategy for the performance evaluation. This strategy focuses on selecting the top-k relevant documents. The potential of the model in understanding and matching query context can be determined using the top-k document retrieval strategy. The standard evaluation metrics, including accuracy, precision, recall, and F1-score, are used to evaluate the model’s performance. Accuracy guarantees the model’s ability to retrieve accurate responses to user queries. The precision value ensures that the model generates less incorrect medical information. Recall measures the potential of the proposed model in extracting relevant responses. F1-score provides a comprehensive view of the performance of the proposed MQAS by balancing precision and recall. In order to evaluate the bilingual response generation capability, we use the BLEU metric. BLEU evaluates the proximity of the model’s outcomes to the actual responses. ROUGE assesses the extent to which the model’s response reflects the exact response. Mean reciprocal rank (MRR) evaluates the consistency of the model in returning appropriate responses at the top of the list. In addition, we compute the average number of floating-point operations per second (FLOPs) to identify the impact of quantization on the proposed model. The statistical evaluation metrics are used to ensure the model’s reliability in responding to real-time medical queries. These metrics provide insights into the MQAS’s performance and guide the iterative process of model refinement. We identify areas of improvement, validate enhancements, and guarantee the accuracy and efficiency of the final system in the bilingual healthcare environment.
4. Results
We implement the proposed MQAS using Windows 11 Pro, Intel (Santa Clara, CA, USA) Xeon Sliver-4309Y@ 3.6 GHz, and NVIDIA (Santa Clara, CA, USA) H100 GPU. PyTorch v2.3.0 and Huggingface frameworks are used for building the T5 V1.1, GPT-Neo (v2.7B), and RoBERTa models. The libraries, including Sentencepiece v0.2.1, Elasticsearch V9.0. (for document retrieval using RAG), NumPy v2.3.4, SpaCy V2.0, CAMeL tools v1.5.7, PyArabic v0.6.15, and Pandas v2.3.1, are used in the model implementation.
Table 4 outlines the computational configurations for developing the proposed MQAS.
Figure 2 offers insights into the training phase using the enriched training sets (datasets 1 and 3). The progressive improvement in the training accuracy indicates the model’s ability to capture the intricate patterns in the dataset. The decline in the training loss shows the error minimization in the model’s prediction. It indicates the convergence of the proposed model, suggesting an optimal point to end the training and avoid overfitting. The exceptional performance during the training phase reflects the effectiveness of the model in handling the complexities in the Arabic and English datasets, leading to better performance on novel data.
Figure 3 highlights the model’s performance on the test set of dataset 1. It illustrates the trade-offs between accuracy and recall at varying document retrieval depths. For instance, at K = 5, the model achieves the highest accuracy of 94.5% with a slightly lower recall of 91.9%. This indicates that the proposed model produced precise predictions on fewer documents. The outcomes for top K = 10 and K = 20 reveal that the model maintained high accuracy with enhanced recall and F1-scores, achieving comprehensive and reliable answers.
Figure 4 reveals the exceptional outcome of the proposed MQAS on the test set of dataset 3. It demonstrates the ability of the model to prioritize relevant documents and generate a precise outcome for the user. The model’s robust design, comprising advanced query optimization, RAG-T5 integration, and a fine-tuning approach, enhanced its understanding of the context of a user query. The findings of the top-k strategy highlight the model’s attention on small and high-quality subsets of data. It indicates that the model retrieved and ranked the relevant documents. In addition, the weighted majority voting approach supported the model in delivering an optimal response. The integration of KANs and LOD enables the model to comprehend the complex query and retrieve meaningful response.
Table 5 outlines the findings of generalization. The model achieved a remarkable accuracy with optimal BLEU and ROUGE scores. These scores represent the model’s ability to generate responses with high linguistic quality. The outcomes suggest that the model’s architecture maintains effective generalization on datasets 2 and 4, supporting robust and quality output. The recommended BOHB-based hyperparameter optimization plays a crucial role in enhancing the model’s information retrieval capabilities. The significant improvement in recall and F1-score can be attributed to RAG’s ability to retrieve external documents to add context and depth to the generated responses.
Table 6 presents an overview of the computational efficiency of the proposed MQAS. The low inference time signifies a prompt response time, suitable for healthcare applications. The enriched datasets 1 and 3 contain semantically expanded and normalized QA pairs, enabling the model to generate contextually richer and meaningful responses to the user queries. Although the training set comprises a limited set of enriched QA pairs, the proposed model delivers strong performance on unseen data (datasets 2 and 4). The data quality, diversity, and enrichment mechanisms improve the model’s performance. For dataset 3, the proposed model achieved the lowest inference time of 1.5 s, indicating its potential to deliver reliable responses in a limited time. The token throughput remains consistent throughout the datasets, representing better processing efficiency and response time. Across the datasets, hallucination rates remained below 5% and bias variance under 3%, ensuring robustness and reliability. These values underscore the combined effect of RAG-based factual grounding, KAN-driven semantic normalization, and SHAP-guided interpretability. The proposed MQAS’s efficiency and computational feasibility in terms of latency, token throughput, and FLOPs requirements render it ideal for real-world healthcare applications.
Table 7 demonstrates the performance and robustness of the proposed model. By achieving a significant improvement of 3–5% in F1-score, the proposed model surpasses the strongest multilingual baseline models. The BLEU and ROUGE scores are improved by an average of 8–10%, indicating substantial semantic alignment and fluency in generated responses. In the English-only test, the F1-score of the proposed model increases by 4.3%, which is higher than BioBERT. The findings demonstrate consistent and significant improvements across the evaluation metrics. The proposed model outperforms biomedical-specific models such as BioBERT and ClinicalBERT. The proposed hybrid integration of RAG-enhanced retrieval, KANs-based semantic expansion, and BOHB-optimized ensemble weighting enables the model to balance contextual precision and language-adaptive fluency. The model achieved an F1-score of 90.00% and BLEU score of 0.810, which reflects the Arabic-only model’s ability to handle rich morphological variations. The potential of cross-lingual transfer learning of the bilingual model is confirmed by the F1-score, BLEU score, and ROUGE score of 89.62%, 0.806, and 0.632, respectively. The validation of the statistical significance through the McNemar χ
2 test reveals that the proposed MQAS’s improvements are significant at the 95% confidence level. The hallucination reduction from 12.4 (GPT-Neo) to 3.4 (English-only model) shows the reliability of the proposed model. The bias variance of 2.3% indicates the contribution of the KANs-LOD pipeline in mitigating the over-representation of high-frequency entities. The low mean uncertainty value of 0.18 affirms the generation of reliable and consistent responses.
The time complexity of the proposed model is ( where denotes the number of integrated sub-models, r is the retrieved documents, n is the number of tokens, and d is the embedding size. The inference time of 3.4 s per query incurred by the bilingual model reflects its ability to maintain the trade-off between response generation and computational efficiency. The quantitative metrics and inferential statistics reveal that the proposed model achieves statistically reliable, interpretable, and cross-lingually consistent performance. The dynamic re-scaling of model’s contribution using dropout variance and SHAP-based interpretability presents transparent reasoning, quantifying the rationale contribution of clinical, linguistic, and contextual features. The low hallucination and bias variance address the ethical concerns of misinformation propagation in automated MQAS. The bilingual capability of the proposed model allows Arabic-speaking populations to access evidence-based insights.
Table 8 highlights the uncertainty analysis of the proposed model. It outlines the key responses of the proposed model for the queries of datasets 2 and 4. Each query batch corresponds to a semantically grouped set of medical questions in order to evaluate the model’s performance across diverse contexts. These batches represent distinct medical subdomains and reasoning types. The diversity across batches reflects real-world user interactions in clinical and patient-education settings. The higher confidence score suggests reliable predictions. For instance, “type 2 diabetes”, “
ارتفاع ضغط الدم” (hypertension), “
الربو” (asthma), and “flu vaccination” predictions can be trusted without substantial verification. The low dropout and ensemble variances affirm the model’s stability. In addition, the SHAP values of +0.12 represent the influence of “age” and “risk factors” in predicting “type 2 diabetes”. These values enhance the model’s transparency, allowing the user to understand the model’s reasoning. The higher confidence score of 0.82 for “flu vaccination” indicates the effectiveness of the proposed ensemble-learning approach. The action “recommended human review” highlights the high-uncertainty predictions. It demonstrates the necessity of human review to ensure the accuracy of the response.
Figure 5 provides a comprehensive view of the global feature contributions that support the model’s interpretability. The features are categorized into three domains: clinical, linguistic, and contextual, highlighting the reasoning process of the proposed model. The highest SHAP values of the clinical features, such as age, blood pressure history, and cholesterol, indicate the influence of patient-specific medical attributes on response generation. The potential of integrating preventive and lifestyle-based determinants is underscored by environmental exposure and vaccination history. Linguistic features, such as language domain specialists and answer length, reflect the model’s sensitivity to bilingual query structure and domain phrasing. Contextual factors, including retrieval count and prior response, exhibit the crucial role of RAG in providing meaningful information. The findings of the SHAP analysis show that the model’s decision aligns with domain-plausible medical reasoning while maintaining linguistic fairness across languages.
Table 9 reports the outcomes of the proposed model on the test sets of datasets 1–4, underscoring its strength and efficiency. The superior performance showcases the model’s potential in handling bilingual queries. The balance between computational efficiency and high BLEU and ROUGE scores renders the feasibility of the model’s deployment in a real-time setting. The existing bilingual models, including Sengupta et al. (2023) [
36] and Preri et al. (2024) [
41], produced low-accuracy results due to their lack of contextual understanding of the user query. The recommended language identification model supported the proposed model in achieving a better ROUGE score. It assisted the model in identifying the user intent and delivering an appropriate response. The proposed model required low computational costs compared to the Wu et al. [
42] model, enabling the proposed model to be ideal for fast-paced and low-resource environments.
5. Discussion
In this study, we offered an effective bilingual MQAS to generate a reliable response for a query. The process of accessing trustworthy medical information is streamlined through the integration of advanced NLP techniques. The model’s interpretability is enabled in order to generate a reliable response. The proposed model supports Arabic-speaking individuals who experience difficulties due to the limited Arabic medical resources and databases. It facilitates users to obtain precise medical responses by addressing the existing challenges. It provides an efficient mechanism for retrieving medical information in a limited time, minimizing the demand for manual consultations. Individuals with limited medical expertise can understand complex medical terms in their preferred language. This promotes awareness of health issues and allows individuals to improve their quality of life.
The innovative integration of a generative, discriminative, and knowledge-aware neural architecture distinguishes the proposed MQAS from the existing models. By leveraging the complementary strengths of generative language models and extractive architectures, the proposed model effectively handles syntactic complexity and linguistic variations. KANs harmonize GPT-Neo and RoBERTa, enabling the model to align its outputs with medically validated concepts, reducing hallucination and improving trustworthiness. Furthermore, the deployment of an ensembling strategy adds a critical layer of intelligence to the decision-making process. This layered architecture generates adaptability across different query types, languages, and complexity levels.
The inability of traditional models to straightforwardly integrate external information into their design renders them inadequate for the healthcare domain. Compared to Zhang et al. (2022) [
29] (accuracy of 84.7%) and Van Sonsbeek et al. (2023) [
33] (accuracy of 84.3%), the proposed MQAS offered a remarkable accuracy in a bilingual setting. The application of a top-k retrieval approach enabled it to focus on the relevant information. The limited adaptability to bilingual datasets and limited query optimization reduced the performance of the Sengupta et al. (2023) [
36] and Preri et al. (2024) [
41] models. The integration of transformers facilitated the proposed model to generate responses with minimal latency, outperforming the Zeming et al. [
38] model. Unlike Jiang et al. (2023) [
37] (accuracy of 40.6%) and Zeming et al. (2023) [
38] (accuracy of 60.7%), the proposed model addressed healthcare queries in diverse linguistic contexts. The complexities of medical terminologies affected the Van Sonsbeek et al. (2023) [
33] and Preri et al. (2024) [
41] models’ ability, leading to suboptimal performance. The lack of advanced retrieval strategies affected the performance of the Wu et al. (2024) [
42] model.
The higher BLEU and F1-score values underscore the practical strengths of the proposed model in generating high-quality, contextually relevant responses. The generated responses match human-written responses in terms of structure and phrasing. This is essential in applications, including patient portals, digital health assistants, and triage chatbots. The improved BLEU scores show that the model is approachable and trustworthy to end-users. Similarly, the high F1-score reflects the model’s potential to consistently return reliable and comprehensive medical information, reducing the risk of misinformation and assisting clinicians as a decision-support tool. Healthcare providers in telemedicine settings may benefit from the proposed MQAS’s ability to pre-answer routine queries, triage symptom descriptions, and deliver contextually appropriate medical advice. This guarantees that patients receive prompt, evidence-based advice before a live consultation, improving efficiency and minimizing physician burden. The MQAS can be integrated into digital front-desk systems or patient-facing kiosks in healthcare facilities such as hospitals and clinics, facilitating bilingual assistance to patients who may not be confident in their ability to communicate in English. With the model’s dual-language understanding and response generation capabilities, healthcare practitioners and their various patient groups may communicate effectively, fostering inclusion. Utilizing natural-language queries, the system may additionally offer clinical decision assistance in settings where electronic health records (EHR) are linked. This could be accomplished by accessing essential clinical information, medical procedures, or summarizing patient conditions.
Unlike the existing models, the proposed MQAS extracts context from an external medical knowledge source to guarantee proper and suitable responses. RAG allows the model to access and retrieve current medical information, assisting the proposed model in generating a meaningful response. This synergy considerably enhances the model’s performance, enabling the proposed model to be suitable for the healthcare industry. The recommended approach addresses a crucial gap in bilingual QAS systems by providing high-quality responses with minimal computational resources. In cross-language contexts, existing models face challenges in handling the structure of specific languages and the nature of medical queries. The proposed model leveraged KANs, RAG, and effective fine-tuning strategies, rendering it capable of answering a wide range of medical queries, including straightforward and complex diagnostic queries, outperforming traditional models. In terms of BLEU and ROUGE, the model showed a significant improvement in response quality and relevance. The use of synthetic datasets supported the proposed model in achieving an exceptional outcome. The integration of the capabilities of GPT-Neo and RoBERTa reduced the response time, facilitating the user to retrieve faster responses. The BOHB optimization assisted the model in maintaining a trade-off between response time and reliable outcome.
The model’s interpretation ability is enhanced through the integration of LOD and KANs, responding to medical queries with greater semantic precision. For instance, in the query “ما أسباب أمراض القلب التاجية؟” (“What are the causes of coronary heart disease?”), the term “أمراض القلب التاجية” (“coronary heart disease”) is mapped to the standard medical concept using LOD resources, enriching the response through the incorporation of information, including underlying causes, clinical terminology, and risk factors. Similarly, the model distinguishes between the broader category “الخرف” (“dementia”) and the specific condition “الزهايمر” (“Alzheimer’s disease”), for the query “هل الزهايمر هو نفسه الخرف؟” (is Alzheimer’s the same as dementia?”). Using LOD resources, such as DBpedia and AraWiki, the model accesses hierarchical and relational knowledge, enabling it to explain that Alzheimer’s is a subtype of dementia. This feature enhances the accuracy, clarity, and educational value of the response.
The proposed context-aware bilingual MQAS exhibits strong performance across complex Arabic and English queries. However, we encounter challenges and limitations in implementing the model. The lack of accessible, high-quality Arabic medical literature is a significant challenge. While the proposed model generated a reliable outcome for Arabic and English queries, the lack of substantial, domain-specific Arabic datasets may influence precision and reliability. The integration of RAG into the proposed model may address this limitation by incorporating relevant external documents. However, this approach depends significantly on the variety and quality of the retrieved literature. A closer analysis suggests opportunities for further refinement, particularly in processing nuanced aspects of Arabic queries, highlighting edge cases where additional linguistic and domain tuning could further performance. For instance, the Arabic query, “ما هو أفضل علاج طبيعي لضغط الدم المرتفع؟” (“What is the best natural remedy for high blood pressure?”) obtains a clinically accurate and relevant response from the proposed model. However, the model is unable to emphasize the “natural remedy” aspect as strongly as intended. This suggests the requirement of additional training on alternative medical terminology in the Arabic language, assisting the model to better align with user-specific intent in lifestyle-oriented queries.
Within the context of multilingual healthcare settings, the proposed bilingual MQAS has the potential to revolutionize the accessibility of medical information, the level of patient interaction, and the clinical decision support. The limited availability of trustworthy medical content in Arabic is one of the persistent challenges in the Middle East and other multilingual regions. The proposed model addresses the challenges by integrating generative, discriminative, and knowledge-aware architectures, delivering contextually accurate and linguistically adaptive responses. By bridging the gap between Arabic-speaking individuals and medical information expressed in English, healthcare facilities, patient education platforms, and telemedicine systems are able to provide culturally and linguistically appropriate medical guidance. The proposed MQAS has the ability to serve as a decision-support system in clinical practice, assisting clinicians in retrieving evidence-based information. The interpretability layer, which is driven by SHAP values and uncertainty estimates, fosters clinician confidence by offering transparent justification for each generated response. In the context of AI-assisted diagnostics and education, this interpretability guarantees that responses can be inspected, confirmed, and modified, contributing to enhanced accountability. With credible, source-based, and multilingual responses, the proposed model can combat online medical misinformation for public health. By providing medical students and practitioners with reliable Arabic translations of complex English medical literature, the proposed MQAS provides assistance for cross-lingual learning in educational contexts.
We reduced the proposed model’s computational requirement in responding to a user query. However, healthcare centers may require substantial computational resources for model implementation. The specificity of the training data may limit the model’s generalization capability to rare medical conditions. The SHAP value interpretation may vary significantly in different medical contexts. This may affect the model’s consistency across English and Arabic languages. Periodic retraining with updated data is required to guarantee the model’s reliability and accuracy. We follow privacy and ethical guidelines in the model development. The implementation of the proposed multilingual MQAS poses substantial ethical challenges relating to medical misinformation, algorithmic bias, and user trust. Large language models typically generate probabilistic outputs, which may lead to hallucinations. The integration of RAG and KANs in generating responses has mitigated the hallucination rate, guaranteeing the generation of contextually relevant and evidence-based responses. SHAP-based interpretability and uncertainty quantification improve transparency, enabling end-users and physicians to understand prediction rationales and flag instances with high epistemic uncertainty that require human supervision. Moreover, bias mitigation is one of the ethical responsibilities in developing MQASs. The system balances Arabic and English datasets and uses KAN-based synonym enrichment to minimize semantic biases that might disfavor one language group over another.
Furthermore, the bias variance under 3% reveals that the model retains comparable reliability across language contexts, reducing the likelihood of biased performance or misunderstandings of medical terminology. From an ethical standpoint, the system follows the rules of transparency and human-in-the-loop decision-making. The proposed model is intended to promote informed patient education and clinician decision support. As a result of the proposed framework’s emphasis on interpretability, factual foundation, and human review for high-uncertainty scenarios, it guarantees that AI-assisted healthcare will continue to be trustworthy and ethically compliant. However, the proposed model requires careful handling to adhere to healthcare data regulations.
In the future, the proposed model can be extended to include multiple languages. The model’s adaptability can be improved by integrating various datasets and dialect variations. By leveraging regional Arabic dialects and rare medical terms, the model’s understanding of the context of the user query can be enhanced. Future work should explore other interpretability approaches for improving the model’s trustworthiness. Incorporation of patient history or demographic information could enable the model to provide personalized responses.