Next Article in Journal
Approximation of Dynamic Systems Using Deep Neural Networks and Laguerre Functions
Previous Article in Journal
A Proposal of “Echo Read” as an Interactive AI Reading System to Support Rereading
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Proceeding Paper

FairCXRnet: A Multi-Task Learning Model for Domain Adaptation in Chest X-Ray Classification for Low Resource Settings †

1
Department of Computer Science, African University of Science and Technology, Abuja 900107, Nigeria
2
Computer Science and Engineering Department, Ajay Kumar Garg Engineering College, Ghazibad 201015, India
3
Software Engineering Lab, University of Aizu, Fukushima 965-8580, Japan
4
Software Engineering Department, Bayero University, Kano 700006, Nigeria
*
Author to whom correspondence should be addressed.
Presented at the 7th International Global Conference Series on ICT Integration in Technical Education & Smart Society, Aizuwakamatsu City, Japan, 20–26 January 2025.
Eng. Proc. 2025, 107(1), 16; https://doi.org/10.3390/engproc2025107016
Published: 22 August 2025

Abstract

Medical imaging analysis plays a pivotal role in modern healthcare, with physicians relying heavily on radiologists for disease diagnosis. However, many hospitals face a shortage of radiologists, leading to long queues at radiology centers and delays in diagnosis. Advances in artificial intelligence (AI) have made it possible for AI models to analyze medical images and provide insights similar to those of radiologists. Despite their successes, these models face significant challenges that hinder widespread adoption. One major issue is the inability of AI models to generalize data from new populations, as performance tends to degrade when evaluated on datasets with different or shifted distributions, a problem known as domain shift. Additionally, the large size of these models requires substantial computational resources for training and deployment. In this study, we address these challenges by investigating domain shifts using ChestXray-14 and a Nigerian chest X-ray dataset. We propose a multi-task learning (MTL) approach that jointly trains the model on both datasets for two tasks, classification and segmentation, to minimize the domain gap. Furthermore, we replace traditional convolutional layers in the backbone model (Densenet-201) architecture with depthwise separable convolutions, reducing the model’s number of parameters and computational requirements. Our proposed model demonstrated remarkable improvements in both accuracy and AUC, achieving 93% accuracy and 96% AUC when tested across both datasets, significantly outperforming traditional transfer learning methods.

1. Introduction

The use of medical imaging in contemporary healthcare systems is essential, enabling physicians to diagnose and monitor diseases with high precision. Among the various imaging modalities, chest X-rays (CXRs) are one of the most widely used tools for detecting conditions such as pneumonia, tuberculosis, and COVID-19 [1]. However, the reliance on radiologists for interpretation poses difficulties, particularly in areas where qualified medical workers are lacking [2]. This shortage often leads to delayed diagnosis, negatively impacting patient outcomes. In light of this, deep learning-based artificial intelligence (AI) methods have emerged as promising solutions to aid in automatic medical image analysis, potentially alleviating the burden on radiologists [3].
Despite the notable progress in AI-based CXR analysis, two key challenges hinder adopting these models in real-world clinical settings: domain shift and high computational complexity [4]. Domain shift refers to the significant performance degradation that occurs when a model trained in one data set is evaluated in a data set from a different population or hospital with varying imaging protocols, equipment, or patient demographics [5]. This phenomenon is particularly concerning in medical imaging analysis, where models must generalize across diverse populations and healthcare systems. For instance, a model trained on the widely used ChestXray-14 dataset may not perform well on data collected in Nigerian hospitals, where differences in image acquisition and population characteristics create a substantial domain gap. Addressing this domain shift is crucial for building robust models that can be deployed globally [6].
Additionally, state-of-the-art deep learning models often require vast computational resources for training and inference, limiting their accessibility, particularly in low-resource settings [7]. The large number of parameters in these models makes them both memory-intensive and computationally expensive, further complicating their practical deployment in resource-constrained environments [8].
To address these challenges, we propose the MTL approach that performs both classification and segmentation tasks to improve the generalizability of deep learning models across different populations. Specifically, we aim to minimize the domain gap between the ChestXray-14 dataset and a Nigerian chest X-ray dataset. By training a model on multiple related tasks (segmentation and classification) simultaneously, MTL encourages the model to learn shared representations that are more robust to domain shifts. Additionally, we replace traditional convolutional layers with depthwise separable convolutions to reduce the number of trainable parameters and computational requirements, building a model that is more efficient for deployment in low-resource environments.
Traditionally, models are trained sequentially on single tasks, such as segmentation, followed by parameter tuning for a different task, like classification, until desirable performance is achieved. This approach allows the model to handle multiple tasks but lacks efficiency and cohesion. In contrast, MTL leverages domain-specific features from related tasks to improve generalization, enabling the model to perform better across several tasks simultaneously. By allowing one task to reinforce another, MTL enhances overall model performance with a single, unified architecture.
Therefore, in this work, we propose a novel method for joint segmentation and classification that aligns domain differences between two datasets, improving both classification accuracy and segmentation quality. Our main technical contribution is the development of a lightweight multi-task learning model, capable of utilizing small, labeled medical image datasets for efficient classification and segmentation.
The remainder of this paper is structured as follows: Section 2 reviews related works addressing domain shifts in chest X-ray analysis and medical imaging. Section 3 outlines the methodology employed in this study, detailing the multi-task learning approach and architectural improvements. Section 4 presents the experimental results, followed by an in-depth discussion of the findings. Finally, Section 5 concludes the paper with a summary of key insights and future research directions.

2. Related Works

In recent years, the rapid advancements in AI and deep learning have enabled significant progress in medical imaging analysis, particularly in tasks such as classification, segmentation, and disease detection from chest X-rays (CXRs) [9]. However, despite the success of these models, challenges related to domain shift and computational complexity remain critical obstacles to their widespread adoption in real-world clinical settings [10]. This section provides an overview of relevant works addressing these issues, with a particular focus on methods for handling domain shifts in chest X-ray classification as well as strategies for improving computational efficiency.

2.1. Domain Shift in Medical Imaging

Domain shift refers to the degradation in model performance when it is trained on one dataset and evaluated on a different dataset with varying characteristics, such as differences in patient demographics, imaging devices, and clinical protocols [11]. This issue is especially pronounced in medical imaging, where diverse data distributions can arise due to variations in hospital equipment, image acquisition techniques, and population health conditions [12]. Several studies have explored solutions to mitigate the effects of domain shift in chest X-ray analysis.
One popular approach to address domain shift is through domain adaptation techniques [13]. For unsupervised settings where the dataset is not labeled, Unsupervised Domain Adaptation (UDA) techniques were used to mitigate the effect of domain shift. The technique performed well in a series of studies, and it performed well on the famous office-31 dataset for domain adaptation and achieved an adaptive effect in the event of large domain shifts [14]. Motiian explored supervised domain adaption techniques where both data from source and target domains are labeled, by learning a discriminative embedding subspace with mapped visual domains that are maximally segregated but semantically linked using Siamese architecture [15]. Tzeng proposed a Domain-Adversarial Neural Network (DANN) as a technique to handle domain shift, where an adversarial objective is used to align the feature distributions of the source and target domains, thereby improving model generalization across different datasets [16]. Similarly, Adversarial Discriminative Domain Adaptation (ADDA) is proposed to minimize domain discrepancies by optimizing both classification and adversarial loss functions. These approaches have shown promise, but their performance can still degrade when there is significant variability between domains [17].
Another method to tackle domain shift is transfer learning, which has been widely used in medical imaging tasks [18]. Models pre-trained on large datasets, such as ImageNet [19] or ChestXray-14 [20], are fine-tuned on a smaller, target domain dataset to adapt the learned features to the new domain. Transfer learning approaches for CXR classification were used and achieved considerable improvements, but these methods still rely heavily on large labeled datasets and do not fully eliminate the domain shift when applied to datasets with distinct characteristics, such as those from low-resource healthcare settings [21].
Despite these advances, most existing domain adaptation methods focus on aligning distributions between datasets but fail to consider the underlying task complexity and the potential benefits of incorporating additional tasks. This motivates the use of MTL as a more holistic approach to mitigate domain shift.

2.2. Multi-Task Learning in Medical Imaging

MTL has emerged as a powerful framework in deep learning, where multiple related tasks are learned jointly [22]. By sharing representations between tasks, MTL encourages the model to learn more robust and generalized features, making it less susceptible to domain-specific variations [23]. Caruana first introduced the concept of MTL, demonstrating that learning-related tasks together improve generalization and reduce overfitting [24].
In the context of medical imaging, several studies have successfully applied MTL to tasks such as disease classification and segmentation. Gai et al. [25] proposed an MTL-based architecture for simultaneous segmentation and classification of lung nodules in CT scans, showing that the model learned complementary features from both tasks, resulting in improved accuracy and generalization. Others explored the use of MTL for COVID-19 detection from chest X-rays, combining the classification of pneumonia, COVID-19, and normal cases with a segmentation task to identify affected regions in the lungs. Their results showed significant improvements over single-task learning approaches [26].
However, while MTL has demonstrated its effectiveness in medical imaging, no studies have applied this approach to address domain shifts across populations, to the best of our knowledge. By training on multiple tasks, such as classification and segmentation, MTL can potentially improve feature extraction across different domains, providing a more robust solution to the cross-population domain shift problem. This research builds on the successes of MTL by investigating its role in mitigating domain shifts between two distinct datasets: the publicly available ChestXray-14 dataset and the Nigerian chest X-ray dataset.

2.3. Lightweight and Efficient Architectures

Another key challenge in deploying deep learning models in medical imaging is the computational cost associated with training and inference. State-of-the-art deep learning models often contain millions of parameters, making them computationally expensive to train and requiring significant memory for deployment [27]. This is a critical issue, especially in low-resource settings, where access to high-performance computing infrastructure is limited [28].
Several architectural innovations have been proposed to address the computational complexity of deep learning models. Howard et al. [29] introduced MobileNet, a lightweight architecture that utilizes depthwise separable convolutions to significantly reduce the number of parameters without sacrificing performance. The concept was extended with Xception, which demonstrated the effectiveness of depthwise separable convolutions in large-scale image classification tasks [30]. These models have proven particularly useful in resource-constrained environments, as they achieve state-of-the-art performance with reduced computational requirements.
In the medical imaging domain, lightweight architectures such as MobileNet have been explored for tasks like disease detection and segmentation. Rajpurkar [20] applied MobileNet for pneumonia detection from chest X-rays and achieved competitive performance compared to large models. Additionally, the use of depthwise separable convolutions in skin cancer detection was seen in the literature, showing that such architectures could deliver strong performance while being efficient in terms of computational and memory requirements [31].
Therefore, we leverage the power of depthwise separable convolutions in our proposed multi-task learning architecture to reduce the computational burden while maintaining high accuracy and generalization across domains. By combining MTL with a lightweight architecture, we aim to create a model that is not only robust to domain shift but also practical for deployment in resource-constrained settings.

3. Methodology

This section details the methodology used to address the problem of cross-population domain shift problem in chest X-ray analysis with the MTL approach. Our goal is to develop a model capable of jointly performing classification and segmentation tasks across two datasets, namely ChestXray-14 and the Nigerian chest X-ray dataset, while also optimizing computational efficiency. To achieve this, we designed a lightweight MTL architecture using depthwise separable convolutions and Densenet-201 as the backbone model, aiming to reduce the domain gap and generalize effectively across different populations.

3.1. Dataset Description

We utilized two datasets in this study. ChestXray-14 Dataset: A publicly available dataset with 112,120 chest X-ray images labeled for 14 thoracic diseases, collected in the United States (US), and organized by the National Institutes of Health (NIH). This dataset was used as the source domain. Nigerian Chest X-ray Dataset (Nigeria CXR): Collected from Aminu Kano Teaching Hospital in Nigeria, this dataset consists of 6345 chest X-ray images labeled into four classes: Pneumonia, COVID-19, Tuberculosis (TB), and Normal. This dataset was used as the target domain.
Figure 1 shows sample images from the dataset. Both datasets were annotated by radiologists and pre-processed to ensure consistency in image size and quality. We applied standard image augmentation techniques such as rotation, flipping, and contrast adjustment to increase the diversity of the training data and improve the model’s robustness to variations in image quality and acquisition settings.

3.2. Proposed Multi-Task Learning Model

The proposed model is based on MTL architecture that performs both classification and segmentation tasks. By learning these two related tasks simultaneously, the model can share representations between tasks, improving generalization and reducing domain shift. The proposed model is well depicted in Figure 2. It shows that the model architecture consists of two main branches:
To enhance robustness against domain variations, an attention-guided fusion mechanism was utilized to exploits segmentation features that refine the classification pathway in shared feature space. The decoder output f seg is transformed into an attention mask through a 1 × 1 convolution and sigmoid activation, i.e.,
A = σ Conv 1 × 1 ( f seg ) , f att = A f s ,
where f s denotes the shared feature representation and ⊙ the elementwise product. The gated representation f att emphasizes anatomically relevant regions and is concatenated with the original features to form z = [ f s f att ] , which is subsequently fed into the classifier. In addition, to mitigate domain shift, we incorporate a Maximum Mean Discrepancy (MMD) regularizer that minimizes the distributional gap between source and target features in the reproducing kernel Hilbert space:
MMD 2 ( f s ( S ) , f s ( T ) ) = 1 n S 2 i , j k ( x i S , x j S ) + 1 n T 2 i , j k ( x i T , x j T ) 2 n S n T i , j k ( x i S , x j T ) ,
with k ( · , · ) defined as a mixture of RBF kernels. The overall training objective is then expressed as
L = L cls + λ seg L seg + λ mmd L MMD ,
ensuring joint optimization of classification accuracy, segmentation quality, and domain-invariant feature alignment.
1-Classification head: This branch is responsible for disease classification, predicting one of the four classes (Pneumonia, COVID-19, TB, Normal) in both datasets.
2-Segmentation head: The segmentation branch is designed to identify and localize regions of interest in the chest X-rays, such as areas affected by pneumonia or TB. This branch outputs a pixel-wise mask that highlights the regions of abnormality.
The two tasks share a common encoder, which extracts features from the input images. The shared encoder allows the model to learn useful representations for both tasks, encouraging it to capture more general features that are less sensitive to domain-specific variations. The training procedure is best desribed in Algorithm 1, where the proposed MTL model integrates segmentation-guided attention with distribution alignment to improve classification under cross-population settings.
Algorithm 1 Attention-Guided Fusion + RBF–MMD
Require:  f s , f seg , f s ( S ) , f s ( T ) , y cls , y mask , λ seg , λ mmd , Σ
  1:
A σ ( Conv 1 × 1 ( f seg ) )
  2:
f att A f s
  3:
z Concat ( f s , f att )
  4:
logits Head cls ( z )
  5:
L cls CE / Focal ( logits , y cls )
  6:
L seg Dice + CE ( SegPred , y mask )
  7:
X Flatten ( f s ( S ) ) ,    Y Flatten ( f s ( T ) )
  8:
L MMD MMD _ RBF ( X , Y , Σ )
  9:
L L cls + λ seg L seg + λ mmd L MMD
10:
return  L

3.3. Depthwise Separable Convolutions

To address the computational challenges associated with training deep learning models on large medical datasets, we integrated depthwise separable convolutions into the MTL architecture. Depthwise separable convolutions break down standard convolutions into two smaller operations:
  • Depthwise Convolution: This operation independently applies a single filter to each input channel, reducing the computational cost.
  • Pointwise Convolution: A 1 × 1 convolution is then applied to combine the outputs of the depthwise convolution.
By using depthwise separable convolutions, the model parameters are significantly reduced while maintaining its capacity to learn rich feature representations. This design choice made the model lightweight and computationally efficient, making it suitable for deployment in low-resource settings.

3.4. Model Training Strategy

The proposed MTL model training is a twofold training process: 1. Joint Training: The model was trained on both tasks (classification and segmentation) simultaneously using a multi-task loss function. The overall loss function, L_total, is a weighted sum of the classification loss and the segmentation loss:
L total = α · L classification + β · L segmentation
where α and β are hyperparameters that control the contribution of each task to the total loss.
2. Domain Alignment: To reduce the domain gap between the ChestXray-14 and the Nigeria CXR datasets, we applied feature sharing during model optimization between features extracted from the source and target domains. The model’s encoder was trained to learn domain-invariant features in the shared feature space. This helped align the feature distributions of the two datasets, improving the model’s generalization to the target domain.
All experimental results are evaluated on the standard performance metrics, such as Accuracy, Area Under the Receiver Operating Characteristic Curve (AUC), Precision, Recall, and F1-score.
The model’s performance was compared against several baseline methods, including transfer learning models and single-task learning approaches. We conducted extensive experiments to assess the model’s robustness to domain shift, particularly focusing on its ability to generalize to the Nigerian chest X-ray dataset.

3.5. Implementation Details

The model was implemented in Pytorch 2.7.1 and trained on an A100 GPU. An Adam optimizer with an initial learning rate of 0.0001 and a batch size 32 was used. Early stopping was employed to prevent overfitting, and a learning rate scheduler was utilized to reduce the learning rate when the validation loss converged gradually.

4. Results

The Results Section provides a detailed account of the performance of the proposed multi-task learning model on both classification and segmentation tasks. The evaluation focuses on key metrics such as accuracy, AUC, and F1-score, and comparisons are made against baseline models to demonstrate the effectiveness of our approach.

4.1. Experiment 1: Establishing Baselines

In this experiment, we used Densenet-201 [32] as the backbone model. The model is trained and evaluated on both ChestXray-14 and the Nigerian dataset. The goal was to establish a baseline for training using a conventional architecture. Table 1 presents the result of experiment 1, which shows the traditional CNN model’s performance on both datasets. The experiment demonstrated DenseNet-201’s baseline performance when trained on the two datasets, offering a reference point for comparing with our proposed model.

4.2. Investigation of Domain Shift Between ChestXray-14 and Nigerian Dataset

In the second experiment, we trained DenseNet-201 solely on the ChestXray-14 dataset and tested it on the Nigeria CXR dataset to investigate the impact of domain shift. The goal was to observe the model’s performance when evaluated on a new population dataset without domain adaptation techniques.
Experiment 2 revealed significant performance degradation when DenseNet-201 was trained on ChestXray-14 and tested on the Nigerian dataset. Table 2 summarizes our quantitative findings, including the Accuracy and average AUC values for experiment 2. The findings following the model’s evaluation on Nigeria CXR show reduced performance on all the performance metrics, indicating that the model did not generalize well to new datasets. This suggests that the training data from the source domain does not sufficiently represent the distribution of data from the target domain.

4.3. Proposed FairCXRnet Results

In the experiment 3, we trained the proposed MTL model on the ChestXray-14 and Nigeria CXR datasets for joint classification and segmentation. The model was tested on the Nigerian dataset to evaluate its ability to generalize and reduce domain shift.
Figure 3 shows the performance in terms of accuracy of the proposed model compared with the transfer learning/fine tuning and continual learning technique, and Table 3 summarizes the findings.
From Table 3, it could be inferred that the proposed FairCXRnet model outperformed the transfer learning approach using Densenet-201 across all metrics, demonstrating its effectiveness in reducing domain shift and improving overall performance across tasks. Continual learning (CL) also shows promising performance, striking a good balance between precision and recall. However, it suffers from catastrophic forgetting: as it continues learning new tasks, it tends to forget the old task. Additionally, while Densenet has over 20 M parameters, and a 77 MB size on disk, the integration of depthwise separable convolutions in our model reduces the total parameters to just 197,848, and model size to 31 MB, achieving 59% parameter reduction with more generalization capability, making it suitable for low-resource settings.

Discussion

The experimental results demonstrate the effectiveness of the proposed FairCXRnet model in addressing the challenges of cross-population domain shift and computational efficiency in chest X-ray analysis. In the first experiment, DenseNet201, trained on both the ChestXray-14 and Nigerian datasets, exhibited reasonable performance. However, when the model was trained solely on ChestXray-14 and tested on the Nigerian dataset (experiment 2), we observed a significant drop in accuracy and AUC, confirming the presence of domain shift. This performance degradation stresses the difficulty in applying models trained on one population to data from a different region or demographic without adaptation.
In contrast, the third experiment, where the proposed FairCXRnet model was trained on both datasets, demonstrated notable improvements. The FairCXRnet model achieved an accuracy of 93% and an AUC of 96% on the Nigerian dataset, outperforming the DenseNet-201 baseline by a substantial margin, further validating the benefit of jointly learning-related tasks. The use of depthwise separable convolutions not only reduced the model’s parameter count but also reduced training time significantly from 8 h to 2 h, improving computational efficiency and making the model more suitable for deployment in resource-constrained settings. These results suggest that multi-task learning, combined with domain alignment strategies, is a promising approach for improving generalization in medical imaging, particularly in diverse and underrepresented populations.

5. Conclusions and Future Work

In this paper, we addressed the challenges of domain shift and computational efficiency in chest X-ray analysis by proposing an MTL model that jointly performs classification and segmentation. By leveraging the ChestXray-14 and Nigerian chest X-ray datasets, we demonstrated that our model successfully reduces the domain gap and improves performance on unseen data from a different population. The experimental results show that our MTL approach outperformed traditional single-task models, achieving higher classification accuracy and better segmentation quality while significantly reducing the number of model parameters through the use of depthwise separable convolutions. This makes the model not only effective in mitigating domain shift but also computationally efficient, which is crucial for deployment in resource-limited settings such as hospitals with minimal computing infrastructure.
Our findings suggest that multi-task learning, combined with domain alignment strategies, holds promise for improving generalization in medical imaging applications across diverse populations. The results also highlight the importance of considering domain-specific variations when developing AI models for healthcare, ensuring that they are adaptable and robust across different regions and patient demographics.
While our proposed model achieved significant improvements, there are several avenues for future research. First, we plan to explore more advanced domain adaptation techniques, such as self-supervised learning, to further improve the model’s generalization to new datasets. Additionally, incorporating larger and more diverse datasets could help to better understand the effects of demographic variations on model performance.

Author Contributions

Conceptualization, A.M., R.P., M.H. (Mohammed Hassan) and M.H. (Mohamed Hamada); methodology, A.M.; software, A.M.; validation, A.M., S.Y.I., R.P., M.H. (Mohammed Hassan) and M.H. (Mohamed Hamada); formal analysis, A.M.; investigation, A.M. and S.Y.I.; resources, R.P., M.H. (Mohammed Hassan), and M.H. (Mohamed Hamada); data curation, A.M. and S.Y.I.; writing—original draft preparation, A.M.; writing—review and editing, A.M., R.P., M.H. (Mohammed Hassan), M.H. (Mohamed Hamada), and S.Y.I.; visualization, A.M.; supervision, R.P., M.H. (Mohammed Hassan) and M.H. (Mohamed Hamada); project administration, R.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets used in this study include NIH ChestX-ray14 dataset (available via NIH Clinical Center). In addition, a locally collected dataset from Nigeria was used, available on kaggle.com at: https://www.kaggle.com/datasets/aminumusa/nigeria-chest-x-ray-dataset (accessed on 26 January 2025). The source code for implementing FairCXRnet is openly available on GitHub at: https://www.github.com/El-amin/FairCXRnet-A-Multi-Task-Learning-Model-for-Chest-X-Ray-Classification-for-Low-Resource-Settings (accessed on 26 January 2025).

Acknowledgments

The authors gratefully acknowledge the support of the CRAI AI Lab, Federal University Dutse, Nigeria, for providing computational resources, technical support, and a collaborative research environment that enabled the successful completion of this work.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Musa, A.; Adam, F.M.; Ibrahim, U.; Zandam, A.Y. Learning from Small Datasets: An Efficient Deep Learning Model for Covid-19 Detection from Chest X-ray Using Dataset Distillation Technique. In Proceedings of the 2022 IEEE Nigeria 4th International Conference on Disruptive Technologies for Sustainable Development (NIGERCON), Lagos, Nigeria, 5–7 April 2022; pp. 1–6. [Google Scholar]
  2. Musa, A.; Ibrahim Adamu, M.; Kakudi, H.A.; Hernandez, M.; Lawal, Y. Analyzing Cross-Population Domain Shift in Chest X-Ray Image Classification and Mitigating the Gap with Deep Supervised Domain Adaptation. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2024; Linguraru, M.G., Dou, Q., Feragen, A., Giannarou, S., Glocker, B., Lekadir, K., Schnabel, J.A., Eds.; Springer: Cham, Switzerland, 2024; pp. 585–595. [Google Scholar]
  3. Wang, F.; Casalino, L.P.; Khullar, D. Deep learning in medicine—promise, progress, and challenges. JAMA Intern. Med. 2019, 179, 293–294. [Google Scholar] [CrossRef] [PubMed]
  4. Jiang, L.; Wu, Z.; Xu, X.; Zhan, Y.; Jin, X.; Wang, L.; Qiu, Y. Opportunities and challenges of artificial intelligence in the medical field: Current application, emerging problems, and problem-solving strategies. J. Int. Med. Res. 2021, 49, 03000605211000157. [Google Scholar] [CrossRef] [PubMed]
  5. Pooch, E.H.; Ballester, P.; Barros, R.C. Can we trust deep learning based diagnosis? the impact of domain shift in chest radiograph classification. In Thoracic Image Analysis: Second International Workshop, TIA 2020, Held in Conjunction with MICCAI 2020, Lima, Peru, October 8, 2020, Proceedings 2; Springer: Berlin/Heidelberg, Germany, 2020; pp. 74–83. [Google Scholar]
  6. Hassan, M.; Hamada, M. Performance comparison of feed-forward neural networks trained with different learning algorithms for recommender systems. Computation 2017, 5, 40. [Google Scholar] [CrossRef]
  7. Alippi, C.; Disabato, S.; Roveri, M. Moving convolutional neural networks to embedded systems: The alexnet and VGG-16 case. In Proceedings of the 2018 17th ACM/IEEE International Conference on Information Processing in Sensor Networks (IPSN), Porto, Portugal, 11–13 April 2018; pp. 212–223. [Google Scholar]
  8. Salau, L.; Hamada, M.; Prasad, R.; Hassan, M.; Mahendran, A.; Watanobe, Y. State-of-the-art survey on deep learning-based recommender systems for e-learning. Appl. Sci. 2022, 12, 11996. [Google Scholar] [CrossRef]
  9. Shen, D.; Wu, G.; Suk, H.I. Deep learning in medical image analysis. Annu. Rev. Biomed. Eng. 2017, 19, 221–248. [Google Scholar] [CrossRef] [PubMed]
  10. Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M.; Van Der Laak, J.A.; Van Ginneken, B.; Sánchez, C.I. A survey on deep learning in medical image analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef] [PubMed]
  11. Kondrateva, E.; Pominova, M.; Popova, E.; Sharaev, M.; Bernstein, A.; Burnaev, E. Domain shift in computer vision models for MRI data analysis: An overview. In Proceedings of the Thirteenth International Conference on Machine Vision, Rome, Italy, 2–6 November 2020; pp. 126–133. [Google Scholar]
  12. Li, H.; Wang, Y.; Wan, R.; Wang, S.; Li, T.Q.; Kot, A. Domain generalization for medical imaging classification with linear-dependency regularization. Adv. Neural Inf. Process. Syst. 2020, 33, 3118–3129. [Google Scholar]
  13. Gulrajani, I.; Lopez-Paz, D. In search of lost domain generalization. arXiv 2020, arXiv:2007.01434. [Google Scholar]
  14. Ganin, Y.; Lempitsky, V. Unsupervised domain adaptation by backpropagation. In Proceedings of the International Conference on Machine Learning, PMLR, Lille, France, 6–11 July 2015; pp. 1180–1189. [Google Scholar]
  15. Motiian, S.; Piccirilli, M.; Adjeroh, D.A.; Doretto, G. Unified deep supervised domain adaptation and generalization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 5715–5725. [Google Scholar]
  16. Tzeng, E.; Hoffman, J.; Saenko, K.; Darrell, T. Adversarial discriminative domain adaptation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7167–7176. [Google Scholar]
  17. Xu, M.; Zhang, J.; Ni, B.; Li, T.; Wang, C.; Tian, Q.; Zhang, W. Adversarial domain adaptation with domain mixup. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 6502–6509. [Google Scholar]
  18. Kouw, W.M.; Loog, M. An introduction to domain adaptation and transfer learning. arXiv 2018, arXiv:1812.11806. [Google Scholar]
  19. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  20. Rajpurkar, P. CheXNet: Radiologist-Level Pneumonia Detection on Chest X-Rays with Deep Learning. arXiv 2017, arXiv:1711.05225. [Google Scholar]
  21. Irvin, J.; Rajpurkar, P.; Ko, M.; Yu, Y.; Ciurea-Ilcus, S.; Chute, C.; Marklund, H.; Haghgoo, B.; Ball, R.; Shpanskaya, K.; et al. Chexpert: A large chest radiograph dataset with uncertainty labels and expert comparison. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 590–597. [Google Scholar]
  22. Zhang, Y.; Yang, Q. An overview of multi-task learning. Natl. Sci. Rev. 2018, 5, 30–43. [Google Scholar] [CrossRef]
  23. Zhang, W.; Li, R.; Zeng, T.; Sun, Q.; Kumar, S.; Ye, J.; Ji, S. Deep model based transfer and multi-task learning for biological image analysis. In Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Sydney, NSW, Australia, 10–13 August 2015; pp. 1475–1484. [Google Scholar]
  24. Caruana, R. Multitask learning. Mach. Learn. 1997, 28, 41–75. [Google Scholar] [CrossRef]
  25. Li, G.; Zhou, W.; Chen, W.; Sun, F.; Fu, Y.; Gong, F.; Zhang, H. Study on the detection of pulmonary nodules in CT images based on deep learning. IEEE Access 2020, 8, 67300–67309. [Google Scholar] [CrossRef]
  26. Bao, G.; Chen, H.; Liu, T.; Gong, G.; Yin, Y.; Wang, L.; Wang, X. COVID-MTL: Multitask learning with Shift3D and random-weighted loss for COVID-19 diagnosis and severity assessment. Pattern Recognit. 2022, 124, 108499. [Google Scholar] [CrossRef] [PubMed]
  27. Alzubaidi, L.; Zhang, J.; Humaidi, A.J.; Al-Dujaili, A.; Duan, Y.; Al-Shamma, O.; Santamaría, J.; Fadhel, M.A.; Al-Amidie, M.; Farhan, L. Review of deep learning: Concepts, CNN architectures, challenges, applications, future directions. J. Big Data 2021, 8, 1–74. [Google Scholar] [CrossRef] [PubMed]
  28. Rahman, M.A.; Hamada, M. Burrows–wheeler transform based lossless text compression using keys and Huffman coding. Symmetry 2020, 12, 1654. [Google Scholar] [CrossRef]
  29. Howard, A.G. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar] [CrossRef]
  30. Xie, G.; Wang, J.; Zhang, T.; Lai, J.; Hong, R.; Qi, G.J. Interleaved structured sparse convolutional neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8847–8856. [Google Scholar]
  31. Li, Y.; Esteva, A.; Kuprel, B.; Novoa, R.; Ko, J.; Thrun, S. Skin cancer detection and tracking using data synthesis and deep learning. In Proceedings of the Workshops at the Thirty-first AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017. [Google Scholar]
  32. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  33. Sanchez, K.; Hinojosa, C.; Arguello, H.; Kouamé, D.; Meyrignac, O.; Basarab, A. CX-DaGAN: Domain adaptation for pneumonia diagnosis on a small chest X-ray dataset. IEEE Trans. Med. Imaging 2022, 41, 3278–3288. [Google Scholar] [CrossRef] [PubMed]
  34. Lenga, M.; Schulz, H.; Saalbach, A. Continual learning for domain adaptation in chest x-ray classification. In Proceedings of the Medical Imaging with Deep Learning, PMLR, Montreal, QC, Canada, 6–8 July 2020; pp. 413–423. [Google Scholar]
Figure 1. Sample X-ray images from dataset.
Figure 1. Sample X-ray images from dataset.
Engproc 107 00016 g001
Figure 2. Proposed FairCXRnet methodology.
Figure 2. Proposed FairCXRnet methodology.
Engproc 107 00016 g002
Figure 3. Performance comparison of FairCXRnet with other techniques.
Figure 3. Performance comparison of FairCXRnet with other techniques.
Engproc 107 00016 g003
Table 1. Experiment I results.
Table 1. Experiment I results.
DatasetsAccuracyPrecisionRecallF1-ScoreAUC
ChestXray1496.1086.7383.3396.000.95
Nigerian-CXR91.3391.6991.3391.320.98
Table 2. Experiment II results.
Table 2. Experiment II results.
DatasetsAccuracyPrecisionRecallF1-ScoreAUC
ChestXray14 –>Nigerian CXR61.5851.1650.0359.870.47
Table 3. Proposed FairCXRnet results.
Table 3. Proposed FairCXRnet results.
TechniquesAccuracyPrecisionRecallF1-ScoreAUC
Transfer learning and fine-tuning [33]88.3682.8879.1688.470.87
Continual learning [34]72.4174.6672.0271.970.71
Proposed FairCXRnet93.2689.8891.7693.070.96
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Musa, A.; Prasad, R.; Hassan, M.; Hamada, M.; Ilu, S.Y. FairCXRnet: A Multi-Task Learning Model for Domain Adaptation in Chest X-Ray Classification for Low Resource Settings. Eng. Proc. 2025, 107, 16. https://doi.org/10.3390/engproc2025107016

AMA Style

Musa A, Prasad R, Hassan M, Hamada M, Ilu SY. FairCXRnet: A Multi-Task Learning Model for Domain Adaptation in Chest X-Ray Classification for Low Resource Settings. Engineering Proceedings. 2025; 107(1):16. https://doi.org/10.3390/engproc2025107016

Chicago/Turabian Style

Musa, Aminu, Rajesh Prasad, Mohammed Hassan, Mohamed Hamada, and Saratu Yusuf Ilu. 2025. "FairCXRnet: A Multi-Task Learning Model for Domain Adaptation in Chest X-Ray Classification for Low Resource Settings" Engineering Proceedings 107, no. 1: 16. https://doi.org/10.3390/engproc2025107016

APA Style

Musa, A., Prasad, R., Hassan, M., Hamada, M., & Ilu, S. Y. (2025). FairCXRnet: A Multi-Task Learning Model for Domain Adaptation in Chest X-Ray Classification for Low Resource Settings. Engineering Proceedings, 107(1), 16. https://doi.org/10.3390/engproc2025107016

Article Metrics

Back to TopTop