Next Article in Journal
Balancing Complexity and Performance in Convolutional Neural Network Models for QUIC Traffic Classification
Previous Article in Journal
Medical Data over Sound—CardiaWhisper Concept
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Eye-Guided Multimodal Fusion: Toward an Adaptive Learning Framework Using Explainable Artificial Intelligence

1
Computational NeuroSurgery (CNS) Lab, Macquarie Medical School, Faculty of Medicine, Health and Human Sciences, Macquarie University, Sydney 2113, Australia
2
Centre for Applied Artificial Intelligence, School of Computing, Faculty of Science and Engineering, Macquarie University, Sydney 2113, Australia
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(15), 4575; https://doi.org/10.3390/s25154575
Submission received: 19 May 2025 / Revised: 10 July 2025 / Accepted: 21 July 2025 / Published: 24 July 2025
(This article belongs to the Section Sensing and Imaging)

Abstract

Interpreting diagnostic imaging and identifying clinically relevant features remain challenging tasks, particularly for novice radiologists who often lack structured guidance and expert feedback. To bridge this gap, we propose an Eye-Gaze Guided Multimodal Fusion framework that leverages expert eye-tracking data to enhance learning and decision-making in medical image interpretation. By integrating chest X-ray (CXR) images with expert fixation maps, our approach captures radiologists’ visual attention patterns and highlights regions of interest (ROIs) critical for accurate diagnosis. The fusion model utilizes a shared backbone architecture to jointly process image and gaze modalities, thereby minimizing the impact of noise in fixation data. We validate the system’s interpretability using Gradient-weighted Class Activation Mapping (Grad-CAM) and assess both classification performance and explanation alignment with expert annotations. Comprehensive evaluations, including robustness under gaze noise and expert clinical review, demonstrate the framework’s effectiveness in improving model reliability and interpretability. This work offers a promising pathway toward intelligent, human-centered AI systems that support both diagnostic accuracy and medical training.

1. Introduction

Generally, eye-tracking reflects a person’s attentional behavior by measuring their eye movements. This technology provides precise insights into how attention is allocated, making it especially useful in fields that heavily depend on visual information processing, such as medicine and education [1,2]. In medical imaging, eye-tracking data, characterized by specific metrics such as fixation, reflex, and saccades, can significantly enhance deep learning (DL) models by guiding them to areas of interest, thereby improving the accuracy of abnormality detection. This approach addresses a notable challenge in many medical imaging tasks, where the scarcity of large, annotated datasets limits the model’s ability to learn directly from extensive labeled examples. For instance, analysis of gaze data indicates that radiologists allocate greater attention to specific areas indicative of abnormalities during verbal reporting of findings, compared to silent examination of the image [3]. However, despite their effectiveness, DL models are criticized as “black boxes” due to their lack of transparency, which highlights the need for explainable artificial intelligence (XAI) [4]. Additionally, traditional radiology training relies on extensive practical experience, which can be enhanced by integrating experts’ gaze data to accelerate skill acquisition among novice radiologists [5].
Recent literature highlights the potential of eye-tracking in medical education, underscoring its ability to differentiate experts from novices through distinct patterns of gaze [6]. Experienced clinicians rapidly form holistic impressions when interpreting medical images, whereas novices typically struggle without substantial practice [7]. With the shift towards online and blended educational environments, integrating eye-tracking data into training programs could bridge this expertise gap. Prior studies indicate that visual characteristics captured through eye movements correlate with cognitive processes involved in clinical reasoning, suggesting a promising approach for knowledge transfer between experts and novices [8]. However, challenges remain, such as handling noisy fixation data and ensuring that gaze-based methodologies align with real-world diagnostic scenarios [9,10]. To address these challenges, this study introduces an Eye-Gaze Guided Multimodal Fusion Framework designed to enhance radiology education and diagnostic accuracy. The overall model and the evaluation steps are illustrated in Figure 1. The key contributions are as follows:
  • Eye-Guided Framework: The Multimodal Fusion Framework incorporates attention mechanisms to capture the most crucial aspects of the stimulus. The model with a shared backbone mitigates the impact of noisy fixation data and separates the processing of modalities to improve the performance of automatic abnormality detection in CXRs.
  • Explanatory Support for Transparency: We provide post-hoc feature attribution explanations to help radiology trainees better understand lesion classification in chest X-rays.
  • Evaluation of the Approach. We evaluate our approach and demonstrate robust performance under noisy conditions, showing resilience to misaligned fixation maps. We further assess the interpretability of the model by utilizing Grad-CAM, ensuring that the generated visual explanations align with expert-annotated ROIs. This alignment enhances the clinical reliability of the model’s predictions.

2. Related Work

2.1. Eye-Gaze Tracking in Radiology

Eye-tracking technology has been extensively explored for over a decade to study human visual perception, yielding numerous methods for tracking eye movements in various applications. Typically, eye tracking is applied in two main areas: interactive and diagnostic applications [11]. In an interactive context, users’ eye movement data is an input method, whereas in a diagnostic context, it provides insights into their intentions and cognitive processes [1]. While AI advancements have enabled medical deep learning to utilize prior knowledge in diagnostic tools, early studies indicate that eye-tracking technology can enhance clinician interaction and improve AI systems [12,13]. Eye tracking studies in stimulus analysis primarily involve exploring where and how medical professionals examine different modalities of images. Eye movement parameters, such as fixation (the process where the eye remains stationary on a specific point to gather and process visual information), saccades (the rapid eye movements that shift the focus from one point to another quickly), and scanpath (sequence of fixation and saccadic that an individual follows while observing a visual scene), present valuable insights into diagnostic processes and expert decision-making relevant to training [14]. Eye-tracking data influences image interpretation, analysis, and diagnosis. Many studies have discussed utilizing eye-gaze data across diverse applications, including (i) Classification: replicating different search strategies and exploring their potential to improve model performance and interpretability across various modalities [15,16,17]. (ii) Data annotation: to highlight the expert’s search patterns while labeling medical data, including images, text, and genetic information [18,19]. (iii) Object Detection and Segmentation: providing a unique form of supervision format for training DL/ML-based approaches and identifying the precise location of the object [20]. (iv) Workload and Skill: focuses on understanding human factors in medical image interpretation and diagnosis, including the effects of fatigue and skill level on diagnostic accuracy [21].

2.2. Multimodal in Medical Data

Multimodal learning in medical applications accelerated in the mid-2010s when integration of various clinical data with multiple types of radiological imaging (e.g., CT scans and MRI) was shown to improve diagnostic precision. Early studies by Li et al. [22] exemplified this paradigm shift, utilizing multiple data sources to provide a more comprehensive analysis. More recently, various medical imaging applications have emerged to integrate eye-gaze tracking data into stimulus analysis. With an increasing focus on multimodal deep learning models, the incorporation of eye-gaze tracking enables a deeper understanding of human search patterns, thereby enhancing the model’s ability to analyze visual information effectively [23]. Drew [24] noted that experts rapidly identify potential lesions, utilize a wider functional field of vision, and draw on more conceptual knowledge than novices when detecting abnormalities. Ma et al. [25] utilized radiologists’ visual attention maps to guide the model in concentrating on task-related objects or features rather than taking harmful shortcuts.
Furthermore, Wang et al. [12] indicate that radiologists’ gaze patterns are strongly linked to diagnostic accuracy in the detection of mammogram lesions. They propose an attention-aware augmentation method that consistently improves focusing contrast and highlights the importance of where radiologists focus, thereby reducing errors in the evaluation phase. Hsieh et al. [26] present a parallel framework that processes chest X-ray images and expert fixation masks through Convolutional Neural Network (CNN) models. This multimodal approach predicts abnormality classes, refines bounding boxes, and optionally generates binary masks for more precise lesion localization. Although analyzing the CXR images and heatmap can cause noise in the model, we discuss in Section 3 the Proposed Model that our approach can mitigate this issue.

2.3. Explainable Artificial Intelligence

Several studies have discussed Explainable Artificial Intelligence (XAI) approaches and presented general frameworks for XAI [27,28,29,30,31]. This section discusses studies that develop XAI methods using multimodal data that incorporate eye-tracking information. Currently, only a limited number of studies within the context of multimodal learning have explored explainable systems, despite the recognized significance of XAI systems. In medical imaging, various explanation approaches are visual-based explanations [32]. The core idea of these approaches is to present the information maintained through the model to analyze image regions that contribute most to its predictions. Generally, these methods present attribution maps, which can be used to diagnose a supportive and transparent system. These saliency-based explanations can be categorized into perturbation-based, activation-based, and backpropagation-based techniques [33].
Perturbation-Based Techniques: These methods assess the importance of each input by modifying the image and observing the effect on the model’s output. Perturbation techniques can be applied broadly to classification and regression tasks if the distance between model outputs can be computed. These methods focus on understanding how changes to input features affect the neural activity and predictions of the model.
Activation-Based Techniques: These techniques leverage the feature maps generated in the last layer of a CNN network to explain the model’s predictions. By weighting each feature map and summing them, these methods create class activation maps (CAMs), which highlight areas of the image corresponding to the predicted class. The final activation map is upscaled to the original image’s size, often resulting in a coarse resolution. CAMs help localize regions of interest by detecting neuronal activity associated with specific classes of neurons. Shallow CAMs capture finer details, while deeper CAMs identify broader areas of objects.
Backpropagation-Based Techniques: These methods propagate the output gradients backward to the input image, creating a high-resolution saliency map that highlights which pixels most influence the model’s prediction. Gradients indicate the extent to which each pixel influences the final decision, either increasing or decreasing the predicted class score. These techniques focus on the impact of individual pixels and how changes affect neuron activity.

3. The Proposed Model

This work introduces an eye-guided multimodal fusion with shared parameters to facilitate interactive exploration and visualization techniques. Our system enables us to use the collected knowledge base on the expert’s fixation map to quickly navigate to reading stimuli while receiving feedback. As shown in Figure 2, the framework of our proposed method consists of three main components. Firstly, after preprocessing raw CXR and the expert’s fixation heatmap, the fusion of feature maps from both modalities enables the model to learn correlations between the eye-tracking heatmaps’ spatial patterns and the X-ray images’ visual features. Secondly, this framework features a unified core between the CXR and fixation map to mitigate the impact of noisy fixation data and prevent the need for separate processing of modalities, which assumes that both sources contribute independently to improved abnormality detection in CXRs. Finally, we evaluate the model’s explanation interpretability through Grad-CAM, its usability, and its impact on radiologist decision-making.

3.1. Multimodal Input Data

The input image used in this framework consists of the CXR images. After collecting the raw eye gaze, we generate a related expert’s fixation heatmap using the coordinated points. The eye-tracking heatmap data captures the CXR regions that received the most attention from the radiologist. Then, we applied preprocessing, such as resizing, cropping, and augmentation, including a Gaussian blur to provide a slight blur and avoid overfitting, and color jitter as a brightness and contrast augmentation.
Input data represented as:
I C X R R W × H
H E y e R W × H
where W and H represent the width and height of the image, the single channel corresponds to the grayscale chest X-ray data, and the three channels correspond to the RGB in heatmap overlay data.

3.2. Mode

Multimodal Data Processing: The model initially constructs a multichannel input by stacking the chest X-ray images and eye-tracking data along the channel dimension. This early fusion of feature maps from both modalities enables the model to capture and learn correlations between the spatial patterns of the eye-tracking heatmaps and the visual features of the X-ray images. This fusion allows the model to process both data types simultaneously.
I M u l t i = Concat I C X R , H E y e R W × H
Attention Heads: The attention mechanism enables the model to focus on the most relevant regions of the input, acting as a filter and aligning the network with human decision-making, potentially improving performance. It processes the input, applies attention scores, and outputs the weighted feature maps. The attention head consists of two 1 × 1 convolutional layer. The first convolution reduces the feature map dimensions, applying ReLU for non-linearity, and the second convolution computes attention scores, followed by a sigmoid activation to normalize the attention weights to the [0, 1] range, and finally, consider adding a dropout layer after attention-scores, as this can help prevent overfitting. The output is element-wise multiplied by the input to generate attended features. The attention head applies the following operations:
F conv 1 = ReLU W 1     F input + b 1
where W1 and b1 are the weights and biases of the first convolutional layer, and ∗ denotes convolution. The second layer computes attention scores, followed by a sigmoid activation to constrain the attention weights to the range [0, 1]:
A weights = σ W 2     F conv 1 + b 2
W2 and b2 are the weights and biases of the second convolutional layer, representing the sigmoid function. Finally, the attended features are computed by element-wise multiplying the input feature maps with the attention weights. The attention weights A (weights) are element-wise multiplied by the original input feature map F (input) to produce the attended feature map:
F attended = F input A weights
Unified Backbone: We use a convolutional neural network model as a shared backbone network. The shared backbone module processes the attended features extracted from the attention head and outputs deep, high-level feature maps. These feature maps are then passed to the fully connected layer, which predicts six labels in the classification task. Since each image can have multiple labels, the fully connected (FC) layer produces independent predictions for each of the six labels. The attended input F attention from the attention head is passed through the ResNet-50 backbone, which creates the high-level feature representation, denoted as:
F backbone = BackboneModel F attention
where F attention R W × H × 4   is the multimodal input (combining the X-ray and eye-tracking heatmap) and the backbone output, F backbone R 2048 , which contains the high-level extracted features.
Fully Connected Layer: The final fully connected layer (FC) of the CNN backbone maps the high-dimensional feature vector F backbone into six output logits for multilabel classification:
y pred = W fc F backbone + b fc
where σ is the sigmoid function that produces a probability for each class. The network then assigns a label to each class based on a predefined threshold (e.g., 0.5). Each of the six labels in the multilabel classification task is converted into a probability using the sigmoid activation function [34]:
σ x = 1 1 + e x   ,
where
P   c l a s s = σ y pred

3.3. Explanation

In the first step, the heatmap generation process involves capturing activations and gradients from the convolutional layer of the backbone model (here, we used the fourth layer). In this study, we employ Grad-CAM as the Post-Hoc XAI component of our framework. Specifically, Grad-CAM provides model interpretability after training by visually highlighting the image regions that most influenced the model’s decisions. A forward hook is registered on the target layer to store the output activations and gradients. During the backward pass, gradients are accumulated with respect to a specific target class, enabling focused visualization. The captured gradients y c A i j k are pooled across spatial dimensions by averaging:
α k c = 1 Z i j y c A i j k
Here y c is the score for the target class c , A i j k ,   w h i c h refers to the activation at spatial location i , j for the k t h   feature map, and Z is the total number of spatial locations i . e . , i × j . These pooled gradients α k c are used to weight the corresponding activations.
The weighted activations are averaged across channels to generate the initial heatmap using the Grad-CAM method [35].
L Grad - CAM c = ReLU k α k c A k
A ReLU operation ensures that only positive values are retained, highlighting the relevant areas of interest. In the next step, we normalized and vertically flipped the image to correct orientation issues and match the original image dimensions, thereby building consistent scaling for the heatmaps. Finally, we converted it to an RGB format using a colormap for visual clarity and generated an interpretable visual output.

4. Experiments and Results

In this section, we discuss the experimental phase of our research and provide an overview of the key components that contribute to the study. We conclude with a comprehensive discussion that synthesizes our findings and offers a detailed interpretation of the results, aiming to provide a nuanced understanding of our research outcomes.

4.1. Dataset

Our experimental validation utilized the REFLACX dataset (Reports and Eye-tracking Data for Localization of Abnormalities in CXR) [36], which is derived from the MIMIC-CXR dataset [37]. Eye-tracking data for this dataset was collected using an Eyelink 1000 Plus system (SR Research, Oakville, ON, Canada) at a 1000 Hz resolution. The system tracked the radiologists’ pupil positions, and the fixation data were synchronized with timestamps from the dictations. Five radiologists provided manual labels for abnormalities and drew ellipses around localized findings over three phases. In the initial testing phase, radiologists reviewed a shared set of 59 CXRs. In the refinement phase, instructions were provided to standardize the labeling process, improving clarity and reliability. It also sets the stage for larger-scale data collection. Eventually, in the primary data collection phase, each radiologist independently reviewed around 500 CXRs. The final dataset was created, which contained eye-tracking data, transcription data, and manual annotations.
To validate the approach, we focus on the most frequently occurring lesions in this dataset, including pleural abnormalities, consolidation, pulmonary edema, enlarged cardiac silhouette, atelectasis, and X-rays showing no specific disease findings. The pleural abnormality represents an abnormal condition of the pleura, the thin tissue that lines the chest cavity and surrounds the lungs. Pulmonary edema is a condition characterized by the accumulation of excess fluid in the lungs. The fluid accumulates in the air sacs, making it difficult to breathe. An enlarged cardiac silhouette is evident when the heart appears more prominent than usual on imaging tests, such as a chest X-ray. Atelectasis is a partial or complete collapse of one or both lungs, which can lead to shortness of breath and difficulty breathing. Finally, consolidation is filling alveolar airspaces with fluid (exudate, transudate, or blood), inflammatory cells, tissue, or other materials. Figure 3 represents CXRs for selected pulmonary conditions analyzed in our research, and Figure 4 displays the distribution of these selected lesion cases across different lung conditions. These visual examples, shown in Figure 3, illustrate the typical radiological appearances, while Figure 4 quantifies their occurrence across the dataset. All data were sourced from the REFLACX dataset, which we describe in detail in Section 4.1.

4.2. Implementation Details

In this study, we tuned our model for 20 epochs. This value was selected based on an analysis of the training and validation loss curves, which showed that after 20 epochs, the model began to overfit, as indicated by an increasing validation loss while the training loss continued to decline. To prevent overfitting and ensure optimal generalization, we stopped training at this point. The loss curves for the proposed Eye-Guided Multimodal Fusion model are illustrated in Figure 5. We employed an initial learning rate of 5 × 10−5 and a weight decay of 1 × 10−3. The Adam optimizer was utilized with a batch size of 32 for optimization. The images were cropped and resized to 224 × 224 pixels, aligning with the input requirements of the ResNet-50 architecture. The experiments leveraged open-source model weights pre-trained on ImageNet, which were then fine-tuned on the REFLACX datasets. The experimental setup involved computational resources from a local environment, where we trained and tested the model using version 2.4.1 of the PyTorch framework. The training was conducted on an internal server with an NVIDIA RTX 6000 GPU and an Intel(R) Xeon(R) w7-3465X CPU. Initially, we divided the dataset into training, validation, and test sets using an 80/10/10 split, consistent with our division strategy. A random seed of 42 was used to guarantee reproducibility of the results.

4.3. Evaluation Results

This section presents the evaluations performed and the corresponding results. The Eye-Gaze Guided Fusion System was evaluated through a two-phase assessment, comprising both model performance analysis and interpretability assessment using Grad-CAM. Clinical experts subsequently reviewed these evaluations to validate the system’s clinical relevance.
Multimodal Fusion Performance: The effectiveness of the Eye-Guided Fusion System was evaluated by comparing its performance across two configurations: (1) using only CXR images, and (2) combining CXR images with fixation maps as input. The comparison focused on the system’s ability to detect abnormalities, as measured by key metrics including accuracy, precision, recall, and F1-score. To assess the system’s practical reliability, robustness evaluations were conducted under varying conditions by introducing artificial noise into the fixation data, simulating potential misalignments or noise typical in real-world applications. This analysis aimed to determine the system’s capacity to maintain performance despite noisy or degraded inputs, reflecting its robustness for clinical deployment.
Ablation Study on Modality Contribution: The system’s performance, integrating CXR and fixation map modalities, was assessed through an ablation study using ResNet-18 [38] and ResNet-50 [39]. This study compared the effectiveness of various input combinations. The results, summarized in Table 1, indicate that utilizing both modalities in the average eye-guided framework improved metrics. This table presents the results, including per-class metrics for all six findings, as well as the average accuracy scores across all classes and metrics.
As shown in the confusion matrices in Figure 6, for most categories, such as Pulmonary Edema, Atelectasis, Consolidation, and Pleural Abnormality, the number of true positives is substantially higher than the number of false positives and false negatives. Additionally, categories like Pulmonary Edema and Enlarged Cardiac Silhouette exhibit few false positives, indicating the model’s precision and ability to avoid overpredictions. Furthermore, the low false negative rate in these categories suggests that the model effectively captures true positives.
Noise Robustness: Artificial noise was introduced into both models to measure how well the system and the explanation component handle noisy or misaligned data. In real-world settings, eye-tracking devices may not perfectly align with the displayed content or the target region of interest (e.g., an X-ray). Minor calibration errors, head movements, or device drift can cause slight shifts in fixation points. In practical applications, eye-tracking data often encounters challenges such as slight misalignment due to device calibration errors, minor head movements, or sensor drift. In this context, we used striped line noise at 10% and 50% levels. A sample of this alignment is shown in Figure 7. This alignment helped test the model’s robustness to minor misalignments that may occur in practice. As shown in Table 2, the eye-guided explanation system demonstrated better noise tolerance results than models without shared parameters, maintaining AUC values at increasing noise levels during testing.
Grad-CAM Explanation Evaluation: In this section, we will present the Intersection Over Union (IoU) performance evaluation of the explanation generated using the Grad-CAM.
Explanation Quality: The Grad-CAM heatmaps were compared to expert-labeled ROI and fixation maps. The overlap between the Grad-CAM activations and experts’ ROIs was measured in multiple classes using the mean Intersection over Union (mIoU) metric, which is defined as:
mIoU = 1 N i = 1 N IoU i  
where N is the total number of classes, and IoU i is the Intersection over Union for class i. The results are illustrated in Table 3.
A radiologist with over 12 years of experience in interpreting radiological imaging evaluated the interpretability and clinical relevance of the Grad-CAM explanations. Evaluations were conducted using a 5-point Likert scale, focusing on clarity, clinical applicability, and diagnostic utility. The radiologist’s repeated assessments across various imaging tasks provided quantitative ratings and in-depth qualitative feedback, highlighting the system’s strengths and limitations in real-world settings. As the radiologist’s familiarity with the eye-guided system increased, there was a marked enhancement in workflow efficiency, evidenced by improvements in diagnostic speed, accuracy, and confidence. Notably, the analysis revealed a progressive alignment between the radiologist’s preferred regions of interest and those identified by the system, further validating its clinical utility. The results of this evaluation are summarized in Table 4.

5. Discussion

The experimental findings demonstrate the effectiveness of our Eye-Gaze Guided Multimodal Fusion framework in enhancing abnormality detection in chest X-rays by incorporating radiologists’ eye-tracking data. Compared to the baseline models that rely solely on image data, the inclusion of expert fixation maps significantly improved classification metrics across all six targeted conditions. This suggests that gaze-based guidance provides valuable prior knowledge about salient diagnostic regions, enabling the model to focus its attention more effectively. One significant finding was the framework’s robustness to noise in the gaze data. When artificial noise was introduced to the fixation maps, the system maintained high AUC scores, outperforming the baseline models that did not share parameters. This highlights the value of using a shared backbone architecture, which helps mitigate the influence of noisy or slightly misaligned input data. This property is critical for real-world clinical deployment, where perfect eye-tracking alignment cannot always be guaranteed. Furthermore, the Grad-CAM explanation module demonstrated strong alignment between the model’s visual focus and expert-annotated ROIs. Both quantitative evaluations using IoU and qualitative feedback from experienced radiologists confirmed that the model’s attention maps were consistent with expert diagnostic reasoning. This not only supports the reliability of the system but also enhances its interpretability, a key requirement for the adoption of AI tools in clinical practice. Compared to prior studies in multimodal medical AI, our approach provides a unified and interpretable architecture that leverages human gaze as a form of weak supervision. This approach avoids the need for extensive pixel-level annotation, reflecting how clinicians naturally interpret images. However, the model still has limitations. The fixation data used in training are limited to a small number of radiologists, and only one expert reviews each image.

6. Conclusions

In this study, we introduced an Eye-Gaze Guided Multimodal Fusion framework that incorporates radiologists’ eye-tracking data to improve the classification of abnormalities in chest X-rays. By integrating visual attention with image data, the model leverages expert domain knowledge and achieves improved accuracy across various conditions. The proposed approach enhances the interpretability of model predictions through post hoc visual explanations, supporting trust and usability in clinical workflows. This work highlights the potential of human-centered AI systems to support diagnostic imaging and radiology education, particularly in contexts where labeled data or training resources are limited.

6.1. Potential Impacts

Although the REFLACX dataset, one of the most valuable resources, used five different radiologists for data preparation, each radiologist examined approximately 20% of the CXRs based on their strategy. However, there is still no public dataset where many experts work on the same images. This issue becomes prominent when different radiologists employ varying search strategies, particularly for images with no significant findings. As we realized, this could introduce bias in these cases.

6.2. Future Work

In the future, we will continue to optimize the proposed system by exploring multimodal approaches, particularly by integrating clinical reports alongside image and gaze data. This approach could further improve the model’s understanding and interpretation of its predictions. Generating human-like explanations by simulating how an expert might describe their focus while analyzing an image could also help in correcting biases. Meanwhile, developing a visual dashboard that showcases a real-time, interactive system will illustrate how radiologists can leverage the fusion of eye-tracking and X-ray data to improve diagnosis.

Author Contributions

Conceptualization, S.M.; Methodology, S.M. and A.D.I.; Validation, S.M., A.H., S.L., Y.Q. and A.D.I.; Formal analysis, S.M. and A.H.; Investigation, S.M.; Writing—original draft, S.M. and A.H.; Writing—review & editing, S.L., Y.Q., A.B. and A.D.I.; Visualization, S.M.; Supervision, A.B. and A.D.I.; Project administration, A.D.I. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Ethical review and approval were waived for this study because all analyses were performed on a de-identified, publicly available dataset. REFLACX (Published: 27 September 2021, Version: 1.0.0). No new human or animal subjects were recruited, and no identifiable information was accessed.

Informed Consent Statement

Not applicable. The source datasets were released with patient identifiers removed, and the original data custodians obtained the necessary consent waivers. The present study involved only secondary analysis of the de-identified data.

Data Availability Statement

We utilized the REFLACX dataset, available at Physionet (https://physionet.org/content/reflacx-xray-localization/1.0.0/). Published: 27 September 2021, Version: 1.0.0.

Acknowledgments

We sincerely thank all members of the Computational NeuroSurgery (CNS) Lab and the Centre for Applied AI at Macquarie University for their invaluable suggestions and support.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Holmqvist, K.; Nystr, M.; Andersson, R.; Dewhurst, R.; Jarodzka, H.; Van de Weijer, J. Eye Tracking: A Comprehensive Guide to Methods and Measures; OUP: Oxford, UK, 2011. [Google Scholar]
  2. O’Brien, H.L.; Cairns, P.; Hall, M. A practical approach to measuring user engagement with the refined user engagement scale (ues) and new ues short form. Int. J. Hum.-Comput. Stud. 2018, 112, 28–39. [Google Scholar] [CrossRef]
  3. Moreira, C.; Nobre, I.B.; Sousa, S.C.; Pereira, J.M.; Jorge, J. Improving x-ray diagnostics through eye-tracking and xr. In Proceedings of the 2022 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW), Christchurch, New Zealand, 12–16 March 2022; pp. 450–453. [Google Scholar]
  4. Chen, X.; Wang, X.; Zhang, K.; Fung, K.-M.; Thai, T.C.; Moore, K.; Mannel, R.S.; Liu, H.; Zheng, B.; Qiu, Y. Recent advances and clinical applications of deep learning in medical image analysis. Med. Image Anal. 2022, 79, 102444. [Google Scholar] [CrossRef] [PubMed]
  5. Borys, K.; Schmitt, Y.A.; Nauta, M.; Seifert, C.; Kr, N.; Friedrich, C.M.; Nensa, F. Explainable ai in medical imaging: An overview for clinical practitioners–beyond saliency-based xai approaches. Eur. J. Radiol. 2023, 162, 110786. [Google Scholar] [CrossRef] [PubMed]
  6. Mohajir, B.E.E. Identifying learning style through eye tracking technology in adaptive learning systems. Int. J. Electr. Comput. Eng. 2019, 9, 4408–4416. [Google Scholar]
  7. Castner, N.; Eivazi, S.; Scheiter, K.; Kasneci, E. Using Eye Tracking to Evaluate and Develop Innovative Teaching Strategies for Fostering Image Reading Skills of Novices in Medical Training. Eye Tracking Enhanced Learning (ETEL2017). 2017. Available online: https://www.hci.uni-tuebingen.de/assets/pdf/publications/NSKE092017.pdf (accessed on 20 July 2025).
  8. Kok, E.M.; Jarodzka, H. Before your very eyes: The value and limitations of eye tracking in medical education. Med. Educ. 2017, 51, 114–122. [Google Scholar] [CrossRef] [PubMed]
  9. Ashraf, H.; Sodergren, M.H.; Merali, N.; Mylonas, G.; Singh, H.; Darzi, A. Eye-tracking technology in medical education: A systematic review. Med. Teach. 2018, 40, 62–69. [Google Scholar] [CrossRef] [PubMed]
  10. Newport, R.A.; Liu, S.; Di Ieva, A. Analyzing eye paths using fractals. In The Fractal Geometry of the Brain; Springer: Berlin/Heidelberg, Germany, 2024; pp. 827–848. [Google Scholar]
  11. Duchowski, A.T. A breadth-first survey of eye-tracking applications. Behav. Res. Methods Instrum. Comput. 2002, 34, 455–470. [Google Scholar] [CrossRef] [PubMed]
  12. Wang, S.; Zhuang, Z.; Ouyang, X.; Zhang, L.; Li, Z.; Ma, C.; Liu, T.; Shen, D.; Wang, Q. Learning better contrastive view from radiologist’s gaze. arXiv 2023, arXiv:2305.08826. [Google Scholar] [CrossRef]
  13. Ma, C.; Jiang, H.; Chen, W.; Li, Y.; Wu, Z.; Yu, X.; Liu, Z.; Guo, L.; Zhu, D.; Zhang, T.; et al. Eye-gaze guided multimodal alignment for medical representation learning. In Proceedings of the The Thirty-Eighth Annual Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 9–15 December 2024. [Google Scholar]
  14. Moradizeyveh, S.; Tabassum, M.; Liu, S.; Newport, R.A.; Beheshti, A.; Di Ieva, A. When eye-tracking meets machine learning: A systematic review on applications in medical image analysis. arXiv 2024, arXiv:2403.07834. [Google Scholar] [CrossRef]
  15. Zhu, H.; Salcudean, S.; Rohling, R. Gaze-guided class activation mapping: Leverage human visual attention for network attention in chest x-rays classification. In Proceedings of the 15th International Symposium on Visual Information Communication and Interaction, Chur, Switzerland, 16–18 August 2022; pp. 1–8. [Google Scholar]
  16. Ji, C.; Du, C.; Zhang, Q.; Wang, S.; Ma, C.; Xie, J.; Zhou, Y.; He, H.; Shen, D. Mammo-net: Integrating gaze supervision and interactive information in multi-view mammogram classification. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Berlin/Heidelberg, Germany, 2023; pp. 68–78. [Google Scholar]
  17. Zhu, H.; Rohling, R.; Salcudean, S. Jointly boosting saliency prediction and disease classification on chest x-ray images with multi-task unet. In Annual Conference on Medical Image Understanding and Analysis; Springer: Berlin/Heidelberg, Germany, 2022; pp. 594–608. [Google Scholar]
  18. Teng, C.; Lee, L.H.; Lander, J.; Drukker, L.; Papageorghiou, A.T.; Noble, J.A. Skill characterisation of sonographer gaze patterns during second trimester clinical fetal ultrasounds using time curves. In Proceedings of the 2022 Symposium on Eye Tracking Research and Applications, Seattle, WA, USA, 8–11 June 2022; pp. 1–7. [Google Scholar]
  19. Mariam, K.; Afzal, O.M.; Hussain, W.; Javed, M.U.; Kiyani, A.; Rajpoot, N.; Khurram, S.A.; Khan, H.A. On smart gaze-based annotation of histopathology images for training of deep convolutional neural networks. IEEE J. Biomed. Health Inform. 2022, 26, 3025–3036. [Google Scholar] [CrossRef] [PubMed]
  20. Stember, J.N.; Celik, H.; Gutman, D.; Swinburne, N.; Young, R.; Eskreis-Winkler, S.; Holodny, A.; Jambawalikar, S.; Wood, B.J.; Chang, P.D.; et al. Integrating eye tracking and speech recognition accurately annotates mr brain images for deep learning: Proof of principle. Radiol. Artif. Intell. 2020, 3, e200047. [Google Scholar] [CrossRef] [PubMed]
  21. Pershin, I.; Mustafaev, T.; Ibragimova, D.; Ibragimov, B. Changes in radiologists’ gaze patterns against lung x-rays with different abnormalities: A randomized experiment. J. Digit. Imaging 2023, 36, 767–775. [Google Scholar] [CrossRef] [PubMed]
  22. Li, W.; Jia, F.; Hu, Q. Automatic segmentation of liver tumor in ct images with deep convolutional neural networks. J. Comput. Commun. 2015, 3, 146–151. [Google Scholar] [CrossRef]
  23. Peng, P.; Fan, W.; Shen, Y.; Liu, W.; Yang, X.; Zhang, Q.; Wei, X.; Zhou, D. Eye gaze guided cross-modal alignment network for radiology report generation. IEEE J. Biomed. Health Inform. 2024, 28, 7406–7419. [Google Scholar] [CrossRef] [PubMed]
  24. Drew, T.; Evans, K.; Võ, M.L.-H.; Jacobson, F.L.; Wolfe, J.M. Informatics in radiology: What can you see in a single glance and how might this guide visual search in medical images? Radiographics 2013, 33, 263–274. [Google Scholar] [CrossRef] [PubMed]
  25. Ma, C.; Zhao, L.; Chen, Y.; Wang, S.; Guo, L.; Zhang, T.; Shen, D.; Jiang, X.; Liu, T. Eye-gaze-guided vision transformer for rectifying shortcut learning. IEEE Trans. Med. Imaging 2023, 42, 3384–3394. [Google Scholar] [CrossRef] [PubMed]
  26. Hsieh, C.; Luís, A.; Neves, J.; Nobre, I.B.; Sousa, S.C.; Ouyang, C.; Jorge, J.; Moreira, C. Eyexnet: Enhancing abnormality detection and diagnosis via eye-tracking and x-ray fusion. Mach. Learn. Knowl. Extr. 2024, 6, 1055–1071. [Google Scholar] [CrossRef]
  27. Adadi, A.; Berrada, M. Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI). IEEE Access 2018, 6, 52138–52160. [Google Scholar] [CrossRef]
  28. Carvalho, D.V.; Pereira, E.M.; Cardoso, J.S. Machine Learning Interpretability: A Survey on Methods and Metrics. Electronics 2019, 8, 832. [Google Scholar] [CrossRef]
  29. Guidotti, R.; Monreale, A.; Ruggieri, S.; Turini, F.; Pedreschi, D.; Giannotti, F. A Survey of Methods For Explaining Black Box Models. ACM Comput. Surv. 2018, 51, 1–42. [Google Scholar] [CrossRef]
  30. Miller, T. Explanation in artificial intelligence: Insights from the social sciences. Artif. Intell. 2019, 267, 1–38. [Google Scholar] [CrossRef]
  31. Tjoa, E.; Guan, C. A survey on explainable artificial intelligence (xai): Toward medical xai. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 4793–4813. [Google Scholar] [CrossRef] [PubMed]
  32. Van der Velden, B.H.; Kuijf, H.J.; Gilhuijs, K.G.; Viergever, M.A. Explainable artificial intelligence (xai) in deep learning-based medical image analysis. Med. Image Anal. 2022, 79, 102470. [Google Scholar] [CrossRef] [PubMed]
  33. Gomez, T.; Mouch, H. Computing and evaluating saliency maps for image classification: A tutorial. J. Electron. Imaging 2023, 32, 020801. [Google Scholar] [CrossRef]
  34. Rumelhart, D.E.; Hinton, G.E.; Williams, R.J. Learning representations by back-propagating errors. Nature 1986, 323, 533–536. [Google Scholar] [CrossRef]
  35. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-cam: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
  36. Bigolin Lanfredi, R.; Zhang, M.; Auffermann, W.F.; Chan, J.; Duong, P.A.T.; Srikumar, V.; Drew, T.; Schroeder, J.D.; Tasdizen, T. REFLACX, a dataset of reports and eye-tracking data for localization of abnormalities in chest x-rays. Sci. Data 2022, 9, 350. [Google Scholar] [CrossRef] [PubMed]
  37. Johnson, A.E.; Pollard, T.J.; Greenbaum, N.R.; Lungren, M.P.; Deng, C.Y.; Peng, Y.; Lu, Z.; Mark, R.G.; Berkowitz, S.J.; Horng, S. MIMIC-CXR-JPG, a large publicly available database of labeled chest radiographs. arXiv 2019, arXiv:1901.07042. [Google Scholar]
  38. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  39. Koonce, B. ResNet 50. In Convolutional Neural Networks with Swift for Tensorflow: Image Recognition and Dataset Categorization; Apress: Berkeley, CA, USA, 2021; pp. 63–72. [Google Scholar]
Figure 1. The visual abstract of the eye-guided framework includes the inputs, model, explanation, and evaluation phases.
Figure 1. The visual abstract of the eye-guided framework includes the inputs, model, explanation, and evaluation phases.
Sensors 25 04575 g001
Figure 2. An eye-guided architecture for multimodal learning to enhance abnormality detection in chest X-rays by incorporating radiologists’ eye-tracking data as additional input. The framework combines raw chest X-rays and a heatmap derived from eye fixation data (where blue denotes low fixation density and red denotes the highest concentration of fixations), creating a multimodal input that is processed through an attention head to guide the model’s focus. The unified backbone processes this guided input, followed by a fully connected layer and a sigmoid activation for multilabel lesion classification. The Explainable AI module, based on Gradient-weighted Class Activation Mapping (Grad-CAM), provides post hoc visual explanations. Specifically, after model training, Grad-CAM highlights the critical regions within the chest X-rays that influenced the model’s predictions, enhancing interpretability and transparency for clinical users.
Figure 2. An eye-guided architecture for multimodal learning to enhance abnormality detection in chest X-rays by incorporating radiologists’ eye-tracking data as additional input. The framework combines raw chest X-rays and a heatmap derived from eye fixation data (where blue denotes low fixation density and red denotes the highest concentration of fixations), creating a multimodal input that is processed through an attention head to guide the model’s focus. The unified backbone processes this guided input, followed by a fully connected layer and a sigmoid activation for multilabel lesion classification. The Explainable AI module, based on Gradient-weighted Class Activation Mapping (Grad-CAM), provides post hoc visual explanations. Specifically, after model training, Grad-CAM highlights the critical regions within the chest X-rays that influenced the model’s predictions, enhancing interpretability and transparency for clinical users.
Sensors 25 04575 g002
Figure 3. Representative Chest X-ray Images for Different Pulmonary Conditions.
Figure 3. Representative Chest X-ray Images for Different Pulmonary Conditions.
Sensors 25 04575 g003
Figure 4. Distribution of the selected most frequently occurring lesion cases across different lung conditions.
Figure 4. Distribution of the selected most frequently occurring lesion cases across different lung conditions.
Sensors 25 04575 g004
Figure 5. Training and validation loss curves for the proposed model (CXR + Fixation Map). The vertical dashed line marks the point where overfitting begins, approximately after epoch 20.
Figure 5. Training and validation loss curves for the proposed model (CXR + Fixation Map). The vertical dashed line marks the point where overfitting begins, approximately after epoch 20.
Sensors 25 04575 g005
Figure 6. Confusion matrices for multilabel classification across the medical conditions.
Figure 6. Confusion matrices for multilabel classification across the medical conditions.
Sensors 25 04575 g006
Figure 7. Effect of Striped Noise on Fixation Map with Varying Noise Levels and Stripe Thickness. The blue denotes low fixation density and red denotes the highest concentration of fixations.
Figure 7. Effect of Striped Noise on Fixation Map with Varying Noise Levels and Stripe Thickness. The blue denotes low fixation density and red denotes the highest concentration of fixations.
Sensors 25 04575 g007
Table 1. Ablation study on the efficacy comparison of different modalities. This comparison of per-class and average performance between ResNet18 and ResNet50 models, using chest X-ray (CXR) images alone and in combination with expert fixation maps. Metrics reported include accuracy (ACC), precision, recall, and F1 score for six clinically relevant categories. The average (AVG) across all classes is provided to summarize overall model performance. This table addresses class-level variations and highlights the impact of incorporating gaze data under conditions of class imbalance.
Table 1. Ablation study on the efficacy comparison of different modalities. This comparison of per-class and average performance between ResNet18 and ResNet50 models, using chest X-ray (CXR) images alone and in combination with expert fixation maps. Metrics reported include accuracy (ACC), precision, recall, and F1 score for six clinically relevant categories. The average (AVG) across all classes is provided to summarize overall model performance. This table addresses class-level variations and highlights the impact of incorporating gaze data under conditions of class imbalance.
Pulmonary ConditionsCXR OnlyCXR + Fixation Map
ACCPrecisionRecallF1ACCPrecisionRecallF1
Model: ResNet18
Atelectasis0.7240.5750.5340.5540.7490.5270.5840.554
Consolidation0.7390.6630.5870.6230.7820.6190.6090.614
Pleural Abnormality0.8120.4250.4960.4580.8010.4570.4960.476
Pulmonary Edema0.7570.4540.4850.4690.7940.3510.4850.407
No-Finding0.8350.4000.5350.4580.8060.4070.5350.462
Enlarged Cardiac Silhouette0.7960.6300.5670.5970.8260.7840.5670.658
AVG0.7770.5250.5340.5270.7900.5240.5460.529
Model: ResNet50
Atelectasis0.7680.4670.5640.5110.7690.5610.5810.571
Consolidation0.7810.5250.5250.5250.7880.5790.5850.582
Pleural Abnormality0.8180.4830.5220.5020.8530.5510.6180.583
Pulmonary Edema0.7930.5260.5420.5340.8280.6870.5920.636
No-Finding0.8340.6840.5360.6010.8470.7190.6430.679
Enlarged Cardiac Silhouette0.8160.5230.5630.5420.8490.6370.6170.627
AVG0.8010.5290.5420.5350.8240.6200.6060.613
Table 2. Noise robustness of the gaze-guided system.
Table 2. Noise robustness of the gaze-guided system.
Noise Level (%)CXR AUCCXR+ Fixation Map AUC (No Shared Params)
081.4183.30
1080.2881.08
5074.6175.86
Table 3. Grad-CAM explanation quality (IoU with expert ROI).
Table 3. Grad-CAM explanation quality (IoU with expert ROI).
ModelIoU Score (Mean ± Std. Dev.)
CXR Only 0.56 ± 0.16
CXR + FixationMap 0.61 ± 0.05
Table 4. Expert review of the Grad-CAM explanation.
Table 4. Expert review of the Grad-CAM explanation.
CriteriaAverage Rating (1 to 5)
Interpretability4.25
Clinical Relevance4.00
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Moradizeyveh, S.; Hanif, A.; Liu, S.; Qi, Y.; Beheshti, A.; Di Ieva, A. Eye-Guided Multimodal Fusion: Toward an Adaptive Learning Framework Using Explainable Artificial Intelligence. Sensors 2025, 25, 4575. https://doi.org/10.3390/s25154575

AMA Style

Moradizeyveh S, Hanif A, Liu S, Qi Y, Beheshti A, Di Ieva A. Eye-Guided Multimodal Fusion: Toward an Adaptive Learning Framework Using Explainable Artificial Intelligence. Sensors. 2025; 25(15):4575. https://doi.org/10.3390/s25154575

Chicago/Turabian Style

Moradizeyveh, Sahar, Ambreen Hanif, Sidong Liu, Yuankai Qi, Amin Beheshti, and Antonio Di Ieva. 2025. "Eye-Guided Multimodal Fusion: Toward an Adaptive Learning Framework Using Explainable Artificial Intelligence" Sensors 25, no. 15: 4575. https://doi.org/10.3390/s25154575

APA Style

Moradizeyveh, S., Hanif, A., Liu, S., Qi, Y., Beheshti, A., & Di Ieva, A. (2025). Eye-Guided Multimodal Fusion: Toward an Adaptive Learning Framework Using Explainable Artificial Intelligence. Sensors, 25(15), 4575. https://doi.org/10.3390/s25154575

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop