Next Article in Journal
Peripheral Nerve Ultrasound Findings in Hereditary Transthyretin Amyloidosis in Brazil
Previous Article in Journal
Is Digital Maxillary Model Scanning Reliable in Individuals with Unilateral Cleft Lip and Palate?
Previous Article in Special Issue
Deep Learning-Based Aortic Diameter Measurement in Traumatic Hemorrhage Using Shallow Attention Network: A Path Forward
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Balanced Multimodal Multi-Task Deep Learning Framework for Robust Patient-Specific Quality Assurance

1
School of Computer Science and Engineering, University of Electronic Science and Technology of China—UESTC, Chengdu 611731, China
2
School of Computer Science, China West Normal University, Nanchong 637009, China
3
Institute of Computer Science, Shah Abdul Latif University, Khairpur 66111, Pakistan
*
Author to whom correspondence should be addressed.
Diagnostics 2025, 15(20), 2555; https://doi.org/10.3390/diagnostics15202555
Submission received: 9 September 2025 / Revised: 29 September 2025 / Accepted: 2 October 2025 / Published: 10 October 2025
(This article belongs to the Special Issue Deep Learning in Medical and Biomedical Image Processing)

Abstract

Background: Multimodal Deep learning has emerged as a crucial method for automated patient-specific quality assurance (PSQA) in radiotherapy research. Integrating image-based dose matrices with tabular plan complexity metrics enables more accurate prediction of quality indicators, including the Gamma Passing Rate (GPR) and dose difference (DD). However, modality imbalance remains a significant challenge, as tabular encoders often dominate training, suppressing image encoders and reducing model robustness. This issue becomes more pronounced under task heterogeneity, with GPR prediction relying more on tabular data, whereas dose difference prediction (DDP) depends heavily on image features. Methods: We propose BMMQA (Balanced Multi-modal Quality Assurance), a novel framework that achieves modality balance by adjusting modality-specific loss factors to control convergence dynamics. The framework introduces four key innovations: (1) task-specific fusion strategies (softmax-weighted attention for GPR regression and spatial cascading for DD prediction); (2) a balancing mechanism supported by Shapley values to quantify modality contributions; (3) a fast network forward mechanism for efficient computation of different modality combinations; and (4) a modality-contribution-based task weighting scheme for multi-task multimodal learning. A large-scale multimodal dataset comprising 1370 IMRT plans was curated in collaboration with Peking Union Medical College Hospital (PUMCH). Results: Experimental results demonstrate that, under the standard 2%/3 mm GPR criterion, BMMQA outperforms existing fusion baselines. Under the stricter 2%/2 mm criterion, it achieves a 15.7% reduction in mean absolute error (MAE). The framework also enhances robustness in critical failure cases (GPR < 90%) and achieves a peak SSIM of 0.964 in dose distribution prediction. Conclusions: Explicit modality balancing improves predictive accuracy and strengthens clinical trustworthiness by mitigating overreliance on a single modality. This work highlights the importance of addressing modality imbalance for building trustworthy and robust AI systems in PSQA and establishes a pioneering framework for multi-task multimodal learning.

1. Introduction

Cancer remains one of the leading causes of mortality worldwide, with radiation therapy serving as a cornerstone in its clinical management. Among advanced techniques, Intensity-Modulated Radiation Therapy (IMRT) has emerged as a pivotal technique in the clinical management of cancer, owing to its advanced capabilities in delivering precise and highly conformal dose distributions that are meticulously tailored to the tumor’s shape and location [1,2]. Nevertheless, the efficacy of IMRT fundamentally depends on the accuracy of dose delivery. This necessitates rigorous verification procedures to ensure the safe and accurate execution of each treatment plan. Patient-specific quality assurance (PSQA) fulfills this need by providing a comprehensive assessment of the treatment plan, thereby ensuring that each treatment plan can be executed both safely and accurately. In clinical practice, two key metrics are widely employed in PSQA: Gamma Passing Rate (GPR) and dose difference (DD) [3,4]. The GPR quantifies the percentage of dose points that satisfy predefined spatial and dosimetric tolerances, providing a coarse-grained measure of treatment plan fidelity. Conversely, the DD offers a detailed spatial distribution of dose discrepancies, yielding fine-grained insight into potential discrepancies between planned and delivered doses. Together, these metrics form the foundation for evaluating and improving the reliability of radiotherapy delivery.
Traditional PSQA methodologies are often labor-intensive, relying significantly on physical measurements and gamma analysis [5,6,7,8,9]. Recent advancements can be classified as follows: manually defined complexity metrics, which tend to have limited clinical relevance; (2) classical machine learning strategies that utilize tabular features, which are constrained by a dependence on low-dimensional representations; and (3) deep learning methods that exploit dose matrices, although these often neglect the advantages of multimodal fusion. Emerging multimodal frameworks seek to enhance the prediction process by integrating dose distribution images with tabular data concerning plan complexity [10,11,12,13,14,15].
Nevertheless, prevailing fusion methodologies often neglect a critical imbalance issue, wherein the tabular modality converges at a significantly swifter pace than its image counterpart due to its lower dimensionality and simpler encoder architecture. This phenomenon is referred to as the modality imbalance problem, where the faster-converging modality tends to dominate the joint training process, thereby suppressing the contributions of image-based features and resulting in suboptimal fusion and degraded performance [16]. The issue of modality imbalance in multimodal learning has been extensively explored, especially within classification tasks, where methodologies typically involve the use of an imbalance indicator (such as output logits or gradient magnitudes) alongside an adjustment strategy (including gradient scaling, adaptive sampling, or loss reweighting)  [2,17,18,19,20,21,22,23,24,25,26]. However, PSQA introduces unique challenges as a regression problem characterized by continuous outputs and multi-task requirements (e.g., Gamma Passing Rate prediction and dose difference maps). Unlike classification tasks, regression endeavors exhibit distinct optimization dynamics, rendering traditional balancing strategies less effective.
Figure 1 illustrates an introductory experiment that visualizes the convergence dynamics of different modalities. The solid lines indicate the performance of each modality when trained independently, while the dashed lines represent the performance of unimodal representations during multimodal joint training. Notably, the tabular modality converges as early as the second epoch in both scenarios but subsequently succumbs to overfitting, as evidenced by a gradual increase in loss. In contrast, the image modality continues to improve and remains unconverged even by the twentieth epoch. Importantly, the performance of each modality during joint training consistently lags behind that achieved through independent training. In this experiment, the image modality exhibits the highest performance upon convergence when trained alone, with the blue dashed line significantly surpassing its solid-line counterpart. Additionally, these results reveal a challenge related to modality preference divergence: while the tabular data primarily supports GPR prediction, the image data is more effective for fine-grained dose difference prediction. This heterogeneity complicates joint modeling but also presents an opportunity for targeted balancing.
To mitigate the above-mentioned challenges in multi-task regression for PSQA, we introduce BMMQA (Balanced Multi-modal Quality Assurance), a robust framework designed to harmonize modality- and task-specific dynamics. BMMQA integrates modality-specific learning rate adjustments to decelerate dominant modality convergence, alongside gradient normalization for stabilized joint training. Central to our approach is a hierarchical fusion mechanism: for coarse-grained GPR regression, we deploy attention-based softmax weighting to adaptively integrate modality embeddings, while spatial concatenation preserves fine-grained structures for dose difference prediction (DDP). Crucially, we derive Shapley-value-driven imbalance indicators to quantify modality contributions in continuous-output regression, enabling dynamic task-weight adjustment via gradient normalization and learning rate modulation. This dual-strategy framework not only resolves convergence asynchrony but also optimizes feature collaboration, ensuring clinically reliable predictions. With the help of Peking Union Medical College Hospital, we created a large multimodal dataset of 1370 IMRT plans. Extensive validation on it demonstrates that BMMQA achieves state-of-the-art performance, significantly reducing GPR MAE by 15.7% under 2%/2 mm criteria while attaining a peak SSIM of 0.964 for dose distribution prediction.
Our main contributions are as follows:
  • BMMQA is the first dedicated solution to modality imbalance for radiotherapy QA and multi-task multimodal learning, explicitly addressing convergence asynchrony where tabular features dominate training. A network forward mechanism enabling fast computation of outputs from different modality combinations is also proposed.
  • BMMQA introduces task-specific fusion protocols: attention-based softmax weighting for GPR regression dynamically balances modality contributions, while spatial concatenation for DDP preserves structural fidelity. This dual-path design resolves representational conflicts between global scalar prediction and fine-grained spatial mapping.
  • BMMQA develops a theoretically grounded Shapley value framework for quantifying modality contributions, tailored explicitly for continuous-output regression. This framework underpins a dynamic balancing mechanism that modulates training to prevent modality suppression.
  • BMMQA innovatively addresses the problem of multi-task multimodal learning by introducing a modality-contribution-based task weighting scheme: assigning different task-specific weights according to modality contributions to resolve the modality imbalance problem in multi-task learning.
  • Extensive experiments on clinically relevant datasets validate that BMMQA outperforms existing methods on both coarse-grained and fine-grained PSQA tasks, significantly improving modality collaboration and task generalization in real clinical settings.
The rest of the study is structured as follows: Section 2 discusses recent related work, followed by Section 3, which describes the dataset and includes relevant discussions, as well as a preliminary task. Further, in Section 4, we present the methodology in a detailed manner, whereas Section 5 provides a thorough discussion of the experimental results. Lastly, Section 6 summarizes this work, and in addition, we highlight the limitations of the current work and discuss future work.

2. Related Work

2.1. Multimodal PSQA

Multimodal deep learning frameworks have emerged to enhance PSQA by integrating dose distribution matrices (image modality) with plan complexity features (tabular modality), enabling richer and more informative predictions [10,11,12,13,14,15]. These approaches can be broadly categorized based on the stage of modality fusion. Some studies adopt early fusion models, where features from different modalities are extracted separately, concatenated, and then jointly processed by an algorithm, which is typically a machine learning model in existing studies, such as Random Forest [11] or logistic regression [13]. Other works employ a late fusion strategy using ensemble models, where each modality is trained independently to produce its own prediction. These outputs are then combined—commonly via weighted averaging, majority voting, or meta-learning—to yield the final decision. In contrast, other works [10,12] adopt intermediate fusion using end-to-end deep learning architectures, where the multimodal fusion module is jointly optimized with the encoder module during training.
For early fusion methods, existing studies typically employ traditional machine learning algorithms as the modality fusion model. Han et al. [11] incorporated a broad set of tabular features beyond MU values, including MU per control point (MU/CP), the proportion of control points with MU < 3 (%MU/CP < 3), and small segment area per control point (SA/CP). Deep features were extracted from dose matrices using a DenseNet-121 backbone. All modalities—deep, dosimetric, and plan complexity—were independently processed using dimensionality reduction techniques such as PCA and Lasso, and the resulting fused feature vector was used as input to a Random Forest (RF) model for GPR prediction. Li et al. [13] followed a similar approach, performing Lasso-based selection on dosimetric and plan complexity features, and employing logistic regression as the final prediction model. These early fusion approaches generally outperform unimodal baselines and offer relative interpretability. However, their reliance on static feature concatenation and low-dimensional input hinders fine-grained regression tasks such as DDP.
In contrast, ref. [14] proposed a late fusion approach by training two separate support vector machine models—one based on plan complexity features and the other on radiomics features—and combining their outputs to enhance beam-level GPR prediction and classification performance in SBRT (VMAT) plans. The late fusion method is typically included as a baseline in existing studies, such as in the work by Hu et al. [10], where outputs from individual modalities are aggregated at the decision level.
For intermediate fusion methods, Hu et al. [10] proposed a 3D ResNet-based architecture that encodes MLC aperture images to extract spatial features, which are then concatenated with corresponding MU values—a tabular feature representing plan complexity—in a one-to-one manner. This fused representation is used for GPR prediction and is referred to as Feature–Data Fusion (FDF). Huang et al. [12] adopted a similar input structure. Still, they employed a simpler fusion mechanism: compressing the dose matrix into a 99-dimensional vector and concatenating it with a single MU value before passing the result through fully connected layers. Ref. [15] converted delivery log files into MU-weighted fluence maps and used them as inputs to a DenseNet model to predict GPRs under multiple gamma criteria. While these intermediate fusion approaches effectively capture high-dimensional cross-modal embeddings and are well-suited for fine-grained prediction tasks such as dose deviation prediction (DDP), they tend to overlook a key challenge in the PSQA context: the modality imbalance problem, where the tabular modality converges significantly faster than the image modality, leading to sub-optimal optimization and degraded model performance [16]. Table 1 provides a summary of multimodal PSQA methods.

2.2. Imbalanced Multimodal Learning

The issue of modality imbalance has also drawn increasing attention in multimodal learning. In intermediate fusion settings, faster-converging modalities often dominate the joint training process, effectively suppressing the contribution of slower, more complex modalities. As a result, rich but harder-to-optimize inputs, such as high-dimensional image features, may be underutilized, leading to sub-optimal multimodal representations. Researchers have proposed modality-specific learning rate scheduling, gradient normalization (e.g., GradNorm), and adaptive weighting of modality contributions during training to mitigate this effect.
The multimodal model should do better than the unimodal model since it works with data that has information from more than one view. However, the widely used multimodal joint training model does not always work well based on existing studies [26], prompting researchers to investigate the reasons for this. Recent studies point out that the jointly trained multimodal model cannot effectively improve the performance with more information as expected due to the discrepancy between modalities [23,26,27,28,29]. Wang et al. [26] found that multiple modalities often converge and generalize at different rates, thus training them jointly with a uniform learning objective is sub-optimal, leading to the multimodal model sometimes is inferior to the unimodal ones. Also, Winterbottom et al. [27] indicated an inherent bias in the TVQA dataset towards the textual subtitle modality. Besides the empirical observation, Huang et al. [29] further theoretically proved that the jointly trained multimodal model cannot efficiently learn features of all modalities, and only a subset of them can capture sufficient representation. They called this process “Modality Competition”. Recently, several methods have emerged attempting to alleviate this problem [23,24,26,30]. Wang et al. [26] proposed to add additional unimodal loss functions besides the original multimodal objective to balance the training of each modality. Du et al. [30] utilized well-trained unimodal encoders to improve the multimodal model by knowledge distillation. Wu et al. [23] found out how quickly the model learns from one type of input compared to the other types of input, and then proposed to guide the model to learn from previously underutilized modalities. Wei et al. [31] introduced a Shapley-based sample-level modality valuation metric to observe and alleviate the fine-grained modality discrepancy. Wei et al. [32] further considered the possible limited capacity of modality and utilized the re-initialization strategy to control unimodal learning. In contrast, Yang et al. [33] focused on the influence of imbalanced multimodal learning on multimodal robustness, and proposed a robustness enhancement strategy.
While these methods have improved multimodal learning, a comprehensive analysis of imbalanced multimodal learning is still lacking. Most previous studies have addressed modality imbalance in single-task scenarios. However, in real-world applications—such as PSQA in the medical domain—multimodal models are often designed for multi-task settings, where modality imbalance typically results in the degradation of the performance of an entire task. This is the central problem our work aims to address.

3. Foundations

This section establishes the methodological foundations for our study.

3.1. Clinically Validated Data Curation

For this study, we utilized the private data set obtained in collaboration with the Department of “Radiation Oncology” at “Peking Union Medical College Hospital” https://www.pumch.cn/en/introduction.html, accessed on 3 October 2025. The dataset was collected between December 2020 and July 2021. It comprises 210 FF-IMRT treatment plans, encompassing 1370 beam fields for various treatment sites, and the treatment plans by site are detailed in Table 2.
All plans were generated using Eclipse TPS version 15.6 (Varian Medical System) and delivered via Halcyon 2.0 linac with SX2 dual-layer MLC. The sliding window was employed for plan optimization, with dose calculation performed using the Anisotropic Analytic Algorithm (AAA, version 15.6.06) at a 2.5 mm grid resolution.
Under the TG-218 report recommendation [34], PSQA measurements were conducted before treatment delivery using actual beam angles and employing portal dosimetry. Dose calibration was performed daily before data collection. Gamma analyses were executed with criteria of 1 % / 1  mm, 2 % / 2 mm, and 2 % / 3 mm at a 10% threshold of maximum dose, analyzing only points with doses greater than 10% of the global maximum dose per beam, as detailed in Table 3. The gamma analyses were carried out in absolute dose mode, applying global normalization to the results. The treatment planning system calculated fluence maps, exported in DICOM format, which served as inputs for the deep learning network. The raw fluence maps underwent preprocessing to standardize input data: resampling to 1 × 1 mm 2 resolution, cropping to 512 × 512 pixels, and Min-Max normalization to scale pixel values to [0, 1]. This comprehensive dataset, alongside rigorous PSQA measurements, underpins the proposed Virtual Dose Verification.
The dataset’s GPR distribution, categorized into clinical intervals (<85%, 85–90%, 90–95%, >95%), revealed site-specific trends as shown in Figure 2; for example, head and neck (H&N) plans exhibited higher complexity (51 beams below 85% GPR). In comparison, chest (C) plans demonstrated optimal compliance (106 beams > 95% GPR). Qualitative PSQA validation as shown in Figure 3, further compared TPS-calculated doses, measured doses, and absolute dose differences across four representative IMRT plans, with the pixel intensity (darker = higher dose Gy) highlighting discrepancies, particularly in high-gradient regions. These visualizations collectively underscore the dataset’s variability and the critical role of rigorous QA, especially for an anatomically complex site like H&N.

3.2. Task Formulation

In our PSQA task, each input sample X comprises two modalities: an image modality X img and a tabular modality X tab . The image modality X img is a 512 × 512 ( H × W ) matrix representing relative dose intensity values normalized to the range [ 0 , 1 ] . The tabular modality X tab is a 33-dimensional vector composed of plan complexity metrics.
We consider two prediction tasks. The first is GPR prediction, where the target variable Y gpr is a three-dimensional vector corresponding to three standard clinical criteria. The second is DDP, where the target variable Y ddp is a 512 × 512 matrix aligned with the spatial resolution of the image input.
Table 4 summarizes the mathematical symbols used throughout this study. Superscripts indicate modalities (e.g., img, tab), while subscripts refer to task-specific roles (e.g., gpr, ddp). These definitions provide a consistent multimodal framework for modeling the PSQA task, enabling both coarse-grained and fine-grained prediction objectives.

3.3. Shapley Value

The Shapley value [42] was originally formulated in coalitional game theory to fairly allocate the total payoff among players based on their contributions. A multi-player game is typically described by a real-valued utility function V defined over a set of players N : = { 1 , , n } . Players can form various coalitions (i.e., subsets of N ), and their collective interactions can be either cooperative or competitive.
To quantify the contribution of a specific player i, consider a coalition S N { i } . The marginal contribution of player i to coalition S is defined as
m i ( S ; V ) = V ( S { i } ) V ( S ) .
Since a player’s influence depends on all possible coalitions it may join, the Shapley value ϕ i is defined as the average of its marginal contributions across all such subsets, weighted by the size of the coalitions:
ϕ i ( N ; V ) : = S N { i } | S | ! ( n | S | 1 ) ! n ! m i ( S ; V ) .
A fair allocation of payouts in cooperative game theory is characterized by four fundamental axioms: Efficiency, Symmetry, Dummy, and Additivity. Among attribution methods, the Shapley value is uniquely defined as the only solution that satisfies all four properties [43]. More formally:
Efficiency: The total payout is fully distributed among all players, i.e., i ϕ i = V ( N ) V ( ) , where V ( ) represents the utility of the empty set (baseline with no players).
Symmetry: If two players i and j contribute equally to every coalition, that is, V ( S { i } ) = V ( S { j } ) for all S N { i , j } , then their attributions must be identical: ϕ i = ϕ j .
Dummy: A player i is considered a dummy if its inclusion does not affect any coalition beyond its standalone contribution, i.e., V ( S { i } ) = V ( S ) + V ( { i } ) for all S N { i } . In this case, the player’s attribution equals its value: ϕ i = V ( { i } ) .
Additivity: For any two games with utility functions V and W, their combined game is defined as U ( S ) = V ( S ) + W ( S ) for all S N . The attribution in the combined game satisfies linearity: ϕ i ( U ) = ϕ i ( V ) + ϕ i ( W ) .
These axioms collectively define the criteria for fair attribution, and it has been mathematically proven that the Shapley value is the only allocation method that satisfies them all [44].
Game-theoretic principles have been widely adopted in machine learning research [45,46,47]. For instance, game theory has been used to explain the effectiveness of ensemble methods such as AdaBoost [45]. In a related line of work, Hu et al. [48] utilize the Shapley value to assess the overall contribution of individual modalities across an entire dataset. However, their approach does not address modality contributions at the sample level and lacks a deeper analysis of underperforming modalities.
In this work, we propose a method for sample-level modality valuation, providing a more fine-grained understanding of modality importance. Furthermore, we analyze and propose solutions to mitigate the impact of low-contributing modalities.

4. Methodology

4.1. Overview

As Figure 4 depicts, the BMMQA framework integrates four modular components for robust radiotherapy quality assurance: Input Layer processes dual-modality data (spatial dose matrices and tabular complexity metrics); modality encoders transform inputs into high-level features via pre-trained image encoding and MLP-based tabular embedding; task-specific feature fusion dynamically balances contributions using attention-based weighting for GPR regression and spatial concatenation for DDP mapping; and decoders generate task-specific predictions (scalar GPR values and 512 × 512 DDP matrices via U-Net decoding). This structured pipeline ensures modality-aware feature collaboration while preserving spatial integrity for clinical reliability.
To further mitigate modality imbalance during training, BMMQA incorporates a dynamic balancing mechanism that integrates multiple synergistic components: (1) Task-specific fusion strategies: a softmax-weighted attention mechanism is used for GPR regression, while spatial cascading is applied for DD prediction. (2) A theoretically supported balancing mechanism: Shapley values, particularly suited for multi-task regression, are employed to quantify the contribution of each modality. (3) A fast network forward mechanism designed to compute the output of different modality combinations efficiently. (4) A modality-contribution-based task weighting scheme developed explicitly for multi-task multimodal learning: assigning different task-specific weights according to modality contributions to address the modality imbalance problem in multi-task learning.
This closed-loop design enhances robustness, particularly in critical failure cases (e.g., GPR < 90%), by preserving feature collaboration integrity across both coarse-grained and fine-grained PSQA tasks.

4.2. Modality Encoders

The model input consists of two modalities: image modality dose plan arrays X i m g , and tabular modality plan complexity metrics X t a b . The image modality is passed through a convolutional or Transformer-based pretrained encoder (e.g., ResNet) to extract spatial features, yielding a high-dimensional embedding Z img R N c × W × H , where N c is the number of channels, and W , H denote the spatial dimensions.
To ensure dimensional compatibility for subsequent fusion, the tabular modality is processed by an MLP encoder to extract structured feature embeddings. These embeddings are initially represented as Z tab R ( N t · W · H ) × 1 , where · denotes scalar multiplication.
The encoding process can be formalized as:
Z img = Enc img ( X img ) ,
Z tab = Enc tab ( X tab ) ,

4.3. Task-Specific Modality Fusion

The intrinsic task heterogeneity between Gamma Passing Rate (GPR) regression (scalar output) and dose difference prediction (DDP) (spatial matrix output) necessitates divergent fusion strategies. This duality manifests in two critical dimensions: Representational Conflict: the GPR requires compact, modality-agnostic feature abstraction for global dose compliance assessment, while DD demands high-resolution spatial preservation for localized discrepancy mapping. Optimization Incompatibility: Naive uniform fusion (e.g., concatenation or averaging) causes representational interference—spatial details critical for DD are diluted by low-dimensional tabular features optimized for GPRs. To resolve this duality, BMMQA introduces a task-specific modality fusion mechanism, GPR-Oriented Fusion prioritizes attention-based modality weighting to distill task-relevant signals while suppressing noise through differentiable softmax gates. DDP-Oriented Fusion maintains structural integrity via channel-wise spatial alignment, preserving high-resolution image embeddings without destructive compression.

4.3.1. GPR-Oriented Fusion

In the GPR task, the objective is to predict coarse-grained scalar values rather than spatial maps. We first apply spatial pooling to the high-dimensional feature maps extracted from the image modality to align the modalities for this point-wise regression. This yields a compact image embedding that is dimensionally compatible with the tabular embedding. The two embeddings are then concatenated to form a joint representation for regression.
Due to the opaque nature of deep networks, assessing modality imbalance during training requires visibility into how each modality contributes to the final output. While prior work has proposed feed-forward architectures that isolate modal contributions, these are typically designed for classification tasks, where the output is normalized via softmax. In regression tasks like PSQA, the outputs are not inherently normalized, which leads to scale shift problems when isolating modality contributions, rendering many imbalance indicators ineffective.
To address this, we adopt an attention-based fusion mechanism on the modality embeddings. Instead of normalizing at the output layer, we apply a softmax weighting to the modality embeddings during fusion. This design enables transparent and differentiable attribution of modality contributions, making it suitable for imbalance assessment in regression settings.
Each modality embedding is projected to a scalar score s via a learnable linear transformation:
s img = W s img Z img , s tab = W s tab Z tab
For notational simplicity, we write W Z img as a matrix multiplication, where Z img R C × H × W are feature maps. In practice, spatial pooling (e.g., global average pooling) is applied to ensure shape compatibility before the projection. This convention is adopted throughout the remainder of this paper.
The scores s are normalized using a softmax function to obtain attention weights α that reflect the relative contribution of each modality:
α img = exp ( s img ) exp ( s img ) + exp ( s tab ) α tab = exp ( s tab ) exp ( s img ) + exp ( s tab )
The final fused representation is computed as a weighted combination of the modality-specific embeddings:
Z gpr = Z gpr ( img , tab ) = α img · W img Z img + α tab · W tab Z tab
We denote Z gpr ( img , tab ) as the joint representation derived from both image and tabular modalities. For clarity, we simplify this as Z gpr when both modalities are present.

4.3.2. DDP-Oriented Fusion

Since dose difference prediction requires generating spatial outputs, it is essential to preserve spatial information from the high-dimensional image modality. We argue that applying modality attention in this task is suboptimal. Therefore, this stage does not use the attention-based modality fusion mechanism described in the previous section.
To align the tabular features with the spatial structure of the image features, we first fold the tabular embedding vector Z tab R N t · H · W into feature maps R N t × H × W . The reshaped Z tab is then concatenated with the image feature map Z img along the channel dimension to form the joint representation Z ddp R ( N c + N t ) × H × W . Notably, the number of tabular channels N t is intentionally set much smaller than the number of image channels N c , reflecting our prior assumption that the DDP task should place greater emphasis on the image modality, which inherently preserves spatial information.
Z ddp = Concat ( Fold ( Z tab ) , Z img )

4.4. Task-Specific Decoders

The fused multimodal representations ( Z g p r and Z d d p ) require distinct decoding pathways tailored to the dimensional characteristics of their respective prediction tasks. The scalar output of GPRs necessitates progressive feature compression through fully connected layers, which transform high-dimensional inputs into compact regression targets. Conversely, the matrix output of DDP requires the maintenance of topological consistency through convolutional decoders that preserve spatial relationships.

4.4.1. GPR Decoder

After attention-based fusion, the resulting joint representation is further processed to enhance non-linearity and enable prediction. Specifically, the fused vector Z gpr is first passed through a ReLU activation function, followed by an MLP to generate the final GPR prediction:
Y ^ gpr = Y ^ gpr ( i m g , t a b ) = f g p r ( i m g , t a b ) ( Z g p r ) = MLP ( ReLU ( Z gpr ) )
Here, the MLP consists of one or more fully connected layers that transform the activated fusion output into a scalar regression value. This structure allows the model to capture complex interactions between modalities beyond the initial linear fusion.
To optimize the GPR regression model, we adopt the mean absolute error (MAE) as the training loss:
L g p r = MAE ( Y g p r , Y ^ g p r ) = 1 n i = 1 n ( Y g p r ) i ( Y ^ g p r ) i

4.4.2. Dose Difference Decoder

To generate matrix-form outputs, we adopt a U-Net-style decoder, commonly used in image segmentation tasks. The joint representation Z ddp is passed through a series of convolutional layers that progressively reduce the channel dimension to 1, resulting in a continuous-valued prediction of the dose difference map Y ^ ddp .
Y ^ ddp = f ddp ( Z ddp )
To ensure a consistent loss scale for multi-task training, we also adopt the MAE as the training loss for the DDP matrix regression task:
L ddp = MAE ( Y ddp , Y ^ ddp ) = 1 n i = 1 n ( Y ddp ) i ( Y ^ ddp ) i
Here, Y ddp R H × W and Y ^ ddp R H × W are the ground-truth and predicted dose difference matrices, respectively, and n = H × W denotes the total number of spatial elements.

4.5. Dynamic Balancing Mechanism

The persistent dominance of tabular modalities during joint optimization necessitates a rigorous quantification framework to address two fundamental manifestations of imbalance: (1) gradient suppression, where earlier-converging modalities produce vanishing gradients that hinder updates to slower modality encoders, and (2) representational collapse, wherein non-dominant modalities fail to achieve optimal capacity due to inhibited feature learning. To mitigate these issues, we base our balancing mechanism on coalitional game theory, utilizing Shapley values to provide an axiomatically fair attribution metric suitable for continuous-output regression scenarios.

4.5.1. Shapley-Based Imbalance Indicator

To reveal the extent of modality imbalance, we formalize modality ablation within the feed-forward process by defining the following single-modality and null-modality variants:
Image-only Forward. When only the image modality is used, the fusion reduces to:
Z gpr img = 1 · W img Z img + 0 · W tab Z tab = W img Z img
Tabular-only Forward Similarly, when only the tabular modality is used, the fused representation becomes:
Z gpr tab = 0 · W img Z img + 1 · W tab Z tab = W tab Z tab
Null-modality Forward To formulate a no-input reference, we define the null-modality fusion as:
Z gpr = 0 · W img Z img + 0 · W tab Z tab = 0
Since Shapley values are typically concerned with relative differences in contribution, and this null case is shared across all Shapley value calculations, it can be safely assigned a fixed value of zero in practice without affecting the consistency of the final attribution results.
Building on this foundation, we adopt a Shapley value framework as a principled attribution method to quantify the degree of modality imbalance in our multimodal regression setting. Given a set of modalities, we define the power set of possible modality combinations as:
P = , { X img } , { X tab } , { X img , X tab }
For each subset S P , we compute a value function v ( S ) as the negative mean absolute error (MAE) over the validation set when the model is evaluated with only the modalities in S:
v ( S ) = MAE ( Y , Y ^ S ) = 1 n i = 1 n Y i Y ^ i S , S P
where Y ^ S denotes the model prediction with only modality subset S available, as defined in our modality ablation forward formulation.
Following the standard two-modality Shapley decomposition, the contribution (or marginal utility) of each modality can be computed as:
ϕ img = 1 2 v ( { X img } ) v ( ) + 1 2 v ( { X img , X tab } ) v ( { X tab } ) ϕ tab = 1 2 v ( { X tab } ) v ( ) + 1 2 v ( { X img , X tab } ) v ( { X img } )
These values reflect each modality’s average marginal contribution across all possible input subsets. In our implementation, since v ( ) is a constant term shared across all Shapley computations, it is set to zero without loss of generality. The resulting Shapley scores serve as modality imbalance indicators: lower values indicate weaker modality utility during prediction, highlighting convergence imbalance or representation dominance issues.
Equation (16) provides the formal definition of modality-level Shapley values. In practice, we substitute the contribution function defined in Equation (15) into the expression in Equation (16), yielding the following MAE-based formulation:
ϕ img = 1 2 MAE ( Y , Y ^ img ) + MAE ( Y , Y ^ ) + 1 2 MAE ( Y , Y ^ ( img , tab ) ) + MAE ( Y , Y ^ tab )
ϕ tab = 1 2 MAE ( Y , Y ^ tab ) + MAE ( Y , Y ^ ) + 1 2 MAE ( Y , Y ^ ( img , tab ) ) + MAE ( Y , Y ^ img )
This formulation allows us to interpret the marginal contribution of each modality as the average reduction in prediction error across all possible inclusion orders.

4.5.2. Adjustment Strategy

Given that the two tasks exhibit distinct modality preferences, we introduce a loss weighting strategy to regulate modality balance during training. Specifically, in the GPR prediction task, the low-dimensional tabular features tend to dominate early convergence due to their high information density. In contrast, the dose difference task, which requires fine-grained spatial matrix regression, relies more heavily on image modality to capture spatial context.
Thus, we propose a dynamic loss weighting strategy based on modality attribution to address modality imbalance. We define the DDP loss coefficient as:
λ DDP = 1 + max ( ϕ tab ϕ img , 0 ) · r 2
where ϕ tab is the Shapley value quantifying the contribution of the tabular modality in GPR prediction. r is the Modality Balance Factor as the hyperparameter. The overall training objective is thus formulated as:
L = L gpr + λ DDP · L ddp
This formulation increases the emphasis on the image modality (via the DDP task) when the tabular modality exhibits dominance, helping to balance modality convergence across tasks.
Algorithm 1 shows the per-epoch training procedure of BMMQA.
Algorithm 1: Per-epoch Training Procedure of BMMQA
Diagnostics 15 02555 i001

5. Evaluation and Results

This section comprehensively analyzes our proposed BMMQA for achieving robust automated PSQA.

5.1. Implementation Setting

The framework processes 512 × 512 resized dose arrays through a dual-task architecture: (1) a GPR regression head (AdaptiveAvgPool2d → Flatten → Dropout(0.1) → Linear(512, 3)) predicting three gamma criteria, and (2) a U-Net decoder with channels [768, 384, 192, 128, 32] for dose prediction. Training used Adam ( η = 5 × 10 4 ) with a 32-batch/30-epoch configuration and a 7:1:2 Train:Validation:Test split, with regularization applied via dropout ( p = 0.1 ) throughout the network as detailed in Table 5. To ensure representativeness, we employed multi-factor stratified sampling (Figure 2), taking into account both binned Gamma Passing Rates (continuous regression targets) and lesion categories (categorical feature). This approach prevents biased splits (e.g., training only on success cases or rare lesion types in isolation) and ensures that each subset covers the full difficulty spectrum and lesion diversity, thereby improving fairness and reproducibility.

5.2. Evaluation Metrics

We evaluate BMMQA performance using mean absolute error (MAE), Root Mean Square Error (RMSE) and R 2 for the GPR task, and both MAE and Structural Similarity Index Measure (SSIM) for the DDP task.
MAE = 1 n i = 1 n | y i y ^ i |
RMSE = 1 n i = 1 n ( y i y ^ i ) 2
R 2 = 1 i = 1 n ( y i y ^ i ) 2 i = 1 n ( y i y ¯ ) 2
SSIM ( X 1 , X 2 ) = ( 2 μ X 1 μ X 2 + C 1 ) ( 2 σ X 1 X 2 + C 2 ) ( μ X 1 2 + μ X 2 2 + C 1 ) ( σ X 1 2 + σ X 2 2 + C 2 )

5.3. Baseline Methods

This subsection presents the baseline PSQA methods as benchmarks for comparison with the proposed BMMQA approach.
Previous studies have predominantly focused on single-modal approaches, which form the foundation of the single-modal baseline methods selected in this work.
To comprehensively evaluate the impact of different modalities in PSQA, we categorized our experimental methods into tabular, image, and unbalanced (a combination of tabular and image modalities) and balanced multimodalities. For each setting, both tasks (GPR and DDP) were trained simultaneously. We then analyzed the performance across various metrics and visualized the results using diagrams and tables.
Methods based on the tabular modality have primarily relied on traditional machine learning approaches in previous studies. In particular, regression tree models and linear regression models have been the most widely adopted in PSQA research: tree-based models have been employed in several works [41,49,50,51], while linear regression models have also been tested in related studies [41,50,52], including Poisson Lasso regression as a representative approach. Following Zhu et al. (2023) [41], we therefore employed Gradient Boosting Decision Tree (GBDT), Random Forest (RF), and Poisson Lasso (PL) models as representative baselines. These models utilize manually designed complexity metrics as input features to predict the Gamma Passing Rate (GPR), ensuring consistency with prior work, enabling direct comparability, and supporting reproducibility. Detailed descriptions of their implementation can be found in [41].
For image modality baselines, ref. [53] proposed a ResNet-based UNet++ architecture for predicting GPR and dose differences, while [54] introduced TransQA, a Transformer-based model designed for the same tasks. To address data imbalance, ref. [8] combined RankLoss with a DenseNet backbone to enhance the robustness of GPR predictions. For consistency in comparison, we also incorporated RegNet [55] to evaluate the performance of a more recent CNN backbone.
In our study, the last two categories of methods—unbalanced and balanced multimodal approaches—share the same network architecture, with the key distinction being whether a modality balancing algorithm is applied. Since the tabular modality is relatively simple, it typically employs a fixed encoder structure [10]; therefore, the main variation across settings lies in the choice of the image modality encoder. Specifically, we selected ResNet18 [53,56], DenseNet121 [8,15,57], MobileVit [54,58], and RegNet [55,59]. ResNet and DenseNet represent widely adopted classical CNN backbones in prior PSQA-related studies, MobileVit provides a lightweight variant, while RegNet is a relatively recent CNN family proposed around the same period as Transformer-based approaches and has demonstrated promising potential in medical imaging applications. Together, these encoders cover classical, lightweight, and modern high-performing architectures, ensuring consistency with prior baselines and enabling fair comparisons.
The main architectures employed in PSQA prediction are convolutional neural networks (CNNs) and, more recently, Transformer-based models. Among CNNs, ResNet and DenseNet have been widely adopted in prior PSQA-related studies [8,15,53,56,60], while MobileVit [54,58] represents a lightweight variant. To complement these, we also included RegNet [61], a relatively recent CNN family proposed around the same period as Transformer-based approaches, which has demonstrated promising potential in medical image generation tasks [59]. Together, these encoders provide a representative spectrum across classical, lightweight, and modern high-performing architectures, ensuring consistency with prior baselines and enabling fair comparisons.
The overall experimental settings remain largely consistent for the DDP task. However, since the tabular modality alone cannot independently accomplish the DDP task, it is not included in the corresponding results table.

5.4. EXP-A: Comparative Experiment with PSQA Methods

5.4.1. GPR Result Analysis

Table 6 reports the GPR prediction performance of various models under three commonly used gamma criteria (1%/1 mm, 2%/2 mm, and 2%/3 mm), with results further stratified by ground truth GPR intervals: All, 95–100, 90–95, and <90. The choice of intervals follows the recommendations of the AAPM Task Group 218 report [34], which defines a universal tolerance limit of GPR ≥ 95% (with 3%/2mm criteria and 10% dose threshold) for clinical acceptance, and a universal action limit of GPR ≥ 90% under the same criteria. Therefore, we stratified the results into three clinically meaningful categories: 95–100% (passing and acceptable), 90–95% (borderline, close to the action limit), and <90% (failing plans requiring further investigation). Since R 2 is known to be overly sensitive under sparse subgroup data distributions, we only report the overall ( A l l ) R 2 values to ensure meaningful and stable interpretation.
This stratification allows for a more nuanced evaluation beyond global accuracy, especially regarding clinically relevant but underrepresented failure cases. In practical radiotherapy quality assurance, plans with GPR < 90% are usually considered to have failed. Misclassifying these failing plans as acceptable poses a significant clinical risk, far greater than conservative rejection of marginally acceptable plans. Moreover, such failure cases account for only a small proportion of the dataset, making the <90 subgroup rare and highly valuable for evaluating model robustness and practical deployment readiness.
To complement the tabulated results, Figure 5 and Figure 6 show GPR scatter plots under the 2%/2mm criterion, with ground-truth GPR on the x-axis and predicted GPR on the y-axis. Light-blue dots represent individual samples. The diagonal dashed lines indicate prediction bands: the red line denotes the top 5% correct predictions, the green line the upper 5% deviation, and the dark-blue line the lower 5% deviation.
From the overall trend, it is clear that models trained solely on the tabular modality underperform across nearly all settings. Among them, RF and PL show lower overall accuracy and higher prediction errors in the <90 subgroup, indicating poor generalization and limited robustness. While the GBDT model demonstrates relatively better robustness in identifying failure cases under relaxed criteria, namely, 2%/2 mm and 2%/3 mm, this comes at the cost of diminished performance in the 95–100 and overall categories, suggesting an overfitting tendency toward rare cases or a lack of adaptability across GPR ranges. This reflects a broader limitation of relying on low-dimensional tabular features alone: such features may capture plan complexity but lack the spatial context necessary for comprehensive dose distribution assessment.
Image-only models, on the other hand, benefit from richer spatial information and demonstrate superior overall accuracy across nearly all criteria. Notably, models such as UNet++, TransQA, and RegNet consistently outperform tabular baselines in the “All” and 90–100 interval. However, their performance in the critical <90 subgroup is not always optimal. For instance, in some settings, the RMSEs in this subgroup are higher than those of GBDT, indicating a potential weakness in detecting outliers or failure cases. This suggests that while image modality provides strong general pattern learning capabilities, it may require additional mechanisms to capture minority failure modes effectively.
Multimodal models integrate image and tabular information and offer a compelling balance between overall performance and robustness. Across all settings, multimodal models outperform their unimodal counterparts, confirming the complementary nature of these modalities. More importantly, the proposed modality-balanced variants demonstrate consistent improvements over unbalanced versions. This is particularly evident in the <90 subgroup, where balanced models reduce the MAE and RMSE for nearly all encoder types. The benefit is pronounced in ResNet18, DenseNet121, and MobileViT-based variants, where balancing helps prevent the tabular modality from prematurely dominating the learning process. This validates our hypothesis that early convergence of simpler modalities can hinder effective feature fusion, and that regulating this imbalance can substantially enhance the model’s ability to improve overall prediction fidelity and remain sensitive to critical failure cases.
Although RegNet-based balanced models show slightly less consistent improvements—especially in RMSE under strict criteria—their performance remains competitive and still offers benefits in terms of MAE. This variation may stem from architectural sensitivity or optimization dynamics, suggesting that the effect of balancing may interact with encoder complexity.
These findings highlight the necessity of addressing modality imbalance in multimodal learning for PSQA. Simple fusion of modalities without considering training dynamics can lead to suboptimal utilization of complementary information. Our approach, which combines learning rate modulation and gradient normalization, offers an effective solution. In particular, the observed robustness improvements in the <90 subgroup emphasize the clinical importance of balanced learning for enhancing average accuracy and ensuring patient safety by reducing the likelihood of high-risk misclassifications. As such, modality balancing should be considered a key component in building safe, trustworthy AI systems for clinical deployment in radiotherapy quality assurance.
In addition to MAE and RMSE, the R 2 values in Table 6 further corroborate these observations. Tabular-only models generally yield a lower R 2 , often below 0.60, confirming their limited explanatory power and inability to capture the variance in GPR outcomes. Image-only models achieve moderately higher R 2 (around 0.62–0.64), reflecting their stronger alignment with ground-truth variations but with persistent weaknesses in the <90 subgroup. In contrast, multimodal models—particularly the balanced variants—consistently attain the highest R 2 values across nearly all criteria, with RegNet-based models exceeding 0.67 in some settings. These improvements indicate not only reduced absolute errors but also enhanced consistency in capturing the underlying variance structure of GPR. Importantly, the higher R 2 in the balanced multimodal models suggests improved generalization, supporting their reliability for both routine QA and the detection of clinically critical outliers.

5.4.2. DDP Result Analysis

Table 7 reports the DDP performance under different modality configurations and encoder backbones. We evaluate MAE (%), SSIM and R 2 , with further stratification by ground truth GPR intervals: 95–100, 90–95, and <90, in addition to the overall average. These strata correspond to varying clinical acceptance levels, with lower GPR values indicating more challenging cases. Since R 2 is highly sensitive to small sample sizes and may yield unstable or even misleading values when computed on narrow strata, we report R 2 values only for the overall (all samples) case to ensure robustness and interpretability.
Since the tabular modality lacks spatial resolution, we focus on three representative settings: (1) image-only baselines, (2) unbalanced multimodality (MM) models without contribution control, and (3) balanced multimodality (BMM) models using our proposed modality regulation mechanism.
From Table 7, we observe that unbalanced fusion can degrade DDP performance, sometimes even underperforming image-only models—especially in high-GPR regions. This suggests that DDP is primarily image-sensitive, and that naive modality fusion can introduce interference when less-suitable modalities dominate. In contrast, BMM models consistently achieve the lowest MAE, highest SSIM, and strongest R 2 values across backbones, highlighting the value of modality balancing in spatially structured prediction. The consistently higher R 2 further indicates that BMM models explain a greater proportion of variance in voxel-wise dose differences, reinforcing their predictive robustness.
To better interpret these quantitative results, Figure 7 shows residual distributions across all modality-backbone combinations. Each row corresponds to a backbone (ResNet, DenseNet, MobileViT, RegNet), and columns represent BMM, MM, and image-only settings. The histograms confirm that BMM models not only reduce average errors, but also produce more symmetric, narrow residual distributions centered near zero—indicating better generalization and reduced voxel-level variance. In contrast, MM models show broader and often skewed distributions, reflecting prediction instability due to unregulated modality mixing.
To further explore spatial behavior, Figure 8 visualizes voxel-wise error maps on selected test cases. Consistently, BMM models yield cleaner and more localized error patterns. Notably, MobileViT-MM exhibits strong overestimation artefacts (bright yellow) in the center, which are suppressed mainly in its BMM counterpart. Similar spatial refinements are observed across other backbones, especially in clinically critical regions.
Together, these results demonstrate that modality regulation not only improves quantitative metrics like MAE and SSIM, but also enhances spatial fidelity—reducing high-error artifacts and increasing clinical reliability. The complementary insights from Table 7 and Figure 7 and Figure 8 strongly validate the effectiveness of our balanced multimodal strategy in spatially sensitive PSQA tasks.

5.4.3. Modality Contribution Dynamics During Training

Figure 9 presents the evolution of modality contributions across training epochs for four representative backbone encoders. In the unbalanced setting (dashed lines), the tabular modality consistently dominates, exhibiting significantly higher contribution scores than the image modality across all backbones. This imbalance persists throughout training and suggests that the network overfits to the easier-to-learn tabular features while underutilizing spatial information, which is critical for DDP.
In contrast, under our balanced multimodality strategy (solid lines), modality contributions become more equitable. The image modality sees a clear increase in contribution, while the tabular modality’s influence is moderated, resulting in a more stable and task-appropriate fusion. These results empirically demonstrate that our balancing mechanism effectively mitigates modality dominance during joint training.
When viewed alongside Table 7, it is evident that balanced modality contributions lead to superior DDP performance, especially for image-sensitive tasks. This further validates our claim that addressing modality imbalance is essential for achieving both accurate and robust predictions in multimodal PSQA models.

5.5. EXP-B: Ablation Experiment with Modality Balance Factor r

We conduct an ablation experiment to analyze the influence of the hyperparameter r in the formulation of λ DDP . This parameter serves as an amplification factor, dynamically rebalancing the contributions of different modalities, particularly when one modality dominates the learning process. In this experiment, for each fixed r, we train a separate model and record its final DDP prediction error (MAE) under three standard GPR thresholds.
As shown in Figure 10, the error curves generally decrease as r increases, reaching their minima around r = 14 to r = 17 for all three criteria. This indicates that moderate amplification of modality imbalance yields the best DDP accuracy. When r is too small (e.g., close to 0), modality imbalance is insufficiently addressed, while overly large values (e.g., r > 18 ) may lead to overcompensation and degrade performance. The consistent minima across criteria demonstrate the robustness of our balancing strategy.

5.6. Computational Resource Utilization

SHAP-based models are often associated with concerns regarding computational cost, as exact SHAP computation typically requires exponential complexity. However, in our proposed framework, we circumvent this limitation by enabling the simultaneous derivation of multiple modality ablation results within a single forward pass, thus significantly reducing the overhead traditionally linked to SHAP.
Table 8 compares the computational resource usage of our BMMQA framework against several deep learning-based PSQA baselines. The comparison includes model names, floating-point operations (FLOPs in GMac), and parameter counts (in millions). Among the baselines, Unet++, TranQA, and RegNet represent PSQA models with varying computational demands.
Notably, BMMQA achieves competitive performance with substantially lower FLOPs and parameter size, underscoring its efficiency and suitability for resource-constrained or computation-sensitive clinical scenarios.

6. Conclusions

Through a comprehensive review and empirical analysis of existing multimodal deep learning approaches for PSQA, we identified modality imbalance—where faster-converging tabular features dominate training and suppress image feature learning—as a key factor limiting model robustness and clinical reliability. To address this, we propose BMMQA, a modality- and task-aware multimodal framework that explicitly regulates modality contributions during training. BMMQA leverages SHAP-derived imbalance indicators with task-specific weighting strategies to balance convergence across modalities, and incorporates a fast network forward mechanism along with tailored fusion protocols—attention-based weighting for GPR and spatial concatenation for DDP—to enhance feature collaboration. Experiments on a large clinical IMRT dataset show that BMMQA consistently outperforms state-of-the-art fusion baselines in both coarse-grained (GPR) and fine-grained (DDP) tasks, while improving interpretability and reliability. These results highlight the importance of explicitly addressing modality imbalance for the safe and effective deployment of AI in radiotherapy quality assurance, establishing a novel example in imbalanced multi-task multimodal learning.
This study has two main limitations. First, the dataset was collected from a single center (PUMCH) within a limited timeframe, which may restrict the generalizability of the results and introduce potential biases associated with technological evolution. Second, the current framework relies on only two modalities: image-based dose distributions and tabular plan complexity features, while excluding other potentially informative clinical data, such as CT scans or electronic health records (EHRs). Future work will aim to validate the framework on multi-center, multi-device datasets to enhance robustness and to extend the multimodal design by incorporating additional clinical information, thereby further improving predictive performance and clinical applicability.

Author Contributions

X.Z. contributed to conceptualization, methodology, software, validation, data curation, and writing—original draft preparation; A.A. handled formal analysis, supervision, investigation, and writing—review and editing, visualization; M.H.T. was involved in writing—reviewing and editing and visualization. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the China West Normal University Doctoral Start-up Fund under Grant No. 24KE033.

Institutional Review Board Statement

This study was conducted in accordance with the Declaration of Helsinki (as revised in 2013) and was approved by the Institutional Ethic Review Board of the Peking Union Medical College Hospital (no. I-23PJ2059).

Informed Consent Statement

Informed consent was obtained from all participants involved in the study. Participants were informed about the study’s purpose, procedures, and their right to withdraw at any time without any consequences.

Data Availability Statement

The data supporting the findings of this study are available upon reasonable request from the first or corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Meijers, A.; Marmitt, G.G.; Siang, K.N.W.; van der Schaaf, A.; Knopf, A.C.; Langendijk, J.A.; Both, S. Feasibility of patient specific quality assurance for proton therapy based on independent dose calculation and predicted outcomes. Radiother. Oncol. 2020, 150, 136–141. [Google Scholar] [CrossRef] [PubMed]
  2. Zeng, X.; Zhu, Q.; Ahmed, A.; Hanif, M.; Hou, M.; Jie, Q.; Xi, R.; Shah, S.A. Multi-granularity prior networks for uncertainty-informed patient-specific quality assurance. Comput. Biol. Med. 2024, 179, 108925. [Google Scholar] [CrossRef] [PubMed]
  3. Park, J.M.; Kim, J.I.; Park, S.Y.; Oh, D.H.; Kim, S.T. Reliability of the gamma index analysis as a verification method of volumetric modulated arc therapy plans. Radiat. Oncol. 2018, 13, 175. [Google Scholar] [CrossRef] [PubMed]
  4. Huang, Y.; Pi, Y.; Ma, K.; Miao, X.; Fu, S.; Chen, H.; Wang, H.; Gu, H.; Shao, Y.; Duan, Y.; et al. Image-based features in machine learning to identify delivery errors and predict error magnitude for patient-specific IMRT quality assurance. Strahlenther. Onkol. 2023, 199, 498–510. [Google Scholar] [CrossRef]
  5. Kui, X.; Liu, F.; Yang, M.; Wang, H.; Liu, C.; Huang, D.; Li, Q.; Chen, L.; Zou, B. A review of dose prediction methods for tumor radiation therapy. Meta-Radiol. 2024, 2, 100057. [Google Scholar] [CrossRef]
  6. Tan, H.S.; Wang, K.; Mcbeth, R. Deep Evidential Learning for Dose Prediction. arXiv 2024, arXiv:2404.17126. [Google Scholar] [CrossRef]
  7. Bi, Q.; Lian, X.; Shen, J.; Zhang, F.; Xu, T. Exploration of radiotherapy strategy for brain metastasis patients with driver gene positivity in lung cancer. J. Cancer 2024, 15, 1994. [Google Scholar] [CrossRef]
  8. Liu, W.; Zhang, L.; Xie, L.; Hu, T.; Li, G.; Bai, S.; Yi, Z. Multilayer perceptron neural network with regression and ranking loss for patient-specific quality assurance. Knowl.-Based Syst. 2023, 271, 110549. [Google Scholar] [CrossRef]
  9. Li, H.; Peng, X.; Zeng, J.; Xiao, J.; Nie, D.; Zu, C.; Wu, X.; Zhou, J.; Wang, Y. Explainable attention guided adversarial deep network for 3D radiotherapy dose distribution prediction. Knowl.-Based Syst. 2022, 241, 108324. [Google Scholar] [CrossRef]
  10. Hu, T.; Xie, L.; Zhang, L.; Li, G.; Yi, Z. Deep multimodal neural network based on data-feature fusion for patient-specific quality assurance. Int. J. Neural Syst. 2022, 32, 2150055. [Google Scholar] [CrossRef]
  11. Han, C.; Zhang, J.; Yu, B.; Zheng, H.; Wu, Y.; Lin, Z.; Ning, B.; Yi, J.; Xie, C.; Jin, X. Integrating plan complexity and dosiomics features with deep learning in patient-specific quality assurance for volumetric modulated arc therapy. Radiat. Oncol. 2023, 18, 116. [Google Scholar] [CrossRef]
  12. Huang, Y.; Pi, Y.; Ma, K.; Miao, X.; Fu, S.; Zhu, Z.; Cheng, Y.; Zhang, Z.; Chen, H.; Wang, H.; et al. Deep learning for patient-specific quality assurance: Predicting gamma passing rates for IMRT based on delivery fluence informed by log files. Technol. Cancer Res. Treat. 2022, 21, 15330338221104881. [Google Scholar] [CrossRef]
  13. Li, C.; Su, Z.; Li, B.; Sun, W.; Wu, D.; Zhang, Y.; Li, X.; Xie, Z.; Huang, J.; Wei, Q. Plan complexity and dosiomics signatures for gamma passing rate classification in volumetric modulated arc therapy: External validation across different LINACs. Phys. Med. 2025, 133, 104962. [Google Scholar] [CrossRef] [PubMed]
  14. Sun, W.; Mo, Z.; Li, Y.; Xiao, J.; Jia, L.; Huang, S.; Liao, C.; Du, J.; He, S.; Chen, L.; et al. Machine learning-based ensemble prediction model for the gamma passing rate of VMAT-SBRT plan. Phys. Med. 2024, 117, 103204. [Google Scholar] [CrossRef] [PubMed]
  15. Huang, Y.; Cai, R.; Pi, Y.; Ma, K.; Kong, Q.; Zhuo, W.; Kong, Y. A feasibility study to predict 3D dose delivery accuracy for IMRT using DenseNet with log files. J. X-Ray Sci. Technol. 2024, 32, 1199–1208. [Google Scholar] [CrossRef] [PubMed]
  16. Xu, S.; Cui, M.; Huang, C.; Wang, H.; Hu, D. BalanceBenchmark: A Survey for Multimodal Imbalance Learning. arXiv 2025, arXiv:2502.10816. [Google Scholar]
  17. Du, C.; Teng, J.; Li, T.; Liu, Y.; Yuan, T.; Wang, Y.; Yuan, Y.; Zhao, H. On Uni-Modal Feature Learning in Supervised Multi-Modal Learning. In Proceedings of the 40th International Conference on Machine Learning, Honolulu, HI, USA, 23–29 July 2023; Krause, A., Brunskill, E., Cho, K., Engelhardt, B., Sabato, S., Scarlett, J., Eds.; PMLR, Proceedings of Machine Learning Research. Volume 202, pp. 8632–8656. [Google Scholar]
  18. Hua, C.; Xu, Q.; Bao, S.; Yang, Z.; Huang, Q. ReconBoost: Boosting Can Achieve Modality Reconcilement. arXiv 2024, arXiv:2405.09321. [Google Scholar] [CrossRef]
  19. Peng, X.; Wei, Y.; Deng, A.; Wang, D.; Hu, D. Balanced Multimodal Learning via On-the-Fly Gradient Modulation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 21–24 June 2022; pp. 8238–8247. [Google Scholar]
  20. Xu, R.; Feng, R.; Zhang, S.X.; Hu, D. MMCosine: Multi-Modal Cosine Loss Towards Balanced Audio-Visual Fine-Grained Learning. In Proceedings of the ICASSP 2023—2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Rhodes Island, Greece, 4–10 June 2023; pp. 1–5. [Google Scholar] [CrossRef]
  21. Li, H.; Li, X.; Hu, P.; Lei, Y.; Li, C.; Zhou, Y. Boosting Multi-modal Model Performance with Adaptive Gradient Modulation. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 2–3 October 2023; pp. 22214–22224. [Google Scholar]
  22. Ma, H.; Zhang, Q.; Zhang, C.; Wu, B.; Fu, H.; Zhou, J.T.; Hu, Q. Calibrating Multimodal Learning. In Proceedings of the 40th International Conference on Machine Learning, Honolulu, HI, USA, 23–29 July 2023; Proceedings of Machine Learning Research. Krause, A., Brunskill, E., Cho, K., Engelhardt, B., Sabato, S., Scarlett, J., Eds.; PMLR. 2023; Volume 202, pp. 23429–23450. [Google Scholar]
  23. Wu, N.; Jastrzebski, S.; Cho, K.; Geras, K.J. Characterizing and Overcoming the Greedy Nature of Learning in Multi-modal Deep Neural Networks. In Proceedings of the 39th International Conference on Machine Learning, Baltimore, MD, USA, 17–23 July 2022; Proceedings of Machine Learning Research. Chaudhuri, K., Jegelka, S., Song, L., Szepesvari, C., Niu, G., Sabato, S., Eds.; PMLR. 2022; Volume 162, pp. 24043–24055. [Google Scholar]
  24. Wei, Y.; Hu, D. MMPareto: Boosting Multimodal Learning with Innocent Unimodal Assistance. arXiv 2024, arXiv:2405.17730. [Google Scholar] [CrossRef]
  25. Fan, Y.; Xu, W.; Wang, H.; Wang, J.; Guo, S. PMR: Prototypical Modal Rebalance for Multimodal Learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, USA, 18–22 June 2023; pp. 20029–20038. [Google Scholar]
  26. Wang, W.; Tran, D.; Feiszli, M. What makes training multi-modal classification networks hard? In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 14–19 June 2020; pp. 12695–12705. [Google Scholar]
  27. Winterbottom, T.; Xiao, S.; McLean, A.; Moubayed, N.A. On modality bias in the tvqa dataset. arXiv 2020, arXiv:2012.10210. [Google Scholar] [CrossRef]
  28. Sun, Y.; Mai, S.; Hu, H. Learning to Balance the Learning Rates Between Various Modalities via Adaptive Tracking Factor. IEEE Signal Process. Lett. 2021, 28, 1650–1654. [Google Scholar] [CrossRef]
  29. Huang, Y.; Lin, J.; Zhou, C.; Yang, H.; Huang, L. Modality Competition: What Makes Joint Training of Multi-modal Network Fail in Deep Learning? (Provably). arXiv 2022, arXiv:2203.12221. [Google Scholar] [CrossRef]
  30. Du, C.; Li, T.; Liu, Y.; Wen, Z.; Hua, T.; Wang, Y.; Zhao, H. Improving Multi-Modal Learning with Uni-Modal Teachers. arXiv 2021, arXiv:2106.11059. [Google Scholar]
  31. Wei, Y.; Feng, R.; Wang, Z.; Hu, D. Enhancing multimodal Cooperation via Sample-level Modality Valuation. In Proceedings of the CVPR, Seattle, WA, USA, 17–21 June 2024. [Google Scholar]
  32. Wei, Y.; Li, S.; Feng, R.; Hu, D. Diagnosing and Re-learning for Balanced Multimodal Learning. In Proceedings of the ECCV, Milan, Italy, 29 September–4 October 2024. [Google Scholar]
  33. Yang, Z.; Wei, Y.; Liang, C.; Hu, D. Quantifying and Enhancing Multi-modal Robustness with Modality Preference. arXiv 2024, arXiv:2402.06244. [Google Scholar] [CrossRef]
  34. Miften, M.; Olch, A.; Mihailidis, D.; Moran, J.; Pawlicki, T.; Molineu, A.; Li, H.; Wijesooriya, K.; Shi, J.; Xia, P.; et al. TG 218: Tolerance limits and methodologies for IMRT measurement-based verification QA: Recommendations of AAPM Task Group No. 218. Med. Phys. 2018, 45, e53–e83. [Google Scholar] [CrossRef]
  35. McNiven, A.L.; Sharpe, M.B.; Purdie, T.G. A new metric for assessing IMRT modulation complexity and plan deliverability. Med. Phys. 2010, 37, 505–515. [Google Scholar] [CrossRef]
  36. Du, W.; Cho, S.H.; Zhang, X.; Hoffman, K.E.; Kudchadker, R.J. Quantification of beam complexity in intensity-modulated radiation therapy treatment plans. Med. Phys. 2014, 41, 021716. [Google Scholar] [CrossRef]
  37. Götstedt, J.; Karlsson Hauer, A.; Bäck, A. Development and evaluation of aperture-based complexity metrics using film and EPID measurements of static MLC openings. Med. Phys. 2015, 42, 3911–3921. [Google Scholar] [CrossRef]
  38. Crowe, S.; Kairn, T.; Kenny, J.; Knight, R.; Hill, B.; Langton, C.M.; Trapp, J. Treatment plan complexity metrics for predicting IMRT pre-treatment quality assurance results. Australas. Phys. Eng. Sci. Med. 2014, 37, 475–482. [Google Scholar] [CrossRef]
  39. Crowe, S.; Kairn, T.; Middlebrook, N.; Sutherland, B.; Hill, B.; Kenny, J.; Langton, C.M.; Trapp, J. Examination of the properties of IMRT and VMAT beams and evaluation against pre-treatment quality assurance results. Phys. Med. Biol. 2015, 60, 2587. [Google Scholar] [CrossRef]
  40. Nauta, M.; Villarreal-Barajas, J.E.; Tambasco, M. Fractal analysis for assessing the level of modulation of IMRT fields. Med. Phys. 2011, 38, 5385–5393. [Google Scholar] [CrossRef] [PubMed]
  41. Zhu, H.; Zhu, Q.; Wang, Z.; Yang, B.; Zhang, W.; Qiu, J. Patient-specific quality assurance prediction models based on machine learning for novel dual-layered MLC linac. Med. Phys. 2023, 50, 1205–1214. [Google Scholar] [CrossRef] [PubMed]
  42. Shapley, L. The value of n-person games. Ann. Math. Stud. 1953, 28, 307–317. [Google Scholar]
  43. Molnar, C. Interpretable Machine Learning; Lulu. Press: Morrisville, NC, USA, 2020. [Google Scholar]
  44. Weber, R.J. Probabilistic values for games. The Shapley Value. Essays in Honor of Lloyd S. Shapley; Cambridge University Press: New York, NY, USA, 1988; pp. 101–119. [Google Scholar]
  45. Freund, Y.; Schapire, R.E. Game theory, on-line prediction and boosting. In Proceedings of the Ninth Annual Conference on Computational Learning Theory, Garda, Italy, 28 June–1 July 1996; pp. 325–332. [Google Scholar]
  46. Wang, K.; Gou, C.; Duan, Y.; Lin, Y.; Zheng, X.; Wang, F.Y. Generative adversarial networks: Introduction and outlook. IEEE/CAA J. Autom. Sin. 2017, 4, 588–598. [Google Scholar] [CrossRef]
  47. Gemp, I.; McWilliams, B.; Vernade, C.; Graepel, T. EigenGame: PCA as a Nash Equilibrium. In Proceedings of the International Conference on Learning Representations, Addis Ababa, Ethiopia, 30 April 2020. [Google Scholar]
  48. Hu, P.; Li, X.; Zhou, Y. Shape: An unified approach to evaluate the contribution and cooperation of individual modalities. arXiv 2022, arXiv:2205.00302. [Google Scholar] [CrossRef]
  49. Ono, T.; Hirashima, H.; Iramina, H.; Mukumoto, N.; Miyabe, Y.; Nakamura, M.; Mizowaki, T. Prediction of dosimetric accuracy for VMAT plans using plan complexity parameters via machine learning. Med. Phys. 2019, 46, 3823–3832. [Google Scholar] [CrossRef]
  50. Lay, L.M.; Chuang, K.C.; Wu, Y.; Giles, W.; Adamson, J. Virtual patient-specific QA with DVH-based metrics. J. Appl. Clin. Med. Phys. 2022, 23, e13639. [Google Scholar] [CrossRef]
  51. Thongsawad, S.; Srisatit, S.; Fuangrod, T. Predicting gamma evaluation results of patient-specific head and neck volumetric-modulated arc therapy quality assurance based on multileaf collimator patterns and fluence map features: A feasibility study. J. Appl. Clin. Med. Phys. 2022, 23, e13622. [Google Scholar] [CrossRef]
  52. Valdes, G.; Scheuermann, R.; Hung, C.; Olszanski, A.; Bellerive, M.; Solberg, T. A mathematical framework for virtual IMRT QA using machine learning. Med. Phys. 2016, 43, 4323–4334. [Google Scholar] [CrossRef]
  53. Huang, Y.; Pi, Y.; Ma, K.; Miao, X.; Fu, S.; Chen, H.; Wang, H.; Gu, H.; Shan, Y.; Duan, Y.; et al. Virtual patient-specific quality assurance of IMRT using UNet++: Classification, gamma passing rates prediction, and dose difference prediction. Front. Oncol. 2021, 11, 700343. [Google Scholar] [CrossRef]
  54. Yoganathan, S.; Ahmed, S.; Paloor, S.; Torfeh, T.; Aouadi, S.; Al-Hammadi, N.; Hammoud, R. Virtual pretreatment patient-specific quality assurance of volumetric modulated arc therapy using deep learning. Med. Phys. 2023, 50, 7891–7903. [Google Scholar] [CrossRef]
  55. Radosavovic, I.; Kosaraju, R.P.; Girshick, R.; He, K.; Dollár, P. Designing network design spaces. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 10428–10436. [Google Scholar]
  56. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar] [CrossRef]
  57. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 22–25 July 2017; pp. 4700–4708. [Google Scholar]
  58. Mehta, S.; Rastegari, M. Mobilevit: Light-weight, general-purpose, and mobile-friendly vision transformer. arXiv 2021, arXiv:2110.02178. [Google Scholar]
  59. Suwanraksa, C.; Bridhikitti, J.; Liamsuwan, T.; Chaichulee, S. CBCT-to-CT translation using registration-based generative adversarial networks in patients with head and neck cancer. Cancers 2023, 15, 2017. [Google Scholar] [CrossRef] [PubMed]
  60. Liu, F.; Liu, J.; Fang, Z.; Hong, R.; Lu, H. Densely Connected Attention Flow for Visual Question Answering. In Proceedings of the IJCAI, Macao, China, 10–16 August 2019; pp. 869–875. [Google Scholar]
  61. Xu, J.; Pan, Y.; Pan, X.; Hoi, S.; Yi, Z.; Xu, Z. RegNet: Self-regulated network for image classification. IEEE Trans. Neural Netw. Learn. Syst. 2022, 34, 9562–9567. [Google Scholar] [CrossRef]
  62. Zeng, L.; Zhang, M.; Zhang, Y.; Zou, Z.; Guan, Y.; Huang, B.; Yu, X.; Ding, S.; Liu, Q.; Gong, C. TransQA: Deep hybrid transformer network for measurement-guided volumetric dose prediction of pre-treatment patient-specific quality assurance. Phys. Med. Biol. 2023, 68, 205010. [Google Scholar] [CrossRef]
  63. Yang, X.; Li, S.; Shao, Q.; Cao, Y.; Yang, Z.; Zhao, Y.Q. Uncertainty-guided man–machine integrated patient-specific quality assurance. Radiother. Oncol. 2022, 173, 1–9. [Google Scholar] [CrossRef]
Figure 1. The training loss curve of different modalities.
Figure 1. The training loss curve of different modalities.
Diagnostics 15 02555 g001
Figure 2. Sample GPR and lesion distribution.
Figure 2. Sample GPR and lesion distribution.
Diagnostics 15 02555 g002
Figure 3. Representative IMRT QA results for four treatment plans. (Left) TPS-calculated dose distribution. (Middle) Measured dose via portal dosimetry. (Right) Absolute dose differences. Pixel intensity corresponds to delivered dose (Gy).
Figure 3. Representative IMRT QA results for four treatment plans. (Left) TPS-calculated dose distribution. (Middle) Measured dose via portal dosimetry. (Right) Absolute dose differences. Pixel intensity corresponds to delivered dose (Gy).
Diagnostics 15 02555 g003
Figure 4. Architecture diagram of the proposed BMMQA.
Figure 4. Architecture diagram of the proposed BMMQA.
Diagnostics 15 02555 g004
Figure 5. Comparison across the remaining two model types: (A) tabular only and (B) image only.
Figure 5. Comparison across the remaining two model types: (A) tabular only and (B) image only.
Diagnostics 15 02555 g005
Figure 6. Comparison across the remaining two model types: (A) unbalanced multimodal and (B) balanced multimodal.
Figure 6. Comparison across the remaining two model types: (A) unbalanced multimodal and (B) balanced multimodal.
Diagnostics 15 02555 g006
Figure 7. Residual distributions of different models in the DDP task. Each row corresponds to a different backbone (ResNet, DenseNet, MobileViT, RegNet), and each column represents a different modality configuration: balanced multimodal (BMM), unbalanced multimodal (MM), and image-only.
Figure 7. Residual distributions of different models in the DDP task. Each row corresponds to a different backbone (ResNet, DenseNet, MobileViT, RegNet), and each column represents a different modality configuration: balanced multimodal (BMM), unbalanced multimodal (MM), and image-only.
Diagnostics 15 02555 g007
Figure 8. Dose difference error maps for selected samples in the DDP task. Each row corresponds to a different backbone (ResNet, DenseNet, MobileViT, RegNet), and each column represents a different modality configuration: balanced multimodal (BMM), unbalanced multimodal (MM), and image-only. Darker red areas indicate larger differences, whereas lighter blue areas indicate smaller differences.
Figure 8. Dose difference error maps for selected samples in the DDP task. Each row corresponds to a different backbone (ResNet, DenseNet, MobileViT, RegNet), and each column represents a different modality configuration: balanced multimodal (BMM), unbalanced multimodal (MM), and image-only. Darker red areas indicate larger differences, whereas lighter blue areas indicate smaller differences.
Diagnostics 15 02555 g008
Figure 9. Modality contribution curves for different backbones.
Figure 9. Modality contribution curves for different backbones.
Diagnostics 15 02555 g009
Figure 10. DDP error (MAE) across different values of r in the computation of λ DDP . Each curve corresponds to a different GPR threshold.
Figure 10. DDP error (MAE) across different values of r in the computation of λ DDP . Each curve corresponds to a different GPR threshold.
Diagnostics 15 02555 g010
Table 1. Summary of multimodal PSQA methods. Symbols: ✓ = Yes; ✗ = No.
Table 1. Summary of multimodal PSQA methods. Symbols: ✓ = Yes; ✗ = No.
ReferenceCategoryDescription (Modalities + Fusion Strategy)Fine-Grained Task?Modality Balance?
[11]Early FusionDose matrices encoded via DenseNet-121 (image) and plan complexity features (tabular) are separately reduced using PCA/Lasso and concatenated for Random Forest regression.
[13]Early FusionPlan and dosimetric features (tabular) are selected via Lasso and used as input to a logistic regression model for classification.
[14]Late FusionTwo SVMs are trained separately on radiomics (image-like features) and plan complexity features; predictions are aggregated to improve GPR classification and regression.
[10]Late FusionDose distribution matrices (image) and MU values (tabular) are processed by separate models; outputs are aggregated at the decision level as a baseline fusion strategy.
[10]Intermediate Fusion3D ResNet encodes MLC aperture images; spatial features are concatenated one-to-one with MU values (tabular), jointly optimized in an end-to-end model (FDF).
[12]Intermediate FusionDose distribution is compressed into a 99D vector and concatenated with MU value; the fused vector is passed to fully connected layers.
[15]Intermediate FusionMU-weighted fluence maps generated from delivery log files are used as image input to DenseNet for GPR prediction under multiple gamma criteria.
BMMQAIntermediate FusionDual-modality input (dose matrix + tabular metrics); image encoded by a pre-trained CNN, tabular by MLP; fusion uses task-specific heads: attention-weighted feature fusion for GPR, spatial concatenation for DDP; Shapley-based diagnostics and adaptive loss scaling address modality imbalance.
Table 2. Treatment plans by site.
Table 2. Treatment plans by site.
Treatment SiteAbbreviationNumber of Plans
Head and NeckH&N19
ChestC138
AbdomenA31
PelvisP22
Table 3. Selected complexity metrics, their references, and one sample in each lesion, namely, Head and Neck (H&N), Chest (C), Abdominal (A), and Pelvic (P).
Table 3. Selected complexity metrics, their references, and one sample in each lesion, namely, Head and Neck (H&N), Chest (C), Abdominal (A), and Pelvic (P).
TypeRefMetric NameH&NCAP
Complexity
Metric
(Input
Tabular
Modality)
[35]Aperture area variability of distal MLC (d-AAV)0.300.420.350.43
Aperture area variability of proximal MLC (p-AAV)0.310.440.410.43
Leaf sequence variability of distal MLC (d-LSV)2.493.176.814.71
Leaf sequence variability of proximal MLC (p-LSV)1.441.663.481.88
Modulation complex score of distal MLC (d-MCS)1.380.900.970.82
Modulation complex score of proximal MLC (p-MCS)1.210.700.680.73
[36]Beam area of distal MLC (d-BA)0.270.560.340.38
Beam area of proximal MLC (p-BA)0.220.530.370.35
Union aperture areas of distal MLC (d-UAA)20.4522.8416.8935.90
Union aperture areas of proximal MLC (p-UAA)15.4320.4518.4432.26
Plan irregularity of distal MLC (d-PI)32.9866.9988.1766.18
Plan irregularity of proximal MLC (p-PI)31.5766.8492.3368.85
Plan modulation of distal MLC (d-PM)0.700.580.650.57
Plan modulation of proximal MLC (p-PM)0.690.560.590.57
Complexity
Metric
(Input
Tabular
Modality)
[37]Circumference/area of distal MLC (d-C/A)0.361.011.121.00
Circumference/area of proximal MLC (p-C/A)0.341.021.321.01
[38]Closed leaf score of distal MLC (d-CLS)0.651.071.791.54
Closed leaf score of proximal MLC (p-CLS)0.661.082.111.55
Mean asymmetry distance of distal MLC (d-MAD)0.780.790.780.80
Mean asymmetry distance of proximal MLC (p-MAD)0.770.810.790.79
[39]Small aperture score 2 mm of distal MLC (d-SAS-2 mm)0.230.340.260.34
Small aperture score 5 mm of distal MLC (d-SAS-5 mm)0.240.360.310.34
Small aperture score 10 mm of distal MLC (d-SAS-10 mm)263.53195.03191.07172.01
Small aperture score 15 mm of distal MLC (d-SAS-15 mm)10.0128.4331.0728.29
Small aperture score 2 mm of proximal MLC (p-SAS-2 mm)9.8729.6238.1729.59
Small aperture score 5 mm of proximal MLC (p-SAS-5 mm)0.190.150.110.13
Small aperture score 10 mm of proximal MLC (p-SAS-10 mm)0.100.080.110.08
Small aperture score 15 mm of proximal MLC (p-SAS-15 mm)0.260.190.130.17
[40]Average leaf gap of distal MLC (d-ALG)0.160.140.150.13
Average leaf gap of proximal MLC (p-ALG)0.490.370.230.25
Standard deviation of leaf gap of distal MLC (d-SLG)0.310.330.200.22
Standard deviation of leaf gap of proximal MLC (p-SLG)0.680.530.310.39
[41]MU value of each beam (MU)0.590.480.310.37
QA Metric
(Target)
Gamma Passing Rate (1%/1mm)91.8093.7299.1597.90
Gamma Passing Rate (2%/2mm)96.5896.7899.7899.05
Gamma Passing Rate (3%/2mm)98.3198.1399.9599.37
Table 4. Mathematical symbol table.
Table 4. Mathematical symbol table.
SymbolDescription
X = { X img , X tab } A single input sample, consisting of an image modality X img R H × W and a tabular modality X tab R 33 .
Y gpr , Y ddp Ground truths for GPR and DDP tasks, respectively. Y gpr R 3 ; Y ddp R H × W .
Z img Latent feature map from the image encoder; shape: N c × H z × W z . Requires pooling before projection.
Z tab Latent tabular embedding vector produced by the tabular encoder; shape: N t .
Z gpr , Z ddp Fused latent representations used in GPR and DDP heads, respectively.
W Z tab Matrix multiplication between projection matrix and tabular embedding vector.
W Z img Matrix multiplication between the projection matrix and the pooled image feature map. Due to the complexity introduced by both multi-task and multimodal notation, spatial pooling over Z img R N c × H × W is applied implicitly for notational simplicity and consistency, resulting in an N c -dimensional vector. This convention is used throughout the paper, including expressions such as W s img Z img .
α img , α tab Attention weights for the image and tabular modalities, computed via softmax over scalar scores.
P Power set of modality subsets considered for ablation: , { X img } , { X tab } , { X img , X tab } .
Y ^ gpr S GPR prediction computed using only the modality subset S P .
Y ^ gpr , Y ^ ddp Model predictions for GPR and DDP tasks using all available modalities by default, namely Y ^ gpr = Y ^ gpr ( img , tab ) R 3 ; Y ^ ddp = Y ^ ddp ( img , tab ) R H × W .
v ( S ) Value function for modality subset S P , defined as the negative MAE of predictions Y ^ S over the validation set. Used to compute Shapley-style contributions.
ϕ img , ϕ tab Modality contributions computed via two-modality Shapley decomposition based on v ( S ) .
λ ddp Adaptive weighting factor for the DDP loss, derived from modality contributions.
L gpr , L ddp Loss terms for the GPR and DDP tasks.
L Total training loss: L = L gpr + λ ddp · L ddp .
Table 5. Parameter settings.
Table 5. Parameter settings.
ParameterDescriptionValue
Resized Dose Plan Array ShapeTo organize dose plan array samples into a batch, we perform ResizeWithPadOrCrop preprocessing[512, 512]
Regression Head ArchitectureArchitecture of GPR prediction output; the output dimension is 3 since we select three GPR criteria [1%/1 mm, 2%/2 mm, 2%/3 mm]AdaptiveAvgPool2d[1, 1]
Flatten()
Dropout(p = 0.1)
Linear[512, 3]
Decoder (DDP) ArchitectureU-Net decoder channel dimensions[768, 384, 192, 128, 32]
Dropout RateThe dropout rate employed in network0.1
OptimizerType of optimizer usedAdam
rThe hyperparameters used in the computation of λ ddp are set to their default values unless otherwise specified16
EpochsNumber of training epochs30
Batch SizeSize of the training batches32
Learning RateLearning rate for the optimizer 5 × 10 4
Dataset Split Ratio (EXP-A and EXP-B)Train:Validation:Test split ratioMulti-factor stratified sampling based on Figure 2, with a 7:1:2 split.
Table 6. GPR performance comparison.
Table 6. GPR performance comparison.
CategoryMethodCriteriaGPR MAE (%)GPR RMSE (%) R 2
All95–10090–95<90All95–10090–95<90
Tabular ModalityGBDT [41]1%/1 mm1.86661.62461.48053.30342.67502.34971.90994.37050.6055
2%/2 mm1.23710.91541.95784.34611.83171.28052.35005.29140.5992
2%/3 mm0.87980.71882.20477.88811.34341.00642.63067.89110.5740
RF [41]1%/1 mm2.09111.75361.52554.16213.01462.28121.89695.51190.5305
2%/2 mm1.35540.93891.94127.05232.11101.26552.22757.77290.5030
2%/3 mm0.95150.69373.080410.49881.56280.94913.277710.50930.4784
PL [41]1%/1 mm2.06051.68591.79213.46423.10272.40752.69534.79860.4821
2%/2 mm1.36731.02891.99155.28112.13531.47062.62796.50240.4697
2%/3 mm0.96640.78332.29659.83071.57071.10722.91729.83890.4349
Image ModalityUNET++ [53]1%/1 mm1.77831.38751.53263.17092.72712.21032.07264.49580.6165
2%/2 mm1.12160.82691.46845.44981.86831.28491.83516.44540.6218
2%/3 mm0.77040.61261.81089.24421.35040.93242.28409.29300.6258
Ranking Loss [8]1%/1 mm1.76041.24481.42863.61562.73921.85431.90045.01890.6286
2%/2 mm1.11310.80301.55275.30851.89411.25021.96556.59410.6157
2%/3 mm0.75120.59391.76439.46021.37030.89792.46799.60210.6129
TransQA [62]1%/1 mm1.73081.43481.35193.24772.67302.31001.77164.52370.6344
2%/2 mm1.10600.77951.56275.55611.87721.23541.91026.60280.6297
2%/3 mm0.77870.60491.97319.53951.38740.91782.50519.59510.6171
RegNet [61]1%/1 mm1.75761.51071.26433.44722.70982.17451.69294.84230.6302
2%/2 mm1.07290.73541.58065.49911.83291.14091.92076.60900.6423
2%/3 mm0.72320.56991.68699.49601.31910.88562.19249.49940.638
Unbalanced
Multimodality
ResNet18 [63]1%/1 mm1.75901.27811.58383.16782.66951.99352.03684.55460.6435
2%/2 mm1.08410.80231.41185.24291.83831.22701.91206.34090.6371
2%/3 mm0.75870.59001.93599.06211.36040.91052.50849.11530.6204
DenseNet121 [53]1%/1 mm1.71631.43121.26473.38492.59832.08651.68894.58530.6701
2%/2 mm1.09430.71431.76795.58571.84881.12962.10926.48630.6681
2%/3 mm0.79780.60842.26948.40391.37430.90642.65258.99470.6399
MobileVit [62]1%/1 mm1.73831.39951.41733.20422.61782.05521.86794.50780.6577
2%/2 mm1.10210.76561.62215.45031.83381.16821.96836.44900.6443
2%/3 mm0.76660.58652.01169.76781.37610.89952.44439.76810.6208
RegNet [61]1%/1 mm1.72311.59931.26903.06602.63642.26521.74084.48180.6493
2%/2 mm1.07090.78641.32645.63301.84951.24931.71696.60100.6314
2%/3 mm0.74620.56412.06159.21111.34940.86602.53749.34950.6356
Balanced
Multimodality
ResNet18 [63]1%/1 mm1.70471.36241.28072.97092.57992.01781.67994.32620.6629
2%/2 mm1.04370.76191.23724.97181.78941.15331.67036.09510.6512
2%/3 mm0.73230.58171.75988.42761.29790.89492.28058.70050.6453
DenseNet121 [53]1%/1 mm1.65541.48551.16003.15542.61452.14501.54944.69660.6631
2%/2 mm1.03850.75991.26115.37751.85311.19581.71336.77080.6357
2%/3 mm0.74460.56131.98639.32011.37670.87442.57599.68520.6224
MobileVit [62]1%/1 mm1.73221.21821.27592.80212.57361.91301.66954.19850.6609
2%/2 mm1.07790.70031.30065.01961.78861.04641.74846.06300.6551
2%/3 mm0.76720.61191.74898.06221.32830.90182.24188.53630.6354
RegNet [61]1%/1 mm1.69251.31511.05873.34352.64481.97371.37714.79890.6689
2%/2 mm1.08140.73071.48905.60021.83451.08261.87536.65430.6770
2%/3 mm0.75820.55532.02039.15341.31980.82082.52499.26360.6814
Table 7. DDP performance comparison.
Table 7. DDP performance comparison.
CategoryMethodDDP MAE (%)DDP SSIM R 2
All95–10090–95<90All95–10090–95<90
Image ModalityUNET++ [53]4.76765.30734.82173.52830.95270.95260.95180.95510.9753
Ranking Loss [8]4.59945.15284.63683.37190.95590.95550.95520.95850.9784
TransQA [62]4.64815.21124.69733.37180.95990.95910.95920.96340.9793
RegNet [61]5.08105.65545.12563.79270.93700.93470.93640.94320.9793
Unbalanced
Multimodality
ResNet18 [63]4.86045.23714.79994.23090.94900.95530.95310.92600.9748
DenseNet121 [53]4.80915.28194.77403.92070.95510.96090.95830.93560.9764
MobileVit [62]4.87845.44694.93943.56310.95680.95540.95610.96130.9774
RegNet [61]4.78145.35584.81833.51160.95540.95330.95510.96040.9805
Balanced
Multimodality
ResNet18 [63]4.30274.72724.35883.29530.95350.96000.95750.93670.9773
DenseNet121 [53]4.02204.45453.99423.19930.96400.96500.96560.95810.9751
MobileVit [62]4.43874.74514.35193.82500.95430.95980.95650.94530.9701
RegNet [61]4.27454.75804.31123.19210.95790.95810.95730.95890.9778
Table 8. Comparative analysis of computational resource utilization by deep learning baselines and the proposed BMMQA.
Table 8. Comparative analysis of computational resource utilization by deep learning baselines and the proposed BMMQA.
Method NameFlops (GMac)Param Size (M)
Unet++63.715.9
TranQA240.34215.38
DenseNet6.414.62
BMMQA6.514.72
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zeng, X.; Ahmed, A.; Tunio, M.H. A Balanced Multimodal Multi-Task Deep Learning Framework for Robust Patient-Specific Quality Assurance. Diagnostics 2025, 15, 2555. https://doi.org/10.3390/diagnostics15202555

AMA Style

Zeng X, Ahmed A, Tunio MH. A Balanced Multimodal Multi-Task Deep Learning Framework for Robust Patient-Specific Quality Assurance. Diagnostics. 2025; 15(20):2555. https://doi.org/10.3390/diagnostics15202555

Chicago/Turabian Style

Zeng, Xiaoyang, Awais Ahmed, and Muhammad Hanif Tunio. 2025. "A Balanced Multimodal Multi-Task Deep Learning Framework for Robust Patient-Specific Quality Assurance" Diagnostics 15, no. 20: 2555. https://doi.org/10.3390/diagnostics15202555

APA Style

Zeng, X., Ahmed, A., & Tunio, M. H. (2025). A Balanced Multimodal Multi-Task Deep Learning Framework for Robust Patient-Specific Quality Assurance. Diagnostics, 15(20), 2555. https://doi.org/10.3390/diagnostics15202555

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop