Next Article in Journal
Geometric Reasoning in the Embedding Space
Previous Article in Journal
Machine Learning Prediction Models of Beneficial and Toxicological Effects of Zinc Oxide Nanoparticles in Rat Feed
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Prediction Model for Multimodal Medical Data Based on Graph Neural Networks

1
School of Public Health, Peking University, Beijing 100871, China
2
Department of Medical Oncology, National Cancer Center/National Clinical Research Center for Cancer/Cancer Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing 100730, China
3
State Key Laboratory of Networking & Switching Technology, Beijing University of the Posts and Telecommunications, Beijing 100876, China
4
International School, Beijing University of Posts and Telecommunications, Beijing 100876, China
5
Petroleum Institute, China University of Petroleum (Beijing), Karamay Campus, Karamay 834000, China
6
Sloan School of Management, Massachusetts Institute of Technology, Cambridge, MA 02139, USA
7
Center for Statistics and Data Science, Massachusetts Institute of Technology, Cambridge, MA 02139, USA
*
Author to whom correspondence should be addressed.
Mach. Learn. Knowl. Extr. 2025, 7(3), 92; https://doi.org/10.3390/make7030092
Submission received: 10 June 2025 / Revised: 24 August 2025 / Accepted: 26 August 2025 / Published: 2 September 2025

Abstract

Multimodal medical data provides a wide and real basis for disease diagnosis. Computer-aided diagnosis (CAD) powered by artificial intelligence (AI) is becoming increasingly prominent in disease diagnosis. CAD for multimodal medical data requires addressing the issues of data fusion and prediction. Traditionally, the prediction performance of CAD models has not been good enough due to the complicated dimensionality reduction. Therefore, this paper proposes a fusion and prediction model—EPGC—for multimodal medical data based on graph neural networks. Firstly, we select features from unstructured multimodal medical data and quantify them. Then, we transform the multimodal medical data into a graph data structure by establishing each patient as a node, and establishing edges based on the similarity of features between the patients. Normalization of data is also essential in this process. Finally, we build a node prediction model based on graph neural networks and predict the node classification, which predicts the patients’ diseases. The model is validated on two publicly available datasets of heart diseases. Compared to the existing models that typically involve dimensionality reduction, classification, or the establishment of complex deep learning networks, the proposed model achieves outstanding results with the experimental dataset. This demonstrates that the fusion and diagnosis of multimodal data can be effectively achieved without dimension reduction or intricate deep learning networks. We take pride in exploring unstructured multimodal medical data using deep learning and hope to make breakthroughs in various fields.

1. Introduction

Advancements in medical technology are making disease detection more scientific and refined, providing crucial data that accurately reflects the true and direct state of diseases [1,2]. This is particularly important for diagnosing complex diseases with multiple pathogenic factors. In cases where complex diseases with multiple contributing factors and high pathogenic hazards are involved, combining multiple detection methods is essential to ensuring a safe and accurate diagnosis process [3]. By leveraging the strengths of each detection method and avoiding their potential limitations, healthcare professionals can achieve more reliable and comprehensive diagnostic results, reducing the risk of misdiagnosis or overlooking critical information. Inevitably, numerous multimodal medical data will be generated by various aspects of this medical process, such as intelligent decision support systems [4], Clinical Concept [5], oncology [6], medical imageology [7], and so on.
Multimodal medical data refers to medical data from two or more sources or storage formats, such as clinical data, imaging, genetics, pathology, molecular biology, etc. [8]. Some researchers have studied the characteristics of multimodal medical data. Kline et al. [9] described the diversity of forms and modalities of multimodal medical data in their study. Stahlschmidt et al. [10] discussed the complexity and multidimensionality of multimodal medical data. Azam et al. [11] and Guo et al. [12] also emphasized the diversity of formats of multimodal medical data in medical imageology research. The processing of multimodal medical data faces challenges, such as data fusion, dimensionality reduction, and feature complexity reduction [13]. Chen et al. [14] proposed a prediction model for breast cancer prognosis based on the clinical and gene expression data. This model combines multiple feature selection with an attention mechanism to extract features and reduce feature complexity, then builds a deep neural network to predict the breast cancer prognosis. Yang et al. [15] proposed a multimodal medical data diagnosis model for clinical and breast cancer medical image data. The model design reduces the complexity of clinical data through feature selection and logical regression classification based on a random forest, and combines a multi-level deep neural network to extract the medical image features to finally predict the recurrence and metastasis risk of HER2-positive breast cancer. Alizadehsani et al. [16] used a paradigm of “feature selection + dimensionality reduction + prediction” to study multimodal data fusion and diagnosis of coronary artery disease, which represents the mainstream approach in the current research. In a 2018 study [16], Alizadehsani et al. firstly extracted features from ultrasound and echo data and then used a “dimensionality reduction + prediction” approach to sort the important features to achieve the goal of reducing complexity. Based on the sorting results, the important feature combinations were selected and input into a machine learning classifier for coronary artery disease classification. Golovenkin et al. [17] developed an intelligent evaluation system for predicting early complications of myocardial infarction. After feature selection, the system reduced the dimensionality of high-dimensional and multimodal cardiovascular detection data, and built a classification model based on multi-layer deep networks. Finally, the classification of complications of myocardial infarction was achieved, with an average accuracy of over 90%. In a 2022 study [18], Yavru et al. also used a paradigm of “feature selection + dimensionality reduction + prediction” to study the classification of complications on the same dataset used by Golovenkin et al. They designed a multimodal data fusion and disease classification model based on deep learning. The overall accuracy of the model classification was improved to 92%. In addition, Hu et al. [19] developed an Enterprise Service Bus (ESB) integration platform for multimodal medical data fusion from the perspective of hardware research on multimodal medical data fusion. This platform starts from the interface and data format standardization of various medical data platforms, builds a unified data output integration platform at the source of data generation, realizes the fusion of multimodal diagnosis and treatment data, and thus builds an intelligent decision-making system that improves the utilization of medical resources. This approach requires large-scale upgrades and iterations of existing hardware devices, which will face significant challenges in terms of economic cost and operability.
To summarize, the current methods for processing multimodal medical data typically involve initially extracting features from all the data, followed by applying traditional machine learning algorithms or deep learning networks for tasks such as classification and clustering. These methods can be summarized as following a pattern of “feature selection + dimensionality reduction + prediction”. However, these methods present several issues. Traditional machine learning algorithms or neural network models often exhibit poor classification performance when applied to high-dimensional data that has undergone feature selection, necessitating the need for data dimensionality reduction. The outcome of the dimensionality reduction process possesses a certain degree of randomness and is influenced by the final classification result. As a result, the interpretability of the dimensionality reduction result is limited when evaluated in reverse based on the final disease classification result. Additionally, the dimensionality reduction may lead to the loss of critical information within the data. Furthermore, the data’s dimension is often overlooked during processing, potentially resulting in the loss of physical information within the data. In addition, complex deep learning networks require a large amount of model training data and training resources (time resources, computing power resources) to support them. This paper proposes a new algorithm based on a graph neural network (GNN) for processing multimodal medical data. Compared to traditional methods, it can leverage the expressive advantage of graph data to preserve the information in the data as much as possible without a dimensionality reduction, and achieves the fusion and disease prediction of multimodal medical data.

2. Materials and Methods

As mentioned above, we need to solve the problems of diverse data modalities and complex data structures, while trying to avoid the loss of useful information as much as possible. We will introduce graph-structured data theory for solving the above problems.

2.1. Graph Data Structure for Feature Expression

Conventional AI algorithms, including traditional neural networks and machine learning algorithms, usually process Euclidean space data with typical spatial structures. In Euclidean space data, for example, text and speech data are one-dimensional sequence data structures [20], and image data [21,22] are typically rectangular structures. However, in reality, much data do not have the typical spatial structures belonging to non-Euclidean spaces, namely a graph data structure (GDS), for example, knowledge maps, sensor networks in communications, regulatory networks in genetics, and so on. In a GDS, the connections between nodes may not be fixed, forming a complex topology without a fixed node order, or the data present a dynamic multimodal structure.
A GDS typically consists of two important components, i.e., nodes and edges, which are used to represent the relationships between objects (or entities). The graph structure contains a set of nodes and edges that connect these nodes (the edges here can be directional or undirected). Graph structures have a stronger ability to express features [23,24,25]. A node can serve as a unit for storing entity information; that is to say, the stored information contains the feature description of the entity. And edges can also store information; the storage method is represented by the direction and thickness of the edges. Taking an undirected graph as an example, the feature expression ability of the GDS is shown in Equation (1). Specifically, the undirected graph G = ( V , E ) , where V and E represent the nodes and the edges in the graph, respectively. Vn describes the n-th node in the GDS with n nodes, v f n describes the n-th feature of n-th node, e m n describes the edge between the m-th and n-th nodes, and w m n describes the connection weight between the m-th and n-th nodes.
G = ( V , E ) V = { V 1 , V 2 , , V n } E = { E 1 , E 2 , ,   E n } V n = { v f 1 , v f 2 , , v f n ) E n = e 11 e 1 m e n 1 e nm w 11 w 1 m w n 1 w n m

2.2. GDS Construction Based on Similarity Measurement

In this paper, we propose a GDS construction method based on a similarity measurement, which constructs a GDS where each patient is represented as a node, and the similarity between patients is described by the edges, as shown in Figure 1. All patients exist in a GDS as the nodes. However, the GDS is not fully connected, and whether a node has neighbors, that is, whether there are edges between the nodes, is based on their similarity measure. If they have a strong correlation, edges between them exist; otherwise, they do not exist.
In statistics, a similarity measure (or similarity function) is crucial for quantifying the similarity between different samples or sets [26]. Common methods include the Pearson correlation coefficient, Euclidean Distance [27], Manhattan Distance, Chebyshev Distance, and Information Entropy [28]. The Pearson correlation coefficient assesses the degree of correlation between two variables and is widely used across various fields, such as machine learning, statistics, and data analysis, for evaluating similarity. The range of the Pearson correlation coefficient r is r [ 1,1 ] . When r = 1 , it indicates that the two variables (sets) are completely negatively correlated. When the Pearson correlation coefficient r = 1 , it indicates that the two variables (sets) are completely positively correlated. The Pearson correlation is shown in Equation (2), where X and Y represent two variables, and X ¯ and Y ¯ , respectively, represent the standard deviation of variable X and variable Y.
r = i = 1 n ( X i X ¯ ) ( Y i Y ¯ ) i = 1 n ( X i X ¯ ) 2 i = 1 n ( Y i Y ¯ ) 2

2.3. The Learning of GDS Using a Graph Neural Network

A GNN is a type of neural network proposed by Gori et al. in 2005 [29]. Compared with conventional neural networks, its advantage is its ability to process a GDS. A GNN can collect and extract the features from nodes and edges in a GDS, and construct a corresponding feature expression model. A GNN can achieve model learning tasks [30], including classification [31], prediction [32], segmentation [33,34], etc. GNNs have been proven to be effective for disease prediction. Mohanraj et al. [35] proposed a new method for weighted feature extraction and GNN classification based on a Graph Wavelet Transform. By calculating the weighted correlation between the data and utilizing machine learning algorithms, it can effectively predict the severity of Parkinson’s disease and significantly improve the prediction accuracy. Hossain et al. [36] introduced a comorbidity network approach to assess the risk of cardiovascular disease in patients with type 2 diabetes, utilizing features derived from foundational networks. Meanwhile, the research by Xuan et al. [37] and Lee et al. [38] suggests that GNNs can be applied to analyze medical images, such as X-rays, CT scans, and MRIs, to detect potential abnormalities or diseases.
GraphSAGE (Graph Sample and aggreGatE) is a type of spatial-based GCN [39] with a strong generalization ability and aggregation operation. A GraphSAGE model mainly completes three stages of learning tasks, namely, sampling, aggregation, and node embedding. Object node(s) are determined through multi-level sampling and aggregating of adjacent nodes, then the features are captured and fused through the aggregate function. The sampling task collects the node information and connection information between the nodes. The aggregation function performs the aggregation task with a certain rule and generates a feature vector. Finally, the feature vector is processed as the node embedding, and input into the downstream classifiers to achieve the learning task. Figure 2 depicts the node classification task of GraphSAGE, and the following is an introduction to the detailed learning process of the model.
  • Sampling: The model first samples the neighborhood node information of the original object node. The rule of sampling is as follows: firstly, set one is the original object node, and the remaining n nodes are the neighboring nodes. Set m as a fixed number of nodes for each sampling, and if n > = m, perform sampling without dropout; otherwise, perform sampling with dropout until m nodes are sampled. It is worth noting that GraphSage can achieve significant generalization of non-adjacent nodes without the involvement of the whole graph structure during the sampling process.
  • Aggregation: For the sampled information, the next step is to aggregate it according to a specific rule. This process involves combining the feature vectors of a node and its neighbors, along with their respective weights, to capture the graph’s structure and generate new neighborhood embeddings. The aggregator function is described by Equation (3). In a GraphSAGE model, there are three primary types of aggregators: the mean aggregator, LSTM aggregator, and pooling aggregator. Additionally, convolution-based GCN operators can also be utilized as aggregators.
    h v k = σ [ W ( k ) c o n c a t ( h v k 1 , A G G R E G A T E k ( h u k 1 u N ( v ) ) ) ]
    where σ represents the non-linear activation function, W k represents the weight matrix of k aggregators, c o n c a t ( · ) represents vector concatenation, and h v k represents the feature vector of the node.
  • Prediction: The results will be input into the downstream machine learning classifiers for training, enabling node prediction.

2.4. Disease Prediction Model Based on GraphSAGE

A new algorithm, EPGC (feature Extraction, Pretreatment, Generate graph, and node Classification), is proposed in this paper for the fusion and disease prediction of multimodal medical data. The EPGC consists of two parts: multimodal data fusion and prediction, as shown in Figure 3. Firstly, fusing the multimodal data, the model preprocesses the diverse modalities and complex structures of data, and is mainly responsible for feature selection and quantification of non-numerical data. The preprocessing method effectively addresses the challenges related to unstructured data processing and inconsistent format standardization, and the purpose is to convert the multimodal data or its features into numerical data for the next step of constructing the GDS. The constructed GDS can better preserve the feature information of multimodal medical data. The feature extraction is based on AI algorithms to extract the effective features. For example, whether there are abnormal bands in an ECG, whether there are regions of interest in the image data, etc. Meanwhile, normalization is also essential. Then, the preprocessed data is used to construct a GDS using the patients as the nodes and the similarity between the patients as the edges, thereby fusing the multimodal data into the GDS, that is, merging some scattered multimodal data into one gathered GDS. Finally, a graph neural network learning task based on node classification is applied to the fused graph data to achieve the diagnosis of patient diseases.
EPGC algorithm described in Algorithm 1. For multimodal medical data N ( X i ) = X 1 , X 2 , X m (m samples), as the graph can only process numerical data, the EPGC first needs to convert the non-numerical multimodal medical data N ( X i ) into numerical data, which requires AI-based feature extraction algorithms to obtain the numerical data N i . Then, N i is normalized to avoid inaccurate classification caused by inconsistent dimensions. Next, calculate the Pearson coefficient between each sample in N i , and determine whether there are edges between the samples based on the value of Pearson coefficient r i j to construct the GDS. Next, select any target node in the GDS for sampling; collect information on the nodes and edges in the GDS; generate a feature vector h v k , aggregate and embed the feature vectors; and finally use the downstream classifiers to predict the node types, i.e., patient disease prediction.
Algorithm 1: EPGC algorithm
Input: Multimodal data N ( X i ) = { X 1 , X 2 , X m } ; the depth of graph; weight matrix W k , k { 1,2 , , K } ; non-linear activation function σ ; aggregator.
Adjacency function: N : υ 2 υ .
Output: z v , v V , the result of the following prediction:
1: Extract features from non-numerical data in N ( X i ) to obtain numerical data N i .
2: Normalize N i , N i N ~ i .
3: Calculate the Pearson correlation coefficient r i j between each sample in N ~ i according to Equation (2).
4: Construct the graph G = ( V , E ) .
5: Sample the node information of G = ( V , E ) and generate feature vector h v k .
6: Aggregate feature information, h N ~ ( v ) k A G G R E G A T E k ( { h v k 1 , v N ~ ( v ) } ) , and generate new feature vectors.
7: Normalize feature vectors, h v k h v k h v k 2 , v V .
8: Generate new node embedding, z v h v k , v V .
9: Classify feature vectors and output the prediction result of diseases.
The advantages and innovations of the EPGC lie in introducing a GDS into the fusion of multimodal medical data, leveraging the feature expression advantages of GDS’s nodes and edges to solve the problem of multi-source heterogeneous and high-dimensional multimodal medical data fusion. It establishes a classification model based on graph learning using the fused multimodal data to achieve the classification and diagnosis of patient disease types. Compared to existing multimodal medical data processing models, the EPGC model preserves the association rule information of features in the data samples as far as possible and does not require redundant work in dimensionality reduction.

3. Experimental Section

3.1. Experimental Data

In this paper, we evaluate the model’s performance using a cardiovascular diseases dataset. Cardiovascular diseases are the leading cause of death globally, taking an estimated 17.9 million lives each year. Cardiovascular diseases include disorders of the heart and blood vessels and include coronary heart disease, cerebrovascular disease, rheumatic heart disease, and other conditions [40]. This paper focuses on the prediction of two types of diseases from the data, namely coronary artery disease and myocardial infarction.
Coronary artery disease, also known as coronary heart disease or ischemic heart disease, is the most prevalent type of cardiovascular disease. In some developed countries, its incidence among individuals over 65 can reach up to 20%, with even higher rates in developing nations. Coronary artery disease is a disease with a complex etiology; it is typically diagnosed using invasive coronary angiography, which offers a high accuracy exceeding 90%. However, this method carries certain risks, as it involves injecting a contrast agent into the body. Such procedures can lead to complications, including bleeding, infection, hematoma, and, in rare cases, death [41]. In contrast, traditional diagnostic methods, a combined medical examination with an electrocardiogram (ECG), electroencephalogram (EEG), and laboratory tests, are safer and more widely used [42]. These methods provide a comprehensive assessment of a patient’s condition and can be effectively used even in regions with less developed medical infrastructure, offering patients a safer and more accessible diagnostic experience.
Myocardial infarction is a prevalent and highly detrimental cardiac condition, typically resulting from myocardial ischemia due to coronary artery occlusion [43]. Myocardial infarction can lead to severe complications, such as atrial fibrillation, arrhythmia [44], cardiogenic shock, and congestive heart failure, with the potential for fatal outcomes in severe cases [45]. Even experienced clinicians may struggle with accurately predicting these complications. Therefore, it is crucial to reliably predict potential complications to implement timely and appropriate interventions. In clinical practice, multimodal data are collected from patients, including demographic information, blood test results, and diagnosis and treatment details during hospitalization.
The two experimental datasets are sourced from UCI. The first dataset is an extension of the Z-Alizadeh Sani dataset, specifically designed for coronary artery disease detection [46]. It comprises the examination records from 303 patients, with each patient’s data including 54 features. This dataset encompasses five types of data—demographic information, symptoms and related disease screening, electrocardiogram data, blood test results, and cardiac echo characteristics—along with the disease types. It includes numerical data, sequential data, and textual data, with complete data and balanced positive and negative samples. A detailed description of the data is provided in Table 1. The dataset has undergone comprehensive feature extraction, including the electrocardiogram, vector electrocardiogram, and echocardiographic data. It is of high quality, with no missing values and clearly labeled data.
The second dataset selected for this paper study is the myocardial infarction complications dataset [47], which comprises examination records of 1700 myocardial infarction patients. Each patient’s examination record has 124 feature attributes, of which 1 attribute is the patient’s ID, 111 feature attributes are the clinical descriptions of myocardial infarction, and 12 feature attributes are the complications of myocardial infarction. The data are presented in Table 2. It includes numerical data, time-series data, and textual data. The dataset has undergone the feature extraction steps, mainly for the electrocardiogram data. The data quality of this dataset is poor, with numerous missing data points and inconsistent data labels. We selected 202 complete data samples from it to form the experimental data, with balanced positive and negative samples.

3.2. Experimental Design

During the construction of a diagnostic model for multimodal medical data, we designed the following experimental steps: data fusion and disease prediction, model evaluation, and control group experimental design.

3.2.1. Data Fusion and Disease Prediction

  • Data preprocessing: This mainly involves standardizing the format and dimensions of experimental data and converting the multimodal data into numerical data. We use AI-based feature extraction to convert these multimodal data into numerical data. This needs to be implemented based on AI algorithms, as shown in Figure 3. It mainly focuses on image data (CT, MRI, etc.), textual data (symptom description, etc.), and sequential data (electrocardiogram, electroencephalogram, etc.). The ECG data describe the features of a patient’s ECG-related bands and was preprocessed by the dataset provider. Additionally, some individual features in the original data include descriptive text data, such as the severity of heart valve disease, which is categorized as N, Mild, Moderate, and Severe to describe the patient’s condition. For consistency and to perform a numerical analysis, we convert these descriptive categories into numerical values: 0 for N, 1 for Mild, 2 for Moderate, and 3 for Severe. In addition, the multimodal medical data are characterized by dimensional inconsistency. To address this, we apply mean normalization to the original data. This step helps to mitigate the impact of varying dimensions on model learning and ensures that the classifier accuracy is not adversely affected by these differences.
  • GDS construction: Each patient is represented as a node, and the similarity between patients is represented as an edge. The edges are determined using the Pearson correlation coefficient. According to the Pearson correlation coefficient, an edge is drawn between two nodes if their correlation coefficient is greater than or equal to 0.5, i.e., r 0.5 , indicating a strong similarity and thus a connection between them. Conversely, if the coefficient is 0.5 or less, i.e., r   < 0.5 , this indicates a weak similarity and no connection edge is drawn between the nodes.
  • Disease prediction—designing the network learning models: The constructed GDS is undirected. We propose using a GraphSAGE network, which samples the node information with random walks on the graph and generates the node features. Given that the types and sizes of experimental data are relatively simple, we propose using a two-layer GNN network, utilizing a GCN aggregator and mean aggregator, respectively. This design approach enhances the extraction of features from the graph structure. The expression of the GCN aggregator is shown in Equation (4), where h k represents the feature vector of node, σ represents the non-linear activation function, W k 1 represents the adjacency weight matrix of node, and D ~ 1 2 A ~ D ~ 1 2 is the normalized Laplacian matrix. The mean aggregator is shown in Equation (5), where h i ( k ) represents the feature vector of node N ( v ) , and g represents the i , which is the adjacent point of the node.
h k = σ ( D ˜ 1 2 A ˜ D ˜ 1 2 h k 1 W k 1 )
h v k σ ( W m e a n ( { h v k 1 } { h u k 1 , u N ( v ) } )
In addition, we also add a control group for the experiment, which comes from the experimental results in the references. The experimental results are currently the best-known classification model for this dataset. The result is a disease classification model using feature selection and machine learning algorithms on the same dataset.

3.2.2. Control Group Experiment

In order to highlight the generalization ability of the model, two datasets were used for model validation, i.e., the coronary artery disease detection dataset and the myocardial infarction complication dataset.
The dataset on coronary artery disease detection is openly available from researchers’ papers. Some researchers have conducted extensive studies on disease prediction based on this dataset. We selected the results with the best classification performance as the control group for the EPGC model experiment. Meanwhile, in order to demonstrate the comparability of our experimental results, we referred to the reported experimental results for the existing best-in-class methods for comparison with ours, and used the same evaluation indicators. This model uses the traditional “feature selection + dimensionality reduction + prediction” approach, referred to as the “EMC” (Feature Extraction, Dimensionality Reduction, Classification) method. Specifically, reference [16] proposed a feature engineering method to select, rank, and combine the features in the dataset to predict coronary artery disease using the optimal feature combination. This model is referred to as EMC1. Reference [48] conducted a study using the BESO (Bald Eagle Search Optimization) for dimensionality reduction. Ten optimal features were selected from the original data, then combined with a random forest to predict coronary artery disease. This model is referred to as EMC2. Reference [49] proposed a combination of genetic algorithms and neural networks for feature selection in their research, obtaining an optimal prediction combination containing 24 features through various parameter settings. This model is referred to as EMC3.
The myocardial infarction complications dataset is openly available from researchers’ papers too. We referred to the dataset on myocardial infarction complications selected the reported experimental results for the existing best-in-class methods for comparison with ours, and used the same evaluation indicators. The control group included models developed by three different research groups. All the groups utilized deep neural network frameworks and dimensionality reduction techniques. Specifically, reference [18] introduced a five-layer deep neural network model, which included three hidden layers for feature extraction and complexity reduction. This model is termed the DL1 (Deep Learning 1) method. Reference [17] initially performed dimensionality reduction on the high-dimensional data and then designed a four-layer deep neural network model with two hidden layers for feature learning and extraction. This model is referred to as the DL2 (Deep Learning 2) method. Reference [50] presented a multitask deep learning model to predict myocardial infarction complications. This model is referred to as DL3 (Deep Learning 3).

3.2.3. Model Evaluation

In terms of model evaluation, this study was based on a control group experiment. In the model validation experiment using the coronary artery disease data, the same evaluation indicators as in references [16,48,49] were used, and the results were evaluated using two indicators: accuracy and recall. The equations for accuracy and recall are shown in Equations (6) and (7), respectively.
In the model validation experiment using the myocardial infarction complications dataset, we used the same accuracy evaluation index as in references [17,18,50] to ensure consistency with the same dataset. The accuracy served as the index for assessing model performance.
A c c u r a c y = T P + T N T P + T N + F N + F P
R e c a l l = T P T P + F N
where TP (True Positive) represents the number of originally positive samples correctly identified as positive, TN (True Negative) represents the number of originally negative samples correctly identified as negative, FN (False Negative) represents the number of originally positive samples incorrectly classified as negative, and FP (False Positive) represents the number of negative samples correctly identified as positive.

4. Results

The experimental design used a two-layer GNN network structure, utilizing a GCN aggregator and a mean aggregator for aggregating the sampling information. The model’s loss function was the entropy loss function. In the experiment, 303 samples from the coronary artery disease dataset were divided into training, validation, and testing sets at an 8:1:1 ratio. The 10-fold cross-validation method was implemented in model training, which means that at each compromise, the data was divided into a ratio of 80% (training), 10% (validation), and 10% (testing), and the experiment was repeated for all folds to ensure the minimization of the risk of insufficient single division and resulting randomness. The training set and validation set were used to train the model, and optimize and adjust the parameters, while the test set was used to test the performance of the model. We conducted numerous repeated experiments, and the model had the best convergence performance when the learning rate for the model was set to 0.1. Given the advantages of GraphSage in small sample learning [51], 202 available samples from the myocardial infarction complications dataset samples underwent the same operation.
The experimental results for the coronary artery disease dataset are displayed in Figure 4. As illustrated in Figure 4, the EPGC model proposed in this study achieved <Accuracy, Recall> = <96.70%, 100%>. In contrast, the EMC1 model achieved performance indexes for <Accuracy, Recall> = <96.40%, 100%>. The EMC2 model achieved performance indexes for <Accuracy, Recall> = <92.20%, 92.20%>. The EMC3 model achieved performance indexes for <Accuracy, Recall> = <94.71%, 96.29%>. The results demonstrate that the EPGC model provides superior prediction performance for the multimodal coronary artery disease data, as reflected in the accuracy and recall.
The experimental results for the myocardial infarction complications dataset are displayed in Figure 5. The prediction accuracies of DL1, DL2, and DL3 are 92.09%, 90.99%, and 91.98%, respectively. The prediction accuracy of EPGC model proposed in this study is 94.22%. The results demonstrate that the EPGC model outperforms the DL1, DL2, and DL3 methods, which are based on complex deep neural network frameworks, on the myocardial infarction complications dataset. The prediction accuracy is improved by 2.13%, 3.23%, and 2.24%, respectively.

5. Discussion

In this study, we propose a data fusion approach to design a GDS construction method based on sample similarity. The model avoids conventional dimensionality reduction processes, such as complex feature selection or extraction during the fusion of multimodal medical data, and helps to preserve the integrity of the data as much as possible. Subsequently, a GraphSAGE network is employed to sample and aggregate the information from the nodes and edges within the GDS, ultimately allowing for the prediction of a patient’s disease type. The proposed model, the EPGC, demonstrates strong applicability to high-dimensional data, and in the disease prediction experiments on two different multimodal medical datasets, the model outperformed the existing control group models.
The fusion and intelligent decision-making of multimodal medical data face challenges, such as diverse data formats, high dimensionality, and data complexity. The data format of multimodal medical data is both multi-source and complex. To address this, data format transformation and feature extraction are essential. In this paper, we have drawn on the experience of other researchers, converting multimodal data into numerical data in preprocessing, to meet the requirements of deep algorithms that require a large number of iterations and computational power to converge a model. Therefore, transforming data formats and extracting features are crucial for effective learning and model training.
Meanwhile, multimodal medical data often has high dimensionality, which can reduce the accuracy of disease prediction when using traditional machine learning or deep neural network classifiers due to the curse of dimensionality. Most existing research on this topic has attempted disease prediction by constructing complex networks or using a “feature selection + dimensionality reduction + prediction” approach. These methods require lots of dimensionality reduction work, which can be result-oriented, blind, and random, increasing a model’s complexity and leading to information loss. For example, researchers have often tried various dimensionality reduction methods and then evaluated them based on the final predicted results, using a reverse approach to identify the best method. This undoubtedly impacts the prediction performances of the models. Compared with the existing research, our method does not require dimensionality reduction, which preserves as much information as possible in the original data. It reduces the loss of useful information and improves the performance of subsequent model training. This is demonstrated by the experimental results on the two datasets. At the same time, leveraging the powerful feature expression capability of a GDS, dispersed multimodal medical data is fused into one GDS, which not only solves the problem of high data dimensionality and complexity but also addresses the fusion issue of multimodal medical data. In addition, GraphSage has been proven to have an excellent learning ability on massive transactional network data [52] and large-scale multi-entity heterogeneous graph data [53], which also provides support for the EPGC to deal with the complexity of multimodal medical data.

6. Conclusions

Intelligent diagnosis using multimodal medical data represents a significant trend in the development of medical big data. In this paper, we propose a fusion and disease prediction model for multimodal medical data; the construction of a GDS is innovatively introduced into the fusion of multimodal medical data. The multimodal medical data initially undergoes preprocessing steps, including format transformation and feature extraction. Meanwhile, normalization is applied to address the dimensional inconsistencies and improve the classifier accuracy. Subsequently, each patient sample is represented as a node, and the similarities between the samples are represented as edges, constructing a GDS. Following the construction of the GDS, node sampling and aggregation are employed to generate the feature vectors associated with the nodes. Node embedding is subsequently applied to these feature vectors. Finally, the machine learning classifier utilizes the embedded features to predict the node labels, thus enabling disease prediction. In the experiment, we compared the proposed EPGC model with the current mainstream and best-performing methods on the same datasets. The results reveal that the EPGC shows strong applicability to high-dimensional data and outperforms the reference models at disease prediction on various multimodal medical datasets.
As shown in Figure 3, there are some AI-based feature extraction modules in the EPGC model that handle multimodal data, such as image data, textual data, sequential data, etc. These modules can address the challenges of common medical data modalities. However, during this process, the AI-based feature extraction may be biased due to the performance of AI algorithms themselves, which can affect the subsequent prediction performance. In addition, constructing the GDS based on the Pearson coefficient (hyperparameters) ( r 0.5 ), may also lead to bias in the results. The above are all limitations of the EPGC. Correspondingly, with an improvement in the AI algorithm’s performance and the optimization of selection of hyperparameters, the performance of the EPGC model will also be further improved. In future research, the EPGC will be applied to other areas by adding feature extraction modules, such as video stream data and sensor data.

Author Contributions

Conceptualization, L.Z., and H.C.; methodology, L.Z., H.C., and R.E.W.; software, L.Z., Q.Z., and J.L.; validation, T.L., J.L., and Z.J. (Zijie Jiang); formal analysis, T.L., Q.Z., Z.J. (Zijie Jiang) and J.L.; investigation, L.Z., and T.L.; data curation, L.Z., and Q.Z.; writing—original draft preparation, L.Z. and R.E.W.; writing—review and editing, L.Z., Q.Z., Z.J. (Zijie Jiang), and R.E.W.; visualization, L.Z.; supervision, H.C. and Z.J. (Zhongwei Jia); project administration, Z.J. (Zhongwei Jia); funding acquisition, Z.J. (Zhongwei Jia). All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by [Research on Vaccination Strategy and Policy of COVID-19 Pneumonia Based on Big Data Modeling] grant number [72174004], [National Key Research and Development Program of China] grant number [2023YFC2308703], [High-level Public Health Talent Development Program of Beijing] grant number [Discipline Backbone-02-12], [Research on Early Warning Signal Recognition and Evaluation of Acute Respiratory Infectious Diseases in Hospitals Based on Medical Prevention Integration] grant number [72274210].

Acknowledgments

Thanks to Zhaoxia Zheng, Yunze Zhang, and Senyao Zhang for assisting in the paper revision.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Jiang, S.; Wang, T.; Zhang, K. Data-driven decision-making for precision diagnosis of digestive diseases. Biomed. Eng. Online 2023, 22, 87. [Google Scholar] [CrossRef] [PubMed]
  2. Khoa, L.D.V.; Yen, L.M.; Nhat, P.T.H.; Duong, H.T.H.; Thuy, D.B.; Zhu, T.; Greeff, H.; Clifton, D.; Thwaites, C.L. Vital sign monitoring using wearable devices in a Vietnamese intensive care unit. BMJ Innov. 2021, 7, s1–s5. [Google Scholar] [CrossRef]
  3. Diagnosis—Coronary Heart Disease. Available online: https://www.nhs.uk/conditions/coronary-heart-disease/diagnosis/ (accessed on 17 January 2024).
  4. Cai, Q.; Wang, H.; Li, Z.; Liu, X. A Survey on Multimodal Data-Driven Smart Healthcare Systems: Approaches and Applications. IEEE Access 2019, 7, 133583–133599. [Google Scholar] [CrossRef]
  5. Beam, A.L.; Kompa, B.; Schmaltz, A.; Fried, I.; Weber, G.; Palmer, N.; Shi, X.; Cai, T.; Kohane, I.S. Clinical Concept Embeddings Learned from Massive Sources of Multimodal Medical Data. Pac. Symp. Biocomput. 2020, 25, 295–306. [Google Scholar]
  6. Lipkova, J.; Chen, R.J.; Chen, B.; Lu, M.Y.; Barbieri, M.; Shao, D.; Vaidya, A.J.; Chen, C.; Zhuang, L.; Williamson, D.F.K.; et al. Artificial intelligence for multimodal data integration in oncology. Cancer Cell 2022, 40, 1095–1110. [Google Scholar] [CrossRef]
  7. Wei, T.; Tiwari, P.; Pandey, H.M.; Moreira, C.; Jaiswal, A.K. Multimodal medical image fusion algorithm in the era of big data. In Neural Computing & Applications; Springer: Berlin/Heidelberg, Germany, 2020. [Google Scholar] [CrossRef]
  8. Acosta, J.N.; Falcone, G.J.; Rajpurkar, P.; Topol, F.J. Multimodal biomedical AI. Nat. Med. 2022, 28, 773–1784. [Google Scholar] [CrossRef]
  9. Kline, A.; Wang, H.; Li, Y.; Dennis, S.; Hutch, M.; Xu, Z.; Wang, F.; Cheng, F.; Luo, Y. Multimodal machine learning in precision health: A scoping review. Npj Digit. Med. 2022, 5, 171. [Google Scholar] [CrossRef] [PubMed]
  10. Stahlschmidt, S.R.; Ulfenborg, B.; Synnergren, J. Multimodal deep learning for biomedical data fusion: A review. Brief. Bioinform. 2022, 23, bbab569. [Google Scholar] [CrossRef]
  11. Azam, M.A.; Khan, K.B.; Salahuddin, S.; Rehman, E.; Khan, S.A.; Khan, M.A.; Kadry, S.; Gandomi, A.H. A review on multimodal medical image fusion: Compendious analysis of medical modalities, multimodal databases, fusion techniques and quality metrics. Comput. Biol. Med. 2022, 144, 105253. [Google Scholar] [CrossRef] [PubMed]
  12. Guo, Z.; Li, X.; Huang, H.; Guo, N.; Li, Q. Deep Learning-Based Image Segmentation on Multimodal Medical Imaging. IEEE Trans. Radiat. Plasma Med. Sci. 2019, 3, 162–169. [Google Scholar] [CrossRef]
  13. Duan, J.; Xiong, J.; Li, Y.; Ding, W. Deep learning based multimodal biomedical data fusion: An overview and comparative review. Inf. Fusion 2024, 112, 102536. [Google Scholar] [CrossRef]
  14. Chen, H.; Gao, M.; Zhang, Y. Attention-Based Multi-NMF Deep Neural Network with Multimodality Data for Breast Cancer Prognosis Model. Biomed. Res. Int. 2019, 2019, 9523719. [Google Scholar] [CrossRef] [PubMed]
  15. Yang, J.; Ju, J.; Guo, L.; Ji, B.; Shi, S.; Yang, Z.; Gao, S.; Yuan, X.; Tian, G.; Liang, Y.; et al. Prediction of HER2-positive breast cancer recurrence and metastasis risk from histopathological images and clinical information via multimodal deep learning. Comput. Struct. Biotechnol. J. 2021, 20, 333–342. [Google Scholar] [CrossRef]
  16. Alizadehsani, R.; Hosseini, M.J.; Khosravi, A.; Khozeimeh, F.; Roshanzamir, M.; Sarrafzadegan, M. Non-invasive detection of coronary artery disease in high-risk patients based on the stenosis prediction of separate coronary arteries. Comput. Methods Programs Biomed. 2018, 162, 119–127. [Google Scholar] [CrossRef]
  17. Golovenkin, S.E.; Dorrer, M.G.; Nikulina, S.Y.; Orlova, Y.V.; Pelipeckaya, E.V. Evaluation of the effectiveness of using artificial intelligence to predict the response of the human body to cardiovascular diseases. J. Phys. Conf. Ser. 2020, 1679, 042017. [Google Scholar] [CrossRef]
  18. Yavru, I.B.; Gunduz, S.Y. Predicting Myocardial Infarction Complications and Outcomes with Deep Learning. Eskişehir Tech. Univ. J. Sci. Technol. A-Appl. Sci. Eng. 2022, 23, 184–194. [Google Scholar] [CrossRef]
  19. Hu, S.L. Research on Medical Multi-Source Data Fusion Based on Big Data. Recent Adv. Comput. Sci. Commun. 2022, 15, 376–387. [Google Scholar] [CrossRef]
  20. Bronstein, M.M.; Bruna, J.; LeCun, Y.; Szlam, A.; Vandergheynst, P. Geometric Deep Learning: Going beyond Euclidean data. IEEE Signal Process. Mag. 2017, 34, 18–42. [Google Scholar] [CrossRef]
  21. Wang, L.; Zhang, Y.; Feng, J. On the Euclidean distance of images. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1334–1339. [Google Scholar] [CrossRef]
  22. He, H.; Wu, D. Transfer Learning for Brain—Computer Interfaces: A Euclidean Space Data Alignment Approach. IEEE Trans. Biomed. Eng. 2020, 67, 399–410. [Google Scholar] [CrossRef]
  23. Tahirovic, A.A.; Angeli, D.; Tahirovic, A.; Strbac, G. Petri graph neural networks advance learning higher order multimodal complex interactions in graph structured data. Sci. Rep. 2025, 15, 17540. [Google Scholar] [CrossRef] [PubMed]
  24. Walke, D.; Micheel, D.; Schallert, K.; Muth, T.; Broneske, D.; Saake, G.; Heyer, R. The importance of graph databases and graph learning for clinical applications. Database 2023, 2024, baad045. [Google Scholar] [CrossRef]
  25. Shao, S.; Ribeiro, P.H.; Ramirez, C.M.; Moore, J.H. A review of feature selection strategies utilizing graph data structures and Knowledge Graphs. Brief. Bioinform. 2024, 25, bbae52. [Google Scholar] [CrossRef]
  26. Jarvis, R.A.; Patrick, E.A. Clustering Using a Similarity Measure Based on Shared Near Neighbors. IEEE Trans. Comput. 1973, C-22, 1025–1034. [Google Scholar] [CrossRef]
  27. Lin, Y.S.; Jiang, J.Y.; Lee, S.J. A Similarity Measure for Text Classification and Clustering. IEEE Trans. Knowl. Data Eng. 2013, 26, 1575–1590. [Google Scholar] [CrossRef]
  28. Liu, X. Entropy, distance measure and similarity measure of fuzzy sets and their relations. Fuzzy Sets Syst. 1992, 52, 305–318. [Google Scholar] [CrossRef]
  29. Gori, M.; Monfardini, G.; Scarselli, F. A New Model for Learning in Graph Domains. In Proceedings of the IEEE International Joint Conference on Neural Network (IJCNN), Montreal, QC, Canada, 31 July–4 August 2005; pp. 729–734. [Google Scholar]
  30. Zhou, J.; Cui, G.; Hu, S.; Zhang, Z.; Yang, C.; Liu, Z.; Wang, L.; Li, C.; Sun, M. Graph neural networks: A review of methods and applications. AI Open 2020, 1, 57–81. [Google Scholar] [CrossRef]
  31. Qiu, J.; Zhang, X.; Wang, T.; Hou, H.; Wang, S.; Yang, T. A GNN-Based False Data Detection Scheme for Smart Grids. Algorithms 2025, 18, 166. [Google Scholar] [CrossRef]
  32. Mateo, S.; Danijel, K.; Mršić, L. Graph Convolutional Networks for Predicting Cancer Outcomes and Stage: A Focus on cGAS-STING Pathway Activation. Mach. Learn. Knowl. Extr. 2024, 6, 2033–2048. [Google Scholar]
  33. Shao, J.; Zhang, H.; Mao, Y.; Zhang, J. Branchy-GNN: A Device-Edge Co-Inference Framework for Efficient Point Cloud Processing. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021; pp. 8488–8492. [Google Scholar]
  34. Alawad, D.M.; Katebi, A.; Hoque, M.T. Enhanced Graph Representation Convolution: Effective Inferring Gene Regulatory Network Using Graph Convolution Network with Self-Attention Graph Pooling Layer. Mach. Learn. Knowl. Extr. 2024, 6, 1818–1839. [Google Scholar] [CrossRef]
  35. Mohanraj, P.; Raman, V.; Ramanathan, S. Deep Learning for Parkinson’s Disease Diagnosis: A Graph Neural Network (GNN) Based Classification Approach with Graph Wavelet Transform (GWT) Using Protein—Peptide Datasets. Diagnostics 2024, 14, 2181. [Google Scholar] [CrossRef]
  36. Hossain, M.E.; Uddin, S.; Khan, A. Network analytics and machine learning for predictive risk modelling of cardiovascular disease in patients with type 2 diabetes. Expert Syst. Appl. 2020, 164, 113918. [Google Scholar] [CrossRef]
  37. Xuan, P.; Wu, X.; Cui, H.; Jin, Q.; Wang, L.; Zhang, T.; Nakaguchi, T.; Duh, H.B.L. Multi-scale random walk driven adaptive graph neural network with dual-head neighboring node attention for CT segmentation. Appl. Soft Comput. 2023, 133, 109905. [Google Scholar] [CrossRef]
  38. Lee, Y.W.; Huang, S.K.; Chang, R.F. CheXGAT: A disease correlation-aware network for thorax disease diagnosis from chest X-ray images. Artif. Intell. Med. 2022, 132, 102382. [Google Scholar] [CrossRef] [PubMed]
  39. Zhang, T.; Shan, H.R.; Little, M.A. Causal GraphSAGE: A robust graph method for classification based on causal sampling. Pattern Recognit. 2022, 128, 108696. [Google Scholar] [CrossRef]
  40. Cardiovascular Diseases. Available online: https://www.who.int/health-topics/cardiovascular-diseases#tab=tab_1 (accessed on 17 January 2024).
  41. The DISCHARGE Trial Group. CT or Invasive Coronary Angiography in Stable Chest Pain. N. Engl. J. Med. 2022, 386, 1591–1602. [Google Scholar] [CrossRef]
  42. Mann, D.L.; Zipes, D.P.; Libby, P.; Bonow, R.O. Braunwald’s Heart Disease: A Textbook of Cardiovascular Medicine, 10th ed.; Elsevier Health Sciences: Amsterdam, The Netherlands, 2018; pp. 636–661. [Google Scholar]
  43. Reed, G.W.; Rossi, J.E.; Cannon, C.P. Acute myocardial infarction. Lancet 2017, 389, 197–210. [Google Scholar] [CrossRef]
  44. Moras, E.; Yakkali, S.; Gandhi, K.D.; Virk, H.U.H.; Alam, M.; Zaid, S.; Barman, N.; Jneid, H.; Vallabhajosyula, S.; Sharma, S.K.; et al. Complications in Acute Myocardial Infarction: Navigating Challenges in Diagnosis and Management. Hearts 2024, 5, 122–141. [Google Scholar] [CrossRef]
  45. Complications of Myocardial Infarction. Available online: https://emedicine.medscape.com/article/164924-overview (accessed on 21 July 2022).
  46. Alizadehsani, R.; Roshanzamir, M.; Sani, Z. Extention of Z-Alizadeh Sani Dataset Data Set; UCI Irvine-Machine Leaming-Repository: Irvine, CA, USA, 2013. [Google Scholar] [CrossRef]
  47. Golovenkin, S.E.; Shulman, V.A.; Rossiev, D.A.; Shesternya, P.A.; Nikulina, S.Y.; Orlova, Y.V. Myocardial Infarction Complications Data Set; UC Irvine-Machine Leaming-Repository: Irvine, CA, USA, 2020. [Google Scholar] [CrossRef]
  48. Olawade, D.B.; Soladoye, A.A.; Omodunbi, B.A.; Aderinto, N.; Adeyanju, I.A. Comparative analysis of machine learning models for coronary artery disease prediction with optimized feature selection. Int. J. Cardiol. 2025, 436, 133443. [Google Scholar] [CrossRef] [PubMed]
  49. Hashemi, M.; Komamardakhi, S.S.S.; Maftoun, M.; Zare, O.; Joloudari, J.H.; Nematollahi, M.A.; Alizadehsani, R.; Sala, P.; Gorriz, J.M. Enhancing Coronary Artery Disease Classification Using Optimized MLP Based on Genetic Algorithm. In Proceedings of the Artificial Intelligence for Neuroscience and Emotional Systems (IWINAC 2024), Olhao, Algarve, Portugal, 4–7 June 2024; p. 14674. [Google Scholar]
  50. Makhmudov, F.; Ravshanov, N.; Akhmedov, D.; Pekos, O.; Turimov, D.; Cho, Y.I. A Multitask Deep Learning Model for Predicting Myocardial Infarction Complications. Bioengineering 2025, 12, 520. [Google Scholar] [CrossRef]
  51. Bajaj, S.; Son, H.; Liu, J.; Guan, H.; Serafini, M. Graph Neural Network Training Systems: A Performance Comparison of Full-Graph and Mini-Batch. arXiv 2024, arXiv:2406.00552v3. [Google Scholar] [CrossRef]
  52. Tare, M.; Rattasits, C.; Wu, Y.; Wielewski, E. Harnessing GraphSAGE for Learning Representations of Massive Transactional Networks. In Proceedings of the 14th IAPR-TC-15 International Workshop, Graph-Based Representations in Pattern Recognition (GbRPR 2025), Caen, France, 25–27 June 2025; pp. 179–188. [Google Scholar]
  53. Badrinath, A.; Yang, A.; Rajesh, K.; Agarwal, P.; Yang, J.; Chen, H.; Xu, J.; Rosenberg, C. OmniSage: Large Scale, Multi-Entity Heterogeneous Graph Representation Learning. In Proceedings of the 31st ACM SIGKDD Conference on Knowledge Discovery and Data Mining (SIGKDD), Toronto, ON, Canada, 3–7 August 2025; pp. 4261–4272. [Google Scholar]
Figure 1. GDS construction based on similarity measurement. Each node represents a patient, and each edge represents the similarity between them. Each patient exists as a node (entity) in the GDS, and whether there are edge connections between the nodes depends on their degree of similarity.
Figure 1. GDS construction based on similarity measurement. Each node represents a patient, and each edge represents the similarity between them. Each patient exists as a node (entity) in the GDS, and whether there are edge connections between the nodes depends on their degree of similarity.
Make 07 00092 g001
Figure 2. Node prediction model based on GraphSAGE, which integrates neighborhood node sampling, sampling information aggregation, node embeddings, and prediction. The learning task of GraphSAGE involves sampling the information of the nodes and edges, aggregating and embedding based on the information, and ultimately achieving node prediction.
Figure 2. Node prediction model based on GraphSAGE, which integrates neighborhood node sampling, sampling information aggregation, node embeddings, and prediction. The learning task of GraphSAGE involves sampling the information of the nodes and edges, aggregating and embedding based on the information, and ultimately achieving node prediction.
Make 07 00092 g002
Figure 3. EPGC algorithm, which integrates multimodal data fusion and prediction. The data fusion module consists of AI-based feature extraction and a GDS construction based on the extracted features. The prediction module completes the learning and prediction tasks of GDS.
Figure 3. EPGC algorithm, which integrates multimodal data fusion and prediction. The data fusion module consists of AI-based feature extraction and a GDS construction based on the extracted features. The prediction module completes the learning and prediction tasks of GDS.
Make 07 00092 g003
Figure 4. The experimental results for the coronary artery disease dataset for the experimental and control groups. The horizontal axis represents the experimental model. The vertical axis represents the values of evaluation indicators. The blue bar chart represents the accuracy. The orange bar chart represents the recall. The results indicate that the EPGC performs better.
Figure 4. The experimental results for the coronary artery disease dataset for the experimental and control groups. The horizontal axis represents the experimental model. The vertical axis represents the values of evaluation indicators. The blue bar chart represents the accuracy. The orange bar chart represents the recall. The results indicate that the EPGC performs better.
Make 07 00092 g004
Figure 5. The experimental results for the myocardial infarction complications dataset for the experimental and control groups. The horizontal axis represents the experimental model. The vertical axis represents the values of evaluation indicator. The results indicate that the EPGC has the best performance.
Figure 5. The experimental results for the myocardial infarction complications dataset for the experimental and control groups. The horizontal axis represents the experimental model. The vertical axis represents the values of evaluation indicator. The results indicate that the EPGC has the best performance.
Make 07 00092 g005
Table 1. Introduction of coronary artery disease datasets.
Table 1. Introduction of coronary artery disease datasets.
FeatureDetails
DemographicAge, Weight, Sex, BMI (Body Mass Index), DM (Diabetes Mellitus), Current Smoker, Ex-Smoker, FH (Family History), CRF (Chronic Renal Failure), CVA (Cerebrovascular Accident), Thyroid Disease, CHF (Congestive Heart Failure), DLP (Dyslipidemia), etc.
Symptoms and examinationBP (Blood Pressure), PR (Pulse Rate), Edema, Weak Peripheral Pulse, Lung Rales, Systolic Murmur, Diastolic Murmur, Typical Chest Pain, Dyspnea, Function Class, Atypical, Nonanginal CP, Exertional CP (Exertional Chest Pain), Low Th Ang (Low Threshold angina).
ECG and vectorcardiogramRhythm, Q Wave, ST Elevation, ST Depression, T Inversion, LVH (Left Ventricular Hypertrophy), Poor R Progression (Poor R Wave Progression), LAD (Left Anterior Descending), LCX (Left Circumflex), RCA (Right Coronary Artery).
Laboratory and echoFBS (Fasting Blood Sugar), Cr (Creatine) (mg/dl), TG (Triglyceride), LDL (Low-Density Lipoprotein), HDL (High-Density Lipoprotein), BUN (Blood Urea Nitrogen), ESR (Erythrocyte Sedimentation Rate), HB (Hemoglobin), K (Potassium), Na (Sodium), WBC (White Blood Cell), Lymph (Lymphocyte), Neut (Neutrophil), PLT (Platelet), EF (Ejection Fraction), Region with RWMAa (Regional Wall Motion Abnormality), VHD (Valvular Heart Disease).
Type of coronary artery diseaseYes or No
Table 2. Introduction of myocardial infarction complications dataset.
Table 2. Introduction of myocardial infarction complications dataset.
FeatureDetails
The clinical description of myocardial infarction.Demography, electrocardiogram, laboratory data, hospitalization records, medication records, etc.
The types of myocardial infarction complications.Atrial fibrillation (AF), supraventricular tachycardia (ST), ventricular tachycardia (VT), ventricular fibrillation (VF), third-degree AV block (TA), pulmonary edema (PE), myocardial rupture (MR), Dressler syndrome (DS), chronic heart failure (CH), relapse of the myocardial infarction (RM), post-infarction angina (PA), lethal outcome (LO).
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, L.; Li, T.; Cui, H.; Zhang, Q.; Jiang, Z.; Li, J.; Welsch, R.E.; Jia, Z. A Novel Prediction Model for Multimodal Medical Data Based on Graph Neural Networks. Mach. Learn. Knowl. Extr. 2025, 7, 92. https://doi.org/10.3390/make7030092

AMA Style

Zhang L, Li T, Cui H, Zhang Q, Jiang Z, Li J, Welsch RE, Jia Z. A Novel Prediction Model for Multimodal Medical Data Based on Graph Neural Networks. Machine Learning and Knowledge Extraction. 2025; 7(3):92. https://doi.org/10.3390/make7030092

Chicago/Turabian Style

Zhang, Lifeng, Teng Li, Hongyan Cui, Quan Zhang, Zijie Jiang, Jiadong Li, Roy E. Welsch, and Zhongwei Jia. 2025. "A Novel Prediction Model for Multimodal Medical Data Based on Graph Neural Networks" Machine Learning and Knowledge Extraction 7, no. 3: 92. https://doi.org/10.3390/make7030092

APA Style

Zhang, L., Li, T., Cui, H., Zhang, Q., Jiang, Z., Li, J., Welsch, R. E., & Jia, Z. (2025). A Novel Prediction Model for Multimodal Medical Data Based on Graph Neural Networks. Machine Learning and Knowledge Extraction, 7(3), 92. https://doi.org/10.3390/make7030092

Article Metrics

Back to TopTop