You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

15 April 2022

A Principal Neighborhood Aggregation-Based Graph Convolutional Network for Pneumonia Detection

,
,
and
School of Computer Science and Engineering, Central South University, Changsha 410083, China
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Advances in Artificial Intelligence for Biomedical Signal and Image Analysis

Abstract

Pneumonia is one of the main causes of child mortality in the world and has been reported by the World Health Organization (WHO) to be the cause of one-third of child deaths in India. Designing an automated classification system to detect pneumonia has become a worthwhile research topic. Numerous deep learning models have attempted to detect pneumonia by applying convolutional neural networks (CNNs) to X-ray radiographs, as they are essentially images and have achieved great performances. However, they failed to capture higher-order feature information of all objects based on the X-ray images because the topology of the X-ray images’ dimensions does not always come with some spatially regular locality properties, which makes defining a spatial kernel filter in X-ray images non-trivial. This paper proposes a principal neighborhood aggregation-based graph convolutional network (PNA-GCN) for pneumonia detection. In PNA-GCN, we propose a new graph-based feature construction utilizing the transfer learning technique to extract features and then construct the graph from images. Then, we propose a graph convolutional network with principal neighborhood aggregation. We integrate multiple aggregation functions in a single layer with degree-scalers to capture more effective information in a single layer to exploit the underlying properties of the graph structure. The experimental results show that PNA-GCN can perform best in the pneumonia detection task on a real-world dataset against the state-of-the-art baseline methods.

1. Introduction

Pneumonia is an infection of the lower respiratory tract. It is caused by several pathogens, mainly viruses or bacteria. Pneumonia is more common in underdeveloped and developing countries, where overcrowding, pollution, and unsanitary environmental states make the situation more menacing, and medical resources are limited. Moreover, it is one of the leading causes of child mortality worldwide. It has been reported by the World Health Organization (WHO) to be the cause of one-third of child deaths in India [1]. Pneumonia is characterized by the presence of an abnormal area compared to the surrounding tissues in an X-ray image (see Figure 1 and Figure 2 for examples). Early diagnosis and treatment are critical in preventing further fatalities due to pneumonia. Analysis of the lungs by computed tomography (CT), magnetic resonance imaging (MRI), or X-rays is used for the diagnosis of pneumonia. Chest X-rays are commonly used to detect malignancy in the lungs, mostly because X-ray analysis constitutes is a relatively inexpensive and non-invasive lung exam [2]. However, as stated in [3], slight dissimilarities in terms of shape, scale, textures, and intensity, can complicate X-ray-based pneumonia detection, especially for patients below five years old. Other illnesses, such as congestive heart failure and lung scarring, could also be misclassified as pneumonia [1]. Thus, pneumonia detection requires an expert to use additional patient information to detect pneumonia symptoms from chest X-ray radiography, which is time-consuming for the radiologist, costly to hospitals, and not necessarily affordable for the users in need. Thus, designing an automated classification system to aid in detecting pneumonia has become a valuable research topic.
Figure 1. Examples of chest X-ray images of healthy people.
Figure 2. Examples of chest X-ray images of people with pneumonia presenting a bandlike opaqueness with a peripheral, mid-to-lower lung zone distribution—white infiltrates, a hallmark of pneumonia.
In the last few decades, researchers and medical practitioners have investigated the possibility of using deep learning in automated disease diagnosis systems. Deep learning techniques utilize the activation states of neurons to gradually assemble low-level features and automatically acquire higher-order abstract representations, avoiding complex feature engineering. Many deep learning models attempt to detect pneumonia by applying convolutional neural networks (CNNs) to X-ray radiographs, as they are essentially images [4]. The convolutional network approach involves hidden convolution and pooling layers to determinate spatially localized attributes through a set of receptive fields in kernel form. The convolution operator, also known as the feature detector of a CNN, is a filter in which the kernel filters input data, and the pooling operator is used to condense the resolution of each attribute map in the spatial dimensions, keeping the sizes of the attribute maps unchanged. Thus, CNNs are a class of neural networks with architecture that can achieve mappings between spatially distributed grid data of random lengths, making it flawless for classifying images. The standard CNN model pipeline consists of an input layer, a set of convolution layers, optional pooling layers with a fully connected neural network, and a final layer. The convolutional neural networks applied for pneumonia detection include VGGNet, Inception, ResNet, DenseNet, and AlexNet. Although these convolution-based approaches have achieved great performances in pneumonia detection to some extent, they can only achieve the higher-order features of ground objects in a distribution region, and fail to adaptively catch the geometric changes of different object regions in the X-ray images. Meanwhile, the boundary pixel classification attribute information may be mislaid while extracting the features, impacting the overall prediction. Hence, they cannot capture the higher-order feature information of all objects based on X-ray images. This is because the topology of the X-ray image’s dimensions does not always come with some spatially regular locality properties, which makes defining a spatial kernel filter for X-ray images non-trivial [5]. One solution to this limitation is to use graph signal computation-based convolutions.
In the last couple of years, graph neural networks (GNNs) have become increasingly important and have been a very popular topic for both academia and industry in the processing of non-Euclidean data because of their ability to learn useful features from a graph. GNN can operate directly on the graph structure and aggregate and transfer neighborhood information [6]. At the same time, it can also grasp higher-order representations of irregularly distributed data. Advances in computer vision show that Graph Neural Network (GNN) architectures perform very well in several image-processing tasks, such as object detection, segmentation, classification, and medical image computing. Shen et al. [7] improved the performance accuracy by applying the graph convolutional neural network to hyper-spectral remote sensing image classification. Cai et al. [8] combined graph convolution integration and a cross-attention mechanism for classification of remote sensing images. They first obtained low-dimensional attributes, which are more expressive. Next, they performed the prediction on the hyperspectral data using attributes and the relationship between them, which were generated by the graph convolution integration algorithm. Although these methods effectively learn pixel distribution in an image, they cannot capture enough information from the neighbors of the node in a single layer, leading to limitations in their expressive power and learning ability.
In this paper, we propose an efficient principal neighborhood aggregation-based graph convolutional network framework for pneumonia detection (PNA-GCN). Our proposed model has three essential components: feature extraction, graph feature reconstruction, and representation learning. Unlike typical ML approaches, which are heavily based on manually-crafted features, our model extracts abstract data features in two main steps: transferring state-of-the-art CNNs to extract features, and graph reconstruction based on the extracted features. In our method, we utilize deep-learning based algorithms that are effective at generalization. Specifically, we achieve feature extraction with the use of a transfer learning approach. After extracting features using a trained CNNs network, we build the features graph, in which each feature extracted from each image is represented as a node of the graphs. Finally, we propose a GCN with principal neighborhood aggregation, which combines multiple aggregators with degree-scalers.
To sum up, the key contributions of the work are summarized as follows:
  • To the best of our knowledge, we made the first attempt to use the graph convolutional network approach for pneumonia detection.
  • We propose PNA-GCN, an efficient principal neighborhood aggregation-based graph convolutional network framework for pneumonia detection. In PNA-GCN, we propose a new graph-based feature construction utilizing the transfer learning technique to extract features and construct the graph from images. Then, we propose a principal neighborhood aggregation-based graph convolutional network, in which we integrate multiple aggregation functions in a single layer with degree-scalers to enable each node to gain a better understanding of the distribution of messages it receives.
  • The performance of the proposed method is evaluated on the publicly available chest X-ray datasets. The accuracy, precision, recall, and F1 score have been utilized to evaluate the effectiveness of the proposed method compared to existing work in the literature.

3. Materials and Methods

In this section, we describe the framework of the proposed model, as shown in Figure 3. In classical machine learning (ML) methods, data features are first extracted and then classified by classifiers. Our proposed PNA-GCN model has four essential components: data augmentation and preprocessing, feature extraction, graph feature reconstruction, and representation learning. We first apply the image preprocessing and data augmentation strategy on a pneumonia dataset. Unlike typical ML approaches which heavily based on manually-crafted features, our model extracts abstract data features utilizing two main steps: transferring state-of-the-art CNNs to extract features, and graph reconstruction based on the extracted features. In the following, we discuss the four main components of our proposed method for pneumonia disease detection.
Figure 3. The methodology flowchart.

3.1. Data Augmentation and Preprocessing

We used the data augmentation strategies that have been published in the literature to alleviate the problem of overfitting and improve the model’s capacity to generalize during the training phase, and to raise the amount and quality of the data [42]. The parameters that were used in data augmentation are listed in Table 1 on the right. The rescaling of the image is the first step in the process (reduction or magnification during the augmentation process). Following that, we perform rotation of the images, which is rotated at random throughout the training. It is the width shift that determines how far the images are offset horizontally, and it is the height shift that determines how far the images are offset vertically. In our situation, the width shift and the height shift were both 10 percent. Finally, the images were rotated horizontally and zoomed by a factor of 20 percent at random.
Table 1. The settings for the image augmentation process.

3.2. Feature Extraction

Feature extraction plays an important role in classification tasks, which affects the overall performance of classifiers. In our method, we utilized deep-learning-based algorithms that are effective at generalization. Specifically, we achieved feature extraction with the use of a transfer learning approach. In the feature extraction step, we first transferred state-of-the-art networks to the binary classification problem by replacing the top layers with new ones. After training on the training set, CNNs can generate preliminary findings and features. Generally, CNNs are trained using ImageNet [43], which provides classification results for 1000 categories. We used the general architecture of CNNs [44]. After transferring the state-of-the-art networks, we chose the CNN that performed the best on the test set to serve as a feature extractor for PNA-GCN.
In our proposed architecture, the transferring process was done by removing the top layers of the general architecture of CNNs and replacing them with dropout, transitional fully-connected (256 channels), and classification layers.
Figure 4 shows the architecture of the transferred state-of-the-art CNNs after removing the top layers of CNNs and adding the new layers. After the final pooling layer, we added one dropout layer, one 256-channel transitional fully-connected layer, and a final 2-channel fully-connected layer for classification. In order to avoid overfitting problems during the training period, a dropout layer was included. If the size of a feature drops rapidly, the information contained within the feature will be significantly reduced. A transitional fully-connected layer is therefore on top of the dropout layer in order to prevent significant information loss from occurring. FC256 and FC2 are two fully-connected layers with 256 and 2 channels, respectively. As illustrated in Figure 4, the link between the final pooling layer and the softmax layer has been replaced with a connection between the last pooling layer and the newly added dropout layer. The parameters inside the CNNs were fine-tuned to offer better representations of the dataset after training with the pneumonia dataset for a limited number of epochs.
Figure 4. The architecture of the transferred CNNs. FC256 and FC2 are the fully connected layers with 256 and 2 channels, respectively.
The details of how features are acquired in our architecture can be seen in the following steps: network transferring and feature extraction, where the acquired features were analyzed for underlying graph representation.
  • In network transferring step, we first load a pre-trained convolution neural networks, which has been trained on the ImageNet dataset. Then, we remove the softmax and classification layers. After that, we add new layers, including a dropout layer and fully-connected layers with 256 and 2 channels (FC2 and FC256 in Figure 4). Using predefined parameters, we train the new networks on the training set of pneumonia dataset and save the networks and parameters.
  • In feature extraction step, we first load the trained networks in the first step. Then, the target dataset is used as input to the network for feature extraction. We extract the features generated by the fully-connected layer (FC256 in Figure 4).

3.3. Graph Construction

As shown in the overall architecture of the proposed model in Figure 5, first, the images were used as input to the pretrained deep CNNs described in in Section 3.2 to extract the image vector features. After extracting image vector features using trained deep CNNs network, we built the features graph, in which each feature extracted from an image is represented as a node of the graph, and the edges were built by calculating the distances between vector features, as described below. After that, the proposed principal neighborhood aggregation GCN was applied to the constructed graph. Finally, we applied the multi-layer perceptron MLP to pneumonia detection.
Figure 5. The overall architecture of the proposed model. In the architecture, we first use deep CNNs to extract the image vector features. Then, we construct graphs by calculating the distances between image vector features. Next is multiple principal neighborhood aggregation GCN. Each PNA-GCN is followed by a GRU layer. Finally, a MLP is used to classify features and images are labeled.
For faster computation, features are broken up into batches. Given the features F R D × M , where D is the number of images and M is the feature dimension, let the batch size be B. Then the number of batches n can be defined as:
n = D / B
where · is the ceiling operation.
For each batch F i , the graph ( V i , E i ) is constructed, which represents the underlying relationship between image vector features (nodes). V i indicates the nodes of the graph, which represents the feature vector for an image i in the batch, and E i represents the edges between nodes (image feature vectors). We utilize Euclidean distance to construct the edges among nodes. We build the edges between each node and its k neighbors if the k smallest Euclidean distance is found. Then, we build the adjacency matrix A R B × B . When the node f i and its neighbor f i + 1 are related, the value in A i at the position i , i + 1 is set to a positive number.
Given a batch of features, wherein each feature represents an image, we construct the graph according to the following process:
First, we initialize the adjacency, distance, and index matrices as follows:
A i = z e r o s ( B , B )
D i s t a n c e = I n f i n i t y ( B , B )
S o r t e d _ D i s t a n c e = z e r o s ( B , B )
I n d e x = z e r o s ( B , B )
where B is the number of samples in the batch. Note that A i , D i s t a n c e , S o r t e d _ D i s t a n c e , and I n d e x are initialized variables.
The Distance is calculated as follows:
D i s t a n c e ( j , m ) = h = 1 M ( f j ( h , 1 ) f m ( h , 1 ) ) 2
After calculating A i , each feature f j in batch F i is recomputed as follows:
f j = A ˜ i j F i
where A ˜ i is the normalized adjacency matrix A i and A ˜ i j is the j th row of A ˜ i .
To build the graph G i for each batch of features F i , we follow the following steps:
  • Calculate the distances between each feature and the other features in the batch. Then, the distance matrix D i s t a n c e R B × B is computed.
  • Sort Each row of the distance matrix in a ascending order.
  • Generate the corresponding index matrix I n d e x R B × B , in which the nearest k features are recorded in the batch F i .
  • Set the value of A i at the position ( j , j ) to 1 if the features B i and B j are nearest to each other based on the distance matrix.

3.4. Principal Neighborhood Aggregation-Based Graph Convolutional Network

In this section, first, we introduce GCN, and then we introduce the proposed principal neighborhood aggregation, which combines multiple aggregators with degree-scalers.

3.4.1. Graph Convolutional Networks

GNNs are used to transfer a non-linear mapping g from a graph to a feature vector for the graph classification task.
g : G F G ,
where F G is a feature vector of the entire graph G that is used for estimating the label of the graph. Based on the neighborhood aggregating techniques, a new perspective divides GNNs into two groups [45]. The spectral-based convolutional GNNs [46,47] are the first group (spectral GNN). The spectral decomposition of graphs inspired this group of GNNs, which try to approximate the spectral filters in each aggregating layer [48,49]. The spatial-based convolutional GNNs are the other type of GNN (spatial GNN). They do not set out to learn spectral properties of graphs; instead, they execute neighborhood aggregation based on the spatial relationships between nodes. The message passing neural network is a well-known example of a spatial GNN (MPNN) [50], and the GIN is another [51].
Inspired by CNNs, the GCN [49] is a multi-layer neural network that works on a graph and tries to find high-level features by combining information from the neighborhoods of graph nodes with information from other graph nodes. Formally, the undirected graph G = ( V ; E ) is defined in GCN as the set of nodes and edges. In G , the adjacency matrix ( A ) is used to show the presence of an edge between each pair of nodes. Specifically, the spatial domain convolution operation may be used to implement the first-order approximation of the Chebyshev expansion of the spectral convolution operation:
X ( l ) = σ D ˜ 1 2 A ˜ D ˜ 1 2 X ˜ ( l 1 ) W ˜ ( l ) R N × C ( l )
where A ˜ = A + I is the adjacency matrix with the recurring loop, D ˜ is the degree matrix, and X ˜ ( l 1 ) is calculated as follows:
X ( l ) = x 1 ( l ) x C ( l ) ( l ) R N × C ( l )
where C ( l ) represents the channel signals at the l th layer. This means that GCN implements the node feature with its neighbors via a layer-specific learnable weight matrix W ( l ) and non-linearity σ .
In contrast to spectral GNNs, spatial-based approaches define graph convolutions based on the spatial relations of a node. In general, this operation consists of the AGGREGATE and COMBINE functions:
a v ( l ) = A G G R E G A T E ( l ) p v ( l 1 ) : u N ( v ) ,
p v ( l ) = C O M B I N E ( l ) p v ( l 1 ) , a v ( l ) ,
where p v ( l ) R C ( l ) is the l th layer feature at the v th node. This means that the AGGREGATE function collects features from nearby nodes to produce an aggregated feature vector for the layer l, and the COMBINE function then combines the previous node feature p v ( l 1 ) with the aggregated node features a v ( l ) to produce the current layer’s node feature p v ( l ) . The mapping is defined after this spatial process by:
P G = R E A D O U T p v ( l ) | v G

3.4.2. Principal Neighborhood Aggregation

Most work in the literature uses only a single aggregation method (Equation (11)) Mean, sum, and max aggregators are the most used in the state-of-the-art models. However, in the literature, we observed how various aggregators fail to discriminate between different messages when using a single GNN layer. In our method, we first apply degree-based scalers [52]. We use the logarithmic scaler S a m p , which is computed as follows:
S a m p ( d ) = l o g ( d + 1 ) δ , δ = 1 | t r a i n | i t r a i n l o g ( d i + 1 )
where δ is a normalization parameter that was calculated over the training set. d is the degree of the node that is getting the message. Then, we generalize this scaler as follows:
S ( d , α ) = l o g ( d + 1 ) δ α , d > 0 , 1 α 1
where α is the value of a variable parameter that can be negative for attenuation, positive for amplification, or zero for complete lack of scaling. Other definitions of S ( d ) can be used, such as a linear scaling definition, as long as the function is injective for S ( d ) greater than zero.
The principal neighborhood aggregation function is created by combining the aggregators and scalers described in the preceding sections (PNA). As detailed in following equation, we employed four neighbor-aggregations with three degree-scalers each to evaluate this general and flexible architecture.
a v ( l ) = I S ( D ˜ , α = 1 ) S ( D ˜ , α = 1 ) s c a l e r s μ σ m a x m i n a g g r e g a t o r s
where mean ( μ ), standard deviation ( σ ), max, and min are the aggregators defined in [52]; scalers are defined in Equation (15); and ⊗ is the product.
M convolutions were used for the experiments, and then three fully-connected layers were used to label nodes. This architecture, shown in Figure 5, was used for the experiments. After each layer’s update function, gated recurrent units (GRUs) [53] were added to each layer to help it run faster. Their ability to keep information from previous layers worked well when we added more convolutional layers M.

4. Experiments

4.1. Dataset

In this research, we used chest X-ray images from Kaggle [54]. The dataset is available online. The dataset contains 5856 chest X-ray images in the JPEG format of children under six years old that were captured during routine clinical care of patients. As shown in Table 2, the training set contained 4232 images divided between 3883 images classified as depicting pneumonia and 1349 images classified as normal. It should be noticed the training set is imbalanced, as shown in Figure 6. The model was evaluated with 390 images classified as depicting pneumonia and 234 images classified as normal. Examples of healthy chest X-ray images and pneumonia infected chest X-ray images are shown in Figure 1 and Figure 2, respectively.
Table 2. Dataset description.
Figure 6. The training set distribution of image classes showing the disparity in data used for training.

4.2. Experimental Settings and Evaluation Criteria

For the experiment, we used a consistent setup. We trained the model for 70 epochs with the Adam optimizer with an initial learning rate of 5 × 10 4 and weight decay of 10 6 . The mini-batch sizes were set to be 32 and we used add as an aggregation operator. To reduce over-fitting, we utilized dropout with a rate of 0.6 for embeddings and used a rate of 0.3 for aggregation module outputs. We set the number of multi-head to be 4. We chose the image size to be 150 × 150 × 3 . All our code was implemented in Pytorch [55].

Evaluation Criteria

We tested a number of models on the test dataset for a fair comparison. For model evaluation, we used four performance metrics: accuracy, precision, recall, and F1 score.
  • The accuracy metric is the ratio of the number correctly predicted images to the total number of images examined.
    A c c u r a c y = ( T r u e   P o s i t i v e + T r u e   N e g a t i v e ) t o t a l   n u m b e r   o f   i m a g e s
  • The precision metric is the ratio of the number correctly predicted pneumonia images to the sum of the number correctly predicted pneumonia images and the number of normal images incorrectly identified as pneumonia images.
    P r e c i s i o n = T r u e   P o s i t i v e T r u e   P o s i t i v e + F a l s e   P o s i t i v e
  • The recall metric is the ratio of the number of correctly predicted pneumonia images to the sum of the number correctly predicted pneumonia images and the number of pneumonia images incorrectly identified as normal.
    R e c a l l = ( T r u e   p o s i t i v e ) T r u e   p o s i t i v e + F a l s e   N e g a t i v e
  • F1 score is the weighted harmonic mean of recall and precision.
    F 1   s c o r e = 2 × ( R e c a l l × P r e c i s i o n ) R e c a l l + P r e c i s i o n

4.3. Results and Discussion

This section represents the experimental results achieved for the considered X-ray images with chosen dimensions. The equivalent convergence for accuracy and the loss function of the outcomes achieved from various convolution layers for training and validation process are shown in Figure 7, and the confusion matrix is shown in Figure 8. The merits of our method were further verified by comparing it with other algorithms from the literature, including those of Kermany et al. [54] and Lahsaini et al. [56]. The outcomes shown in Table 3 substantiate that the results achieved with the proposed method were better result compared to the other methods on the chosen dataset.
Figure 7. Training and validation performance.
Figure 8. Confusion matrix of PNA-GCN on test data.
Table 3. Comparison of results.

4.4. Ablation Study

4.4.1. Input Size

The dimensions of processed images were fixed. To evaluate the validation performance of our model, we resized the chest X-ray images at 300 × 300 × 3, 250 × 250 × 3, and 150 × 150 × 3, respectively, and we trained them for 30 epochs each. The results obtained showed that the shape of the 150 × 150 image gave a better result than the others, as the model achieved higher validation accuracy with it. The experimental results are presented in Figure 9.
Figure 9. Performance of PNA-GCN with different data sizes. (a) Data size = 300 × 300 × 3. (b) Data size = 250 × 250 × 3. (c) Data size = 150 × 150 × 3.

4.4.2. The Influence of Batch Size N

We explored the effects of different values for the batch size N on the PNA-GCN performance. We chose 16, 32, 50, and 64, respectively. The experiment results plotted in Figure 10 show that PNA-GCN performs best with N = 32 , alleviating the memory problem during training.
Figure 10. Performance comparison with different values for batch size.

4.4.3. Effectiveness of the Multi-Head Attention

Rather than using the attention function just once, we use a multi-head attention layer, as [58] showed that the learning process can be stabilized using multiple independent attention operators. We collected the number M of multi-total heads within the set (1, 2, 4, 6, 8) to show the effectiveness of the multi-head attention. The experimental results plotted in Table 4 show that PNA-GCN performs best with the number of multi-heads being 4.
Table 4. Results comparison with different values for multi-head attention.

4.5. Influence of the Number of Aggregators

We compare the performance of PNA-GCN when making use of multiple aggregators along with degree-based scalers with single aggregators, as shown in Figure 11. PNA-GCN performs better with multiple aggregators. Specifically, PNA-GCN performs well with a combination of mean aggregation, standard deviation aggregation, max aggregation, and min aggregation. That is because the model can benefit from the strengths of each of them.
Figure 11. Performance of PNA-GCN with variable numbers of aggregators. (a) One aggregator; (b) two aggregators; (c) three aggregators; (d) four aggregators.

5. Conclusions

In this work, we proposed PNA-GCN, a principal neighborhood aggregation-based graph convolutional network for detecting pneumonia. The proposed model was developed taking into account the criteria of parameters and the time cost and memory cost of the training step, in contrast to other approaches that are based on transfer learning or use a more complex architecture. In this work, binary classification is executed and four different evaluation metrics have been used to evaluate the proposed method. The experimental results confirmed that the PNA-GCN achieved an improved outcome compared with other related approaches existing in the literature.

Author Contributions

Conceptualization, A.A.A.G.; funding acquisition, G.J.; methodology, A.A.A.G.; project administration, G.J.; supervision, G.J.; investigation, A.A.A.G. and G.J.; resources, A.A.A.G.; data curation, A.A.A.G.; writing—original draft preparation, A.A.A.G.; writing—review and editing, A.A.A.G., G.J., B.M.O. and R.A.-S.; visualization, A.A.A.G., R.A.-S. and B.M.O.; software, A.A.A.G., B.M.O. and R.A.-S.; validation, A.A.A.G., B.M.O. and R.A.-S. All authors have read and agreed to the published version of the manuscript.

Funding

This research has been sponsored by the Natural Science Foundation of China (number 61873352).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Data is contained within the article.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
PNAprincipal neighborhood aggregation
GCNgraph convolutional network
GRUsgated recurrent units
GNNsgraph neural network
CNNconvolutional neural network
WHOworld health organization

References

  1. Varshni, D.; Thakral, K.; Agarwal, L.; Nijhawan, R.; Mittal, A. Pneumonia Detection Using CNN based Feature Extraction. In Proceedings of the 2019 IEEE International Conference on Electrical, Computer and Communication Technologies (ICECCT), Coimbatore, India, 20–22 February 2019; pp. 1–7. [Google Scholar] [CrossRef]
  2. Sharma, A.; Raju, D.; Ranjan, S. Detection of pneumonia clouds in chest X-ray using image processing approach. In Proceedings of the 2017 Nirma University International Conference on Engineering (NUiCONE), Ahmedabad, India, 23–25 November 2017; pp. 1–4. [Google Scholar]
  3. de Melo, G.; Macedo, S.O.; Vieira, S.L.; Oliveira, L.L. Classification of images and enhancement of performance using parallel algorithm to detection of pneumonia. In Proceedings of the 2018 IEEE International Conference on Automation/XXIII Congress of the Chilean Association of Automatic Control (ICA-ACCA), Concepción, Chile, 17–19 October 2018; pp. 1–5. [Google Scholar]
  4. Chaudhary, P.K.; Pachori, R.B. Automatic diagnosis of COVID-19 and pneumonia using FBD method. In Proceedings of the IEEE International Conference on Bioinformatics and Biomedicine (BIBM 2020), Seoul, Korea, 16–19 December 2020; pp. 2257–2263. [Google Scholar] [CrossRef]
  5. Edwards, M.; Xie, X. Graph based convolutional neural network. arXiv 2016, arXiv:1609.08965. [Google Scholar]
  6. Ma, T.; Wang, H.; Zhang, L.; Tian, Y.; Al-Nabhan, N. Graph classification based on structural features of significant nodes and spatial convolutional neural networks. Neurocomputing 2021, 423, 639–650. [Google Scholar] [CrossRef]
  7. Wan, S.; Gong, C.; Zhong, P.; Du, B.; Zhang, L.; Yang, J. Multiscale Dynamic Graph Convolutional Network for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote. Sens. 2020, 58, 3162–3177. [Google Scholar] [CrossRef] [Green Version]
  8. Cai, W.; Wei, Z. Remote sensing image classification based on a cross-attention mechanism and graph convolution. IEEE Geosci. Remote Sens. Lett. 2020. [Google Scholar] [CrossRef]
  9. Razzak, M.I.; Naz, S.; Zaib, A. Deep Learning for Medical Image Processing: Overview, Challenges and Future. arXiv 2017, arXiv:1704.06825. [Google Scholar]
  10. Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M.; van der Laak, J.A.W.M.; van Ginneken, B.; Sánchez, C.I. A survey on deep learning in medical image analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef] [Green Version]
  11. Stephen, O.; Sain, M.; Maduh, U.J.; Jeong, D.U. An efficient deep learning approach to pneumonia classification in healthcare. J. Healthc. Eng. 2019, 2019, 4180949. [Google Scholar] [CrossRef] [Green Version]
  12. Esteva, A.; Kuprel, B.; Novoa, R.A.; Ko, J.; Swetter, S.M.; Blau, H.M.; Thrun, S. Dermatologist-level classification of skin cancer with deep neural networks. Nature 2017, 542, 115–118. [Google Scholar] [CrossRef]
  13. Milletari, F.; Navab, N.; Ahmadi, S. V-Net: Fully Convolutional Neural Networks for Volumetric Medical Image Segmentation. In Proceedings of the Fourth International Conference on 3D Vision (3DV 2016), Stanford, CA, USA, 25–28 October 2016; pp. 565–571. [Google Scholar] [CrossRef] [Green Version]
  14. Grewal, M.; Srivastava, M.M.; Kumar, P.; Varadarajan, S. RADnet: Radiologist level accuracy using deep learning for hemorrhage detection in CT scans. In Proceedings of the 15th IEEE International Symposium on Biomedical Imaging (ISBI 2018), Washington, DC, USA, 4–7 April 2018; pp. 281–284. [Google Scholar] [CrossRef] [Green Version]
  15. Voets, M.; Møllersen, K.; Bongo, L.A. Replication study: Development and validation of deep learning algorithm for detection of diabetic retinopathy in retinal fundus photographs. arXiv 2018, arXiv:1803.04337. [Google Scholar]
  16. Bar, Y.; Diamant, I.; Wolf, L.; Lieberman, S.; Konen, E.; Greenspan, H. Chest pathology detection using deep learning with non-medical training. In Proceedings of the 12th IEEE International Symposium on Biomedical Imaging (ISBI 2015), Brooklyn, NY, USA, 16–19 April 2015; pp. 294–297. [Google Scholar] [CrossRef] [Green Version]
  17. Avni, U.; Greenspan, H.; Konen, E.; Sharon, M.; Goldberger, J. X-ray Categorization and Retrieval on the Organ and Pathology Level, Using Patch-Based Visual Words. IEEE Trans. Med. Imaging 2011, 30, 733–746. [Google Scholar] [CrossRef]
  18. Melendez, J.; van Ginneken, B.; Maduskar, P.; Philipsen, R.H.H.M.; Reither, K.; Breuninger, M.; Adetifa, I.M.O.; Maane, R.; Ayles, H.; Sánchez, C.I. A Novel Multiple-Instance Learning-Based Approach to Computer-Aided Detection of Tuberculosis on Chest X-Rays. IEEE Trans. Med. Imaging 2015, 34, 179–192. [Google Scholar] [CrossRef] [PubMed]
  19. Jaeger, S.; Karargyris, A.; Candemir, S.; Folio, L.R.; Siegelman, J.; Callaghan, F.M.; Xue, Z.; Palaniappan, K.; Singh, R.K.; Antani, S.K.; et al. Automatic Tuberculosis Screening Using Chest Radiographs. IEEE Trans. Med. Imaging 2014, 33, 233–245. [Google Scholar] [CrossRef] [PubMed]
  20. Hermann, S. Evaluation of Scan-Line Optimization for 3D Medical Image Registration. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2014), Columbus, OH, USA, 23–28 June 2014; pp. 3073–3080. [Google Scholar] [CrossRef]
  21. Nasrullah, N.; Sang, J.; Alam, M.S.; Mateen, M.; Cai, B.; Hu, H. Automated Lung Nodule Detection and Classification Using Deep Learning Combined with Multiple Strategies. Sensors 2019, 19, 3722. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. Yao, L.; Poblenz, E.; Dagunts, D.; Covington, B.; Bernard, D.; Lyman, K. Learning to diagnose from scratch by exploiting dependencies among labels. arXiv 2017, arXiv:1710.10501. [Google Scholar]
  23. Khatri, A.; Jain, R.; Vashista, H.; Mittal, N.; Ranjan, P.; Janardhanan, R. Pneumonia identification in chest X-ray images using EMD. In Trends in Communication, Cloud, and Big Data; Springer: Singapore, 2020; pp. 87–98. [Google Scholar]
  24. Abiyev, R.H.; Ma’aitah, M.K.S. Deep convolutional neural networks for chest diseases detection. J. Healthc. Eng. 2018, 2018, 4168538. [Google Scholar] [CrossRef] [Green Version]
  25. Rajaraman, S.; Candemir, S.; Kim, I.; Thoma, G.; Antani, S. Visualization and interpretation of convolutional neural network predictions in detecting pneumonia in pediatric chest radiographs. Appl. Sci. 2018, 8, 1715. [Google Scholar] [CrossRef] [Green Version]
  26. Sirazitdinov, I.; Kholiavchenko, M.; Mustafaev, T.; Yixuan, Y.; Kuleev, R.; Ibragimov, B. Deep neural network ensemble for pneumonia localization from a large-scale chest x-ray database. Comput. Electr. Eng. 2019, 78, 388–399. [Google Scholar] [CrossRef]
  27. Abbas, A.; Abdelsamea, M.M. Learning transformations for automated classification of manifestation of tuberculosis using convolutional neural network. In Proceedings of the 2018 13th International Conference on Computer Engineering and Systems (ICCES), Cairo, Egypt, 18–19 December 2018; pp. 122–126. [Google Scholar]
  28. Rajpurkar, P.; Irvin, J.; Ball, R.L.; Zhu, K.; Yang, B.; Mehta, H.; Duan, T.; Ding, D.; Bagul, A.; Langlotz, C.P.; et al. Deep learning for chest radiograph diagnosis: A retrospective comparison of the CheXNeXt algorithm to practicing radiologists. PLoS Med. 2018, 15, e1002686. [Google Scholar] [CrossRef]
  29. Ho, T.K.K.; Gwak, J. Multiple feature integration for classification of thoracic disease in chest radiography. Appl. Sci. 2019, 9, 4130. [Google Scholar] [CrossRef] [Green Version]
  30. Saraiva, A.A.; Santos, D.; Costa, N.J.C.; Sousa, J.V.M.; Ferreira, N.M.F.; Valente, A.; Soares, S. Models of Learning to Classify X-ray Images for the Detection of Pneumonia using Neural Networks. Bioimaging 2019, 76–83. [Google Scholar] [CrossRef]
  31. Ayan, E.; Ünver, H.M. Diagnosis of pneumonia from chest X-ray images using deep learning. In Proceedings of the 2019 Scientific Meeting on Electrical-Electronics & Biomedical Engineering and Computer Science (EBBT), Istanbul, Turkey, 24–26 April 2019; pp. 1–5. [Google Scholar]
  32. Rahman, T.; Chowdhury, M.E.; Khandakar, A.; Islam, K.R.; Islam, K.F.; Mahbub, Z.B.; Kadir, M.A.; Kashem, S. Transfer learning with deep convolutional neural network (CNN) for pneumonia detection using chest X-ray. Appl. Sci. 2020, 10, 3233. [Google Scholar] [CrossRef]
  33. Xiao, Z.; Du, N.; Geng, L.; Zhang, F.; Wu, J.; Liu, Y. Multi-scale heterogeneous 3D CNN for false-positive reduction in pulmonary nodule detection, based on chest CT images. Appl. Sci. 2019, 9, 3261. [Google Scholar] [CrossRef] [Green Version]
  34. Xu, S.; Wu, H.; Bie, R. CXNet-m1: Anomaly detection on chest X-rays with image-based deep learning. IEEE Access 2018, 7, 4466–4477. [Google Scholar] [CrossRef]
  35. Jaiswal, A.K.; Tiwari, P.; Kumar, S.; Gupta, D.; Khanna, A.; Rodrigues, J.J. Identifying pneumonia in chest X-rays: A deep learning approach. Measurement 2019, 145, 511–518. [Google Scholar] [CrossRef]
  36. Jung, H.; Kim, B.; Lee, I.; Lee, J.; Kang, J. Classification of lung nodules in CT scans using three-dimensional deep convolutional neural networks with a checkpoint ensemble method. BMC Med. Imaging 2018, 18, 48. [Google Scholar] [CrossRef]
  37. Chouhan, V.; Singh, S.K.; Khamparia, A.; Gupta, D.; Tiwari, P.; Moreira, C.; Damaševičius, R.; De Albuquerque, V.H.C. A novel transfer learning based approach for pneumonia detection in chest X-ray images. Appl. Sci. 2020, 10, 559. [Google Scholar] [CrossRef] [Green Version]
  38. Liang, X.; Zhang, Y.; Wang, J.; Ye, Q.; Liu, Y.; Tong, J. Diagnosis of COVID-19 pneumonia based on graph convolutional network. Front. Med. 2021, 7, 1071. [Google Scholar] [CrossRef]
  39. Keicher, M.; Burwinkel, H.; Bani-Harouni, D.; Paschali, M.; Czempiel, T.; Burian, E.; Makowski, M.R.; Braren, R.; Navab, N.; Wendler, T. U-GAT: Multimodal Graph Attention Network for COVID-19 Outcome Prediction. arXiv 2021, arXiv:2108.00860. [Google Scholar]
  40. Chen, B.; Zhang, Z.; Lu, Y.; Chen, F.; Lu, G.; Zhang, D. Semantic-interactive graph convolutional network for multilabel image recognition. IEEE Trans. Syst. Man Cybern. Syst. 2021, 1, 13. [Google Scholar] [CrossRef]
  41. Mondal, R.; Mukherjee, D.; Singh, P.K.; Bhateja, V.; Sarkar, R. A New Framework for Smartphone Sensor-Based Human Activity Recognition Using Graph Neural Network. IEEE Sensors J. 2020, 21, 11461–11468. [Google Scholar] [CrossRef]
  42. Roth, H.R.; Lu, L.; Seff, A.; Cherry, K.M.; Hoffman, J.; Wang, S.; Liu, J.; Turkbey, E.; Summers, R.M. A new 2.5 D representation for lymph node detection using random sets of deep convolutional neural network observations. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Boston, MA, USA, 14–18 September 2014; pp. 520–527. [Google Scholar]
  43. Deng, J.; Dong, W.; Socher, R.; Li, L.; Li, K.; Fei-Fei, L. ImageNet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR 2009), Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar] [CrossRef] [Green Version]
  44. Yu, X.; Wang, S.; Zhang, Y. CGNet: A graph-knowledge embedded convolutional neural network for detection of pneumonia. Inf. Process. Manag. 2021, 58, 102411. [Google Scholar] [CrossRef] [PubMed]
  45. Wu, Z.; Pan, S.; Chen, F.; Long, G.; Zhang, C.; Philip, S.Y. A comprehensive survey on graph neural networks. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 4–24. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  46. Bianchi, F.M.; Grattarola, D.; Alippi, C. Spectral clustering with graph neural networks for graph pooling. In Proceedings of the 37th International Conference on Machine Learning Research, Virtual Event, 13–18 July 2020; pp. 874–883. [Google Scholar]
  47. Zhang, Y.; Wu, B.; Liu, Y.; Lv, J. Local community detection based on network motifs. Tsinghua Sci. Technol. 2019, 24, 716–727. [Google Scholar] [CrossRef]
  48. Bruna, J.; Zaremba, W.; Szlam, A.; LeCun, Y. Spectral Networks and Locally Connected Networks on Graphs. In Proceedings of the 2nd International Conference on Learning Representations (ICLR 2014), Banff, AB, Canada, 14–16 April 2014. [Google Scholar]
  49. Kipf, T.N.; Welling, M. Semi-Supervised Classification with Graph Convolutional Networks. In Proceedings of the 5th International Conference on Learning Representations (ICLR 2017), Toulon, France, 24–26 April 2017. [Google Scholar]
  50. Gilmer, J.; Schoenholz, S.S.; Riley, P.F.; Vinyals, O.; Dahl, G.E. Neural Message Passing for Quantum Chemistry. In Proceedings of the 34th International Conference on Machine Learning (ICML 2017), Sydney, Australia, 6–11 August 2017; pp. 1263–1272. [Google Scholar]
  51. Xu, K.; Hu, W.; Leskovec, J.; Jegelka, S. How Powerful are Graph Neural Networks? In Proceedings of the 7th International Conference on Learning Representations (ICLR 2019), New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  52. Corso, G.; Cavalleri, L.; Beaini, D.; Liò, P.; Velickovic, P. Principal Neighbourhood Aggregation for Graph Nets. In Proceedings of the Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020 (NeurIPS 2020), Virtual Event, 6–12 December 2020. [Google Scholar]
  53. Cho, K.; van Merrienboer, B.; Gülçehre, Ç.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning Phrase Representations using RNN Encoder-Decoder for Statistical Machine Translation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP 2014), Doha, Qatar, 25–29 October 2014; pp. 1724–1734. [Google Scholar] [CrossRef]
  54. Kermany, D.S.; Goldbaum, M.; Cai, W.; Valentim, C.C.; Liang, H.; Baxter, S.L.; McKeown, A.; Yang, G.; Wu, X.; Yan, F.; et al. Identifying medical diagnoses and treatable diseases by image-based deep learning. Cell 2018, 172, 1122–1131. [Google Scholar] [CrossRef] [PubMed]
  55. Fey, M.; Lenssen, J.E. Fast graph representation learning with PyTorch Geometric. In Proceedings of the 7th International Conference on Learning Representations, Workshop, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  56. Lahsaini, I.; El Habib Daho, M.; Chikh, M.A. Convolutional neural network for chest x-ray pneumonia detection. In Proceedings of the 1st International Conference on Intelligent Systems and Pattern Recognition, Virtual Event, 16–18 October 2020; pp. 55–59. [Google Scholar]
  57. Arunmozhi, S.; Rajinikanth, V.; Rajakumar, M. Deep-Learning based Automated Detection of Pneumonia in Chest Radiographs. In Proceedings of the 2021 International Conference on System, Computation, Automation and Networking (ICSCAN), Puducherry, India, 30–31 July 2021; pp. 1–4. [Google Scholar]
  58. Veličković, P.; Cucurull, G.; Casanova, A.; Romero, A.; Lio, P.; Bengio, Y. Graph attention networks. arXiv 2017, arXiv:1710.10903. [Google Scholar]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.