Next Article in Journal
Understanding Historical Demographic Processes to Inform Contemporary Conservation of an Arid Zone Specialist: The Yellow-Footed Rock-Wallaby
Previous Article in Journal
Genetic and Genomic Tools in Sunflower Breeding for Broomrape Resistance
Previous Article in Special Issue
Detection of Microaneurysms in Fundus Images Based on an Attention Mechanism
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Network Embedding the Protein–Protein Interaction Network for Human Essential Genes Identification

1
Faculty of Information Engineering and Automation, Kunming University of Science and Technology, Kunming 650050, China
2
Computer Technology Application Key Lab of Yunnan Province, Kunming University of Science and Technology, Kunming 650050, China
3
College of Engineering and Design, Hunan Normal University, Changsha 410081, China
*
Author to whom correspondence should be addressed.
Genes 2020, 11(2), 153; https://doi.org/10.3390/genes11020153
Submission received: 17 December 2019 / Revised: 27 January 2020 / Accepted: 29 January 2020 / Published: 31 January 2020

Abstract

:
Essential genes are a group of genes that are indispensable for cell survival and cell fertility. Studying human essential genes helps scientists reveal the underlying biological mechanisms of a human cell but also guides disease treatment. Recently, the publication of human essential gene data makes it possible for researchers to train a machine-learning classifier by using some features of the known human essential genes and to use the classifier to predict new human essential genes. Previous studies have found that the essentiality of genes closely relates to their properties in the protein–protein interaction (PPI) network. In this work, we propose a novel supervised method to predict human essential genes by network embedding the PPI network. Our approach implements a bias random walk on the network to get the node network context. Then, the node pairs are input into an artificial neural network to learn their representation vectors that maximally preserves network structure and the properties of the nodes in the network. Finally, the features are put into an SVM classifier to predict human essential genes. The prediction results on two human PPI networks show that our method achieves better performance than those that refer to either genes’ sequence information or genes’ centrality properties in the network as input features. Moreover, it also outperforms the methods that represent the PPI network by other previous approaches.

1. Introduction

Essential genes are a group of genes that are indispensable for cell survival and cell fertility [1]. Scientists focus on studying essential genes in bacteria to find potential drug targets for new antibiotics [2]. Identifying essential genes in prokaryotic or simple eukaryotic organisms helps us understand the primary requirement for cell life, but also points the way to synthetic biology research that aims to create a cell with a minimal genome. The development of wet-lab experimental techniques, such as single-gene deletions [3], RNA interference [4], and conditional knockouts [5], has accumulated several essential genes in some simple eukaryotic organisms such as S. cerevisiae [6], C. elegans [7], and A. thaliana [8], in the past few years. However, it is hard to apply these techniques to explore human essential genes due to technical difficulties and species limitations [9].
The close relationships between essential genes and human diseases found by previous researches has motivated scientists to identify essential genes in humans to seek guidance for disease treatment [10,11,12]. However, it is hard to apply the computational methods that have successfully predicted essential genes of some simple eukaryotic organisms to humans. The biggest challenge is the lack of known human essential genes, because a human is a high and complex organism, which requires more than basic cell survival and fertility to be completely functional and adapted to the environment. The techniques mentioned above are difficult to use to explore human essential genes correctly. Recently, with the emergence of the CRISPR-Cas gene-editing system, several landmark studies [13,14,15] have individually disrupted every human gene in several cell lines in culture and examined how the loss of each gene affects cell viability by using the gene-editing system. Together, these studies provide the first accurate picture of the sets of genes that are absolutely required for the viability of individual human cell types, a long-awaited analysis that presages efficient screens for tumor-specific essential genes. They have published about 2000 human essential genes [13,14,15] that are indispensable for the viability of individual human cancer cell lines.
The publication of human essential gene data has motivated researchers to deeply study the functions of human genes. The popular way is to train a machine-learning classifier by inputting the features of some known human essential genes and use the classifier to predict new human essential genes. The outputs are lists of genes with the probabilities to be essential. Then, we set a threshold, i.e., 0.5. If the probabilities are above the threshold, the genes are predicted as essential genes. Guo et al. [16] extracted features from genes’ DNA sequence by a Z-curve method and input these features into an SVM classifier to predict human essential genes. As far as we know, in addition to DNA sequences, there are other biological properties related to the essentiality of genes. Guo’s work inspired us to design a more effective method and extract powerful features to improve the prediction accuracy of human essential genes.
One of the most important features related to the genes’ essentiality is their topology properties in biological networks. The studies on some simple eukaryotic organisms have found that the proteins (genes) with a larger number of interactive partners in the protein–protein interaction (PPI) network are highly likely to be essential proteins (genes), because removing them from the networks would cause the collapse of the network [17]. Therefore, some centrality methods are frequently employed to measure the importance of the genes in the network [9], including degree centrality (DC) [18], closeness centrality (CC) [19], betweenness centrality (BC) [20], information centrality (IC) [21], eigenvector centrality (EC) [22], subgraph centrality (SC) [23] and edge clustering coefficient centrality (NC) [24]. Since the PPI data currently available are incomplete and error prone, some researchers design the composited features by integrating PPI networks with other biological information, such as gene functional annotation data [25], gene expression data (PeC [26], WDC [27], FDP [28]), subcellular information [29], protein domain information [29], and orthologous information [30].
The features mentioned above display the genes’ essentiality from different perspectives. Previous researchers have made great efforts to improve the essential gene prediction accuracy in some simple eukaryotic organisms by employing machine-learning methods and combining some topological and biological features [9]. For example, Chen et al. [6] trained an SVM classifier by the features of some known yeast essential proteins, including the combination of phyletic retention, proteins’ evolutionary rate, paralogs, protein size, and the degree centrality in PPI networks and gene-expression networks. Then, the classifier was adopted to predict other potential yeast essential genes. Zhong et al. [31] developed an XGBoost-based classifier to integrate multiple topology properties of essential genes in the PPI network to predict yeast and E. coli essential genes.
As for human essential gene identification, we can draw a lesson from previous studies on other species. However, a machine-learning method that combines multiple biological and topological properties to make predictions still cannot take full advantage of the characteristics of network topology. Genes or their protein products interact with each other to perform their functions in cells together [12,32]. Their interactions in the PPI network play an important role in essential gene prediction. Recently, with the development of biological techniques, the human PPI network becomes even more complete and correct. Therefore, it is imperative to find an effective method to extract the characteristics of genes in the network and represent the human PPI network as much as possible.
Network embedding is an efficient method to mine the useful information from the network, which converts the network nodes into a vector of low dimensional space while maximally preserves the network structural information and the network properties. There are many network embedding methods [33,34,35,36,37]. In this work, we learn the vector representation of the nodes in the human PPI network by using a random walk-based network embedding method. This kind of method, first, implements random walk on the network to get the network context of a starting node. Then, a word2vec [33] method that originates for natural language processing inputs the node pairs into an artificial neural network and learns the representation vectors for the nodes in the PPI network. The random walk-based network embedding methods different in the way of walking on the network. The DeepWalk [34] method implements a depth-first search on the network, whereas The LINE [35] method implements a breadth-first search strategy to generate the node context. The vertices that share similar neighbors in the network tend to be identical to each other [32]. Hence, in this work, we conduct a bias random walk on the human PPI network to generate the note context, which means a node more likely walks to the connecting nodes sharing similar neighbors with it [38]. After that, the node pairs are input into a neural network to represent the features of genes in the PPI network and, then, these features are put into a variety of classifiers to label whether or not the genes are essential. In contrast to previous supervised machine-learning methods, our method conducts network embedding to extract the features for the genes from the PPI network through mapping them to latent space vectors that maximally preserve the relationships between the nodes and their network neighborhoods. The prediction results on two human PPI networks show that our method achieves better performance that either refers to genes’ sequence information [16] or selects genes’ centrality properties in the network as input features. It also outperforms the methods that represent the PPI network by other previous approaches, i.e., Deepwalk [34] and Line [35]. Part of the contents appeared in the conference paper [39]. In this work, we comprehensively discuss the advantage of bias random walk in the course of network embedding learning, but also investigate what the feature representations of the human essential genes are.

2. Methods

Our method involved two steps to predict humane essential genes. First, every node in the human PPI network represents a vector that can preserve the structure of the input network. Secondly, the vector is put into a classifier to train the classifier and make predictions. Similar to the node2vec model [38], our method learns the feature representation for every node in the PPI network by the following steps: The first step is to conduct a bias random walk on the PPI network starting from a single node and, then, to obtain a node sequence that records the network context of the starting node. Next, a window slides along the node sequence to get the node pairs that have a close relationship in the network. These node pairs are put into a skip-gram-based architecture to learn the feature representation of every node. Figure 1 shows an overview of our method for identifying essential genes.

2.1. Bias Random Walk

Network embedding methods originate for learning features for words in the natural language processing field. The input of the methods is a word sequence that represents the neighborhood relationship of the word in the language context. The word and its neighbor are the input and output layers of an artificial neural network. The weights of the hidden layer are the “node vectors” that we are trying to learn. In this work, the input is a PPI network, which is not linear and is composed of nodes with complicated neighborhood relationships. Hence, to resolve this issue, a bias random walk is simulated on the network to get an ordered node sequence starting from every node in the network.
Let G = (V,E,W) denote a PPI network, where a vertex v∈ V is a gene and an edge e (u,v) ∈ E connects the genes v and u and w (u,v) ∈ W is the weight of edge e (u,v), which measures the reliability of the connection between u and v. Here, all edges in the PPI network are treated equally, and their weights are set to one. Considering the module nature of the PPI network, a vertex walks to its neighbors at different probabilities. That is, given a vertex v, there is an edge e (u,v), where vertex u is visited by vertex v at the previous step. The vertex v walks to one of its neighbor k by evaluating the transition probabilities T (v,k) on edge e (v,k). The transition probabilities T (v,k) is defined as follows:
T (v,k) = π (u,k) * w (v,x),
with   π ( u , k ) { 1 p ,     d uk = 0   1 ,   d uk = 1   1 q ,   d uk = 2 ,
where duk denotes the shortest path distance from the previous vertex u to the next vertex k. Especially, duk = 0 means that the vertices u and k are the same vertex and the vertex v jumps back to its previous vertex u, duk = 1 means that the vertex k is the common neighbor of the vertices u and v, and duk = 2 means that the u and k connect indirectly, and the vertex k is not their common neighbor. The parameter p controls the likelihood of the vertex v traveling back to its previous vertex u. The parameter q controls the probability of the vertex v jumping to the next vertex k. Setting different values for the parameters p and q guides a bias walk on the network. If the parameters p and q are larger than one, the vertex v is more likely to walk to the vertex that is the common neighbor of it and its previous step, i.e., the vertex u. After computing the transition probabilities for every edge in the PPI network, we row normalize the transition probability matrix T to ensure the sum of the walk-out probabilities for each node to be one.
  T n o r m ( v , k ) = T ( v ,   k ) j N ( v ) T ( v ,   j ) ,
In bias random walk, a node tends to walk to its neighbor along the edge with the highest transition probability. Suppose a node has number N of neighbors, it usually needs O(N) time to find the walk-out edge with the highest transition probability. This work adopts an alias sampling method to find the edge in O(1) time [38]. The outcome of the bias random walk is a vertex sequence that records the walking trace in the network starting from a vertex, which reserves the neighborhood relationship of the starting vertex in the network. Algorithm 1 gives the pseudocode of the bias random walk.
Algorithm 1. Bias random walk algorithm.
Input G = (V, E, W), Len_walkLists, parameters w, p and q;
Output vertex sequence lists: walkLists
T = computing transition probabilities (G, p, q, w)//computing transition probabilities for every edge in the network
Tnorm = normalizing T by Equation (2) G’ = (V, E, Tnorm)
walkLists = {}
for iter = 1 to Len_walkLists do
    for every node u ∈ V do
      Append u to seq
      while len(seq) < w:
        t = seq [-1] // getting the last node of the set seq
        N(t) = sort (GetNeighbors(t, G’)) // sorting neighbor list of current vertex in alphabetic
                      order
        n = AliasSampling(N(t), Tnorm) //applying alias sampling with respect to the normalized
            transition probabilities to select a next visiting neighbor node
      Append n to seq
    Append seq to walkLists
return walkLists

2.2. Feature Learning

After getting the neighbor sequence of every starting vertex in the network by a bias random walk, a window (a parameter c controls the window size) slides along the sequence to get the neighbors of an input vertex. We set a window size two, for example. If the sequence of vertices is {1,2,3,4,5…} and Vertex 1 is the input vertex. The training samples should be {1,2}, {1,3}, {2,1}, {2,3}, {2,4}, {3,2}, {3,1}, {3,4}, {3,5}, {4,3}, {4,2}, {4,5}, {5,4}, and {5,3} (see Figure 1). Then these vertices and their neighbors are input into a skip-gram with negative sampling architecture (SGNS, also called word2vec method) [33] to finish the feature learning for every vertex in the network. Following is a brief overview of the SGNS method.
Give an undirected graph G = (V,E,W), for every vertex v∈V, we can get its sequence of neighbor {vm} in the network by a bias random walk. Suppose there are number m of vertices in the sequence. Let f: v∈V->Rd be the function that maps a vertex v to its feature representation vector and f(v) denotes the feature representation of the vertex v. Here, d is the size of the feature representation vector, which is empirically set to 64 [33]. Given a vertex v in the input vertex sequence, N(v) denotes its neighborhood set. The SGNS architecture aims to build a model that maximizes the log-probability of neighbors N(v) around a vertex v under the assumption that its feature representation is f (v). The objective function is formally defined as follows:
max f v V log P ( N ( v ) | f ( v ) ) ,
Given the feature representation of the node vertex v, we factorize the likelihood by assuming that the probability of observing a neighbor is independent of observing any other neighborhood node [38].
P ( N ( v ) | f ( v ) ) = u N ( v ) P ( u | f ( v ) ) ,  
where P(u|f(v)) is defined as follows:
P ( u | f ( v ) ) = exp ( f ( u ) · f ( v ) ) v s V exp ( f ( v s ) · f ( v ) ) ,
The underlying assumption of this objective function is the vertices that are close in the original network should have similar feature representations in latent space. The SGNS uses a two-layer artificial neural network to train the feature representation of every vertex. Figure 1 shows the workflow of SGNS architecture. Suppose there are m distinct vertices in an input sequence, the SGNS applies an m-dimensional one-hot vector to represent the input sequence. The one-hot vector of the ith vertex in an input sequence is a binary vector, whose elements are all zeros except for the ith element is one. The input vertex and its neighbors are encoded into corresponding one-hot vectors. Then, these vectors are placed at the input layer and the output layer of the artificial neural network, respectively. The stochastic gradient descent (SGD) technique is adopted to train the weights of the hidden layer. Finally, merely multiplying the 1 × m one-hot vector from left with an m × d weight matrix results in the feature representation vector of an input vertex. Since computing v s V exp ( f ( v s ) · f ( v ) ) in Equation (4) is too time-consuming for the large scale network, negative sampling is adopted to address the computational problem [33].

2.3. Classification

The previous steps help us learn feature vectors for genes while maintaining their topological properties in the PPI network. The features of some labeled genes are put into a classifier to build a model. Then, the classifier determines the unlabeled genes to be essential according to their feature vectors. A variety of classifiers are available to finish the prediction task. This work aims to verify that using a network embedding method to learn genes’ features in the PPI network is helpful to predict human essential genes. Hence, one of popular classifiers is selected to predict the human essential genes, including random forest (RF) [40], support vector machine (SVM, RBF kernel) [6], decision tree (DT) [41], logistic regression (LR) [42], extra tree (ET), k-nearest neighbor (KNN) [43] and Naive Bayes (NB) [44].

3. Results

3.1. Datasets

Two kinds of datasets were involved in this work. The one dataset was the human essential gene dataset, which was from the supplementary files of [16]. The human essential genes came from three recent works [13,14,15], which consist of 1516 essential genes and 10,499 non-essential genes.
The other dataset was the human PPI network. We tested our method on two different human PPI networks. One network (namely FIs) was downloaded from the Reactome database [45], involving 12,277 genes and 230,243 interactions. The other network was from the InBio Map database (called InWeb_IM dataset) [46], consisting of 17,428 genes and 625,641 interactions. To compare with the Guo’s method [16], the genes which appeared in both the PPI network and the supplementary files of [16] remained in the training and testing datasets. Consequently, the FIs dataset included 1359 essential genes and 5388 non-essential genes, and the InWeb_IM dataset included 1512 essential genes and 9036 non-essential genes. Table 1 shows the details of the two human PPI networks.

3.2. Evaluation Metrics

The five-fold cross-validation was selected to test the prediction performance of our method. All genes in the benchmark dataset partitioned into five parts at random. Four of the five parts were the training set, and the remaining part was the testing set. According to the benchmark dataset, some popular statistical metrics measured the prediction performance. The metrics included precision, recall, Sp (specificity), NPV (negative predictive value), F-measure, ACC (accuracy), and MCC (Matthews correlation coefficient). Their definitions are as follows:
P r e c i s i o n = T P T P + F P   ,
R e c a l l = T P T P + F N   ,
F m e a s u r e = 2 · p r e c i s i o n · r e c a l l p r e c i s i o n + r e c a l l   ,
S p = T N T N + F P   ,
N P V = T N T N + F N   ,  
M C C = T P × T N F P × F N ( T P + F P ) ( T P + F N ) ( T N + F P ) ( T N + F N )   ,
ACC = TP + TN TP + TN + FP + FN   ,  
where TP (true positive) and FP (false positive) measure the number of predicted essential genes that match or not match with known essential genes. FN denotes the number of known essential genes that are missed by predictions. TN indicates the number of true-negative genes. In addition, the area under the Receiver Operating Characteristic (ROC) Curve (AUC) and the area under precision-recall curve (AP) measure the overall performance of each method with selecting different thresholds to get candidate essential genes.
Due to the unbalance between essential genes and non-essential genes, we took two different strategies in the course of validation. In one strategy, each fold data maintains the same ratio of essential genes to non-essential genes in the original data, i.e., about 1:4 in the FIs dataset and 1:6 in the InWeb_IM dataset. The other strategy keeps the ratio of essential genes to non-essential genes as 1:1 in each fold data.

3.3. Parameter Selection

As mentioned in the Method section, the transition probability T (v,k) mainly depends on the value of π (u,k) (see Equation (1)), since, in this work, we only focus on the unweighted human PPI network. Equation (1) shows that the value of π (u,k) changes with different values of the parameters p and q. If the p and q equal one, the node v has an equal probability of jumping back to its previous node u or walking to the common neighbor of it and its last step node u or guiding to its new neighbors. On the one hand, if the p and q exceed one, the node v more likely walks to the node that is the common neighbor of it and its last step node u. On the other hand, if the value of q is less than one, the node has more likelihood of jumping to the nodes far away from it and if the value of p is less than one, the node is more likely to be guided back to its previous nodes. To investigate the effect of parameters on the prediction performance of our method, for the FIs dataset, we fix the values of w, c, and p to 20, 10, and one respectively and set the values of q to 0.5, 0.8 and the integers ranging from one to 10. Figure 2a shows that, when the p and q equal one, the AUC values and F-measure values by keeping the ratio of essential genes to non-essential genes as 1:1 or their original rate on this dataset are 0.9179, 0.9227, 0.8415, 0.7098, respectively. Under this condition, the performance of our method is lower than when the q is 0.8 or ranges from two to six. It suggests that the bias random walk on the human PPI network can improve the performance of human essential gene predictions. Moreover, the AUC values and F-measure values when q ranging from two to 10 vary within a very narrow band and are higher than that when q is less than one. In this work, the q is three on this dataset because the performance of our method receives the best. Similarly, for the parameter p, we fix the values of w, c, and q to 20, 10 and three, respectively, and set the p to 0.5, 0.8 and the integers ranging from one to 10. We notice from Figure 2b, that the AUC values of our method drop quickly when the p is less than one, which suggests that setting more probability to jump back to the previous node reduces the prediction performance. When the parameter p has the same value three as the parameter q, the AUC values are lower than that when the p is one. Under most conditions, setting the p to one achieves better performance than setting p to other values. Hence, in this work, the parameter p is set to one on this dataset.
The parameter w decides the walk distance in the network. The larger the w value is, the farther the node walks away from the start node and the more global information of the network is explored. To investigate the effect of the parameter w, the values of c, p, and q are fixed to 10, one and three, respectively, and the w ranges from 10 to 45 on the FIs dataset. Figure 2c shows that the AUC and the F-measure values of our method achieve relatively high values when the w is 20 on this dataset. Consequently, in the following tests, w was set to 20 on this dataset.
The parameter c controls the size of the sliding window along the node sequence. It decides which two nodes in the node sequence are regarded to have a close relationship in the PPI network and will be the input and output of the artificial neural network. Setting the c values too small ignores many truly related gene pairs and setting the c values too large introduces many unrelated gene pairs. To investigate the effect of the parameter c, we fix the values of w, p, and q to 20, one and three, respectively, and the range of c from two to 20 on the FIs dataset. Figure 2d shows that the AUC values and F-measure values of our method achieve relatively high points when the c is ten on this dataset. Consequently, the c is set to 10 on the FIs dataset in this work.
For the InWeb_IM dataset, we fixed the values of w, c, and p to 25, four, and one, respectively, and set the q to 0.5, 0.8 and the integers ranging from one to 10. Figure 3a shows that when the values of q equal four, the AUC values and the F-measure values achieve the highest. Hence, the q is four on this dataset. For parameter p, we fixed the values of w, c, and q to 25, four, and four, respectively and set the p to 0.5, 0.8 and the integers ranging from one to 10. Figure 3b shows that setting p to six helps our method achieve the best performance. Similarly, we test the parameters w and c and set the w and c to 25 and four, respectively, on the InWeb_IM dataset.

3.4. Comparison with Existing Methods

To test the effectiveness of our method, we compared it with four other existing approaches, including the Z-curve method [16], the centrality-based method, the DeepWalk [34] method, and the LINE [35] method. The Z-curve method is the latest method to predict human essential genes, which inputs DNA sequence features extracted by a λ-interval Z-curve method into an SVM classifier to make predictions [16]. The centrality-based method makes use of genes’ central indices in the network as input features, including DC, BC, CC, SC, EC, IC, and NC. CytoNCA [47] calculated these centrality indices. The remaining two are the DeepWalk [34] and the LINE [35] that learn the feature representation for every node in the human PPI network by different network embedding methods. All these methods select the SVM classifier to make predictions. Table 2; Table 3 show their prediction performances on two different human network datasets, including the FIs dataset and the InWeb_IM dataset. The ratio of essential genes to non-essential genes in the course of validation shows in the brackets at the first column of the tables. For example, we maintained the ratio of the two kinds of genes as 1:4 on the FIs dataset or 1:6 on the InWeb_IM dataset. The best and comparable results are in bold font.
Table 2 and Table 3 show that our method has the best performance among all the methods compared on the two datasets, no matter maintaining which ratio of essential genes to non-essential genes in the course of validation. When the ratio of the two types of genes was 1:4 on the FIs dataset, the F-measure, MCC, AUC, and AP values of our method were 0.692, 0.641, 0.913, and 0.769, which achieved 31%, 57%, 9.5%, and 47.3% improvements as compared with the Z-curve method and achieved 23%, 38%, 21.2%, and 39.8% improvements as compared with the centrality method. For the comparison with the DeepWalk and LINE, the four indicators showed improvements of 0.6%, 2.6%, 0.9%, 4.8% and 28.4%, 25%, 6.7%, 11%, respectively. When the ratio of the two kinds of genes was 1:6 on the InWeb_IM dataset, the F-measure, MCC, AUC, and AP values of our method were 0.665, 0.641, 0.915 0.762, respectively, which were 40.6%, 65.2%, 8.98%, 67.1% higher than that of the Z -curve method, and were 22.2%, 37.8%, 10.5%, 49.1% higher than that of the centrality method. For the comparison with DeepWalk and LINE, the F-measure, MCC, AUC, AP values of our method showed improvements of 9%, 12.3%, 1.2%, 14.2%, and 141.8%, 93.1%, 9.8%, 39.3%, respectively. When the ratio of the two kinds of genes was kept as 1:1 in the course of cross-validation, the evaluation metrics of most comparing methods, including precision, recall, F-measure, ACC, MCC, AUC, and AP, increase their values to some degree. Compared with DeepWalk that has the highest performance among the other existing methods, the F-measure, MCC, AUC and AP values of our method were 0.847, 0.699, 0.914, 0.902 on the FIs dataset and reached up to 0.857, 0.713, 0.928, 0.921 on the InWeb_IM dataset. Our method still possesses the highest F-measure, MCC, AUC, and AP values on the two datasets.
The Z-curve method used the gene’s sequence information as prediction features. The centrality method combined seven genes’ central indices in the network into feature vectors to make predictions. The success of our method proved that learning the genes’ latent features from the PPI network, which extremely maintains the structure of the network, was able to improve the identification of human essential genes. Compared with DeepWalk and LINE that also extracted genes’ features by network-embedding, our methods showed better performance, which proves that our method takes better network-embedding strategies to predict human essential genes.

3.5. Comparison of Different Classifiers

In this work, we extract the gene’s features in the PPI network by using a network-embedding method. To assess the effectiveness of the features on human essential gene predictions, besides the SVM classifier, the other six frequently used classifiers are selected to predict human essential genes in our method. These classifiers include deep neural network (DNN), decision tree (DT), Naive Bayes (NB), k-nearest neighbor (KNN), logistic regression (LR), random forest (RF), and extra tree (ET). The batch_size, epochs, learning rate, and activation function of the DNN model are 64, 20, 0.0001, and sigmoid, respectively. For the DNN with one hidden layer, the number of neurons is 32. For the DNN with three hidden lays, the numbers of neurons are 32, 16, and one, respectively. Table 4 and Table 5 list the performance comparisons of our method on two different datasets with different classifiers and different ratios of essential genes to non-essential genes in cross-validation. Among all comparing classifiers, the extra tree algorithm (ET) based on random achieved the highest prediction performance. By maintaining the original essential and non-essential gene ratio in the training and testing dataset, the F-measure, MCC, AUC, and AP values of our method by using ET classifier achieved 0.727, 0.676, 0.932, 0.806 on the FIs dataset, and 0.692, 0.659, 0.943, 0.779 on the InWeb_IM dataset. When keeping the same ratio of essential gene to non-essential, the F-measure, MCC, AUC, and AP values of our method using ET classifier are 0.7%, 1.9%, 0.98%, and 2.2% higher than that by using the SVM classifier on the FIs dataset, and are 1.05%, 2.2%, 0.65%, and 0.76% higher than that by using the SVM classifier on the InWeb_IM dataset. The results suggest that choosing other efficient classifiers can further improve the prediction performance of our method.

3.6. Feature Representation of Human Essential Genes

Our method represents every node in the human PPI network as a latent vector and uses the vector as the features of the node to predict essential proteins. To investigate whether the feature representation learned by our method can preserve network structure and the properties of the nodes in the network, we rebuild the network according to the Euclidean distance between the nodes’ feature vectors. Then, we calculate some centrality indexes for nodes in the PPI network, including DC, BC, CC, NC, and IC. Table 6 shows the Pearson's correlation coefficient between these centrality indexes of the human essential genes in the original PPI network and the rebuilt network. We note that on the FIs dataset and InWeb_IM dataset, the Pearson's correlation coefficients between the centrality indexes of the human essential genes in the original and the rebuilt network are above 0.8 and some values even close to one. It suggests the essential genes have very similar topology properties in the two PPI networks and the feature representation learned by our method can preserve network structure.
To probe why the feature representations can improve the performance of the human essential genes’ prediction, we clustered the essential genes into 20 subgroups according to their features by a k-means method. For comparison, we select two different kinds of features for clustering. One type of gene feature is the connection relationship between a gene and the other genes in the PPI network, which is a row of the adjacent matrix of the network. The other type of gene feature is a gene’s feature representation learned by the network embedding method. The clustering results on two human PPI networks were visualized by a t-SNE tool. Figure 4; Figure 5 show that the human essential genes can be separated into several subgroups well, when using the genes’ feature representations. On the FIs dataset, the largest subgroup consists of 289 essential genes, and the smallest one consists of 21 essential genes. On the InWeb_IM dataset, the largest subgroup consists of 176 essential genes, and the smallest one consists of 20 essential genes. However, when the clustering features are selected as the connection relationship between a gene and the other genes in the PPI network, most of the human essential genes are aggregated in a large subgroup while the remaining subgroups have small sizes. On the FIs dataset, the largest subgroup consists of 902 essential genes, and the smallest one consists of six essential genes. On the InWeb_IM dataset, the largest subgroup consists of 1022 essential genes, and the smallest one consists of one essential gene. Moreover, we also leverage silhouette value and Dunn value to evaluate the quality of the clustering results under two different clustering features. Table 7 lists the detailed information about these subgroups. The silhouette and Dunn values of the subgroups generated by the feature representation are higher than that of the subgroups generated by the connection relationship features. It suggests that selecting the feature representation as clustering features can successfully divide human essential genes into the subgroups with the high compactness within clusters and the high separation among clusters.
To further investigate the biological functions of these essential gene subgroups, we used the DAVID on-line database to perform the GO functional enrichment analysis for these subgroups in the GOTERM_BP_FAT category. In this work, we only present the enrichment analysis based on the GO biological process (BP) annotation. Because the module enriches in BP annotations indicates that the genes in the module have diverse molecular functions but work together to perform a particular biological process, such as signaling in a pathway. An essential gene subgroup is regarded as being biological significance if its p-value is less than 0.01. All of the essential gene clusters generated by two different clustering features on the two human PPI networks have significant biological functions (see Supplementary Files). As we can see from Table 7, the essential gene subgroups clustered by the feature representation have higher vital biological functions (with higher Avg(-log(p-value)) values) than those gathered by connection relationship features. Table 8 lists ten example essential gene clusters with the smallest p-values on the FIs dataset. We notice that the feature representation can cluster a group of essential genes with the p-value of 6.64E-191. It consists of 116 essential genes, and 110 of them enrich in the function of RNA splicing (GO:0000377). A cluster includes 50 essential genes, and all of the genes in it have the function of rRNA processing (GO:0006364). However, these 50 essential genes are separated into three clusters by connection relationship features (see Supplementary Files). The bias random walk in the course of learning feature representation for genes can catch the modularity structure of the network. Therefore, these features contribute to the performance improvement of essential gene prediction because essential genes tend to cluster together.

4. Conclusions

Extracting the features related to essential genes is the critical step for designing a machine-learning method with powerful prediction performance. Compared with previous approaches that refer to the topological features in the PPI network or learn features from the DNA sequence, this work adopts a network embedding method to represent the nodes in the PPI network as latent feature vectors that maximally preserve their interactive relationships in the network. To measure the power of the latent features on predicting human essential proteins, we learn features from two different human PPI networks and these features are input into several popular classifiers. In the course of learning, a bias random walk is, first, implemented on the PPI network to get the network context of a given node and, then, the node pairs in the network context are input into an artificial neural network to learn their representation in the latent space. The human essential gene prediction results based on the two PPI networks show that our method using the SVM classifier outperforms the methods [16] that select network topological properties or the DNA sequences as input features. It proved that the network embedding method could learn feature vectors for the nodes from the PPI network to efficiently predict human essential genes. The features represented by our method have better performance on human essential gene prediction than that by other network embedding approaches, i.e., Deepwalk [34] and LINE [35]. It suggests that our approach is an efficient way to learn the feature representations for nodes in the human PPI network by adding bias in exploring neighborhoods. Moreover, using the ET classifier can further improve the prediction performance of our method.
In the future, we will focus on designing a more robust network embedding method to find the latent feature representation of the nodes in the PPI network, and developing more effective machine learning methods to predict human essential genes based on the features of genes. In addition, one human gene can produce many different protein isoforms and there exists noise in the human PPI network. Hence, in future work, establishing reliable gene–gene networks would be a potential solution to improve the essential gene prediction.

Supplementary Materials

The following are available online at https://www.mdpi.com/2073-4425/11/2/153/s1, Table S1: GO enrichment analysis for the essential gene subgroups clustered by the feature representation on the FIs dataset, Table S2: GO enrichment analysis for the essential gene subgroups clustered by the connection relationship features on the FIs dataset, Table S3: GO enrichment analysis for the essential gene subgroups clustered by the feature representation on the InWeb_IM dataset, Table S4: GO enrichment analysis for the essential gene subgroups clustered by the connection relationship features on the InWeb_IM dataset.

Author Contributions

Software, Q.C.; resources, Y.L.; writing—original draft preparation, W.P.; writing—review and editing, W.D.; funding acquisition, J.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported in part by the National Natural Science Foundation of China under grant nos. 61972185, 31560317, 61502214, 61472133, 61502166, and 81560221, the Natural Science Foundation of Yunnan Province of China (2019FA024), the Natural Science Foundation of Hunan Province of China under grant no. 2018JJ2262, the Yunnan Key Research and Development Program (2018IA054), and the Yunnan Ten Thousand Talents Plan young.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Zhang, R.; Lin, Y. DEG 5.0, a database of essential genes in both prokaryotes and eukaryotes. Nucleic Acids Res. 2009, 37, D455–D458. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. E Clatworthy, A.; Pierson, E.; Hung, D.T. Targeting virulence: A new paradigm for antimicrobial therapy. Nat. Methods 2007, 3, 541–548. [Google Scholar] [CrossRef] [PubMed]
  3. Giaever, G.; Chu, A.M.; Ni, L.; Connelly, C.; Riles, L.; Véronneau, S.; Dow, S.; Lucau-Danila, A.; Anderson, K.; André, B.; et al. Functional profiling of the Saccharomyces cerevisiae genome. Nat. 2002, 418, 387–391. [Google Scholar] [CrossRef] [PubMed]
  4. Cullen, L.M.; Arndt, G.M. Genome-wide screening for gene function using RNAi in mammalian cells. Immunol. Cell Boil. 2005, 83, 217–223. [Google Scholar] [CrossRef] [PubMed]
  5. Roemer, T.; Jiang, B.; Davison, J.; Ketela, T.; Veillette, K.; Breton, A.; Tandia, F.; Linteau, A.; Sillaots, S.; Marta, C.; et al. Large-scale essential gene identification in Candida albicans and applications to antifungal drug discovery. Mol. Microbiol. 2010, 50, 167–181. [Google Scholar] [CrossRef] [PubMed]
  6. Chen, Y.; Xu, D. Understanding protein dispensability through machine-learning analysis of high-throughput data. Bioinform 2005, 21, 575–581. [Google Scholar] [CrossRef] [Green Version]
  7. Yuan, Y.; Xu, Y.; Xu, J.; Ball, R.L.; Liang, H. Predicting the lethal phenotype of the knockout mouse by integrating comprehensive genomic data. Bioinform. 2012, 28, 1246–1252. [Google Scholar] [CrossRef]
  8. Lloyd, J.P.; Seddon, A.E.; Moghe, G.D.; Simenc, M.C.; Shiu, S.-H. Characteristics of plant essential genes allow for within- and between-species prediction of lethal mutant phenotypes. Plant Cell 2015, 27, 2133–2147. [Google Scholar] [CrossRef]
  9. Wang, J.; Peng, W.; Wu, F.-X. Computational approaches to predicting essential proteins: A survey. Proteom. Clin. Appl. 2013, 7, 181–192. [Google Scholar] [CrossRef]
  10. Furney, S.J.; Albà, M.M.; López-Bigas, N. Differences in the evolutionary history of disease genes affected by dominant or recessive mutations. BMC Genom. 2006, 7, 165. [Google Scholar]
  11. Song, J.; Peng, W.; Wang, F.; Zhang, X.; Tao, L.; Yan, F.; Sung, D.K. An entropy-based method for identifying mutual exclusive driver genes in cancer. IEEE/ACM Trans. Comput. Boil. Bioinform. 2019, 1. [Google Scholar] [CrossRef] [PubMed]
  12. Song, J.; Peng, W.; Wang, F. A random walk-based method to identify driver genes by integrating the subcellular localization and variation frequency into bipartite graph. BMC Bioinform. 2019, 20, 238. [Google Scholar] [CrossRef] [PubMed]
  13. Fraser, A. Essential human genes. Cell Syst. 2015, 1, 381–382. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Hart, T.; Chandrashekhar, M.; Aregger, M.; Steinhart, Z.; Brown, K.R.; MacLeod, G.; Mis, M.; Zimmermann, M.; Fradet-Turcotte, A.; Sun, S.; et al. High-resolution CRISPR screens reveal fitness genes and genotype-specific cancer liabilities. Cell 2015, 163, 1515–1526. [Google Scholar] [CrossRef] [Green Version]
  15. Wang, T.; Birsoy, K.; Hughes, N.W.; Krupczak, K.M.; Post, Y.; Wei, J.J.; Lander, E.S.; Sabatini, D.M. Identification and characterization of essential genes in the human genome. Sci. 2015, 350, 1096–1101. [Google Scholar] [CrossRef] [Green Version]
  16. Guo, F.-B.; Dong, C.; Hua, H.-L.; Liu, S.; Luo, H.; Zhang, H.-W.; Jin, Y.-T.; Zhang, K.-Y. Accurate prediction of human essential genes using only nucleotide composition and association information. Bioinform. 2017, 33, 1758–1764. [Google Scholar] [CrossRef] [Green Version]
  17. Jeong, H.; Mason, S.P.; Barabási, A.-L.; Oltvai, Z.N. Lethality and centrality in protein networks. Nat. 2001, 411, 41–42. [Google Scholar] [CrossRef] [Green Version]
  18. Vallabhajosyula, R.R.; Chakravarti, D.; Lutfeali, S.; Ray, A.; Raval, A. Identifying hubs in protein interaction networks. PloS One 2009, 4, 5344. [Google Scholar] [CrossRef] [Green Version]
  19. Wuchty, S.; Stadler, P.F. Centers of complex networks. J. Theor. Boil. 2003, 223, 45–53. [Google Scholar] [CrossRef] [Green Version]
  20. Joy, M.P.; Brock, A.; Ingber, N.E.; Huang, S. High-betweenness proteins in the yeast protein interaction network. J. Biomed. Biotechnol. 2005, 2005, 96–103. [Google Scholar] [CrossRef]
  21. Stephenson, K.; Zelen, M. Rethinking centrality: Methods and examples. Soc. Networks 1989, 11, 1–37. [Google Scholar] [CrossRef]
  22. Bonacich, P. Power and centrality: A family of measures. Am. J. Sociol. 1987, 92, 1170–1182. [Google Scholar] [CrossRef]
  23. Estrada, E.; Rodríguez-Velázquez, J.A. Subgraph centrality in complex networks. Phys. Rev. E 2005, 71, 056103. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Wang, J.; Li, M.; Wang, H.; Pan, Y. Identification of essential proteins based on edge clustering coefficient. IEEE/ACM Trans. Comput. Biol. Bioinform. 2012, 9, 1070–1080. [Google Scholar] [CrossRef]
  25. Li, M.; Wang, J.; Wang, H.; Pan, Y. Essential proteins discovery from weighted protein interaction networks. Lect. Notes Comput Sc. 2010, 6053, 89–100. [Google Scholar]
  26. Li, M.; Zhang, H.; Wang, J.-X.; Pan, Y. A new essential protein discovery method based on the integration of protein-protein interaction and gene expression data. BMC Syst. Boil. 2012, 6, 15. [Google Scholar] [CrossRef] [Green Version]
  27. Tang, X.; Wang, J.; Zhong, J.; Pan, Y. Predicting essential proteins based on weighted degree centrality. IEEE/ACM Trans. Comput. Biol. Bioinf. 2014, 11, 407–418. [Google Scholar] [CrossRef]
  28. Zhang, F.; Peng, W.; Yang, Y.; Dai, W.; Song, J. A novel method for identifying essential genes by fusing dynamic protein–protein interactive networks. Genes 2019, 10, 31. [Google Scholar] [CrossRef] [Green Version]
  29. Peng, W.; Wang, J.; Cheng, Y.; Lu, Y.; Wu, F.; Pan, Y. UDoNC: An algorithm for identifying essential proteins based on protein domains and protein-protein interaction networks. IEEE/ACM Trans. Comput. Biol. Bioinf. 2015, 12, 276–288. [Google Scholar] [CrossRef]
  30. Peng, W.; Wang, J.; Wang, W.; Liu, Q.; Wu, F.-X.; Pan, Y. Iteration method for predicting essential proteins based on orthology and protein-protein interaction networks. BMC Syst. Boil. 2012, 6, 87. [Google Scholar] [CrossRef] [Green Version]
  31. Zhong, J.; Sun, Y.; Peng, W.; Xie, M.; Yang, J.; Tang, X. XGBFEMF: An XGBoost-based framework for essential protein prediction. IEEE Trans. NanoBioscience 2018, 17, 243–250. [Google Scholar] [CrossRef] [PubMed]
  32. Peng, W.; Li, M.; Chen, L.; Wang, L. Predicting protein functions by using unbalanced random walk algorithm on three biological networks. IEEE/ACM Trans. Comput. Biol. Bioinf. 2017, 14, 360. [Google Scholar] [CrossRef] [PubMed]
  33. Mikolov, T.; Sutskever, I.; Chen, K.; Corrado, G.; Dean, J. Distributed Representations of Words and Phrases and their Compositionality. In Proceedings of the International Conference on Neural Information Processing Systems, Lake Tahoe, NV, USA, 5–8 December 2013. [Google Scholar]
  34. Perozzi, B.; Al-Rfou, R.; Skiena, S. DeepWalk: Online Learning of Social Representations. In Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, New York, NY, USA, 24–27 August 2014; ACM Digital Library: New York, NY, USA, 2014; pp. 701–710. [Google Scholar]
  35. Tang, J.; Qu, M.; Wang, M.; Zhang, M.; Yan, J.; Mei, Q. LINE: Large-scale information network embedding. In Proceedings of the 24th International Conference on World Wide Web, WWW 2015, Florence, Italy, 18–22 May 2015. [Google Scholar]
  36. Wang, Y.; Yao, Y.; Tong, H.; Xu, F.; Lu, J. A brief review of network embedding. Big Data Min. Anal. 2019, 2, 35–47. [Google Scholar] [CrossRef] [Green Version]
  37. Ye, Z.; Zhao, H.; Zhang, K.; Wang, Z.; Zhu, Y. Network representation based on the joint learning of three feature views. Big Data Min. Anal. 2019, 2, 248–260. [Google Scholar] [CrossRef]
  38. Grover, A.; Leskovec, J. node2vec: Scalable feature learning for networks. KDD 2016, 855–864. [Google Scholar]
  39. Dai, W.; Chang, Q.; Peng, W.; Zhong, J.; Li, Y. Identifying human essential genes by network embedding protein-protein interaction network. Lect. Notes Comput Sc. 2019, 11490, 127–137. [Google Scholar]
  40. Wu, J.; Zhang, Q.; Wu, W.; Pang, T.; Hu, H.; Chan, W.K.B.; Ke, X.; Zhang, Y. WDL-RF: Predicting bioactivities of ligand molecules acting with G protein-coupled receptors by combining weighted deep learning and random forest. Bioinform. 2018, 34, 2271–2282. [Google Scholar] [CrossRef]
  41. Acencio, M.L.; Lemke, N. Towards the prediction of essential genes by integration of network topology, cellular localization and biological process information. BMC Bioinform. 2009, 10, 290. [Google Scholar] [CrossRef] [Green Version]
  42. Liao, J.; Chin, K.-V. Logistic regression for disease classification using microarray data: Model selection in a large p and small n case. Bioinform. 2007, 23, 1945–1951. [Google Scholar] [CrossRef]
  43. Chou, K.C.; Shen, H.B. Predicting eukaryotic protein subcellular location by fusing optimized evidence-theoretic K-Nearest Neighbor classifiers. J. Proteome. Res. 2006, 5, 1888–1897. [Google Scholar] [CrossRef]
  44. Cheng, J.; Xu, Z.; Wu, W.; Zhao, L.; Li, X.; Liu, Y.; Tao, S. Training set selection for the prediction of essential genes. PloS One 2014, 9, e86805. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  45. Wu, G.; Feng, X.; Stein, L. A human functional protein interaction network and its application to cancer data analysis. Genome Boil. 2010, 11, R53. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  46. Li, T.; Wernersson, R.; Hansen, R.B.; Horn, H.; Mercer, J.; Slodkowicz, G.; Workman, C.T.; Rigina, O.; Rapacki, K.; Stærfeldt, H.H.; et al. A scored human protein–protein interaction network to catalyze genomic interpretation. Nat. Methods 2016, 14, 61–64. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  47. Tang, Y.; Li, M.; Wang, J.; Pan, Y.; Wu, F.-X. CytoNCA: A cytoscape plugin for centrality analysis and evaluation of protein interaction networks. Biosyst. 2015, 127, 67–72. [Google Scholar] [CrossRef] [PubMed]
Figure 1. An overview of identifying human essential genes by the network embedding protein–protein interaction (PPI) network.
Figure 1. An overview of identifying human essential genes by the network embedding protein–protein interaction (PPI) network.
Genes 11 00153 g001
Figure 2. Parameter adjustments on FIs dataset. Different colors indicate different data ratios. (a) Fixing the p, w, and c and adjusting q; (b) fixing the q, w, and c and adjusting p; (c) fixing the p, q, and c and adjusting w; and (d) fixing the p, q, and wand adjusting c.
Figure 2. Parameter adjustments on FIs dataset. Different colors indicate different data ratios. (a) Fixing the p, w, and c and adjusting q; (b) fixing the q, w, and c and adjusting p; (c) fixing the p, q, and c and adjusting w; and (d) fixing the p, q, and wand adjusting c.
Genes 11 00153 g002
Figure 3. Parameter adjustments on InWeb_IM dataset. Different colors indicate different data ratios. (a) Fixing the p, w ,and c and adjusting q; (b) fixing the q, w, and c and adjusting p; (c) fixing the p, q, and c and adjusting w; and (d) fixing the p, q, and w and adjusting c.
Figure 3. Parameter adjustments on InWeb_IM dataset. Different colors indicate different data ratios. (a) Fixing the p, w ,and c and adjusting q; (b) fixing the q, w, and c and adjusting p; (c) fixing the p, q, and c and adjusting w; and (d) fixing the p, q, and w and adjusting c.
Genes 11 00153 g003
Figure 4. The k-means clustering essential genes by two different features on the FIs dataset. (a) By connection relationship features; (b) by feature representation.
Figure 4. The k-means clustering essential genes by two different features on the FIs dataset. (a) By connection relationship features; (b) by feature representation.
Genes 11 00153 g004
Figure 5. The k-means clustering essential genes by two different features on the InWeb_IM dataset. (a) By connection relationship features; (b) by feature representation.
Figure 5. The k-means clustering essential genes by two different features on the InWeb_IM dataset. (a) By connection relationship features; (b) by feature representation.
Genes 11 00153 g005
Table 1. Details of two human PPI networks.
Table 1. Details of two human PPI networks.
Data setGenesInteractionsEssential genesTraining and testing genes
FIs1227723024313596747
InWeb_IM17428625641151210548
Table 2. Performance comparison between our method and existing methods with different essential to non-essential gene ratios on FIs dataset.
Table 2. Performance comparison between our method and existing methods with different essential to non-essential gene ratios on FIs dataset.
MethodsPrecisionRecallSPNPVF-measureMCCACCAUCAP
DeepWalk (1:4)0.7710.6210.9540.9090.6880.6250.8870.9050.734
LINE (1:4)0.8370.3940.9810.8660.5390.5130.8630.8560.693
Centrality (1:4)0.6080.5230.9170.8790.5620.4640.8360.7530.550
Z-curve (1:4)0.5260.5300.8800.8820.5280.4090.8100.8340.522
Our method (1:4)0.8220.5970.9680.9050.6920.6410.8930.9130.769
DeepWalk (1:1)0.8330.8460.8300.8430.8390.6760.8380.9070.896
LINE (1:1)0.7050.8470.6470.8080.7700.5030.7470.8390.848
Centrality (1:1)0.8520.5530.9040.6690.6710.4880.7280.7600.797
Z-curve (1:1)0.7330.8000.7080.7800.7650.5110.7540.8240.783
Our method (1:1)0.8590.8360.8630.8400.8470.6990.8490.9140.902
Table 3. Performance comparison between our method and existing methods with different essential to non-essential gene ratios on InWeb_IM dataset.
Table 3. Performance comparison between our method and existing methods with different essential to non-essential gene ratios on InWeb_IM dataset.
MethodsPrecisionRecallSPNPVF-measureMCCACCAUCAP
DeepWalk (1:6)0.7490.5150.9710.9230.6100.5710.9060.9040.667
LINE (1:6)0.8190.1650.9940.8770.2750.3320.8750.8320.547
Centrality (1:6)0.5270.5620.9160.9260.5440.4650.8650.8280.511
Z-curve (1:6)0.4850.4620.9180.9110.4730.3880.8530.8400.456
Our method (1:6)0.8410.5500.9830.9290.6650.6410.9210.9150.762
DeepWalk (1:1)0.8170.8480.8110.8420.8320.6590.8290.9080.894
LINE (1:1)0.6410.9050.4920.8390.7500.4360.6990.8040.792
Centrality (1:1)0.8160.7130.8390.7450.7610.5570.7760.8510.841
Z-curve (1:1)0.7300.7770.7130.7620.7530.4910.7450.8270.801
Our method (1:1)0.8550.8580.8540.8580.8570.7130.8560.9280.921
Table 4. Performance comparison of our method with different classifiers and different essential and non-essential gene rates on FIs dataset.
Table 4. Performance comparison of our method with different classifiers and different essential and non-essential gene rates on FIs dataset.
MethodsPrecisionRecallSPNPVF-measureMCCAccuracyAUCAP
DNN (1 layer, 1:4)0.2580.7700.2200.8050.2340.0260.6670.4600.242
DNN (3 layers, 1:4)0.2480.4070.6900.8220.3080.0830.6330.5340.245
DT (1:4)0.6230.6190.9060.9040.6210.5260.8480.7630.660
NB (1:4)0.5530.7370.8500.9280.6320.5310.8270.8800.683
KNN (1:4)0.7950.6280.9590.9110.7020.6440.8930.8890.782
LR (1:4)0.7870.5850.9600.9020.6710.6130.8850.9140.755
SVM (1:4)0.8220.5970.9680.9050.6920.6410.8930.9130.769
RF (1:4)0.8260.6460.9660.9160.7260.6740.9020.9270.799
ET (1:4)0.8280.6480.9660.9160.7270.6760.9020.9320.806
DNN (1 layer, 1:1)0.5540.4520.5030.5040.5270.0070.5030.5000.519
DNN (3 layers, 1:1)0.5350.5380.5320.5360.5360.0700.5350.5620.553
DT (1:1)0.7680.7880.7620.7830.7780.5510.7750.7750.831
NB (1:1)0.8220.7650.8340.7800.7920.6000.7990.8760.873
KNN (1:1)0.8370.8050.8440.8120.8210.6490.8240.8950.906
LR (1:1)0.8390.8270.8410.8290.8330.6680.8340.9100.907
SVM (1:1)0.8590.8360.8630.8400.8470.6990.8490.9140.902
RF (1:1)0.8590.8440.8610.8460.8510.7050.8520.9210.921
ET (1:1)0.8670.8400.8720.8450.8530.7120.8560.9230.922
Table 5. Performance comparison of our method with different classifiers and different essential and non-essential gene rates on InWeb_IM dataset.
Table 5. Performance comparison of our method with different classifiers and different essential and non-essential gene rates on InWeb_IM dataset.
MethodsPrecisionRecallSPNPVF-measureMCCAccuracyAUCAP
DNN (1 layer, 1:6)0.3550.7720.2060.8770.2610.1030.7120.4990.206
DNN (3 layers, 1:6)0.3500.2550.9210.8810.2950.2020.8260.4830.221
DT (1:6)0.5840.5880.9300.9310.5860.5170.8810.7590.616
NB (1:6)0.4560.6970.8610.9450.5510.4730.8370.8770.615
KNN (1:6)0.7780.5640.9730.9300.6540.6170.9140.8880.740
LR (1:6)0.7850.5910.9730.9340.6750.6370.9180.9310.749
SVM (1:6)0.8410.5500.9830.9290.6650.6410.9210.9150.762
RF (1:6)0.7990.6150.9740.9380.6950.6590.9230.9400.776
ET (1:6)0.8160.6000.9770.9360.6920.6590.9250.9430.779
DNN (1 layer, 1:1)0.6520.5040.5680.5910.6070.1570.5780.6030.640
DNN (3 layers, 1:1)0.7370.4970.8230.6200.5930.3380.6590.6370.692
DT (1:1)0.8020.7910.8050.7940.7970.5960.7980.7980.849
NB (1:1)0.8360.7080.8620.7470.7670.5760.7850.8740.872
KNN (1:1)0.8530.8430.8540.8450.8480.6970.8490.9040.906
LR (1:1)0.8650.8340.8700.8400.8490.7040.8520.9250.920
SVM (1:1)0.8550.8580.8540.8580.8570.7130.8560.9280.921
RF (1:1)0.8440.8860.8360.8800.8640.7230.8610.9320.920
ET (1:1)0.8530.8790.8490.8760.8660.7290.8640.9340.928
Table 6. The Pearson's correlation coefficient between the centrality indexes of the human essential genes in the original PPI network and the rebuilt network.
Table 6. The Pearson's correlation coefficient between the centrality indexes of the human essential genes in the original PPI network and the rebuilt network.
Data setDCBCCCNCIC
FIs0.92620.80400.99980.99110.9794
InWeb_IM0.96170.83720.99990.99380.9839
Table 7. Details of the subgroups aggregated by two different features.
Table 7. Details of the subgroups aggregated by two different features.
Data SetFeaturesMax SizeMin SizeMedian SizeSilhouetteDunnAvg(-log(p-value))
FIsFeature representation2892151.50.32420.5963.50
Connection relationship902619.50.24480.5845.01
InWeb_IMFeature representation1762061.50.22790.6354.95
Connection relationship10221130.16190.2530.11
Table 8. Ten example essential gene clusters with the smallest p-values gathered by feature representation on FIs dataset.
Table 8. Ten example essential gene clusters with the smallest p-values gathered by feature representation on FIs dataset.
GO IDDescriptionp-valueGenes in ClusterGene Ratio
GO:0000377RNA splicing, via transesterification reactions with bulged adenosine as nucleophile6.64·10−191PLRG1/PABPN1/SNRNP35/HNRNPM/RBMX/RBM22/DHX9/MAGOH/XAB2/SRRM2/SNRPF/SMNDC1/SRRM1/SF3B5/PPIE/CTNNBL1/PRPF40A/SNRNP70/PRPF4B/EIF4A3/PRPF19/BUD31/HNRNPL/NCBP1/DHX8/SNRPC/CWC22/CPSF1/RNPC3/HNRNPC/TFIP11/SNU13/CLP1/SNRNP200/ISY1/TXNL4A/CPSF2/USP39/SNRNP27/ALYREF/DDX23/CPSF3/PCBP1/DDX39B/PCF11/SYMPK/SF3B1/BCAS2/SRSF1/SF3A2/WDR33/SRSF11/DHX38/SNRNP25/SNRPE/HNRNPH1/ZMAT5/RBM17/SNRNP48/CHERP/PUF60/FIP1L1/HNRNPK/HNRNPA2B1/CSTF3/LSM7/CDC40/SNRPD3/PRPF3/PRPF31/NUDT21/SART3/AQR/CRNKL1/U2AF2/PPWD1/PRPF6/SRSF2/LSM4/SRRT/DHX16/SART1/CSTF1/SF1/SRSF7/SNRPB/DHX15/EFTUD2/NCBP2/PRPF4/SF3B2/SLU7/CDC5L/SF3A3/LSM2/GPKOW/SUGP1/HNRNPU/SF3B4/CPSF4/PRPF8/SRSF3/SNRPD2/HSPA8/SNW1/U2SURP/DDX46/SF3A1/SF3B3/PDCD7110/116
GO:0070125mitochondrial translational elongation6.97·10−158MRPL28/MRPL22/MRPL17/MRPL48/MRPL37/MRPL4/MRPL47/TUFM/MRPL23/MRPL9/MRPL24/MRPS12/MRPL11/MRPL10/MRPL38/ERAL1/MRPL46/MRPS6/MRPS25/MRPL41/MRPS27/MRPL12/MRPS16/MRPS23/MRPL34/MRPS34/MRPL43/MRPL15/MRPS24/MRPL35/MRPL40/MRPL57/MRPL21/GFM1/MRPS7/PTCD3/MRPS22/MRPL13/MRPL51/MRPL53/MRPS2/MRPL14/MRPS5/MRPL45/MRPL18/DAP3/AURKAIP1/MRPL19/MRPS15/MRPL20/MRPL39/MRPL44/GADD45GIP1/MRPS18A/MRPS14/MRPS11/MRPS31/MRPS18C/MRPS18B/MRPS30/MRPL16/MRPS35/MRPS10/MRPL3364/67
GO:0000184nuclear-transcribed mRNA catabolic process, nonsense-mediated decay7.98·10−107UPF1/RPS16/RPSA/SMG5/RPS14/RPL10A/RPL8/RPS11/SMG7/RPL13A/RPLP1/ETF1/RPL27A/EIF4G1/GSPT1/RPLP2/RPL4/RPS13/RPL18/RPS4X/RPL36/RPS10/RPL23A/RPS12/RPS5/RPL9/SMG6/RPS18/RPS21/RPS7/RPL29/RPL31/RPL12/RPL10/RPL7/PABPC1/RPL3/RPS15A/RPL37A/RPL18A/RPL19/RPS25/SMG1/RPL11/RPL7A/RPS15/RPS9/RPS2/UPF2/RPL27/RPL13/RPL14/RPL15/RPS29/RPL38/RPLP0/RPS8/RPS658/96
GO:0000819sister chromatid segregation1.97·10−94PMF1/NSL1/BUB3/ESPL1/NUP43/BIRC5/SMC2/XPO1/CENPC/NUF2/KIF18A/SEH1L/SKA1/NUP62/SEC13/NCAPD3/NCAPD2/TTK/MAU2/SMC5/NSMCE2/CENPA/CDCA8/RAD21/AHCTF1/CENPM/RANGAP1/SPC24/PLK1/SMC1A/AURKB/CENPK/INCENP/RANBP2/NUP85/SPC25/CKAP5/NUP107/CENPE/SMC3/NUP98/CENPL/PAFAH1B1/MAD2L1/CENPN/SPDL1/NDC80/CDC20/DSN1/BUB1B/NUP133/TPR/KPNB1/CENPI/SMC4/NUDC/SGO1/CCNB1/RAN/NUP160/CDCA561/91
GO:0006364rRNA processing5.08·10−94WDR75/BMS1/EXOSC9/UTP6/NOP56/HEATR1/EXOSC1/UTP20/UTP4/UTP14A/EXOSC8/RRP36/NOL6/MPHOSPH10/DDX47/RRP9/TBL3/PWP2/EXOSC7/WDR3/WDR18/IMP4/EXOSC5/EXOSC4/WDR46/UTP18/DIEXF/UTP11/LAS1L/UTP3/UTP15/NOC4L/FBL/IMP3/NOP14/PDCD11/TEX10/NOP58/RCL1/XRN2/KRR1/DDX49/EXOSC3/EXOSC10/WDR43/DDX52/EXOSC6/NOL11/EXOSC2/WDR3650/50
GO:0031145anaphase-promoting complex-dependent catabolic process5.71·10−89ANAPC11/PSMD7/PSMD14/PSMA7/PSMB3/PSMB6/PSMA3/PSMD12/PSMA4/PSMB5/PSMA2/ANAPC2/PSMB1/ANAPC4/PSMD6/ANAPC5/PSMA5/PSMD2/PSMD4/PSMC6/PSMC1/CDC16/ANAPC15/PSMD1/PSMD11/PSMB4/PSMD3/PSMD8/CUL3/ANAPC10/PSMC4/PSMA1/PSMC2/PSMC3/PSMB7/AURKA/PSMC5/PSMD13/PSMB2/CDC2340/51
GO:0098781ncRNA transcription2.19·10−81GTF2E1/PHAX/INTS5/TAF13/TAF6/POLR2E/ZC3H8/ELL/POLR2H/TAF8/INTS7/POLR2F/BRF1/GTF2A1/POLR2B/SNAPC2/POLR2I/GTF3C2/INTS9/TAF5/INTS8/GTF2A2/INTS2/POLR2G/POLR2C/GTF3C4/INTS3/SNAPC4/GTF3C3/GTF2B/POLR2D/INTS6/SNAPC3/INTS1/CCNK/CDK9/ICE1/POLR2L/RPAP2/GTF3C5/INTS4/CDK7/GTF2E2/GTF3C144/80
GO:0006270DNA replication initiation2.12·10−66CDC7/MCM2/POLA1/MCM4/POLE2/ORC5/ORC2/CDC45/MCM3/ORC4/MCM7/ORC3/MCM10/CDC6/MCM6/ORC6/ORC1/MCM5/PRIM1/POLA2/GINS2/CDT1/CDK2/GINS4/POLE25/28
GO:0048193Golgi vesicle transport1.46·10−61GBF1/TRAPPC1/VPS52/COG4/KIF11/COPB1/DCTN5/VPS54/YKT6/DCTN6/STX18/COPA/COPB2/NSF/COG8/PREB/TMED2/TMED10/TRAPPC4/DCTN4/KIF23/COPG1/TRAPPC5/ARFRP1/ARCN1/NAPA/COPE/TRAPPC8/RINT1/SCFD1/COPZ1/STX5/TRAPPC11/SYS1/NBAS/COG3/SEC16A/TRAPPC3/RACGAP1/GOSR240/ 47
GO:0042254ribosome biogenesis5.70·10−53WDR12/FTSJ3/NOP2/DDX56/MAK16/BRIX1/RRP1/NAT10/ABCE1/GTPBP4/PPAN/MDN1/PES1/EBNA1BP2/DIS3/NOP16/POP4/DDX27/EFL1/SDAD1/NSA2/HEATR3/RPF1/RSL1D1/RRS1/TSR2/TRMT112/LSG1/NHP2/NOP10/NOL9/DKC1/NIP7/GNL2/ISG20L2/RPL7L1/SURF6/DDX51/RRP15/PELP1/NGDN/NOC2L/WDR7443/ 81

Share and Cite

MDPI and ACS Style

Dai, W.; Chang, Q.; Peng, W.; Zhong, J.; Li, Y. Network Embedding the Protein–Protein Interaction Network for Human Essential Genes Identification. Genes 2020, 11, 153. https://doi.org/10.3390/genes11020153

AMA Style

Dai W, Chang Q, Peng W, Zhong J, Li Y. Network Embedding the Protein–Protein Interaction Network for Human Essential Genes Identification. Genes. 2020; 11(2):153. https://doi.org/10.3390/genes11020153

Chicago/Turabian Style

Dai, Wei, Qi Chang, Wei Peng, Jiancheng Zhong, and Yongjiang Li. 2020. "Network Embedding the Protein–Protein Interaction Network for Human Essential Genes Identification" Genes 11, no. 2: 153. https://doi.org/10.3390/genes11020153

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop