Next Article in Journal
Combining OSMAC Approach and Untargeted Metabolomics for the Identification of New Glycolipids with Potent Antiviral Activity Produced by a Marine Rhodococcus
Previous Article in Journal
Immunomodulatory Arming Factors—The Current Paradigm for Oncolytic Vectors Relies on Immune Stimulating Molecules
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SecProCT: In Silico Prediction of Human Secretory Proteins Based on Capsule Network and Transformer

Key Laboratory of Symbol Computation and Knowledge Engineering of the Ministry of Education, College of Computer Science and Technology, Jilin University, Changchun 130012, China
*
Author to whom correspondence should be addressed.
Int. J. Mol. Sci. 2021, 22(16), 9054; https://doi.org/10.3390/ijms22169054
Submission received: 13 July 2021 / Revised: 12 August 2021 / Accepted: 20 August 2021 / Published: 22 August 2021
(This article belongs to the Section Molecular Informatics)

Abstract

:
Identifying secretory proteins from blood, saliva or other body fluids has become an effective method of diagnosing diseases. Existing secretory protein prediction methods are mainly based on conventional machine learning algorithms and are highly dependent on the feature set from the protein. In this article, we propose a deep learning model based on the capsule network and transformer architecture, SecProCT, to predict secretory proteins using only amino acid sequences. The proposed model was validated using cross-validation and achieved 0.921 and 0.892 accuracy for predicting blood-secretory proteins and saliva-secretory proteins, respectively. Meanwhile, the proposed model was validated on an independent test set and achieved 0.917 and 0.905 accuracy for predicting blood-secretory proteins and saliva-secretory proteins, respectively, which are better than conventional machine learning methods and other deep learning methods for biological sequence analysis. The main contributions of this article are as follows: (1) a deep learning model based on a capsule network and transformer architecture is proposed for predicting secretory proteins. The results of this model are better than the those of existing conventional machine learning methods and deep learning methods for biological sequence analysis; (2) only amino acid sequences are used in the proposed model, which overcomes the high dependence of existing methods on the annotated protein features; (3) the proposed model can accurately predict most experimentally verified secretory proteins and cancer protein biomarkers in blood and saliva.

1. Introduction

Human secretory proteins can enter the blood, saliva or other body fluids through various complex secretory pathways and can be used as protein markers for the detection of blood, saliva or other body fluids [1]. The complex blood circulation system of the human body has many biomarkers that can indicate physiological conditions and disease conditions. Therefore, most current studies on biomarkers in body fluids use blood as the main research object [2]. Similar to other human body fluids, saliva is rich in biomolecules secreted from salivary glands or leaked from nearby tissues [3]. In addition, biomolecules can also be released into the blood circulatory system through various organs far away from the salivary glands into the human body and then be secreted into saliva [4]. Therefore, the biomolecules in saliva can reflect the health of specific organs to a certain extent, including organs both near and far away from the salivary glands.
At present, there have been many studies conducted to predict which proteins are located on the cell surface or secreted into the extracellular environment [5,6,7]. The methods proposed in these studies are mostly based on annotation information such as the amino acid compositions of protein, protein domains and protein functions. These existing methods play a very important role in the diagnosis of early cancer and other diseases, and to a certain extent, they solve the problem of predicting human secretory proteins. However, most of these methods are based on conventional machine learning algorithms. The prediction process in most machine learning algorithms can be described in two steps: first select the features from the constructed feature set, and then train the machine learning prediction algorithm using the divided training set [8]. However, incomplete features or selection bias are very common in the feature selection process, which may have a negative impact on the final prediction result [9].
The existing studies are shown in Table 1, which are mostly based on annotation information such as the amino acid compositions of protein, protein domains and protein functions. Cui et al. proposed a prediction model based on machine learning algorithms using data mining techniques to predict which human proteins can be secreted into the blood [10]. Liu et al. proposed a method for predicting blood-secretory proteins based on manifold ordering [11]. Wang et al. presented a novel computational method for predicting salivary proteins that come from the circulation based on a set of physiochemical and sequence features [12]. Sun et al. proposed a computational method for predicting secretory proteins in human saliva [8]. Zhang et al. proposed a predictor based on sequence features for the high-throughput and accurate identification of mammalian secretory proteins [13]. Recently, Zhang et al. also proposed a prediction method for blood-secretory proteins based on the optimisation of the discrete firefly algorithm [14].
In this article, we propose an end-to-end prediction model based on a deep learning framework, which is mainly comprised of a capsule network and transformer architecture, to predict secretory proteins using only amino acid sequences. Our model can accurately identify human blood and salivary secretory proteins only based on sequence information. How to effectively represent the protein data for the input of the neural network is another key issue that urgently needs to be solved. The data input representation of the network directly determines the effectiveness of a series of settings, such as the network structure, loss function, and hyperparameters, and determines the upper limit of the final predictive performance of the model. Therefore, in the pre-treatment process, our main task is to design effective protein sequence encoding. A simple and widely used protein sequence encoding method is one-of-K encoding. The data for the protein encoding consist of a matrix of N rows and K columns, where K is usually the number of amino acid types and N is the length of the input protein sequence [9]. In the matrix, each column corresponds to a type of amino acid, and each row represents the position in the protein sequence. The K-dimensional vector represented by the row has a value of 1 at the index where the corresponding amino acid appears and has a value of 0 at the other positions. However, one-of-K encoding does not consider the evolutionary relationships among different proteins. Therefore, in this article, we used PSI-BLAST [15] to perform sequence alignment in the “Uniref50” database [16] to obtain evolutionary profile information for protein sequence encoding. To solve the problem of unbalanced data, the bagging ensemble learning method is applied during the training process. Then, the PSSM of the training data is input into the proposed model to train the model parameters.
The proposed model achieves high accuracies using cross-validation (0.921 and 0.892 for predicting blood-secretory proteins and saliva-secretory proteins, respectively) and on an independent test set (0.917 and 0.905 for predicting blood-secretory proteins and saliva-secretory proteins, respectively), thus outperforming existing methods based on conventional machine learning algorithms. By comparing the results of the proposed model with experimentally detected blood-secretory proteins and saliva-secretory proteins, our model can achieve true positive rates of 0.909 and 0.935, respectively. By comparing the results of the proposed model with known cancer biomarkers in blood and saliva, our model can achieve true positive rates of 0.884 and 0.946, respectively. A web server for predicting secretory proteins was developed, which can be accessed via the following website: http://www.csbg-jlu.info/SecProCT/, accessed on 21 August 2021. We believe that the proposed model and web server are useful for biomedical researchers who are interested in finding protein biomarkers in blood and saliva, especially when they have candidate proteins obtained from transcriptome or proteome data.
The main contributions of this article are as follows: (1) a deep learning model based on a capsule network and transformer architecture is proposed for predicting secretory proteins. The results of the model are better than the existing conventional machine learning methods and deep learning methods for biological sequence analysis; (2) only amino acid sequences are used in the proposed model, which overcomes the high dependence of existing methods on the annotated protein features; (3) the proposed model can accurately predict most experimentally verified secretory proteins and cancer protein biomarkers in blood and saliva.
The rest of this paper is organised as follows: the Related Work Section surveys the related work for predicting secretory protein. The Materials and Methods Section presents the datasets and technical details of the proposed method. The Results Section gives the performance evaluation. Finally, the Conclusions Section concludes this paper and pinpoints future works.

2. Results

2.1. Evaluating the Performance of the Binary Classification

To evaluate the performance of the proposed model on the test set, we used bagging ensemble learning for 10 iterations to train our proposed model. For blood-secretory proteins, 380 blood-secretory proteins and 380 non-blood-secretory proteins were used to train the model in each iteration, and the model training was evaluated on an independent test set consisting of 106 blood-secretory proteins and 100 non-blood-secretory proteins. For saliva-secretory proteins, 350 saliva-secretory proteins and 350 non-salivary secretory proteins were used to train the model in each iteration, and the model training was then evaluated on an independent test set consisting of 100 salivary secretory proteins and 100 non-salivary secretory proteins. Then, the average of the results of 10 iterations was calculated as the final prediction result. The performance metrics of SecProCT and the other methods are shown in Table 2 and Table 3. On the independent test set, the average accuracy, sensitivity, specificity, MCC and AUC for predicting blood-secretory proteins and saliva-secretory proteins were 0.917, 0.906, 0.930, 0.835, 0.967 and 0.909, 0.898, 0.919, 0.817, 0.956, respectively.
The performance metrics of SecProCT and the other methods are shown in Table 2 and Table 3. Among them, the SVM methods for predicting blood-secretory protein and saliva-secretory protein were proposed by Cui et al. [10] and Sun et al. [8]. To ensure a comprehensive and systematic comparison, we also constructed several other prediction models based on the features selected in the literature [8,10], including the K-nearest neighbour (KNN), decision tree (DT), random forest (RF) and adaptive boosting (AdaBoost) methods. According to Table 2 and Table 3, it can be seen that on the independent test set that the performance of SecProCT was better than that of the other conventional machine learning methods.
To better evaluate the performance of our model, we also compared our model with the existing deep learning architecture on an independent test set. In recent years, there have been many excellent works applying deep learning methods to biological sequence analysis. For example, a convolutional neural network was used in DeepSig to detect signal peptides in proteins [17]. DanQ used a hybrid CNN and a bidirectional long-term short-term memory network to predict the characteristics and functions of DNA sequences [18]. In DeepLoc, an end-to-end model based on a convolutional neural network, bidirectional long short-term memory network and an attention mechanism was used to predict the protein subcellular localisation [19]. Du et al. used a novel end-to-end deep learning model based on multilane capsule network (CapsNet-SSP) with differently sized convolution kernels to identify saliva-secretory proteins [20]. To ensure a fair comparison with other deep learning architectures, we used the same balanced dataset and training strategy as the proposed SecProCT model to train these deep learning models.
Specifically, we used the deep learning architecture proposed in DeepSig, DanQ, DeepLoc and CapsNet-SSP to replace the network architecture part of the code. Table 4 and Table 5 describe the performance comparison results between different deep learning architectures for predicting blood-secretory proteins and saliva-secretory proteins. As shown in Table 4 and Table 5, the performance of SecProCT on the independent test set was significantly better than the other deep learning architectures.
In addition, we also compared our method with other methods on the iMSP test dataset [13]. Zhang et al. recently proposed a method for predicting secretory proteins in plasma based on discrete firefly optimisation, which had satisfactory results [14]. Since the authors did not provide the relevant source code and program, we evaluated the performance of our model on the iMSP test dataset and compared the experimental results with those in article [14]. Table 6 shows the performance metrics of SecProCT and these methods applied to the iMSP test dataset. Table 6 shows that the performance of SecProCT was better than that of the other methods.

2.2. Ablation Study on Binary Classification

To understand what makes our framework effective, we compared the performance of different combinations of convolutional (Conv) blocks, capsule network (CapsNet) blocks and transformer encoder (Trans) blocks using 10-fold cross-validation on the training dataset. For all combinations, the same training set and validation set were used. The performance comparison results of different combinations for predicting blood-secretory protein and saliva-secretory protein are shown in Table 7 and Table 8. The prediction results of all combinations on blood-secretory proteins and saliva-secretory proteins were satisfactory, with accuracies of 0.921 and 0.892 and MCCs of 0.686 and 0.759, respectively.
Comparing these results with the prediction results without the CapsNet block and without the Trans block (MCCs of 0.674, 0.722 and 0.647, 0.750, respectively), it can be found that the CapsNet block and Trans block can improve the prediction performance. In addition, by comparing the architecture using all blocks and without the Conv block, we can see that the Conv block also improved the prediction performance. As shown in Table 7 and Table 8, the average accuracy, sensitivity, specificity, MCC and AUC using all combinations for predicting blood-secretory protein and saliva-secretory protein were 0.921, 0.911, 0.931, 0.686, 0.950 and 0.892, 0.834, 0.950, 0.759, 0.920, respectively, which were better than those of the other architectures.

2.3. Evaluating the Performance of Human Secretory Protein Prediction

To further validate the effectiveness of the proposed model, we collected additional human secretory proteins that do not overlap with our training set. We collected known blood-secretory proteins from the published literature [21] and removed the proteins in the training set. Finally, 968 new blood-secretory proteins were obtained to evaluate the performance of the blood-secretory protein prediction. There are 3449 proteins that have been detected in saliva by using LC–MS/MS analyses in several published studies [22,23,24,25,26,27], among which 1628 have been detected in more than one study. Of these 1628 proteins, 529 are secretory proteins listed in the secretory protein database. Then, we removed the proteins in the training set from the 529 proteins and obtained 215 saliva-secretory proteins.
Figure 1 and Figure 2 show the prediction results for the 968 blood-secretory proteins and 215 saliva-secretory proteins using the model based on DeepSig, DanQ, DeepLoc, CapsNet-SSP and SecProCT. Figure 1 and Figure 2 show the recall when using different threshold values in the prediction. As shown in Figure 1 and Figure 2, the prediction results of SecProCT are better than those of the other methods.

2.4. Evaluating the Performance of Cancer Biomarker Prediction

To further confirm the effectiveness of the proposed model in detecting disease markers, we also evaluated the prediction results of existing cancer biomarkers in blood and saliva. The existing blood biomarkers were collected from published research [28,29]. In this study [28], Ahn et al. analysed biomarkers in the blood of patients with colorectal cancer and obtained 41 biomarkers for the diagnosis of colorectal cancer. In study [29], Ahn et al. analysed the blood biomarkers of patients with small-cell lung cancer and obtained 106 biomarkers for the diagnosis of lung cancer. We combined the cancer markers of the above two studies and removed the overlapping proteins and finally obtained 129 cancer markers in blood. We also collected saliva in head and neck squamous cell carcinoma (HNSCC) [8], oral squamous cell carcinoma (OSCC) [30], lung cancer (LC) [31] and breast cancer (BC) [32] and obtained cancer markers in saliva.
Figure 3 and Figure 4 show the prediction results of 129 cancer markers in blood and 215 cancer markers in saliva using the model based on DeepSig, DanQ, DeepLoc, CapsNet-SSP and SecProCT. Figure 3 and Figure 4 show the recall by using different threshold values in the prediction. As shown in Figure 3 and Figure 4, the prediction results of SecProCT are better than those of the other methods.

3. Discussion

The main contribution of this paper was the proposal of an end-to-end prediction model based on deep learning, which can accurately identify secretory proteins only using protein sequence information and obtain statistical significance for identifying existing cancer biomarkers in blood and saliva. Compared with traditional machine learning methods, our model can automatically learn feature representations and predict results from amino acid sequences, which can reduce feature incompleteness or bias that may result from feature engineering and the feature selection process of traditional machine learning methods. The prediction results of the proposed model are better than those of existing conventional machine learning methods and deep learning methods for biological sequence analysis. The disadvantage of the proposed model is that the training time is longer than that of traditional machine learning methods.

4. Materials and Methods

4.1. Data Collection

First, we collected human secretory proteins from the secreted protein database (SPD)  [33], the mammalian protein subcellular localisation database (LOCATE) [34], the Universal Protein Resource (UniProt) [35] and the Subcellular Proteome KnowledgeBase (MetazSecKB) [36]. At present, there are already many databases that have collected proteins that can be detected in blood and saliva. In addition, some studies have reported the proteins that appear in blood and saliva. We collected the detectable proteins in blood and saliva from the Sys-BodyFluid database [37], which contains proteins from different human tissues collected from many proteomics studies. We collected blood proteins from the Plasma Proteome Project (PPP) database [38] and other studies of blood proteins [10]. We also collected saliva proteins from two other saliva proteomics studies [24,39]. Then, the intersections of the human secretory proteins and the blood and saliva proteins were calculated to obtain the blood-secretory proteins and saliva-secretory proteins, respectively. In addition, to prevent learning bias due to protein redundancies, we removed the proteins that had a mutual sequence similarity above 30% using the CD-HIT tool [40]. Finally, 486 blood-secretory proteins and 450 saliva-secretory proteins remained, which were regarded as positive data. Among these blood-secretory proteins, 380 proteins were classified into the positive training set, and the other 106 proteins were classified into the positive test set. Among these saliva-secretory proteins, 350 proteins were classified into the positive training set, and the other 100 proteins were classified into the positive test set.
Since no protein is clearly defined as a non-blood-secretory protein and non-saliva-secretory protein in existing studies, generating a negative dataset was a difficult step in the data collection process. In this article, we used a method similar to that proposed by Cui et al. [10] to select proteins from the Pfam family that do not contain positive data, suggesting they are non-blood-secretory proteins and non-saliva-secretory proteins. To reduce the influence of protein families that only contain a small number of proteins, we only selected proteins from protein families that have more than ten proteins. For each protein family, we selected three members to construct the negative data. Then, we used the CD-HIT tool to remove the proteins with a mutual sequence similarity higher than 30% from the negative data. Among the 3900 obtained non-blood-secretory proteins, 3800 proteins were divided into the negative training set, and the other 100 proteins were divided into the negative test set. Among the 1850 non-saliva-secretory proteins, 1750 proteins were divided into the negative training set, and the other 100 proteins were divided into the negative test set.

4.2. Input Sequence Encoding

After completing the data collection, we used PSI-BLAST [15] to perform sequence alignment in the “Uniref50” database [16] to obtain evolutionary profile information for protein sequence encoding. Specifically, the profile of each protein is actually a normalised position-specific scoring matrix (PSSM) [41], which is generated based on the amino acid frequencies at every position of the multiple alignment using PROFILpro [42]. The sequence lengths of the different proteins are different, and the lengths of the obtained position-specific scoring matrices are also different. If the longest protein coding length is used uniformly, the overall training time of the model will be too long. Therefore, we appropriately cropped the encoded data and set 1000 as the maximum length of the input data. When the length of the PSSM of a certain protein was less than 1000, 0 was filled at the end of it. Since most of the information about blood-secretory proteins and saliva-secretory proteins is stored at the beginning (N-terminus) and the end (C-terminus) of the sequence [8,10], when the length of the PSSM of a certain protein exceeds 1000, we selected 500 amino acids from the beginning (N-terminus) of the protein and 500 amino acids from the end (C-terminus) of the protein to avoid losing the N-terminus and C-terminus classification signals. When applying this rule, only 11.74% and 16.32% of human blood-secretory proteins and saliva-secretory proteins are truncated.

4.3. Architecture Design

The prediction architecture of the proposed SecProCT model is shown in Figure 5. The model contains a feature extraction subnetwork ( N f ) and a classification subnetwork ( N c ). The input of the model is a 1000 × 20 profile of the PSSM for each protein. The feature extraction subnetwork ( N f ) consists of multiple feature extraction blocks, each of which contains a convolutional (Conv) block, a capsule network (CapsNet) [43] block and a transformer encoder (Trans) [44] block, and one concatenation layer which is used to concatenate the extracted features. The classification subnetwork ( N c ) contains two fully connected dense layers, and one softmax output layer. In this article, the feature extraction subnetwork ( N f ) contains eight feature extraction blocks, which have different Conv layers containing 1D convolution kernels of a specified size with a stride of 1 and a rectified linear unit (ReLU) activation function [45].
The size of the kernel in the Conv layer determines the length of the motif that can be extracted from the protein sequence. In the network architecture proposed by Armenteros et al. for predicting protein subcellular localisation [19], the kernel size of each channel is 1, 3, 5, 9, 15, and 21. In this paper, to better identify the important characteristics of the signal peptides and transmembrane domains for predicting secretory proteins [10], we added kernel sizes of 27 and 33 to the network architecture. These Conv layers can convert the initial PSSM profile of each protein to intermediate-level features, which are then fed into the CapsNet block and the Trans block for further feature abstraction.
In each feature extraction block, the CapsNet block and Trans block are parallel connections. Each CapsNet block contains the PrimaryCaps layer and the HiddenCaps layer. The PrimaryCaps layer is also a 1D convolutional layer containing eight convolutional capsule channels. Each capsule contains 64 convolution units, and each capsule contains a 1D convolution kernel with a stride of 1. The squashing activation function is used in the PrimaryCaps layer to scale the lengths of the capsules to [0, 1] as follows:
v j = s j 1 + s j 2 s j s j ,
where s j is the input vector of capsule j and v j is its output vector.
There are eight 16D capsules in the HiddenCaps layer, which can map the input protein to different states. The computation between the PrimaryCaps layer and HiddenCaps layer is the same as the calculation in the original CapsNet paper [43]. The structure of the Trans block is similar to the encoder part in the original Transformer paper [44], which contains two sub-layers. The first layer is a multi-head self-attention mechanism layer, and the second layer is a fully connected feed-forward network. For each of the two sub-layers, a residual connection is employed followed by layer normalisation. In the Trans block, the output dimension of all sub-layers and embedding layers is defined as 16 to facilitate these residual connections.

4.4. Model Training

For predictions based on deep learning frameworks, one of the challenging problems that must be solved is how to train highly accurate generalised models with small samples. If the number of training samples is much smaller than the number of parameters in the deep learning model, there is a high risk of overfitting during the training process [9]. Transfer learning refers to the transfer of knowledge learned in related tasks to new tasks [46], thereby improving the learning effect, which has been successfully applied to a deep learning framework for small sample problems [47]. To solve the problem of small samples in secretory protein prediction, we used the concept of transfer learning to fine-tune the prediction model of secretory proteins by the pre-trained prediction model of protein subcellular localisation. The dataset of protein subcellular localisation is from DeepLoc, which contains 13,858 proteins with ten main locations [19]. We trained the framework of the proposed model on the dataset of protein subcellular localisation and then transferred all layers except the output layer to the secretory protein prediction model. Finally, we used the blood-secretory protein and saliva-secretory protein data to fine-tune the entire model. At the same time, the bagging ensemble learning method [48] was used in the training process to reduce the impact of unbalanced data. By training the model on multiple selected balanced training subsets, we obtained multiple independent classifiers. Then, the final prediction result was calculated by averaging the results of these independent classifiers. The bagging ensemble learning algorithm used to train our proposed model is given below:
Algorithm 1: Bagging Algorithm [49]
Input:
   S + : Training set with positive samples
   S : Training set with negative samples
  T: Number of iterations
n: The size of a random selection   
I : Weak classifier
1:
for  t = 1 to T do
2:
S t = Random Sample Selection ( n , S )
3:
h t = I S t S +
4:
end for
Output:
  Bagged classifier: H x = sgn t = 1 T h t x where h t 1 , 1
Here, S + contains 380 proteins and 350 proteins for blood and saliva, respectively. S contains 3800 proteins and 1750 proteins for blood and saliva, respectively. The number of iterations T is 10, and the size of the random selection n is 380 and 350 for blood and saliva, respectively.
Dropout, early stopping and L2 regularisation strategies were used to prevent overfitting during the training of complex deep learning models. The dropout strategy prevents overfitting and optimises the generalisation ability of the model by adding multiple dropout layers to the prediction model. Another strategy to reduce overfitting was early stopping, that is, stopping training early at the right time during the training process. Specifically, when the loss of validation data does not decrease within a preset iteration period, the training process stops early [50]. The third strategy we used to prevent overfitting is to use regularisation in the training process of neural networks. The final optimisation function contains two items: one is a loss term used to quantify the degree of fitting the model to the data, and the other is a regular term used to quantify the complexity of the model and to prevent overfitting. In this study, we used L2 regularisation to prevent the overfitting of the deep learning model.
In the model training process, the proposed deep learning model and other deep learning models used the same training strategies. These deep learning models were optimised using the Adam stochastic optimisation method [51] and used the following parameters: the exponential decay rate of the first moment estimation was 0.9, the exponential decay rate of the second moment estimation was 0.999, and the learning rate was 0.0005. We used cross-entropy loss as the loss function to measure the difference between the true distribution and the predicted distribution of secretory proteins and non-secretory proteins. The proposed model and the comparative models used in the experiment were all executed on a workstation equipped with Ubuntu 18.04.2 LTS operating system, Intel Core i7-7800X CPU, 128 GB RAM and NVIDIA GeForce RTX 2080 Ti GPU. The software environment for our proposed deep learning model is Keras 2.2.4 and TensorFlow 1.13.1.

4.5. Performance Measurements

To compare the performance of different prediction models, the accuracy, sensitivity, specificity, Matthews correlation coefficient (MCC) and area under the receiver operating characteristic (ROC) curve (AUC) were used as performance metrics. The corresponding formulas are as follows:
accuracy = T P + T N N t o t a l ,
sensitivity = recall = T P T P + F N ,
specificity = T N T N + F P ,
MCC = T P × T N F P × F N ( T P + F N ) ( T P + F P ) ( T N + F P ) ( T N + F N ) ,
AUC = p 0 D 0 p 1 D 1 g ( p 0 , p 1 ) | D 0 | · | D 1 | , where g ( p 0 , p 1 ) = 1 if ( f ( p 0 ) < f ( p 1 ) ) 0 otherwise ,
where T P represents the true positive, T N represents the true negative, F P represents the false positive, F N represents the false negative, and N t o t a l is the total number of samples in the validation or test set. D 0 and D 1 are the positive dataset and the negative dataset, respectively. f(.) represents the prediction function. In machine learning, MCC has a distribution range of [−1, 1], which is a measurement used to quantify the quality of the binary classification. When the value is 1, it means that the prediction is exactly the same as the label. When the value is −1, it means that the prediction result and the label are completely opposite. Since the accuracy, sensitivity, specificity and MCC depend on the threshold, it is necessary to choose a threshold to calculate their value. In the evaluation of binary classification, MCC produces more informative and truthful scores than accuracy and AUC. Therefore, in this article, the threshold is set where the MCC reaches its maximum value. AUC is one of the most important evaluation indexes to measure the performance of any classification model. It is a performance measurement of classification problems under different threshold settings.

5. Conclusions

In this study, we propose an end-to-end model based on capsule network and transformer architecture, which can accurately identify human secretory proteins from only amino acid sequence information. The first step in constructing the model was to use PSI-BLAST to transform the input protein sequence into a normalised position-specific scoring matrix (PSSM). In addition, to solve the problem of unbalanced datasets, the bagging ensemble learning method was applied during the training process. Then, the PSSM of the training set was fed into the model to train the model parameters. Finally, the trained model was used to predict the secretory proteins in the test set, and its performance was verified to be better than the existing methods. At the same time, our model can accurately detect known human secretory proteins and cancer biomarkers.

Author Contributions

Conceptualisation, W.D. and Y.Z.; methodology, W.D., Y.L. and Y.Z.; validation, W.D., X.Z., Y.S. and L.Z.; formal analysis, W.D., X.Z. and Y.Z.; writing—original draft preparation, W.D., X.Z. and Y.S.; writing—review and editing, W.D., Y.L. and Y.Z.; project administration, W.D.; funding acquisition, W.D., Y.L. and Y.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (61872418 and 61972175), the Natural Science Foundation of Jilin Province (20180101050JC and 20180101331JC).

Data Availability Statement

The pipeline of the model, prediction data and results can be accessed at the following URL: http://www.csbg-jlu.info/SecProCT/, accessed on 15 August 2021.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Nagpal, M.; Singh, S.; Singh, P.; Chauhan, P.; Zaidi, M.A. Tumor markers: A diagnostic tool. Natl. J. Maxillofac. Surg. 2016, 7, 17–20. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Loke, S.Y.; Lee, A.S.G. The future of blood-based biomarkers for the early detection of breast cancer. Eur. J. Cancer 2018, 92, 54–68. [Google Scholar] [CrossRef] [PubMed]
  3. Pfaffe, T.; Cooper-White, J.; Beyerlein, P.; Kostner, K.; Punyadeera, C. Diagnostic potential of saliva: Current state and future applications. Clin. Chem. 2011, 57, 675–687. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Castagnola, M.; Scarano, E.; Passali, G.C.; Messana, I.; Cabras, T.; Iavarone, F.; Di Cintio, G.; Fiorita, A.; De Corso, E.; Paludetti, G. Salivary biomarkers and proteomics: Future diagnostic and clinical utilities. Acta Otorhinolaryngol. Ital. 2017, 37, 94–101. [Google Scholar] [CrossRef]
  5. Robinson, J.L.; Feizi, A.; Uhlen, M.; Nielsen, J. A Systematic Investigation of the Malignant Functions and Diagnostic Potential of the Cancer Secretome. Cell Rep. 2019, 26, 2622–2635.e5. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Geyer, P.E.; Holdt, L.M.; Teupser, D.; Mann, M. Revisiting biomarker discovery by plasma proteomics. Mol. Syst. Biol. 2017, 13, 942. [Google Scholar] [CrossRef]
  7. Huang, L.; Shao, D.; Wang, Y.; Cui, X.; Li, Y.; Chen, Q.; Cui, J. Human body-fluid proteome: Quantitative profiling and computational prediction. Brief. Bioinform. 2021, 22, 315–333. [Google Scholar] [CrossRef] [Green Version]
  8. Sun, Y.; Du, W.; Zhou, C.; Zhou, Y.; Cao, Z.; Tian, Y.; Wang, Y. A computational method for prediction of saliva-secretory proteins and its application to identification of head and neck cancer biomarkers for salivary diagnosis. IEEE Trans. Nanobiosci. 2015, 14, 167–174. [Google Scholar] [CrossRef] [PubMed]
  9. Wang, D.; Zeng, S.; Xu, C.; Qiu, W.; Liang, Y.; Joshi, T.; Xu, D. MusiteDeep: A deep-learning framework for general and kinase-specific phosphorylation site prediction. Bioinformatics 2017, 33, 3909–3916. [Google Scholar] [CrossRef] [Green Version]
  10. Cui, J.; Liu, Q.; Puett, D.; Xu, Y. Computational prediction of human proteins that can be secreted into the bloodstream. Bioinformatics 2008, 24, 2370–2375. [Google Scholar] [CrossRef]
  11. Liu, Q.; Cui, J.; Yang, Q.; Xu, Y. In-silico prediction of blood-secretory human proteins using a ranking algorithm. BMC Bioinform. 2010, 11, 250. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  12. Wang, J.; Liang, Y.; Wang, Y.; Cui, J.; Liu, M.; Du, W.; Xu, Y. Computational prediction of human salivary proteins from blood circulation and application to diagnostic biomarker identification. PLoS ONE 2013, 8, e80211. [Google Scholar] [CrossRef] [PubMed]
  13. Zhang, J.; Chai, H.; Guo, S.; Guo, H.; Li, Y. High-Throughput Identification of Mammalian Secreted Proteins Using Species-Specific Scheme and Application to Human Proteome. Molecules 2018, 23, 1448. [Google Scholar] [CrossRef] [Green Version]
  14. Zhang, J.; Zhang, Y.; Ma, Z. In silico Prediction of Human Secretory Proteins in Plasma Based on Discrete Firefly Optimization and Application to Cancer Biomarkers Identification. Front. Genet. 2019, 10, 542. [Google Scholar] [CrossRef] [Green Version]
  15. Altschul, S.F.; Madden, T.L.; Schaffer, A.A.; Zhang, J.; Zhang, Z.; Miller, W.; Lipman, D.J. Gapped BLAST and PSI-BLAST: A new generation of protein database search programs. Nucleic Acids Res. 1997, 25, 3389–3402. [Google Scholar] [CrossRef] [Green Version]
  16. Suzek, B.E.; Wang, Y.; Huang, H.; McGarvey, P.B.; Wu, C.H.; UniProt, C. UniRef clusters: A comprehensive and scalable alternative for improving sequence similarity searches. Bioinformatics 2015, 31, 926–932. [Google Scholar] [CrossRef] [Green Version]
  17. Savojardo, C.; Martelli, P.L.; Fariselli, P.; Casadio, R. DeepSig: Deep learning improves signal peptide detection in proteins. Bioinformatics 2018, 34, 1690–1696. [Google Scholar] [CrossRef] [Green Version]
  18. Quang, D.; Xie, X. DanQ: A hybrid convolutional and recurrent deep neural network for quantifying the function of DNA sequences. Nucleic Acids Res. 2016, 44, e107. [Google Scholar] [CrossRef] [Green Version]
  19. Almagro Armenteros, J.J.; Sonderby, C.K.; Sonderby, S.K.; Nielsen, H.; Winther, O. DeepLoc: Prediction of protein subcellular localization using deep learning. Bioinformatics 2017, 33, 3387–3395. [Google Scholar] [CrossRef]
  20. Du, W.; Sun, Y.; Li, G.; Cao, H.; Pang, R.; Li, Y. CapsNet-SSP: Multilane capsule network for predicting human saliva-secretory proteins. BMC Bioinform. 2020, 21, 237. [Google Scholar] [CrossRef]
  21. Emilsson, V.; Ilkov, M.; Lamb, J.R.; Finkel, N.; Gudmundsson, E.F.; Pitts, R.; Hoover, H.; Gudmundsdottir, V.; Horman, S.R.; Aspelund, T.; et al. Co-regulatory networks of human serum proteins link genetics to disease. Science 2018, 361, 769–773. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. Yan, W.; Apweiler, R.; Balgley, B.M.; Boontheung, P.; Bundy, J.L.; Cargile, B.J.; Cole, S.; Fang, X.; Gonzalez-Begne, M.; Griffin, T.J.; et al. Systematic comparison of the human saliva and plasma proteomes. Proteom. Clin. Appl. 2009, 3, 116–134. [Google Scholar] [CrossRef] [Green Version]
  23. Xie, H.; Rhodus, N.L.; Griffin, R.J.; Carlis, J.V.; Griffin, T.J. A catalogue of human saliva proteins identified by free flow electrophoresis-based peptide separation and tandem mass spectrometry. Mol. Cell Proteom. 2005, 4, 1826–1830. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Denny, P.; Hagen, F.K.; Hardt, M.; Liao, L.; Yan, W.; Arellanno, M.; Bassilian, S.; Bedi, G.S.; Boontheung, P.; Cociorva, D.; et al. The proteomes of human parotid and submandibular/sublingual gland salivas collected as the ductal secretions. J. Proteome Res. 2008, 7, 1994–2006. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Fang, X.; Yang, L.; Wang, W.; Song, T.; Lee, C.S.; DeVoe, D.L.; Balgley, B.M. Comparison of electrokinetics-based multidimensional separations coupled with electrospray ionization-tandem mass spectrometry for characterization of human salivary proteins. Anal. Chem. 2007, 79, 5785–5792. [Google Scholar] [CrossRef]
  26. Bandhakavi, S.; Stone, M.D.; Onsongo, G.; Van Riper, S.K.; Griffin, T.J. A dynamic range compression and three-dimensional peptide fractionation analysis platform expands proteome coverage and the diagnostic potential of whole saliva. J. Proteome Res. 2009, 8, 5590–5600. [Google Scholar] [CrossRef] [Green Version]
  27. Sivadasan, P.; Kumar Gupta, M.; Sathe, G.J.; Balakrishnan, L.; Palit, P.; Gowda, H.; Suresh, A.; Abraham Kuriakose, M.; Sirdeshmukh, R. Data from human salivary proteome—A resource of potential biomarkers for oral cancer. Data Brief 2015, 4, 374–378. [Google Scholar] [CrossRef] [Green Version]
  28. Ahn, S.B.; Sharma, S.; Mohamedali, A.; Mahboob, S.; Redmond, W.J.; Pascovici, D.; Wu, J.X.; Zaw, T.; Adhikari, S.; Vaibhav, V.; et al. Potential early clinical stage colorectal cancer diagnosis using a proteomics blood test panel. Clin. Proteom. 2019, 16, 34. [Google Scholar] [CrossRef]
  29. Ahn, J.M.; Sung, H.J.; Yoon, Y.H.; Kim, B.G.; Yang, W.S.; Lee, C.; Park, H.M.; Kim, B.J.; Kim, B.G.; Lee, S.Y.; et al. Integrated glycoproteomics demonstrates fucosylated serum paraoxonase 1 alterations in small cell lung cancer. Mol. Cell Proteom. 2014, 13, 30–48. [Google Scholar] [CrossRef] [Green Version]
  30. Chu, H.W.; Chang, K.P.; Hsu, C.W.; Chang, I.Y.; Liu, H.P.; Chen, Y.T.; Wu, C.C. Identification of Salivary Biomarkers for Oral Cancer Detection with Untargeted and Targeted Quantitative Proteomics Approaches. Mol. Cell Proteom. 2019, 18, 1796–1806. [Google Scholar] [CrossRef]
  31. Xiao, H.; Zhang, L.; Zhou, H.; Lee, J.M.; Garon, E.B.; Wong, D.T. Proteomic analysis of human saliva from lung cancer patients using two-dimensional difference gel electrophoresis and mass spectrometry. Mol. Cell Proteom. 2012, 11, M111.012112. [Google Scholar] [CrossRef] [Green Version]
  32. Porto-Mascarenhas, E.C.; Assad, D.X.; Chardin, H.; Gozal, D.; De Luca Canto, G.; Acevedo, A.C.; Guerra, E.N. Salivary biomarkers in the diagnosis of breast cancer: A review. Crit. Rev. Oncol. Hematol. 2017, 110, 62–73. [Google Scholar] [CrossRef] [PubMed]
  33. Chen, Y.; Zhang, Y.; Yin, Y.; Gao, G.; Li, S.; Jiang, Y.; Gu, X.; Luo, J. SPD–a web-based secreted protein database. Nucleic Acids Res. 2005, 33, D169–D173. [Google Scholar] [CrossRef] [Green Version]
  34. Sprenger, J.; Lynn Fink, J.; Karunaratne, S.; Hanson, K.; Hamilton, N.A.; Teasdale, R.D. LOCATE: A mammalian protein subcellular localization database. Nucleic Acids Res. 2008, 36, D230–D233. [Google Scholar] [CrossRef] [PubMed]
  35. UniProt, C. UniProt: A worldwide hub of protein knowledge. Nucleic Acids Res. 2019, 47, D506–D515. [Google Scholar] [CrossRef] [Green Version]
  36. Meinken, J.; Walker, G.; Cooper, C.R.; Min, X.J. MetazSecKB: The human and animal secretome and subcellular proteome knowledgebase. Database 2015, 2015, bav077. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. Li, S.J.; Peng, M.; Li, H.; Liu, B.S.; Wang, C.; Wu, J.R.; Li, Y.X.; Zeng, R. Sys-BodyFluid: A systematical database for human body fluid proteome research. Nucleic Acids Res. 2009, 37, D907–D912. [Google Scholar] [CrossRef]
  38. Omenn, G.S. The HUPO Human Plasma Proteome Project. Proteom. Clin. Appl. 2007, 1, 769–779. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  39. Hu, S.; Loo, J.A.; Wong, D.T. Human saliva proteome analysis and disease biomarker discovery. Expert Rev. Proteom. 2007, 4, 531–538. [Google Scholar] [CrossRef]
  40. Huang, Y.; Niu, B.; Gao, Y.; Fu, L.; Li, W. CD-HIT Suite: A web server for clustering and comparing biological sequences. Bioinformatics 2010, 26, 680–682. [Google Scholar] [CrossRef] [PubMed]
  41. Maurer-Stroh, S.; Debulpaep, M.; Kuemmerer, N.; Lopez de la Paz, M.; Martins, I.C.; Reumers, J.; Morris, K.L.; Copland, A.; Serpell, L.; Serrano, L.; et al. Exploring the sequence determinants of amyloid structure using position-specific scoring matrices. Nat. Methods 2010, 7, 237–242. [Google Scholar] [CrossRef] [PubMed]
  42. Magnan, C.N.; Baldi, P. SSpro/ACCpro 5: Almost perfect prediction of protein secondary structure and relative solvent accessibility using profiles, machine learning and structural similarity. Bioinformatics 2014, 30, 2592–2597. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  43. Sabour, S.; Frosst, N.; Hinton, G.E. Dynamic routing between capsules. arXiv 2017, arXiv:1710.09829. [Google Scholar]
  44. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  45. Li, Y.; Yuan, Y. Convergence analysis of two-layer neural networks with relu activation. arXiv 2017, arXiv:1705.09886. [Google Scholar]
  46. Caruana, R. Learning many related tasks at the same time with backpropagation. In Advances in Neural Information Processing Systems; Massachusetts Institute of Technology Press: Cambridge, MA, USA, 1995; pp. 657–664. [Google Scholar]
  47. Ng, H.W.; Nguyen, V.D.; Vonikakis, V.; Winkler, S. Deep learning for emotion recognition on small datasets using transfer learning. In Proceedings of the 2015 ACM on International Conference on Multimodal Interaction, Seattle, WA, USA, 9–13 November 2015; pp. 443–449. [Google Scholar]
  48. Galar, M.; Fernandez, A.; Barrenechea, E.; Bustince, H.; Herrera, F. A review on ensembles for the class imbalance problem: Bagging-, boosting-, and hybrid-based approaches. IEEE Trans. Syst. Man Cybern. Part C 2011, 42, 463–484. [Google Scholar] [CrossRef]
  49. Breiman, L. Bagging predictors. Mach. Learn. 1996, 24, 123–140. [Google Scholar] [CrossRef] [Green Version]
  50. Yao, Y.; Rosasco, L.; Caponnetto, A. On early stopping in gradient descent learning. Construct. Approx. 2007, 26, 289–315. [Google Scholar] [CrossRef]
  51. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2017, arXiv:cs.LG/1412.6980. [Google Scholar]
Figure 1. Prediction results of known blood-secretory proteins.
Figure 1. Prediction results of known blood-secretory proteins.
Ijms 22 09054 g001
Figure 2. Prediction results of known saliva-secretory proteins.
Figure 2. Prediction results of known saliva-secretory proteins.
Ijms 22 09054 g002
Figure 3. Prediction results of known cancer markers in blood.
Figure 3. Prediction results of known cancer markers in blood.
Ijms 22 09054 g003
Figure 4. Prediction results of known cancer markers in saliva.
Figure 4. Prediction results of known cancer markers in saliva.
Ijms 22 09054 g004
Figure 5. Architecture of the proposed model.
Figure 5. Architecture of the proposed model.
Ijms 22 09054 g005
Table 1. Existing studies for predicting secretory proteins.
Table 1. Existing studies for predicting secretory proteins.
StudyBody FluidAlgorithmAdvantageDisadvantageRef
Cui et al. (2008)BloodSVM15 [10]
Liu et al. (2010)BloodRanking25 [11]
Wang et al. (2013)SalivaSVM45 [12]
Sun et al. (2015)SalivaSVM15 [8]
Zhang et al. (2018)Body fluidsSVM35 [13]
Zhang et al. (2019)BloodDFO *15 [14]
1: High prediction accuracy; 2: Not need negative training dataset; 3: Strong biological correlation; 4: For all body fluids; 5: Need feature construction and feature selection. * DFO: Discrete firefly optimisation.
Table 2. Performance comparison with machine learning methods for predicting blood-secretory proteins.
Table 2. Performance comparison with machine learning methods for predicting blood-secretory proteins.
MethodsAccuracySensitivitySpecificityMCCAUC
KNN0.8790.8480.9110.7590.873
Decision Tree0.8580.8520.8640.7160.838
Random Forest0.8660.8180.9160.7360.922
AdaBoost0.8590.8380.8810.7190.912
SVM0.8930.8770.9100.7870.925
SecProCT0.9170.9060.9300.8350.967
Table 3. Performance comparison with machine learning methods for predicting saliva-secretory proteins.
Table 3. Performance comparison with machine learning methods for predicting saliva-secretory proteins.
MethodsAccuracySensitivitySpecificityMCCAUC
KNN0.7780.6490.9070.5750.809
Decision Tree0.7720.6920.8510.5500.740
Random Forest0.7810.8040.7580.5630.836
AdaBoost0.7920.7030.8810.5930.847
SVM0.7810.7840.7780.5620.857
SecProCT0.9090.8980.9190.8170.956
Table 4. Performance comparison with deep learning architectures for predicting blood-secretory proteins.
Table 4. Performance comparison with deep learning architectures for predicting blood-secretory proteins.
MethodsAccuracySensitivitySpecificityMCCAUC
DeepSig0.8930.9060.8800.7860.939
DanQ0.8500.7830.9200.7080.927
DeepLoc0.8980.8770.9200.7970.940
CapsNet-SSP0.8930.8970.8900.7860.942
SecProCT0.9170.9060.9300.8350.967
Table 5. Performance comparison with deep learning architectures for predicting saliva-secretory proteins.
Table 5. Performance comparison with deep learning architectures for predicting saliva-secretory proteins.
MethodsAccuracySensitivitySpecificityMCCAUC
DeepSig0.7920.7450.8380.5860.867
DanQ0.8020.7450.8590.6080.886
DeepLoc0.8430.7550.9290.6950.891
CapsNet-SSP0.8880.8470.9290.7790.948
SecProCT0.9090.8980.9190.8170.956
Table 6. Performance of SecProCT and other methods on the iMSP test set.
Table 6. Performance of SecProCT and other methods on the iMSP test set.
MethodsAccuracySensitivitySpecificityMCCAUC
SecretomeP0.7620.6320.7870.3400.764
SRTpred0.7820.6780.8020.3920.770
iMSP-U0.8290.6310.8660.4430.821
iMSP-H0.8500.5380.9080.4410.817
SCRIP0.8580.7160.8840.5370.865
SecProCT0.8730.8090.8840.5960.920
Table 7. Performance of different architectures for predicting blood-secretory proteins on the training set.
Table 7. Performance of different architectures for predicting blood-secretory proteins on the training set.
ArchitecturesAccuracySensitivitySpecificityMCCAUC
Without Conv block0.8820.8530.9120.6040.915
Without CapsNet block0.9160.9030.9280.6740.947
Without Trans block0.9050.8890.9210.6470.938
All0.9210.9110.9310.6860.950
Table 8. Performance of different architectures for predicting saliva-secretory proteins on the training set.
Table 8. Performance of different architectures for predicting saliva-secretory proteins on the training set.
ArchitecturesAccuracySensitivitySpecificityMCCAUC
Without Conv block0.8750.8480.9020.6710.909
Without CapsNet block0.8910.8570.9250.7220.918
Without Trans block0.8880.8280.9480.7500.915
All0.8920.8340.9500.7590.920
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Du, W.; Zhao, X.; Sun, Y.; Zheng, L.; Li, Y.; Zhang, Y. SecProCT: In Silico Prediction of Human Secretory Proteins Based on Capsule Network and Transformer. Int. J. Mol. Sci. 2021, 22, 9054. https://doi.org/10.3390/ijms22169054

AMA Style

Du W, Zhao X, Sun Y, Zheng L, Li Y, Zhang Y. SecProCT: In Silico Prediction of Human Secretory Proteins Based on Capsule Network and Transformer. International Journal of Molecular Sciences. 2021; 22(16):9054. https://doi.org/10.3390/ijms22169054

Chicago/Turabian Style

Du, Wei, Xuan Zhao, Yu Sun, Lei Zheng, Ying Li, and Yu Zhang. 2021. "SecProCT: In Silico Prediction of Human Secretory Proteins Based on Capsule Network and Transformer" International Journal of Molecular Sciences 22, no. 16: 9054. https://doi.org/10.3390/ijms22169054

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop