Next Article in Journal
PUL-Mediated Plant Cell Wall Polysaccharide Utilization in the Gut Bacteroidetes
Next Article in Special Issue
Rare Variants in Autophagy and Non-Autophagy Genes in Late-Onset Pompe Disease: Suggestions of Their Disease-Modifying Role in Two Italian Families
Previous Article in Journal
Accurate Receptor-Ligand Binding Free Energies from Fast QM Conformational Chemical Space Sampling
Previous Article in Special Issue
SARS-CoV-2 Entry Related Viral and Host Genetic Variations: Implications on COVID-19 Severity, Immune Escape, and Infectivity
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Position-Specific Encoding Algorithm (SeqPose) of Nucleotide Sequences and Its Application for Detecting Enhancers

1
Health Informatics Lab, College of Computer Science and Technology, Key Laboratory of Symbolic Computation and Knowledge Engineering of the Ministry of Education, Jilin University, Changchun 130012, China
2
School of Mathematics, Jilin University, Changchun 130012, China
3
Department of Epidemiology and Biostatistics, School of Public Health, Jilin University, Changchun 130021, China
*
Author to whom correspondence should be addressed.
Int. J. Mol. Sci. 2021, 22(6), 3079; https://doi.org/10.3390/ijms22063079
Submission received: 29 November 2020 / Revised: 4 March 2021 / Accepted: 11 March 2021 / Published: 17 March 2021
(This article belongs to the Collection Feature Papers in Molecular Genetics and Genomics)

Abstract

:
Enhancers are short genomic regions exerting tissue-specific regulatory roles, usually for remote coding regions. Enhancers are observed in both prokaryotic and eukaryotic genomes, and their detections facilitate a better understanding of the transcriptional regulation mechanism. The accurate detection and transcriptional regulation strength evaluation of the enhancers remain a major bioinformatics challenge. Most of the current studies utilized the statistical features of short fixed-length nucleotide sequences. This study introduces the location information of each k-mer (SeqPose) into the encoding strategy of a DNA sequence and employs the attention mechanism in the two-layer bi-directional long-short term memory (BD-LSTM) model (spEnhancer) for the enhancer detection problem. The first layer of the delivered classifier discriminates between enhancers and non-enhancers, and the second layer evaluates the transcriptional regulation strength of the detected enhancer. The SeqPose-encoded features are selected by the Chi-squared test, and 45 positions are removed from further analysis. The existing studies may focus on selecting the statistical DNA sequence descriptors with large contributions to the prediction models. This study does not utilize these statistical DNA sequence descriptors. Then the word vector of the SeqPose-encoded features is obtained by using the word embedding layer. This study hypothesizes that different word vector features may contribute differently to the enhancer detection model, and assigns different weights to these word vectors through the attention mechanism in the BD-LSTM model. The previous study generously provided the training and independent test datasets, and the proposed spEnhancer is compared with the three existing state-of-the-art studies using the same experimental procedure. The leave-one-out validation data on the training dataset shows that the proposed spEnhancer achieves similar detection performances as the three existing studies. While spEnhancer achieves the best overall performance metric MCC for both of the two binary classification problems on the independent test dataset. The experimental data shows that the strategy of removing redundant positions (SeqPose) may help improve the DNA sequence-based prediction models. spEnhancer may serve well as a complementary model to the existing studies, especially for the novel query enhancers that are not included in the training dataset.

1. Introduction

The innovative technologies and comprehensive biological investigations show that the non-coding genomic regions are not functionally inactive as previously hypothesized and play essential roles in transcriptional regulations [1]. An enhancer is a small genomic region that binds to transcription factors and exerts its regulatory roles to the target genes [2,3]. Enhancers are different from promoters that are always upstream to the target genes and may even reside in the introns [4]. The burst frequency of a target gene may be significantly increased by enhancers [5]. Therefore, the functional investigation of enhancers will improve our understanding of the transcription regulation mechanism [6].
Enhancers may be detected through in vivo animal experiments. Heintzman and Ren identified novel enhancers through the binding affinities to the transcription factors like P300 [7]. Boyle et al. detected enhancers by investigating the DNaseI hypersensitivity [8]. However, the wet-lab experiments are time-consuming and labor-intensive, and many enhancers cannot be detected in this way due to their condition-specific activities [9].
Various machine learning methods are proposed for the enhancer detection and evaluation problem. Firipi et al. proposed the artificial neural network-based algorithm CSI-ANN to efficiently extract the enhancers’ sequence features and accurately detect novel ones [10]. The tool EnhancerFinder combined multiple learning kernels based on the evolutionary conservation patterns, sequence motifs, and cell type-specific functional information for the detection, and genomic distribution characterization of enhancers [3]. A random forest (RF) model was trained using the chromatin status to construct the enhancer maps in multiple cell types [11]. Deep learning is another powerful tool to detect enhancers, and the tool EnhancedDBN achieved the enhancer detection task using a deep belief network (DBN) [12].
The strength type is another important feature of enhancers. Liu et al. utilized the pseudo-k-set nucleotide composition (PseKNC) algorithm [13] as the sequence features and proposed a two-layer classifier iEnhancer-2L for both the enhancer detection and the enhancer strength type determination [14]. Jia et al. utilized a two-step wrapper feature selection algorithm to find the best features from the useful information of bot bi-profile Bayes and PseKNC, and their model EnhancerPred by 0.01 and 0.12 in the metric Matthews Correlation Coefficient (MCC) for the two layers of iEnhancer-2L [15]. Nguyen et al. combined the one-hot-encoding of and the statistical k-mer descriptors neighboring to each nucleotide, and trained convolutional neural network (CNN) models for the enhancer detection problem (iEnhancer-ECNN) [16]. Their ensembled models demonstrated much improved prediction performance on the independent test dataset. Both the ensembled prediction models and the CNN classifiers are notorious for the high computational requirements. This might be the reason that the leave-one-out validation was not carried out.
This study proposes a novel position-specific encoding algorithm (SeqPose) of nucleotide sequences and selects a subset of the SeqPose features to build the two-layer enhancer classification model. The first layer separates the enhancers from the non-enhancers, and the second layer determines whether an enhancer is strong or weak. The experimental data shows that the position-specific patterns contribute useful information to the accurate enhancer detection and strength evaluation problem.
There are two major contributions of this study. Firstly, the DNA sequence encoding strategy in this study utilizes the location information of each k-mer. In this paper, a novel sequence preprocessing strategy (SeqPose) is proposed. We map the original DNA sequences into numerical sequences by using the encoding rules (SeqPoseDict), and then select the redundant positions with low correlation with sample label by using the Chi-squared test, and remove them from the sequences altogether. In this study, we look at the different positions of the encoded sequence as features, trying to find unimportant positions and delete them from the sample, which was ignored by most of the existing studies. The correlations between different positions in a DNA sequence also shows phenotype associations. This is different from the previous position-specific encoding strategy of DNA sequences, which calculates the one-hot encoding of and the other statistical descriptors neighboring to the nucleotide A/C/G/T in each position and formulates these engineered features in the same orders of their corresponding nucleotides [16].
Secondly, the attention mechanism is widely used in the deep learning-based text classification studies, and this study hypothesizes that the attention mechanism might be able to highlight the nucleotide ‘words’ with large contributions to the enhancer prediction models. Our experimental data in the following sections demonstrates that the proposed SeqPose features achieves similar leave-one-out validation and independent test performances compared with the existing studies. The feature selection strategy and the attention mechanism ensure that the training and prediction of the proposed prediction models may be completed within reasonable time and have the potential to be deployed to the situations with limited computing power.

2. Results and Discussion

This section evaluates the parameters of the proposed model spEnhancer, and then compares spEnhancer with the existing studies on the same datasets. In the first three subsections, we divide 10% of the 2968 training DNA sequences into test sets, then 10% of the remaining data sets into verification sets, and the rest were all used as training sets for training models. All experiments in this study the random number seed 75. The results of the first layer structure of the model on the test set are used as the criteria for parameter selection.

2.1. Evaluating the Length of K-mers

It is anticipated that different length of k-mers makes different SeqPose features and may have large impacts on the final prediction models. This section investigates the binary classification between enhancers and non-enhancers. The three parameters are initially set as pBatchSize = 100, pLSTMSize = 128, and pDropoutRatio = 0.2.
The enhancer detection model of 7-mers achieves the worst prediction accuracy (Acc), as shown in Table 1. The data suggests that the model of 6-mers performed reasonably well on the metric specificity (Sp), but its sensitivity (Sn) is only 0.6338, which is at least 0.0775 worse than the models using the other k-mers. Therefore, 6-mers are excluded from further evaluation. The data suggests that the model using 2-mers performed the best Acc = 0.8047 in Table 1. The following sections focus on the models using 2-mers.

2.2. Selecting the Best SeqPose Features

We hypothesize that some of the SeqPose features had no contributions to the classification problem in this study. We use the Chi2 measurement to evaluate the features and calculate the prediction performances of the binary classification problem between enhancers and non-enhancers on the test dataset after removing some features, as shown in Figure 1.
Firstly, we carry out a procedure of coarse-grain feature selection, through removing five features with the largest p-value measurement in each iteration, as shown in Figure 1a. There are 398 2-mer SeqPose features constructed from each 200-bp nucleotide sequences. The best classification accuracy is Acc = 0.8249 after removing 45 features (“−45”). At the same time, the classification model based on the feature list also obtains the best parameter-independent metric AUC = 0.8986.
Secondly, we carry out a fine-grain feature selection procedure by removing one feature in each iteration, as shown in Figure 1b. The experimental data suggests that the classification model with 45 removed features is not improved with more features being removed. The experimental data of Table 1 and Figure 1 show that 45 of the SeqPose features do not contribute to the enhancer prediction models. The best prediction model in Table 1 was substantially improved for Acc (from 0.8047 to 0.8249), MCC (from 0.5673 to 0.5936), and AUC (from 0.8781 to 0.8986). Therefore, the removal of the redundant positions in the SeqPose encoding features may improve the DNA sequence-based prediction models like the enhancer predictions. The following section will use the 2-mer SeqPose features after the removal of those 45 position-specific features.

2.3. Symbolic Interpretation of SeqPoseDict

This section demonstrates through experimental comparison that SeqPoseDict is only used to convert DNA sequences into computer-recognizable numerical representations, and its selection will not have a decisive impact on the final experimental results (see Table 2). We sequentially: (1) use all data sets to construct SeqPoseDict, see Res1; (2) use 95% of all data sets to construct SeqPoseDict, see Res2; (3) use 90% of all data sets to construct SeqPoseDict, see Res3; (4) use 85% of all data sets to construct SeqPoseDict, see Res4; (5) use 80% of all data sets to construct SeqPoseDict, see Res5.
The data shows that under the framework proposed in this paper, the size of SeqPoseDict does not have a decisive influence on the prediction effect of the model. It can be seen that when the data set used to construct SeqPoseDict fluctuates within 20%, the range of the Acc indicator does not exceed 0.03, which is within the acceptable range.

2.4. Optimizing the Best Choices of the Three Parameters

The three parameters pLSTMSize, pDropoutRatio, and pBatchSize are optimized through eight-fold cross-validation, as shown in Table 3. The overall prediction accuracy Acc is used as a parameter selection goal for the binary classification problem between enhancers and non-enhancers.
The set of pLSTMSize is 64 , 128 , 192 ; the set of pDropoutRatio is 0.1 , 0.2 , 0.3 , 0.4 , 0.5 , 0.6 , 0.7 ; the set of pBatchSize is 16 , 32 , 64 , 128 , 256 , 512 . Every time we take out a set of hyperparameters from these three sets as training parameters, until we get all possible combinations, compare their evaluation indicators on the same test set (See Table 3). The data shows that when the hyperparameter combination is pLSTMSize = 64, pBatchSize = 64, and pDropoutRatio = 0.5, the model training effect is the best (Acc = 0.7884, AUC = 0.8506). Therefore, the following sections are conducted on the basis of this set of hyperparameters.

2.5. Comparing spEnhancer with the Existing Models

This section compares the experimental results of the proposed spEnhancer models with the existing models, as shown in Table 4 and Table 5. The parameters of spEnhancer models are set by the optimization procedure in the above sections. The SeqPose encoding features use 2-mers. The 45 redundant positions are removed according to the above sections. The parameters pLSTMSize, pBatchSize, and pDropoutRatio are set as 64, 64, and 0.5, respectively. The random seed is set to 75. The SeqPoseDict is calculated based on the training dataset. The threshold of the prediction probability uses the default value 0.5.
The performance metrics Sn and Sp measure a binary classification model from different aspects, and one metric may be increased at the cost of the other one [17,18]. The overall prediction performance metrics Acc and MCC may be used for a fair comparison of two classification models [14]. Leave-one-out validation tends to give optimistic results for a classification model and the model generalization on future samples is usually evaluated by the prediction performances on independent datasets that are not involved in the model training process. The same independent dataset is publicly available and used in this study [14].
Table 4 shows that the proposed method spEnhancer achieves similar prediction performances for both of the two binary classification problem using the leave-one-out validation. The spEnhancer model achieves Acc = 0.7793 for the classification problem between enhancers and non-enhancers, which is 0.001 smaller than the best accuracy of the iEnhancer-EL model. Due to the difference in experimental equipment and the influence of the randomness of the parameters, we consider the difference in accuracy to be negligible and acceptable. The spEnhancer model achieved the second-best accuracy Acc = 0.6413, which is slightly smaller than the best model iEnhancer-EL’s Acc = 0.6503.
Table 5 gives the comparative results between the proposed spEnhancer and the four existing state-of-the-art models on the independent test dataset. The prediction performances of the four existing state-of-the-art studies are also retrieved from the previous study. The prediction accuracy on the independent dataset is supposed to represent how well a prediction model may achieve on the future query nucleotide sequences.
The proposed spEnhancer models and another algorithm iEnhancer-ECNN achieve the best two prediction accuracies and MCC on both enhancer detection problems, as shown in Table 5. The spEnhancer model achieves the best prediction Acc = 0.7725 and MCC = 0.5793 for the first layer of the enhancer detection problem, suggesting that spEnhancer may accurately separate enhancers from non-enhancers even on novel query sequences. For the second layer of the enhancer detection problem, spEnhancer achieves Acc = 0.6200, which is slightly worse than the Acc = 0.6780 of the iEnhancer-ECNN model [16]. However, spEnhancer outperforms iEnhancer-ECNN in MCC by an improvement of 0.0023.
The previously best model iEnhancer-ECNN achieves a better strong/weak enhancer prediction accuracy than the proposed spEnhancer with a sacrifice in the training and prediction time. The iEnhancer-ECNN integrates the prediction results of five convolutional neural networks, each of which is trained by 20 epochs [16]. While this study completes the training of the proposed spEnhancer within six epochs. This might be the reason that the iEnhancer-ECNN model does not provide the leave-one-out validation results.
Therefore, the overall prediction performances of spEnhancer are better than or comparable to the four state-of-the-art models. This suggests that spEnhancer may serve well as a complementary model to the existing enhancer detection studies.

2.6. Evaluating the Three-Class Classification Model

The BD-LSTM algorithm may also handle the three-class classification problem. The above sections formulate the enhancer detection problem as a two-layer setting for a consistent comparison with the network structure of Liu et al. [17]. This section collects the three classes of samples and directly trains a three-class BD-LSTM prediction model with the same parameter values for the enhancer detection problem. The model training is on the training dataset and the performance is calculated using the independent test dataset. The three-class BD-LSTM model achieves 0.6402 in the overall prediction accuracy, while the binary classification BD-LSTM model achieves Acc = 0.7725 for separating enhancers from non-enhancers, as shown in Table 5. However, the three-class BD-LSTM model outperforms the binary classification model between strong and weak enhancers.
Therefore, the setting of the two-layer model delivers similar prediction performances as the three-class classification model. In order to carry out a direct comparison with the existing studies, the two-layer model is recommended.

2.7. Evaluating Different Word Vector Dimensions

The dimension size of the word vectors may affect the prediction performances, and this study evaluates different dimensions of the word vectors, as shown in Table 6. Eight values are evaluated for the dimension of the word vector, i.e., 12, 24, 48, 96, 192, 394, 768, and 1536. The values 48, 768, and 1536 achieve the best three Acc and MCC for the binary classification problem between enhancers and non-enhancers. The values 48, 96, 394, and 768 achieve the best four Acc for separating the strong enhancers from the weak ones, but the best MCC = 0.3703 is achieved by the dimension of word vector 768. At the same time, it can be seen that although the Acc with vector dimension equal to 1546 reached the optimal when classifying the enhancers and non-enhancers, the result was only 0.0039 higher than that of the Acc with dimension of 768. However, when classifying strong enhancers and weak enhancers, the Acc with a vector dimension of 768 was higher than the Acc with a vector of 1536, which was 0.04, which was a significant improvement. Therefore, this study sets 768 as the dimension of the word vector.

3. Materials and Methods

This section introduces a two-layer classifier for the detection and strength determination of enhancers. Firstly, the position-specific encoding algorithm (SeqPose) of nucleotide sequences is described. Then the Chi-squared test (Chi2) is used to select a subset of the SeqPose features. Lastly, the selected SeqPose features are loaded to the embedding layer and mapped to a three-dimensional tensor. The two-layer classifier spEnhancer is optimized for the two binary classification problems.

3.1. Datasets and Performance Metrics

This study retrieves the publicly available training and independent test datasets released by the previous study [14]. Three existing state-of-the-art models are evaluated using the same datasets. Therefore, the proposed model spEnhancer is also evaluated on these two datasets. A fair comparison is carried out for the proposed model spEnhancer and the three existing models using the leave-one-out validation on the training dataset and the validation on the independent test dataset. The following describes how the training dataset is generated. The independent test dataset is generated using the same procedure and has no overlapped samples with the training dataset.
This article cites the standard data set S constructed by Prof. Bin Liu [14] based on nine kinds of cell chromosome status information (2968 sequences in total)
S = S s t r o n g + + S w e a k + + S
where,   S s t r o n g + contains only strong enhancer sequences, with a total of 742; S w e a k + contains only weak enhancer sequences, with a total of 742; S contains only non-enhancer sequences, with a total of 1484.
All data consists of a positive sample data set and a negative sample data set. Among them, the enhancer type in the first layer structure is a positive sample, and the non-enhancer type is a negative sample; the strong enhancer type in the second layer structure is a positive sample, and the weak enhancer type is a negative sample.
This study uses the five widely-used classification performance metrics [13], i.e., accuracy (Acc), Matthews correlation coefficient (MCC), sensitivity (Sn), specificity (Sp), and the area under the ROC curve (AUC). Let the samples with the class labels 1 and 0 be positive and negative ones, respectively. The metrics Sn and Sp describe the ratios of correctly predicted positive and negative samples, respectively. Acc is the ratio of the correctly predicted samples. MCC describes the correlation between the real and predicted class labels. AUC is a threshold-independent metric for a classification model, and a good model tends to have a large AUC value [19]. These five metrics are defined in the following formula.
S n = 1 S + / S +
S p = 1 S + / S
A c c = 1 S + + S + / S + S
M C C = 1 S + + S + / S + + S 1 + S + S + / S + 1 + S + S + / S
The notation S + is the number of positive samples, while S + is the number of incorrectly predicted positive samples. S is the number of negative samples, and S + is the number of incorrectly negative samples.

3.2. K-mer Indexing and SeqPose Feature Extraction

This study hypothesizes that the position-specific k-mer patterns may deliver important information to discriminate enhancers from the other nucleotide sequences. A nucleotide sequence is a vector of letters and may be formulated as the one-hot integer vectors. Another feature extraction strategy is to summarize the statistical metrics of k-mers in a nucleotide sequence. All the k-mer instances are usually collected through a sliding window with step size 1, as shown in Figure 2.
The collected list of unique k-mers from all the training sequences is shuffled into a set, which is used as an enhancer dictionary (denoted as SeqPoseDict). Each k-mer has a unique ID (starting from 1) in this dictionary, SeqPoseDict. A query nucleotide sequence is encoded by the same dictionary generated from the training sequences. Note that SeqPoseDict ID is not in the actual meaning. It is used to map the DNA sequence for numerical vectors, we will in the next chapter by the experimental results show that different SeqPoseDict model does not cause large fluctuations, it also illustrates the model of our proposed architecture as a whole is robust. Here we only give one of the methods of constructing SeqPoseDict: according to the order of appearance of k-mer in the training set, assign values from 1 until all k-mers in the training set are traversed. For k-mers that are not in the SeqPoseDict in the unknown data set, we specify that the k-mer ID is 0.

3.3. Selecting the Subset of Best Features

A feature selection step is carried out to remove the positions extracted in the above section if these have low phenotype associations. Chi2 assumes the null hypothesis that under the Chi-square distribution, a given feature is independent of the class label. A statistical p-value is calculated to describe the null hypothesis. A small p-value rejects the null hypothesis and supports that the given position is correlated with the class label. Next, use Chi2 to determine the position label with low correlation with the label on the training set (T) as the redundant feature and remove it from all samples, including the test set.
Each sample is a fixed-length (n) nucleotide sequence and is sliced as multiple k-mers through the sliding window with step size 1. Each k-mer is regarded as a ‘word’. Therefore, each sample can be viewed as a ‘sentence’ composed of multiple ‘words’ (See step 1 in Figure 3).
The k-mer ‘word’ in the extracted feature of a query nucleotide sequence is replaced by its ID in the dictionary SeqPoseDict (See step 2 in Figure 3).
The samples in the training dataset and the labels of these corresponding samples are formed into a matrix M. That is, the first C-1 column of M indicates position-specific feature, the last column is the label column, and each row of M is a sample. The correlation between each feature and the class label is calculated by Chi2. The detailed calculations are described in detail in the following sub-steps.
Take out the ith column and the label column Y in M. We can treat the different IDs in the ith column as different k-mer categories and extend the column vectors into matrix X by one-hot coding, where each row represents the one-hot encoding of IDs, and each column represents the k-mer category. Then do the same for the label variable Y with one-hot coding to make a two-dimensional matrix Y_label, where the rows represent the one-hot code of the sample’s class label, and the column represents the sample’s class label (see Step 3.1 in Figure 3).
Calculate the observed value by formula (6) and record it as vObserved (see Step 3.2 in Figure 3).
v O b s e r v e d = Y T X
Summing each column of the data matrix X makes the variable vFeatureSum. Then the formulas (7) and (8) are used to calculate the proportion of positive (vProbP) and negative (vProbN) samples in the column vector Y_label, respectively. The variables vProbP and vProbN are combined as vProbClass. The theoretical value is calculated according to the formula (9), denoted as vExpected. Please be noted that the calculation uses floating-point values with eight digits after the decimal point. Figure 3 shows only two digits after the decimal point for an easy view (see Step 3.3 in Figure 3).
v P r o b P = N 1 N
v P r o b N = N 0 N
v E x p e c t e d = v P r o b C l a s s T × v F e a t u r e S u m
where N1, N0, and N represent the numbers of the samples in category 1, the samples in category 0 and all the samples, respectively.
Formula (10) calculates the Chi2 value of each k-mer category in a column of the data matrix M. Please be noted that the calculated Chi2 values reach 16 digits after the decimal point. Figure 3 rounds these values for a better view (see Step 3.4 in Figure 3).
Chi 2 = v O b s e r v e d v E x p e c t e d 2 v E x p e c t e d
The corresponding p-value is calculated using the tool Python package scikit-learn version 0.20.3. This paper defines the p-value of each column in the matrix M as the sum of p-values of each k-mer in this column, denoted as vFeaturePvalue (See step 3.5 in Figure 3).
Repeat the steps 3.1–3.5 until all the feature columns are evaluated (See step 4 in Figure 3).
The features are sorted by vFeaturePvalue in the descendent order, and the top K features with the largest vFeaturePvalue are removed for their small correlations with the class labels.

3.4. Vectorization

The features of a sample nucleotide sequence from the above section form a list of consecutive k-mer ‘words’, and this list is regarded as a ‘sentence’. The traditional one-hot coding strategy makes the feature matrix very sparse, so the step of text vectorization is carried out through the word embedding technique. A numeric tensor is generated by the text vectorization of a sample (a nucleotide sequence).
The dimension of the original one-hot coding feature vector of each k-mer is 1 × P , and we project each k-mer into a Q-dimensional word vector by embedding layer. P is also the size of the dictionary SeqPoseDict. This paper uses the word embedding layer to map each k-mer to a word vector of Q dimensions. The recently proposed BERT model performs very well on the text-based classification problems. After evaluating eight values {12, 24, 48, 96, 192, 394, 768, 1536} for the dimension of word vectors, this study uses the same parameter settings of the token embedding layer from the BERTBASE model [20], which assigns Q = 768. The detailed experimental data may be found in the following sections.
The word embedding layer is randomly initialized and updated during the training process of the deep learning classifier. After each word passes through the word embedding layer, it will change from an ID value to a 768-dimensional row vector, as shown in Figure 4. This step generates the SeqPose features from the samples.

3.5. Structural Description of the Classifier SpEnhancer

The two-layer enhancer prediction model spEnhancer is illustrated in Figure 5. All the nucleotide sequences are converted to the SeqPose features. Both of the two layers of the proposed classifier spEnhancer use the bi-directional long-short term memory (BD-LSTM) model combined with attention mechanism as the classifier because that the reverse complementary nature of a DNA molecule. The output of each neuron in the BD-LSTM is calculated from the input of the current neuron and its two neighboring neurons, and then passed as the input of batch normalization. The reason for introducing batch normalization is to prevent the problem of gradient disappearance. In addition, when extracting features, we hope that the model will pay more attention to features with high importance. Therefore, we introduce the attention mechanism in the model to assign weight coefficients to features to indicate the importance of the features, and then linearly combine the weight coefficients with the original input features and output to the next layer. The dropout layer is then utilized to mask a certain proportion of neurons from calculations and may effectively prevent over-fitting. This BD-LSTM layer delivers its output to the dropout layer, and then the full connection layer generates the final prediction results using one neuron with the value 1 or 0. The two layers of the proposed model spEnhancer have the same neural network architecture, as shown in Figure 5. The first layer is trained to predict whether a nucleotide sequence is an enhancer (1) or not (0). The second layer of the proposed model is trained to describe whether an enhancer is a strong (1) or weak (0) one.

3.6. Training Procedure of SpEnhancer

This study implements the deep learning model’s training through the packages keras version 2.2.4 and tensorflow-gpu version 1.14.0 in the Python programming language version 3.7.7. The working environment is equipped with the GPU card Nvidia GeForce RTX 2060. In a multi-CPU server environment, we use 8-fold cross-validation to determine three sets of hyperparameters in turn: batch size (pBatchSize), the number of neural units in the LSTM layer (pLSTMSize), and the dropout ratio (pDropoutRatio). Moreover, the results obtained on the leave-one-out method are compared with existing research.
The dataset is retrieved from the database [17].
The model uses the Adam [21] optimizer to guide the model training process, and the optimization goal is the prediction accuracy on the validation dataset.

4. Conclusions

This proof-of-principle study demonstrates that the SeqPose features generated by the natural language processing (NLP) technologies achieved similar detection performances of enhancers and their enhancing strengths, compared with the existing best models. The experimental data suggests that the genomic sequences may be regarded as the language of lives, and the functional roles of genomic sequences may be investigated through the NLP technologies.
Our experimental data also demonstrates the importance of removing the unassociated positions from training a DNA sequence-based prediction model. The retaining of some positions in the DNA sequences may even reduce the overall model prediction performances. A previous study showed that many deep learning models may be improved by removing the features without contributions to the models [22]. The time-consuming training process of a deep learning model may be sped up by removing the unassociated features.
It is anticipated that more applications of the NLP technologies will be conducted to investigate genomic functional elements.

Author Contributions

Conceptualization, X.M. and F.Z.; methodology, X.M., Y.W., M.D., X.W., K.Z. and F.Z.; software, X.M., M.D., S.L., F.L. and X.W.; validation, X.M., Y.W. and M.D.; investigation, X.M., Y.W., M.D. and L.H; writing—original draft preparation, X.M. and F.Z.; writing—review and editing, X.M., L.H. and F.Z.; funding acquisition, L.H. and F.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Jilin Provincial Key Laboratory of Big Data Intelligent Computing (20180622002JC), the Education Department of Jilin Province (JJKH20180145KJ, JJKH20190091KJ), and the startup grant of the Jilin University. This work was also partially supported by the Bioknow MedAI Institute (BMCPP-2018-001), the High Performance Computing Center of Jilin University, and by the Fundamental Research Funds for the Central Universities, JLU.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The Python source code and data can be obtained from the data folder in the link http://www.healthinformaticslab.org/supp/resources.php, accessed on 23 October 2020.

Acknowledgments

Insightful and constructive comments from the anonymous reviewers are greatly appreciated.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wierzbicki, A.T. The role of long non-coding RNA in transcriptional gene silencing. Curr. Opin. Plant Biol. 2012, 15, 517–522. [Google Scholar] [CrossRef] [PubMed]
  2. Ramji, D.P.; Foka, P. CCAAT/enhancer-binding proteins: Structure, function and regulation. Biochem. J. 2002, 365, 561–575. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Erwin, G.D.; Oksenberg, N.; Truty, R.M.; Kostka, D.; Murphy, K.K.; Ahituv, N.; Pollard, K.S.; Capra, J.A. Integrating Diverse Datasets Improves Developmental Enhancer Prediction. PLoS Comput. Biol. 2014, 10, e1003677. [Google Scholar] [CrossRef] [Green Version]
  4. Gillies, S.D.; Morrison, S.L.; Oi, V.T.; Tonegawa, S. A tissue-specific transcription enhancer element is located in the major intron of a rearranged immunoglobulin heavy chain gene. Cell 1983, 33, 717–728. [Google Scholar] [CrossRef]
  5. Larsson, A.J.M.; Johnsson, P.; Hagemann-Jensen, M.; Hartmanis, L.; Faridani, O.R.; Reinius, B.; Segerstolpe, Å.; Rivera, C.M.; Ren, B.; Sandberg, R. Genomic encoding of transcriptional burst kinetics. Nat. Cell Biol. 2019, 565, 251–254. [Google Scholar] [CrossRef] [PubMed]
  6. Kim, T.-K.; Hemberg, M.; Gray, J.M.; Costa, A.M.; Bear, D.M.; Wu, J.; Harmin, D.A.; Laptewicz, M.; Barbara-Haley, K.; Kuersten, S.; et al. Widespread transcription at neuronal activity-regulated enhancers. Nat. Cell Biol. 2010, 465, 182–187. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Heintzman, N.D.; Ren, B. Finding distal regulatory elements in the human genome. Curr. Opin. Genet. Dev. 2009, 19, 541–549. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Boyle, A.P.; Song, L.; Lee, B.-K.; London, D.; Keefe, D.; Birney, E.; Iyer, V.R.; Crawford, G.E.; Furey, T.S. High-resolution genome-wide in vivo footprinting of diverse transcription factors in human cells. Genome Res. 2010, 21, 456–464. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  9. Davis, H.L.; Weeratna, R.; Waldschmidt, T.J.; Tygrett, L.; Schorr, J.; Krieg, A.M. CpG DNA is a potent enhancer of specific immunity in mice immunized with recombinant hepatitis B surface antigen. J. Immunol. 1998, 160, 870–876. [Google Scholar] [PubMed]
  10. Firpi, H.A.; Ucar, D.; Tan, K. Discover regulatory DNA elements using chromatin signatures and artificial neural network. Bioinformatics 2010, 26, 1579–1586. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  11. Rajagopal, N.; Xie, W.; Li, Y.; Wagner, U.; Wang, W.; Stamatoyannopoulos, J.; Ernst, J.; Kellis, M.; Ren, B. RFECS: A Random-Forest Based Algorithm for Enhancer Identification from Chromatin State. PLoS Comput. Biol. 2013, 9, e1002968. [Google Scholar] [CrossRef] [PubMed]
  12. Bu, H.; Gan, Y.; Wang, Y.; Zhou, S.; Guan, J. A new method for enhancer prediction based on deep belief network. BMC Bioinform. 2017, 18, 418. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Chen, J.; Liu, H.; Yang, J.; Chou, K.-C. Prediction of linear B-cell epitopes using amino acid pair antigenicity scale. Amino Acids 2007, 33, 423–428. [Google Scholar] [CrossRef] [PubMed]
  14. Liu, B.; Fang, L.; Longyun, F.; Lan, X.; Chou, K.-C. iEnhancer-2L: A two-layer predictor for identifying enhancers and their strength by pseudok-tuple nucleotide composition. Bioinformatics 2016, 32, 362–369. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Jia, C.; He, W. EnhancerPred: A predictor for discovering enhancers based on the combination and selection of multiple features. Sci. Rep. 2016, 6, 38741. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Nguyen, Q.H.; Nguyen-Vo, T.-H.; Le, N.Q.K.; Do, T.T.; Rahardja, S.; Nguyen, B.P. iEnhancer-ECNN: Identifying enhancers and their strength using ensembles of convolutional neural networks. BMC Genom. 2019, 20, 951. [Google Scholar] [CrossRef] [PubMed]
  17. Liu, B.; Li, K.; Huang, D.-S.; Chou, K.-C. iEnhancer-EL: Identifying enhancers and their strength with ensemble learning approach. Bioinformatics 2018, 34, 3835–3842. [Google Scholar] [CrossRef] [PubMed]
  18. Chou, K. A vectorized sequence-coupling model for predicting HIV protease cleavage sites in proteins. J. Biol. Chem. 1993, 268, 16938–16948. [Google Scholar] [CrossRef]
  19. Fawcett, T. An introduction to ROC analysis. Pattern Recognit. Lett. 2006, 27, 861–874. [Google Scholar] [CrossRef]
  20. Devlin, J.; Chang, M.-W.; Lee, K.; Toutanova, K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the North American Association for Computational Linguistics-Human Language Technologies 2019, Minneapolis, MN, USA, 2–7 June 2019; pp. 4171–4186. [Google Scholar]
  21. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. In Proceedings of the International Conference on Learnning Representations (ICLR), San Diego, CA, USA, 5–8 May 2015. [Google Scholar]
  22. Chen, Z.; Pang, M.; Zhao, Z.; Li, S.; Miao, R.; Zhang, Y.; Feng, X.; Feng, X.; Zhang, Y.; Duan, M.; et al. Feature selection may improve deep neural networks for the bioinformatics problems. Bioinformatics 2019, 36, 1542–1552. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Performance evaluation of the models using the 2-mer SeqPose features after filtering out some redundant features. The vertical axis gives the values of the five performance metrics, including Sn, Sp, MCC, AUC, and Acc. The horizontal axis is the number of features removed from the 2-mer SeqPose features, i.e., “−5” means that the five features with the largest p-values are removed from the list of all 2-mer SeqPose features. The feature removal is carried out in two steps, i.e., (a) five features are removed in each iteration; and then (b) one feature is removed in each iteration from the list of features “−45”.
Figure 1. Performance evaluation of the models using the 2-mer SeqPose features after filtering out some redundant features. The vertical axis gives the values of the five performance metrics, including Sn, Sp, MCC, AUC, and Acc. The horizontal axis is the number of features removed from the 2-mer SeqPose features, i.e., “−5” means that the five features with the largest p-values are removed from the list of all 2-mer SeqPose features. The feature removal is carried out in two steps, i.e., (a) five features are removed in each iteration; and then (b) one feature is removed in each iteration from the list of features “−45”.
Ijms 22 03079 g001
Figure 2. Extracted SeqPose features of a nucleotide sequence. A, T, G, C represent the bases in the nucleotide sequence.
Figure 2. Extracted SeqPose features of a nucleotide sequence. A, T, G, C represent the bases in the nucleotide sequence.
Ijms 22 03079 g002
Figure 3. Flow chart of Chi2-based SeqPose feature selection.
Figure 3. Flow chart of Chi2-based SeqPose feature selection.
Ijms 22 03079 g003
Figure 4. Illustration of the nucleotide sequence vectorization process.
Figure 4. Illustration of the nucleotide sequence vectorization process.
Ijms 22 03079 g004
Figure 5. Workflow of the proposed two-layer model spEnhancer. A schematic diagram depicting the steps of feature extraction, feature selection, and prediction.
Figure 5. Workflow of the proposed two-layer model spEnhancer. A schematic diagram depicting the steps of feature extraction, feature selection, and prediction.
Ijms 22 03079 g005
Table 1. Performance comparison of different k-mers on the test dataset. The performance of individual k-mer setting.
Table 1. Performance comparison of different k-mers on the test dataset. The performance of individual k-mer setting.
k-merAccSnSpMCCAUC
1-mer0.78110.71830.83870.53190.8783
2-mer0.80470.72540.87740.56730.8781
3-mer0.79120.70420.87100.53830.8693
4-mer0.80130.71130.88390.55540.8717
5-mer0.76770.78870.74840.55340.8637
6-mer0.77780.63380.90970.48700.8652
7-mer0.73060.69720.76130.44980.8322
Table 2. Performance of different SeqPoseDict on the same test set. The first row gives the values of the five performance metrics—including Sn, Sp, MCC, AUC, and Acc. The first column is different SeqPoseDict.
Table 2. Performance of different SeqPoseDict on the same test set. The first row gives the values of the five performance metrics—including Sn, Sp, MCC, AUC, and Acc. The first column is different SeqPoseDict.
AccSnSpMCCAUC
Res10.79460.66200.91610.52210.8827
Res20.79120.80280.78060.59510.8797
Res30.81140.73940.87740.87740.8920
Res40.78450.80990.76130.59050.8827
Res50.78450.83800.73550.61000.8885
Table 3. Performance of different hyperparameters on the test dataset. ‘Upper left corner’: When pBatchSize = 16, traverse the collection of pDropoutRatio and pLSTMSize. ‘Upper right corner’: When pBatchSize = 32, traverse the collection of pDropoutRatio and pLSTMSize. ‘Middle left’: When pBatchSize = 64, traverse the collection of pDropoutRatio and pLSTMSize Collection. ‘Middle right’: When pBatchSize = 128, traverse the collections of pDropoutRatio and pLSTMSize. ‘Lower left corner’: When pBatchSize = 256, traverse the collections of pDropoutRatio and pLSTMSize. ‘Bottom right corner’: When pBatchSize = 512, traverse the collections of pDropoutRatio and pLSTMSize. In each subgraph, the first row gives the values of the five prediction performance metrics—i.e., Sn, Sp, MCC, Acc, and AUC. The first column gives the values of each pair pLSTMSize and pDropoutRatio, separated by a comma.
Table 3. Performance of different hyperparameters on the test dataset. ‘Upper left corner’: When pBatchSize = 16, traverse the collection of pDropoutRatio and pLSTMSize. ‘Upper right corner’: When pBatchSize = 32, traverse the collection of pDropoutRatio and pLSTMSize. ‘Middle left’: When pBatchSize = 64, traverse the collection of pDropoutRatio and pLSTMSize Collection. ‘Middle right’: When pBatchSize = 128, traverse the collections of pDropoutRatio and pLSTMSize. ‘Lower left corner’: When pBatchSize = 256, traverse the collections of pDropoutRatio and pLSTMSize. ‘Bottom right corner’: When pBatchSize = 512, traverse the collections of pDropoutRatio and pLSTMSize. In each subgraph, the first row gives the values of the five prediction performance metrics—i.e., Sn, Sp, MCC, Acc, and AUC. The first column gives the values of each pair pLSTMSize and pDropoutRatio, separated by a comma.
SnSpMCCAccAUC SnSpMCCAccAUC
64,0.10.71800.82190.51610.76920.844164,0.10.70570.84290.51730.77290.8476
64,0.20.74890.80320.53840.77590.849664,0.20.74680.79930.53270.77290.8471
64,0.30.71350.83950.52240.77730.851664,0.30.73010.78950.50610.75710.8484
64,0.40.73090.82070.53070.77560.850464,0.40.71450.83170.51720.77120.8508
64,0.50.74080.79760.52520.76850.847564,0.50.70010.85970.52050.78000.8430
64,0.60.73930.81020.53070.77430.848564,0.60.72420.80750.51290.76520.8519
64,0.70.73190.82860.53660.78100.851464,0.70.72630.82540.52820.77560.8478
128,0.10.67460.86020.49080.76620.8427128,0.10.72950.81470.52500.77290.8433
128,0.20.72560.77580.48940.75070.8347128,0.20.67430.85890.49700.76420.8459
128,0.30.69190.84260.49830.76650.8441128,0.30.68940.83040.49460.76420.8361
128,0.40.68540.85340.50190.77090.8479128,0.40.76200.75700.52190.75710.8418
128,0.50.69470.83370.49820.76380.8438128,0.50.73220.79870.51870.76550.8452
128,0.60.72890.80280.51540.76550.8441128,0.60.75390.78200.53020.76820.8384
128,0.70.68660.85090.49900.76720.8373128,0.70.72550.78940.50150.75510.8416
192,0.10.69690.77450.46740.73850.8364192,0.10.68250.82200.47530.75200.8276
192,0.20.68680.82280.48210.75370.8291192,0.20.64800.87390.47520.76180.8364
192,0.30.70470.80980.49130.75740.8357192,0.30.72500.82370.52550.77560.8467
192,0.40.64060.85550.45470.74700.8329192,0.40.64300.84820.45760.74390.8352
192,0.50.62860.85000.44270.73580.8146192,0.50.68540.85270.50000.76920.8484
192,0.60.75670.78000.53120.76950.8420192,0.60.74680.76040.51050.75640.8383
192,0.70.70890.83220.51860.77460.8438192,0.70.73730.76080.49240.74560.8377
SnSpMCCAccAUC SnSpMCCAccAUC
64,0.10.71990.82500.51850.77090.849764,0.10.71520.83720.52330.77730.8527
64,0.20.73770.83930.55090.78770.856764,0.20.71910.82760.51880.77290.8507
64,0.30.74270.81060.53580.77560.852164,0.30.68680.86180.50740.77320.8530
64,0.40.68960.84440.49870.76650.846464,0.40.67220.82880.46640.74830.8316
64,0.50.74030.83610.55210.78840.850664,0.50.71520.82330.51330.76890.8428
64,0.60.68980.84110.49680.76310.850764,0.60.69950.81540.48740.75770.8401
64,0.70.72650.80490.51360.76420.848664,0.70.71360.81910.50810.76650.8497
128,0.10.69090.83900.49460.76480.8340128,0.10.69200.84640.50290.77090.8466
128,0.20.72600.81660.52380.77220.8479128,0.20.69730.84280.50800.76920.8469
128,0.30.67540.82990.47640.75440.8393128,0.30.67190.84150.47550.75470.8447
128,0.40.67650.84090.48630.75400.8443128,0.40.69160.81680.48580.75340.8444
128,0.50.70600.83000.50750.76750.8470128,0.50.69940.83430.50480.76680.8392
128,0.60.68520.86110.50640.77320.8459128,0.60.70850.83660.51520.77120.8498
128,0.70.69180.85070.50910.77020.8430128,0.70.69620.81890.49290.75980.8518
192,0.10.67670.84620.48600.76310.8378192,0.10.65290.87870.48240.76790.8488
192,0.20.68480.83930.49390.76180.8480192,0.20.67930.84450.48840.76250.8345
192,0.30.66540.85380.47680.75840.8283192,0.30.73500.81560.53250.77490.8528
192,0.40.69890.83490.50500.76580.8437192,0.40.65920.86230.47810.75940.8444
192,0.50.68120.84990.49330.76380.8472192,0.50.60460.89260.44390.74800.8441
192,0.60.68010.83740.48360.75740.8313192,0.60.70900.79490.49440.74800.8405
192,0.70.70440.83050.50650.76650.8422192,0.70.71770.82100.51450.76750.8467
SnSpMCCAccAUC SnSpMCCAccAUC
64,0.10.68570.83430.49100.75980.836364,0.10.68570.79480.45810.74330.8242
64,0.20.69790.84120.50520.77020.841064,0.20.65140.84880.45910.74830.8217
64,0.30.69230.84830.50270.77060.842264,0.30.66140.80050.43190.73380.8102
64,0.40.68400.85240.49750.76820.843664,0.40.67810.78350.43980.73150.8109
64,0.50.73040.81480.52520.77360.846064,0.50.67050.82060.45930.74560.8203
64,0.60.71130.81470.50400.76180.843364,0.60.64900.77620.39860.71130.8037
64,0.70.71320.84210.52450.77460.853864,0.70.62530.82110.40910.72170.8011
128,0.10.69820.84510.51230.77020.8494128,0.10.66010.84690.47010.75340.8383
128,0.20.71240.81490.50690.76750.8423128,0.20.64120.87300.46550.75640.8237
128,0.30.66000.87390.48690.76890.8400128,0.30.67700.82920.47360.75270.8303
128,0.40.64070.87990.47080.75810.8456128,0.40.70360.83610.50970.76750.8444
128,0.50.67260.84940.48570.75810.8469128,0.50.72530.80130.51140.76210.8374
128,0.60.69950.81900.49000.75910.8343128,0.60.69670.82300.48870.76010.8281
128,0.70.71940.82760.52280.77320.8418128,0.70.66800.86680.49120.76790.8455
192,0.10.72080.79460.49640.75670.8375192,0.10.71180.80420.49610.75640.8404
192,0.20.67940.80200.45560.74190.8239192,0.20.70250.79040.47190.74730.8363
192,0.30.68930.78190.45260.73550.8287192,0.30.69490.80850.47790.75200.8358
192,0.40.66510.84350.47150.75370.8392192,0.40.71630.79010.49080.75240.8314
192,0.50.69420.81220.48360.75270.8317192,0.50.68930.83210.48850.76210.8426
192,0.60.68470.81960.47390.74760.8278192,0.60.69700.80090.47500.75070.8334
192,0.70.68400.83090.48160.75670.8436192,0.70.66880.85320.48080.76250.8395
Table 4. Comparison of the results of the leave-one-out method on the training dataset between spEnhancer and the three existing methods.
Table 4. Comparison of the results of the leave-one-out method on the training dataset between spEnhancer and the three existing methods.
MethodsAccSnSpMCCAUC
enhancers
vs
non-enhancers
spEnhancer0.77930.70820.85040.52270.8468
iEnhancer-EL0.78030.75670.80390.56130.8547
iEnhancer-2L0.76890.78090.75880.54000.8500
EnhancerPred0.73180.72570.73790.46360.8082
strong enhancers
vs
weak enhancers
spEnhancer0.64130.85030.30520.21050.6148
iEnhancer-EL0.65030.69000.61050.31490.6957
iEnhancer-2L0.61930.62210.61820.24000.6600
EnhancerPred0.62060.62670.61460.24130.6601
Table 5. Comparison of results between spEnhancer and the three existing methods on the independent test dataset.
Table 5. Comparison of results between spEnhancer and the three existing methods on the independent test dataset.
MethodsAccSnSpMCCAUC
enhancers
vs
non-enhancers
spEnhancer0.77250.83000.71500.57930.8235
iEnhancer-ECNN0.76900.78500.75200.53700.8320
iEnhancer-EL0.74750.71000.78500.49640.8173
iEnhancer-2L0.73000.71000.75000.46040.8062
EnhancerPred0.74000.73500.74500.48000.8013
strong enhancers
vs
weak enhancers
spEnhancer0.62000.91000.33000.37030.6253
iEnhancer-ECNN0.67800.79100.74800.36800.7480
iEnhancer-EL0.61000.54000.68000.22220.6801
iEnhancer-2L0.60500.47000.74000.21810.6678
EnhancerPred0.55000.45000.65000.10210.5790
Table 6. Comparison of the spEnhancer model performances using different dimensions of word vectors on the independent test dataset. The first column gives the binary classification problem. The column “WV” is the dimension of the word vector. The other five columns give the prediction performances Acc, Sn, Sp, MCC, and AUC.
Table 6. Comparison of the spEnhancer model performances using different dimensions of word vectors on the independent test dataset. The first column gives the binary classification problem. The column “WV” is the dimension of the word vector. The other five columns give the prediction performances Acc, Sn, Sp, MCC, and AUC.
Word Vector DimensionAccSnSpMCCAUC
enhancers
vs.
non-enhancers
120.70850.85500.56060.49430.8177
240.66580.91500.41410.46550.8094
480.75380.81500.69190.54080.8167
960.70600.86500.54550.49710.7359
1920.71860.76500.67170.45800.8078
3940.73370.84000.62630.52530.8172
768 (this study)0.77250.83000.71500.57930.8235
15360.77640.75500.79800.54080.8281
strong enhancers
vs.
weak enhancers
120.55500.43000.68000.09840.6351
240.50001.00000.00000.00000.6324
480.60000.71000.49000.22650.6342
960.62500.80000.45000.31010.6275
1920.57000.67000.47000.15650.6279
3940.59500.71000.48000.21650.5987
768 (this study)0.62000.91000.33000.37030.6253
15360.58000.87000.29000.24690.5972
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Mu, X.; Wang, Y.; Duan, M.; Liu, S.; Li, F.; Wang, X.; Zhang, K.; Huang, L.; Zhou, F. A Novel Position-Specific Encoding Algorithm (SeqPose) of Nucleotide Sequences and Its Application for Detecting Enhancers. Int. J. Mol. Sci. 2021, 22, 3079. https://doi.org/10.3390/ijms22063079

AMA Style

Mu X, Wang Y, Duan M, Liu S, Li F, Wang X, Zhang K, Huang L, Zhou F. A Novel Position-Specific Encoding Algorithm (SeqPose) of Nucleotide Sequences and Its Application for Detecting Enhancers. International Journal of Molecular Sciences. 2021; 22(6):3079. https://doi.org/10.3390/ijms22063079

Chicago/Turabian Style

Mu, Xuechen, Yueying Wang, Meiyu Duan, Shuai Liu, Fei Li, Xiuli Wang, Kai Zhang, Lan Huang, and Fengfeng Zhou. 2021. "A Novel Position-Specific Encoding Algorithm (SeqPose) of Nucleotide Sequences and Its Application for Detecting Enhancers" International Journal of Molecular Sciences 22, no. 6: 3079. https://doi.org/10.3390/ijms22063079

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop