Next Article in Journal
Controversies and Perspectives in the Current Management of Patients with Locally Advanced Rectal Cancer—A Systematic Review
Previous Article in Journal
Dual-Task Gait Analysis: Combined Cognitive–Motor Demands Most Severely Impact Walking Patterns and Joint Kinematics
Previous Article in Special Issue
Inherited Hypertrabeculation? Genetic and Clinical Insights in Blood Relatives of Genetically Affected Left Ventricular Excessive Trabeculation Patients
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

EnsembleNPPred: A Robust Approach to Neuropeptide Prediction and Recognition Using Ensemble Machine Learning and Deep Learning Methods

by
Supatcha Lertampaiporn
1,
Warin Wattanapornprom
2,
Chinae Thammarongtham
1 and
Apiradee Hongsthong
1,*
1
Biochemical Engineering and Systems Biology Research Group, National Center for Genetic Engineering and Biotechnology, National Science and Technology Development Agency at King Mongkut’s University of Technology Thonburi, Bangkok 10150, Thailand
2
Applied Computer Science Program, Department of Mathematics, Faculty of Science, King Mongkut’s University of Technology Thonburi, Bangkok 10150, Thailand
*
Author to whom correspondence should be addressed.
Life 2025, 15(7), 1010; https://doi.org/10.3390/life15071010
Submission received: 21 May 2025 / Revised: 20 June 2025 / Accepted: 23 June 2025 / Published: 25 June 2025

Abstract

Neuropeptides (NPs) are a diverse group of signaling molecules involved in regulating key physiological processes such as pain perception, stress response, mood, appetite, and circadian rhythms. Acting as neurotransmitters, neuromodulators, or neurohormones, they play a critical role in modulating and fine-tuning neural signaling networks. Despite their biological significance, identifying NPs through experimental techniques remains time-consuming and resource-intensive. To support this effort, computational prediction tools have emerged as a cost-effective approach for prioritizing candidate sequences for experimental validation. In this study, we propose EnsembleNPPred, an ensemble learning framework that integrates traditional machine learning (ML) models with a deep learning (DL) component. By combining the complementary strengths of these approaches, the model aims to improve generalization and predictive robustness. EnsembleNPPred employs a majority voting mechanism to aggregate the outputs from three classifiers: Support Vector Machine (SVM), Extra Trees (ET), and a CNN-based DL model. When evaluated on independent datasets, EnsembleNPPred demonstrated consistently competitive performance, achieving improvements in both accuracy and sensitivity-specificity balance compared to several existing methods. Furthermore, testing on multiple neuropeptide families from the NeuroPep database yielded an average accuracy of 91.92%, suggesting the model’s potential to generalize across diverse peptide classes. These results suggest that EnsembleNPPred may be a useful tool for early-stage neuropeptide candidate identification and for supporting downstream experimental validation.

1. Introduction

Neuropeptides are intercellular signaling molecules that function as neurotransmitters, neuromodulators, or neurohormones, playing crucial roles in regulating physiological processes such as pain perception, stress response, mood regulation, appetite control, and circadian rhythm maintenance [1,2]. By modulating the release of classical neurotransmitters and interacting with specific receptors—primarily G protein-coupled receptors (GPCRs)—they fine-tune neural activity across diverse systems [3,4,5]. Typically derived from larger precursor proteins (prepropeptides), neuropeptides undergo proteolytic processing and post-translational modifications within the regulated secretory pathway of neurons. These peptides are stored in dense-core vesicles and released in a stimulus-dependent manner. A single precursor can encode multiple copies of the same neuropeptide or distinct neuropeptides with different biological actions [4,5,6]. Notably, secreted peptides (e.g., hormones, growth factors) may share sequence similarity but are not considered classical neuropeptides unless they meet key criteria: neuronal origin, regulated release, and receptor-mediated neural signaling [2,4].
Neuropeptides have been identified in both vertebrates and invertebrates. Well-known examples include oxytocin, vasopressin, orexins, bradykinin, neuropeptide Y, tachykinin, opioids, somatostatin, angiotensin, calcitonin, gastrin, and galanin, which regulate functions such as growth, glucose homeostasis, inflammation, stress, reproduction, and memory [7,8]. Dysregulation of neuropeptides has been implicated in several diseases, including Alzheimer’s disease, Parkinson’s disease, epilepsy, diabetes, hypertension, and dermatological disorders [9]. Recent studies have also identified dual-function neuropeptides that exhibit both canonical signaling roles and antimicrobial activity, often linked to immune modulation [10,11]. These peptides participate in host immunity through bidirectional communication between the brain and gut, frequently involving microbiome interactions under biotic- and/or abiotic stress conditions [12]. In addition to their conserved functional domains, many of these peptides contain N-terminal signal sequences or domains that facilitate secretion and cellular targeting [13,14].
With the advancement of high-throughput sequencing technologies, the number of sequenced genomes has expanded rapidly, offering new opportunities to identify putative neuropeptides across diverse species. Although mass spectrometry remains the gold standard for neuropeptide characterization, it is labor-intensive and resource demanding. In this context, computational prediction methods offer an efficient and cost-effective complement, facilitating high-throughput candidate screening and accelerating the pace of neuropeptide discovery.
The availability of curated resources such as NeuroPep [15], which has expanded from 5949 peptides in 2015 to over 11,000 entries in its latest release [16], has facilitated the development of machine learning (ML) and deep learning (DL) models for neuropeptide prediction. Previous methods, including NeuroPIpred [17], PredNeuroP [18], NeuroPpred-Fuse [19], NeuroPred-FRL [20], NeuroPred-CLQ [21], and NeuroPred-PLM [22], have introduced innovative strategies in feature engineering and learning frameworks, contributing significantly to computational neuropeptide discovery. These models have laid a strong foundation, though challenges remain, particularly in improving generalizability and minimizing false positives across diverse peptide families.
Recent studies highlight the growing integration of deep learning (DL), emphasizing its ability to capture complex sequential dependencies and hierarchical representations that traditional ML may overlook. DL models such as convolutional neural networks (CNNs), recurrent neural networks (RNNs), long short-term memory (LSTM), and bidirectional LSTM (BiLSTM) have been increasingly utilized with minimal manual feature engineering [23,24,25,26,27]. Additionally, hybrid approaches that combine DL with traditional ML techniques have emerged, leveraging both the abstraction capability of deep models and the interpretability of classical algorithms [23,26,27].
In this study, we propose EnsembleNPPred, a hybrid neuropeptide prediction framework that integrates interpretable handcrafted ML features with CNN-derived representations from Word2Vec-encoded peptide sequences. By combining the complementary strengths of handcrafted features and automatically learned deep embeddings, the ensemble model aims to improve both predictive accuracy and interpretability. The ensemble strategy also enhances generalization, improves robustness to feature variability, and supports flexible deployment across different computational environments by selectively utilizing either ML or DL components. The integration may also aid in prioritizing candidates for downstream experimental validation.
Evaluation on two independent test sets demonstrates that EnsembleNPPred consistently achieves strong predictive performance in terms of accuracy, Matthews correlation coefficient (MCC), and AUC, while maintaining competitive sensitivity and specificity. Additionally, performance on a third dataset encompassing a wide range of neuropeptide families yielded an overall accuracy of 91.92%, underscoring its potential for broad applicability in neuropeptide prediction.

2. Materials and Methods

The key steps in the proposed EnsembleNPPred workflow are illustrated in Figure 1 and summarized as follows. (1) Data collecting involves assembling both training and testing datasets. (2) Feature extraction and selection are then performed to generate representative features suitable for machine learning. (3) A range of machine learning (ML) models, including Extra Trees (ET), Random Forest (RF), Decision Trees (DT), Support Vector Machine (SVM), k-Nearest Neighbors (KNN), and XGBoost (XGB), are trained and evaluated using 10-fold cross-validation to identify the top-performing models. (4) In parallel, a Word2Vec embedding approach is applied to encode peptide sequences for deep learning. (5) A convolutional neural network (CNN) model is constructed, trained, and fine-tuned for optimal performance. (6) Finally, an ensemble model is developed by combining the predictions from the selected ML models and the CNN model using a voting mechanism. The final prediction output is generated by voting strategy. This integrative pipeline leverages both traditional ML and DL techniques to enhance predictive performance for neuropeptide classification. The complete workflow is illustrated in Figure 1.

2.1. Dataset Preparation

Collecting a high-quality dataset is a crucial initial step in developing a reliable machine learning model. In this study, we utilized datasets from two sources: NeuroPred-CLQ [21] and NeuroPred-PLM [22]. To construct our training data, we adopted a strategy to minimize redundancy across datasets. As illustrated in Step 1 of the EnsembleNPPred workflow (Figure 1), we defined the training datasets from NeuroPred-CLQ and NeuroPred-PLM as D1 and D2, respectively, and their corresponding testing sets as T1 and T2. Since D1 may contain sequences also present in T2, and D2 may include sequences from T1, we defined the full dataset as the union:
D = D1D2
T = T1T2
The final training dataset was obtained by removing all sequences in the testing sets from the union of the training datasets, formally represented as:
CombinedTrainingSet = D\T = (D1D2)\(T1T2)
where “\” denotes set subtraction, ensuring that overlapping sequences between training and testing sets were excluded to prevent data leakage.
To further reduce redundancy, we applied CD-HIT with a 90% sequence identity threshold. This clustered highly similar sequences and retained only representatives with less than or equal to 90% identity, resulting in a non-redundant training dataset denoted as:
FilteredTrainingSet = CD-HIT90%(CombinedTrainingSet)
For model evaluation, we employed two independent test sets. Testing dataset 1, sourced from NeuroPred-CLQ [21], comprises 485 NPs and 485 non-NPs, and has been widely adopted in prior studies, enabling direct performance comparisons. Testing dataset 2, derived from NeuroPred-PLM [22], includes 444 unique NPs and 444 unique non-NPs.
To ensure independence of training and testing datasets and mitigate the risk of data leakage, we conducted BLASTP searches using the training set as the database and each test set as the query, applying a stringent E-value cutoff of 0.00001. For the NeuroPred-PLM test set, only 17.1% of queries returned significant hits (identity scores: 41.2–89.7%), while 82.9% showed no detectable similarity. Similarly, for the NeuroPred-CLQ test set, 27.8% had matches (identity scores: 36.8–89.9%), and 72.2% returned no significant hits. Importantly, no test sequences exhibited over 90% identity with any training sequence, validating the non-redundancy and independence of the datasets.
A detailed analysis of sequence length and physicochemical properties of the training dataset is presented in Section 3.1 (Amino acid composition and positional residue analysis). All datasets used in this work are publicly available at: http://www.ncrna-pred.com/EnsembleNPPred.htm (accessed on 24 May 2025) (http://www.ncrna-pred.com/Data.tar.xz (accessed on 24 May 2025)).

2.2. Feature Extraction and Feature Engineering

Various numerical representation schemes were employed to characterize peptide sequences, resulting in a feature vector comprising 982 numerical descriptors. These features are categorized into eight main types as follows:
(1)
AAC descriptors: These represent the relative abundance of each amino acid type in a protein sequence. The proportions of all 20 standard amino acids [A, C, D, E, F, G, H, I, K, L, M, N, P, Q, R, S, T, V, W, Y] are calculated (AAC1-AAC20).
(2)
Chou’s pseudo amino acid composition (PseAAC): PseAAC converts protein sequences of varying lengths into fixed-length numerical feature vectors, incorporating sequence-order information. Unlike AAC, PseAAC captures more detailed information, making it suitable for various sequence-based prediction tasks [28,29]. In this study, PseAAC was computed with parameters λ = 3 and weight = 0.05, resulting in 23 dimensions (PAAC1-PAAC23). Additional PseAAC variants were also calculated: parallel correlations (PsePC1-PsePC22), series correlations (PseSC1-PseSC26), and amphiphilic pseudo AACs based on hydrophobicity (APAAC1-APAAC23) and hydrophilicity correlation functions (APAAC24-APAAC46).
(3)
CTD descriptors: These are derived from grouped amino acid compositions [30,31]. These include composition descriptors (CTDC1-CTDC21), transition descriptors (CTDT1-CTDT21), and distribution descriptors (CTDD1-CTDD105). These descriptors were calculated using the protr R package (1.7.4) [32], with amino acids classified into three groups based on seven physicochemical properties: normalized van der Waals volume, charge, hydrophobicity, polarity, secondary structure, and solvent accessibility.
(4)
Quasi-sequence-order descriptors: These are based on the distance matrix of the 20 amino acids [33] and include sequence-order-coupling numbers (SOCN1-SOCN6) and quasi-sequence-order descriptors (QSO1-QSO46), computed with lag = 3 and weight = 0.1.
(5)
Physicochemical and topological property-related features: These encompass the Crucian properties covariance index (Crucian1–Crucian3) [34], Z-scales (zscales1–zscales5) [35], factor analysis scales of generalized amino acid information (fasgai1–fasgai6) [36], T-scales (tScales1–tScales5) [37], VHSE-scales (vhsescales1–vhsescales8) [38], protFPs (protFP1–protFP8) [39], ST-scales (stscales1–stscales8) [40], MS-WHIM scores (mswhimscore1–mswhimscore3) [41], the aliphatic index (aIndex) [42], Geary autocorrelations (geary1–geary12), autocovariance index (autocov) [42], potential protein interaction index (Boman) [43], cross-covariance indices (Crosscov1–Crosscov2), net charge (Charge), instability index (Instaindex) [44], hydrophobic moment for alpha helices (Hmoment1), hydrophobic moment for beta sheets (Hmoment2), BLOSUM matrix-derived descriptors (Blosum1–8), and isoelectric point (pI) calculated using the peptide R package [45].
(6)
Occurrence of 2-mer and selected 3-mer motifs: Initially, all possible 2-mers (400 dimensions) were generated and retained. Then, 3-mers (8000 dimensions) were generated, and only those significantly different between positive and negative data, as determined by log-odds and MERCI [46] scores, were selected. The selected 3-mer motifs include: ALP, DFI, DTD, ENL, ETI, FLP, FYP, GLQ, GPF, HLP, HPF, IAW, IFP, IKW, IPA, IPP, IYP, KDQ, KRI, KVL, LAV, LHL, LLE, LMR, MFL, NPC, NVP, NWN, PAG, PEV, PFP, PGA, PIP, PIT, PKH, PLP, PSE, PTH, PVP, PYP, QTP, RLN, RND, STC, TKE, TLE, TLV, TST, VKE, VLP, VPP, VPQ, VRP, VYP, WLP, YNP, and YST motifs.
(7)
Secondary structure conformation features: The propensities for aggregation, amyloid formation, turns, alpha-helices, helical aggregation, and beta-strand structures were calculated using the Tango program [47] (tango1-tango6).
(8)
Composite features for neuropeptides: To enhance prediction with more informative features, we implemented a method for generating composite features by combining significant attributes using a logistic regression model. Multiple composite features were developed and evaluated through a 10-fold cross-validation process (referred to as logistic1-logistic15). The detailed process for constructing these composite features is outlined in the hybrid feature section of ensemble-AMPPred [48]. A set of selected features was used to fit a logistic regression model, which is expressed by the following equation:
Prob .   ( Y = NPs | x ) = logistic   ( x ) = e β 0 + β 1 X 1 + β 2 X 2 + β 3 X 3 + + β n X n   1 + e β 0 + β 1   X 1 + β 2 X 2 + β 3 X 3 + + β n X n
A logit transformation (natural logarithm of the odds ratio for Y being in the NPs category) was applied to establish a link function with the logistic regression model. The logit function is defined as follows:
Logit ( x ) = log ( P Y = N P s   X = X ) P Y = n o n N P s x = x ) = β 0 + β 1 X 1 + β 2 X 2 + β 3 X 3 + + β n X n
As a result, the composite feature is represented by the following equation:
Composite Feature = β0 + β1 feature1 + β2feature2 + β3feature3 + ⋯ + βn featureN
In this equation, β0 denotes the intercept, while β1, β2, β3, and βn correspond to the regression coefficients of each selected feature. The variables feature1, feature2, …, and featureN represent the individual features that make up the composite feature. Among the selected composite features, logistic1 demonstrated the highest sensitivity and is expressed as follows:
logistic 1 = β 0 + β 1 m e r 5 + β 2 k m e r 195 + β 3 K m e r 285 + β 4 K m e r 92 + β 5 P s e _ S C 16 + β 6 C h a r g e
where β0 is the intercept, β1, β2, β3, β4, β5, and β6 represent the regression coefficients; mer5 is the amino acid composition of Phe (F) in a peptide; kmer195 is the di-amino-acid composition of LR; kmer285 is the di-amino-acid composition of RF; and kmer92 is the di-amino-acid composition of FN. Pse_SC16 is a pseudo amino acid composition (PseAAC) descriptor in the generalized mode of amino acid Ser (S). Charge is a computed net charge of a peptide.

2.3. Feature Selection

To identify a discriminative subset of features, a feature selection (FS) method was employed to select relevant and informative features that efficiently discriminate NPs from non-NPs. The ReliefF [49] method was applied as a preprocessing step to eliminate irrelevant features. This method assigns a ReliefF score to each feature based on its ability to differentiate between neighboring instances from different classes. Features that contribute more significantly to separating samples were assigned higher weights, prioritizing those that enhance local class discrimination. To assess potential multicollinearity among features, we computed Pearson correlation coefficients for all possible feature pairs. Among 481,671 possible combinations, only 621 pairs (0.13%) exhibited high correlation (|r| ≥ 0.8), indicating minimal redundancy. This supports the suitability of the ReliefF method, which prioritizes feature relevance over correlation.

2.4. Base Classifier Selection and Model Implementation

To construct a robust ensemble classifier, we evaluated six machine learning (ML) algorithms: Support Vector Machine (SVM) [50], k-Nearest Neighbors (kNN) [51], Decision Tree (DT) [52], Random Forest (RF) [53], Extreme Gradient Boost (XGB) [54], and Extra Trees (ET) [55]. Each algorithm offers unique inductive biases and learning mechanisms, which, when combined in an ensemble, can contribute to improved prediction diversity and generalization. Hyperparameter tuning was conducted using grid search to determine the optimal configuration for each model.
The SVM model is a supervised learning algorithm used for both classification and regression tasks. It maps data into a high-dimensional space and identifies a hyperplane that maximizes the margin between classes. In this study, we used a radial basis function (RBF) kernel with the following parameters: C = 19 and gamma = 0.125.
The kNN classifier is a non-parametric method used in statistical pattern recognition. It classifies a data point based on the majority label among its k-Nearest Neighbors. We used k = 9 and applied inverse distance weighting to compute similarity.
The Decision Tree (DT) is a tree-based classifier that uses a set of hierarchical decision rules to partition the feature space. It is valued for its interpretability and low computational cost. The confidence factor was set to 0.25 in our implementation.
The Random Forest (RF) algorithm is an ensemble of decision trees that aggregates predictions via majority voting. It is known for high accuracy and robustness to overfitting, especially on datasets with mixed feature types. We used 300 trees in our model.
The Extra Trees (ET) algorithm, or Extremely Randomized Trees, is another ensemble technique that introduces additional randomness during training. Unlike RF, which selects optimal split from a random subset of features, ET randomly selects both features and split points, resulting in greater model diversity. This stochastic approach reduces variance and training time, making ET especially suitable for large or noisy datasets. In this study, we used 900 trees.
The Extreme Gradient Boosting (XGB) model is based on the gradient boosting framework. It iteratively trains weak learners, optimizing for a differentiable loss function by adjusting weights at each iteration. The following parameters were used: nrounds = 150, max_depth = 5, learning rate (eta) = 0.03, and subsample = 0.6.
We also implemented a deep learning model using a Convolutional neural network (CNN) [56,57], which is effective for identifying spatial patterns in grid-like biological data. Input peptide sequences were encoded using Word2Vec embeddings [58] with the skip-gram approach to represent k-mers in a 120-dimensional vector space.
We empirically explored several CNN architectures with 2 to 6 convolutional layers and also evaluated LSTM-based models. Although deeper networks had higher capacity, they showed diminishing returns in performance and increased overfitting risk on the training data. The final CNN configuration was selected based on empirical results and tuned via grid search with five-fold cross-validation. The selected architecture consisted of two convolutional layers (64 and 32 filters; kernel size = 3), each followed by a max pooling layer (pool size = 2), and ReLU activation functions, followed by a fully connected dense layer (32 units, ReLU activation) and a dropout rate of 0.3. The model was optimized using Adam’s optimizer with a learning rate of 0.0001 for up to 100 epochs with early stopping applied. This CNN module automatically learns hierarchical, discriminative sequence features analogous to autoencoding, effectively reducing the need for manual feature engineering. It complements traditional handcrafted features with learned nonlinear representations to improve classification.
All implementations were performed using a combination of Perl (v5.38.2), Python (v3.12.6), Weka (v3.8.6; University of Waikato, Hamilton, New Zealand), and R (v4.2.2; R Foundation for Statistical Computing, Vienna, Austria) on a Fedora Linux-based system (Fedora 39; Red Hat, Inc., Raleigh, NC, USA) equipped with an Intel Core i7 CPU @ 2.30 GHz, 16 GB RAM, and an NVIDIA GeForce RTX3070 GPU. The trained models and standalone program are publicly available at: http://www.ncrna-pred.com/EnsembleNPPred.htm (accessed on 24 June 2025).
A 10-fold cross-validation was conducted on the training dataset to evaluate classification performance and identify the best-performing models for integration into the final program. Subsequently, two independent test datasets were used to validate the selected models and benchmark performance against existing neuropeptide prediction tools.
The model’s performance was assessed using the following evaluation metrics:
A c c u r a c y ( A C C ) = T P + T N ( T P + T N + F P + F N )
S e n s i t i v i t y ( R e c a l l , S n ) = T P ( T P + F N )
S p e c i f i c i t y ( S p ) = T N ( T N + F P )
P r e c i s i o n = T P ( T P + F P )
M a t t h e w s C o r r e l a t i o n C o e f f i c i e n t ( M C C ) = T P × T N F P × F N ( T P + F P ) × ( T P + F N ) × ( T N + F P ) × ( T N + F N )
F 1 s c o r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
where TP, TN, FP, and FN represent true positives, true negatives, false positives, and false negatives, respectively. These metrics were used to evaluate the model’s overall accuracy, class balance, and predictive reliability.
In addition, a Receiver Operating Characteristic (ROC) curve was generated to illustrate the trade-off between sensitivity and specificity across various classification thresholds. The Area Under the Curve (AUC) was calculated to assess the classifier’s ability to distinguish between classes, with an AUC of 1.0 indicating a perfect classifier.

3. Results and Discussion

3.1. Amino Acid Composition and Positional Residue Analysis

The amino acid composition of neuropeptides (NPs) reveals distinguishing distinctive properties that differentiate them from non-neuropeptides (non-NPs). Table 1 summarizes the percentage of amino acid groups, categorized by physicochemical properties, as calculated using COPid [59]. Neuropeptides tend to be shorter in length and are enriched in aromatic, polar, and negatively charged residues, which may be crucial for bioactivity, receptor interaction, and molecular recognition. These compositional features provide valuable input for machine learning models aimed at distinguishing neuropeptides from other peptide sequences.
Figure 2A presents a comparison of the average amino acid composition between neuropeptides (NPs) and non-neuropeptides (non-NPs). The analysis reveals notable differences in the relative abundance of specific residues. Amino acids such as Glycine (G), Proline (P), Phenylalanine (F), Serine (S), and Arginine (R) are more abundant in NPs, suggesting their potential involvement in neuropeptide functionality, possibly contributing to flexibility, receptor interaction, or bioactivity. Conversely, residues including Valine (V), Threonine (T), Methionine (M), Cysteine (C), Lysine (K), Isoleucine (I), and Leucine (L) are more prevalent in non-NPs, which may reflect their association with structural stability or non-signaling peptide segments. These compositional differences emphasize underlying physicochemical properties that may influence neuropeptide behavior and can be informative for classification tasks.
Positional residue analysis of both the N-terminal and C-terminal regions was conducted by calculating the average amino acid composition at positions 1 to 5 for each terminus in neuropeptides (NPs, positive) and non-neuropeptides (non-NPs, negative). As illustrated in Figure 2B,C, sequence logos were generated using ggseqlogo [60]. In these logos, amino acids are color-coded as follows: red for acidic, blue for basic, purple for neutral, black for hydrophobic, and green for polar residues. The height of each letter reflects the relative frequency of that amino acid at the corresponding position.
Figure 2B shows the first five N-terminal residues for both NPs (left panel) and non-NPs (right panel). The NP panel shows a notable preference for polar (green) and acidic (red) residues, suggesting sequence conservation at the N-terminus. In contrast, the non-NP panel displays a more uniform distribution without a clear dominant residue pattern, suggesting less positional conservation.
Figure 2C displays the amino acid distribution at the C-terminal region. The NP panel shows a clear enrichment of hydrophobic (black), polar (green), and basic (blue) residues, particularly arginine (R) and glycine (G), along with phenylalanine (F), leucine (L), proline (P), and serine (S). By contrast, the non-NP panel shows a broader, less structured distribution, further suggesting a lack of motif conservation at the C-terminus in non-neuropeptides.
The most prevalent amino acids at the N-terminus of NPs were Serine (S, 11.70%), Glycine (G, 9.32%), Alanine (A, 8.40%), Proline (P, 8.37%), Asparagine (N, 5.18%), and Tyrosine (Y, 4.07%). At the C-terminus, the dominant residues were Phenylalanine (F, 13.80%), Leucine (L, 10.45%), Arginine (R, 10.29%), and Glycine (G, 9.95%).
In addition, Figure 3 presents a heatmap of log-odds scores for 2-mer (dipeptide) motifs comparing NPs vs. non-NPs. In this heatmap, each cell represents a specific amino acid pair, with the first residue on the y-axis and the second residue on the x-axis. The color intensity indicates the log-odds ratio; red shading indicates dipeptides overrepresented in NPs, while lighter red to white shading indicates lower or underrepresented occurrences in NPs relative to non-NPs. The log-odds values range approximately from −3 to +3, with zero indicating no difference between the two groups.
Notably, several dipeptides such as Gly-Phe (GF; 1.51), Phe-Gly (FG; 1.67), Arg-Phe (RF; 1.88), Tyr-Gly (YG; 1.67), Gly-Met (GM; 1.95), and Pro-Arg (PR; 1.67) were found to be overrepresented in NPs. These motifs suggest potential sequence features contributing to neuropeptide specificity. Conversely, dipeptides such as Met-Cys (MC; −2.98), Trp-Cys (WC; −2.49), Met-Phe (MF; −2.02), Met-Ile (MI; −2.91), Met-Lys (MK; −2.29), Met-Leu (ML; −2.08), Ile-Ile (II = −2.20), and Cys-Ile (CI; −2.53) were more prevalent in non-NPs, possibly indicating these pairings may be more characteristic of non-functional or structural peptides. Taken together, these findings highlight amino acid and dipeptide-level features that distinguish NPs from non-NPs, offering useful insights for improving predictive modeling and sequence-based classification of neuropeptides.

3.2. 10-Fold Cross-Validation of Predictive Performance with the Training Dataset

To evaluate the generalization performance of individual classifiers, we conducted 10-fold cross-validation on the training dataset. The key evaluation metrics including accuracy (ACC), Matthews correlation coefficient (MCC), sensitivity (Sn), specificity (Sp), area under the ROC curve (AUC), and 95% confidence intervals for the ROC (CI) are summarized in Table 2.
Among the individual models, the DL classifier achieved the highest accuracy (93.918%) and MCC (0.878), demonstrating strong overall performance. The ET and SVM models also performed competitively, with accuracy exceeding 93.7% and AUC values of 0.986. The RF and XGB classifiers followed closely, while the DT and KNN models showed relatively lower performance.
Based on these results, we constructed a heterogeneous ensemble model combining SVM, DL, and ET. This integration leveraged the complementary strengths of each classifier: DL contributed high overall metrics, SVM provided the highest sensitivity, and ET enhanced predictive diversity through its randomized feature subspaces.
The final ensemble model, which integrated predictions through a voting mechanism, achieved an accuracy of 93.978%, an MCC of 0.880, and an AUC of 0.987. It outperformed all individual models in terms of MCC and specificity (0.941), suggesting more balanced and robust predictions, particularly in reducing false positives, which is critical for distinguishing neuropeptides from non-neuropeptides. The ensemble’s narrow AUC confidence interval further supports its performance consistency across validation folds. These results align with the intended mitigation strategy embedded in our model design by combining diverse learners (SVM, ET, CNN) to reduce individual model bias and variance while enhancing predictive reliability. Additionally, the hybrid ensemble integrates interpretable handcrafted features with deep learning-derived sequence embeddings. While handcrafted features offer domain relevance and model transparency, the CNN applied to Word2Vec embeddings captures complex, nonlinear sequence patterns analogous to autoencoding. This complementary combination balances interpretability with feature richness, contributing to a performance-oriented and interpretable framework for neuropeptide classification.
Taken together, these findings suggest that the ensemble approach effectively combines high-performing base classifiers and demonstrates enhanced generalization capability. A comparison of ROC curves and cross-validation performance metrics is presented in Figure 4.

3.3. Feature Interpretability and Importance

Interpreting model behavior and identifying influential features are essential for validating predictions and gaining biological insights. Understanding which features contribute most to the model’s predictions can provide biological insights and help validate model behavior. Feature importance was initially assessed using the built-in feature ranking mechanisms of the RF and XGB models. In both models, the composite feature logistic1 consistently emerged as the most informative, as shown in Figure 5A,B.
To further investigate feature contributions, we conducted Shapley Additive exPlanations (SHAP) analysis. SHAP is a widely used game-theoretic approach that quantifies the impact of each feature on individual predictions [61]. The SHAP summary plot in Figure 5C supports the RF and XGB results, again highlighting logistic1 as the most influential feature.
The SHAP framework estimates the contribution of each input feature to the final prediction for each individual sample. By assigning importance values based on the mean absolute SHAP value across all predictions, SHAP provides a more detailed and individualized interpretation than traditional feature importance methods. These consistent findings across both tree-based methods (RF and XGB) and SHAP analysis indicate that composite features such as logistic1, along with CTD-derived descriptors, play a critical role in distinguishing neuropeptides from non-neuropeptides in the EnsembleNPPred model.

3.4. Performance Comparison of Various Existing Predictive Models

To assess the performance of the proposed EnsembleNPPred method, two independent testing datasets were used. The performance of our model was evaluated against several previously published neuropeptide prediction methods, using testing dataset 1 and testing dataset 2, as summarized in Table 3 and Table 4, respectively.
As shown in Table 3, the performance of various neuropeptide prediction models was assessed using five key metrics. NeuroPred-FRL achieved an accuracy (ACC) of 0.900, MCC of 0.803, sensitivity (Sn) of 0.946, specificity (Sp) of 0.854, and an AUC of 0.965. NeuroPpred-Fuse demonstrated good specificity (0.930) but had slightly lower sensitivity compared to NeuroPred-CLQ. NeuroPIpred showed the lowest performance across all metrics, with an ACC of 0.536, MCC of 0.074, Sn of 0.331, Sp of 0.736, and an AUC of 0.581. These results highlight the limited effectiveness of NeuroPIpred, which was specifically trained and designed for insect neuropeptides, when applied to broader datasets encompassing neuropeptides from various animal phyla. NeuroPred-CLQ exhibited strong performance with an ACC of 0.936, MCC of 0.875, Sn of 0.897, Sp of 0.975, and an AUC of 0.988. This model’s high specificity and AUC indicate its ability to correctly identify neuropeptides, making it one of the top performers among the models evaluated.
On testing dataset 1 (Table 3), EnsembleNPPred achieved an accuracy (ACC) of 0.940, which was higher than that of all other models, including NeuroPred-CLQ, which had the next highest ACC at 0.936. Additionally, our model outperformed the other models in terms of the Matthews correlation coefficient (MCC), which was 0.881. The high MCC suggests that EnsembleNPPred has a strong predictive capability with a good balance between sensitivity (Sn = 0.962) and specificity (Sp = 0.918). Notably, the area under the curve (AUC) was the highest among all methods at 0.990, further indicating the robustness of our model in distinguishing neuropeptides from non-neuropeptides.
When evaluated on testing dataset 2 (Table 4), which includes more recent neuropeptide sequences from the NeuroPep 2.0 database, EnsembleNPPred again achieved top-tier performance. It reached an ACC of 0.929 and an MCC of 0.859, surpassing the performance of NeuroPred-PLM, which achieved an ACC of 0.922 and an MCC of 0.845. The precision and recall of our model were both high, at 0.930 and 0.929, respectively, resulting in an F1 score of 0.929, the highest among all compared methods.
Across both independent datasets, EnsembleNPPred consistently demonstrated a better performance, highlighting its robustness and reliability in neuropeptide prediction. Its high accuracy, MCC, and AUC scores suggest not only strong predictive capability but also a well-balanced performance between precision and recall, making it highly effective across different datasets.
The EnsembleNPPred framework combining interpretable handcrafted features with sequence patterns automatically learned through CNN embedding. This hybrid design helps mitigate potential issues such as overfitting, feature redundancy, and limited interpretability, particularly when classifying heterogeneous peptide families. By integrating domain-specific knowledge and data-driven representations, the final model architecture was selected based on its stability, interpretability, and computational efficiency. While promising, the current approach could benefit from future enhancements, such as adopting advanced deep learning paradigms that better capture long-range dependencies and structural complexity in peptide sequences.

3.5. Performance Across Diverse Neuropeptide Families

To further assess generalizability, we evaluated the model on a diverse range of neuropeptide families using testing dataset 3, as detailed in Supplementary Data S1. This additional evaluation served to mitigate dataset-specific bias and further assessed the model’s robustness, particularly for well-defined peptide classes. The results demonstrate that the model performs well across a broad range of neuropeptide families. High accuracy was observed for families such as AKH/HRTH/RPCH (99.62%), FMRFamide-related peptide (99.65%), insulin (98.26%), and NPY (99.18%). The overall average accuracy across all families was 91.92%, although performance varied depending on the specific family. Challenges observed in structurally complex, or neuropeptide-like families highlight potential areas for future model improvement through structural feature integration. Certain families, such as chemokine (44%), serpin (37.61%), and nucleobindins (66.67%), exhibited reduced accuracy. These discrepancies may be attributed to factors such as complex post-translational modifications, the presence of multiple isoforms, and dynamic structural conformation [62,63,64,65,66,67,68,69], which complicate classification. It is also worth noting that some families such as chemokines, chromogranin/secretogranin, nucleobindin, serpins, and tenascins are often categorized as “neuropeptide-like” rather than canonical neuropeptides. According to Burbach [1,4], neuropeptide-like peptides may influence neural or physiological activity in similar ways to classical neuropeptides but fail to meet all defining criteria, such as originating from neurons, being secreted in a stimulus-dependent manner, or acting via well-characterized neuropeptide receptors. These deviations are likely to contribute to the reduced classification performance observed in these families.
For benchmarking purposes, we compared our model against the NeuroPred-FRL web server [20], which provides a widely accessible and user-friendly interface. A detailed comparison of predictions generated by NeuroPred-FRL across all neuropeptide families is presented in Supplementary Data S2 and summarized in Supplementary Data S1. On the same evaluation dataset, our proposed model achieved an overall accuracy of 91.92%, while NeuroPred-FRL achieved 71.97%. These results suggest that EnsembleNPPred offers competitive performance under the conditions tested.
Overall, the proposed ensemble approach, which integrates both machine learning and deep learning techniques, demonstrates robust performance across most neuropeptide families and appears particularly effective for classifying well-characterized peptides. The average accuracy of 91.92% across all families suggests that the model is generally useful for neuropeptide prediction. However, the reduced performance observed for structurally complex or atypical families highlights key limitations. Addressing these challenges in future work may require incorporating more advanced structural feature representations and expanding the training dataset to include greater sequence diversity, thereby enhancing generalizability and predictive accuracy.

3.6. Evaluation of False Positive Rates for EnsembleNPPred

To assess the false positive rate of EnsembleNPPred, we used antimicrobial peptides from the DRAMP Database [70] as a negative test set (see Supplementary Data S3 for details). EnsembleNPPred demonstrates consistently low false positive rates across multiple AMP categories. Specifically, the model achieved false positive rates of 6.68% for antibacterial, 12.13% for antifungal, and 11.86% for antiviral peptides. These results suggest that EnsembleNPPred provides relatively high specificity and robustness in distinguishing neuropeptides from non-neuropeptide sequences.
To further investigate the nature of the misclassified sequences, a subset of AMP sequences that were predicted as neuropeptides was analyzed for conserved domains using the InterPro domain database (Supplementary Data S3, column L). Interestingly, the majority of these peptides were found to contain domains associated with neuropeptide precursors or neuroendocrine signaling in addition to their known antimicrobial functions. In accordance with the report by Wei et al., neuropeptides and AMPs share common characteristics, i.e., signal sequence, mostly at the N-terminus, and size [12]. Notable examples include the Chromogranin A/B/C domain, neuropeptide-like protein domains, and the Tachykinin domain [71], all of which are associated with classical neuropeptide signaling pathways. Moreover, several other domains identified in the misclassified sequences—e.g., Acyl-CoA-binding protein [72], Chemokine beta/gamma/delta [73], and Pancreatic hormone-like domain [74]—are not traditionally classified as neuropeptide-related but have been reported to be involved in the regulation of the immune system [75]. While the presence of these domains does not confirm neuropeptide function, it raises the possibility that some misclassified AMP sequences may exhibit a domain structure common to that of neuropeptides, which may possibly lead to dual-functionality, as previously reported [11,12].
Taken together, these observations suggest the possibility that certain peptides may exhibit dual-functionality—potentially contributing to both host defense and neuroimmune modulation, particularly in relation with bidirectional gut–brain communication to regulate microbiome population in the gut and immune response mechanisms of the host. However, experimental validation will be necessary to determine the biological relevance of these predictions and to confirm whether these peptides indeed possess neuropeptide-like functions.

4. Conclusions

This study presents EnsembleNPPred, a hybrid neuropeptide prediction framework that integrates traditional machine learning (ML) techniques with a deep learning (DL) component using an ensemble voting strategy. By integrating interpretable handcrafted features with CNN-derived embeddings from k-mer-based Word2Vec encoding, the model leverages both domain knowledge and data-driven representations to improve predictive performance and interpretability. The model was evaluated on two independent test sets and demonstrated consistent and competitive results. On the first test set, EnsembleNPPred achieved an accuracy of 0.940 and a Matthews correlation coefficient (MCC) of 0.881. On the second test set, it attained an accuracy of 0.929, an MCC of 0.859, and an F1 score of 0.929, surpassing several existing methods under the tested conditions.
When applied across a wide range of neuropeptide families, EnsembleNPPred maintained strong performance, particularly for well-characterized families such as AKH/HRTH/RPCH, FMRFamide-related peptides, insulin, and NPY. However, reduced accuracy was observed for families such as chemokines, leptin, and serpins. This may be attributed to their complex structural features, including diverse receptor interactions, conformational flexibility, and the presence of multiple isoforms. To address these challenges, future enhancement may focus on incorporating more advanced structural representations and expanding training datasets to better capture conformational diversity and post-translational modifications. One possible direction includes integrating three-dimensional structural data (e.g., predicted tertiary conformations or receptor-binding motifs) potentially using pretrained transformer-based models such as ESM or ProtBERT for representing tertiary conformations or receptor-binding properties. EnsembleNPPred demonstrates strong performance in neuropeptide classification, particularly in canonical families; however, further refinement and broader validation on novel or experimentally derived datasets are still needed. Overall, EnsembleNPPred provides a flexible and interpretable framework for large-scale peptide annotation, with the potential to support research in functional genomics, neurobiology, and therapeutic peptide discovery.

Supplementary Materials

The following supporting data are available online at https://www.mdpi.com/article/10.3390/life15071010/s1, Supplementary Data S1: Assessment of model accuracy across neuropeptide families, Supplementary Data S2: Detailed comparison of predictions generated by NeuroPred-FRL across all neuropeptide families, Supplementary Data S3: Evaluation of false positive rates for EnsembleNPPred across multiple AMP categories.

Author Contributions

Conceptualization, A.H. and S.L.; methodology, A.H. and S.L.; software, S.L. and W.W.; validation, S.L., A.H., W.W. and C.T.; formal analysis, S.L.; investigation, S.L., A.H., W.W. and C.T.; resources, A.H.; data curation, S.L.; writing—original draft preparation, S.L.; writing—review and editing, S.L., A.H., W.W. and C.T.; funding acquisition, A.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Thailand Science Research and Innovation, grant numbers FRB670016/0164 and FRB680074/0164.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available in the supplementary materials of this article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Schoofs, L.; De Loof, A.; Van Hiel, M.B. Neuropeptides as Regulators of Behavior in Insects. Annu. Rev. Entomol. 2017, 62, 35–52. [Google Scholar] [CrossRef] [PubMed]
  2. Burbach, J.P. Neuropeptides from concept to online database www.neuropeptides.nl. Eur. J. Pharmacol. 2010, 626, 27–48. [Google Scholar] [CrossRef] [PubMed]
  3. Kupcova, I.; Danisovic, L.; Grgac, I.; Harsanyi, S. Anxiety and Depression: What Do We Know of Neuropeptides? Behav. Sci. 2022, 12, 262. [Google Scholar] [CrossRef]
  4. Burbach, J. What are neuropeptides? Methods Mol. Biol. 2011, 789, 1–36. [Google Scholar] [CrossRef] [PubMed]
  5. Elphick, M.R.; Mirabeau, O.; Larhammar, D. Evolution of neuropeptide signalling systems. J. Exp. Biol. 2018, 221, jeb151092. [Google Scholar] [CrossRef] [PubMed]
  6. Ofer, D.; Linial, M. NeuroPID: A predictor for identifying neuropeptide precursors from metazoan proteomes. Bioinformatics 2014, 30, 931–940. [Google Scholar] [CrossRef] [PubMed]
  7. Nässel, D.R.; Zandawala, M. Recent advances in neuropeptide signaling in Drosophila, from genes to physiology and behavior. Prog. Neurobiol. 2019, 179, 101607. [Google Scholar] [CrossRef]
  8. Bhat, U.S.; Shahi, N.; Surendran, S.; Babu, K. Neuropeptides and Behaviors: How Small Peptides Regulate Nervous System Function and Behavioral Outputs. Front. Mol. Neurosci. 2021, 14, 786471. [Google Scholar] [CrossRef]
  9. Sharma, D.; Kumar, K.; Bisht, G.S. A Mini-Review on Potential of Neuropeptides as Future Therapeutics. Int. J. Pept. Res. Ther. 2022, 28, 39. [Google Scholar] [CrossRef]
  10. An, M.Y.; Gao, J.; Zhao, X.F.; Wang, J.X. A new subfamily of penaeidin with an additional serine-rich region from kuruma shrimp (Marsupenaeus japonicus) contributes to antimicrobial and phagocytic activities. Dev. Comp. Immunol. 2016, 59, 186–198. [Google Scholar] [CrossRef]
  11. Blanchet, X.; Weber, C.; von Hundelshausen, P. Chemokine Heteromers and Their Impact on Cellular Function-A Conceptual Framework. Int. J. Mol. Sci. 2023, 24, 10925. [Google Scholar] [CrossRef] [PubMed] [PubMed Central]
  12. Wei, P.; Keller, C.; Li, L. Neuropeptides in gut-brain axis and their influence on host immunity and stress. Comput. Struct. Biotechnol. J. 2020, 18, 843–851. [Google Scholar] [CrossRef] [PubMed] [PubMed Central]
  13. Florea, G.; Tudorache, I.F.; Fuior, E.V.; Ionita, R.; Dumitrescu, M.; Fenyo, I.M.; Bivol, V.G.; Gafencu, A.V. Apolipoprotein A-II, a Player in Multiple Processes and Diseases. Biomedicines 2022, 10, 1578. [Google Scholar] [CrossRef]
  14. Zhuang, J.; Zhang, Y.D.; Sun, W.X.; Zong, J.; Li, J.; Dai, X.; Klosterman, S.J. The acyl-CoA-binding protein VdAcb1 is essential for carbon starvation response and contributes to virulence in Verticillium dahliae. aBIOTECH 2024, 5, 431–448. [Google Scholar] [CrossRef] [PubMed]
  15. Wang, Y.; Wang, M.; Yin, S.; Jang, R.; Wang, J.; Xue, Z.; Xu, T. NeuroPep: A comprehensive resource of neuropeptides. Database 2015, 2015, bav038. [Google Scholar] [CrossRef] [PubMed]
  16. Wang, M.; Wang, L.; Xu, W.; Chu, Z.; Wang, H.; Lu, J.; Xue, Z.; Wang, Y. NeuroPep 2.0: An Updated Database Dedicated to Neuropeptide and Its Receptor Annotations. J. Mol. Biol. 2024, 436, 168416. [Google Scholar] [CrossRef]
  17. Agrawal, P.; Kumar, S.; Singh, A.; Raghava, G.; Singh, I.K. NeuroPIpred: A tool to predict, design and scan insect neuropeptides. Sci. Rep. 2019, 9, 5129. [Google Scholar] [CrossRef]
  18. Bin, Y.; Zhang, W.; Tang, W.; Dai, R.; Li, M.; Zhu, Q.; Xia, J. Prediction of Neuropeptides from Sequence Information Using Ensemble Classifier and Hybrid Features. J. Proteome Res. 2020, 19, 3732–3740. [Google Scholar] [CrossRef]
  19. Jiang, M.; Zhao, B.; Luo, S.; Wang, Q.; Chu, Y.; Chen, T.; Mao, X.; Liu, y.; Wang, y.; Jiang, X.; et al. NeuroPpred-Fuse: An interpretable stacking model for prediction of neuropeptides by fusing sequence information and feature selection methods. Brief Bioinform. 2021, 22, bbab310. [Google Scholar] [CrossRef]
  20. Hasan, M.M.; Alam, M.A.; Shoombuatong, W.; Deng, H.W.; Manavalan, B.; Kurata, H. NeuroPred-FRL: An interpretable prediction model for identifying neuropeptide using feature representation learning. Brief Bioinform. 2021, 22, bbab167. [Google Scholar] [CrossRef]
  21. Chen, S.; Li, Q.; Zhao, J.; Bin, Y.; Zheng, C. NeuroPred-CLQ: Incorporating deep temporal convolutional networks and multi-head attention mechanism to predict neuropeptides. Brief Bioinform. 2022, 23, bbac319. [Google Scholar] [CrossRef] [PubMed]
  22. Wang, L.; Huang, C.; Wang, M.; Xue, Z.; Wang, Y. NeuroPred-PLM: An interpretable and robust model for neuropeptide prediction by protein language model. Brief Bioinform. 2023, 24, bbad077. [Google Scholar] [CrossRef]
  23. Farias, J.G.; Herrera-Belén, L.; Jimenez, L.; Beltrán, J.F. PROTA: A Robust Tool for Protamine Prediction Using a Hybrid Approach of Machine Learning and Deep Learning. Int. J. Mol. Sci. 2024, 25, 10267. [Google Scholar] [CrossRef] [PubMed]
  24. Li, C.; Wang, H.; Wen, Y.; Yin, R.; Zeng, X.; Li, K. GenoM7GNet: An Efficient N7-Methylguanosine Site Prediction Approach Based on a Nucleotide Language Model. IEEE/ACM Trans. Comput. Biol. Bioinf. 2024, 21, 6. [Google Scholar] [CrossRef]
  25. Zhang, R.; Lin, Y.; Wu, Y.; Deng, L.; Zhang, H.; Liao, M. MvMRL: Multi-view molecular representation learning with cross-attention for bioactivity prediction. Brief Bioinform. 2024, 25, bbae298. [Google Scholar] [CrossRef]
  26. Alarfaj, F.K.; Khan, J.A. Deep Dive into Fake News Detection: Feature-Centric Classification with Ensemble and Deep Learning Methods. Algorithms 2023, 16, 507. [Google Scholar] [CrossRef]
  27. Borandag, E. Software Fault Prediction Using an RNN-Based Deep Learning Approach and Ensemble Machine Learning Techniques. Appl. Sci. 2023, 13, 1639. [Google Scholar] [CrossRef]
  28. Chou, K.C. Prediction of protein subcellular locations by incorporating quasi-sequence-order effect. Biochem. Biophys. Res. Commun. 2000, 278, 477–483. [Google Scholar] [CrossRef]
  29. Chou, K.C. Using amphiphilic pseudo amino acid composition to predict enzyme subfamily classes. Bioinformatics 2005, 21, 10–19. [Google Scholar] [CrossRef]
  30. Dubchak, I.; Muchnik, I.; Holbrook, S.R.; Kim, S.H. Prediction of protein folding class using global description of amino acid sequence. Proc. Natl. Acad. Sci. USA 1995, 92, 8700–8704. [Google Scholar] [CrossRef]
  31. Dubchak, I.; Muchnik, I.; Mayor, C.; Dralyuk, I.; Kim, S. Recognition of a protein fold in the context of the scop classification. Proteins Struct. Funct. Genet. 1999, 35, 401–407. [Google Scholar] [CrossRef]
  32. Xiao, N.; Cao, D.S.; Zhu, M.F.; Xu, Q.S. protr/ProtrWeb: R package and web server for generating various numerical representation schemes of protein sequences. Bioinformatics 2015, 31, 1857–1859. [Google Scholar] [CrossRef] [PubMed]
  33. Chou, K.C. Some remarks on protein attribute prediction and pseudo amino acid composition (50th anniversary year re-view). J. Theor. Biol. 2011, 273, 236–247. [Google Scholar] [CrossRef] [PubMed]
  34. Cruciani, G.; Baroni, M.; Carosati, E.; Clementi, M.; Valigi, R.; Clementi, S. Peptide studies by means of principal properties of amino acids derived from MIF descriptors. J. Chemom. 2004, 18, 146–155. [Google Scholar] [CrossRef]
  35. Sandberg, M.; Eriksson, L.; Jonsson, J.; Sjostrom, M.; Wold, S. New chemical descriptors relevant for the design of biologically active peptides. A multivariate characterization of 87 amino acids. J. Med. Chem. 1998, 41, 2481–2491. [Google Scholar] [CrossRef]
  36. Liang, G.; Li, Z. Factor analysis scale of generalized amino acid information as the source of a new set of descriptors for elucidating the structure and activity relationships of cationic antimicrobial peptides. Mol. Inform. 2007, 26, 754–763. [Google Scholar] [CrossRef]
  37. Tian, F.; Zhou, P.; Li, Z. T-scale as a novel vector of topological descriptors for amino acids and its application in QSARs of peptides. J. Mol. Struct. 2007, 830, 106–115. [Google Scholar] [CrossRef]
  38. Mei, H.U.; Liao, Z.H.; Zhou, Y.; Li, S.Z. A new set of amino acid descriptors and its application in peptide QSARs. Pept. Sci. 2005, 80, 775–786. [Google Scholar] [CrossRef]
  39. van Westen, G.J.; Swier, R.F.; Wegner, J.K.; IJzerman, A.P.; van Vlijmen, H.W.; Bender, A. Benchmarking of protein descriptor sets in proteochemometric modeling (part 1): Comparative study of 13 amino acid descriptor sets. J. Cheminformatics 2013, 5, 41. [Google Scholar] [CrossRef]
  40. Yang, L.; Shu, M.; Ma, K.; Mei, H.; Jiang, Y.; Li, Z. ST-scale as a novel amino acid descriptor and its application in QSAM of peptides and analogues. Amino Acids 2010, 38, 805–816. [Google Scholar] [CrossRef]
  41. Zaliani, A.; Gancia, E. MS-WHIM scores for amino acids: A new 3D-description for peptide QSAR and QSPR studies. J. Chem. Inf. Comput. Sci. 1999, 39, 525–533. [Google Scholar] [CrossRef]
  42. Ikai, A. Thermostability and aliphatic index of globular proteins. J. Biochem. 1980, 88, 1895–1898. [Google Scholar] [PubMed]
  43. Boman, H.G. Antibacterial peptides: Basic facts and emerging concepts. J. InternalMedicine 2003, 254, 197–215. [Google Scholar] [CrossRef] [PubMed]
  44. Guruprasad, K.; Reddy, B.V.; Pandit, M.W. Correlation between stability of a protein and its dipeptide composition: A novel approach for predicting in vivo stability of a protein from its primary sequence. Protein Eng. 1990, 4, 155–161. [Google Scholar] [CrossRef] [PubMed]
  45. Osorio, D.; Rondon-Villarreal, P.; Torres, R. Peptides: A package for data mining of antimicrobial peptides. R J. 2015, 7, 4–14. [Google Scholar] [CrossRef]
  46. Vens, C.; Rosso, M.; Danchin, E. Identifying discriminative classification-based motifs in biological sequences. Bioinformatics 2011, 27, 1231–1238. [Google Scholar] [CrossRef]
  47. Fernandez-Escamilla, A.M.; Rousseau, F.; Schymkowitz, J.; Serrano, L. Prediction of sequence-dependent and mutational effects on the aggregation of peptides and proteins. Nat. Biotech. 2004, 22, 1302–1306. [Google Scholar] [CrossRef]
  48. Lertampaiporn, S.; Vorapreeda, T.; Hongsthong, A.; Thammarongtham, C. Ensemble-AMPPred: Robust AMP Prediction and Recognition Using the Ensemble Learning Method with a New Hybrid Feature for Differentiating AMPs. Genes 2021, 12, 137. [Google Scholar] [CrossRef]
  49. Kononenko, I. Estimating attributes: Analysis and extensions of RELIEF. In Machine Learning: ECML-94, Proceedings of the European Conference on Machine Learning, Catania, Italy, 6–8 April 1994; Bergadano, F., De Raedt, L., Eds.; Lecture Notes in Computer Science (Lecture Notes in Artificial Intelligence); Springer: Berlin/Heidelberg, Germany, 1994; Volume 784. [Google Scholar] [CrossRef]
  50. Vapnik, V.N. An overview of statistical learning theory. IEEE Trans. Neural Netw. 1999, 10, 988–999. [Google Scholar] [CrossRef]
  51. Weinberger, K.Q.; Saul, L.K. Distance metric learning for large margin nearest neighbor classification. J. Mach. Learn Res. 2009, 10, 207–244. [Google Scholar]
  52. Quinlan, J.R. Induction of decision trees. Mach. Learn. 1986, 1, 81–106. [Google Scholar] [CrossRef]
  53. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  54. Chen, T.; Guestrin, C. Xgboost: A scalable tree boosting system. In Proceedings of the 22nd ACM Sigkdd International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar]
  55. Geurts, P.; Ernst, D.; Wehenkel, L. Extremely randomized trees. Mach. Learn. 2006, 63, 3–42. [Google Scholar] [CrossRef]
  56. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  57. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  58. Mikolov, T.; Chen, K.; Corrado, G.; Dean, J. Efficient Estimation of Word Representations in Vector Space. arXiv 2013, arXiv:1301.3781. [Google Scholar]
  59. Kumar, M.; Thakur, V.; Raghava, G.P. COPid: Composition-based protein identification. Silico Biol. 2008, 8, 121–128. [Google Scholar] [CrossRef]
  60. Wagih, O. ggseqlogo: A versatile R package for drawing sequence logos. Bioinformatics 2017, 33, 3645–3647. [Google Scholar] [CrossRef]
  61. Lundberg, S.M.; Lee, S.-I. A unified approach to interpreting model predictions. Adv. Neural Inf. Process. Syst. 2017, 30, 4765–4774. [Google Scholar]
  62. De Zutter, A.; Van Damme, J.; Struyf, S. The Role of Post-Translational Modifications of Chemokines by CD26 in Cancer. Cancers 2021, 13, 4247. [Google Scholar] [CrossRef]
  63. Vanheule, V.; Metzemaekers, M.; Janssens, R.; Struyf, S.; Proost, P. How post-translational modifications influence the biological activity of chemokines. Cytokine 2018, 109, 29–51. [Google Scholar] [CrossRef] [PubMed]
  64. Londraville, R.; Prokop, J.; Duff, R.; Liu, Q.; Tuttle, M. On the Molecular Evolution of Leptin, Leptin Receptor, and Endospanin. Front. Endocrinol. 2017, 8, 58. [Google Scholar] [CrossRef] [PubMed]
  65. Wardman, J.H.; Berezniuk, I.; Di, S.; Tasker, J.G.; Fricker, L.D. ProSAAS-Derived Peptides are Colocalized with Neuropeptide Y and Function as Neuropeptides in the Regulation of Food Intake. PLoS ONE 2011, 6, e28152. [Google Scholar] [CrossRef]
  66. Meng, X.; McGraw, C.M.; Wang, W.; Jing, J.; Yeh, S.; Wang, L.; Lopez, J.; Brown, A.M.; Lin, T.; Chen, W.; et al. Neurexophilin4 is a selectively expressed α-neurexin ligand that modulates specific cerebellar synapses and motor functions. eLife 2019, 8, e46773. [Google Scholar] [CrossRef] [PubMed]
  67. Spence, M.A.; Mortimer, M.D.; Buckle, A.M.; Minh, B.Q.; Jackson, C.J. A Comprehensive Phylogenetic Analysis of the Serpin Superfamily. Mol. Biol. Evol. 2021, 38, 2915–2929. [Google Scholar] [CrossRef]
  68. Nillni, E.A. Regulation of Prohormone Convertases in Hypothalamic Neurons: Implications for ProThyrotropin-Releasing Hormone and Proopiomelanocortin. Endocrinology 2007, 148, 4191–4200. [Google Scholar] [CrossRef]
  69. Southey, B.R.; Romanova, E.V.; Rodriguez-Zas, S.L.; Sweedler, J.V. Bioinformatics for Prohormone and Neuropeptide Discovery. Methods Mol. Biol. 2018, 1719, 71–96. [Google Scholar] [CrossRef]
  70. Kang, X.; Dong, F.; Shi, C.; Liu, S.; Sun, J.; Chen, J.; Li, H.; Xu, H.; Lao, X.; Zheng, H. DRAMP 2.0, an updated data repository of antimicrobial peptides. Sci. Data 2019, 6, 148. [Google Scholar] [CrossRef]
  71. Nässel, D.R.; Zandawala, M.; Kawada, T.; Satake, H. Tachykinins: Neuropeptides That Are Ancient, Diverse, Widespread and Functionally Pleiotropic. Front. Neurosci. 2019, 13, 1262. [Google Scholar] [CrossRef] [PubMed] [PubMed Central]
  72. Lai, S.H.; Chye, M.L. Plant Acyl-CoA-Binding Proteins-Their Lipid and Protein Interactors in Abiotic and Biotic Stresses. Cells 2021, 10, 1064. [Google Scholar] [CrossRef] [PubMed] [PubMed Central]
  73. Liu, J.; Chu, S.; Zhou, X.; Zhang, D.; Chen, N. Role of chemokines in Parkinson’s disease. Brain Res. Bull. 2019, 152, 11–18. [Google Scholar] [CrossRef] [PubMed]
  74. Holzer, P.; Reichmann, F.; Farzi, A. Neuropeptide Y, peptide YY and pancreatic polypeptide in the gut-brain axis. Neuropeptides 2012, 46, 261–274. [Google Scholar] [CrossRef] [PubMed] [PubMed Central]
  75. Wu, S.; Bekhit, A.E.D.A.; Wu, Q.; Chen, M.; Liao, X.; Wang, J.; Ding, Y. Bioactive peptides and gut microbiota: Candidates for a novel strategy for reduction and control of neurodegenerative diseases. Trends Food Sci. Technol. 2021, 108, 164–176. [Google Scholar] [CrossRef]
Figure 1. Workflow of EnsembleNPPred. Blue arrows indicate training flow; red arrows indicate testing or evaluation flow. Different colored borders represent individual ML models.
Figure 1. Workflow of EnsembleNPPred. Blue arrows indicate training flow; red arrows indicate testing or evaluation flow. Different colored borders represent individual ML models.
Life 15 01010 g001
Figure 2. Composite analysis of amino acid composition and conserved sequence motifs in neuropeptides (NPs) vs. non-neuropeptides (non-NPs). (A) Average percent composition of amino acid residues in NP and non-NP datasets. (B) Sequence logos plots showing amino acid enrichment at positions 1–5 of the N-terminal region in the training data. Left panel: NP-positive dataset. Right panel: NP-negative dataset. (C) Sequence logos plots showing amino acid enrichment at positions 1–5 of the C-terminal region in the training data. Left panel: NP-positive dataset. Right panel: NP-negative dataset. Color codes for amino acids in (B,C): Red: acidic residues (D, E); Blue: basic residues (K, R, H); Purple: neutral residues (Q, N); Black: hydrophobic residues (A, V, L, I, P, W, F, M); Green: polar residues (G, S, T, Y, C). The height of each letter represents the relative frequency (information content) of the amino acid at the designated position.
Figure 2. Composite analysis of amino acid composition and conserved sequence motifs in neuropeptides (NPs) vs. non-neuropeptides (non-NPs). (A) Average percent composition of amino acid residues in NP and non-NP datasets. (B) Sequence logos plots showing amino acid enrichment at positions 1–5 of the N-terminal region in the training data. Left panel: NP-positive dataset. Right panel: NP-negative dataset. (C) Sequence logos plots showing amino acid enrichment at positions 1–5 of the C-terminal region in the training data. Left panel: NP-positive dataset. Right panel: NP-negative dataset. Color codes for amino acids in (B,C): Red: acidic residues (D, E); Blue: basic residues (K, R, H); Purple: neutral residues (Q, N); Black: hydrophobic residues (A, V, L, I, P, W, F, M); Green: polar residues (G, S, T, Y, C). The height of each letter represents the relative frequency (information content) of the amino acid at the designated position.
Life 15 01010 g002
Figure 3. Heatmap of log-odds ratios for dipeptide (2-mer) frequencies comparing NP-positive and NP-negative peptide sequences. Each cell represents a dipeptide, with the amino acid at position 1 on the y-axis and the amino acid at position 2 on the x-axis. Color legend: red shades indicate 2-mers that are overrepresented in NPs relative to non-NPs (positive log-odds values), while lighter red to white shades indicate underrepresentation in NPs (negative log-odds values). The color scale ranges approximately from −3 (underrepresented in NPs) to +3 (strongly enriched in NPs).
Figure 3. Heatmap of log-odds ratios for dipeptide (2-mer) frequencies comparing NP-positive and NP-negative peptide sequences. Each cell represents a dipeptide, with the amino acid at position 1 on the y-axis and the amino acid at position 2 on the x-axis. Color legend: red shades indicate 2-mers that are overrepresented in NPs relative to non-NPs (positive log-odds values), while lighter red to white shades indicate underrepresentation in NPs (negative log-odds values). The color scale ranges approximately from −3 (underrepresented in NPs) to +3 (strongly enriched in NPs).
Life 15 01010 g003
Figure 4. (A) Receiver operating characteristic (ROC) curves for individual classifiers and the ensemble models, evaluated using 10-fold cross-validation. The ensemble model shows the highest overall performance, followed closely by the Deep Learning (DL), Support Vector Machine (SVM), Extra Trees (ET), and Random Forest (RF) models. In contrast, the Decision Tree (DT) and k-Nearest Neighbors (KNN) models demonstrate lower AUC and sensitivity, indicating reduced classification reliability. (B) Bar chart comparing four key evaluation metrics: Matthews correlation coefficient (MCC), sensitivity (Sn, or true positive rate), specificity (Sp, or true negative rate), and area under the ROC curve (AUC) across all models. The x-axis represents the evaluation metrics, while the y-axis indicates the metric value ranging from 0 to 1. Colored bars correspond to each classifier as indicated in the legend. The ensemble model consistently achieves strong and balanced performance across all metrics, with notable improvements in MCC and specificity. SVM, DL, and ET models also perform competitively, while DT and KNN show weaker overall performance, especially in MCC.
Figure 4. (A) Receiver operating characteristic (ROC) curves for individual classifiers and the ensemble models, evaluated using 10-fold cross-validation. The ensemble model shows the highest overall performance, followed closely by the Deep Learning (DL), Support Vector Machine (SVM), Extra Trees (ET), and Random Forest (RF) models. In contrast, the Decision Tree (DT) and k-Nearest Neighbors (KNN) models demonstrate lower AUC and sensitivity, indicating reduced classification reliability. (B) Bar chart comparing four key evaluation metrics: Matthews correlation coefficient (MCC), sensitivity (Sn, or true positive rate), specificity (Sp, or true negative rate), and area under the ROC curve (AUC) across all models. The x-axis represents the evaluation metrics, while the y-axis indicates the metric value ranging from 0 to 1. Colored bars correspond to each classifier as indicated in the legend. The ensemble model consistently achieves strong and balanced performance across all metrics, with notable improvements in MCC and specificity. SVM, DL, and ET models also perform competitively, while DT and KNN show weaker overall performance, especially in MCC.
Life 15 01010 g004
Figure 5. Feature importance analysis. (A) Feature importance rankings from the Random Forest model, measured by percent increase in mean squared error (%IncMSE) and node purity (IncNodePurity). (B) Relative importance scores of top-ranked features from the XGBoost model. (C) SHAP summary plot of the top 15 features. The y-axis lists features sorted by importance (highest to lowest). The x-axis indicates SHAP values, which reflect the change in log-odds used to derive predicted probabilities. Colors represent feature values. Logistic1 consistently ranked highest across all methods, followed by other composite logistic features and CTD-derived descriptors, supporting their critical role in distinguishing neuropeptides from non-neuropeptides.
Figure 5. Feature importance analysis. (A) Feature importance rankings from the Random Forest model, measured by percent increase in mean squared error (%IncMSE) and node purity (IncNodePurity). (B) Relative importance scores of top-ranked features from the XGBoost model. (C) SHAP summary plot of the top 15 features. The y-axis lists features sorted by importance (highest to lowest). The x-axis indicates SHAP values, which reflect the change in log-odds used to derive predicted probabilities. Colors represent feature values. Logistic1 consistently ranked highest across all methods, followed by other composite logistic features and CTD-derived descriptors, supporting their critical role in distinguishing neuropeptides from non-neuropeptides.
Life 15 01010 g005
Table 1. Percent composition of amino acid based on physicochemical properties.
Table 1. Percent composition of amino acid based on physicochemical properties.
Physicochemical PropertyPositive Data (NPs)Negative Data (Non-NPs)
Average Length (amino acid residue)23.9326.93
% Charged Residues (DEKHR)24.4122.81
% Aliphatic Residues (ILV)16.0421.78
% Aromatic Residues (FHWY)12.8411.22
% Polar Residues (DERKQN)30.7928.25
% Neutral Residues (AGHPSTY)41.2735.01
% Hydrophobic Residues (CVLIMFW)27.9436.74
% Positively Charged Residues (HKR)12.9814.82
% Negatively Charged Residues (DE)11.437.99
% Tiny Residues (ACDGST)36.5832.01
% Small Residues (EHILKMNPQV)45.8553.04
% Large Residues (FRWY)17.5714.95
Table 2. Classification performance on the training dataset using 10-fold cross-validation.
Table 2. Classification performance on the training dataset using 10-fold cross-validation.
ModelACCMCCSnSpAUC95% ROC CI
SVM93.8190.8760.9400.9360.986[0.973–0.995]
KNN90.5700.8140.8700.9420.968[0.954–0.981]
DT88.5690.7720.8760.8950.877[0.853–0.899]
RF93.3240.8660.9300.9360.985[0.971–0.994]
ET93.7790.8760.9380.9370.986[0.973–0.996]
XGB91.6600.8330.9150.9180.972[0.963–0.984]
DL93.9180.8780.9390.9380.986[0.973–0.996]
Ensemble93.9780.8800.9390.9410.987[0.975–0.996]
Table 3. Comparison of performance on independent testing dataset 1.
Table 3. Comparison of performance on independent testing dataset 1.
MethodACCMCCSnSpAUC
NeuroPpred-Fuse0.9060.8130.8820.9300.958
PredNeuroP0.8970.7940.8860.9070.954
NeuroPred-FRL0.9000.8030.9460.8540.965
NeuroPIpred0.5360.0740.3310.7360.581
NeuroPred-CLQ0.9360.8750.8970.9750.988
EnsembleNPPred0.9400.8810.9620.9180.990
Table 4. Comparison of performance on independent testing dataset 2.
Table 4. Comparison of performance on independent testing dataset 2.
MethodACCMCCPrecisionRecallF1
PredNeuroP0.8640.7380.9350.7820.852
NeuroPred-FRL0.8610.7400.9600.7570.847
NeuroPpred-Fuse0.9050.8130.9060.9080.907
NeuroPred-PLM0.9220.8450.9070.9410.924
EnsembleNPPred0.9290.8590.9300.9290.929
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lertampaiporn, S.; Wattanapornprom, W.; Thammarongtham, C.; Hongsthong, A. EnsembleNPPred: A Robust Approach to Neuropeptide Prediction and Recognition Using Ensemble Machine Learning and Deep Learning Methods. Life 2025, 15, 1010. https://doi.org/10.3390/life15071010

AMA Style

Lertampaiporn S, Wattanapornprom W, Thammarongtham C, Hongsthong A. EnsembleNPPred: A Robust Approach to Neuropeptide Prediction and Recognition Using Ensemble Machine Learning and Deep Learning Methods. Life. 2025; 15(7):1010. https://doi.org/10.3390/life15071010

Chicago/Turabian Style

Lertampaiporn, Supatcha, Warin Wattanapornprom, Chinae Thammarongtham, and Apiradee Hongsthong. 2025. "EnsembleNPPred: A Robust Approach to Neuropeptide Prediction and Recognition Using Ensemble Machine Learning and Deep Learning Methods" Life 15, no. 7: 1010. https://doi.org/10.3390/life15071010

APA Style

Lertampaiporn, S., Wattanapornprom, W., Thammarongtham, C., & Hongsthong, A. (2025). EnsembleNPPred: A Robust Approach to Neuropeptide Prediction and Recognition Using Ensemble Machine Learning and Deep Learning Methods. Life, 15(7), 1010. https://doi.org/10.3390/life15071010

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop