Next Article in Journal
Identification of Candidate Genes and Functional Pathways Associated with Body Size Traits in Hulunbuir Sheep Through GWAS Analysis
Previous Article in Journal
Resequencing and Transcriptome Analyses Reveal Variations and Expression Patterns of the RR Gene Family in Cucumber
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Biological Prior Knowledge-Embedded Deep Neural Network for Plant Genomic Prediction

1
Agricultural Bioinformatics Key Laboratory of Hubei Province, College of Informatics, Huazhong Agricultural University, Wuhan 430070, China
2
Hubei Hongshan Laboratory, Wuhan 430070, China
3
State Key Laboratory of Plant Diversity and Specialty Crops, Wuhan Botanical Garden, Chinese Academy of Sciences, Wuhan 430074, China
4
School of Computer, BaoJi University of Arts and Sciences, Baoji 721016, China
5
Department of Biostatistics, School of Public Health, Peking University, Beijing 100191, China
6
Beijing International Center for Mathematical Research, Peking University, Beijing 100871, China
*
Authors to whom correspondence should be addressed.
Genes 2025, 16(4), 411; https://doi.org/10.3390/genes16040411
Submission received: 24 February 2025 / Revised: 23 March 2025 / Accepted: 26 March 2025 / Published: 31 March 2025
(This article belongs to the Section Plant Genetics and Genomics)

Abstract

:
Background/Objectives: Genomic prediction is a powerful approach that predicts phenotypic traits from genotypic information, enabling the acceleration of trait improvement in plant breeding. Traditional genomic prediction methods have primarily relied on linear mixed models, such as Genomic Best Linear Unbiased Prediction (GBLUP), and conventional machine learning methods like Support Vector Regression (SVR). Traditional methods are limited in handling high-dimensional data and nonlinear relationships. Thus, deep learning methods have also been applied to genomic prediction in recent years. Methods: We proposed iADEP, Integrated Additive, Dominant, and Epistatic Prediction model based on deep learning. Specifically, single nucleotide polymorphism (SNP) data integrating latent genetic interactions and genome-wide association study results as biological prior knowledge are fused to an SNP embedding block, which is then input to a local encoder. The local encoder is fused with an omic-data-incorporated global decoder through a multi-head attention mechanism, followed by multilayer perceptrons. Results: Firstly, we demonstrated through experiments on four datasets that iADEP outperforms existing methods in genotype-to-phenotype prediction. Secondly, we validated the effectiveness of SNP embedding through ablation experiments. Third, we provided an available module for combining other omics data in iADEP and propose a novel method for fusing them. Fourthly, we explored the impact of feature selection on iADEP performance and conclude that utilizing the full set of SNPs generally provides optimal results. Finally, by altering the partition of training and testing sets, we investigated the differences between transductive learning and inductive learning. Conclusions: iADEP provides a new approach for AI breeding, a promising method that integrates biological prior knowledge and enables combination with other omics data.

1. Introduction

Genomic prediction (GP) was initially proposed as a new method of selection decision that leverages high-density genomic markers [1]. GP builds a prediction model using a training set consisting of known phenotypes and genotypes. With this model, individuals with known genotypes but unknown phenotypes can be predicted, enabling early individual selection based on genomic estimated phenotypic values. This approach reduces the cost of field phenotype identification, shortens breeding years, and improves genetic gain [2].
Traditional genomic prediction methods are primarily based on statistical models, which have been divided into two categories: the direct method and the indirect method [3]. The direct method treats individuals as random effects and constructs their genetic relationship matrix as the covariance matrix based on genetic information from the training and testing sets [4]. The phenotype values for the testing set are obtained by solving the mixed linear model, represented by methods such as genomic best linear unbiased prediction (GBLUP) [5]. The indirect method estimates markers’ effects on the training set [6]. The phenotype values of the testing set are obtained by accumulating the marker’s effects based on the training set. Common methods include BayesA, BayesB, BayesC, etc. [1,7]. However, the traditional GP methods face several challenges. Specifically, the direct method requires the genotype data of the testing set, necessitating the reconstruction of the GP model for new individuals. The indirect method relies on the prior distribution of the biological marker effect and its variance, which limits its ability to capture the genotype-phenotype relationship. Furthermore, both methods struggle to effectively capture nonlinear relationships or handle multi-dimensional and multi-omic data. To overcome these limitations, machine learning (ML) methods have been further developed [8]. Random forest (RF), ridge regression (RR), support vector regression (SVR), and other methods have yielded great prediction accuracy in breeding wheat for rust resistance [9]. A comparison of the prediction accuracy of different machine learning methods for the reproductive traits of Nellore cattle reveals that SVR is the most suitable method for these traits [10].
Deep learning (DL), a subfield of ML, is an advanced technique that excels in handling high-dimensional big data and analyzing involved interaction or relationship [11,12]. The deep convolutional neural network (DeepGS), the first DL model in GP proposed by Ma et al., is a model composed of convolutional neural networks (CNN) and uses wheat DArT (Diversity Array Technology) markers to predict phenotypes [13]. Similarly, based on CNN, deep neural network genomic prediction (DNNGP) takes high-dimensional SNPs (single nucleotide polymorphisms) markers through PCA (principal component analysis) as partial input and integrates multi-omic data to jointly predict phenotypes [14]. In contrast to the CNN architecture, GPformer is a modified model of the Encoder in Transformer, which can greatly expand the features extraction receptive field and adopts an attention mechanism to capture global information [3,15]. Furthermore, the integration of genome-wide association study (GWAS) results as prior knowledge can improve GPformer prediction performance.
So far, the markers (usually SNPs) are commonly encoded by 0, 1, and 2 to distinguish homozygous or heterozygous states in the aforementioned DL-based methods. However, this encoding approach lacks discrimination for different SNPs, which contradicts the inductive bias of local similarity in CNN. DNNGP allows for the application of a variety of omics data to predict. However, at the input layer, DNNGP only employs one data type, which does not fully utilize the information of multi-omic. Additionally, the auto-correlation attention mechanism used in GPformer is derived from the series periodicity, introducing an inductive bias inconsistent with biological markers although this mechanism effectively handles long-range sequences [16]. Moreover, the effects of genes themselves, including additive effects, dominant effects, and epistatic effects (interactions between genes), have been partially overlooked [17].
To address these issues, we proposed the Integrated Additive, Dominant, and Epistatic Prediction (iADEP) model, a deep learning-based model designed to improve genomic prediction performance by integrating biological prior knowledge and optional multi-omic data. Specifically, iADEP mainly consists of three parts: SNP embedding, local encoder-global decoder and multilayer perceptrons (MLP), involving ResNet [18] and multi-head attention mechanism [15]. In the SNP embedding block, SNP data are inputted and integrated with GWAS and Latent Interaction Testing (LIT) [19] as biological prior knowledge to generate SNP embedding. In the global decoder block, SNP data can be inputted and merged with the output of the local encoder block, and additional omics data can be inputted and merged with the SNP output. The prediction performance of iADEP was compared with other representative methods like GBLUP, RR, SVR, RF, XGBoost (eXtreme Gradient Boosting), and DNNGP from various models in diverse datasets. The results demonstrate that iADEP is a promising method that can successfully predict multiple traits in diverse datasets. Subsequently, we validated the effectiveness of SNP embedding as biological prior knowledge by ablation experiments. Furthermore, we evaluated the performance of incorporating other omics data into the global decoder block. Finally, we investigated the impact of feature selection on iADEP and explored the distinctions between transductive learning and inductive learning.

2. Materials and Methods

2.1. Dataset Used for Genomic Prediction

Five datasets in which the genetic markers are SNPs and have large sample sizes were selected in this study, including millet827, spruce1722, soy5014, maize5820, and springbarley2463. Detailed information on the SNP and traits for all datasets in provided in Table S1. Millet827, consisting of 827 foxtail millet cultivars collected from China, was used to explore genotype-dependent microbial effects [20]. Each line has a sequence of 161,562 SNPs and can be downloaded from CropGS-Hub [21]. The SNPs were filtered out by PLINK [22] with minor allele frequency (MAF) < 0.05 and genotype calling rate < 0.1, yielding a total of 75,780 SNPs. This dataset consists of twelve agronomic traits, including six growth traits and six yield traits. Among these traits, the hundred kernel weight (HKW) trait was excluded due to its extremely low heritability ( h s n p 2 = 0.0069 ).
Spruce1722, derived from the SmartForests project team [23], consists of 1722 white spruce samples with a sequence of 6930 SNPs without missing data and three traits. The SNPs were filtered out by PLINK with MAF < 0.05, yielding a total of 5679 SNPs.
Soy5014 was obtained from the SoyNAM population, which contains recombinant lines (RILs) derived from 40 biparental populations [24]. Soy5014 consists of 5014 samples with a sequence of 4234 SNPs and three traits. Spruce1722 and Soy5014 datasets can be downloaded from the previous study [25].
Maize5820 is a maize NCII population of 5820 F1 hybrids created from the cross of 30 diverse elite paternal lines and maternal inbred lines which were a subset of the CUBIC population [26]. The SNPs called in all 224 inbred lines by whole-genome resequencing were filtered out by PLINK with MAF < 0.05 or expected missing rate > 10% and was pruned with an LD threshold of 0.3. The maize5820, containing the resulting 156,269 SNPs and 18 agronomic traits, can be downloaded from CropGS-Hub or a previous study [27].
Springbarley2463 was collected from 2628 plots of 565 spring barley lines from Nordic Seed A/S, including genotype, phenotype, and metabolome data, which are available in a public accessible repository [28]. Each line had 3889 SNP markers and five malting quality traits. In addition, 24,018 metabolomic features were obtained for each sample from nuclear magnetic resonance spectra for wort samples produced from each experimental plot. The detailed descriptions of these traits and metabolomic features are reported in Guo et al. [29]. Samples with missing phenotypic data were filtered out, and missing SNP data were imputed using mean imputation followed by rounding. Finally, spring barley dataset consists of 563 lines and 2463 samples. This dataset was only used to study the application of iADEP in predicting phenotypes incorporating other omics data.

2.2. Heritability Estimation

The heritability for each trait was estimated using an SNP-based method, denoted as h s n p 2 [30]. The genetics complex trait analysis (GCTA) was used to estimate the phenotypic variance explained by SNP through genome-based restricted maximum likelihood (GREML) method [31]. The estimated heritability provides a reference significance for the genomic prediction results.

2.3. iADEP Algorithm

iADEP consists of three parts: SNP embedding, local encoder-global decoder, and MLP regression. SNP embedding block distinguished different SNPs based on biological prior knowledge of GWAS and LIT. In the local encoder block, it performed dimensionality reduction and encoded locally fused information, while the global decoder block employed multi-head attention mechanism to extract the globally fused features. In the global decoder block, it is possible to add other omics data to fuse genomic and other omics features. Finally, using MLP as regressor, the predictive results were obtained. Each part is described as follows.
In classical quantitative genetic methods, the genotypic value can be decomposed into the sum of additive, dominant, and epistatic parts as follows [32]:
G = A + D + I
where G is the genotypic value, A is the sum of the additive effects attributable to the separate loci, D is the sum of the dominance deviations, and I is the sum of interaction deviation or epistatic deviation.
Based on this hypothesis, we employed an embedding methodology to encode each SNP that contain these three parts. Specifically, GWAS based on a linear mixed model can capture the additive effects between SNPs and phenotypes. Each SNP encoded as heterozygous state may exhibit within-locus interactions, known as dominant effect. Additionally, the interactions between different SNPs can be captured using the LIT algorithm [19]. LIT uses a flexible kernel-based framework to test for latent genetic interactions in genome-wide association studies, yielding output formats that are similar to GWAS. Therefore, the representation of a SNP depends on which homozygous and whether heterozygous and its initial representation can be encoded as follows:
log G W A S   p v a l u e ,   0 , log L I T   p v a l u e ,   i f   S N P = 1 0,1 , log L I T   p v a l u e ,   i f   S N P = 0 log G W A S   p v a l u e ,   0 , log L I T   p v a l u e ,   i f   S N P = 1
Figure 1A provides a detailed description of the process of SNP embedding, where GWAS can be performed in R package “rrBLUP” [6] or EMMAX (beta) [33] software, and LIT can be performed in R package “lit” [19].
Local encoder-global decoder mainly consists of a series of residue connections and a multi-head attention mechanism. The output of SNP embedding can be considered as making each SNP contain more information. In contrast to traditional CNN architectures employed in previous studies, ResNet block demonstrates advantages in preventing gradient explosion and vanishing, while concurrently preserving more low-level features [18]. Therefore, we extracted local features by ResNet block. The interaction information between local features is extracted through multi-head self-attention mechanism. Multi-head self-attention creates a high-level understanding of the data, including interaction among local features, as illustrated below:
A t t e n t i o n Q ,   K ,   V = s o f t m a x Q K T d k V
where Q, K, and V are query, key, and value, respectively, and d k is the dimension of K. Q, K, and V are the same vectors in self-attention [15]. Eventually, a high-level feature containing local information and interaction information is obtained in the local encoder block.
Residue connection and multi-head attention mechanism are also employed in the global decoder block. First, we applied two-layer linear mappings with LeakyReLU as the activation function to the raw SNP, which helps simplify the global information to create a low-dimensional representation. This low-dimensional representation with the output of local encoder block is integrated through a multi-head attention layer to fuse local and global information. Additionally, residual connections were employed to supplement the fused features with global information (Figure 1B). The LeakyReLU activation function is as follows [34]:
y i = x i ,   i f   x 0 x i a i , i f   x < 0
where i represents the index of the input or output, and a i is a fixed parameter, defaulting to 100.
If other omics data are available, the global decoder block provides a module for combining them. Taking metabolomic features as an example, similar to raw SNP data, a two-layer linear mapping is used to obtain a low-dimensional representation. This low-dimensional representation is then combined with the output of the global decoder for the SNP data using the following feature fusion method:
F u s i o n O u t p u t S N P ,   O u t p u t m e t = O u t p u t m e t + S i g m o i d M a x O u t p u t S N P ,   O u t p u t m e t
where the O u t p u t S N P is the output of the global decoder for the SNP data and O u t p u t m e t a is the low-dimensional representation. The reason for employing such fusion approach is that, according to the central dogma, the features of the metabolome manifest later than genome and are closer to phenotypic traits. Therefore, the features of the metabolome not only exhibit intersections with features on the genome, but also contain more hidden phenotypic information. To capture these metabolic features more effectively, we used non-linear function max and activation function sigmoid to avoid redundancy in the features. Additionally, residue connection by adding the metabolomic feature to capture more metabolomic information.
The final features were used to predict the phenotype value through a three-layer MLP that incorporates LeakyReLU activation function and dropout layers. We chose mean squared error (MSE) as the initial loss function and used L2 regularization to adjust the training process to reduce generalization errors. Figure 1 shows the entire architecture of iADEP.

2.4. Model Construction of Existing Representative Methods

To assess the effectiveness of iADEP more objectively, we selected representative methods of different types on the same dataset. Specially, we selected the classical method GBLUP based on linear mixed model, the linear regression method ridge regression, random forest as the representative method of tree models, SVR as the classical ML method, XGBoost as the ensemble learning method, and the latest deep method DNNGP. Here is a brief introduction to these methods.
GBLUP, genomic best linear unbiased prediction, is a linear mixed model derived from BLUP. It replaces the pedigree relationship matrix with a genomic relationship matrix [5], which is formulated as follows:
G = Z Z T 2 p j 1 p j
where Z represents the standardized genotype matrix, and p j denotes the allele frequency of locus j . GBLUP can be implemented using the R package “rrBLUP” [6].
RR, ridge regression, introduces a regularization term to the ordinary least squares regression to improve the stability and generalization ability of the model [35]. This regularization term penalizes the sum of squared coefficients. The formula for RR is as follows:
y = X β ^ + λ j = 1 p β j 2
where y is the predicted response variable, X is the design matrix of predictors, β ^ is the vector of estimated regression coefficients, λ is the ridge parameter, j is an index used to refer to the j -th coefficient in the regression coefficient vector β ^ , β j is the coefficients of the predictors, and p is the number of predictors.
RF, random forest, is an ensemble tree model that combines multiple decision trees to make predictions [36]. It achieves diversity by training each decision tree on a random subset of data and features. The final prediction is obtained by voting or averaging the predictions of individual tress.
SVR, support vector regression, is a non-linear regression model based on support vector machines [37]. It maps the training sample to a high-dimensional feature space and constructs an optimal hyperplane for regression prediction. SVR establishes a boundary by finding support vectors to minimize the error between predicted values and observed values. If linearity is not possible in the high-dimensional space, it can introduce the relaxation variables ε i   0 . The formula is as follows:
min w , b 1 2 w T w + C i = 1 N ε i s . t .   y i   ( w T x i + b ) 1 ε i ,   ε i 0
where w is the weight vector, b is the intercept, C is the regularization constant, N is the sample size, i is the index of the sample, ε i is the relaxation variables, and y i and x i are paired training set features and labels. RR, RF, and SVM can be implemented in Python3.10.9 using “scikit-learn1.2.0” [38].
XGBoost, eXtreme Gradient Boosting, is based on the gradient boosting framework and combines multiple weak learners to create a strong ensemble model [39]. XGboost can be implemented in Python using “xgboost2.0.0”.
DNNGP, deep neural network genomic prediction, is based on deep CNN [14]. It contains one input layer, three convolutional layers, one batch normalization layer, two dropout layers, one flattening layer, one dense layer, and one output layer. The core block convolutional layer is presented as follows [40]:
s t = f k t = x k t x f x
where k represents the kernel size, convolution is the transformation of f into s ( t ) . In DNNGP, the kernel size is set to 4, which is much smaller than the number of SNPs. To address this, PCA is applied to the SNP data, and the top 95% of feature variance is extracted. We have also adopted this approach when using DNNGP. On springbarley2463 dataset, the CNN kernel size is much smaller than the length of the metabolomic features (24,018). Thus, we applied PCA to the metabolomic features, selecting the top three PCs that accounted for 99% of the feature variance. These three features were then concatenated with the SNP data.

2.5. Model Evaluation

In this study, we implemented a 5-fold cross-validation scheme for a more rigorous evaluation of different methods. Each dataset was randomly divided into five folds, with four folds used as the training set and the remaining fold as the testing set. To make sure that phenotype data were not leaked owing to the fact that the p-values of GWAS and LIT are functions of the phenotype, GWAS and LIT were only performed on the training set, where only the springbarley2463 dataset is based the averaged values of different plots on the same line. We used Pearson correlation coefficient (PCC) as an evaluation metric and there are two schemes to evaluate prediction performance. The first scheme involves using PCC between the observed phenotypes and corresponding predicted values on each test fold. The average and standard error of the PCC are used as performance metrics. The second scheme involves recording the predicted values for all test sets and calculating a single PCC across all observed phenotypes. In this study, only the second scheme is used to investigate the differences between transductive learning and inductive learning, while the first scheme is used for all other experiments. The formula for PCC is as follows:
P C C = C o v X , Y σ X σ Y
where C o v X , Y is the covariance between variable X and Y , σ X is the standard deviation of variable X , σ Y is the standard deviation of variable Y , X and Y correspond to phenotype observations and predictions, respectively.

3. Results

3.1. Overview of iADEP Architecture

iADEP consists of three parts: SNP embedding, local encoder-global decoder, and MLP regression (Figure 1). SNP embedding block distinguishes different SNPs based on biological prior knowledge of GWAS and LIT (Figure 1A). The local encoder block contains ResNet, multi-head self-attention layer and layer normalization, which are used to perform dimensionality reduction and encode locally fused information. The global decoder block employs a multi-head attention mechanism to extract the globally fused features and residue connection to retain partial global information. Additionally, the global decoder block provides a module that allows the addition of other omics data to fuse genomic and other omics features. Finally, the prediction results are obtained by an MLP as a regressor (Figure 1B).
The prediction performance is evaluated using the PCC between the predicted values and the observed values. The overall prediction performance of iADEP is positively correlated with heritability (Figure S1). SNP-based heritability ( h s n p 2 ), which is smaller than h 2 , was calculated to provide a reference for GP [41,42]. The heritability differences between different species may be due to differences in planting environments, which will not be discussed here. In general, iADEP exhibits higher prediction performance for traits with higher h s n p 2 . Specially, the days to anther (DTA) trait ( h s n p 2 = 0.8263 ) in maize5820 dataset achieves the highest prediction performance of 0.940, whereas the diameter at breast height (DBH) trait ( h s n p 2 = 0.1674 ) in spruce1722 dataset has the lowest prediction performance of 0.286.

3.2. Prediction Performance of iADEP and Comparison with Other Methods

We compared the prediction accuracy for GBLUP, RR, SVR, RF, XGBoost, DNNGP and iADEP on four datasets: millet827, spruce 1722, soy5014 and maize5820. Firstly, we compared all methods on the millet827 dataset, which is a relatively small dataset, containing six growth traits and five yield traits. In terms of growth traits, iADEP demonstrated the best overall prediction performance, outperforming all other methods except for the main stem height (MSH) trait and fringe neck length (FNL), where their performance was lower than that of RF and SVR. GBLUP and RF were found to be suboptimal methods. However, for example, the performance of GBLUP and RF was 4.8% and 4.4% lower than iADEP in main stem width (MSW) trait, respectively (Figure 2A). In terms of yield trait, iADEP also exhibited the superior comprehensive prediction performance across all traits, with the exception of the grain number per spike (SGN) trait, where it exhibited a slightly lower performance compared to GBLUP (0.4%) (Figure 2B).
We then tested and determined the prediction performance of all methods on another relatively small dataset, spruce1722, which contains three traits. Across the three traits, iADEP consistently achieved the highest prediction performance, followed by SVR and GBLUP (Figure 2C). Compared to SVR, iADEP showed the highest improvement in the diameter at breast height (DBH) trait, with an increase of 1.9%. The next highest improvement was observed in the height (HT) trait, with an increase of 1.1%, while there was almost no improvement in the wood density (DE) trait (0.1%).
The third dataset, soy5014, is a relatively large dataset that includes three traits. In terms of the height (HT) and yield (YLD) traits, iADEP demonstrates superior predictive performance compared to the second-best methods, XGBoost and RF, with improvements of 0.3% and 0.7%, respectively. However, for the time to R8 developmental stage (R8) trait, iADEP is slightly lower than GBLUP by 0.2% (Figure 2D). Overall, on the soy5014 dataset, iADEP exhibited the best overall performance, although its advantage over smaller dataset is not as significant.
The final dataset, maize5820, is the largest dataset containing the most samples, SNPs and traits. It includes 18 traits that can be categorized into three groups: flowering traits, plant architecture traits and yield traits. In terms of plant architecture traits, iADEP showed the best prediction performance except for the ear leaf width (ELW) trait, followed by SVR. For flowering traits, SVR exhibited the highest prediction performance, followed by iADEP. As for yield traits, iADEP, SVR and GBLUP all demonstrated satisfactory performance (Figure 3). Together, for 9 out of 18 traits, iADEP achieves the best performance, although the differences among the methods are minimal, which could be attributed to the high heritability of each trait on this dataset, indicating their inherent predictability.
In summary, for most of the traits tested, iADEP performs better than other methods with small to moderate magnitudes. The statistical analysis results are shown in Figure S5.

3.3. Ablation Experiments of SNP Embedding

To validate the effectiveness of biological prior knowledge, we conducted ablation experiments focusing on SNP embedding. Two conventional approaches were designed for this purpose (Figure S2):
Approach 1: one-hot encoding was employed, whereby the three possible states of SNPs, namely [1, 0, −1], were encoded as [[1, 0, 0], [0,1, 0], [0, 0, 1]], respectively.
Approach 2: Non-encoding was employed, but an additional channel was introduced to incorporate the SNPs information into the ResNet architecture. In this case, the states [1, 0, −1] were encoded as [[1], [0], [−1]], which is similar to methods such as DNNGP.
Ablation experiments were performed on four datasets to assess the effectiveness of SNP embedding (Table 1). Firstly, validation was conducted on millet827 dataset. In terms of growth traits, both the one-hot encoding and the non-encoding approaches resulted in a decrease in the prediction performance of iADEP. The most significant decrease was observed in the main stem width (MSW) trait, with a decrease of 2.7% and 3.2% for the two encoding approaches, respectively. As for the yield traits, only the main stem panicle weight (MSPW) trait showed a 1.2% improvement when employing one-hot encoding. However, for the other traits, both encoding methods led to a decrease in prediction performance.
On the spruce1722 dataset, SNP embedding consistently showed the highest prediction performance across the three traits, while one-hot encoding and non-encoding resulted in performance declines. Overall, one-hot encoding led to a decrease of 1% to 2%, while non-encoding resulted in a decrease of 2% to 6% in performance.
On the soy5014 dataset, SNP embedding also exhibits the highest prediction performance. Specifically, for the HT and R8 traits, both one-hot encoding and non-encoding result in significant performance declines, ranging from 1.8% to 4.4%. However, for the YLD trait, the differences among the three methods are minimal, with only a slight decrease of 0.3% observed.
On the maize5820 dataset, both one-hot encoding and non-encoding led to a decrease in prediction performance, and the results showed a larger standard deviation comprehensively. For flowering traits, this decrease is relatively small, ranging from 0 to 1%. However, for plant architecture traits, the decrease is more pronounced, except for the ear leaf width (ELW) trait, where the decrease is minimal. Other plant architecture traits show a decrease of 1% to 3%. As for yield traits, except for the kernel number per ear (KNPE) trait, which shows a slight decrease, the decreasing trend is more significant for other yield traits.

3.4. Genomic Prediction Incorporating Other Omics Data

In the global decoder block, we designed a module and feature fusion approach for other omics data (Figure 1B). In existing methods, such as DNNGP, multi-omic data including SNP are treated as markers without differentiation. Traditional machine learning methods primarily handle unidimensional data rather than multi-dimensional data. Therefore, for both traditional machine learning methods and DNNGP, we employed a multi-omic data integration strategy through feature concatenation. In contrast, the GBLUP method is incompatible with multi-omic data as it requires the construction of a genomic relationship matrix, which cannot be directly applied to multi-omic data. We compared the prediction performance of iADEP and other methods when combining other omics data on the springbarley2463 dataset, which contains SNP, metabolomic and phenotypic data.
On the Springbarley2463 dataset, iADEP achieved the highest prediction performance across all traits tested, followed by RF (Figure 4). The statistical analysis results are shown in Figure S6. Particularly, iADEP exhibited a significant advantage in predicting the beta glucan (BG) trait, surpassing RR by 14.4%. It also performed well in predicting the wort color (WC) and extract yield (EY) traits. For the wort viscosity (WV) and filtering speed (FS) traits, although the improvement of iADEP was not as pronounced, it still outperformed RF by 0.75% and 0.33%, respectively. Together, iADEP outperforms other state-of-the-art methods when conducting genome prediction based on multi-omic data.

3.5. Effects of Feature Selection on iADEP

In previous studies, researchers have explored the potential of feature selection to improve the performance of genomic prediction. In terms of feature selection, these studies can be categorized into two approaches. The first approach randomly selects a certain percentage of markers, such as selecting randomly 20% of the markers [14,43]. The second approach considers GWAS-associated markers, such as selecting the top-k GWAS-associated markers [25,44]. In this study, we simultaneously considered markers associated with GWAS or LIT. We extracted the top 20%, 10%, and 5% SNPs associated with GWAS or LIT due to the variations in the number of SNPs across different datasets. Then, the union of these SNPs was taken to explore the impact of feature selection on prediction performance.
On the millet827 dataset, the impact of feature selection on prediction results can be roughly categorized into three types (Figure 5A). The first type is characterized by a positive correlation between model performance and the number of markers used. These traits included TSLL, TSLW, MSW, MSPD, and SGD trait, with TSLW showing the strongest trend (Figure 5A). The second type exhibits minimal changes with the variation in the number of markers (change less than 0.5%), such as the FNL trait. The third type presents a potential improvement in prediction performance as the number of markers decreased, although these improvements were generally small. For example, the MSSN trait showed a slight improvement, while only the MSH trait exhibited a significant increase. Overall, when the number of markers decreased to top 5–10%, except for the MSH trait, the performance tended to decline (Figure 5A).
On the spruce1722 and soy5014 datasets, all six traits exhibited a declining trend. Among them, the YLD trait in soy5014 dataset showed a slower decline, while the remaining traits displayed a sharper trend (Figure 5B,C).
On the maize5820 dataset, the overall prediction performance exhibited minimal changes with the variation in the number of markers, for example, all flowering traits showed a variation of less than 0.2%. Among other traits, the LBT and KNPR traits demonstrated a declining trend, while for the remaining traits, there might be a slight improvement in performance as the number of markers decreased, as observed in the case of the ED trait (Figure 5D). The results of all datasets are recorded in Table S4.
Together, for three out of the four datasets, model performance improves with the increase in SNPs used, which might be attributed to the number of markers in the dataset and SNP embedding. For maize5820 dataset, it shows no significant correlation between model performance and the number of SNP used. Out of the 35 traits examined, 23 traits show optimal prediction performance when all SNPs are used, 3 traits give optimal performance when the top 20% or top 10% SNPs are used, and the other 9 traits show minimal performance changes upon feature selection. Hence, in the majority of cases, selection of SNP does not improve prediction performance. Instead, iADEP tends to perform best when all markers are inputted as a whole.

3.6. Impact of Transductive Learning and Inductive Learning on Genomic Prediction

Transductive learning and inductive learning are terms used in ML, and they are frequently discussed in graph neural networks. In transductive learning, the models predict the labels of the training samples it has already seen, while the models of inductive learning predict the labels of unknown samples [45]. In traditional methods, GBLUP needs to construct a genetic relationship matrix based on the training and testing set when building the model, making it similar to transductive learning. However, iADEP uses a trained model to predict labels for unknown samples, thus falling in the paradigm of inductive learning.
To explore the differences between these two learning paradigms, we conducted 100-fold cross-validation on the millet827 dataset (Figure S3). In this setup, GBLUP still sees and uses all the samples, while iADEP has seen 99% of the samples, almost reaching a uniform level. We collected the predicted values for all test sets and calculated the prediction performance across the entire dataset. Figure 6 illustrates the changes in performance for both methods. Compared to the five-fold cross-validation, GBLUP showed only a modest improvement of around 1% for all traits, whereas iADEP exhibited significant improvement. This indicates that GBLUP is insensitive to the training and testing split of the samples since it needs to see all the samples, while the performance of iADEP is influenced by the partition of the training set.

4. Discussion

4.1. The Potential Role of SNP Embedding

With the recent breakthroughs in artificial intelligence, researchers are applying deep learning methods to build GP models. Currently, deep learning methods used for GP primarily rely on CNN architectures. For example, Zingaretti et al. built a CNN-based deep learning model that outperformed traditional linear statistical models in predicting traits with epistatic variances in allopolyploid strawberry species and blueberries [46]. Wang et al. built a DNNGP model containing three CNN layers, one batch normalization layer, and two dropout layers to predict quantitative traits in plants [14]. These CNN-based GP models have a common limitation. CNN models are based on two inductive biases: locality and translation equivariance [47], which are not compatible with the nature of biological markers, especially SNP (Figure S4). To address this issue, we introduced SNP embedding, an encoding strategy designed to be more consistent with biological logic.
Specifically, we proposed SNP embedding that integrates GWAS and LIT (Figure 1A). The primary potential role of SNP embedding is to differentiate SNPs at different loci through the statistical methods of GWAS and LIT. Therefore, in ablation experiments, the design of one-hot encoding and non-encoding approaches (both of which fail to differentiate SNPs) led to a decline in the prediction performance of nearly all traits compared to SNP embedding (Table 1 and Figure S2). Another potential effect of SNP embedding is to serve as a biologically meaningful form of pre-training, guiding the model toward a more accurate learning direction. In the feature selection experiments, we found that only a few traits exhibited improved prediction performance through feature selection in iADEP. This may be attributed to the fact that before model training, we have already marked different SNPs by SNP embedding. For instance, some uncorrelated SNPs may have initial embeddings close to [0, 0, 0], and their influence during the forward propagation of deep learning may be minimal, but not entirely disregarded.

4.2. Scalability of Multi-Omic Data Fusion

In the local encoder-global decoder block, we employed two rounds of multi-head attention mechanisms to extract local and global interaction information. In the global decoder block, we provided an available module for combining other omics data. We compared iADEP with other methods on the springbarley2463 dataset, excluding GBLUP due to its incompatibility with other omics data. Following the central dogma, we designed a method to fuse SNP features and metabolomic features. In this study, we only tested the combination of metabolomic data, but this block can be extended to combine additional information beyond omics data. For instance, in the future when more environmental data is collected, we could supplement the model with environmental data and design various fusion methods for combining genotypic and environmental data. This aligns with the purpose of the integrated genomic-enviromic prediction (iGEP) model proposed by Xu et al. [48].

4.3. Differences Between Transductive Learning and Inductive Learning

We explored the differences between transductive learning and inductive learning in GP. GBLUP is more like transductive learning as it requires accessing both training and testing samples to create a genomic relationship matrix. It is insensitive to the partitioning between training and testing data, making it a stable method and one of the most commonly used approaches in animal and plant breeding. However, its drawback is that for a new sample, it has to rebuild the model and make new prediction because the genomic relationship matrix needs to be reconstructed. On the other hand, machine learning and deep learning methods like iADEP are examples of inductive learning. They learn from the training set and apply that knowledge to unknown samples. Therefore, they are sensitive to the training sample size. Once a well-trained model is built, it can be applied to other scenarios, making it more flexible in practical applications. Moreover, the inductive learning framework of deep learning provides a strong foundation for future development, particularly in interpretability, as it excels at summarizing and generalizing information. Many biologists prefer to explore the mechanisms at the genetic level, while deep learning methods appear more like black box models. However, with the development of interpretable methods in deep learning, such as Shapley Value [49], Gradient-weighted Class Activation Mapping (GradCAM) [50], it is possible to identify the features that models considered most significant, thereby providing new biological understanding and novel research direction.

4.4. Limitations of iADEP

Although deep learning methods like iADEP have numerous advantages in terms of prediction performance and application scenarios, they also come with several limitations. Firstly, the model-building process is less flexible and relies on the empirical setting of hyperparameters. In iADEP, specific hyperparameters include learning rate, training epochs, dropout rate, activation functions, the head number of multi-head attention mechanism, and in the case of ResNet, the kernel size of the first residual block and the number of residual blocks. These settings are primarily adjusted based on the number of SNPs in the dataset and the range of phenotypic values for different traits. For example, millet827 dataset with a larger number of SNPs may require larger kernel sizes and more residual blocks to reduce to length of SNPs to reduce the dimension of SNPs. Thus, the training process is complex and involves more human intervention although a well-trained model is flexible in applications. In comparison, methods like GBLUP and SVR have simpler model-building processes and require fewer parameter settings. The hyperparameters of these methods are shown in Tables S2 and S3. Secondly, iADEP is sensitive to the sample size of the training data. Having a larger number of training samples often leads to better prediction results, as evident from the experiments with 100-fold cross-validation. Therefore, iADEP has higher requirements for training sample size, which might limit its application in practical breeding scenarios. However, the data-demanding feature of iADEP will profit from and benefit from this era of big data, especially the collection and organization of big data and large projects for crop breeding. Thirdly, deep learning methods need higher hardware requirements. In this study, for millet827, spruce1722 and soy5014 datasets, which have smaller sample sizes or SNP numbers, we conducted training and testing on CPUs. For maize5820, we performed training and testing on an A100 GPU. Compared to GBLUP, iADEP is significantly more time-consuming.

5. Conclusions

In summary, we built a deep learning model, iADEP, which effectively integrates biological prior knowledge and allows for the incorporation of other omics data. Our results show the prediction performance of iADEP to be promising compared to existing methods on four datasets from various plant species. Through ablation experiments, we validated the effectiveness of SNP embedding. Moreover, we provided a module to combine information from other omics data, even environmental data. Additionally, we investigated the impact of feature selection on iADEP and found that for the majority of the traits tested, model performance is better when using full SNP sites. Finally, we explored the differences between transductive learning and inductive learning in GP. We found that increasing the size of the training set significantly improved the performance of iADEP. This behavior indicates that large datasets coming out in the future will empower such deep learning-based approaches. In the era with great advances of AI, the insights gained by developing, evaluating, and applying iADEP in this study would contribute to the future widespread application of deep learning in plant breeding. Genomic prediction can facilitate the selection of superior plant varieties in contemporary breeding, thereby enhancing production efficiency and agricultural yield. Furthermore, as environmental data continue to be collected in the future, it may enable the identification of varieties better suited to withstand environmental impacts, further increasing productivity [51]. Notably, with the development of interpretability in the future, genome prediction based on deep learning may reveal more genetic-related mechanisms.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/genes16040411/s1. Additional File S1: Supplemental Figures S1–S6 and Supplemental Tables S1–S3. Additional File S2: Supplemental Table S4. Figure S1: Prediction performance of iADEP for traits on four datasets millet827, spruce1722, soy5014, and maize5820; Figure S2: Design of different methods in ablation experiments; Figure S3: The different between transductive learning and inductive learning in 5-fold cross-validation and 100-fold cross-validation; Figure S4: The different output results through CNN; Figure S5: p-value of paired sample t-test on four datasets using five-fold cross-validation; Figure S6: p-value of paired sample t-test on springbarley2463 datasets using five-fold cross-validation; Table S1: Overview of the dataset used in this study; Table S2: Hyperparameters used for iADEP in this study; Table S3: Hyperparameters of other methods excluding iADEP in this study; Table S4: The results of feature selection experiment.

Author Contributions

L.L., C.Y. and Y.-J.H. conceived and designed the study; C.Y. and L.L. analyzed the data and wrote the manuscript; K.L., W.S., Y.J., P.Z., W.Z., Y.-J.H. and Y.H. revised the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Fundamental Research Funds for the Central Universities (No. 2662022JC004) and Hubei Hongshan Laboratory (2021hszd017).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The iADEP scripts are available on github https://github.com/Yechonghang/iADEP (accessed on 24 February 2025). The millet827 dataset is available from https://iagr.genomics.cn/CropGS/#/Datasets?species=Millet (accessed on 24 February 2025) [21]. The spruce1722 dataset and soy5014 dataset are available from https://github.com/FelixHeinrich/GP_with_IFS/tree/main/Datasets (accessed on 24 February 2025) [25]. The maize5820 dataset is available from https://github.com/yingjiexiao/TOP (accessed on 24 February 2025) [27]. The springbarley2463 dataset is available from https://data.mendeley.com/datasets/s3s4ft92wj/1 (accessed on 24 February 2025) [28].

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
GPGenomic prediction
GBLUPGenomic best linear unbiased prediction
MLMachine learning
RFRandom forest
RRRidge regression
SVRSupport vector regression
DLDeep learning
DeepGSDeep convolutional neural network
CNNConvolutional neural networks
DArTDiversity Array Technology
DNNGPDeep neural network genomic prediction
SNPSingle nucleotide polymorphisms
GWASGenome-wide association study
iADEPIntegrated Additive, Dominant and Epistatic Prediction
MLPMultilayer perceptrons
LITLatent Interaction Testing
XGBoosteXtreme Gradient Boosting
PCCPearson correlation coefficient
iGEPintegrated genomic-enviromic prediction
GradCAMGradient-weighted Class Activation Mapping
MAFMinor allele frequency
RILRecombinant lines
GCTAGenetics complex trait analysis
GREMLGenome-based restricted maximum likelihood
MSEMean squared error
TSLLTop second leaf length
TSLWTop Second Leaf Width
MSHMain Stem Height
MSWMain Stem Width
MSPDPanicle Diameter of Main Stem
FNLFringe Neck Length
MSPLPanicle Length of Main Stem
MSPWMain Stem Panicle Weight
PGWPer Plant Grain Weight
MSSNSpikelet Number of Main Stem
SGNGrain Number per Spike
DBGDiameter at Breast Height
DEWood Density
HTHeight
R8Time to R8 Developmental Stage
YLDYield
DTADays to Anther
DTSDays to Silk
DTTDays to Tassel
EHEar Height
ELLEar Leaf Length
ELWEar Leaf Width
PHPlant Height
TBNTassel Branch Number
TLTassel Length
CWCob Weight
EDEar Diameter
ELEar Length
ERNEar Row Number
EWEar Weight
KNPEKernel Number per Ear
KNPRKernel Number per Row
KWPEKernel Weight per Ear
LBTLength of Barren Tip
FSFiltering Speed
EYExtract Yield
WCWort Color
BGBeta Glucan
WVWort Viscosity

References

  1. Meuwissen, T.H.; Hayes, B.J.; Goddard, M. Prediction of Total Genetic Value Using Genome-Wide Dense Marker Maps. Genetics 2001, 157, 1819–1829. [Google Scholar] [PubMed]
  2. Heffner, E.L.; Jannink, J.; Iwata, H.; Souza, E.; Sorrells, M.E. Genomic Selection Accuracy for Grain Quality Traits in Biparental Wheat Populations. Crop Sci. 2011, 51, 2597–2606. [Google Scholar] [CrossRef]
  3. Wu, C.; Zhang, Y.; Ying, Z.; Li, L.; Wang, J.; Yu, H.; Zhang, M.; Feng, X.; Wei, X.; Xu, X. A Transformer-Based Genomic Prediction Method Fused with Knowledge-Guided Module. Brief. Bioinform. 2024, 25, bbad438. [Google Scholar]
  4. Zhang, Z.; Zhang, Q.; Ding, X. Advances in Genomic Selection in Domestic Animals. Chin. Sci. Bull. 2011, 56, 2655–2663. [Google Scholar] [CrossRef]
  5. VanRaden, P.M. Efficient Methods to Compute Genomic Predictions. J. Dairy Sci. 2008, 91, 4414–4423. [Google Scholar]
  6. Endelman, J.B. Ridge Regression and Other Kernels for Genomic Selection with R Package rrBLUP. Plant Genome 2011, 4, 250–255. [Google Scholar] [CrossRef]
  7. Kärkkäinen, H.P.; Sillanpää, M.J. Back to Basics for Bayesian Model Building in Genomic Selection. Genetics 2012, 191, 969–987. [Google Scholar]
  8. Yan, J.; Wang, X. Machine Learning Bridges Omics Sciences and Plant Breeding. Trends Plant Sci. 2023, 28, 199–210. [Google Scholar]
  9. González-Camacho, J.M.; Ornella, L.; Pérez-Rodríguez, P.; Gianola, D.; Dreisigacker, S.; Crossa, J. Applications of Machine Learning Methods to Genomic Selection in Breeding Wheat for Rust Resistance. Plant Genome 2018, 11, 170104. [Google Scholar] [CrossRef]
  10. Alves, A.A.C.; Espigolan, R.; Bresolin, T.; Costa, R.M.; Fernandes Júnior, G.A.; Ventura, R.V.; Carvalheiro, R.; Albuquerque, L.G. Genome-enabled Prediction of Reproductive Traits in Nellore Cattle Using Parametric Models and Machine Learning Methods. Anim. Genet. 2021, 52, 32–46. [Google Scholar] [CrossRef]
  11. Zhao, Z.; Gui, J.; Yao, A.; Le, N.Q.K.; Chua, M.C.H. Improved Prediction Model of Protein and Peptide Toxicity by Integrating Channel Attention into a Convolutional Neural Network and Gated Recurrent Units. ACS Omega 2022, 7, 40569–40577. [Google Scholar] [CrossRef]
  12. Le, N.Q.K. Predicting Emerging Drug Interactions Using GNNs. Nat. Comput. Sci. 2023, 3, 1007–1008. [Google Scholar]
  13. Ma, W.; Qiu, Z.; Song, J.; Li, J.; Cheng, Q.; Zhai, J.; Ma, C. A Deep Convolutional Neural Network Approach for Predicting Phenotypes from Genotypes. Planta 2018, 248, 1307–1318. [Google Scholar] [CrossRef] [PubMed]
  14. Wang, K.; Abid, M.A.; Rasheed, A.; Crossa, J.; Hearne, S.; Li, H. DNNGP, a Deep Neural Network-Based Method for Genomic Prediction Using Multi-Omics Data in Plants. Mol. Plant 2023, 16, 279–293. [Google Scholar] [CrossRef]
  15. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention Is All You Need. Adv. Neural Inf. Process. Syst. 2017, 30. [Google Scholar] [CrossRef]
  16. Wu, H.; Xu, J.; Wang, J.; Long, M. Autoformer: Decomposition Transformers with Auto-Correlation for Long-Term Series Forecasting. Adv. Neural Inf. Process. Syst. 2021, 34, 22419–22430. [Google Scholar]
  17. Vitezica, Z.G.; Legarra, A.; Toro, M.A.; Varona, L. Orthogonal Estimates of Variances for Additive, Dominance, and Epistatic Effects in Populations. Genetics 2017, 206, 1297–1307. [Google Scholar] [PubMed]
  18. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  19. Bass, A.J.; Bian, S.; Wingo, A.P.; Wingo, T.S.; Cutler, D.J.; Epstein, M.P. Identifying Latent Genetic Interactions in Genome-Wide Association Studies Using Multiple Traits. Genome Med. 2024, 16, 62. [Google Scholar] [CrossRef]
  20. Wang, Y.; Wang, X.; Sun, S.; Jin, C.; Su, J.; Wei, J.; Luo, X.; Wen, J.; Wei, T.; Sahu, S.K. GWAS, MWAS and mGWAS Provide Insights into Precision Agriculture Based on Genotype-Dependent Microbial Effects in Foxtail Millet. Nat. Commun. 2022, 13, 5913. [Google Scholar]
  21. Chen, J.; Tan, C.; Zhu, M.; Zhang, C.; Wang, Z.; Ni, X.; Liu, Y.; Wei, T.; Wei, X.; Fang, X. CropGS-Hub: A Comprehensive Database of Genotype and Phenotype Resources for Genomic Prediction in Major Crops. Nucleic Acids Res. 2023, 52, D1519–D1529. [Google Scholar]
  22. Chang, C.C.; Chow, C.C.; Tellier, L.C.; Vattikuti, S.; Purcell, S.M.; Lee, J.J. Second-Generation PLINK: Rising to the Challenge of Larger and Richer Datasets. Gigascience 2015, 4, s13742-015. [Google Scholar]
  23. Beaulieu, J.; Doerksen, T.K.; MacKay, J.; Rainville, A.; Bousquet, J. Genomic Selection Accuracies within and between Environments and Small Breeding Groups in White Spruce. BMC Genom. 2014, 15, 1048. [Google Scholar] [CrossRef]
  24. Xavier, A.; Muir, W.M.; Rainey, K.M. Assessing Predictive Properties of Genome-Wide Selection in Soybeans. G3 Genes Genomes Genet. 2016, 6, 2611–2616. [Google Scholar]
  25. Heinrich, F.; Lange, T.M.; Kircher, M.; Ramzan, F.; Schmitt, A.O.; Gültas, M. Exploring the Potential of Incremental Feature Selection to Improve Genomic Prediction Accuracy. Genet. Sel. Evol. 2023, 55, 78. [Google Scholar] [CrossRef]
  26. Liu, H.-J.; Wang, X.; Xiao, Y.; Luo, J.; Qiao, F.; Yang, W.; Zhang, R.; Meng, Y.; Sun, J.; Yan, S.; et al. CUBIC: An Atlas of Genetic Architecture Promises Directed Maize Improvement. Genome Biol. 2020, 21, 20. [Google Scholar] [CrossRef]
  27. Yang, W.; Guo, T.; Luo, J.; Zhang, R.; Zhao, J.; Warburton, M.L.; Xiao, Y.; Yan, J. Target-Oriented Prioritization: Targeted Selection Strategy by Integrating Organismal and Molecular Traits through Predictive Analytics in Breeding. Genome Biol. 2022, 23, 80. [Google Scholar] [CrossRef]
  28. Guo, X. Data for Spring Barley from Nordic Seed A/S. Mendeley Data 2020. [Google Scholar] [CrossRef]
  29. Guo, X.; Sarup, P.; Jahoor, A.; Jensen, J.; Christensen, O.F. Metabolomic-Genomic Prediction Can Improve Prediction Accuracy of Breeding Values for Malting Quality Traits in Barley. Genet Sel. Evol. 2023, 55, 61. [Google Scholar] [CrossRef]
  30. Yang, J.; Benyamin, B.; McEvoy, B.P.; Gordon, S.; Henders, A.K.; Nyholt, D.R.; Madden, P.A.; Heath, A.C.; Martin, N.G.; Montgomery, G.W. Common SNPs Explain a Large Proportion of the Heritability for Human Height. Nat. Genet. 2010, 42, 565–569. [Google Scholar]
  31. Yang, J.; Lee, S.H.; Goddard, M.E.; Visscher, P.M. GCTA: A Tool for Genome-Wide Complex Trait Analysis. Am. J. Hum. Genet. 2011, 88, 76–82. [Google Scholar]
  32. Namkoong, G. Introduction to Quantitative Genetics in Forestry; Department of Agriculture, Forest Service: Washinton, DC, USA, 1979. [Google Scholar]
  33. Kang, H.M.; Sul, J.H.; Service, S.K.; Zaitlen, N.A.; Kong, S.; Freimer, N.B.; Sabatti, C.; Eskin, E. Variance Component Model to Account for Sample Structure in Genome-Wide Association Studies. Nat. Genet. 2010, 42, 348–354. [Google Scholar] [CrossRef]
  34. Xu, B.; Wang, N.; Chen, T.; Li, M. Empirical Evaluation of Rectified Activations in Convolutional Network 2015. arXiv 2015, arXiv:1505.00853. [Google Scholar]
  35. Hoerl, A.E.; Kennard, R.W. Ridge Regression: Biased Estimation for Nonorthogonal Problems. Technometrics 1970, 12, 55–67. [Google Scholar] [CrossRef]
  36. Amit, Y.; Geman, D. Shape Quantization and Recognition with Randomized Trees. Neural Comput. 1997, 9, 1545–1588. [Google Scholar] [CrossRef]
  37. Maenhout, S.; De Baets, B.; Haesaert, G.; Van Bockstaele, E. Support Vector Machine Regression for the Prediction of Maize Hybrid Performance. Theor. Appl. Genet. 2007, 115, 1003–1013. [Google Scholar] [CrossRef]
  38. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V. Scikit-Learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  39. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, ACM, San Francisco, CA, USA, 13 August 2016; pp. 785–794. [Google Scholar]
  40. Sandhu, K.S.; Lozada, D.N.; Zhang, Z.; Pumphrey, M.O.; Carter, A.H. Deep Learning for Predicting Complex Traits in Spring Wheat Breeding Program. Front. Plant Sci. 2021, 11, 613325. [Google Scholar]
  41. Yang, J.; Zeng, J.; Goddard, M.E.; Wray, N.R.; Visscher, P.M. Concepts, Estimation and Interpretation of SNP-Based Heritability. Nat. Genet. 2017, 49, 1304–1310. [Google Scholar]
  42. Zhu, H.; Zhou, X. Statistical Methods for SNP Heritability Estimation and Partition: A Review. Comput. Struct. Biotechnol. J. 2020, 18, 1557–1568. [Google Scholar] [CrossRef]
  43. Lee, H.-J.; Lee, J.H.; Gondro, C.; Koh, Y.J.; Lee, S.H. deepGBLUP: Joint Deep Learning Networks and GBLUP Framework for Accurate Genomic Prediction of Complex Traits in Korean Native Cattle. Genet. Sel. Evol. 2023, 55, 56. [Google Scholar] [CrossRef]
  44. Zhang, F.; Kang, J.; Long, R.; Li, M.; Sun, Y.; He, F.; Jiang, X.; Yang, C.; Yang, X.; Kong, J. Application of Machine Learning to Explore the Genomic Prediction Accuracy of Fall Dormancy in Autotetraploid Alfalfa. Hortic. Res. 2023, 10, uhac225. [Google Scholar] [PubMed]
  45. Lachaud, G.; Conde-Cespedes, P.; Trocan, M. Comparison between Inductive and Transductive Learning in a Real Citation Network Using Graph Neural Networks. In Proceedings of the 2022 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining (ASONAM), Istanbul, Turkey, 10–13 November 2022; IEEE: New York, NY, USA, 2022; pp. 534–540. [Google Scholar]
  46. Zingaretti, L.M.; Gezan, S.A.; Ferrão, L.F.V.; Osorio, L.F.; Monfort, A.; Muñoz, P.R.; Whitaker, V.M.; Pérez-Enciso, M. Exploring Deep Learning for Complex Trait Genomic Prediction in Polyploid Outcrossing Species. Front. Plant Sci. 2020, 11, 25. [Google Scholar] [CrossRef] [PubMed]
  47. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S. An Image Is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  48. Xu, Y.; Zhang, X.; Li, H.; Zheng, H.; Zhang, J.; Olsen, M.S.; Varshney, R.K.; Prasanna, B.M.; Qian, Q. Smart Breeding Driven by Big Data, Artificial Intelligence, and Integrated Genomic-Enviromic Prediction. Mol. Plant 2022, 15, 1664–1695. [Google Scholar] [PubMed]
  49. Castro, J.; Gómez, D.; Tejada, J. Polynomial Calculation of the Shapley Value Based on Sampling. Comput. Oper. Res. 2009, 36, 1726–1730. [Google Scholar]
  50. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-Cam: Visual Explanations from Deep Networks via Gradient-Based Localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
  51. Heslot, N.; Jannink, J.; Sorrells, M.E. Perspectives for Genomic Selection Applications and Research in Plants. Crop Sci. 2015, 55, 1–12. [Google Scholar] [CrossRef]
Figure 1. Overview of iADEP framework. (A) Illustration of SNP embedding using trait MSW as an example. The upper left panel shows the Manhattan plot of GWAS; the lower left panel shows the Manhattan plot of LIT. The right panel gives the result of SNP embedding. (B) The construction of iADEP, where dashed line-boxes indicate sections which are optional for other omics data. MSW, main stem width.
Figure 1. Overview of iADEP framework. (A) Illustration of SNP embedding using trait MSW as an example. The upper left panel shows the Manhattan plot of GWAS; the lower left panel shows the Manhattan plot of LIT. The right panel gives the result of SNP embedding. (B) The construction of iADEP, where dashed line-boxes indicate sections which are optional for other omics data. MSW, main stem width.
Genes 16 00411 g001
Figure 2. Prediction performance comparison of seven methods on three datasets: millet827, spruce1722 and soy5014. (A,B) Prediction performance on millet827 growth traits (A) and yield traits (B). (C) Prediction performance on spruce1722 dataset. (D) Prediction performance on soy5014 dataset. TSLL, top second leaf length; TSLW, top second leaf width; MSH, main stem height; MSW, main stem width; MSPD, panicle diameter of the main stem; FNL, fringe neck length; MSPL, panicle length of the main stem; MSPW, main stem panicle weight; PGW, per plant grain weight; MSSN, spikelet number of the main stem; SGN, grain number per spike; DBH, diameter at breast height; DE, wood density; HT, height; R8, time to R8 developmental stage; YLD, yield.
Figure 2. Prediction performance comparison of seven methods on three datasets: millet827, spruce1722 and soy5014. (A,B) Prediction performance on millet827 growth traits (A) and yield traits (B). (C) Prediction performance on spruce1722 dataset. (D) Prediction performance on soy5014 dataset. TSLL, top second leaf length; TSLW, top second leaf width; MSH, main stem height; MSW, main stem width; MSPD, panicle diameter of the main stem; FNL, fringe neck length; MSPL, panicle length of the main stem; MSPW, main stem panicle weight; PGW, per plant grain weight; MSSN, spikelet number of the main stem; SGN, grain number per spike; DBH, diameter at breast height; DE, wood density; HT, height; R8, time to R8 developmental stage; YLD, yield.
Genes 16 00411 g002
Figure 3. Prediction performance of seven methods on maize5820 dataset. (AC) Prediction performance on maize5820 flowering traits (A), plant architecture traits (B), and yield traits (C). DTA, days to anther; DTS, days to silk; DTT, days to tassel; EH, ear height; ELL, ear leaf length; ELW, ear leaf width; PH, plant height; TBN, tassel branch number; TL, tassel length; CW, cob weight; ED, ear diameter; EL, ear length; ERN, ear row number; EW, ear weight; KNPE, kernel number per ear; KNPR, kernel number per row; KWPE, kernel weight per ear; LBT, length of barren tip.
Figure 3. Prediction performance of seven methods on maize5820 dataset. (AC) Prediction performance on maize5820 flowering traits (A), plant architecture traits (B), and yield traits (C). DTA, days to anther; DTS, days to silk; DTT, days to tassel; EH, ear height; ELL, ear leaf length; ELW, ear leaf width; PH, plant height; TBN, tassel branch number; TL, tassel length; CW, cob weight; ED, ear diameter; EL, ear length; ERN, ear row number; EW, ear weight; KNPE, kernel number per ear; KNPR, kernel number per row; KWPE, kernel weight per ear; LBT, length of barren tip.
Genes 16 00411 g003
Figure 4. Prediction performance comparison of six methods on springbarley2463 dataset when incorporating other omics data. FS, filtering speed; EY, extract yield; WC, wort color; BG, beta glucan; WV, wort viscosity.
Figure 4. Prediction performance comparison of six methods on springbarley2463 dataset when incorporating other omics data. FS, filtering speed; EY, extract yield; WC, wort color; BG, beta glucan; WV, wort viscosity.
Genes 16 00411 g004
Figure 5. The changes of prediction performance with feature selection using thresholds of top-20%, top-10%, and top-5% SNPs for four datasets. (AD) The prediction performance of feature selection under top-20%, top-10%, and top-5% thresholds for SNP selection on four datasets: millet827 (A), spruce1722 (B), soy5014 (C), and maize5820 (D). TSLL, top second leaf length; TSLW, top second leaf width; MSH, main stem height; MSW, main stem width; MSPD, panicle diameter of the main stem; FNL, fringe neck length; MSPL, panicle length of the main stem; MSPW, main stem panicle weight; PGW, per plant grain weight; MSSN, spikelet number of the main stem; SGN, grain number per spike; DBH, diameter at breast height; DE, wood density; HT, height; R8, time to R8 developmental stage; YLD, yield; DTA, days to anther; DTS, days to silk; DTT, days to tassel; EH, ear height; ELL, ear leaf length; ELW, ear leaf width; PH, plant height; TBN, tassel branch number; TL, tassel length; CW, cob weight; ED, ear diameter; EL, ear length; ERN, ear row number; EW, ear weight; KNPE, kernel number per ear; KNPR, kernel number per row; KWPE, kernel weight per ear; LBT, length of barren tip.
Figure 5. The changes of prediction performance with feature selection using thresholds of top-20%, top-10%, and top-5% SNPs for four datasets. (AD) The prediction performance of feature selection under top-20%, top-10%, and top-5% thresholds for SNP selection on four datasets: millet827 (A), spruce1722 (B), soy5014 (C), and maize5820 (D). TSLL, top second leaf length; TSLW, top second leaf width; MSH, main stem height; MSW, main stem width; MSPD, panicle diameter of the main stem; FNL, fringe neck length; MSPL, panicle length of the main stem; MSPW, main stem panicle weight; PGW, per plant grain weight; MSSN, spikelet number of the main stem; SGN, grain number per spike; DBH, diameter at breast height; DE, wood density; HT, height; R8, time to R8 developmental stage; YLD, yield; DTA, days to anther; DTS, days to silk; DTT, days to tassel; EH, ear height; ELL, ear leaf length; ELW, ear leaf width; PH, plant height; TBN, tassel branch number; TL, tassel length; CW, cob weight; ED, ear diameter; EL, ear length; ERN, ear row number; EW, ear weight; KNPE, kernel number per ear; KNPR, kernel number per row; KWPE, kernel weight per ear; LBT, length of barren tip.
Genes 16 00411 g005
Figure 6. Comparison of prediction performance between 5-fold and 100-fold cross-validation for GBLUP and iADEP. For each trait, performance of GBLUP is represented on the left side of the dashed line, while performance of iADEP is represented on the right side. TSLL, top second leaf length; TSLW, top second leaf width; MSH, main stem height; MSW, main stem width; MSPD, panicle diameter of the main stem; FNL, fringe neck length; MSPL, panicle length of the main stem; MSPW, main stem panicle weight; PGW, per plant grain weight; MSSN, spikelet number of the main stem; SGN, grain number per spike.
Figure 6. Comparison of prediction performance between 5-fold and 100-fold cross-validation for GBLUP and iADEP. For each trait, performance of GBLUP is represented on the left side of the dashed line, while performance of iADEP is represented on the right side. TSLL, top second leaf length; TSLW, top second leaf width; MSH, main stem height; MSW, main stem width; MSPD, panicle diameter of the main stem; FNL, fringe neck length; MSPL, panicle length of the main stem; MSPW, main stem panicle weight; PGW, per plant grain weight; MSSN, spikelet number of the main stem; SGN, grain number per spike.
Genes 16 00411 g006
Table 1. Results of ablation experiments on four datasets: millet827, spruce1722, soy5014, and maize5820.
Table 1. Results of ablation experiments on four datasets: millet827, spruce1722, soy5014, and maize5820.
DatasetTraitSNP EmbeddingOne-Hot EncodingNon-Encoding
millet827TSLL0.559 ± 0.0400.535 ± 0.0330.533 ± 0.024
TSLW0.478 ± 0.0530.438 ± 0.0690.457 ± 0.048
MSH0.361 ± 0.0800.348 ± 0.0790.354 ± 0.089
MSW0.666 ± 0.0310.639 ± 0.0380.634 ± 0.042
MSPD0.669 ± 0.0360.657 ± 0.0300.649 ± 0.040
FNL0.545 ± 0.1040.537 ± 0.1120.544 ± 0.111
MSPL0.592 ± 0.0370.582 ± 0.0500.587 ± 0.039
MSPW0.326 ± 0.0590.338 ± 0.0620.291 ± 0.085
PGW0.321 ± 0.0590.299 ± 0.0610.281 ± 0.097
MSSN0.381 ± 0.0410.351 ± 0.0400.341 ± 0.035
SGN0.394 ± 0.0570.393 ± 0.0600.393 ± 0.059
spruce1722DBH0.286 ± 0.0560.268 ± 0.0780.225 ± 0.060
DE0.433 ± 0.0240.424 ± 0.0280.417 ± 0.022
HT0.347 ± 0.0640.334 ± 0.0670.308 ± 0.041
soy5014HT0.476 ± 0.0190.458 ± 0.0260.453 ± 0.030
R80.481 ± 0.0100.437 ± 0.0180.456 ± 0.024
YLD0.662 ± 0.0070.659 ± 0.0100.659 ± 0.008
maize5820DTA0.948 ± 0.0010.940 ± 0.0030.942 ± 0.003
DTS0.946 ± 0.0040.938 ± 0.0050.941 ± 0.004
DTT0.948 ± 0.0020.939 ± 0.0030.944 ± 0.003
EH0.941 ± 0.0020.921 ± 0.0090.929 ± 0.001
ELL0.914 ± 0.0030.884 ± 0.0050.893 ± 0.004
ELW0.887 ± 0.0070.883 ± 0.0080.885 ± 0.008
PH0.941 ± 0.0030.932 ± 0.0040.931 ± 0.005
TBN0.927 ± 0.0040.905 ± 0.0040.913 ± 0.006
TL0.924 ± 0.0020.902 ± 0.0030.910 ± 0.004
CW0.852 ± 0.0070.815 ± 0.0070.826 ± 0.013
ED0.840 ± 0.0130.552 ± 0.2380.796 ± 0.094
EL0.831 ± 0.0050.803 ± 0.0080.807 ± 0.010
ERN0.914 ± 0.0030.908 ± 0.0040.908 ± 0.004
EW0.805 ± 0.0140.776 ± 0.0110.782 ± 0.014
KNPE0.847 ± 0.0110.841 ± 0.0100.842 ± 0.012
KNPR0.705 ± 0.0160.686 ± 0.0180.691 ± 0.019
KWPE0.778 ± 0.0140.747 ± 0.0220.756 ± 0.019
LBT0.683 ± 0.0170.621 ± 0.1060.678 ± 0.020
Each value in the cells are means and standard errors of the prediction performance for 5-fold cross-validation. The best results are highlight in bold.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ye, C.; Li, K.; Sun, W.; Jiang, Y.; Zhang, W.; Zhang, P.; Hu, Y.-J.; Han, Y.; Li, L. Biological Prior Knowledge-Embedded Deep Neural Network for Plant Genomic Prediction. Genes 2025, 16, 411. https://doi.org/10.3390/genes16040411

AMA Style

Ye C, Li K, Sun W, Jiang Y, Zhang W, Zhang P, Hu Y-J, Han Y, Li L. Biological Prior Knowledge-Embedded Deep Neural Network for Plant Genomic Prediction. Genes. 2025; 16(4):411. https://doi.org/10.3390/genes16040411

Chicago/Turabian Style

Ye, Chonghang, Kai Li, Weicheng Sun, Yiwei Jiang, Weihan Zhang, Ping Zhang, Yi-Juan Hu, Yuepeng Han, and Li Li. 2025. "Biological Prior Knowledge-Embedded Deep Neural Network for Plant Genomic Prediction" Genes 16, no. 4: 411. https://doi.org/10.3390/genes16040411

APA Style

Ye, C., Li, K., Sun, W., Jiang, Y., Zhang, W., Zhang, P., Hu, Y.-J., Han, Y., & Li, L. (2025). Biological Prior Knowledge-Embedded Deep Neural Network for Plant Genomic Prediction. Genes, 16(4), 411. https://doi.org/10.3390/genes16040411

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop