Next Article in Journal
Soybean Performance as Affected by Lime and Gypsum Incorporation Through Tillage Versus Surface Application in Pasture-to-Cropland Conversion Areas in Southeast Brazil
Next Article in Special Issue
A Lightweight Plant Disease Detection Model for Long-Tailed Agricultural Scenarios
Previous Article in Journal
Differential Responses and Temporal Lags of Heterotrophic and Autotrophic Respiration to Plant Activity in a Forest Ecosystem
Previous Article in Special Issue
ConvDeiT-Tiny: Adding Local Inductive Bias to DeiT-Ti for Enhanced Maize Leaf Disease Classification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Tree Height Prediction Using a Double Hidden-Layer Neural Network and a Mixed-Effects Model

1
Wenzhou Key Laboratory of AI Agents for Agriculture, Wenzhou Vocational College of Science and Technology, Wenzhou 325006, China
2
Institute of Resources Information Techniques, Chinese Academy of Forestry, Beijing 100091, China
3
Institute of Forestry Inventory and Planning of Jilin Province, Changchun 130022, China
4
Agricultural Information Institute, Chinese Academy of Agricultural Sciences, Beijing 100081, China
5
School of Computer Science and Engineering, Intelligent Collaborative Innovation Studio, Guangzhou Institute of Science and Technology, Guangzhou 510540, China
*
Authors to whom correspondence should be addressed.
Plants 2026, 15(8), 1176; https://doi.org/10.3390/plants15081176
Submission received: 9 February 2026 / Revised: 29 March 2026 / Accepted: 7 April 2026 / Published: 10 April 2026
(This article belongs to the Special Issue AI-Driven Machine Vision Technologies in Plant Science)

Abstract

The double hidden-layer neural network has increasingly been applied in tree height modeling due to its superior performance. To improve the precision of tree height estimation, this study compared the performance of a double hidden-layer neural network with that of a nonlinear mixed-effects model, aiming to provide a new method for tree height prediction. Taking the Larix olgensis forest plantation in Jilin Province as the research object, a double hidden-layer back propagation (BP) neural network was established for tree height prediction by adopting trial and error, k-fold cross-validation, and near-domain optimization strategies. In constructing the nonlinear mixed-effects model, the overall and local differences in forest growth data, as well as the autocorrelation among the various levels of data, were considered. Accordingly, after determining the base model, random effects were introduced, the correlation variance–covariance matrix was calculated, and random parameters were estimated to compare the predictive performance of the two aforementioned models. For the mixed-effects model, the coefficient of determination R2 was 0.8590, the root mean square error (RMSE) was 1.6230, and the mean absolute error (MAE) was 2.2658. For the double hidden-layer BP neural network, the R2 reached 0.9068 (an increase of 5.56%), the RMSE was 1.3197 (a decrease of 18.69%), and the MAE was 1.2736 (a decrease of 43.79%). The results demonstrate that the double hidden-layer BP neural network is superior to the nonlinear mixed-effects model for tree height prediction. Therefore, the results provide a more accurate method for tree height prediction.

1. Introduction

Tree height is one of the important tree-measuring factors in forest surveys, and it is also an important basis for assessing growth status and site quality as well as dividing the forest layers [1,2]. Since it is difficult to measure directly, tree height is calculated with a model based on the relationship between it and DBH (diameter at breast height). The core of this method is formulating various special equations to obtain the height by solving those equations [3,4,5,6]. Because the stand mean height is affected not only by the DBH, but also by other factors, all factors that affect the tree height should be considered.
In general, the factors affecting tree height include DBH, site factor, forest competition factor, and climate factor [7]. A neural network has numerous characteristics such as nonlinear mapping, adaptability, generalization ability, and fault tolerance. In recent years, it has been gradually applied to forest growth prediction and achieved various results. Scrinzi et al. [8] used neural network models to update the tree DBH distributions for managed alpine stands and predict the number of standing trees. Pirotti [9] analyzed the return signals from canopy covers by using an artificial neural network (ANN) to determine if there is an improvement in detecting tree height and position compared to a more classic local-maximum-filter approach. Diamantopoulou [10] used an ANN model to estimate the inside-bark and outside-bark total volume of dominant pine trees (Pinus brutia) in reforestations. A neural network model is a hot method at present, and some scholars [11,12,13,14,15] have adopted a neural network model to predict forest growth. In the field of forest growth research, the application of neural networks is limited in the single hidden-layer, and the research involving more hidden layers is rarely reported. However, it is considered that double hidden layers are better than a single layer for more complex problems [16]. Therefore, the double hidden-layer neural network is selected to establish a tree height model.
Fixed effects models [17] assume that the observed data are independent and identically distributed. However, in many practical scenarios, such as forest growth modeling, data exhibit hierarchical or spatiotemporal correlation, which violates these assumptions. The mixed-effects models, introduced to address these limitations by incorporating random effects into fixed effects, provide more accurate and robust predictions. It has shown improved performance in handling complex data structures with spatiotemporal correlation, can reflect the overall average change trend, and provide various information reflecting inter-individual differences, such as data variance and covariance. In addition, this method can represent the correlation error by specifying the different covariance structures so as to improve the prediction accuracy and explain the source of random error. Compared with other models, it has advantages in the processing of continuous observed data having spatiotemporal sequence correlation [18]. The tree height data were obtained through several consecutive observations, and a certain autocorrelation exists between the levels of these repeated measurement data. The random error of them consists of the individual random effects and the multiple repeated measurement effects within the individual, for which the mixed model is suitable for dealing. Thus, the mixed-effects model is chosen to predict the tree height. The mixed-effects model can be divided into linear and nonlinear models. Because the tree height is affected by many factors and the interaction effect of them is extremely complicated, it is appropriate to adopt the nonlinear mixed-effects model. Therefore, some nonlinear mixed-effects models for estimation of tree height have been developed in recent years. Razali [19] used mixed-effects models for predicting the early height growth of forest trees planted in Sarawak, Malaysia. Kalbi [18] evaluated two linear and eighteen nonlinear height–diameter equations and took the best model as the base mixed-effects model to predict the tree height of Oriental beech in the Hyrcanian Forest in Iran. Lam [20] utilized the taxonomic hierarchy of genus and species as the random effects of the nonlinear mixed-effects model to develop species-specific height–diameter (H–D) models for 842 species representing 295 genera, which is useful for many rare tree species with very few observations.
Neural networks have been increasingly applied to tree height prediction due to their ability to capture complex nonlinear relationships. However, the performance of neural networks can vary significantly depending on their architecture, such as the number of hidden layers and the choice of activation functions. Thus, this study focuses on comparing the performance of a double hidden-layer neural network with that of a nonlinear mixed-effects model, providing insights into the impact of hidden layer architecture on prediction accuracy. The mixed-effects model introduces random effects into the fixed model, thereby enabling description of both general and individual characteristics, which has high prediction accuracy. However, usage of the complex structure of the mixed-effects model requires strong professional knowledge. In addition, the amount of data required by this method is very large, and the data must have obvious differences. Therefore, the above two methods have advantages and disadvantages. In practical applications, both the neural network and mixed-effects model should be established with the given data, and then they can be compared and analyzed according to the evaluation criteria.
In order to compare the application effect in tree height prediction, the tree height models based on the double hidden-layer back propagation (BP) neural network and the nonlinear mixed-effects model were respectively established with the data of the Larix olgensis plantation in Jilin Province, and the fitting accuracy between them was compared. The research aims to provide a reference for tree height model optimization.

2. Materials and Methods

2.1. Data Sources

In this study, data were obtained from the forest inventory in Jilin Province in 2009, covering a total of 7662 trees across 96 sample plots. To ensure the research subjects were valid and stable forest stands, and to exclude interference from non-target plots such as newly afforested but not yet matured areas and sparse forests, this study excluded plots with a forest stand density below 300 trees per hectare. Additionally, abnormal values in measurements such as tree height and age were removed to eliminate the impact of measurement errors and extreme individuals on model fitting, thereby ensuring the homogeneity and reliability of the data. The final dataset included a total of 7001 trees across 86 sample plots. Although the number of sample plots is not large, the number of trees is over 80 times that of the sample plots. Each record of data represents all the information within the corresponding sample plot. Behind this lies a huge amount of data, which is representative and typical, and the data quality is relatively high. In our experiment, 64 records were taken as the training data and others as the testing data. In this study, the DBH, stand basal area, and slope position (the position of the slope where the plot was surveyed, including the ridge, upper slope, middle slope, lower slope, valley, flatland) were taken as the independent variables and the tree height as the target variable, and a double hidden-layer BP neural network and a nonlinear mixed-effects model were established for tree height prediction using the MatLab 2016b software (MathWorks, Natick, MA, USA).
In this study, DBH and stand basal area were both used as input factors. DBH is the most crucial tree measurement indicator for individual trees, directly reflecting the growth status of the tree. It is a fundamental variable for constructing growth models and estimating biomass. The stand basal area of the forest represents the overall spatial occupancy level and competition intensity of the forest, reflecting the characteristics of the population structure. The simultaneous use of forest stand basal area and DBH as input factors for the model has theoretical rationality and practical feasibility, as supported by relevant research [21]. The two correspond to individual scale and forest scale respectively, with complementary information dimensions and no complete linear redundancy. DBH reflects the growth status of individual trees, while stand basal area reflects the overall and density competition of the forest stand. Therefore, it is reasonable to input them as independent variables into the model. Introducing models that can simultaneously capture individual differences and group competition effects in forest growth, site quality evaluation, carbon storage estimation, etc., significantly improves fitting accuracy and explanatory power. In forest growth simulation, yield estimation, and intelligent monitoring models, the joint input of DBH and stand basal area is a scientifically feasible optimal solution.
The dataset covers the sample plots distribution across six slope position categories (ridge: 12.79%; upper slope: 15.12%; middle slope: 22.09%; lower slope: 18.60%; valley: 16.28%; flatland: 15.12%). Descriptive statistics (e.g., mean, standard deviation) were calculated for all categories. The results show that the sample sizes across categories are relatively balanced, and there are certain differences in the distribution of tree height, DBH, and stand basal area. This supports the rationality of incorporating slope position as a random effect into the model.
Introducing slope position as a random effect into tree height growth models can significantly improve model fitting and prediction robustness. Slope position influences the redistribution of water and nutrients on the slope surface, as well as micro-environmental conditions such as light and temperature, thereby creating differentiated habitats like upper, middle, and lower slopes. This directly affects tree root distribution, nutrient uptake, and photosynthetic efficiency, leading to obvious spatial variations in tree height growth. By setting slope position as a random effect, it not only quantitatively captures topographic heterogeneity unexplained by fixed factors, reducing model residuals and heteroscedasticity, but also avoids loss of sample degrees of freedom due to the introduction of excessive categorical variables, making it particularly suitable for small-sample modeling. From an ecological perspective, the random effect of slope position reflects the regulatory role of micro-habitat heterogeneity driven by topography on tree growth and demonstrates the potential constraints of site conditions on tree height growth, enhancing the model’s alignment with the actual physiological and ecological processes of forest growth.

2.2. BP Neural Network

A BP neural network is a machine learning method utilizing a multi-layer feed-forward network trained by a BP error algorithm, usually composed of input, hidden, and output layers [22]. The structural parameters of this network include the number of hidden layers, the number of nodes in each hidden layer, and the transfer function between different layers. According to Kolmogorov’s theory [16], a single hidden-layer can satisfy the requirement of most general problems; however, for a complex system, two or more hidden layers may achieve better results [23]. Thus, the target of our study was optimizing the number of hidden-layer neurons in the network.

2.2.1. Number of Hidden-Layer Nodes and Transfer Function

The calculation of the number of nodes in each hidden layer is an important issue in the field of neural networks. This problem is uncertain and the optimal number depends on various factors such as data size, activation function, network structure, etc. [24]. Since the 1990s until now, researchers have conducted a series of studies on this topic [25,26,27,28]. Among the many available solutions, Formula (1) [25,29] is chosen to calculate the number of nodes in the hidden layer.
C = g + o + r
where C is the number of hidden-layer nodes; g , the number of input-layer nodes; o , the number of output-layer nodes; and r , any integer between 1 and 10.
For the transfer function between the input and the hidden layers, or between the adjacent hidden layers includes tansig and logsig, the corresponding expressions are shown as Equations (2) and (3). The transfer function between the hidden and the output layers is purelin, as expressed in Equation (4):
tan s i g ( x ) = 2 1 + e 2 x 1
log s i g ( x ) = 1 1 + e x 1
p u r e l i n ( x ) = a x + b

2.2.2. Normalization and Denormalization of Input/Output Data

The map-min-max function was used to normalize the input and output data, mapping to [−1, 1]. The normalization procedure is expressed as Equation (5):
y = y max y min × x x min x max x min + y min
where y represents the normalized value, and the maximum value y max = 1 and the minimum value y min = 1 ; x is the input/output value to be normalized, the maximum and minimum values of which are x max and x min respectively.
Denormalization can map the normalized value from [−1, 1] to the actual value.

2.2.3. Model Training

The target of this phase is to minimize the loss function, which is used to measure the difference between the predicted and the observed value [30]. The commonly used loss functions include absolute value loss function, squared error loss function, Huber loss function, Logarithmic hyperbolic cosine (log-cosh) loss function, and Quantile loss function [31]. In terms of usability and efficiency, the squared error loss function was selected.
Newton’s method is an iterative solution method to minimize the loss function, which is based on the second-order Taylor series [32,33,34]. Compared with other optimization methods suitable for neural network training, such as the gradient descent method [35], conjugate gradient method [36], and heuristic method (simulated annealing algorithm, genetic algorithm, etc.) [37], Newton’s method has the advantages of fast convergence and good fitting, the basic formula for which is as follows:
X k + 1 = X k F 1 k · J k
where X k + 1 and X k are the parameter values in the (k+1)-th and the k-th iterations respectively, and F k and J k are the second partial derivative (Hessian matrix) and the first partial derivative (Jacobian matrix) of X k respectively.
The Hessian matrix of Newton’s method is composed of the second partial derivatives, which results in the large calculation load. In some cases, the Newton’s method may be failed if the Hessian matrix could not be positive definite.
In this case, Quasi-Newton method is proposed, which is based on the following “Quasi-Newton equation” [38]:
l k = F k × z k
F - 1 k × l k = z k
The above two formulas are equivalent to each other, where l k = J k + 1 J k , z k = X k + 1 X k .
On the basis of Formulas (7) and (8), the approximate Hessian matrix with positive definite symmetry is constructed with iteration but without the calculation of second partial derivative, so as to reduce the significant computation of the Hessian matrix in Newton’s method.
The commonly used Quasi-Newtonian methods include DFP [39], BFGS [40], and L-BFGS [41]. DFP and BFGS can be used to calculate F - 1 k and F k iteratively according to Formulas (8) and (9), respectively. L-BFGS reduces the space complexity of BFGS from O N 2 to O t N , where N is the order of matrix F k , and t is the parameter far less than N . Through analysis and testing carefully, the L-BFGS method was selected to minimize the loss function.
In addition to the loss function and optimization method, other parameters were set as follows: learning rate, 0.01; maximum iterations, 1000; target precision, 0.001; maximum number of verification failures, 20; and minimum performance gradient, 0.000001.

2.2.4. Model Selection

As described in Section 2.2, a series of BP neural networks was obtained by setting the different structural parameters, such as the number of hidden layers and the transfer function. Then, the most suitable model was selected with the K-fold cross-validation [42].
Suppose the sample set S contains d samples, and the candidate models are M 1 , M 2 , …, M h , then the process of K-fold cross-validation is as follows:
Step 1: The sample set S is randomly divided into K disjoint subsets, and the number of samples in each subset is d / K . These subsets are denoted as S 1 , S 2 , …, S K .
Step 2: The following operation is carried out for every model M e , e = 1 , 2 , , h :
For n = 1 to K
Take the dataset D S n = S 1 S n 1 S n + 1 S K as the training set.
Train the model M e with the training set D S n , so as to get the corresponding fitting function H e , n .
Take the dataset S n as the validation set, then calculate the generalization error ε S n H e , n of H e , n .
Generate the mean generalization error of model M e by calculating the average of the ε S n H e , n , n = 1 , 2 , , K .
Step 3: Choose the model M o p t with the minimal mean generalization error as the optimal model.
Generally speaking, the mean square error (MSE) is used to represent the generalization error ε S n H e , n , as shown in Equation (9).
M S E = 1 w l = 1 w ( Y l Y ^ l ) 2
where w is the number of samples, and Y and Y ^ l are the observed and the predicted values of the l-th samples respectively.
The “trial and error method” [43] may not traverse all the models corresponding to all the parameter combinations because its search step-size is often larger than 1. Therefore, it is necessary to carry out the second optimization based on the first optimization. In practice, in addition to the generalization error, the number of iterations, running time, and other factors are sometimes taken as the evaluation indicators.

2.2.5. Neural Network Modeling Process

The process of tree height prediction modeling was as follows: firstly, the value range of the number of hidden layer nodes, transfer function, and other parameters were initially determined according to the data characteristics and modeling requirement; secondly, the appropriate step-size and value of the above parameters were accurately calculated with the “trial-and-error method” [43], and a series of candidate models was generated subsequently; thirdly, the optimal model was selected with K-fold cross-validation; in the end, the final model was produced by adjusting parameters around one of the previous optimal models with a step-size of 1.
The neural network model was generated according to the following parameters. The neurons of input-layer were DBH, stand basal area, and slope position, and the one in the output-layer was tree height. The transfer function was selected from logsig and tansig. There were two hidden layers, the number of whose neurons was from 3 to 12. The step-size of the “trial-and-error method” was 3, so the candidate numbers of neurons in the hidden-layer were 3, 6, 9, and 12. The MSE and the number of iterations were taken as the evaluation indicators whose priorities were first and second respectively.

2.3. Mixed-Effects Modeling

The mixed-effects model adopts the stand basal area and the DBH as the independent variables, the slope position as the random effect, and the tree height as the target variable. This model has integrated the mixed-effects and the covariance structure in order to describe the tree height and variability at the different slope positions.

2.3.1. Mixed-Effects Model

The mixed-effects model first appeared in the 1970s [44], and was officially proposed by Laird et al. in 1981 [45], whose parameters partially or fully consist of two parts, i.e., fixed and random effects. According to the mixing form between mixed-effects and model, it can be divided into linear and nonlinear mixed-effects models. The model with linear relationship between them is called the linear mixed-effects model; others are known as the nonlinear one. Tree height is affected nonlinearly by many factors, and thus the nonlinear mixed-effects model can better describe tree height.
A mixed-effects model can be considered as the extension of a conventional model, which adds the random effect into the fixed parameters. For applying a mixed-effects model, the basic process is outlined in Figure 1. First of all, the basic model is selected; then, the mixed parameters to which the random effect will be added are determined. Next, the variance–covariance matrix of the random effect and the error effect are calculated successively. After that, the appropriate variance function is chosen to eliminate the heteroscedasticity problem that the error variance is increasing with the increase in the independent variable and results in the failure of the traditional independent equal-variance assumption. Finally, the random parameters of this model are estimated.

2.3.2. Basic Model

Among the factors affecting the tree height, in addition to DBH, stand basal area and slope position are also included. In this study, DBH and stand basal area were considered as the fixed effects, slope position as the random effect, and tree height as the target variable, so as to construct the mixed-effects model. Using random effects to handle spatial heterogeneity and data non-independence issues can effectively improve the fitting accuracy and prediction reliability of tree height growth models.
After thorough analysis and comparison, the basic formula of this model [46,47] is expressed as Equation (10):
H = 1.3 + a + b · B A · D G c
where H represents the tree height; B A , the stand basal area; D G , the DBH; and a , b , and c , the parameters to be estimated.
We calculated the correlation coefficients between fixed effects and random effects, and the results showed that the correlation coefficient between slope position and cross-sectional area was −0.1396, and the correlation coefficient between slope position and DG was −0.0813, indicating that they can be used relatively independently for model construction. This study considered slope position as a random variable to investigate its impact on tree height.
The random effect slope position can be incorporated into different parameters of the model (e.g., a, b, and c). In the mixed-effects model of tree height growth, random effects can be introduced to single or multiple parameters such as asymptotic height, growth rate, and shape parameters to reflect differences among plots and across slope positions, thereby improving fitting and prediction accuracy.

2.3.3. Structure of Mixed-Effects Model

The nonlinear mixed-effects model of tree height reflects the nonlinear dependence of the regression function on the fixed effects (DBH, stand basal area) and the random effects (slope position). The formula is expressed as Equation (11):
H i , j = f ϕ i , d i , j + ε i , j
where i = 1 , 2 , , m , m represents the number of slope positions; j = 1 , 2 , , n i , n i , the number of measurements at the grade i slope position; and ϕ i , the parameter matrix of the grade i slope position. The size of ϕ i is s × 1 ; s is the number of parameters in this model, where s = 3 ; d i , j = D G i , j , B A i , j , D G i , j , and B A i , j are DBH and stand basal area derived from the j-th measurement on the grade i slope position. ε i , j is the prediction error; f , the nonlinear function representing the relationship between tree height and the fixed and the random effects; and H i , j , the predicted value of tree height at the j-th measurement on the grade i slope position.
The parameter matrix ϕ i reflects the change in Larix olgensis at the different slope positions, which is shown as Equation (12):
ϕ i = A i β + B i z i , z i N 0 , D
where β denotes the fixed effect parameter matrix with a size of p × 1 ; and z i is the random effect parameter matrix with a size of q × 1 , which follows the normal distribution with a mean of 0 and variance–covariance matrix of D . p and q represent the number of the fixed and the random effects parameters respectively in this model, where p = 2 and q = 2 ; D , the variance–covariance matrix of random effect (slope position), reflecting the variation and difference between the different slope positions; A i and B i , the designed matrixes of the fixed and the random effects at the grade i slope position.

2.3.4. Calculating Parameter of Mixed-Effects Model

After determining the structure of the mixed-effects model, the relevant parameters were calculated according to the procedure in Figure 2.
(1) 
Determining Parameter
It is necessary to determine which parameter in Equation (11) should be added with the random effect before defining the covariance structure of the different slope positions. Thus, the parameters a, b, and c of this formula need to be tested with the appropriate criterion, such as AIC (Akaike Information Criterion), BIC (Bayesian Information Criterion), LL (Log Likelihood), and LLR (Log Likelihood Ratio). For the first three criteria, the smaller is better, but it is the opposite for the last criterion. It is found that adding the random effect (slope position) into the parameters a and b in Equation (11) had the better result.
(2) 
Calculating Variance–Covariance Matrix of Random Effect
The variance–covariance matrix of the random effect represents the variability of slope position, which is all the same for every slope position. This matrix appeared as D in Section 2.3.3. Since the random effect was added into the parameters a and b, the size of D was 2 × 2 , which is shown as follows:
D = σ u 2 σ u v 2 σ u v 2 σ v 2
where u and v represent the random effect parameters added into the parameters a and b; σ u 2 and σ v 2 , the variances of u and v , respectively; and σ u v 2 , the covariance of the two parameters.
(3) 
Calculating Variance–Covariance Matrix of Error Effect
The variance–covariance matrix of the error effect represents the error of slope position, which is different for every slope position. R i is one of the grade i slope positions, which is expressed as Equation (14).
R i = σ 2 G i 0.5 E i G i 0.5
where σ 2 is the error’s variance, which is determined by the residual variance of this model; G i and E i , with the same size of n i × n i , are used to explain the heteroscedasticity and error’s autocorrelation of the grade i slope position respectively; G i is the diagonal matrix. Since every sample area only has one value in the dataset, there is no autocorrelation of multi-values from the same sample area, and thereby E i is the unit matrix.
In order to eliminate the heteroscedasticity, the following functions were considered: exponential function var ε = σ 2 exp α D G , power function var ε = σ 2 D G α , and power function with constant var ε = σ 2 α 1 + D G α 2 2 . Among them, α , α 1 , and α 2 are parameters to be estimated, and D G is the DBH. The optimal residual variance model was selected from the above functions, which included smaller AIC and BIC, but bigger Log Likelihood (LL) was also possible.
(4) 
Estimating Random Parameter
The random effect parameter matrix of the grade i slope position is estimated according to Equation (15).
z i = D L i T R i + L i D L i T 1 ε i
where D and R i are variance–covariance matrixes as in Equations (14) and (15) respectively; L i , the matrix with a size of n i × q , representing the partial derivatives of different predicted values with respect to random effect parameters at the grade i slope position; L i T , the transpose of L i ; and ε i , the residual vector composed of the errors at the n i measurements of the grade i slope position.

2.4. Evaluation Index of the Two Models

Generally speaking, the coefficient of determination ( R 2 ) is used to characterize the fitting quality through the change in the data. The closer the value is to 1, the stronger the model’s ability is to explain the dependent variable, and the model fits the data better. The smaller the root mean square error (RMSE) and the average absolute error (MAE), the higher the accuracy of the model fitting.
Thus, the R 2 , RMSE, and MAE were used as the evaluation indices of the tree height model, as shown in Equation (16), Equation (17), and Equation (18), respectively:
R 2 = 1 l = 1 w Y l Y ^ l 2 Y l Y ¯ 2
R M S E = 1 w l = 1 w Y l Y ^ l 2
M A E = l = 1 w Y l Y ^ l w
In Equation (16), Equation (17), and Equation (18), w is the number of samples; Y is the mean measurements of the samples; and Y l and Y l are the measurement and predicted values of the l-th sample, respectively.

3. Results

3.1. Data Characteristic

The statistical characteristics of the above data are summarized in Table 1.

3.2. Strucutre of Neural Nework

A series of candidate models was generated according to the method in Section 2.2.4, which are shown in Table 2.
It can be seen that the MSE reached the minimum of 0.00136 when the neuron distribution was “3:3:6:1” and the transfer functions were both “tansig”. The corresponding number of iterations was 30.4, which was also less than the one in the most other situations. Thus, this candidate model was taken as the optimal model according to the K-fold cross-validation in Section 2.2.4, where K = 5.
For the further optimization, the number of neurons in the hidden-layer of the optimal model was adjusted with a step-size of 1. Thus, its testing range in the first and the second hidden layers was “1, 2, 3, 4, 5” and “4, 5, 6, 7, 8” respectively. The corresponding training results are listed in Table 3.
It can be seen that when the neurons distribution was “3:1:5:1”, the MSE obtained the minimum of 0.0133, which was smaller than the one of 0.0136 in Table 2. The final model generated by this parameter had the highest precision. The number of iterations in that network was 30, which was not the minimum. But it was obviously less than the average iterations of 33.475 in Table 3, which fully shows its advantage of speed.
The final neural network is shown in Figure 2. The B A , D G , and P W of the input layers represent the stand basal area, DBH, and slope position respectively. The neuron in the first hidden-layer is I 1 , 1 ; those in the second hidden-layer are I 2 , 1 , I 2 , 2 , I 2 , 3 , I 2 , 4 , and I 2 , 5 . The biases between the input-layer and the first hidden-layer, the first and the second hidden layers, and the second hidden-layer and the output-layer were Bias1, Bias2, and Bias3, respectively, as follows:
B i a s 1 = 2.3042
B i a s 2 = 7.0795 ,   3.3531 ,   0.1008 ,   4.4650 ,   7.0193
B i a s 3 = 0.4914
In Figure 2, the transfer function from the input-layer to the first hidden-layer, as well as that from the first to the second hidden-layer, was tansig, whereas, the one from the second hidden-layer to the output-layer was purelin. According to the weight matrix and the corresponding bias, the transfer function from the input-layer to the first hidden-layer is as follows:
I 1 , 1 = tan s i g 0.0170 B A + 0.1813 D G + 0.0015 P W + 2.3042
The transfer functions from the first to the second hidden-layer are as follows:
I 2 , 1 = t a n s i g 7.1615 I 1 , 1 + 7.0795
I 2 , 2 = t a n s i g 7.1550 I 1 , 1 + 3.3531
I 2 , 3 = t a n s i g 6.9874 I 1 , 1 + 0.1008
I 2 , 4 = t a n s i g 6.5161 I 1 , 1 + 4.4650
I 2 , 5 = t a n s i g 6.9861 I 1 , 1 + 7.0193
The transfer function from the second hidden-layer to the output-layer is as follows:
O = purelin 16.9476 I 2 , 1 + 1.0247 I 2 , 2 1.2429   I 2 , 3 + 0.8921 I 2 , 4 + 2.0557 I 2 , 5 + 0.4914

3.3. Construction of Mixed-Effects Model

The random effect (slope position) was added into the parameters a and b in Equation (11). The heteroscedasticity was eliminated according to the procedure in the section “Calculating Variance–Covariance Matrix of Error Effect”, the result of which is shown in Table 4.
It can be seen that the result of the exponential function had the better AIC, BIC, and LL values than the other two functions. Therefore, the exponential function was selected to eliminate the heteroscedasticity.
All data were divided into training set (64 samples) and test set (22 samples), so as to establish the mixed-effects model, the parameters of which are listed in Table 5.
By using the results in Table 4 and Table 5, the formula for the tree height prediction of Larix olgensis based on the nonlinear mixed-effects model was finally determined as follows:
H i , j = 1.3 + 1 . 2306873 + u i + 0 . 0034150 + v i · B A i , j · D G i , j 0 . 8759386 + ε i , j
where B A i , j ,  D G i , j ,  H i , j , and ε i , j , respectively, are the stand basal area, DBH, predicted value, and error at the j-th measurement on the grade i slope position as in Equation (12); u i and v i are the random effect parameters of the grade i slope position, which follows the normal distribution with a mean of 0 and variance–covariance matrix of D , as shown in Equation (20):
u i v i ~ N 0 , D D = 0.00000460513 0.00000122332 0.00000122332 0.00000044662
ε i , j also follows the similar normal distribution as shown in Equation (21):
ε i , j ~ N 0 , R i R i = 1 . 234501 2 · G i 0.5 I i G i 0.5
where R i , G i , and I i are as described in Equation (15); the number of diagonal elements of G i is equal to the number of measurements at the grade i slope position. The j-th diagonal element G i j is calculated as follows:
G i j = exp - 0 . 04903804 · D G i , j
Then, the formula of ε i , j can be expressed synthetically as follows:
var ε i , j = 1 . 234501 2 exp - 0 . 04903804 · D G i , j
Thus, the tree height of Larix olgensis was predicted by substituting the measurements B A i , j and D G i , j and parameters u i , v i , and ε i , j i into Equations (19)~(22).

3.4. Comparsion Results of Evaluation Index

In this study, the DBH, the stand basal area, and the slope position were used as the input factors, and the tree height was used as the output factor. The tree height model was established by using a double hidden-layer neural network. Concurrently, the slope position was taken as the random effects, whereby the mixed-effects model was used to establish the tree height model. In order to compare the effects of the above two methods, precision analysis was performed.
The numbers of samples in training set and test set were 64 and 22. The neural network and the mixed-effects models were compared by using the evaluation indices defined by Equations (16)–(18). The results are listed in Table 6.
In Table 6, the R2 of the double hidden-layer BP neural network was 0.9068, which is higher than the mixed-effects model. That is to say, the neural network accounted for about 90.68% of the tree height variation, but the mixed-effects model only 85.9%. The RMSE and MAE in the neural network were 18.69% and 43.79% lower than the mixed-effects model, respectively. It can be observed from the overall levels that the double hidden-layer BP neural network is superior to the nonlinear mixed-effects model in tree height prediction. Although the coefficient of determination (R2) for the tree height model increased by 4%, tree height is the basis for calculating forest stock and carbon sinks. Under the current trend of precision forestry, precise tree height plays an important role in calculating forest stock and carbon sinks, and has significant meaning and value for forest management and other aspects.
From Table 3, it can be seen that the best neural network “3:1:5:1” has an RMSE of 0.03688 in the training set (derived from the mean square error in this table: 0.00136). From Table 6, it is known that this network has an RMSE of 1.3197 in the validation set. By comparison, it can be concluded that this network has overfitting, and the reason is that the size of the dataset is relatively small (86 records). However, judging the model’s performance solely based on RMSE is not comprehensive; it is necessary to conduct a comparative analysis through the fitting of “predicted value—observed value” and the residual distribution.
The predicted and measured values of the two tree height models were compared, the results of which are plotted in Figure 3.
In this figure, the circles represent the sample points, and the red dotted lines represent the fitting equations. The left subgraph is the scatter diagram of neural network model, and the right subgraph is the one of mixed-effects model. The fitting equations for the two models are y = 1.0384x − 0.4265 and y = 0.9637x + 0.3631 respectively, where x and y are the predicted and measured values respectively. If the fitting equation is closer to y = x, the predicted value is closer to the measured value, which indicates that the corresponding prediction model is better.
The residual is the difference between the predicted and the measured values. The more concentration around 0, the higher the prediction accuracy. The residual distribution of the predicted values of the two tree height models is plotted in Figure 4.
In this figure, the circles represent the sample points. It can be observed in Figure 4 that within the double hidden-layer BP neural network, there were two plots in a residual interval of (−1, 1), accounting for 9.09%; the residual interval of (−2, 2) had 22 plots, accounting for 100%. When using the mixed-effects model, there were zero plots in a residual interval of (−1, 1) and 21 plots in a residual interval of (−2, 2), accounting for 95.45%. It can also be observed that the difference between the predicted value and observed value of the double hidden-layer BP neural network was significantly lower than that of the nonlinear mixed-effects model.
From the above results, the fitting effect of the neural network model is slightly better than that of the mixed model, and their residual distributions conform to the model assumptions. Thus, it can be concluded that through the reasonable effect setting and model selection, the overfitting risk has been successfully controlled, and the constructed model has a good fitting effect.
In summary, the double hidden-layer BP neural network is superior to the nonlinear mixed-effects model in tree height prediction.

4. Discussion

We established two kinds of models. One is the double hidden layer BP neural network, and the other is the mixed-effects model. Among these models, the best double hidden-layer BP neural network was determined according to MSE and iterations, which is used as the optimal height–diameter model for Larix olgensis forest plantation in Jilin Province, China. Larix olgensis is one of the dominant fast-growing afforestation tree species, which is important in wood production and other aspects and is widely used in the electricity industry, coal mines, manufacturing ships, bridges, railways, etc. Thus, Larix olgensis was taken as the research objective in our study. This study compared the prediction effects of double-layered BP neural network and mixed-effects models. The experimental results indicated that the higher performance was obtained with two hidden layers, which includes higher fitting precision, higher estimating efficiency, and fewer iterations [23]. That is to say, the double hidden-layer BP neural network appears substantially superior to the mixed-effects model in prediction of tree height (Table 6; Figure 3 and Figure 4). Our results are not consistent with those from Castaño-Santamaría et al. [12] and Özçelik et al. [13], which predict tree height of the uneven-aged beech forests in northwestern Spain and Crimean juniper in the southwestern region of Turkey. They compared the neural network model against the mixed-effects model and found the latter was better. I do not think this discrepancy comes from tree species used. The main reason is that we use the double hidden layer neural network, but they only apply a single hidden-layer neural network. For example, Özçelik et al. only used the single hidden-layer with only one or two nodes, and did not investigate the effects of multiple hidden layers on the precision and determine the appropriate transfer functions. In addition, the “trial and error approach”, k-fold cross-validation, and combinatorial optimization [23] were adopted to select the optimal model, which is substantially different from the previous studies [8,10,12,13]. It can help to optimize the structure of the neural network, such as the transfer functions and the number of hidden layers and its nodes, which has substantial effects on the precision of the neural network model. Castaño-Santamaría [12] considered the change in input factors, but did not take into account other factors, such as different transfer functions, and did not introduce the process of determining the optimal neural network. Castro et al. [48] established the multi-layer perceptron neural network growth model for Eucalyptus. They estimated the annual mortality of the best structure with three, four, and one neurons in the input, hidden, and output layers respectively. All these studies [12,48] compared the precisions using different input variables, but did not compare the precisions using different transfer functions and the number of neurons. In our study, the neural network modeling produced the higher fitting precision compared to the mixed-effects models (Table 6). Generally speaking, the neural network has the strong robustness, and the mixed-effects model has biological significance, such as the parameters describing growth rate and pattern. But it is very complex to determine the optimal parameter of the mixed-effects model compared with the procedure of the neural network. Furthermore, the BP neural network model has better generalization ability, and therefore can approximate any nonlinear continuous function with high precision. The BP neural network generally follows nonlinear patterns, which make it suitable for solving the problems caused by factors affecting plant growth. The process of determining the traditional height–diameter model and neural network height–diameter model was also evaluated in our study. The traditional regression modeling needs the comparison and evaluation of the fitting precisions of the candidate models, and the model with the highest precision could be selected as the final model. The neural network model should determine the number of hidden layers and neurons in each layer, and the form of the transfer function according the precision and iterations, and the optimal model can be generated with the best structure.
Due to the small size of the dataset, the trained neural network exhibits overfitting to a certain extent. However, by comparing the “predicted value—observed value” fitting and residual distribution with the mixed-effects model, the fitting effect of the neural network model is slightly better than that of the mixed model, and the residual distribution also conforms to the model assumptions. This indicates that the constructed model can control the risk of overfitting and the prediction effect meets the actual needs.
In addition, this study is different from the authors’ previous study [23]. In previous research, the input factor of the neural network only considered the DBH. In addition to DBH, this research considered two additional factors, i.e., stand basal area and slope position. In the authors’ previous study [23], the input factor of the neural network and mixed-effects models was different. One used DBH, the other DBH and sample plot level. But the input factors in this study were both the same. That is to say, the result of this study is more convincing. In addition, the tree species studied was different. The tree species in the previous study was Poplar, while that in this study was Larix olgensis. But the research results are basically the same, which means that the fitting accuracy of the neural network is higher than that of the mixed-effects model.

5. Conclusions

Improvement of tree height prediction accuracy is not only a necessity for forestry production, but is also key to laying a solid foundation for carbon reduction research and quantifying the ecological value of forests. Deepening the optimization of tree height prediction technology cannot only solve the pain points in forestry resource monitoring, but also provide strong support for national carbon cycle research and climate governance decisions. It has both theoretical innovation significance and practical application value, and is a core research direction in the forestry and ecological fields.
This study proposed a method for optimizing the hidden-layer number, node number, and transfer function of tree height prediction models based on the BP neural network. First, the “trial-and-error method” was used to determine the value of the aforementioned parameters, and several models were established. Then, by using K-fold cross-validation, the optimal neural network was initially screened according to the mean square error and the number of iterations. After that, the near-domain search was performed around the structure of the optimal neural network with a step-size of 1, and a number of models was generated. K-fold cross-validation was again implemented. Finally, the final neural network was determined using the evaluation indicators of mean square error and number of iterations for K-fold cross-validation.
The Larix olgensis plantation in Jilin Province was used to establish and optimize the tree height prediction model. The results indicated that the final neural network structure was 3:1:5:1, and the transfer functions from the input-layer to the first hidden-layer and from the first to the second hidden-layer were both tansig. Thus, this model accounted for about 90.68% of the tree height variation. On the basis of determining the basic model, by introducing random effects, calculating the variance–covariance matrix, and estimating random parameters, the mixed-effects model was established for comparison with the neural network model. By using R2, RMSE, and MAE as evaluation indicators, we compared the fitting precision of the two models in tree height prediction, as well as the residual distribution. The experimental results indicate that the double hidden-layer BP neural network is superior to the nonlinear mixed-effects model in tree height prediction, although it takes a lot of time in model construction (setting number of layers and nodes, and transfer functions) and model training.
In this study, the slope position was regarded as a random effect in the prediction of tree height. This approach can control the spatial heterogeneity and intra-group correlation of the site, save degrees of freedom, improve the efficiency of model prediction, and distinguish between fixed driving factors and random site fluctuations. The experimental results show that this method can suppress overfitting and improve the prediction accuracy. Currently, there are relatively few studies in this area, which is of certain forward-looking and innovative significance.
To sum up the above statement, this paper has compared the tree height prediction models of a double hidden-layer BP neural network and a nonlinear mixed-effects model, which is useful for selecting the appropriate method in tree height prediction. In addition to the slope position, the stand mean height is also affected by other factors. In subsequent studies, these factors need to be incorporated into the tree height prediction model so as to improve the prediction accuracy.

Author Contributions

Conceptualization, J.S., G.W., and X.L.; methodology, J.S., G.W., and Y.P.; software, J.S. and Y.P.; formal analysis, J.S. and G.W.; data curation, J.S., Y.L., and G.W.; writing—original draft preparation, J.S.; writing—review and editing, G.W. and J.S.; visualization, J.S. and G.W.; supervision, G.W. and Y.P.; project administration, X.L. and Y.L.; funding acquisition Y.P. and J.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Zhejiang Provincial Soft Science Research Project (Grant No. 2026C35055), the Forestry Public Welfare Scientific Research Project of China (Grant No. 201504303), 2025 Open Project of Key Laboratory of Agricultural Monitoring and Early Warning Technology, Ministry of Agriculture and Rural Affairs, P.R. China (Grant No. JCYJKFKT2501), Guangdong Province Key Construction Discipline Research Capacity Enhancement Project (Grant No. 2024ZDJS101), Guangdong Province Key Research Project of Colleges and Universities (Grant No. 2025ZDZX3054) and Key Project for Industrial Upgrading in Longwan Modern Agricultural Industry Research Institute (Grant No. 2025LWYJY20).

Data Availability Statement

Data used in this study are available from the National Forestry and Grassland Science Data Center (http://www.cfsdc.org/).

Conflicts of Interest

The authors declare no competing financial interests.

References

  1. Watt, M.S.; Dash, J.P.; Bhandari, S.; Watt, P. Comparing parametric and non-parametric methods of predicting Site Index for radiata pine using combinations of data derived from environmental surfaces, satellite imagery and airborne laser scanning. For. Ecol. Manag. 2015, 357, 1–9. [Google Scholar] [CrossRef]
  2. Burkhart, H.E.; Tomé, M. Modeling Forest Trees and Stands; Springer: Dordrecht, The Netherlands, 2016. [Google Scholar]
  3. Ibrahim, E.M.M.; Osman, E.M.H.; Idris, E.Z.A. Height-Diameter Prediction Models for Some Utilitarian Natural Tree Species. J. For. Prod. Ind. 2013, 2, 31–39. [Google Scholar]
  4. Kearsley, E.; Moonen, P.C.; Hufkens, K.; Doetterl, S.; Lisingo, J.; Bosela, F.B.; Boeckx, P.; Beeckman, H.; Verbeeck, H. Model performance of tree height-diameter relationships in the central Congo Basin. Ann. For. Sci. 2017, 74, 7. [Google Scholar] [CrossRef]
  5. Park, G.; Lee, D.; Seo, Y.; Choi, J. Height Growth Models for Pinus thunbergii in Jeju Island. J. For. Environ. Sci. 2015, 31, 255–260. [Google Scholar] [CrossRef]
  6. Sumida, A.; Miyaura, T.; Torii, H. Relationships of tree height and diameter at breast height revisited: Analyses of stem growth using 20-year data of an even-aged Chamaecyparis obtusa stand. Tree Physiol. 2013, 33, 106–118. [Google Scholar] [CrossRef]
  7. Bontemps, J.D.; Hervé, J.C.; Duplat, P.; Dhôte, J. Shifts in the height-related competitiveness of tree species following recent climate warming and implications for tree community composition: The case of common beech and sessile oak as predominant broadleaved species in Europe. Oikos 2012, 121, 1287–1299. [Google Scholar] [CrossRef]
  8. Scrinzi, G.; Marzullo, L.; Galvagni, D. Development of a neural network model to update forest distribution data for managed alpine stands. Ecol. Model. 2007, 206, 331–346. [Google Scholar] [CrossRef]
  9. Pirotti, F.; Guarnieri, A.; Vettore, A. Neural network and quad-tree approach to extract tree position and height from LiDAR data. In Proceedings of the SilviLaser, The 8th International Conference on LiDAR Applications in Forest Assessment and Inventory, Edinburgh, UK, 17–19 September 2008. [Google Scholar]
  10. Diamantopoulou, M.J.; Milios, E. Modelling total volume of dominant pine trees in reforestations via multivariate analysis and artificial neural network models. Biosyst. Eng. 2010, 105, 306–315. [Google Scholar] [CrossRef]
  11. Leite, H.G.; da Silva, M.L.M.; Binoti, D.H.B.; Fardin, L.; Takizawa, F.H. Estimation of Inside-bark Diameter and Heartwood Diameter for Tectona grandis, Linn. Trees Using Artificial Neural Networks. Eur. J. For. Res. 2011, 130, 263–269. [Google Scholar] [CrossRef]
  12. Castaño-Santamaría, J.; Crecente-Campo, F.; Fernández-Martínez, J.L.; Barrio-Anta, M.; Obeso, J.R. Tree height prediction approaches for uneven-aged beech forests in northwestern Spain. For. Ecol. Manag. 2013, 307, 63–73. [Google Scholar] [CrossRef]
  13. Özçelik, R.; Diamantopoulou, M.J.; Crecentecampo, F. Estimating Crimean Juniper Tree Height Using Nonlinear Regression and Artificial Neural Network Models. For. Ecol. Manag. 2013, 306, 52–60. [Google Scholar] [CrossRef]
  14. Miguel, E.P.; Mota, F.C.M.; Téo, S.J. Artificial intelligence tools in predicting the volume of trees within a forest stand. Afr. J. Agric. Res. 2016, 11, 1914–1923. [Google Scholar] [CrossRef]
  15. Nandy, S.; Singh, R.; Ghosh, S.; Watham, T.; Kushwaha, S.P.S.; Kumar, A.S.; Dadhwal, V.K. Neural Network-based Modelling for Forest Biomass Assessment. Carbon Manag. 2017, 8, 305–317. [Google Scholar] [CrossRef]
  16. Hecht-Nielsen, R. Kolmogorov’s Mapping Neural Network Existence Theorem. In The First IEEE International Conference on Neural Networks; IEEE: New York, NY, USA, 1987. [Google Scholar]
  17. Fisher, L.; McDonald, J.N. Fixed Effects Analysis of Variance; Academic Press: New York, NY, USA, 2014. [Google Scholar]
  18. Kalbi, S.; Fallah, A.; Bettinger, P.; Shataee, S.; Yousefpour, R. Mixed-effects modeling for tree height prediction models of Oriental beech in the Hyrcanian forests. J. For. Res. 2017, 3, 1195–1204. [Google Scholar] [CrossRef]
  19. Razali, W.W.; Razak, T.A.; Azani, A.M.; Kamziah, A.K. Mixed-effects models for predicting early height growth of forest trees planted in Sarawak, Malaysia. J. Trop. For. Sci. 2015, 27, 267–276. [Google Scholar]
  20. Lam, T.Y.; Kershaw, J.A.; Hajar, Z.S.N.; Rahman, K.A.; Weiskittel, A.R.; Potts, M.D. Evaluating and modeling genus and species variation in height-to-diameter relationships for Tropical Hill Forests in Peninsular Malaysia. Forestry 2016, 90, 268–278. [Google Scholar]
  21. Ma, W.; Lei, X. Nonlinear Simultaneous Equations for Individual-Tree Diameter Growth and Mortality Model of Natural Mongolian Oak Forests in Northeast China. Forests 2015, 6, 2261–2280. [Google Scholar] [CrossRef]
  22. Ding, S.; Su, C.; Yu, J. An Optimizing BP Neural Network Algorithm Based on Genetic Algorithm. Artif. Intell. Rev. 2011, 36, 153–162. [Google Scholar] [CrossRef]
  23. Shen, J.; Hu, Z.; Sharma, R.P.; Wang, G.; Meng, X.; Wang, M.; Wang, Q.; Fu, L. Modeling Height–Diameter Relationship for Poplar Plantations Using Combined-Optimization Multiple Hidden Layer Back Propagation Neural Network. Forests 2020, 11, 442. [Google Scholar] [CrossRef]
  24. Thomas, A.J.; Petridis, M.; Walters, S.D.; Gheytassi, S.M.; Morgan, E.R. On predicting the optimal number of hidden nodes. In Proceedings of the International Conference on Computational Science and Computational Intelligence, Las Vegas, NV, USA, 7–9 December 2015; pp. 565–570. [Google Scholar]
  25. FECIT Technological Product Research Center. Neural Network Theory and Implementation with MATLAB 7; Publishing House of Electronics Industry: Beijing, China, 2005. (In Chinese) [Google Scholar]
  26. Stathakis, D. How many hidden layers and nodes? Int. J. Remote Sens. 2009, 30, 2133–2147. [Google Scholar] [CrossRef]
  27. Wanas, N.; Auda, G.; Kamel, M.S.; Karray, F. On the optimal number of hidden nodes in a neural network. In Proceedings of the IEEE Canadian Conference on Electrical and Computer Engineering, Waterloo, ON, Canada, 25–28 May 1998; pp. 918–921. [Google Scholar]
  28. Xu, S.; Chen, L. A novel approach for determining the optimal number of hidden layer neurons for FNN’s and its application in data mining. In Proceedings of the 5th International Conference on Information Technology and Applications, Cairns, QLD, Australia, 23–26 June 2008; pp. 683–686. [Google Scholar]
  29. Tan, X.; Ji, Z.; Zhang, Y. Non-invasive continuous blood pressure measurement based on mean impact value method, BP neural network, and genetic algorithm. Technol. Health Care 2018, 26, 87–101. [Google Scholar] [CrossRef] [PubMed]
  30. Miller, J.W.; Goodman, R.; Smyth, P. On loss functions which minimize to conditional expected values and posterior probabilities. IEEE Trans. Inf. Theory 1993, 39, 1404–1408. [Google Scholar] [CrossRef]
  31. Russell, R.; Robert, J.M. Neural Smithing: Supervised Learning in Feedforward Artificial Neural Networks; MIT Press: Cambridge, MA, USA, 1999. [Google Scholar]
  32. Huang, N.; Lin, L. An Improved BP Neural Network Model Based on Quasic-Newton Algorithm. In Proceedings of the Fifth International Conference on Natural Computation, Tianjin, China, 14–16 August 2009. [Google Scholar]
  33. Lu, P.X. Research on BP Neural Network Algorithm Based on Quasi-Newton Method. Appl. Mech. Mater. 2014, 686, 388–394. [Google Scholar] [CrossRef]
  34. Lewis, A.S.; Overton, M.L. Nonsmooth optimization via quasi-Newton methods. Math. Program. 2013, 141, 135–163. [Google Scholar] [CrossRef]
  35. Du, S.S.; Lee, J.D.; Li, H.; Wang, L.; Zhai, X. Gradient Descent Finds Global Minima of Deep Neural Networks. In Proceedings of the 36th International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019. [Google Scholar]
  36. Martin, F.M. A scaled conjugate gradient algorithm for fast supervised learning. Neural Netw. 1993, 6, 525–533. [Google Scholar]
  37. Chen, L.N.; Aihara, K. Chaotic simulated annealing by a neural network model with transient chaos. Neural Netw. 1995, 8, 915–930. [Google Scholar] [CrossRef]
  38. Shanno, D.F. Conditioning of quasi-Newton methods for function minimization. Math. Comput. 1970, 24, 647–656. [Google Scholar] [CrossRef]
  39. Fletcher, R.; Powell, M.J.D. A rapidly convergent descent method for minimization. Comput. J. 1963, 6, 163–168. [Google Scholar] [CrossRef]
  40. Fletcher, R.; Spedicato, E. An Overview of Unconstrained Optimization. In Algorithms for Continuous Optimization; NATO ASI Series (Series C: Mathematical and Physical Sciences); Springer: Dordrecht, The Netherlands, 1994; Volume 434, pp. 109–143. [Google Scholar]
  41. Nocedal, J. Updating Quasi-Newton Matrices with Limited Storage. Math. Comput. 1980, 35, 773–782. [Google Scholar] [CrossRef]
  42. Zhang, Y.; Yang, Y. Cross-validation for selecting a model selection procedure. J. Econom. 2015, 187, 95–112. [Google Scholar] [CrossRef]
  43. Liu, C.Y.; Ling, J.C.; Kou, L.Y.; Wu, J.Q. Performance Comparison between GA BP Neural Network an BP Neural Network. Chin. J. Health Stat. 2013, 30, 173–176, (In Chinese with English Abstract). [Google Scholar]
  44. Harville, D.A. Maximum likelihood approaches to variance component estimation and to related problems. J. Am. Stat. Assoc. 1977, 72, 320–338. [Google Scholar] [CrossRef]
  45. Laird, N.M.; Ware, J.H. Random effects models for longitudinal data. Biometrics 1982, 38, 963–974. [Google Scholar] [CrossRef] [PubMed]
  46. Hui, G.Y.; Gadow, K.V. Zur entwicklung von einheitshöhenkurven am beispel der baumart cunninghamia lanceolata. Allg. Forstund Jagdztg. 1993, 164, 218–220. [Google Scholar]
  47. Brunori, A.; Nair, P.K.R.; Rockwood, D.L. Performance of two Eucalyptus species at different slope positions and aspects in a contour-ridge planting system in the Negev Desert of Israel. For. Ecol. Manag. 1995, 75, 41–48. [Google Scholar] [CrossRef]
  48. Vinícius Oliveira Castro, R.; Boechat Soares, C.P.; Leite, H.G.; Lopes de Souza, A.; Saraiva Nogueira, G.; Bolzan Martins, F. Individual growth model for Eucalyptus stands in Brazil using artificial neural network. Forestry 2013, 2013, 1–12. [Google Scholar]
Figure 1. Flowchart for constructing the mixed model.
Figure 1. Flowchart for constructing the mixed model.
Plants 15 01176 g001
Figure 2. Tree height model based on the BP neural network.
Figure 2. Tree height model based on the BP neural network.
Plants 15 01176 g002
Figure 3. Comparison between predicted value and measured value of two tree height models.
Figure 3. Comparison between predicted value and measured value of two tree height models.
Plants 15 01176 g003
Figure 4. Residual distribution based on BP neural network and mixed-effects models.
Figure 4. Residual distribution based on BP neural network and mixed-effects models.
Plants 15 01176 g004
Table 1. Statistical characteristics of the Larix olgensis data.
Table 1. Statistical characteristics of the Larix olgensis data.
CharacteristicMinMaxMeanStandard Deviation
DBH/cm5.726.112.44.3
Stand basal area/m2·hm−21.3532.5615.447.97
Tree height/m52212.83.8
Table 2. Training results from neural networks.
Table 2. Training results from neural networks.
Neurons in Each LayerMSEIterationsMSEIterationsMSEIterationsMSEIterations
log:logtan:tanlog:tantan:log
3:3:3:10.0613 35.0 0.0242 35.6 0.0280 53.2 0.0295 42.8
3:3:6:10.0278 35.0 0.013630.40.0210 42.2 0.0234 34.0
3:3:9:10.0146 31.4 0.0188 40.8 0.0416 37.2 0.0163 30.0
3:3:12:10.0244 29.8 0.0180 36.2 0.0357 31.0 0.0255 36.8
3:6:3:10.0262 30.6 0.0283 31.2 0.0310 29.6 0.0561 24.0
3:6:6:10.0304 35.6 0.0279 37.8 0.0360 22.0 0.0216 28.8
3:6:9:10.0243 46.8 0.0190 38.8 0.0368 29.0 0.0278 36.4
3:6:12:10.0232 38.0 0.0242 32.0 0.0642 22.4 0.0322 34.6
3:9:3:10.0545 40.8 0.0250 28.0 0.0412 24.4 0.0184 50.6
3:9:6:10.0362 29.0 0.0271 24.2 0.0265 31.2 0.1304 35.2
3:9:9:10.0357 39.0 0.0142 29.6 0.0385 24.8 0.0390 26.4
3:9:12:10.0360 29.2 0.0401 33.8 0.0631 48.8 0.0361 28.8
3:12:3:10.0310 41.4 0.0326 21.4 0.0423 35.6 0.0244 26.6
3:12:6:10.0553 26.2 0.0208 31.6 0.0356 37.8 0.0330 33.2
3:12:9:10.0402 34.2 0.0400 27.6 0.0211 41.8 0.0271 36.8
3:12:12:10.0200 34.2 0.0697 25.0 0.0205 32.8 0.0273 34.0
Table 3. Training results from local optimized neural networks.
Table 3. Training results from local optimized neural networks.
Neurons in Each LayerMSEIterationsNeurons in Each LayerMSEIterationsNeurons in Each LayerMSEIterations
tan:tantan:tantan:tan
3:1:4:10.015827.63:2:7:10.083837.83:4:6:10.071040.8
3:1:5:10.013330.03:2:8:10.046029.43:4:7:10.026643.6
3:1:6:10.095231.03:3:4:10.016439.23:4:8:10.017241.2
3:1:7:10.062428.03:3:5:10.048024.63:5:4:10.019135.2
3:1:8:10.019229.83:3:7:10.054931.63:5:5:10.031726.6
3:2:4:10.016641.43:3:8:10.033929.43:5:6:10.020031.4
3:2:5:10.060632.43:4:4:10.026528.23:5:7:10.016436.0
3:2:6:10.062028.03:4:5:10.018644.43:5:8:10.017735.8
Table 4. Residual simulation results from nonlinear mixed model for predicting height of Larix olgensis.
Table 4. Residual simulation results from nonlinear mixed model for predicting height of Larix olgensis.
Residual Variance ModelFormulaAICBICLL
Exponential function var ε = σ 2 exp α D G 147.2112164.4823−65.60562
Power function var ε = σ 2 D G α 147.927165.1981−65.9635
Power function with constant var ε = σ 2 α 1 + D G α 2 2 149.927169.357−65.96352
Table 5. Parameters of mixed-effects model for predicting height of Larix olgensis.
Table 5. Parameters of mixed-effects model for predicting height of Larix olgensis.
ItemsNameValueStandard Deviationp Value
fixed parametersa1.23068730.09217312p = 0.0000 < 0.05
b0.00341500.00158609p = 0.0356 < 0.05
c0.87593860.03105736p = 0.0000 < 0.05
variance σ u 2 0.00000460513
σ v 2 0.00000044662
σ u v 2 −0.00000122332
σ 2 1.2345006919
Table 6. Accuracy comparison between BP neural network and mixed-effects models.
Table 6. Accuracy comparison between BP neural network and mixed-effects models.
ModelsR2RMSEMAE
Neural network model0.90681.31971.2736
Mixed-effects model0.85901.62302.2658
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Shen, J.; Lei, X.; Li, Y.; Pan, Y.; Wang, G. Tree Height Prediction Using a Double Hidden-Layer Neural Network and a Mixed-Effects Model. Plants 2026, 15, 1176. https://doi.org/10.3390/plants15081176

AMA Style

Shen J, Lei X, Li Y, Pan Y, Wang G. Tree Height Prediction Using a Double Hidden-Layer Neural Network and a Mixed-Effects Model. Plants. 2026; 15(8):1176. https://doi.org/10.3390/plants15081176

Chicago/Turabian Style

Shen, Jianbo, Xiangdong Lei, Yutang Li, Yuehong Pan, and Gongming Wang. 2026. "Tree Height Prediction Using a Double Hidden-Layer Neural Network and a Mixed-Effects Model" Plants 15, no. 8: 1176. https://doi.org/10.3390/plants15081176

APA Style

Shen, J., Lei, X., Li, Y., Pan, Y., & Wang, G. (2026). Tree Height Prediction Using a Double Hidden-Layer Neural Network and a Mixed-Effects Model. Plants, 15(8), 1176. https://doi.org/10.3390/plants15081176

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop