Next Article in Journal
Discrimination of Mangrove Stages Using Multitemporal Sentinel-1 C-Band Backscatter and Sentinel-2 Data—A Case Study in Samut Songkhram Province, Thailand
Previous Article in Journal
Clonal Diversity and Fine-Scale Genetic Structure of a Keystone Species: Ilex aquifolium
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Exploring the Potential of Machine Learning for Modeling Growth Dynamics in an Uneven-Aged Forest at the Level of Diameter Classes: A Comparative Analysis of Two Modeling Approaches

by
Gaspard Dumollard
School of Agricultural, Forest and Food Sciences HAFL, Bern University of Applied Sciences BFH, 3052 Zollikofen, Switzerland
Forests 2022, 13(9), 1432; https://doi.org/10.3390/f13091432
Submission received: 15 July 2022 / Revised: 1 September 2022 / Accepted: 2 September 2022 / Published: 6 September 2022
(This article belongs to the Section Forest Inventory, Modeling and Remote Sensing)

Abstract

:
Growth models of uneven-aged forests on the diameter class level can support silvicultural decision making. Machine learning brings added value to the modeling of dynamics at the stand or individual tree level based on data from permanent plots. The objective of this study is to explore the potential of machine learning for modeling growth dynamics in uneven-aged forests at the diameter class level based on inventory data from practice. Two main modeling approaches are conducted and compared: (i) fine-tuned linear models differentiated per diameter class, (ii) an artificial neural network (multilayer perceptron) trained on all diameter classes. The models are trained on the inventory data of the Canton of Neuchâtel (Switzerland), which are area-wide data without individual tree-level growth monitoring. Both approaches produce convincing results for predicting future diameter distributions. The linear models perform better at the individual diameter class level with test R2 typically between 50% and 70% for predicting increments in the numbers of stems at the diameter class level. From a methodological perspective, the multilayer perceptron implementation is much simpler than the fine-tuning of linear models. The linear models developed in this study achieve sufficient performance for practical decision support.

1. Introduction

Uneven-aged forests are generally defined as forests whose stands contain several cohorts of trees of different ages, which results in a relative structural heterogeneity of these stands. Despite the diversity of uneven-aged management methods, the management of these stands is often associated with the maintenance of a targeted distribution of diameters [1]. A good understanding of the dynamics within an uneven-aged stand at the global level but also at the level of diameter classes then allows for better planning of cuts, in the sense of choosing the timing and the intensity of the next cut [2], including the intensity of the cut by diameter class. Growth models for uneven-aged stands can therefore support silvicultural decision making.
The literature on existing uneven-aged forest growth models identifies three possible resolutions for this type of model: the whole stand, diameter classes and individual trees [3,4]. It is also possible to make a distinction between empirical and mechanistic (also called process-based) models [3]. This distinction is theoretically relevant, although there is a gradient between these two archetypes, with models estimated on the basis of empirical data but whose structure is to some extent inspired by theoretical mechanisms of forest growth [5].
Stand-level models aim at predicting the increment of a key stand-level variable, typically the increment of the standing volume or basal area. The recent literature has seen the publication of several such studies based on machine learning methods and data from permanent plots [6,7,8]. In particular, some of these studies use artificial neural networks among other regressors [7,8]. All of these models achieve good predictive performance and can be useful in practice, e.g., in management systems where harvested volumes are planned on the basis of increments. However, they do not provide information on stand structure.
Models at the individual tree level aim to predict individual diameter increments. Several recent studies have approached this topic from a machine learning perspective [4,9], for example, by testing several types of models including an artificial neural network [9]. These models perform well and give a maximum of details on growth within a stand, but individual monitoring data on diameter growth, i.e., data from permanent plots, are needed to train them. When this is the case, the data are usually available in large numbers, which fits well to machine learning approaches. For example, one of these studies [4] is based on 16,619 observations from the monitoring of 20 permanent plots.
Models based on diameter classes generally aim to predict increments in the number of stems per diameter class. Some studies [5,10] propose for example an empirical modeling of radial increments per diameter class based on inventory data using a linear model. These radial increments are then integrated into a process-based model to simulate the dynamics of the passage of trees between two successive classes, as well as the recruitment into the first diameter class. Numerous other studies [11,12,13,14] use an approach that is quite similar in principle but generalized and formalized. Passage rates are modeled directly (and empirically with linear models), and the increment in the number of stems in a given diameter class potentially depends on the respective numbers of stems in all other diameter classes. This type of model accounts for passages between two non-successive classes, competition effects between diameter classes, as well as the influence of the diameter distribution on recruitment. These models are formalized as matrix equation systems allowing the recursive simulation of stand dynamics, sometimes called a matrix model or transition matrix model [3]. The predictive performances of those models are limited compared to recent machine learning models developed on the stand or individual tree levels.
The canton of Neuchâtel in Switzerland is historically and professionally recognized as the Swiss high place of the selective felling forest management, which is implemented on a tree-by-tree basis (see Section 2.2 for further details). In the canton of Neuchâtel, selective felling management is implemented through the so-called control method (Méthode du contrôle) [3,15,16]. This method is based on periodic and complete inventories of stands as well as on a direct control of all harvests. Some studies [5,10] use an extract of these data. Some other studies [12,14] use data from the French Jura mountain where a similar management and controlling system prevail. These inventory data from the practice are non-experimental and are not based on individual tree monitoring, which makes their analysis more complex than data from permanent plots.
The objective of this study is to explore the potential of machine learning for modeling growth dynamics at the level of diameter classes from the available inventory data of the Canton of Neuchâtel. From a practical point of view, the aim is to determine if a model of this type can predict the evolution of the distribution of diameters that is specific to a given stand with sufficient accuracy to serve as decision support for the planning and implementation of cuts.
To achieve this goal, a machine learning workflow is implemented for data preparation and preprocessing, modeling, and evaluation of the models for their out-of-the-box predictive performances. This workflow can iteratively refine the data preprocessing and modeling steps. As to the modeling step, two main approaches are compared: (i) an approach based on fine-tuned linear models differentiated per diameter class, analogous in spirit to previous matrix models [11], (ii) an approach based on an artificial neural network trained on all diameter classes. The potential of these two types of models for predicting a future state at the stand level is analyzed in absolute terms, but they are also compared with each other and with previous models found in the literature. Moreover, a comparative analysis of features importance and the impact of data scarcity on model performances is proposed. The best model is also evaluated for its ability to be used in practice, i.e., for its ability to predict the evolution of the overall diameter distribution at the stand level, and to predict increments of important aggregated variables.

2. Materials and Methods

2.1. Geographical and Climatic Context in the Canton of Neuchâtel

The canton of Neuchâtel, in Switzerland, is located in the Jura Mountain range. Its altitude stretches between 429 and 1552 m. Its climate is humid continental [17]. In terms of altitudinal zonation, the territory of the canton is currently located in the submontane, lower montane and upper montane zones [18]. Simulations of two possible climatic futures [18] also tend to show, in the canton, a transition from the lower and upper montane zones to the submontane zone. The southern part of the Canton, located at a lower altitude and in the minority, would even enter the foothill zone.

2.2. Brief Summary on Forest Management and Forest Data Collection in the Canton of Neuchâtel

The selection felling management that is historically applied in the canton of Neuchâtel is an uneven-aged management method where silviculture is implemented on a tree-by-tree basis. This type of management is based on periodic and situative cuts. The selection of trees to be harvested is the result of multiple determinants (wood harvesting, fostering tree vitality, fostering wood quality, continuous regeneration), but these cuts affect all diameter classes and are characterized by a fairly constant harvesting intensity that depends on the periodicity of interventions [2]. At the stand level, selective felling management results in the transformation to or the maintenance of a state of equilibrium that can be expressed in terms of diameter distribution within a given stand. A typical example of a diameter distribution at the stand level is shown in Figure 1.
Information on the state of a stand, including the diameter distribution, also helps with the planning of the next cuts, as management goals on the stand level can be expressed as a desired state of equilibrium in terms of species mixture, standing volume, and stand structure [2].
Cuts take place periodically, typically every 7 to 12 years (in the Canton of Neuchâtel). The maintenance of the equilibrium is made possible by the fact that the harvesting of given trees allows for a “continuous” natural regeneration within the stand and allows for the further growth of trees in the intermediate strata.
Management by selective felling on a tree-by-tree basis is generally based on species that are shade tolerant, at least in their youth, typically but not exclusively white fir, spruce and beech. Since regeneration occurs in small patches and trees continue to grow in intermediate strata in the shade of larger trees, heliophilic species are not suitable for this type of management.
In the canton of Neuchâtel, selective felling management has been implemented through the so-called control method (Méthode du contrôle) [3,15,16] since the end of the 19th century. This method is based on several important elements: complete inventory of the stands (in particular individual diameters and species) between two successive cuts, direct control (i.e., inventory) of all harvests (as well as forced harvests), definition of parcels with fixed boundaries as management units called divisions (a term that we use hereafter in lieu of stand). The diameters are recorded by class of 5 cm width and with an inventory threshold of 17.5 cm. The classes are numbered from 1 to 24, and the center of the class in cm is then obtained by the following formula ((#class − 1) * 5 + 20). From the perspective of selective felling management, the purpose of collecting such systematic and detailed information is to be able to accurately characterize the silvicultural state of a division, particularly with respect to the diameter distribution, so as to better plan the next cuts. This information also allows for a detailed monitoring of forests at the division or landscape level over time and thus helps in better forest planning at the landscape level. There are no direct data on recruitment and natural mortality (excluding windfall cases) that are collected through this control method. The application of this method in Neuchâtel allowed for the collection of the large data set used in this study.

2.3. Data and Data Preprocessing

This study is based on a standard machine learning workflow. The data collection and preprocessing steps are presented in this section. Section 2.4 presents the data sets building and features engineering steps, and Section 2.5 presents the modeling approaches.
The raw data consist of a record of the absolute number of stems in a division for a given inventory year, species and diameter class. These data are subjected to preprocessing, which consists of the following steps.
Step 1. The absolute numbers of stems per division are converted into numbers of stems per hectare. The areas of the divisions, whose boundaries are administratively set and (persistently) fixed, are known and could be found for 98% of the divisions mentioned in the data set. The remaining divisions were discarded. Meanwhile, the main forest site of each division (in terms of area) is integrated into the data set. Forest sites are given according to the typology of Ellenberg and Klötzli [19].
Step 2. Only divisions that have been inventoried at least twice are kept, i.e., divisions for which it is possible to calculate increments over at least one period are kept, which corresponds to 95% of the divisions kept thus far. From this step on, the inventory data are organized by couples (division, inventory year). This couple designates the state of a division at a given date, but it also designates the period between the inventory year and the date of the next inventory to which cuts and windfalls are associated (see step 4) and for which it is possible to determine increments.
Step 3. Only the couples (division, inventory year) whose species are compatible with selective felling management are retained. The three usual species in this management method are white fir, spruce and beech [16], but the presence of other species that are sufficiently tolerant to shade, in the minority or minority, cannot be excluded. Therefore, the following criteria were defined: the majority species (in terms of basal area) must be among fir, spruce or beech, the 2nd majority species must be among fir, spruce, beech, maple, or ash, and the 3rd majority species must be among the same species or flowering ash or be another deciduous species.
Table 1 is a contingency table of the number of couples (division, inventory year) according to their first and second majority species, before application of step 3. The table reveals that the divisions that meet the filtration criteria of step 3 are overwhelmingly in the majority, reflecting the importance of the selective felling management in the Canton of Neuchâtel. The table also reveals that most of the situations that do not meet the defined criteria correspond to couples (division, inventory year) where the 1st or 2nd majority species is oak. These situations may correspond to even-aged stands preferentially located in the south of the canton on less elevated grounds and on forest sites that are better adapted to oak.
Step 4. The inventory data are merged with the respective data on decided cuts and windfalls, which are formatted in the same way. The data on cuts and windfalls are associated with couples (division, inventory year) when the cuts and windfalls occur in this division and between the inventory year and the date of the next inventory.
Step 5. The temporal gap between two successive inventories in the same division is variable, see Figure 2.
There are a large number of successive inventories that are 7 to 12 years apart but also many successive inventories that are 20 to 24 years apart. In the first case, the inventories were carried out before and after the same cut and correspond therefore to the standard application of the control method. In the second case, there are several cuts between two successive inventories. The standard application of the control method is actually historical in the Canton of Neuchâtel, but it has since been adapted, and the current legislation provides for an inventory at least every 25 years for each division, which corresponds to 2 or 3 cutting cycles. In some divisions of historical or technical importance, inventories are maintained on the basis of one inventory per cutting cycle.
The goal of this study is to simulate growth dynamics recursively, i.e., we want to predict the state of a division at date t + Δ t from its state at date t . This is a pragmatic approach that is well suited to the development of the model and its use in practice. The duration Δ t does not need to be constant (see Section 3), but it must be within an interval whose definition meets two constraints: (i) the possibility of finding data describing dynamics over such periods; (ii) the relevance of these durations Δ t for a use of the model as a decision support in practice. Figure 2 shows that the available data correspond to durations from 8 to 26 years. However, this interval is too wide for model development. The diameter data are given in 5 cm diameter classes. For a relatively large annual diameter increase of 5 mm·year−1, a tree takes 10 years to completely cross a diameter class. Over a period of 20 years and more, with such an increase, a tree can pass through two or more diameter classes. The growth mechanisms in the sense of recursive modeling cannot therefore be the same over these two types of periods. We choose to focus on durations Δ t of up to 12 years. This interval allows the model to be used over a sufficiently long period to assist in planning the next cut, and it is possible to use the models recursively to simulate dynamics over longer periods. Data corresponding to gaps of more than 12 years between two successive inventories are left out, so that only 52% of couples (division, inventory year) that had been retained until then are kept.
Step 6. Much of the harvest data are not available. All the data on the cuts, some of which are old, have not yet been compiled and digitized. This gap in the data leads us to keep only the couples (divisions, year of inventory) for which cuts and windfalls are known, i.e., 38% of the couples that had been kept until now. In addition, a few outliers (in small numbers) that stand out by cutting intensities that are much too strong for the selective felling management are also set aside.
Step 7. The number of trees in the upper diameter classes is small or even negligible and therefore does not allow for a reliable modeling of the growth mechanisms among these classes. Therefore, the data must be truncated at a given diameter threshold. The number of trees that are beyond diameter class 16 (class 24 is the maximum class), i.e., trees with a diameter strictly greater than 97.5 cm, are relatively rare compared to the lower classes; they represent about 0.07% of the trees remaining in the data set at this stage. Such a threshold is also high enough that it is not a problem for practical use of the model; thus, we chose this threshold to truncate the data. Nevertheless, simply ignoring trees beyond class 16 could be problematic in some cases. These trees are few in number but large and can therefore have a non-negligible impact on the growth of other trees. For this reason, only couples (division, inventory year) that meet the following criteria are retained: a maximum of 0.5 trees per ha above class 16 in the division and a maximum of 0.125 trees per ha above class 16 for cuts or windfalls. These criteria were defined partly arbitrarily but in the search for a compromise between filtering potentially biased data and keeping a sufficient volume of data. At the end of this step, 77% of the couples (division, year of inventory) that had been kept until then remain.
In the end, after this preprocessing, there are data on 580 couples (division, inventory year), and for each couple, there are data on 16 diameter classes. Individual observations in the data set then correspond to triplets (division, inventory year, diameter class). The mean basal area of remaining couples (division, inventory year) is 28.0 m2·ha−1 (the standard deviation is 7.26). Despite a strong filtration, the volume of data remains important for a study of this type. The data are moreover complete and consistent.

2.4. Data Buildings and Features Engineering

The objective is to develop a model or a set of models to predict the gross annual increment in the number of stems per hectare for a diameter class d (if there were neither cuts nor windfalls, hence the gross increments), noted Δ N d . Δ N d will be the endogenous variables in our models. They can be calculated as follows:
Δ N d = N a f t e r , d N d + N c u t , d + N w i n d f a l l , d Δ t
In Equation (1), N d denotes the number of stems per hectare in diameter class d at the time of the first inventory, N a f t e r , d the number of stems per hectare in diameter class d at the time of the subsequent inventory, N c u t , d   and   N w i n d f a l l , d the respective numbers of stems cut or windfallen per hectare in diameter class d between these two inventories, and Δ t denotes the duration in years between the two successive inventories.
Following exogenous variables (also called features) are defined for use in the models (all of which are expressed per hectare, except the diameter class):
  • The diameter class d ;
  • The number of stems per diameter class at the time of the first inventory, aggregated by species type (resinous/deciduous) N r e s i n o u s , d and N d e c i d u o u s , d , or aggregated for all species N d (for 1 d 16 ) ;
  • The corresponding basal areas (calculated from the N features and the centers of diameter classes): G r e s i n o u s , d , G d e c i d u o u s , d , and G d (for 1 d 16 ) ;
  • The numbers of stems cut and windfallen (aggregated for all species) and expressed on an annual basis:
    N a n n u a l   c u t , d = N c u t , d Δ t ,
    N a n n u a l   w i n d f a l l , d = N w i n d f a l l , d Δ t .
  • The total number of stems per hectare N t o t and the total basal area per hectare G t o t are calculated too;
  • The basal area of all trees overlying a given diameter class is used as an indicator of the competition to which this diameter class is subject. This variable, denoted G c u m (for cumulative basal area), is defined as the sum of the basal areas of every tree larger than the considered diameter class [4,5], in our case:
    G c u m . d = d = d + 1 d m a x G d .
  • The number of stems cut or windfallen annually in the overlying classes (as an indicator of the regulation of competition for a given diameter class):
    N o v e r l y i n g   a n n u a l   c u t , d = d = d + 1 16 N a n n u a l   c u t , d ,
    N o v e r l y i n g   a n n u a l   w i n d f a l l , d = d = d + 1 16 N a n n u a l   w i n d f a l l , d .
This set of features includes indicators of the current state and structure of the division ( N r e s i n o u s , d , N d e c i d u o u s , d , N d , G r e s i n o u s , d , G d e c i d u o u s , d , G d ), indicators of density at the division level ( N t o t , G t o t ), an indicator of competition at the diameter class level ( G c u m . d ), and indicators of sylvicultural management ( N a n n u a l   c u t , d ,   N a n n u a l   w i n d f a l l , d ,   N o v e r l y i n g   a n n u a l   c u t , d ,   N o v e r l y i n g   a n n u a l   w i n d f a l l , d ), all of which can potentially provide additional insight into forest growth. As an additional explanation, the G c u m of a diameter class measures the density of trees in overlying diameter classes. Since competition between trees is on average exerted by larger trees on smaller trees, G c u m effectively measures the competition exerted on a given diameter class within a given division. When some of the trees in these overlying diameter classes are cut or windfallen, the competition exerted on the diameter class may decrease, hence the role of N o v e r l y i n g   a n n u a l   c u t and N o v e r l y i n g   a n n u a l   w i n d f a l l .
These dendrometric variables represent a classical set of variables used in forest sciences to describe the state and dynamics of a forest stand. In addition to these dendrometric variables, the main forest site for each division is also used as a feature [19].

2.5. Modeling Approaches

This study is based on two main modeling approaches: (i) fine-tuned linear models; (ii) an artificial neural network (a multilayer perceptron regressor). The criterion for refining and comparing models is a coefficient of determination R2 calculated on a set of test data that are not used for training the models. This test data set accounts for 20% of total available data. Test data are selected among couples (division, inventory year), not at the level of triplets (division, inventory year, diameter class). Prediction performances are therefore measured on divisions that are completely unknown to the models, which is crucial to test the capacity of the model under practical conditions. The principle of parsimony, which with equal or almost equal performance, leads to favoring the simplest models, is also applied, as well as the search for models that are consistent with theoretical considerations on forest growth.
Algorithms for data transformation, model training and model evaluation are derived from the sklearn library (version 0.23.1, and version 1.0.2 for the SplineTransformer, [20]). Other important libraries are pandas (version 0.24.2, [21]) and numpy (version 1.19.5, [22]).

2.5.1. Detailed Method for Linear Modeling

Linear models are refined following a usual machine learning workflow so as to find the best possible combination of features, data transformation and (linear) regressor. This type of model provides transparent and easily interpretable results from the point of view of forest growth. Moreover, their linearity and the relatively limited number of parameters make them resilient to overfitting problems when sufficient data are available.
Past studies clearly show the essential role of diameter as a predictor of diameter growth [5,9]. However, the effect of diameter is not necessarily separable from the effect of other features. For example, there is no reason that the effect of forest density, competition between trees or harvesting be expressed in the same way on trees of (very) different diameters. In order to take into account these interaction effects (which could turn out to be non-linear) without complicating the modeling process, a distinct model is estimated for each diameter class, resulting in a total of 16 linear models. In order to maintain overall consistency, a common specification (choice of features) of these different models is proposed; only the transformation of the features and the estimation of the coefficients are distinct.
The selection of features is an important and sensitive step in a machine learning or statistical workflow. The principle of parsimony suggests keeping only the features that bring significant additional information to the model. This allows one to obtain a more robust model by limiting multicollinearity problems and facilitates the interpretation of the regression results.
In this study, the features selection is based on a mix of three approaches: (i) Recursive Feature Elimination; (ii) Stepwise Regression by Backward Elimination; (iii) theoretical considerations on forest growth.
Recursive Feature Elimination consists in successively and iteratively removing features from a model from the least to the most important one [23,24]. In the context of a linear model, the relative importance of the features is given by the numerical order of the regression coefficients in terms of absolute value (which are recalculated at each iteration). This purely empirical method makes it possible to obtain a first systematic ranking of the features according to the information they bring to the model. However, this approach is limiting for two reasons. The application of this method is sensitive to scale effects and therefore to the transformation of features (e.g., standardization). Furthermore, this approach does not take into account multicollinearity effects.
In order to complete this approach, a Stepwise Regression method with Backward Elimination is applied [25]. This method is based on the successive and iterative elimination of features, but the elimination is performed on the basis of a model fit criterion. At each iteration, the eliminated features are those that decrease the least (increase the most) fit criterion, here a test R2. This allows one to take into account the multicollinearity effects. In this study, this method is applied at the level of groups of features (e.g., forest sites, density indicators) so as to identify the groups of features that can be removed and those that should be kept in the model. Theoretical considerations help in the definition of groups and further engineering of features. Proceeding by groups also facilitates a common specification for all models.
The models are also optimized from the point of view of the transformation of the features (e.g., standardization) and the choice of a regressor (choice of the loss function, regularization, fine-tuning of hyperparameters). To do so, different specifications are tested in order to find the combination that maximizes the test R2.
In all cases (features selection, estimation of the final models), the fit criterion used is a coefficient of determination (R2) measured on a set of test data not used for training the model and representing 20% of available observations. The root mean squared error (RMSE) is also calculated as additional information (also computed on test data sets). The advantage of the RMSE is that it expresses an average estimation error in absolute value, in the unit of the predicted variable. The evaluations of test R2 and RMSE are always averaged on 10 different train-test splits.

2.5.2. Detailed Method with Multilayer Perceptrons

The use of a multilayer perceptron [26,27], which is a nonlinear model, allows for a more generic and flexible modeling approach. A multilayer perceptron with at least one hidden layer (hence multilayer) can serve as a universal approximator [28], in the sense that it can approximate with a finite number of neurons any continuous function (on particular subdomains, see [28] for further details). The scikit-learn library [27] provides such a ready-to-use regressor. In this case, potential exogenous variables can be all pooled together because this model fits to non-linear effects and interaction effects. However, this type of model, usually based on a large number of parameters (at least larger than linear models), requires large volumes of data to avoid overfitting. In addition, results (the estimated values of parameters) are harder to interpret. The volume of data used in this study is comparatively large, but it is not large in the sense of big data. One of the objectives of this second modeling approach is to analyze the potential of the multilayer perceptron for this modeling problem based on a large but probably limiting data set.
The multilayer perceptron regressor was trained to directly predict the gross numbers of stems (without the removal effect from cuts or windfalls) by diameter class at the end of a growth period of duration Δ t : N a f t e r   g r o s s , d . This direct approach, without going through increments, resulted in better predictive performances. Only one model was trained instead of separate models for each diameter class. Therefore, the training of this model is based on 9280 entries, which is a comparatively large volume of data for this type of application.
The exogenous variables (features) presented in Section 2.4 were all retained, and several features were even added: the diameter (center of the class), the length of the growth period Δ t , which differs from one couple (division, inventory year) to another, as well as the forest site. The model hyperparameters: the number of hidden layers, the number of neurons in hidden layers, and the regularization parameter [27], are tuned by gridsearch and cross-validation. The fit criterion is also a coefficient of determination (R2) measured on test data representing 20% of available observations. The root mean squared error (RMSE) is computed too. Results are all averaged over 10 different train-test-splits.
A permutation importance procedure [29] is also applied to a fitted multilayer perceptron. This procedure determines the importance of each feature to the model by measuring the model score loss (here the test R2) when the entries for this feature are randomly shuffled (the shuffling and the measure are here repeated 30 times for each feature).

3. Results

3.1. Results of the Linear Modeling

3.1.1. Features Selection

The best results with linear models are obtained by training a separate model for each diameter class. These models are of the following form for any diameter class 1 d 16 and for each couple (division, inventory year) i :
Δ N d , i = β 0 d + β d 2 d · N d 2 , i + β d 1 d · N d 1 , i + β d d · N d , i + β d + 1 d · N d + 1 , i + + β d max d d · N d max d , i + β c u t d · N a n n u a l   c u t , d , i + β w i n d f a l l d · N a n n u a l   w i n d f a l l , d , i + ε d . i ,
β j d are the regression coefficients corresponding to the model of diameter class d .
The results of the Recursive Feature Elimination procedure are presented in Table A1 in Appendix A (note that these results are based on the optimal features transformation and the optimal regressor presented in Section 3.1.2). The results show that the most important features for predicting the variation of the number of stems in a given diameter class d , i.e., Δ N d , are the initial number of stems N d and basal area G d in this class, as well as annual cuts N a n n u a l   c u t , d and to a lesser extent annual windfalls N a n n u a l   w i n d f a l l , d occurring in this class. The number of stems ( N r e s i n o u s / d e c i d u o u s , d ) and basal area ( G r e s i n o u s / d e c i d u o u s , d ) in the different classes ( 1 d 16 ) play a highly variable role that differs in part between models. Annual cuts in the overlying diameter classes N o v e r l y i n g   a n n u a l   c u t , d are of intermediate importance. The features G c u m , G t o t , N o v e r l y i n g   a n n u a l   w i n d f a l l , d , N t o t and forest sites play a less important role.
Following this order of importance, a Stepwise Regression with Backward Elimination is conducted by a group of features. The detailed results are presented in Table A2 in Appendix A (note that these results are based on the optimal features transformation and the optimal regressor presented in Section 3.1.2). In this procedure, the features corresponding to the forest sites are eliminated, then N t o t and   G t o t , then G c u m , d , then N o v e r l y i n g   a n n u a l   c u t , d and N o v e r l y i n g   a n n u a l   w i n d f a l l , d . Then, there is the problem of multicollinearity between the features N r e s i n o u s / d e c i d u o u s , d and G r e s i n o u s / d e c i d u o u s , d (due to the identity G = N · π · d 2 2 ) . Since the objective of the models is to predict variations in the number of stems, the G features are eliminated. Then, the N r e s i n o u s , d and N d e c i d u o u s , d features are added together, yielding the N d features (with 1 d 16 ). Finally, the N d features can be adjusted to take into account certain characteristics of forest growth, such as the fact that in the short/medium run (around 10 years), trees of a given diameter normally have little influence on the dynamics of much larger diameter classes (see a full explanation below in this section). The eliminations or adaptations mentioned thus far do not lead to a strong and uniform degradation of the test R2 of the different models compared to the models with all features (see Table A2). However, the elimination of N a n n u a l   c u t , d and N a n n u a l   w i n d f a l l , d leads to a noticeable and uniform degradation of the models. Thus, these latter features are kept, resulting in the models given in Equation (7) above.
As to the adjustment of features N d , when modeling the increment in the number of stems for a diameter class d ( Δ N d ), among underlying classes, only the two preceding ones ( d 1 and d 2 ) provide relevant information. Some stems in these classes may enter class d during the growth periods considered in the data, typically 7 to 12 years. Trees in classes lower than d 2 do not have the time to grow as much (at least 10 cm in diameter) over a period of 7 to 12 years. Moreover, trees in those lower classes normally exert no competition (or limited competition) on trees in class d . The corresponding variables can therefore be eliminated since they would only capture statistical noise (overfitting). This verifies empirically when fitting models on unadjusted features N d . In the same way, it is not always necessary to keep all the variables corresponding to numbers of stems in classes higher than a given class, i.e., d m a x is not necessarily equal to 16 in all cases (see Equation (7)). These d m a x values are adjusted for each model separately so as to optimize the model’s performances (through a trial and error approach). A detailed analysis of the value of the regression coefficients and the role of each feature in the linear models is presented in Section 3.1.4.

3.1.2. Features Transformation and Optimization of the Regression Procedure

Despite their relative simplicity, linear models can be significantly improved by transforming the endogenous and exogenous variables. The exogenous variables (features) are each subjected to a standard scaling (noted by an apostrophe in Equation (7) above). This usual transformation is obtained by removing the mean and scaling to unit variance. The endogenous variable is subjected to a Yeo-Johnson transformation [30]. This transformation is noted by a double apostrophe in Equation (7). This non-linear transformation can reduce the skewness of a distribution and solves the heteroscedasticity problem that was visible on the residuals plots. This transformation is similar to the Box-Cox transformation [31], which is better known, but it accepts negative values as an argument, which is necessary in our case. The Yeo-Johnson transformation improved the performance of the models of classes 1 to 12 (as a reminder, classes are 5 cm wide, and the center of a class in cm = (#class − 1) × 5 + 20). For classes 13 to 16, this transformation did not improve performances, and it even deteriorated them. The transformation was therefore applied only to classes 1 to 12. These two transformations (standard scaling of the exogenous variables and Yeo-Johnson transformation of the endogenous variable) were performed separately for each model. B-splines transformations [32,33] on the exogenous variables were also tested. This type of non-linear transformation, related to the generalized additive modeling (GAM) approach, allows for non-linearity effects between the respective exogenous variables and endogenous variable. The use of this transformation did not produce strong and unambiguous results (performances were slightly better for half of the diameter classes and slightly worse for half of the diameter classes with no particular order among these diameter classes). By principle of parsimony and to keep a common structure to our models, this transformation was left aside.
Finally, best performances were obtained using a Huber loss function [34,35]. This loss function is quadratic below a given error threshold but linear above this threshold, which can limit the weight of points that stand out (e.g., residual outliers). Moreover, a regularization term of type L2 is added to the loss function [36], which can control the overfitting in a systematic way. Note that this regularization term makes the models non-linear in the strict sense. A Stochastic Gradient Descent regressor [36] is used to train the model, and the hyperparameters related to the Huber loss function and the regularization term are tuned by gridsearch and cross-validation.

3.1.3. Performance of Linear Models

The performances obtained for the different linear models are presented in Table 2. As the train-test-split is random, it is repeated 10 times, and the table therefore presents averaged results. Table 2 gives the test R2s for the prediction of transformed annual increments of the number of stems per hectare Δ N p r e d i c t , d , i . The table also shows the test R2s for predicting gross (i.e., if neither cuts nor windfalls occur) stem counts at the end of the growth period, i.e., based on the errors between the N a f t e r   g r o s s , d , i and the N p r e d i c t   a f t e r   g r o s s , d , i [6]:
N a f t e r   g r o s s , d , i = N a f t e r , d , i + N c u t , d , i + N w i n d f a l l , d , i ,
N p r e d i c t   a f t e r   g r o s s , d , i = N d , i + Δ N p r e d i c t , d , i · Δ t .
The plots of the residuals (corresponding to one of the train-test-split) for the 16 models are presented in Figure A1 in Appendix B. A review of the histograms giving the distributions of these residuals (not shown here) shows Gaussian-type distributions centered in 0.
The results for the predictions of Δ N d , i are satisfying except for class 1 ( 17.5   cm d < 22.5   cm ). Since there are no data on the number of stems for diameter classes below 17.5 cm, it is impossible to precisely predict the number of trees that will enter class 1 during the growth period, i.e., the recruitment rate of our set of models, hence the limited performance. For the other models, test R2s range typically from 50% to 70%. These models manage to predict most of the variance of the Δ N d , i based on the initial state of a division N d , i (and on cuts N a n n u a l   c u t , d , i and windfalls N a n n u a l   w i n d f a l l , d , i ) for cases that were not used for training the model. Using these predicted increments (after inverse transformation) to predict stem counts at the end of the growth period also produces satisfying results, mechanically better than the prediction of increments since the initial stem counts are known and largely explain the stem counts 7 to 12 years later. All diameter classes combined, the models predict future stem numbers with a test R2 of 90.3%. The root mean squared error (RMSE) indicates the average difference between predicted values and real values in absolute terms. It is here computed on N a f t e r   g r o s s , d , i real and predicted values and can be compared to mean real values of N a f t e r   g r o s s , d , i . Table 2 shows that RMSE values are satisfyingly small compared to mean values, although performances decrease noticeably for diameter classes above 10 ( d 10 ) and are mediocre for diameter classes above 13 ( d 13 ). These figures can be compared to the result obtained with the artificial neural network approach (see below).
The models are then retrained on the whole data set to determine the final regression coefficients. Table 2 presented above shows corresponding R2s, which are in this case training R2 s. Those training R2s are close to the test R2s presented before, showing an absence of overfitting (or limited in the higher diameter classes), which is normal given that the linear models have relatively few features and have been trained on a fairly large data set (580 observations each).

3.1.4. Analysis of the Regression Coefficients

The regression coefficients corresponding to the training on the whole data set are presented in Table A3 in Appendix B and are shown in Figure 3 as a heatmap for ease of exposition and interpretation. Because of the transformations undergone by the variables (Yeo-Johnson and standard scaling, for each diameter class separately), the coefficients are difficult to interpret in absolute terms, but their sign can be interpreted, as can their relative values within a given model.
For diameter classes greater than 1 (diameters greater than 22.5 cm), a common pattern is visible. This pattern is similar to the results presented in [13], despite the differences due to our data transformations. This pattern is also clearly visible when displaying the coefficients for a given model on a graph, see for example, Figure 4 and Figure 5.
The effects of the respective features are comparable within the same model despite different numbers of stems depending on the class (high for small classes to low for large classes) because these features have been scaled to the unit variance. The increment in the number of stems per ha Δ N d of a given diameter class d is positively influenced by the number of stems per ha in the two directly underlying classes d 1 and d 2 . These are the trees whose growth will allow them to move from class d 1 to d or from class d 2 to d during the growth period.
The number of stems in class d , however, has a negative influence on the increment of class d . This effect corresponds to the trees that will leave class d and move to the overlying classes during the growth period. The effect of the overlying classes is negative but more difficult to interpret. In general, only the directly overlying classes seem to have a slight negative influence on the increment of the number of stems in a given class. It can be assumed that there is some competition between close diameter classes, provided that there is some spatial proximity between trees of these classes. This can be the case in small and intermediate classes. It is more uncommon for the largest trees whose neighborhood at their height is generally more sparse (those trees have already been fostered by previous cuts). Classes much larger than a given diameter class did not seem to play a clear and strong role on the increment in that class, either because of a lack of strong competition between trees of different diameters (e.g., due to spatial differentiation) or because this competition effect is relatively constant across divisions and that without variability, the models fail to quantify this effect. For this reason, the corresponding features were not included in the model. The threshold for not including these features in the model, d m a x in Equation (7), depends on a given diameter class, hence the notation d m a x d .
For class 1, the situation is different for the reasons already stated, which have to do with the lack of information on the underlying classes. In this case, the model performed better by keeping all features corresponding to the overlying classes. Figure 6 shows the value of the regression coefficients associated with features N d (see Equation (7)) for the diameter class 1 model.
The number of stems in class 1 has a negative effect on the increment in the same class. This effect is explained by the trees that will leave this diameter class during the growth period. The effect of the number of stems in the directly overlying diameter classes is also negative. This could be explained by the existing competition between the trees of these diameter classes and the trees of class 1. In contrast, the numbers of stems in the larger diameter classes do not appear to have a significant effect on the increment of class 1. The latter two findings could be explained by the hypotheses made above.
Figure 7 shows the regression coefficients associated with the feature N a n n u a l   c u t , d (see Equation (7)) for the different models.
Figure 7 shows the gross effect of cuts in a given diameter class on the increment in the number of stems in that same class, i.e., the indirect effect of cuts in a class on the growth of trees in that class but without the direct effect of tree removal. The indirect effect of cuts is positive for all diameter classes. Comparing diameter classes with each other is less obvious because increments (a fortiori transformed according to Yeo-Johnson) do not mean the same thing from one class to another; an increment in the number of stems will necessarily be lower in the larger diameter classes. However, it may be that intra-class competition is weaker in the larger diameter classes because the larger trees are more spatially separated from each other (they have been fostered by previous cuts), compared to the smaller classes. The pattern of regression coefficients associated with the feature N a n n u a l   w i n d f a l l , d is similar to that in Figure 7.

3.2. Results of the Modeling with the Multilayer Perceptron

The best topology had two hidden layers with 10 neurons each (among those tested: one or two hidden layers, 5 to 100 neurons by layer in steps of five). The optimal value of the regularization parameter (L2) optimally defined by the gridsearch is 100 in both cases, an intermediate value among the values tested.
Even though a unique model is trained for all diameter classes, its performances are disaggregated by diameter class in order to compare the results with those of the linear models. Results are presented in Table 3.
The analysis of results by diameter class shows that the multilayer perceptron results in lower and more heterogeneous performance than the linear models. The performance is close for some diameter classes (the lower diameter classes for which a large volume of data are available), and it is inferior in the higher diameter classes. For classes 14, 15, and 16, for which there are limited data, the multilayer perceptron is apparently totally unable to predict future stem numbers. In the case presented in Table 3, the model with two hidden layers of 10 neurons each is relatively complex (in comparison to linear models presented in Section 3.1), which gives good results in the lower classes but produces overfitting and thus bad results in the higher classes. This hypothesis is confirmed by comparing the training and test R2 at the level of the different diameter classes. These two scores are close for the low and intermediate diameter classes ( 1 d 9 ) but diverge for the higher diameter classes (and the higher the class). This reflects overfitting in the higher diameter classes. For the two highest classes (15 and 16), even the training R2s become negative, reflecting the fact that the MLP is no longer able to learn from the data even through overfitting. This is due to the scarcity of data in the higher classes. There are few trees this large in the data set and thus many null features for these diameter classes.
The performance of the multilayer perceptron seen on the whole data set is however better than that of the linear models. As the multilayer perceptron has been trained on the whole data set, its objective is to minimize the errors on the whole data set, which leads to this better global result.
The results of the permutation importance procedure are shown in Table 4.
The information extracted by the multilayer perceptron is related to the initial state of the diameter class (above all the number of stems N d but also the basal area G d in this class), the diameter d , an indicator of the competition regulation ( N o v e r l y i n g   a n n u a l   c u t , d ), an indicator of competition ( G c u m , d ), and information on the initial state of the 1st diameter class ( G r e s i n o u s , 1 , N r e s i n o u s , 1 , N d e c i d u o u s , 1 ).

3.3. Results of Linear Modeling at the Forest Division Scale

Linear models perform better for predictions at the diameter class level, which are important from a silvicultural point of view. In particular, linear models also produce good results for larger diameter classes. In addition, due to their simplicity and transparency, linear models are much easier to interpret in terms of forest growth, which facilitates the quality control. Linear models could be used as a decision support tool for the planning of cuts in forests under selective felling management.
The performance of these models is also convincing when tested in real situations. The linear models are applied on whole test divisions (completely unknown to the models) to predict the increment of the number of stems in each diameter class. These increments are then used to calculate the raw stem counts (i.e., without tree removal from cuts or windfalls) at the end of the growth period and can be compared to the actual values. Figure 8, Figure 9, Figure 10 and Figure 11 show the results for four couples (division, inventory year).
Figure 8, Figure 9 and Figure 10 show quite contrasting situations for which the models produce good results. The predictions for the first diameter class are rather approximate due to the lower performance of this model. However, results are particularly convincing for intermediate classes, which correspond to the best performing models. Figure 11 shows an example where the models produce poor results, which happens for a small minority of the divisions tested.
Linear models also perform well for predicting increments of some aggregate features at the division level, e.g., number of stems per ha N, basal area per hectare G, and standing volume V. Prediction tests are this time performed on all couples (division, inventory year) to keep the number of points large. This is not a problem, as the overfitting of linear models is extremely limited.
For each couple (division, inventory year), the linear models are used to predict the Δ N and Δ G at the end of the corresponding growth period. The predicted values are then compared to the actual values. For Δ N , an R2 of 52.3% is obtained, or a mean absolute error of 13.2 stems·ha−1. For Δ G , an R2 of 42.5% is obtained, or a mean absolute error of 1.37 m2·ha−1. Figure 12 and Figure 13 show the corresponding residuals plots.
In selective felling management practice, variables N , G and V are often disaggregated by large diameter classes; in the Canton of Neuchâtel, those classes are: small woods (17.5 to 32.5 cm), medium woods (32.5 to 52.5 cm) and large woods (>52.5 cm). Disaggregating the variables into these three classes provides concise information about the structure of a division. The predictive capacity of our models for Δ G in these three respective classes is tested. For small woods, we obtain an R2 of 66.2%, i.e., a mean absolute error of 0.46 m2·ha−1. For medium woods, we obtain an R2 of 78.9%, that is, a mean absolute error of 0.68 m2·ha−1. For large woods, we obtain an R2 of 41.9%, that is, a mean absolute error of 0.73 m2·ha−1. Figure 14, Figure 15 and Figure 16 show the corresponding residuals plots.

4. Discussion

4.1. Absolute and Comparative Performance of the Models Developed in This Study

The two modeling approaches followed in this study for modeling dynamics of the numbers of stems by diameter class both produce compelling results, which differ from each other. The linear models achieve better performance at the level of individual diameter classes. More specifically, performances of the two model types are close to each other for most diameter classes, but the performance of the linear models is much better for larger diameter classes.
For the linear models, test R2 coefficients typically range from 50% to 70% (excluding class 1). These good performances are made possible by training a specific model for each diameter class (a form of data panelization). The absence of overfitting in the case of linear models indicates that the (large) volume of data available in this study is not limiting for those models. However, the comparatively poor performance of the model for diameter class 1 shows that the lack of data on the underlying classes ( d 17.5   cm ) is naturally limiting for predicting the evolution of the number of stems in this class.
As to the methodology, the implementation of a machine learning workflow, facilitated by existing technologies (in this case the scikit-learn library supported by the Python language), could significantly improve the performance of linear models by iteratively improving the features selection and transformation, as well as the selection of a loss function and a regressor. Train-test splits also allowed for an unbiased evaluation of model performances in terms of predictive capabilities, which was rarely proposed in past inferential statistical studies.
The multilayer perceptron provides the best performance at the global level for predicting the future number of stems with an average test R2 of 96.1%. This is due to the fact that only one model was trained on the whole data set and not independently for each diameter class. The volume of data was considered insufficient to train a multilayer perceptron per diameter class, as there would only be 580 observations per class. The multilayer perceptron shows heterogeneous performance in the diameter classes. The performances are close to the linear models for the lower and intermediate diameter classes (between 17.5 and 62.5 cm), but they decrease quickly toward the large diameters. The analysis of the difference between test and training R2 shows an increasing overfitting for the diameter classes between 62.5 and 82.5 cm. Beyond 82.5 cm, the multilayer perceptron is unable to learn from the available data. It is therefore clear that the volume of data available for this study is limiting for modeling growth with a multilayer perceptron beyond 62.5 cm. From a methodological point of view, the implementation of the multilayer perceptron was much simpler than the fine-tuning of linear models.

4.2. Comparison of Performances with a Literature Benchmark

Both types of models developed in this study perform better than previously developed linear models for predicting dynamics at the diameter class level. For example, the model in [12] achieves an R2 (on the whole data set) of 40% or less for predicting the probabilities of transition of trees between size classes. Our approach, differentiated by diameter class and based on a larger number of features (notably the initial diameter distribution) produces better results. It is not possible to directly compare our results with the results of models predicting individual tree diameter increment, but we can still note that the orders of magnitude of the test R2 obtained are similar to ours, 57% for the best model in [4] and 53% for the best model in [9].
Aggregating the values predicted by our models for predicting increments of the number of stems per hectare, basal area per ha, or standing volume per ha at the division level produces relevant results for the practice of forest management. Nevertheless, the prediction performance is inferior to the performance of models specialized on this approach and based on machine learning methods. One model in [8] based on an artificial neural network achieves a test R2 of 76%, and another model in [7], also based on an artificial neural network, achieves a test R2 of 94% (versus 87% with linear regressions). This better performance is enabled by specialized models that directly predict the aggregate increment but probably also by the fact that these models are based on data from permanent plots.

4.3. Comparative Analysis of Feature Importance

The feature selection conducted in the case of linear models and the feature importance analysis conducted in the case of the multilayer perceptron result in partly different conclusions. It is interesting that the feature selection in the case of linear models has led without this being the initial goal to models based on the concept of passage rates from one diameter class to the next. Linear models based on passage rates are common in the literature [5,10,11,12,13,14], but they are usually specified as such initially. They are simple but general models and are consistent with forest growth theory. The mainly empirical feature selection conducted in this study shows that this is also a powerful and parsimonious type of model when the use of a linear model is imposed.
The analysis of feature importance in the case of the multilayer perceptron shows the high importance of a limited number of features. These results are consistent with the theory on forest growth and with the results of past studies. It is clear that the number of stems ( N d ) or basal area ( G d ) in a diameter class are good predictors of the number of stems in that same class in the near future. The role of diameter d is also intuitive and has been demonstrated many times [5,10]. Most interesting is the important role played by indicators of competition (cumulative basal area G c u m , d ) and of regulation of competition (cuts in overlying diameter classes N o v e r l y i n g   a n n u a l   c u t , d ) between diameter classes. The role of G c u m has already been demonstrated in previous studies [5,10] and is confirmed and quantified here in a flexible and general-value model. The role of cuts in the overlying diameter categories is clearly demonstrated here. The fact that the latter two variables ( G c u m , d and N o v e r l y i n g   a n n u a l   c u t , d ) were not equally important in the linear models shows that the corresponding effects are likely to be non-linear. However, the linear models perform well nonetheless by exploiting the information related to the initial structure at the division level, which also reflects the competitive conditions. Finally, the particular importance of the features related to the initial state of diameter class 1 ( G r e s i n o u s , 1 , N r e s i n o u s , 1 , N d e c i d u o u s , 1 ) is probably due to the difficulty of predicting recruitment in the absence of data on underlying classes, the initial state of this class 1 remaining the best predictor of its future state.
However, one should be careful not to draw too general conclusions about the importance of the features achieved in this study. For example, forest sites have a known influence on forest productivity but are not considered important in the models developed in this study. This does not imply that these parameters are negligible in general. The variability of forest sites is relatively small at the scale of the Neuchâtel forests managed according to selective felling. Furthermore, our models focus on short- to medium-term growth (about 10 years). On a larger spatial or temporal scale, forest sites have an important effect on forest productivity.

4.4. Assessment of the Use of Data from the Canton of Neuchâtel

Our results show that the inventory data from the Canton of Neuchâtel, collected through the Méthode du contrôle, generate valuable information to train convincing growth models at the diameter class level. The use of this type of data is an additional challenge compared to the use of data from permanent plots. Because increments are not tracked on the individual tree level, the calculation of increments is more approximate, for example, because of the lack of data on mortality. The lack of information on stems below 17.5 cm and of direct data on recruitment in the first diameter classes makes it difficult to model their dynamics. From this point of view, the performance of our models is limited. Another particularity of the data collected by the Canton of Neuchâtel compared to permanent plots is related to the inventory areas. The divisions used for this purpose are each about ten hectares in size instead of a few thousand square meters at most for the plots. Data at such a large scale are necessarily averaged, resulting in a probable reduction of variability in the data and reduced information for model training compared to the permanent plots.
The fact that the volume of data provided by the Canton of Neuchâtel was so large was clearly an added value for our study. The volume of data was clearly sufficient to train our linear models with a remarkable absence of overfitting. This large volume of data was also crucial for the training of the multilayer perceptron, which reached convincing performances. Nevertheless, the volume of data remained limiting for the latter, especially in the upper diameter classes.

5. Conclusions

The linear models developed in this study achieve sufficient performance for practical use. The predictions from these models are division specific in that the predicted dynamics depend on the initial diameter distribution and take into account competition between diameter classes to some extent. Linear models also have the advantage of being relatively transparent and therefore convincing for practitioners (coefficients can be directly interpreted by a professional knowledgeable in forest growth). Linear models, at least in this study, also predict fewer outliers than the multilayer perceptron.
Linear models can be used to predict the growth dynamics of a division over a period of 7 to 12 years. They can possibly be used recursively to predict dynamics over longer periods, but it is not recommended to simulate more than two or three growth periods, especially because the prediction of recruitment is limiting. The models can typically be used for better planning and implementation of the next cut in a given division. The model is not well suited to understanding the role of cuts on recruitment.
Before the model can be used in practice, it must first be tested under real conditions, for example on a current case study in the Canton of Neuchâtel.

Implications and Management Options for Policy

The (linear) models developed in this study could be used to assist in the planning of cuttings in uneven-aged stands. In their current state, they are applicable in stands managed according to the selection felling approach such as in the Canton of Neuchâtel, but after adaptation (i.e., re-estimation of the models on different sets of data), they could be applied to other uneven-aged management systems.
More specifically, the models could be used to:
  • help plan the couple (harvest intensity, periodicity of interventions), specifically to a given forest stand (i.e., taking into account the current state of the stand);
  • help in the silvicultural knowledge management, by serving as an analysis tool for training to selection felling (e.g., by simulating the potential effects of a cut on the dynamics of a stand);
  • help concretely with selection felling in the field (for the selection of trees to be cut).
The possibilities of using such a tool would also be increased dramatically if it were possible to determine the initial structure of a stand in terms of diameter distribution from remote sensing data (especially LiDAR data). Data on the initial state of a stand are necessary for the application of the models, and remote sensing data are often available at lower cost and potentially more regularly than inventory data.

Funding

This study was financed by the Swiss Federal Office for the Environment (contract 00.0059.PZ/5933E7220). The author alone is responsible for the content of this article; the Federal Office for the Environment cannot be held responsible.

Data Availability Statement

The data used in this study were provided as a courtesy by the Canton of Neuchâtel (Switzerland). Some of the data are publicly available on the canton’s geoshop (sitn.ne.ch/geoshop), free of charge when used for scientific purposes. Some data were not on the geoshop but were provided on motivated request and as a courtesy by the Service de la faune, des forêts et de la nature of the Canton of Neuchâtel.

Acknowledgments

I thank Christian Rosset from Bern University of Applied Sciences BFH for his feedback on the project design and acquisition.

Conflicts of Interest

The author declares no conflict of interest.

Appendix A

Table A1. Average ranking of features through Recursive Feature Elimination (one feature is eliminated at each iteration until complete elimination of all features); these results are based on the optimal features transformation and the optimal regressor presented in Section 3.1.2; each average is calculated over feature ranks of the respective models for the 16 different diameter classes and over 10 train-test splits for each model (i.e., 160 values in total).
Table A1. Average ranking of features through Recursive Feature Elimination (one feature is eliminated at each iteration until complete elimination of all features); these results are based on the optimal features transformation and the optimal regressor presented in Section 3.1.2; each average is calculated over feature ranks of the respective models for the 16 different diameter classes and over 10 train-test splits for each model (i.e., 160 values in total).
FeatureAverage Ranking (Recursive Feature Elimination) FeatureAverage Ranking (Recursive Feature Elimination)
N a n n u a l   c u t , d 8.2 N r e s i n o u s , 15 43.0
N d 8.9 N d e c i d u o u s , 4 43.4
G d 9.6 G d e c i d u o u s , 2 43.5
G r e s i n o u s , 6 28.9 G r e s i n o u s , 15 43.6
N r e s i n o u s , 6 28.9 G d e c i d u o u s , 10 44.0
N a n n u a l   w i n d f a l l , d 29.1 N r e s i n o u s , 14 44.1
G r e s i n o u s , 10 30.6 G t o t 44.8
N r e s i n o u s , 3 31.5 N d e c i d u o u s , 14 44.9
G r e s i n o u s , 5 32.3 N r e s i n o u s , 1 45.3
G r e s i n o u s , 3 32.4 N d e c i d u o u s , 9 46.0
N r e s i n o u s , 10 32.4 G d e c i d u o u s , 9 46.1
N r e s i n o u s , 5 32.5 N d e c i d u o u s , 10 46.2
G d e c i d u o u s , 8 32.9 G d e c i d u o u s , 14 46.6
N r e s i n o u s , 11 33.2 N d e c i d u o u s , 1 47.0
G r e s i n o u s , 11 33.4 N r e s i n o u s , 16 47.5
G r e s i n o u s , 7 33.5 G r e s i n o u s , 1 47.6
N d e c i d u o u s , 8 33.8 G d e c i d u o u s , 16 47.9
N o v e r l y i n g   a n n u a l   c u t , d 34.9 N o v e r l y i n g   a n n u a l   w i n d f a l l , d 48.0
G r e s i n o u s , 12 35.7Forest_site_1348.1
N r e s i n o u s , 7 35.8 G d e c i d u o u s , 1 48.4
Forest_site_836.0Forest_site_1248.7
N d e c i d u o u s , 3 36.0 G r e s i n o u s , 16 49.5
N r e s i n o u s , 4 36.3Forest_site_249.6
G r e s i n o u s , 4 36.4 N t o t 49.8
N r e s i n o u s , 12 38.5Forest_site_1450.0
G r e s i n o u s , 9 38.5 N d e c i d u o u s , 5 50.0
G r e s i n o u s , 8 38.6 N d e c i d u o u s , 11 50.1
N r e s i n o u s , 9 38.9 G d e c i d u o u s , 5 50.3
N r e s i n o u s , 8 39.0 G d e c i d u o u s , 11 50.8
G r e s i n o u s , 13 39.1Forest_site_2752.6
N r e s i n o u s , 13 39.3 G d e c i d u o u s , 15 53.1
G d e c i d u o u s , 3 39.6 N d e c i d u o u s , 16 53.3
N r e s i n o u s , 2 40.0 N d e c i d u o u s , 15 54.7
G d e c i d u o u s , 7 40.2Forest_site_1154.9
N d e c i d u o u s , 7 40.5Forest_site_955.7
G r e s i n o u s , 2 40.9 N d e c i d u o u s , 12 55.9
G r e s i n o u s , 14 41.0 G d e c i d u o u s , 12 57.6
N d e c i d u o u s , 2 41.3Forest_site_1757.8
N d e c i d u o u s , 6 41.6 N d e c i d u o u s , 13 58.3
G d e c i d u o u s , 6 42.0 G d e c i d u o u s , 13 59.4
G d e c i d u o u s , 4 42.6Forest_site_2360.3
G c u m . d 42.6
Table A2. Results of the Stepwise Regression with Backward Elimination; values are test R2 for the respective models corresponding to the 16 different diameter classes; these results are based on the optimal features transformation and the optimal regressor presented in Section 3.1.2; the columns correspond to different sets of features, starting with all features in the left column and progressively and cumulatively eliminating or modifying groups of features going to the right; the bold column corresponds to the set of features selected for the final models in this study; all values are averaged on 10 different train-test splits.
Table A2. Results of the Stepwise Regression with Backward Elimination; values are test R2 for the respective models corresponding to the 16 different diameter classes; these results are based on the optimal features transformation and the optimal regressor presented in Section 3.1.2; the columns correspond to different sets of features, starting with all features in the left column and progressively and cumulatively eliminating or modifying groups of features going to the right; the bold column corresponds to the set of features selected for the final models in this study; all values are averaged on 10 different train-test splits.
Diameter ClassAll FeaturesThenRemoving Forest SitesThen Removing N t o t / G t o t Then Removing G c u m Then Removing N o v e r l y i n g   a n n u a l   c u t and N o v e r l y i n g   a n n u a l   w i n d f a l l Then Removing G Features Then Merging of Variables N r e s i n o u s and N d e c i d u o u s Then Modifying of N FeaturesThen Removing N a n n u a l   c u t / N a n n u a l   w i n d f a l l
112.413.212.914.612.513.511.911.910.6
250.153.553.353.352.851.653.354.151.5
36163.963.763.662.762.363.564.161
465.967.868.768.769.367.36565.963.7
56568.468.868.769.26764.865.261.9
661.2636464.164.262.459.460.157.1
760.160.760.360.160.756.956.254.650.2
850.953.953.253.153.650.746.648.744.8
948.948.85047.649.347.84648.742.6
1054.754.754.354.955535051.946.8
1148.248.748.847.949.849.149.852.545
1250.751.251.850.150.65251.353.147
1356.956.557.85958.458.459.261.556.4
1455.152.55553.65656.95757.641.5
1551.349.150.150.851.853.254.956.446
1648.250.449.450.95152.853.949.928.5

Appendix B

Figure A1. Residuals plots for linear models differentiated by diameter class presented in Equation (7) (classes from 1 to 16 where class center in cm = (#class − 1) × 5 + 20).
Figure A1. Residuals plots for linear models differentiated by diameter class presented in Equation (7) (classes from 1 to 16 where class center in cm = (#class − 1) × 5 + 20).
Forests 13 01432 g0a1
Table A3. Regression coefficients for linear models trained on the whole data set (class center in cm = (#class − 1) × 5 + 20).
Table A3. Regression coefficients for linear models trained on the whole data set (class center in cm = (#class − 1) × 5 + 20).
ModelModel Class 1Model Class 2Model Class 3Model Class 4Model Class 5Model Class 6Model Class 7Model Class 8Model Class 9Model Class 10Model Class 11Model Class 12Model Class 13Model Class 14Model Class 15Model Class 16
Diameter range[17.5;22.5[[22.5;27.5[[27.5;32.5[[32.5;37.5[[37.5;42.5[[42.5;47.5[[47.5;52.5[[52.5;57.5[[57.5;62.5[[62.5;67.5[[67.5;72.5[[72.5;77.5[[77.5;82.5[[82.5;87.5[[87.5;92.5[[92.5;97.5[
Intercept0.315677830.438084980.473401180.460514550.522863540.420562520.352159250.257051170.202679670.151799390.104450360.071826930.05907850.02889670.015386930.00806861
N_class_1−0.088572240.952625860.304781680000000000000
N_class_2−0.19883118−0.704498740.444696220.26800507000000000000
N_class_3−0.2479029−0.31537605−0.479452680.333492090.1944185400000000000
N_class_4−0.09295262−0.12675351−0.22661129−0.360865380.385623110.218564270000000000
N_class_5−0.030212110−0.10632899−0.21420664−0.447156990.194828130.14910786000000000
N_class_6−0.0255145500−0.11447264−0.12177319−0.303913040.182571230.1081218100000000
N_class_7−0.09059655000−0.05773986−0.1078003−0.246066610.119243280.069070120000000
N_class_8−0.0480426800000.00040884−0.06378476−0.145912350.069920390.07125811000000
N_class_90.03877847000000−0.05121676−0.090017260.067306880.0414401900000
N_class_10−0.030404910000000−0.03252369−0.105021990.050302680.041664530000
N_class_110.00132175000000000−0.068078740.047879120.02745242000
N_class_12−0.003343770000000000−0.065328240.037420490.0187752800
N_class_130.012009670000000000−0.00304633−0.036855990.025489710.013386720
N_class_140.01124542000000000000−0.030287620.013271430.00454961
N_class_15−0.067034460000000000000−0.018338050.0064068
N_class_160.0264934400000000000000−0.00812474
N_annual_cut0.13786310.25311860.218028260.16557760.131142360.106570150.085379380.059795870.051139590.051602610.049967780.037516770.039172840.027880620.015343580.01277216
N_annuel_windfall0.099509840.041074770.032556230.022926890.009901270.030908720.026979620.01470160.009611940.00320460.010487880.005607380.007737450.007130120.006038130.00204978

References

  1. Bettinger, P.; Boston, K.; Siry, J.P.; Grebner, D.L. Estimation and projection of stand and forest conditions. In Forest Management and Planning, 2nd ed.; Academic Press: London, UK, 2017; pp. 87–111. [Google Scholar] [CrossRef]
  2. Susse, R.; Allegrini, C.; Bruciamacchie, M.; Burrus, R. Management of Irregular Forests: Developing the Full Potential of the Forest; Morgan, P., Translator; Azur Multimedia: Saint Maime, France, 2011. [Google Scholar]
  3. Peng, C. Growth and yield models for uneven-aged stands: Past, present and future. For. Ecol. Manag. 2000, 132, 259–279. [Google Scholar] [CrossRef]
  4. Ou, Q.; Lei, X.; Shen, C. Individual Tree Diameter Growth Models of Larch–Spruce–Fir Mixed Forests Based on Machine Learning Algorithms. Forests 2019, 10, 187. [Google Scholar] [CrossRef]
  5. Schütz, J.-P. Modelling the demographic sustainability of pure beech plenter forests in Eastern Germany. Ann. For. Sci. 2006, 63, 93–100. [Google Scholar] [CrossRef]
  6. Eskelson, B.; Barrett, T.; Temesgen, H. Imputing mean annual change to estimate current forest attributes. Silva Fenn. 2009, 43, 649–658. [Google Scholar] [CrossRef]
  7. Hamidi, S.K.; Zenner, E.K.; Bayat, M.; Fallah, A. Analysis of plot-level volume increment models developed from machine learning methods applied to an uneven-aged mixed forest. Ann. For. Sci. 2021, 78, 4. [Google Scholar] [CrossRef]
  8. Bayat, M.; Bettinger, P.; Hassani, M.; Heidari, S. Ten-year estimation of Oriental beech (Fagus orientalis Lipsky) volume increment in natural forests: A comparison of an artificial neural networks model, multiple linear regression and actual increment. For. Int. J. For. Res. 2021, 94, 598–609. [Google Scholar] [CrossRef]
  9. Hamidi, S.K.; Weiskittel, A.; Bayat, M.; Fallah, A. Development of individual tree growth and yield model across multiple contrasting species using nonparametric and parametric methods in the Hyrcanian forests of northern Iran. Forstwiss. Cent. 2021, 140, 421–434. [Google Scholar] [CrossRef]
  10. Schütz, J.P. Dynamique et conditions d’équilibre de peuplements jardinés sur les stations de la hêtraie a sapin. J. For. Suisse 1975, 126, 637–670. [Google Scholar]
  11. Lu, H.-C.; Buongiorno, J. Long- and short-term effects of alternative cutting regimes on economic returns and ecological diversity in mixed-species forests. For. Ecol. Manag. 1993, 58, 173–192. [Google Scholar] [CrossRef]
  12. Buongiorno, J.; Peyron, J.L.; Houllier, F.; Bruciamacchie, M. Growth and management of mixed-species, uneven-aged forests in the French Jura: Implications for economic returns and tree diversity. For. Sci. 1995, 41, 397–429. [Google Scholar] [CrossRef]
  13. Buongiorno, J. Quantifying the implications of transformation from even to uneven-aged forest stands. For. Ecol. Manag. 2001, 151, 121–132. [Google Scholar] [CrossRef]
  14. Rollin, F.; Buongiorno, J.; Zhou, M.; Peyron, J.L. Management of mixed-species, uneven-aged forests in the French Jura: From stochastic growth and price models to decision tables. For. Sci. 2005, 51, 64–75. [Google Scholar] [CrossRef]
  15. Pardé, J. La méthode du Contrôle, d’hier à aujourd’hui. Rev. For. Fr. 1991, 43, 185–202. [Google Scholar] [CrossRef]
  16. Favre, L.-A.; Oberson, J.-M. 111 années d’application de la méthode du contrôle à la forêt de Couvet | 111 years of application of the method of control in the community forest of Couvet. Schweiz. Z. Forstwes. 2002, 153, 298–313. [Google Scholar] [CrossRef]
  17. Beck, H.E.; Zimmermann, N.E.; McVicar, T.R.; Vergopolan, N.; Berg, A.; Wood, E.F. Present and future Köppen-Geiger climate classification maps at 1-km resolution. Sci. Data 2018, 5, 180214. [Google Scholar] [CrossRef] [PubMed]
  18. Allgaier Leuch, B.; Streit, K.; Stillhard, J.; Cioldi, F.; Brang, P. Zukunft der Fichte im Schweizer Mittelland. Wald. Holz 2018, 98, 31–34. [Google Scholar]
  19. Ellenberg, H.; Klötzli, F. Waldgesellschaften und Waldstandorte der Schweiz. Mitteilungen/Schweizerische Anstalt für das Forstliche Versuchswesen; Beer: Zürich, Switzerland, 1972; Volume 48/4. [Google Scholar]
  20. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  21. McKinney, W. Data structures for statistical computing in python. In Proceedings of the 9th Python in Science Conference, Austin, TX, USA, 28 June–3 July 2010; Volume 445, pp. 51–56. [Google Scholar]
  22. Harris, C.R.; Millman, K.J.; Van Der Walt, S.J.; Gommers, R.; Virtanen, P.; Cournapeau, D.; Wieser, E.; Taylor, J.; Berg, S.; Smith, N.J.; et al. Array programming with NumPy. Nature 2020, 585, 357–362. [Google Scholar] [CrossRef]
  23. Guyon, I.; Weston, J.; Barnhill, S.; Vapnik, V. Gene Selection for Cancer Classification using Support Vector Machines. Mach. Learn. 2002, 46, 389–422. [Google Scholar] [CrossRef]
  24. Scikit-Learn. Feature Selection (Recursive Feature Elimination)—Scikit-Learn 1.1.2 Documentation. 2022. Available online: https://scikit-learn.org/stable/modules/feature_selection.html (accessed on 28 August 2022).
  25. Hocking, R.R. A Biometrics Invited Paper. The Analysis and Selection of Variables in Linear Regression. Biometrics 1976, 32, 1–49. [Google Scholar] [CrossRef]
  26. Haykin, S. Neural Networks: A Comprehensive Foundation, 2nd ed.; Prentice Hall: Upper Saddle River, NJ, USA, 2004. [Google Scholar]
  27. Scikit-Learn. Neural Network Models (Supervised)—Scikit-Learn 1.0.2 Documentation. 2021. Available online: https://scikit-learn.org/stable/modules/neural_networks_supervised.html#regression (accessed on 10 March 2022).
  28. Hornik, K.; Stinchcombe, M.; White, H. Multilayer feedforward networks are universal approximators. Neural Netw. 1989, 2, 359–366. [Google Scholar] [CrossRef]
  29. Scikit-Learn. Permutation Feature Importance—Scikit-Learn 1.1.2 Documentation. 2022. Available online: https://scikit-learn.org/stable/modules/permutation_importance.html (accessed on 28 August 2022).
  30. Yeo, I.K.; Johnson, R.A. A new family of power transformations to improve normality or symmetry. Biometrika 2000, 87, 954–959. [Google Scholar] [CrossRef]
  31. Box, G.E.P.; Cox, D.R. An Analysis of Transformations. J. R. Stat. Soc. Ser. B Methodol. 1964, 26, 211–243. [Google Scholar] [CrossRef]
  32. Eilers, P.H.C.; Marx, B.D. Flexible smoothing with B-splines and penalties. Stat. Sci. 1996, 11, 89–121. [Google Scholar] [CrossRef]
  33. Scikit-Learn. Spline Transformer—Scikit-Learn 1.0.2 Documentation. 2021. Available online: https://scikit-learn.org/stable/modules/generated/sklearn.preprocessing.SplineTransformer.html#sklearn.preprocessing.Splin%20Transformer (accessed on 10 March 2022).
  34. Huber, P.J. Robust Estimation of a Location Parameter. Ann. Stat. 1964, 53, 73–101. [Google Scholar] [CrossRef]
  35. Scikit-Learn. Huber Regressor—Scikit-Learn 1.0.2 Documentation. 2021. Available online: https://scikit-learn.org/stable/modules/generated/sklearn.linear_model.HuberRegressor.html (accessed on 10 March 2022).
  36. Scikit-Learn. SGD Regressor—Scikit-Learn 1.0.2 Documentation. 2021. Available online: https://scikit-learn.org/stable/modules/generated/sklearn.linear_model.SGDRegressor.html (accessed on 10 March 2022).
Figure 1. Example of a distribution of diameters in a forest division under selective felling management (division number 2202, inventory of 1971).
Figure 1. Example of a distribution of diameters in a forest division under selective felling management (division number 2202, inventory of 1971).
Forests 13 01432 g001
Figure 2. Histogram of the time gap between two successive inventories in a given division.
Figure 2. Histogram of the time gap between two successive inventories in a given division.
Forests 13 01432 g002
Figure 3. Heatmap of the regression coefficients for the models given by Equation (7) (the values are given in Table A3 in Appendix B).
Figure 3. Heatmap of the regression coefficients for the models given by Equation (7) (the values are given in Table A3 in Appendix B).
Forests 13 01432 g003
Figure 4. Regression coefficients for the model of diameter class 5, associated with transformed features N d .
Figure 4. Regression coefficients for the model of diameter class 5, associated with transformed features N d .
Forests 13 01432 g004
Figure 5. Regression coefficients for the model of diameter class 8, associated with transformed features N d .
Figure 5. Regression coefficients for the model of diameter class 8, associated with transformed features N d .
Forests 13 01432 g005
Figure 6. Regression coefficients for the model of diameter class 1, associated with transformed features N d .
Figure 6. Regression coefficients for the model of diameter class 1, associated with transformed features N d .
Forests 13 01432 g006
Figure 7. Regression coefficients corresponding to the transformed feature N a n n u a l   c u t , d .
Figure 7. Regression coefficients corresponding to the transformed feature N a n n u a l   c u t , d .
Forests 13 01432 g007
Figure 8. Overall prediction quality control for the division 1974 inventoried in the year 2003.
Figure 8. Overall prediction quality control for the division 1974 inventoried in the year 2003.
Forests 13 01432 g008
Figure 9. Overall prediction quality control for the division 2240 inventoried in the year 1947.
Figure 9. Overall prediction quality control for the division 2240 inventoried in the year 1947.
Forests 13 01432 g009
Figure 10. Overall prediction quality control for the division 809 inventoried in the year 1989.
Figure 10. Overall prediction quality control for the division 809 inventoried in the year 1989.
Forests 13 01432 g010
Figure 11. Overall prediction quality control for the division 1878 inventoried in the year 1983.
Figure 11. Overall prediction quality control for the division 1878 inventoried in the year 1983.
Forests 13 01432 g011
Figure 12. Residuals plot for Δ N predictions at the level of couples (division, inventory year).
Figure 12. Residuals plot for Δ N predictions at the level of couples (division, inventory year).
Forests 13 01432 g012
Figure 13. Residuals plot for Δ G predictions at the level of couples (division, inventory year).
Figure 13. Residuals plot for Δ G predictions at the level of couples (division, inventory year).
Forests 13 01432 g013
Figure 14. Residuals plot for Δ G predictions at the level of couples (division, inventory year) for the small woods class.
Figure 14. Residuals plot for Δ G predictions at the level of couples (division, inventory year) for the small woods class.
Forests 13 01432 g014
Figure 15. Residuals plot for Δ G predictions at the level of couples (division, inventory year) for the medium woods class.
Figure 15. Residuals plot for Δ G predictions at the level of couples (division, inventory year) for the medium woods class.
Forests 13 01432 g015
Figure 16. Residuals plot for Δ G predictions at the level of couples (division, inventory year) for the large woods class.
Figure 16. Residuals plot for Δ G predictions at the level of couples (division, inventory year) for the large woods class.
Forests 13 01432 g015
Table 1. Contingency table of couples (division, inventory year) according to the 1st and 2nd majority species (for simplicity of representation, the 2nd majority species that had less than 10 couples for all 1st majority species combined are not reported in this table).
Table 1. Contingency table of couples (division, inventory year) according to the 1st and 2nd majority species (for simplicity of representation, the 2nd majority species that had less than 10 couples for all 1st majority species combined are not reported in this table).
2nd Majority Species
SpruceFirBeechOak (Sessile)Scots PineMaple (Sycamore)AshOther DeciduousAspenOther Resinous
1st majority speciesFir14120544553111150
Spruce010534441568572380
Beech201317014319255202
Oak (sessile)2141240500008
Scots pine516187000200
Larch1000000000
Other resinous10149000400
Other deciduous31200300000
Black pine3003200000
Poplar0000102000
Ash4001000000
Table 2. Performances of the models measured by R2 and RMSE scores for respective prediction of Δ N d , i and N a f t e r   g r o s s , d , i (the overall R2 is measured on the whole data set when the models are trained on the whole data set).
Table 2. Performances of the models measured by R2 and RMSE scores for respective prediction of Δ N d , i and N a f t e r   g r o s s , d , i (the overall R2 is measured on the whole data set when the models are trained on the whole data set).
Diameter ClassDiameter RangeTest R2 for the Prediction of Δ N d , i Test R2 for the Prediction of N a f t e r   g r o s s , d , i Overall R2 for the Prediction of N a f t e r   g r o s s , d , i Test RMSE for the Prediction of N a f t e r   g r o s s , d , i Mean of Real N a f t e r   g r o s s , d , i Values
1[13.6%82.4%82.0%15.3 stems/ha66.2 stems/ha
2[22.5;27.5[55.293.593.27.149.0
3[27.5;32.5[64.593.493.55.740.2
4[32.5;37.5[71.692.992.44.432.2
5[37.5;42.5[66.991.391.73.525.8
6[42.5;47.5[60.589.291.42.919.8
7[47.5;52.5[54.790.392.02.414.9
8[52.5;57.5[49.088.790.82.110.4
9[57.5;62.5[51.290.690.91.67.3
10[62.5;67.5[56.190.490.91.44.7
11[67.5;72.5[58.189.689.71.02.7
12[72.5;77.5[61.588.088.00.71.6
13[77.5;82.5[60.684.285.30.50.8
14[82.5;87.5[50.574.579.70.40.4
15[87.5;92.5[53.869.872.90.20.2
16[92.5;97.5[57.365.970.80.10.1
All classes combined 90.3%96.7%
Table 3. Comparison of performances for predicting N a f t e r   g r o s s , d , i between the linear models and the multilayer perceptron (MLP).
Table 3. Comparison of performances for predicting N a f t e r   g r o s s , d , i between the linear models and the multilayer perceptron (MLP).
Diameter ClassDiameter RangeTest R2 Linear Models (Reminder)Test R2 MLPTrain R2 MLPTest RMSE MLPMean of Real N a f t e r   g r o s s , d , i Values
1[82.4%81.4%83.0%14.3 stems/ha74.9 stems/ha
2[22.5;27.5[93.589.791.58.359.1
3[27.5;32.5[93.490.491.56.449.1
4[32.5;37.5[92.985.888.25.639.1
5[37.5;42.5[91.382.786.24.831.3
6[42.5;47.5[89.281.184.14.024.0
7[47.5;52.5[90.380.383.23.618.4
8[52.5;57.5[88.778.881.23.113.3
9[57.5;62.5[90.679.382.32.59.4
10[62.5;67.5[90.475.580.52.16.1
11[67.5;72.5[89.661.474.61.93.7
12[72.5;77.5[88.047.066.71.42.1
13[77.5;82.5[84.216.155.11.21.3
14[82.5;87.5[74.5−85.520.31.00.6
15[87.5;92.5[69.8−298.8−87.50.80.3
16[92.5;97.5[65.9−1133.7−389.30.80.1
All classes combined 90.3%96.1%96.5%
Table 4. Results of a permutation importance procedure on the multilayer perceptron (with 30 repetitions). Only main features for which the loss in test R2 is above 0.001 are mentioned in the table.
Table 4. Results of a permutation importance procedure on the multilayer perceptron (with 30 repetitions). Only main features for which the loss in test R2 is above 0.001 are mentioned in the table.
FeatureTest R2 Loss (Permutation Importance)
N d 0.6522
d (diameter class)0.2012
G d 0.046
N o v e r l y i n g   a n n u a l   c u t , d 0.0085
G c u m , d 0.0026
G r e s i n o u s , 1 0.0025
N r e s i n o u s , 1 0.0021
N d e c i d u o u s , 1 0.0011
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Dumollard, G. Exploring the Potential of Machine Learning for Modeling Growth Dynamics in an Uneven-Aged Forest at the Level of Diameter Classes: A Comparative Analysis of Two Modeling Approaches. Forests 2022, 13, 1432. https://doi.org/10.3390/f13091432

AMA Style

Dumollard G. Exploring the Potential of Machine Learning for Modeling Growth Dynamics in an Uneven-Aged Forest at the Level of Diameter Classes: A Comparative Analysis of Two Modeling Approaches. Forests. 2022; 13(9):1432. https://doi.org/10.3390/f13091432

Chicago/Turabian Style

Dumollard, Gaspard. 2022. "Exploring the Potential of Machine Learning for Modeling Growth Dynamics in an Uneven-Aged Forest at the Level of Diameter Classes: A Comparative Analysis of Two Modeling Approaches" Forests 13, no. 9: 1432. https://doi.org/10.3390/f13091432

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop