Next Article in Journal
What Children with Neuromotor Disabilities Need to Play with Technological Games
Previous Article in Journal
FONDUE: A Framework for Node Disambiguation and Deduplication Using Network Embeddings
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Tree-Structured Regression Model Using a Projection Pursuit Approach

Department of Statistics, Ewha Womans University, Seoul 03760, Korea
*
Author to whom correspondence should be addressed.
Appl. Sci. 2021, 11(21), 9885; https://doi.org/10.3390/app11219885
Submission received: 2 October 2021 / Revised: 18 October 2021 / Accepted: 20 October 2021 / Published: 22 October 2021

Abstract

:
In this paper, a new tree-structured regression model—the projection pursuit regression tree—is proposed. It combines the projection pursuit classification tree with the projection pursuit regression. The main advantage of the projection pursuit regression tree is exploring the independent variable space in each range of the dependent variable. Additionally, it retains the main properties of the projection pursuit classification tree. The projection pursuit regression tree provides several methods of assigning values to the final node, which enhances predictability. It shows better performance than CART in most cases and sometimes beats random forest with a single tree. This development makes it possible to find a better explainable model with reasonable predictability.

1. Introduction

Classification and regression trees are commonly applied to data, mainly for interpretability. The first regression tree was AID (automatic interaction detection [1]). At each node, AID determines the splitting rules using the sum of squared deviations as an impurity measure. It splits the data recursively and stops when the impurity measure is less than the predefined threshold value. It uses the sample mean of data in the final node as the predicted value. There are several problems with AID: overfitting, selection bias when the variables are highly correlated, and the masking problem. To overcome the overfitting problem, CART (the classification and regression tree [2]) was developed. It uses a pruning method to reduce the size of the tree to prevent overfitting. However, this increases the computational cost.
AID and CART use piecewise constant regression approaches. It is difficult to solve the masking problem with these approaches because the predictors can be used in two ways: for splitting and for fitting the model in the final node [3]. To overcome this problem, CTREE (the conditional inference tree [4]) uses a permutation test, and GUIDE (generalized, unbiased, interaction detection and estimation [5]) uses bootstrap calibration.
For decades, various algorithms have been proposed to apply complex models instead of constants values to the final nodes of the regression tree. M5 [6] and M5P [7] allow the construction of linear models at final nodes. HTL [8] allows for a nonlinear regressor at final nodes. One advantage of predicting using such complex models at final nodes over the traditional constant values is that they are generally much simpler and more accurate. Instead of making a single prediction from a complex model, ensemble models use summarize the predictions from many simple models. Tree-based ensemble models have been developed, e.g., bagging [9], boosting [10], random forests [11], BART (Bayesian Additive Regression Model [12]), and KTBoost [13]. In general, the ensemble technique is known to reduce overfitting. Accordingly, the performance of the model can be expected to be improved. Furthermore, there have been many efforts to apply tree algorithms to various regression problems like logistic regression and quantile regression [14]. Regression tree algorithms have recently extebd to the autoregressive tree model for time-series analysis [15] and the random effect tree model for panel data analysis [16]. The flexibility of fitting various types of data well is a powerful feature of regression trees. Therefore, regression tree algorithms have been developed and widely used in regression problems [17,18,19,20].
Another regression approach is projection pursuit regression (PPR [21]). This approach uses a sum of empirically determined univariate functions of the projected predictor values. To find the regression model, PPR finds the projection of the predictors, fits an empirical model with the projected predictor values and calculates the residuals. This procedure is repeated with the residuals until the residual sum of squares is very small. The final projection pursuit regression model is the sum of the empirical models in each iteration. Even though projection pursuit regression predicts the response variable quite well, it is not easy to understand the model itself, and it is quite difficult to explain patterns in the data. The projection pursuit idea is applied to improve the interpretability of the model, and a new tree with a piecewise regression approach—the projection pursuit regression tree—is proposed.
With the projection pursuit regression tree, the data analyst can explore the partitioned data as well as the data as a whole. The main difference between the general tree-based regression methods and the projection pursuit regression tree is the way they partition data. The general tree-based regression methods are focused on the space of independent variables and each final node tend to have wider-ranging values of the dependent variable. However, the projection pursuit regression tree starts from the dependent variable. This tree divides the range of the dependent variable into the number of final nodes and assigns groups in each interval. After fitting the tree, each group is assigned to one final node, and all the final nodes are ordered by values of the dependent variable. That is, the left most final node has the group with the smallest value of the dependent variable and the rightmost final node has the group with the largest value of dependent variable. Therefore, with the projection pursuit regression tree, the user can easily explore the features of the independent variables in each range of the dependent variable. To explore a more fine-grained range of the dependent variable, the user can increase the depth of the projection pursuit regression tree.
In Section 2, the general tree-based regression methods and the projection pursuit approach are reviewed. The main algorithm of the projection pursuit regression tree is in the Section 3. Our new method, the projection pursuit regression tree, is explored with wine data, and the performance of our method is compared to linear regression, CART, and the random forests in the result section. A discussion follows.

2. Regression Tree and Projection Pursuit

2.1. General Algorithm of the Regression Tree

Let Y be the dependent continuous variable and X 1 , , X p be the independent variables. For the whole procedure of the regression tree, several decisions should be made: the best independent variable X k * among X 1 , , X p and the best cut-off c * in each node, the stopping rule to declare a final node, the pruning method to use to avoid overfitting, and the assignment of numerical values to the final nodes.
In each node, the regression tree uses one independent variable (for example, X k * ) and separates the data in the specific node into two groups using the cut-off c * . If  X k * < c * , it is assigned to the left child node. Otherwise, it is assigned to the right child node. This cutoff is determined by a predetermined rule, usually to minimize the mean squared error (MSE) for the regression tree including CART [2]. This procedure is continued until the node is declared a final node.
The stopping rule is closely connected to the model complexity. If the number of final nodes is increased, there is a high chance that it will become a complex model, and the risk of overfitting will increase. Most tree-structured methods adapt a pruning method after growing a large tree. CART uses the cost-complexity pruning method with a penalty for model complexity.
The most common values to assign to the final node are the mean of the Y values in the final node. Because of the restriction that the one value should be assigned to all observations in the same final node, the tree-structured regression methods usually have poor predictability. To improve the precision of tree-structured regression methods, several methods have suggested using linear combinations of the independent variables instead of one value [8]. In this paper, the projection pursuit method and several different models for the assignment of the final node are adapted.

2.2. Projection Pursuit

Projection pursuit is a method of finding interesting low-dimensional projections of high-dimensional data by optimizing a predetermined criterion function, called a projection pursuit index. This idea originated in [22], and the authors of [23] coined the term “projection pursuit” as a technique for exploring multivariate data. It is useful in an initial data analysis. The method can also be used to reduce multivariate data to a low-dimensional but “interesting” subspace.
The projection pursuit classification tree [24] is a classification tree that uses projection pursuit indices to separate classes. The usual tree-structured classification finds a rule to separate the data into two groups using impurity measures that determine the degree of purity in the two groups in terms of classes. A projection pursuit classification tree, on the other hand, finds a rule to separate classes into two groups. This rule uses the best projection to separate the two groups of classes with various projection pursuit indices for separating classes. One class is assigned to only one final node, and the maximum depth of the projection pursuit classification tree is at most the number of classes. Therefore, the projection pursuit classification tree constructs a simple but more understandable tree for classification. The projection coefficients of each node represent the importance of the variables in separating the classes in each node. The behaviors of the projection coefficients in each node are useful in exploring how to separate classes. This approach is extended to the regression tree.

3. Projection Pursuit Regression Tree

3.1. Construction of the Projection Pursuit Regression Tree

Multiple regression models are widely used, but it is difficult to interpret the final model, especially when there are many independent variables. Additionally, if there are nonlinear patterns between the dependent variable and independent variables, the multiple regression model is more difficult to interpret. Most regression trees have better interpretability than the ordinary multiple regression model. With recursive partitioning of the independent variable space, the regression tree discovers interesting features in each partitioned space of the independent variables. However, one of the main purposes of regression is to explain the dependent variable using the independent variables. In this paper, the dependent variable is focused and a new regression tree—the projection pursuit regression tree—is proposed.
The original idea of the projection pursuit regression tree comes from the projection pursuit classification tree [24]. In the projection pursuit regression tree, a modified approach to split nodes and assign values in the final node is used.
Let ( Y 1 * , X 1 * ) , , ( Y n * , X n * ) be the set of observations in a specific node and X i * be a p-dimensional vector. Then, in each node,
step 1:
Divide Y 1 * , , Y n * into two groups: group 1 with small values and group 2 with large values (eg. Assign Y i * to group 1 if Y i < median ( Y 1 * , , Y n * ) . Otherwise, assign Y i * to group 2.)
step 2:
Find the optimal 1-dimensional projection α * to separate group 1 and group 2 using one of the projection pursuit indices that involves class information—the LDA, PDA, or Lr index.
step 3:
Project the data in the node onto the optimal projection α * , i.e.,  α * T X * .
If a v e ( α * T X i * | Y i * group 1 ) a v e ( α * T X i * | Y i * group 2 ) , use α * = α * as the optimal projection.
step 4:
Find the best cut-off, c, to separate group 1 and group 2.
step 5:
Group 1 is assigned to the left node, and group 2 is assigned to the right node
In step 2, the projection pursuit index is needed to separate group 1 and group 2. The LDA index, Lr index [25], and PDA index [26] were developed to find the projection from a separating view of groups. The LDA index is based on linear discriminant analysis and is useful for finding the view that maximizes between-group variation relative to within-group variation. If the correlations among variables are high or the number of variables is relatively large compared with the number of observations, the PDA index is useful to escape the data piling problem [27] in the LDA index. The amount of penalty on correlations can be controlled with the tuning parameter λ . The PDA index with λ = 0 is the same as the LDA index. Both the LDA and PDA indexes use all information in the variance-covariance matrix of the independent variables. On the other hand, the Lr index ignores the covariances among variables and uses only the variances of the independent variables ( L r i n d e x ( A ) = l = 1 q i = 1 g j = 1 n i ( y ¯ i . l y ¯ . . l ) r l = 1 q i = 1 g j = 1 n i ( y i j l y ¯ i . l ) r 1 / r , where y i j l is the lth variable of the projected data in ith group jth observation and A is the p × q projection matrix). r determines how to calculate the distance between two points in each dimension. With different selections of indices and the parameters associated with indices, the interesting features of our data space can be found.
The projection pursuit regression tree is designed to provide an exploratory data analysis tool, and the tree structure is constructed to retain this feature. In each node, the observations are separated into a group with small Y values and a group with large Y values; the group with small Y values is assigned to the left node, and the group with large Y values is assigned to the right node. To retain this directional convention, the direction of the optimal projection in step 3 is modified.
To explain how the projection pursuit regression tree works, a toy example is used. One-hundred observations with two independent variables, X 1 and X 2 , and one dependent variable, Y, are randomly chosen. In this toy example data, X 1 and X 2 are highly correlated with Y (Figure 1a,b). In step 1, the data are divided into two groups: group 1 (●) and group 2 (▲). In step 2, the LDA index is used to find the optimal projection to separate group 1 and group 2. Figure 1c is the scatter plot of the projected values of X 1 and X 2 vs. Y.
Figure 2 shows the difference between CART and the projection pursuit regression tree in dividing the space of independent variables. The dashed vertical line in the scatter plot of X 1 and X 2 represents the cut-off point in each method. The observations on the left/lower side of this line are assigned to the left node, and those on the right/upper side are assigned to the right node. CART uses the variable X 1 to separate the two groups. Because CART uses one variable at a time, the separating line is parallel to the axes of the variables. In contrast, the projection pursuit regression tree uses a linear combination of the independent variables a 1 X 1 + a 2 X 2 , and this separating line can be in any direction, including the direction parallel to axes.
These procedures continue until a node is declared as a final node. To decide whether a node is final, two different ways are provided: one is for exploring data, and the other is for predicting. For the exploratory data analysis, a tree with a predefined depth can be constructed. With this approach, the projection pursuit regression tree divides the Y values into 2 d e p t h groups and explores the properties of each group. Figure 3 is a simple projection pursuit regression tree for the toy example with depth 1. The data with small Y values have small X 1 and X 2 values, and the data with large Y values have large X 1 and X 2 values. For the prediction purpose, the final node is declared if the number of observations in a node is small or if the between-groups sum of squares is relatively smaller than the within-groups sum of squares. The second criterion is similar to the classical Fisher’s LDA idea.
After constructing the tree structure with the observed data, the numerical values are assigned to the final nodes. Five different models are provided in our projection pursuit regression tree to assign numerical values in the final nodes. Let Y i * and X i * = X i * 1 , , X i * p T be the i * th observations in a specific final node, i * = 1 , , n * , where n * is the number of observations in the node and α * is the best projection of the previous node. Then, for a new observation X * = X 1 * , , X p * T ,
Model 1:
Y ^ * = 1 n * i * = 1 n * Y i *
Model 2:
Y ^ * = median( Y 1 , , Y n * )
Model 3:
Y ^ * = β ^ 0 + β ^ 1 · p r o j X *
where β ^ 0 and β ^ 1 are estimated from the simple linear regression model with new variable p r o j X * = α * T X * that is generated from the optimal projection α * .
Model 4:
Y ^ * = β ^ 0 + β ^ 1 X 1 * + + β ^ p X p *
where β ^ 0 , , β ^ p are estimated from the multiple linear regression model with all independent variables.
Model 5:
Y ^ * = β ^ 0 + β ^ 1 X 1 * * + + β ^ p * X p * *
where β ^ 0 , , β ^ p * are estimated from the multiple linear regression model with the selected p * independent variables.
In Model 5, p * is the predetermined number and the independent variables are selected using the correlations with Y. Figure 4 shows how CART and the projection pursuit regression tree with Models 1–5 assign values to the final node. For Figure 4, the toy example data is fitted to a projection pursuit regression tree with depth 1 (Figure 3). The result of CART has 7 final nodes, and each node is assigned to one value. In Models 1 and 2, the mean and median of the small Y group are assigned to the left node and the mean and median of the large Y group are assigned to the right node. For Model 3, Y ^ * = β ^ 0 + β ^ 1 · p r o j X * is used to predict the Y values, where p r o j X * is the projected data of the first node. The Y ^ * = β ^ 0 + β ^ 1 X 1 + β ^ 2 X 2 model is used for Model 4, and Y ^ * = β ^ 0 + β ^ 1 X 1 is used for Model 5 ( p * = 1 ).
The MSEs of the fitted values with multiple linear regression, CART, the random forest for regression [11], and the projection pursuit regression tree with various models for the final node are summarized in Table 1. For the comparison, the lm, randomForest, and rpart functions in R are used. In this toy example, lm and the projection pursuit regression tree with Model 4 show the best performance. randomForest shows poor performance on these data. The MSE of randomForest is much higher than that of the lm method. The result of rpart is worse than that of randomForest. In the projection pursuit regression tree, the performances of various models are quite different. The MSEs of the Model 1 and 2 are much higher than that of rpart. This result is mainly due to the simple tree with depth 1 (the depth of the rpart model is 3). If a more complex tree with a greater depth is used, the performance of Models 1 and 2 can be improved. Model 5 shows similar performance to rpart, and Model 3 shows better performance than randomForest. The performance of Model 4 is similar to that of lm. From this result, it is confirmed that the predictability with various models of assigning final values can be improved.

3.2. Features of the Projection Pursuit Regression Tree

The projection pursuit regression tree is developed to explore data in each partitioned data space instead of the whole data space by focusing on the value of the dependent variable (Y). This tree always divides the data by Y values; the observations with smaller Y values are assigned to the left node and the observations with larger Y values are assigned to the right node. After fitting the projection pursuit regression tree, all final nodes are sorted by their estimated values—the node that is furthest to the left has the smallest estimated value, and the rightmost node has the largest estimated value. Therefore, with this tree structure, the most important variables for large or small Y values can be determined.
To examine this feature of the projection pursuit regression tree more deeply, 100 bservations with one dependent variable and four independent variables are simulated. Let Q1, Q2, and Q3 be the 25%, 50%, and 75% quartiles of Y. The Y values are divided into four groups—G1, G2, G3 and G4—with these quartiles. X 1 is linearly associated with Y in G1 and does not have any relation with Y in the other groups. Similarly, X 2 shows a linear relationship with Y only in G2, X3 shows an association with Y only in G3, and X4 shows an association with Y only in G4. The structures of the data are presented in Figure 5. The relationships between Y and the Xs depend on the range of Y values.
Figure 6 shows the result of the projection pursuit regression tree with depth 3, and Table 2 shows the coefficients of projection in each node and the overall importance measure of the projection pursuit regression tree as well as the importance measure of the random forest for regression. The length of the projection is stick to one and the value of the projection coefficient depends on the number of variables. For consistent decision, the number of variables in each coefficient is multiplied for comparison. The projection coefficients of projection pursuit in each node show the role of each variable in separating the left and the right nodes. If the absolute value of a coefficient is large (great than 1), the corresponding variable plays an important role in this separation.
In node 1, all Y values are divided into two groups—(G1, G2) and (G3, G4). To separate the smaller Y groups (G1, G2) from the larger Y groups (G3, G4), X 2 and X3 should play an important role, as indicated by the coefficients of node 1. Additionally, X 1 and X 2 are important for separating G1 and G2. This separation occurs in node 2 and the coefficients of node 2 are large in X 1 and X 2 . In separating G3 and G4 in node 3, X 3 and X 4 are important. To separate G1 into smaller and larger groups (node 4), X 1 should primarily be used. Additionally, X 2 should be used for G2 (node 5), X 3 should be used for G3 (node 10), and  X 4 should be used for G4 (node 11). All these features can be found in Table 2 by the corresponding coefficients of each node in the projection pursuit regression tree. Table 2 show the importance measure of the random forest for regression. The order of importance is X 3 , X 2 , X 1 , and X 4 . It shows the same result of the overall importance of the projection pursuit regression tree. Table 3 shows MSE of multiple regression line (lm), randomForest, rpart and 5 models of the projection pursuit regression tree. According to Table 3, all methods in the projection regression tree show better performance than the other three methods, lm, randomForest, and rpart.

4. Results

4.1. Application: White and Red Wine Data

To explain how to explore data and find interesting features with the projection pursuit regression tree, wine data from the UCI Machine Learning repository (“https://archive.ics.uci.edu/ml/datasets.php” [28]; accessed on 15 September 2021) is used. These data are for Portugal wines. Eleven sensory variables were collected from 4898 white wines and 1599 red wines. Additionally, the quality of wines was evaluated. The wine quality is between 0 (very bad) and 10 (excellent). In this paper, the projection pursuit regression tree is fitted to the white wine data and red wine data separately and compare the differences between white wine and red wine.
For this exploration, the projection pursuit regression tree is fitted first. Figure 7 shows the final projection pursuit regression tree. Table 4 shows the best projection coefficients for all internal nodes. Node 1 divides the data into a high-quality group (quality 7 ) and a low-quality group (quality 6 ). In both wines, if the alcohol is high and the density is low, the wine is classified in the high-quality group. Additionally, there are differences in the significant sensory variables between the white and red wine quality. White wines with high residual sugar are classified in the high-quality group. In contrast, red wines with high fixed acidity and high sulphates are classified in the high-quality wine group.
In node 2, the projection pursuit regression tree divides the low-quality group into two small groups, one with quality 6 and the other with the lowest quality (≤5). In white wine, high volatile acidity, low residual sugar, high density, and low alcohol tend to produce lower quality, and red wines with high volatile acidity, high total sulfur dioxide, low sulphates, and low alcohol have the lowest quality. Node 3 divides the top-quality group (≥8) and the group with quality 7. High residual sugar and low density are still important to classify a wine as a top-quality white wine. The red wine shows a bit of complicated sensory evaluation. Low fixed acidity, high volatile acidity, low chlorides, low pH, and high alcohol tend to produce top-quality wine.
Node 4 divides the lowest-quality group (≤4) and the group with quality 5. White wines with low residual sugar, high density, and high alcohol have the lowest quality. In red wine, high acidity, low density, and high pH tend to produce the lowest quality.
Table 5 shows the importance measures of each variable in the white wine data. For the projection pursuit regression tree, the importance measure of each variable is a weighted sum of the absolute value of the projection coefficients of each node. For the weights, the number of observations in each node is used. In the white wine data, four variables—density, residual sugar, alcohol, and volatile acidity—have important roles in dividing five groups; the lowest-quality group, the groups of quality 5, 6, and 7, and the top-quality group. These results are compared with the importance measure in randomForest. Alcohol, density, volatile acidity, and free sulfur dioxide are the four most important variables in randomForest. The roles of residual sugar and free sulfur dioxide are quite different in the two methods. Residual sugar is a very important variable in the projection pursuit regression tree but not in randomForest, and free sulfur dioxide is important in randomForest but not in the projection pursuit regression tree.
Table 6 shows the importance measures for red wine. The most important variable is alcohol in both methods. However, the other patterns are quite different. Random forest shows only two more important variables: sulphates and volatile acidity. However, the importance measures in the projection pursuit regression tree change smoothly up to total sulfur dioxide. This is also a quite different pattern to that of the importance measures of white wine. From this result, it is found that the quality of red wine is determined by more complex tastes than the quality of white wine.

4.2. Comparison of the Performance of Tree-Based Regression Methods

Even though the main purpose of the projection pursuit regression tree is to explore data, prediction is also an important feature of the tree-structured model. In this section, 18 data sets are used to examine the performance of the projection pursuit regression tree. These 18 data sets come from various sources, such as the UCI data repositories, Statlib, or R. For each data set, the data is divided into two-thirds training data and one-third test data, the projection pursuit regression model is fitted with the training data, and  the mean squared error (MSE) and the mean absolute errors (MAE) of the training data and the test data are calculated, respectively. lm, rpart, and randomForest are used to compare the performance. rpart is the most commonly used tree-based model and randomForest is also popular among ensemble tree methods. Usually a single regular tree-structured model cannot outperform ensemble tree methods.
Table 7 and Table 8 show the results of the MSE for the training and test data, respectively. The average of the ratios of the MSE of each method to the MSE of lm are calculated. If this average of ratios is less than 1, the performance of the corresponding method is better than that of the linear model.
For the training data, the projection pursuit regression tree shows better performance than rpart in seven data sets and shows competitive performance in three data sets. Surprisingly, in tecator and yacht data sets, the projection pursuit regression tree shows much better performance than randomForest. The projection pursuit regression tree shows much better performance in the test data. For 15 data sets, the projection pursuit regression tree shows similar or better performance than rpart. For three data sets—mussels, tecator, and yacht, the performance is much better than that of randomForest.
Table 9 and Table 10 show the ratios of MAE in the training and test data, respectively. In the MAE results, the projection pursuit regression tree shows similar or slightly better performance than rpart in most data sets, and it is outperformed in three data sets—mussels, tecator, and yachtdata. Even though the projection pursuit regression tree shows similar or slightly worse performance than randomForest, it is outperformed in these three data sets.
Sometimes the performance of Model 5 is better than the performance of Model 4. The reason is mainly due to the overfitting. In the price data, the final tree has eight final nodes, and each node has about 15 observations. Model 4 fits the linear regression model with 15 variables to the data with 15 observations and causes the overfitting. Therefore, Model 4, the multiple linear regression model with all variables, should be used when the number of observations is enough to fit the model.

5. Conclusions and Discussion

This paper proposed a new exploratory regression method: the projection pursuit regression tree. This tree is an extension of the projection pursuit classification tree to the regression problem. In each node, all observations in the node are divided into two groups—a larger dependent-variable group and a smaller dependent-variable group. Then, the projection pursuit method is applied using indices for classification. One of the projection pursuit indices—the LDA, Lr, or PDA index—is maximized to find a projection that separates the two groups. With the projection pursuit regression tree, the partitioned data as well as the data as a whole can be explored. The main difference between the general tree-based regression method and the projection pursuit regression tree is the way the data is partitioned. The general tree-based regression methods focus on the space of independent variables and each final node tends to have much wider-ranging Y values. However, the projection pursuit regression tree starts from the dependent variable. The data in each node are divided into two groups: group 1, with small dependent values, and group 2, with large dependent values. Then, group 1 is assigned to the left node, and group 2 is assigned to the right node. This procedure is repeated until a node is declared a final node. Therefore, all final nodes are ordered by the averages of Y values. With this property, the features of the independent variables can be explored easily in each range of the dependent variable. To explore more fine-grained ranges of the dependent variable, the depth of the projection pursuit regression tree can be increased.
The projection pursuit regression tree provides five different models to assign final values to the final nodes and these methods improve predictability. It is shown that the predictability of the projection pursuit regression tree is similar or better than that of rpart by testing 18 data sets. Even though our projection pursuit regression tree is not outperformed in all data sets, it shows better performance than rpart in most cases, and sometimes beats randomForest. This projection pursuit regression tree is developed in the R package PPtreereg and this package is available on GitHub (https://github.com/EK-Lee/PPtreereg).
The projection pursuit regression tree is originally from the projection pursuit classification tree, which separates classes using the separating hyperplane. This projection pursuit regression tree can perform poorly in a nonlinearly separable case, especially in the paraboloid case. Therefore, this regression tree also separates the data space using the separating hyperplane, limiting the nonlinearly separable case. Based on this study, it will be possible to extend the method of considering nonlinear transformations or non-linear projections when a growing tree. However, in that case, the computational complexity will increase exponentially, and the model’s interpretability will disappear. Therefore, in the area where the predictive model requires explanatory power, this developed tree can be sufficiently valuable because it can be explained by looking at the projection according to each node.

Author Contributions

This paper had equal contributions of all authors. Conceptualization, H.C. and E.-K.L.; methodology, H.C. and E.-K.L.; software, H.C. and E.-K.L.; validation, H.C. and E.-K.L.; formal analysis, H.C. and E.-K.L.; investigation, H.C. and E.-K.L.; resources, H.C. and E.-K.L.; data curation, H.C. and E.-K.L.; writing—original draft preparation, H.C. and E.-K.L.; writing—review and editing, H.C. and E.-K.L.; visualization, H.C. and E.-K.L.; supervision, E.-K.L.; project administration, E.-K.L.; funding acquisition, E.-K.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea Government (MSIT) (2018R1A2B6001251).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Morgan, J.N.; Sonquist, J.A. Problems in the analysis of survey data, and a proposal. J. Am. Stat. Assoc. 1963, 58, 415–434. [Google Scholar] [CrossRef]
  2. Breiman, L.; Friedman, J.; Stone, C.J.; Olshen, R.A. Classification and Regression Trees; CRC Press: Boca Raton, FL, USA, 1984. [Google Scholar]
  3. Loh, W.Y. Fifty Years of Classification and Regression Trees. Int. Stat. Rev. 2014, 82, 329–348. [Google Scholar] [CrossRef] [Green Version]
  4. Hothorn, T.; Hornik, K.; Zeileis, A. Unbiased recursive partitioning: A conditional inference framework. J. Comput. Graph. Stat. 2006, 15, 651–674. [Google Scholar] [CrossRef] [Green Version]
  5. Kim, H.; Loh, W.Y. Classification trees with bivariate linear discriminant node models. J. Comput. Graph. Stat. 2003, 12, 512–530. [Google Scholar] [CrossRef] [Green Version]
  6. Quinlan, J.R. Learning with continuous classes. In Proceedings of the 5th Australian Joint Conference on Artificial Intelligence, Hobart, TAS, Australia, 16–18 November 1992; Volume 92, pp. 343–348. [Google Scholar]
  7. Wang, Y.; Witten, I.H. Induction of Model Trees for Predicting Continuous Classes; University of Waikato: Hamilton, New Zealand, 1996. [Google Scholar]
  8. Torgo, L. Functional models for regression tree leaves. Int. Conf. Mach. Learn. 1997, 97, 385–393. [Google Scholar]
  9. Breiman, L. Bagging predictors. Mach. Learn. 1996, 24, 123–140. [Google Scholar] [CrossRef] [Green Version]
  10. Friedman, J.; Hastie, T.; Tibshirani, R. Additive logistic regression: A statistical view of boosting (with discussion and a rejoinder by the authors). Ann. Stat. 2000, 28, 337–407. [Google Scholar] [CrossRef]
  11. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  12. Chipman, H.A.; George, E.I.; McCulloch, R.E. BART: Bayesian additive regression trees. Ann. Appl. Stat. 2010, 4, 266–298. [Google Scholar] [CrossRef]
  13. Sigrist, F. KTBoost: Combined kernel and tree boosting. Neural Process. Lett. 2021, 53, 1147–1160. [Google Scholar] [CrossRef]
  14. Chang, Y. Panel data analysis with regression trees. J. Korean Data Inf. Sci. Soc. 2014, 25, 1253–1262. [Google Scholar]
  15. Meek, C.; Chickering, D.M.; Heckerman, D. Autoregressive tree models for time-series analysis. In Proceedings of the 2002 SIAM International Conference on Data Mining, SIAM, Chicago, IL, USA, 11–12 April 2002; pp. 229–244. [Google Scholar]
  16. Sela, R.J.; Simonoff, J.S. RE-EM trees: A data mining approach for longitudinal and clustered data. Mach. Learn. 2012, 86, 169–207. [Google Scholar] [CrossRef] [Green Version]
  17. Li, Z.; Hu, D. Exploring the relationship between the 2D/3D architectural morphology and urban land surface temperature based on a boosted regression tree: A case study of Beijing, China. Sustain. Cities Soc. 2021. [Google Scholar] [CrossRef]
  18. Wang, Y.; Fang, Z.; Hong, H.; Costache, R.; Tang, X. Flood susceptibility mapping by integrating frequency ratio and index of entropy with multilayer perceptron and classification and regression tree. J. Environ. Manag. 2021, 289, 112449. [Google Scholar] [CrossRef]
  19. Strzelecka, A.; Zawadzka, D. Application of classification and regression tree (CRT) analysis to identify the agricultural households at risk of financial exclusion. Procedia Comput. Sci. 2021, 192, 4532–4541. [Google Scholar] [CrossRef]
  20. Tyasi, T.L.; Eyduran, E.; Celik, S. Comparison of tree-based regression tree methods for predicting live body weight from morphological traits in Hy-line silver brown commercial layer and indigenous Potchefstroom Koekoek breeds raised in South Africa. Trop. Anim. Health Prod. 2021, 53, 1–8. [Google Scholar] [CrossRef]
  21. Friedman, J.H.; Stuetzle, W. Projection pursuit regression. J. Am. Stat. Assoc. 1981, 76, 817–823. [Google Scholar] [CrossRef]
  22. Kruskal, J.B. Toward a practical method which helps uncover the structure of a set of multivariate observations by finding the linear transformation which optimizes a new ‘index of condensation’. In Statistical Computation; Academic Press: New York, NY, USA, 1969; pp. 427–440. [Google Scholar]
  23. Friedman, J.H.; Tukey, J.W. A projection pursuit algorithm for exploratory data analysis. IEEE Trans. Comput. 1974, 100, 881–890. [Google Scholar] [CrossRef]
  24. Lee, Y.D.; Cook, D.; Park, J.W.; Lee, E.K. PPtree: Projection pursuit classification tree. Electron. J. Stat. 2013, 7, 1369–1386. [Google Scholar] [CrossRef]
  25. Lee, E.K.; Cook, D.; Klinke, S.; Lumley, T. Projection pursuit for exploratory supervised classification. J. Compuational Graph. Stat. 2005, 14, 831–846. [Google Scholar] [CrossRef] [Green Version]
  26. Lee, E.K.; Cook, D. A projection pursuit index for large p small n data. Stat. Comput. 2010, 20, 381–392. [Google Scholar] [CrossRef]
  27. Marron, J.; Todd, M.J.; Ahn, J. Distance-weighted discrimination. J. Am. Stat. Assoc. 2007, 102, 1267–1271. [Google Scholar] [CrossRef] [Green Version]
  28. Cortez, P.; Cerdeira, A.; Almeida, F.; Matos, T.; Reis, J. Modeling wine preferences by data mining from physicochemical properties. Decis. Support Syst. 2009, 47, 547–553. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Toy example: 100 observations; X 1 and X 2 are highly correlated with Y.
Figure 1. Toy example: 100 observations; X 1 and X 2 are highly correlated with Y.
Applsci 11 09885 g001
Figure 2. How to separate the space of independent variables in CART and in the projection pursuit regression tree.
Figure 2. How to separate the space of independent variables in CART and in the projection pursuit regression tree.
Applsci 11 09885 g002
Figure 3. Simple projection pursuit regression tree for the toy example with depth 1.
Figure 3. Simple projection pursuit regression tree for the toy example with depth 1.
Applsci 11 09885 g003
Figure 4. The scatter plot of Y and the predicted values using CART and the projection pursuit regression model with methods Models 1–5.
Figure 4. The scatter plot of Y and the predicted values using CART and the projection pursuit regression model with methods Models 1–5.
Applsci 11 09885 g004
Figure 5. Plots of the simulated data Y vs. X 1 , X 2 , X 3 and X 4 . G1, G2, G3 and G4 are divided by the dashed lines in each plot.
Figure 5. Plots of the simulated data Y vs. X 1 , X 2 , X 3 and X 4 . G1, G2, G3 and G4 are divided by the dashed lines in each plot.
Applsci 11 09885 g005
Figure 6. The projection pursuit regression tree with depth 2 for the simulated data.
Figure 6. The projection pursuit regression tree with depth 2 for the simulated data.
Applsci 11 09885 g006
Figure 7. The projection pursuit regression tree for the wine data: (a) white wine and (b) red wine.
Figure 7. The projection pursuit regression tree for the wine data: (a) white wine and (b) red wine.
Applsci 11 09885 g007
Table 1. MSEs of the toy example with lm, randomForest, rpart, and the projection pursuit regression tree.
Table 1. MSEs of the toy example with lm, randomForest, rpart, and the projection pursuit regression tree.
Projection Pursuit Regression Tree
MethodLmRandomForestRpartModel 1Model 2Model 3Model 4Model 5
MSE0.002729.83124.83444.02447.166.470.0025121.67
Table 2. The projection coefficients of each node in the projection pursuit regression tree and the importance measure of randomForest.
Table 2. The projection coefficients of each node in the projection pursuit regression tree and the importance measure of randomForest.
Projection Pursuit Regression Tree
CoefficientsX1X2X3X4
node 10.0172.5343.073−0.368
node 22.2313.257−0.360−0.530
node 3−0.023−0.4643.0312.569
node 43.9810.386−0.0730.002
node 50.2753.9710.378−0.100
node 10−0.0740.3253.985−0.080
node 110.2090.2330.4683.960
overall
importance
75.946187.443199.80098.442
randomForest
X1X2X3X4
importance
measure
6.72411.36011.6787.151
Table 3. MSEs of lm, randomForest, rpart, and the projection pursuit regression tree.
Table 3. MSEs of lm, randomForest, rpart, and the projection pursuit regression tree.
Projection Pursuit Regression Tree
MethodLmRandomForestRpartModel 1Model 2Model 3Model 4Model 5
MSE0.03530.00960.00830.00670.00730.00240.00240.0017
Table 4. Projection pursuit coefficients in each node for the white and red wine.
Table 4. Projection pursuit coefficients in each node for the white and red wine.
Variable NameWhite WineRed Wine
Node 1Node 2Node 3Node 4Node 1Node 2Node 3Node 4
fixed acidity1.56−0.241.78−0.044.341.04−4.32−5.05
volatile acidity−1.24−4.710.31−2.71−2.34−4.493.71−4.84
citric acid−0.170.140.510.331.24−1.831.92−1.36
residual sugar6.305.087.205.622.61−0.02−1.00−2.91
chlorides−0.230.122.780.14−2.26−1.47−3.76−0.96
free sulfur dioxide0.981.482.69−0.037−0.422.280.500.52
total sulfur dioxide−0.41−0.74−1.071.63−1.61−4.79−2.022.36
density−7.80−4.18−6.62−7.70−5.080.040.435.51
pH2.030.962.240.250.20−0.61−6.10−4.69
sulphates1.261.48−0.680.354.373.702.83−0.76
alcohol3.107.040.45−4.455.967.224.451.41
Table 5. Importance measures of the white wine data.
Table 5. Importance measures of the white wine data.
Projection Pursuit Regression TreeRandomForest
VariablesImportanceVariablesImportance
density646.23alcohol621.77
residual sugar587.62density413.10
alcohol436.91volatile acidity389.11
volatile acidity252.90free sulfur dioxide385.95
pH143.63chlorides306.12
free sulfur dioxide117.04total sulfur dioxide293.33
sulphates114.89residual sugar277.97
fixed acidity91.90pH258.04
total sulfur dioxide75.78citric acid255.65
chlorides41.67fixed acidity235.73
citric acid21.52sulphates222.10
Table 6. Importance measures of the red wine data.
Table 6. Importance measures of the red wine data.
Projection Pursuit Regression TreeRandomForest
IDVariablesImportanceIDVariablesImportance
X11alcohol545.82X11alcohol194.85
X2volatile acidity364.27X10sulphates135.30
X10sulphates336.86X2volatile acidity127.13
X8fixed acidity331.98X8density86.55
X1density313.67X7total sulfur dioxide79.42
X7total sulfur dioxide288.75X3citric acid69.13
X5chlorides182.27X5chlorides68.24
X4residual sugar166.65X1fixed acidity61.71
X3pH151.30X9pH57.77
X9citric acid150.70X4residual sugar55.05
X6free sulfur dioxide109.87X6free sulfur dioxide50.07
Table 7. Comparison of lm, rpart, randomForest, and the projection pursuit regression tree with various data sets; the mean MSE/MSE(lm)of 200 iterations for the training dataset.
Table 7. Comparison of lm, rpart, randomForest, and the projection pursuit regression tree with various data sets; the mean MSE/MSE(lm)of 200 iterations for the training dataset.
Projection Pursuit Regression Tree Random
DatasetnpModel 1Model 2Model 3Model 4Model 5RpartForest
Abalone417781.2061.2181.1921.1991.2001.0960.214
airfoil150351.2631.2341.2531.2661.2810.7190.414
ais202101.6311.6791.3031.0341.1511.1290.297
alcohol2467171.2151.1831.2441.2301.2351.0220.471
autompg39271.0741.0830.8410.7550.8830.7980.182
baseball263181.3961.3791.3371.2271.3270.8290.186
boston506131.2221.2770.8450.6940.8090.6990.110
budget172990.8180.8970.4790.3200.4650.2350.007
cane3775231.3801.3781.1131.0911.0971.0740.266
mussels8243.2893.4811.1470.9921.1191.2320.423
ozone33091.5301.5451.0110.9221.0590.7150.168
parkinsons5875201.3411.3331.1621.1301.1261.1250.122
price159151.9472.0531.6611.5250.9751.0430.221
strikes62841.0841.0551.1151.1051.1040.8000.377
tecator240227.4127.6923.8200.5002.0495.8671.056
wine-red1599111.2751.2741.3271.3251.3230.9380.168
wine-white4898111.3561.2991.4671.4411.4401.0060.137
yachtdata30860.1890.2000.0270.0090.0200.0630.108
Table 8. Comparison of lm, rpart, randomForest, and the projection pursuit regression tree with various data sets; the mean MSE/MSE(lm)of 200 iterations for the test dataset.
Table 8. Comparison of lm, rpart, randomForest, and the projection pursuit regression tree with various data sets; the mean MSE/MSE(lm)of 200 iterations for the test dataset.
Projection Pursuit Regression Tree Random
DatasetnpModel 1Model 2Model 3Model 4Model 5RpartForest
abalone417781.1661.1791.1571.1581.1611.1620.912
airfoil150351.2811.2481.2641.2761.2940.8600.580
ais202101.7351.7601.5401.4031.3651.4741.221
alcohol2467171.2081.1751.2401.2371.2331.0171.057
autompg39271.0891.1000.8730.8220.9341.1270.695
baseball263181.2061.1771.3121.1901.1521.0520.679
boston506131.2551.2870.9140.8130.8740.9720.471
budget172990.8320.9030.4930.3540.4780.2580.034
cane3775231.3831.3911.1371.1181.1301.2150.870
mussels8242.8412.9961.0521.1131.1201.8731.186
ozone33091.5071.5101.0100.9791.0611.1630.801
parkinsons5875201.3261.3201.1621.1431.1351.1580.681
price159151.6461.6831.5601.6251.0671.0070.525
strikes62841.0751.0461.1021.0971.0941.0820.935
tecator240226.1806.3284.1001.3362.7206.4112.918
wine-red1599111.2751.2751.3251.3301.3231.0890.817
wine-white4898111.3491.2941.4611.4391.4351.0200.673
yachtdata30860.2090.2340.0320.0170.0260.0610.207
Table 9. Comparison of lm, rpart, randomForest, and the projection pursuit regression tree with various data sets; the mean MAE/MAE(lm)of 200 iterations for the training dataset.
Table 9. Comparison of lm, rpart, randomForest, and the projection pursuit regression tree with various data sets; the mean MAE/MAE(lm)of 200 iterations for the training dataset.
Projection Pursuit Regression Tree Random
DatasetnpModel 1Model 2Model 3Model 4Model 5RpartForest
abalone417781.0791.0621.0721.0681.0681.0440.446
airfoil150351.0951.0831.0731.0701.0840.8540.669
ais202101.1911.1741.1190.9571.0291.0280.516
alcohol2467171.0580.9971.0921.0841.0901.0060.707
autompg39271.0181.0100.8880.8210.9100.8950.403
baseball263181.0421.0101.0230.8910.9660.7940.341
boston506131.0521.0400.8800.7410.8230.8430.309
budget172990.6600.6200.5280.4620.5180.5140.054
cane3775231.1311.1221.0601.0341.0441.1040.545
mussels8241.9961.9541.0780.9991.0411.1610.656
ozone33091.1351.1180.9520.8760.9350.8130.383
parkinsons5875201.1071.1001.0531.0331.0361.0640.332
price159151.1331.0861.0630.9090.8340.9380.409
strikes62841.0100.9641.0431.0241.0290.9030.548
tecator240222.4962.4411.6810.5171.3112.4131.060
wine-red1599110.8930.8930.9560.9600.9610.9790.390
wine-white4898111.0190.9781.1101.0921.0931.0210.347
yachtdata30860.2780.2720.1020.0600.0900.2400.271
Table 10. Comparison of lm, rpart, randomForest, and the projection pursuit regression tree with various data sets; the mean MAE/MAE(lm)of 200 iterations for the test dataset.
Table 10. Comparison of lm, rpart, randomForest, and the projection pursuit regression tree with various data sets; the mean MAE/MAE(lm)of 200 iterations for the test dataset.
Projection Pursuit Regression Tree Random
DatasetnpModel 1Model 2Model 3Model 4Model 5RpartForest
abalone417781.0741.0621.0731.0661.0661.0790.944
airfoil150351.1061.0941.0841.0851.0940.9340.791
ais202101.2721.2781.2261.1641.1631.2041.078
alcohol2467171.0671.0061.1031.1021.1031.0021.043
autompg39271.0321.0260.9010.8710.9471.0390.765
baseball263181.0631.0381.0871.0241.0190.9380.706
boston506131.0901.0750.9360.8460.8760.9670.660
budget172990.6700.6280.5390.4930.5310.5340.122
cane3775231.1521.1431.0821.0641.0711.1430.957
mussels8241.8541.8401.0471.0731.0661.4741.111
ozone33091.1441.1350.9560.9330.9531.0050.836
parkinsons5875201.1161.1081.0641.0481.0481.0820.804
price159151.1401.1061.1011.1220.9450.9790.681
strikes62841.0130.9751.0491.0301.0351.0110.869
tecator240222.4582.4251.7620.9071.4942.7261.999
wine-red1599110.9070.9070.9690.9790.9751.0480.864
wine-white4898111.0260.9831.1161.0991.1011.0320.774
yachtdata30860.2980.3030.1140.0840.1040.2400.360
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Cho, H.; Lee, E.-K. Tree-Structured Regression Model Using a Projection Pursuit Approach. Appl. Sci. 2021, 11, 9885. https://doi.org/10.3390/app11219885

AMA Style

Cho H, Lee E-K. Tree-Structured Regression Model Using a Projection Pursuit Approach. Applied Sciences. 2021; 11(21):9885. https://doi.org/10.3390/app11219885

Chicago/Turabian Style

Cho, Hyunsun, and Eun-Kyung Lee. 2021. "Tree-Structured Regression Model Using a Projection Pursuit Approach" Applied Sciences 11, no. 21: 9885. https://doi.org/10.3390/app11219885

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop