Previous Article in Journal
Symmetry-Aware Simulation and Experimental Study of Thin-Wall AA7075 End Milling: From Tooth-Order Force Symmetry to Symmetry-Breaking Dynamic Response and Residual Stress
Previous Article in Special Issue
Artificial Neural Network-Based Optimization of an Inlet Perforated Distributor Plate for Uniform Coolant Entry in 10 kWh 24S24P Cylindrical Battery Module
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Asymmetric Feature Weighting for Diversity-Enhanced Random Forests

1
Department of Statistics and Data Science, Yonsei University, Seoul 03722, Republic of Korea
2
Business Data Division, GS Retail, Seoul 06141, Republic of Korea
*
Author to whom correspondence should be addressed.
Symmetry 2026, 18(1), 73; https://doi.org/10.3390/sym18010073 (registering DOI)
Submission received: 23 October 2025 / Revised: 2 December 2025 / Accepted: 10 December 2025 / Published: 1 January 2026

Abstract

Random Forest (RF) is one of the most widely used ensemble learning algorithms for classification and regression tasks. Its performance, however, depends not only on the accuracy of individual trees but also on the diversity among them. This study proposes a novel ensemble method, Heterogeneous Random Forest (HRF), which enhances ensemble diversity through adaptive and asymmetric feature weighting. Unlike conventional RF that treats all features equally during tree construction, HRF dynamically reduces the sampling probability of features that have been frequently selected—particularly those appearing near the root nodes of previous trees. This mechanism discourages repetitive feature usage and encourages a more balanced and heterogeneous ensemble structure. Simulation studies demonstrate that HRF effectively mitigates feature selection bias, increases structural diversity, and improves classification accuracy, particularly on datasets with low noise ratios and diverse feature cardinalities. Comprehensive experiments on 52 benchmark datasets further confirm that HRF achieves the highest overall performance and significant accuracy gains compared to standard ensemble methods. These results highlight that asymmetric feature weighting provides a simple yet powerful mechanism for promoting diversity and enhancing generalization in ensemble learning.

1. Introduction

Random forest [1] is one of the most prevalent machine learning techniques for classification and regression problems. It is a tree-based ensemble method that aggregates the results of numerous decision trees to generate prediction outcomes. RF employs a strategy to induce diversity among classifiers. This is accomplished by choosing a random subset of candidate features from the pool of available features and then splitting a node of the tree based on the most optimal variable within this subset.
The performance of ensemble methods is largely determined by two factors: the predictive quality of the individual base learners and the degree of correlation among them. When the base learners achieve higher accuracy, the ensemble also tends to perform better. Likewise, when the trees are more heterogeneous—i.e., less correlated—the ensemble generally benefits from increased diversity. Based on these principles, numerous studies have proposed modifications to Random Forests (RF) aimed at improving accuracy, diversity, or both.
Several studies have enhanced the performance of ensemble models by developing more sophisticated trees. Specifically, some approaches have focused on selecting only the most relevant variables or eliminating redundant ones. This can be seen as a feature selection technique aimed at improving the accuracy of individual trees [2,3,4,5,6,7]. In contrast, other research has explored the construction of oblique trees to simultaneously boost accuracy and diversity. These methods involve transforming features, such as through linear combinations, to create splits that are not aligned with the original feature axes [8]. For instance, Refs. [9,10] proposed rotated trees based on techniques like LDA (Linear Discriminant Analysis), CLDA (Canonical LDA), and CCA (Canonical Correlation Analysis).
Instead of focusing solely on the performance of individual trees, some approaches emphasize increasing the diversity among trees. A common strategy involves introducing additional randomness to reduce the correlation between trees. For instance, Double Random Forest (DRF) [11] enhances diversity through bootstrapping at the node level; the Extremely Randomized Trees (ExtTrees) method [12] increases randomness by selecting candidate split points completely at random; and the Random Patches method [13] simultaneously samples both training instances and feature subsets, thereby promoting greater heterogeneity among the individual trees within the ensemble.
Another limitation of conventional Random Forest (RF) lies in its handling of datasets containing features with different cardinalities. In such cases, variables with low cardinality—those having fewer unique values—tend to be selected less frequently for splitting compared to high-cardinality variables. This imbalance introduces a form of feature selection bias, causing RF to undervalue potentially important low-cardinality predictors even when they carry substantial predictive information. As a result, RF not only produces biased estimates of feature importance but also exhibits reduced ensemble diversity. These shortcomings highlight the need for a method that can alleviate feature dominance and promote a more balanced utilization of features across trees.
This paper proposes a novel ensemble method called Heterogeneous Random Forest (HRF), which enhances tree diversity through asymmetric feature weighting. During training, HRF adaptively adjusts feature sampling probabilities based on their usage in previous trees—assigning lower weights to frequently selected features and higher weights to underutilized ones. This intentional asymmetry mitigates feature dominance and promotes structural heterogeneity across trees. While prior works [14,15,16,17] also employed feature weighting, HRF uniquely models feature-level asymmetry to achieve controlled ensemble diversity.
The remainder of this article is organized as follows. Section 2 presents the proposed Heterogeneous Random Forest (HRF), outlining its core ideas and full algorithmic framework. Section 3 summarizes simulation studies designed to analyze HRF’s behavior under controlled settings, including its susceptibility to selection bias, its impact on feature importance, its ability to enhance structural diversity, and its performance trends under noise and correlated feature conditions. Section 4 provides empirical results across 52 benchmark datasets, including overall accuracy comparisons with existing ensemble methods, noise–performance analysis, and observed hyperparameter selection patterns. Finally, Section 5 concludes the paper with key findings and directions for future research.

2. Method

Consider an input (feature, explanatory variable) X j = [ x j 1 , , x j n ] T , where j = 1 , , p , represented by n instances. The training inputs, represented by the matrix X = [ X 1 , , X p ] , are composed of p dimensions and have the form of an n × p matrix. y denotes the class labels (1,…K) of inputs, where y = [ y 1 , , y n ] T . The training set, denoted by D , comprises both inputs and class labels and is defined as D = [ X , y ] . The base classifiers that comprise an ensemble are represented by the symbols h 1 , , h B .

2.1. Random Forest

Random forest (RF) is a type of bagging method that employs bootstrap sampling for each tree. The distinction between bagging and RF lies in the random selection of m candidate features for splitting. At each node of the tree, a new random feature subset is constructed, and the optimal split is selected based on the goodness of split within the subset. RF employs this strategy to induce diversity among individual trees, introducing additional randomness. The detailed algorithm is described in Algorithm 1. In general, the most commonly used value for hyper-parameter m is p .    
Algorithm 1: Random Forest Algorithm
Symmetry 18 00073 i001

2.2. Basic Idea of New Method

The proposed method employs a memory mechanism that retains the structural characteristics of previously constructed trees. This stored information is subsequently utilized to generate more diverse trees in later stages.
Due to the hierarchical nature of decision trees, splits occurring at shallow depths (near the root node) have a broad influence on the overall decision space, whereas splits at deeper depths exert more localized effects and have a comparatively minor impact on the global tree structure.
Accordingly, enhancing ensemble diversity requires regulating the features selected near the root nodes, as these early splits exert the strongest influence on tree structure. In this study, we seek to decrease the likelihood that features heavily used near the roots of earlier trees are repeatedly selected in subsequent trees. To achieve this, we introduce an asymmetric weighted feature-sampling mechanism that incorporates information from previously grown trees when selecting features for the next split. In contrast to conventional Random Forests (RF), where feature sampling is purely random and independent of prior trees, the proposed method adaptively adjusts feature sampling probabilities based on the historical dominance of each feature, thereby discouraging repeated use of overly influential features.

2.3. Feature Depth

Let us consider an individual tree within the ensemble. We define the feature depth vector d b = [ d b , 1 β , d b , 2 β , , d b , p β ] , where each element d b , j β represents the earliest level at which feature X j is used in the b-th tree. Formally, the feature depth of X j is defined as
d b , j β = minimum depth that X j used for split , if X j is used in b - th tree M b 1 + β , otherwise ,
where j = 1 , 2 , , p , b = 1 , 2 , , B , M b represents the depth of h b , and β is a hyperparameter assigned as a positive integer.
To illustrate, consider the case where the first decision tree, denoted as h 1 ( x ) , is constructed following the structure shown in Figure 1a. The feature X 1 is selected as the splitting criterion at the root node and therefore assigned a feature depth of 0. For features that appear multiple times within the same tree, the smallest depth value—corresponding to the occurrence closest to the root node—is recorded. For example, feature X 2 appears at both the first and third levels of h 1 ( x ) and is thus assigned a depth value of 1.
If certain features, such as X 4 and X 5 , are not selected during tree construction, their feature depths are assigned the maximum possible depth of the tree (which equals 4 when β = 1 ). The same procedure is applied to compute the feature depths for the second tree. The resulting feature depth vectors, d 1 for h 1 ( x ) and d 2 for h 2 ( x ) , are summarized in Table 1(a).

2.4. Cumulative Depth

The next step is to compute the cumulative depth vector, denoted as D , for each variable across the previously constructed trees. Since h 1 ( x ) is the first tree, its cumulative depth D 1 is identical to its own depth vector d 1 , as shown in Table 1(b). For subsequent trees, the cumulative depth D t is updated as a weighted combination of the previous cumulative depth and the current tree’s depth, where the preceding depth is multiplied by the hyperparameter α before being added to the current one. This formulation allows recent trees to have greater influence while preserving the historical contribution of earlier ones. For instance, when α = 0.5 , the cumulative depths up to the second tree are calculated as follows: the cumulative depth for X 2 becomes 4.5 , computed as 1 × 0.5 + 4 , whereas that for X 3 is 3, computed as 2 × 0.5 + 2 .
The hyperparameter α can be interpreted as a memory parameter that controls the influence of previous trees, i.e., how much each subsequent tree remembers its predecessors. When α = 1 , all information is fully retained, and the b-th tree is influenced equally by all prior trees. As α approaches zero, the influence of earlier trees diminishes rapidly, and only the most recent trees affect the subsequent ones. The cumulative depth resulting from the b-th construction can be expressed as follows:
D b = d b + α d b 1 + α 2 d b 2 + + α b 1 d 1
= d b + k = 1 b 1 α k d b k ,
where d b = [ d b , 1 β , , d b , p β ] is a vector of feature depths.

2.5. Heterogeneous Random Forest

The next step is to compute the feature weight vector, obtained by normalizing the cumulative depth of each feature by the total sum of cumulative depths from all previous trees, as formulated in Equation (4).
w b + 1 = D b D b 1 ,
where D b denotes the cumulative depth vector aggregated from all previously constructed b trees, and 1 is a column vector of ones used for summation. The resulting weight vector w b + 1 represents the normalized importance of each feature and satisfies w b + 1 1 = 1 .
For instance, the feature weights for h 2 ( x ) , denoted as w 2 , are computed by dividing the cumulative depth of each feature by the total sum of cumulative depths from h 1 ( x ) , as shown in Table 1(c). It can be observed that the feature weight of X 1 is 0, whereas those of X 4 and X 5 are the largest. Consequently, feature X 1 is excluded from the construction of h 2 ( x ) , while X 4 and X 5 , which were not selected in h 1 ( x ) , appear near the root nodes of h 2 ( x ) .
The overall procedure is completed through Algorithm 2, which implements the weighting mechanism described in Equation (4). The resulting weights are subsequently used as feature sampling probabilities for constructing the next tree. The hyperparameter β controls the degree to which unselected features are emphasized; a larger value of β assigns higher sampling weights to features that were not chosen in the previous trees.
Algorithm 2: Weight Updating Algorithm
Symmetry 18 00073 i002
Finally, the complete procedure of the proposed Heterogeneous Random Forest (HRF) algorithm is summarized in Algorithm 3.

2.6. New Measure of Diversity

Our proposed method, HRF, seeks to improve the diversity of trees within an ensemble. In ensemble learning, diversity is often assessed using prediction-oriented measures such as the disagreement measure [18], the double-fault measure [19], or the Kohavi–Wolpert variance [20], all of which quantify how differently individual learners respond to the same input instances. While these metrics are widely used, they rely solely on the outputs of the learners and do not capture how heterogeneous the underlying tree structures are. To complement this perspective, we introduce a new tree-structure-based dissimilarity metric designed to evaluate the internal diversity encouraged by HRF. Specifically, our metric quantifies heterogeneity in terms of feature usage patterns and feature dominance across trees, offering a structural viewpoint that prediction-based metrics cannot provide. This allows us to examine whether HRF successfully induces diverse tree architectures, beyond differences observed only at the prediction level.
The structural characteristics of a decision tree can be represented by its feature depth vector, d b , which specifies the depth at which each feature is used for splitting. Based on this representation, we define the concept of feature dominance, which reflects the relative importance of features within a single tree and enables quantitative comparison between different trees.
Algorithm 3: Heterogeneous Random Forest Algorithm
Symmetry 18 00073 i003
The dominance of the j-th feature in the b-th classifier is defined as
I b , j = M b * + 1 d b , j β if X j is used in b - th tree 0 , otherwise ,
where M b * denotes the maximum feature depth among the split variables in h b . Unused features are assigned a dominance value of 0, whereas features appearing near the root node receive relatively larger values.
To compare two trees, h l and h m , we concatenate their dominance vectors by row to form a 2 × p table, [ I l , I m ] . A χ 2 test of homogeneity is then applied to this table. A large χ 2 statistic indicates that the two trees differ substantially in their feature usage, whereas a small statistic implies structural similarity.
Figure 2 illustrates examples of tree dissimilarity. Trees 1 and 2, as well as Trees 3 and 4, exhibit similar structural patterns. Table 2(a) presents the dominance values for each feature, while Table 2(b) reports the corresponding χ 2 test statistics used to measure similarity between trees.
To enable comparison under varying degrees of freedom, the χ 2 statistic can be transformed into a standard normal variable using the Wilson–Hilferty transformation [21]. The resulting dissimilarity, D S , between two trees, h l and h m , is defined as
D S ( h l , h m ) = X / k 3 1 2 9 k 2 / ( 9 k ) ,
where X represents the χ 2 test statistic with k degrees of freedom. Finally, Table 2(c) summarizes the resulting dissimilarity ( D S ) values. In general, larger positive dissimilarity values indicate greater structural differences between trees. As shown, the D S value between Trees 1 and 2, as well as between Trees 3 and 4, is relatively small, while other tree pairs exhibit considerably higher positive values, validating the effectiveness of the proposed measure.

2.7. Scalability

The computational complexity of RF [22] is known to be O ( B · m · n · log n ) , where B is the number of trees, m is the number of candidate features considered at each split, and n is the number of training samples. HRF follows the same tree-growing process as RF, with the addition of a feature-weight update after each tree is built. This update requires only a single pass through all p features, contributing an additional O ( B · p ) term. Therefore, the total complexity of HRF is
O B · m · n · log n + B · p ,
which differs from RF only by a lower-order linear term.
To validate this theoretical analysis, we empirically examined HRF’s training time while independently varying (i) the number of samples n, (ii) the number of features p, and (iii) the ensemble size B. All experiments were conducted on synthetic datasets in which all features were equally informative, with default settings of n = 10,000 , p = 100 , and B = 100 . Each factor was varied one at a time while the others were held fixed, and the HRF-specific parameters were set to α = 0.5 and β = 1 . The results, summarized in Figure 3, show that HRF scales approximately linearly with n, p, and B, consistent with the theoretical complexity. These findings confirm that HRF introduces negligible computational overhead relative to RF and remains computationally feasible even for high-dimensional datasets.

3. Simulation Study

The objective of this simulation study is to examine several key aspects of the proposed HRF model. Each simulation corresponds to a binary classification task consisting of 1000 samples, of which 70% were used for training and the remaining 30% for evaluation. The ensemble comprised 100 individual trees, with the hyperparameters for tree construction set to their default values. The default values of the HRF-specific parameters α and β were 0.5 and 1, respectively. To ensure robust comparison, the entire simulation procedure was repeated 100 times.

3.1. Selection Bias

We aim to assess the extent to which HRF exhibits feature selection bias, as previous studies have reported that both bagging and RF suffer from such bias due to the greedy search property of decision trees [23].
In this experiment, five input features were generated. Feature X 1 takes integer values between 0 and 128, X 2 between 0 and 64, X 3 between 0 and 32, X 4 between 0 and 16, and X 5 between 0 and 8, all following discrete uniform distributions. The binary target variable was generated as y = Bernoulli ( σ ( η Q 2 ( η ) ) ) , where Q 2 ( η ) denotes the median of η , σ ( · ) is the sigmoid function, and Bernoulli ( · ) represents a binary random variable with the given probability. The linear predictor η is defined as η = X 1 + X 2 + X 3 + X 4 + X 5 , indicating that all five features are equally important in determining y .
The feature depth vector d b = [ d b , 1 1 , , d b , p 1 ] represents the depth of each feature in the b-th tree within the ensemble. If selection bias exists toward features with a larger number of unique values, the feature depth of X 1 (i.e., d b , 1 1 ) will tend to be smaller, as it is more likely to be selected near the root node. Conversely, features with fewer unique values will have larger depth values. For an ensemble of 100 trees, a total of 100 feature depth values are accumulated for each feature.
Figure 4 presents box plots of feature depth values for each feature across different ensemble methods. It can be observed that both bagging and RF tend to select features with a larger number of unique values near the root nodes, a phenomenon attributable to selection bias. In particular, bagging exhibits a pronounced bias, almost always prioritizing the selection of feature X 1 . In contrast, HRF substantially mitigates this bias, as the feature depth distributions of HRF are more balanced across features compared with Bagging and RF.

3.2. Impact on Feature Importance

In this subsection, we investigate how feature selection bias influences the estimated feature importance of different ensemble methods. Ideally, when all input features contribute equally to the target variable, the resulting feature importance values should be similar across features. However, as shown in Figure 5, both bagging and RF display a pronounced imbalance in their importance estimates, heavily emphasizing the X 1 feature while underestimating the others. This indicates that the feature selection bias observed in these methods adversely affects their feature importance estimation, causing them to overvalue features with higher cardinality.
In contrast, HRF produces feature importance values that are much more evenly distributed among all five features, reflecting the true underlying structure of the data.

3.3. Dissimilarity

In this section, we examine whether HRF indeed enhances ensemble diversity relative to bagging and RF. Using the same simulation setup described in Section 3.1, we computed the tree-level dissimilarity measure D S in a pairwise manner for all trees within each ensemble (see Equation (6)). The overall dissimilarity for each method was obtained by averaging all pairwise values, and the entire procedure was repeated 100 times to ensure stability of the estimates.
As shown in Figure 6, bagging exhibits the lowest level of structural diversity, followed by RF. HRF consistently achieves the highest dissimilarity scores among the compared methods, indicating that its asymmetric feature weighting mechanism effectively promotes more diverse tree structures.

3.4. Noise Feature Effect

As HRF assigns feature weights based on feature selection information from preceding trees, there is a potential risk of assigning higher selection weights to noise features. We intend to assess the severity of this problem.
To investigate the impact of noise features on the HRF method, a case with a total of 30 features was considered. Let p * denote the number of informative features. The input features X j , j = 1 , , 30 , were generated from discrete uniform distribution between 0 and 4. The target variable is generated by y = b e r n o u l l i ( σ ( η Q 2 ( η ) ) ) , where η = j = 1 p * X j . In the simulation data, the noise feature ratio was increased in increments of 0.1, starting from 0 and reaching 0.7.
Figure 7 presents box plots showing the accuracy differences between HRF and bagging, as well as between HRF and RF. A positive value indicates that HRF exhibits higher accuracy. HRF outperforms bagging when the proportion of noise features is less than 60% of the total features. Conversely, when the proportion of noise features exceeds 60%, HRF experiences a decline in performance. Similarly, HRF outperforms RF when the proportion of noise features is below 40% of the total features; however, when the proportion exceeds 50%, HRF also sees a decrease in performance.

3.5. Hyperparameter Effect

We further examined how the HRF-specific hyperparameters—the memory factor α and the penalty factor β —influence predictive performance. This experiment was conducted under a simulation design similar to that of Section 3.4, using datasets with 100 features and 5000 samples. Figure 8 displays the accuracy differences between HRF and RF across various combinations of ( α , β ) under different proportions of informative features.
When the proportion of informative variables is moderate to high, HRF generally outperforms RF across most parameter settings, although the magnitude of improvement varies across ( α , β ) ; this indicates that hyperparameter tuning can be beneficial in settings where many features carry meaningful signals. In contrast, when noise features dominate the feature space (i.e., informative proportion below 0.4), HRF tends to underperform RF in a consistent manner across all ( α , β ) configurations. This suggests that the degradation observed in high-noise scenarios does not arise from particular hyperparameter choices but rather reflects an inherent limitation of the HRF weighting mechanism when informative variables are overwhelmed by noise.
Overall, these results indicate that while HRF exhibits mild sensitivity to ( α , β ) in signal-rich environments, its performance drop in extremely noisy settings is largely independent of specific parameter values.

3.6. Effect of Correlated Features

To assess whether penalizing a dominant feature simply transfers its influence to a correlated surrogate or instead promotes exploration of non-redundant predictors, we conducted an additional simulation study. The input features X j ( j = 1 , , 5 ) were generated with 2000 observations, where each feature followed a standard normal distribution independently, except that a strong correlation ( ρ = 0.9 ) was imposed between X 1 and X 2 . The remaining features X 3 , X 4 , and X 5 were independent of X 1 . The binary response was generated according to the linear predictor η = 2 X 1 + X 3 + X 4 + X 5 , such that only X 1 , X 3 , X 4 , and X 5 carried true signal, while X 2 acted purely as a correlated surrogate.
Figure 9 compares feature-depth distributions for RF and HRF.
Under RF, the dominant feature X 1 consistently appears near the root, while the remaining features—including the correlated surrogate X 2 —are typically selected much deeper in the trees. Under HRF, the depth of X 1 increases as expected due to penalization, but this reduction in dominance is not accompanied by a corresponding decrease in the depth of X 2 . Instead, the non-redundant informative features X 3 X 5 tend to exhibit smaller depth values (i.e., they appear closer to the root), indicating that the additional sampling probability is distributed among distinct predictors rather than being absorbed by the correlated surrogate.
These results suggest that even in the presence of a strongly correlated pair, HRF does not merely replace one dominant feature with another; rather, it promotes a broader use of non-redundant features, contributing to increased structural diversity in the ensemble.

4. Empirical Evaluation

In order to investigate the performance of HRF, a series of experiments were conducted. The experiments were based on 52 real or artificial datasets used in previous studies or obtained from the UCI data repository [24]. Table 3 provides a concise overview of the datasets.
We compared the proposed HRF with several representative ensemble algorithms. The candidate hyperparameters and their sources are summarized in Table 4. Extremely Randomized Trees (ExtTrees; [12]) promote diversity by introducing stronger randomization in the tree construction process—specifically, by selecting both the split features and their thresholds completely at random. In contrast, Random Patches [13] enhance diversity by randomly sampling subsets of both training samples and features for each base learner. XGBoost [31] employs gradient boosting with regularization and shrinkage, and CatBoost [32] applies ordered target encoding and robust boosting strategies to mitigate overfitting. To evaluate classification performance, we measured the classification accuracy obtained through ensemble voting.

4.1. Experimental Setting

Any missing values were imputed using the mean for numerical features and the mode for categorical features. As an encoding method for categorical inputs, TargetEncoder [35] was employed for binary classification, while PolynomialWrapper [36] was utilized for multi-class classification to convert them into numerical ones. TargetEncoder was selected instead of one-hot encoding because it reduces dimensionality and sparsity while preserving target-dependent information that can benefit tree-based models. Likewise, PolynomialWrapper was applied to capture nonlinear feature interactions that may not be fully represented by standard tree splits.
Each dataset was randomly divided into a training set (80%) for model fitting and a test set (20%) for performance evaluation. All ensemble methods were constructed with 100 trees, and their hyperparameters were optimized through a 5-fold cross-validation process. For HRF, the same default settings as RF were used for tree construction, while the optimal values of α and β were determined via 5-fold cross-validation. Specifically, α was varied from 0.0 to 0.9 in increments of 0.1, and β was varied from 1 to min ( 10 , p ) —where p denotes the number of features—in increments of 1. To ensure statistical reliability, all experiments were repeated 50 times, and the results were evaluated using the Wilcoxon signed-rank test for pairwise statistical comparison [37].

4.2. Classification Accuracy

Table 5 presents a comparison of the accuracy between HRF and other ensemble methods. The ‘+’ symbol indicates that HRF showed statistically more accurate results than the corresponding method, whereas the ‘−’ symbol indicates less accurate results. The ‘W/T/L’ at the bottom row represents the number of datasets where HRF was better/equal/worse compared to the respective method.
Table 6 displays pairwise comparisons among the methods. The values in the table show the number of instances where the method listed vertically is more accurate than the method listed horizontally, with the number in parentheses denoting statistical significance. To illustrate, a value of 35(22) in the first row and second column indicates that RF outperforms the Bagging method in 35 out of 52 datasets, with 22 of those instances being statistically significant. In cases where the accuracy values are the same, a value of 0.5 was assigned.
Table 7 summarises the ranking of methods according to the results presented in Table 6. The dominance rank is calculated as the difference between the number of significant wins and significant losses. For example, HRF achieves a dominance rank of 164, derived from a total of 206 significant wins and 42 significant losses. The number of significant wins for HRF is equal to the cumulative sum of the values within parentheses in the HRF columns of Table 6. Similarly, the number of significant losses for HRF is the sum of the values within parentheses in the HRF row. Table 6 and Table 7 collectively demonstrate that the HRF method exhibits significant superiority over other methods across a dataset pool comprising 52 instances.

4.3. Noise Feature Effect

In Section 3.4, we investigated the effects of noise features on HRF. Specifically, when the number of noise features is excessively large, HRF tends to perform worse than bagging and RF. In this section, we further examined whether a similar pattern is observed in experiments with real datasets.
To estimate the proportion of noise features in each dataset, a deep decision tree was trained to compute feature importance values for all variables. Let the feature importance vector be denoted as I = [ I 1 , , I p ] , which was normalized as I [ I 1 , , I p ] / j I j . If the scaled feature importance was smaller than 1 / p , that is, I j < 1 p , the corresponding feature was regarded as noise. The resulting proportion of noise features (p_noise = # noise features p ) is summarized in Table 8.
We divided all datasets into two groups based on a noise ratio threshold of 0.75. Then, using the accuracy differences between HRF and other ensemble methods within each group, we generated box plots, as shown in Figure 10. The results indicate that HRF achieves relatively higher accuracy improvements over bagging and RF in datasets with lower proportions of noise features, consistent with the earlier simulation findings.
Unlike bagging and RF, the accuracy difference between HRF and CatBoost does not show a clear dependence on the proportion of noise features. This observation suggests that the difference in accuracy is primarily attributable to fundamental distinctions in model structure and learning mechanisms: CatBoost employs gradient boosting with ordered target encoding and regularization, whereas HRF enhances diversity through adaptive feature weighting within a bagging framework.

4.4. Hyperparameter Selection

We analyzed the stability and distribution of the optimal values selected through cross-validation across the 52 benchmark datasets. Specifically, for each dataset, the ( α , β ) pair was chosen via 5-fold cross-validation, repeated 50 times under identical train–test splits to ensure robustness.
Figure 11 summarizes the empirical distribution of the selected ( α , β ) values across all datasets. Despite variability across data domains, a clear concentration pattern emerges: most optimal values lie within the ranges α [ 0.1 , 0.4 ] and β [ 1 , 4 ] . This clustering indicates that simple near-default configurations (e.g., α = 0.3 , β = 2.5 ) can often provide competitive performance, reducing the practical burden of extensive hyperparameter tuning.

4.5. Discussion

The empirical evaluation demonstrates that the proposed Heterogeneous Random Forest (HRF) consistently achieves competitive or superior classification accuracy compared with conventional ensemble methods across 52 benchmark datasets. As summarized in Table 7, HRF achieved the highest dominance rank among all compared methods, indicating its superiority in predictive performance across diverse datasets. The improvements observed in HRF can be attributed to its adaptive and asymmetric feature weighting mechanism, which promotes feature diversity and mitigates the repetitive selection of dominant features. This mechanism enables HRF to better balance the utilization of informative features, leading to enhanced generalization performance without introducing additional model complexity.
The analysis of noise-feature effects further compared HRF and RF to examine how performance changes under varying levels of irrelevant or noisy variables. The results revealed that HRF shows its greatest advantage in datasets with relatively low proportions of noise features. As the proportion of irrelevant or noisy variables increases, HRF’s benefit tends to diminish, and in some cases, its accuracy becomes comparable to that of RF. This outcome is consistent with HRF’s design philosophy—its adaptive weighting is most effective when meaningful features are present and not overwhelmed by noise. When the same analysis was extended to CatBoost, the accuracy difference between HRF and CatBoost did not show a clear dependence on the proportion of noise features. This suggests that, unlike bagging-based ensembles where diversity plays a central role, CatBoost’s performance is primarily governed by its gradient-based optimization and regularization mechanisms.
Overall, these findings indicate that HRF is particularly effective in improving predictive performance when datasets contain a balanced and informative set of features. However, its relative advantage may be reduced in high-noise environments, suggesting that pre-filtering of irrelevant variables or feature selection techniques could further enhance its robustness in practical applications.

5. Conclusions

In this study, we proposed the Heterogeneous Random Forest (HRF), an ensemble method that extends the Random Forest (RF) framework through adaptive and asymmetric feature weighting designed to promote structural diversity among decision trees. HRF introduces several methodological mechanisms that contribute to improved ensemble performance:
  • Asymmetric Feature Weighting: Sampling probabilities are adaptively updated based on feature usage history, reducing the tendency to repeatedly select dominant features.
  • Mitigation of Feature Selection Bias: HRF alleviates RF’s well-known preference for high-cardinality variables, enabling more balanced utilization of informative low-cardinality features.
  • Enhanced Ensemble Diversity: By discouraging redundant split patterns, HRF yields more heterogeneous tree structures, which in turn improve generalization performance.
Together, these mechanisms reduce repeated reliance on dominant features and encourage exploration of underutilized but potentially meaningful variables, resulting in more balanced feature usage and improved predictive accuracy without increasing model complexity. Simulation studies and empirical analyses further demonstrated that HRF performs particularly well when the dataset contains meaningful signals and diverse feature characteristics. Although its relative advantage diminishes when irrelevant or noisy features dominate, HRF still maintains performance comparable to RF, indicating strong overall stability.
Across 52 benchmark datasets, HRF achieved the highest dominance rank among all compared ensemble algorithms, outperforming Bagging, RF, Extremely Randomized Trees, Random Patches, and boosting-based models such as XGBoost and CatBoost. HRF showed significant improvements over RF in 43 out of 52 datasets, confirming its robustness and consistent predictive advantage in practice.
We also introduced a new structural dissimilarity metric based on feature dominance and the χ 2 test to quantify internal diversity within ensembles. This measure revealed that HRF produces more heterogeneous decision trees than conventional ensemble methods, providing empirical evidence that encouraging structural diversity at the feature-selection level can translate into measurable performance gains.
Overall, HRF offers a practical and theoretically grounded enhancement to traditional ensemble learning by improving feature-level diversity and mitigating selection bias. It represents a promising approach for classification tasks that require robust, diverse ensemble models across a broad range of data domains.
  • Future Work. There remain several directions for extending this work. First, exploring strategies to improve HRF’s robustness in high-noise environments would be valuable. Initial feature filtering—using statistical tests or preliminary feature-importance screening—may help reduce the influence of irrelevant variables, while less aggressive adjustments to the adaptive weight update rule could improve stability.
Second, a comprehensive comparison between the proposed structural dissimilarity measure D S and established prediction-oriented diversity indices (e.g., disagreement, double-fault, and Kohavi–Wolpert variance) would clarify the relationships among different diversity notions and further validate D S as a structural proxy for ensemble diversity.
These directions provide a basis for future methodological development and for adapting HRF to real-world applications where noise levels and feature characteristics are highly variable.

Author Contributions

Conceptualization, Y.E.K., S.Y.K. and H.K.; methodology, Y.E.K. and S.Y.K.; software, Y.E.K.; validation, Y.E.K.; formal analysis, Y.E.K.; investigation, Y.E.K.; writing—original draft preparation, Y.E.K.; writing—review and editing, Y.E.K., S.Y.K. and H.K.; visualization, Y.E.K.; supervision, H.K.; project administration, H.K.; funding acquisition, H.K. All authors have read and agreed to the published version of the manuscript.

Funding

Hyunjoong Kim’s work was supported by the IITP (Institute of Information & Coummunications Technology Planning & Evaluation)-ICAN (ICT Challenge and Advanced Network of HRD) grant funded by the Korea government (Ministry of Science and ICT) (IITP-2023-00259934) and by the National Research Foundation of Korea (NRF) grant funded by the Korean government (RS-2016-NR017145).

Data Availability Statement

The data are publicly available from the relevant source and can also be provided upon request.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [PubMed]
  2. Nanni, L.; Lumini, A. Input Decimated Ensemble based on Neighborhood Preserving Embedding for spectrogram classification. Expert Syst. Appl. 2009, 36, 11257–11261. [Google Scholar] [CrossRef]
  3. Kumar, A.; Yadav, J. A review of feature set partitioning methods for multi-view ensemble learning. Inf. Fusion 2023, 100, 101959. [Google Scholar] [CrossRef]
  4. Campos, Y.; Estrada, R.; Morell, C.; Ferri, F.J. A Feature Set Decomposition Method for the Construction of Multi-classifier Systems Trained with High-Dimensional Data. In Progress in Pattern Recognition, Image Analysis, Computer Vision, and Applications; Ruiz-Shulcloper, J., Sanniti di Baja, G., Eds.; Springer: Berlin/Heidelberg, Germany, 2013; pp. 278–285. [Google Scholar]
  5. Rokach, L. Using recursive feature elimination in random forest to account for correlated variables in high dimensional data. BMC Genet. 2008, 19, 65. [Google Scholar] [CrossRef]
  6. Conn, D.; Ngun, T.; Li, G.; Ramirez, C. Fuzzy Forests: Extending Random Forest Feature Selection for Correlated, High-Dimensional Data. J. Stat. Softw. 2019, 91, 1–25. [Google Scholar] [CrossRef]
  7. Rokach, L. Genetic algorithm-based feature set partitioning for classification problems. Pattern Recognit. 2008, 41, 1676–1700. [Google Scholar] [CrossRef]
  8. Rodriguez, J.; Kuncheva, L.; Alonso, C. Rotation Forest: A New Classifier Ensemble Method. IEEE Trans. Pattern Anal. Mach. Intell. 2006, 28, 1619–1630. [Google Scholar] [CrossRef]
  9. Chen, Y.C.; Ha, H.; Kim, H.; Ahn, H. Canonical Forest. Comput. Stat. 2014, 29, 849–867. [Google Scholar] [CrossRef]
  10. Rainforth, T.; Wood, F. Canonical Correlation Forests. arXiv 2015, arXiv:1507.05444. [Google Scholar] [CrossRef]
  11. Han, S.; Kim, H.; Lee, Y.S. Double random forest. Mach. Learn. 2020, 109, 1569–1586. [Google Scholar] [CrossRef]
  12. Geurts, P.; Ernst, D.; Wehenkel, L. Extremely randomized trees. Mach. Learn. 2006, 63, 3–42. [Google Scholar] [CrossRef]
  13. Louppe, G.; Geurts, P. Ensembles on Random Patches. In Machine Learning and Knowledge Discovery in Databases; Flach, P.A., De Bie, T., Cristianini, N., Eds.; Springer: Berlin/Heidelberg, Germany, 2012; pp. 346–361. [Google Scholar]
  14. Xu, B. Classifying Very High-Dimensional Data with Random Forests Built from Small Subspaces. Int. J. Data Warehous. Min. 2012, 8, 44–63. [Google Scholar] [CrossRef]
  15. Cabrera, J.; Lee, Y.S.; Dhammika, A. Enriched random forests. Bioinformatics 2008, 24, 2010–2014. [Google Scholar] [CrossRef] [PubMed]
  16. Li, X.; Ng, M.K.; Huang, J.Z.; Wu, Q.; Ye, Y. Stratified sampling for feature subspace selection in random forests for high dimensional data. Pattern Recognit. 2013, 46, 769–787. [Google Scholar] [CrossRef]
  17. Nguyen, T.T.; Huang, J.Z.; Nguyen, T.T.; Wang, Q. An efficient random forests algorithm for high dimensional data classification. Adv. Data Anal. Classif. 2018, 12, 953–972. [Google Scholar] [CrossRef]
  18. Giacinto, G.; Roli, F. Design of effective multiple classifier systems by clustering of classifiers. Pattern Recognit. Lett. 2001, 24, 1959–1969. [Google Scholar]
  19. Kuncheva, L.I.; Whitaker, C.J. Measures of Diversity in Classifier Ensembles and Their Relationship with the Ensemble Accuracy. Mach. Learn. 2003, 51, 181–207. [Google Scholar] [CrossRef]
  20. Kohavi, R.; Wolpert, D.H. Bias plus variance decomposition for zero-one loss functions. In Proceedings of the 13th International Conference on Machine Learning, Bari, Italy, 3–6 July 1996; Morgan Kaufmann: Cambridge, MA, USA, 1996; pp. 275–283. [Google Scholar]
  21. Wilson, E.B.; Hilferty, M.M. The Distribution of Chi-Square. Proc. Natl. Acad. Sci. USA 1931, 17, 684–688. [Google Scholar] [CrossRef]
  22. Probst, P.; Wright, M.N.; Boulesteix, A.L. Hyperparameters and tuning strategies for random forest. WIREs Data Min. Knowl. Discov. 2019, 9, e1301. [Google Scholar] [CrossRef]
  23. Loh, W.Y. Improving the precision of classification trees. Ann. Appl. Stat. 2009, 3, 1710–1737. [Google Scholar] [CrossRef]
  24. Dua, D.; Graff, C. UCI Machine Learning Repository. University of California, Irvine, School of Information and Computer Sciences. 2019. Available online: http://archive.ics.uci.edu/ml (accessed on 1 January 2024).
  25. Lim, T.S.; Loh, W.Y.; Shih, Y.S. A comparison of prediction accuracy, complexity, and training time of thirty-three old and new classification algorithms. Mach. Learn. 2000, 40, 203–228. [Google Scholar] [CrossRef]
  26. Heinz, G.; Peterson, L.J.; Johnson, R.W.; Kerk, C.J. Exploring Relationships in Body Dimensions. J. Stat. Educ. 2003, 11. Available online: https://jse.amstat.org/v11n2/datasets.heinz.html (accessed on 9 December 2025). [CrossRef]
  27. Kim, H.; Kim, H.; Moon, H.; Ahn, H. A weight-adjusted voting algorithm for ensembles of classifiers. J. Korean Stat. Soc. 2011, 40, 437–449. [Google Scholar] [CrossRef]
  28. Peng, R.D.; Hengartner, N.W. Quantitative Analysis of Literary Styles. Am. Stat. 2002, 56, 175–185. [Google Scholar] [CrossRef]
  29. Kim, H.; Loh, W.Y. Classification Trees With Unbiased Multiway Splits. J. Am. Stat. Assoc. 2001, 96, 589–604. [Google Scholar] [CrossRef]
  30. Alcalá-Fdez, J.; Fernandez, A.; J. Luengo, J.D.; García, S.; Sánchez, L.; Herrera, F. KEEL Data-Mining Software Tool: Data Set Repository, Integration of Algorithms and Experimental Analysis Framework. J. -Mult.-Valued Log. Soft Comput. 2011, 17, 255–287. [Google Scholar]
  31. Chen, T.; Guestrin, C. XGBoost: A scalable tree boosting system. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; ACM Press: San Francisco, CA, USA, 2016; pp. 785–794. [Google Scholar] [CrossRef]
  32. Dorogush, A.V.; Gulin, A.; Gusev, G.; Kazeev, N.; Prokhorenkova, L.O.; Vorobev, A. Fighting biases with dynamic boosting. arXiv 2017, arXiv:1706.09516. [Google Scholar]
  33. Breiman, L. Bagging predictors. Mach. Learn. 1996, 24, 2010–2014. [Google Scholar] [CrossRef]
  34. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  35. McGinnis, W.D.; Siu, C.; S, A.; Huang, H. Category Encoders: A scikit-learn-contrib package of transformers for encoding categorical data. J. Open Source Softw. 2018, 3, 501. [Google Scholar] [CrossRef]
  36. Raschka, S. MLxtend: Providing machine learning and data science utilities and extensions to Python’s scientific computing stack. J. Open Source Softw. 2018, 3, 638. [Google Scholar] [CrossRef]
  37. Rey, D.; Neuhäuser, M. Wilcoxon-Signed-Rank Test. In International Encyclopedia of Statistical Science; Springer: Berlin/Heidelberg, Germany, 2011; pp. 1658–1659. [Google Scholar] [CrossRef]
Figure 1. Decision tree examples. The feature inside the node is the split variable and the features in brackets next to the node represent candidate features.
Figure 1. Decision tree examples. The feature inside the node is the split variable and the features in brackets next to the node represent candidate features.
Symmetry 18 00073 g001
Figure 2. Illustrative examples of decision trees used for dissimilarity evaluation. Each variable is indicated with a different shading pattern.
Figure 2. Illustrative examples of decision trees used for dissimilarity evaluation. Each variable is indicated with a different shading pattern.
Symmetry 18 00073 g002
Figure 3. Empirical scalability of HRF with respect to sample size, feature size, and ensemble size.
Figure 3. Empirical scalability of HRF with respect to sample size, feature size, and ensemble size.
Symmetry 18 00073 g003
Figure 4. Box plot of feature depths.
Figure 4. Box plot of feature depths.
Symmetry 18 00073 g004
Figure 5. Feature importances by method.
Figure 5. Feature importances by method.
Symmetry 18 00073 g005
Figure 6. Dissimilarity of trees by ensemble methods. Each line represents the result of 100 repeated experiments.
Figure 6. Dissimilarity of trees by ensemble methods. Each line represents the result of 100 repeated experiments.
Symmetry 18 00073 g006
Figure 7. Accuracy differences between HRF and bagging, and between HRF and RF, based on the proportion of noise features. A positive value indicates that the HRF model exhibits superior accuracy.
Figure 7. Accuracy differences between HRF and bagging, and between HRF and RF, based on the proportion of noise features. A positive value indicates that the HRF model exhibits superior accuracy.
Symmetry 18 00073 g007
Figure 8. Accuracy differences between HRF and RF across various ( α , β ) combinations and proportions of informative features.
Figure 8. Accuracy differences between HRF and RF across various ( α , β ) combinations and proportions of informative features.
Symmetry 18 00073 g008
Figure 9. Feature-depth distributions under RF and HRF in the presence of a highly correlated surrogate ( ρ = 0.9 between X 1 and X 2 ).
Figure 9. Feature-depth distributions under RF and HRF in the presence of a highly correlated surrogate ( ρ = 0.9 between X 1 and X 2 ).
Symmetry 18 00073 g009
Figure 10. Box plots for the accuracy difference between HRF and others.
Figure 10. Box plots for the accuracy difference between HRF and others.
Symmetry 18 00073 g010
Figure 11. A scatter plot of the optimal ( α , β ) values for each of the 52 benchmark datasets, with the dataset names shown next to the points.
Figure 11. A scatter plot of the optimal ( α , β ) values for each of the 52 benchmark datasets, with the dataset names shown next to the points.
Symmetry 18 00073 g011
Table 1. Procedure lfor calculating feature weights in the toy example presented in Figure 1.
Table 1. Procedure lfor calculating feature weights in the toy example presented in Figure 1.
(a) feature depth of h 1 ( x ) and h 2 ( x ) when β = 1
d X 1 X 2 X 3 X 4 X 5
d 1 01244
d 2 44201
(b) cumulative feature depth when α = 0.5
D X 1 X 2 X 3 X 4 X 5
D 1 01244
D 2 44.5323
(c) feature weights: w 2 for h 2 ( x ) and w 3 for h 3 ( x )
w X 1 X 2 X 3 X 4 X 5
w 2 0/111/112/114/114/11
w 3 4/16.54.5/16.53/16.52/16.53/16.5
Table 2. Procedure to calculate dissimilarity between trees.
Table 2. Procedure to calculate dissimilarity between trees.
(a) Dominance
X 1 X 2 X 3 X 4 X 5
I 1 31200
I 2 30210
I 3 21043
I 4 12032
(b) χ 2 ( df ) test statistics
χ 2 ( df ) h 1 h 2 h 3 h 4
h 1 -2.000(3)8.747(4)8.215(4)
h 2 -7.467(4)7.875(4)
h 3 -0.797(4)
h 4 -
(c) Dissimilarity
D S h 1 h 2 h 3 h 4
h 1 -−0.1921.5001.386
h 2 -1.2171.310
h 3 -−1.040
h 4 -
Table 3. Description of data sets.
Table 3. Description of data sets.
Data SetSizeInputsClasses# MissingSource
aba *417782 UCI (Abalone)
ail13,750122 [23]
aus *690142 UCI (Australian credit approval)
bal62553 UCI (Balance scale)
ban167252 UCI (Bank note authentication)
bcw68392 [25]
bld34662 UCI (BUPA liver disorders)
blo74852 UCI (Blood transfusion center)
bod507242 [26]
bos *506133 UCI (Boston housing)
bre *69910216UCI (Wisconsin Breast Cancer)
cir1000102 R library mlbench (Circle in as square)            
cmc *147393 UCI (Contraceptive method choice)
cre *69015267UCI (Credit approval)
der358346 UCI (Dermatology)
dia76882 [23]
dna *3187603 R package mlbench (StatLog DNA)
eco33674 [23]
ger *1000202 UCI (German credit)
gla21496 UCI (Glass)
hea *271132 UCI (StatLog heart disease)
hil6061002 UCI (Hil-valley)
imp *20525659UCI (Auto imports)
int1000102 [27]
ion351342 UCI (Ionosphere)
iri15043 UCI (Iris)
led6000710 UCI (LED display)
lib3599015 UCI (Libra Movement)
lit2329699 [28]
mam96152162UCI (Mammographic mass)
mar8777410 [23]
mus8124222 UCI (Mushroom)
pid53272 UCI (PIMA Indian diabetes)
pks195222 UCI (Parkinsons)
pov97666[29]
rng1000102 R library mlbench (Ringnorm)
sat6435366 UCI (StatLog satellite image)
seg2310197 UCI (Image segmentation)
smo *285583 UCI (Attitude towards smoking restrictions)
snr208602 R library mlbench (Sonar)
spa4601572 UCI (Spambase)
spe267442 UCI (SPECETF heart)
tae *15153 KEEL (Teaching) [30]
thy1000102 UCI (Thyroid disease)
trn1000102 R library mlbench (Three norm)
twn1000102 R library mlbench (Two norm)
veh846184 UCI (StatLog vehicle silhouette)
vol *15126640[23]
vot *435162 UCI (Congressional voting records)
vow9901011 UCI (Vowel recognition)
wav3600213 UCI (Wave)
yea1484810 KEEL (Yeast) [30]
zoo101167 R library mlbench (zoo)
* Categorical features have been encoded.
Table 4. Ensemble methods and their associated hyper-parameters. The default values are in bold.
Table 4. Ensemble methods and their associated hyper-parameters. The default values are in bold.
MethodsHyper-Parameter TuningSource
Baggingmax_features = [0.6, …, 1.0][33,34]
RFmax_features = [sqrt, log2, None][1,34]
ExtTreesmax_depth = [4, 6, None][12,34]
min_samples_split = [2, 4]
Random Patchesmax_features = [0.6, …, 1.0][13,34]
XGBmax_depth = [4, 6, 8][31]
min_child_weight = [1, 2]
CatBoostdefault[32]
Table 5. Accuracy of ensemble methods: ‘+’ indicates that HRF is significantly better, while a ‘−’ indicates that HRF is significantly worse at the 0.05 significance level.
Table 5. Accuracy of ensemble methods: ‘+’ indicates that HRF is significantly better, while a ‘−’ indicates that HRF is significantly worse at the 0.05 significance level.
DataBaggingRFExtTreesRndPatchXGBCatBoostHRF
aba0.7743+0.7777+0.7790+0.7722+0.7711+0.78140.7818
ail0.88050.8803−0.8767+0.8728+0.8838−0.8848−0.8798
aus0.8692+0.8676+0.8647+0.8692+0.8698+0.8671+0.8765
bal0.87060.8467+0.8810−0.86520.8868−0.8836−0.8662
ban0.9918+0.9918+0.9988−0.9898+0.99450.99420.9940
bcw0.9717+0.9704+0.9714+0.9719+0.965+0.971+0.9746
bld0.7089+0.7175+0.7146+0.6994+0.7045+0.7159+0.7375
blo0.76570.7450+0.77130.77090.7551+0.7918−0.7696
bod0.9513+0.9566+0.9687−0.9561+0.96360.9711−0.9621
bos0.7911+0.7921+0.7796+0.7899+0.7853+0.7887+0.8005
bre0.9656+0.9670+0.96850.97000.9581+0.9684+0.9701
cir0.8263+0.8269+0.82910.8248+0.8781−0.8618−0.8340
cmc0.5191+0.5167+0.5227+0.5170+0.5372−0.53180.5300
cre0.8731+0.8742+0.8656+0.8726+0.8678+0.8746+0.8825
der0.9677+0.9744+0.98170.9697+0.9602+0.9753+0.9798
dia0.7562+0.7624+0.7647+0.7548+0.7430+0.77180.7706
dna0.9621+0.96780.96700.9654+0.9663+0.9656+0.9679
eco0.9258+0.9368+0.94420.9246+0.9154+0.9308+0.9450
ger0.7648+0.7678+0.7648+0.7611+0.7589+0.77450.7728
gla0.7438+0.7581+0.7403+0.7581+0.7319+0.7541+0.7688
hea0.8195+0.8286+0.8323+0.8200+0.8079+0.8328+0.8407
hil0.5674+0.5625+0.5536+0.5719+0.57440.5118+0.5781
imp0.78850.7810+0.7351+0.79110.79870.78490.7895
int0.7455−0.59460.5571+0.7081−0.8469−0.8050−0.5994
ion0.9310+0.93450.93940.9314+0.9274+0.93280.9375
iri0.9449+0.9493+0.95510.9458+0.9378+0.9489+0.9511
led0.7340+0.7315+0.7334+0.7334+0.7323+0.7311+0.7354
lib0.7748+0.7811+0.8141−0.7757+0.7119+0.7809+0.8006
lit0.8808+0.8835+0.8815+0.8810+0.9055−0.9033−0.8878
mam0.8295−0.7938+0.82270.8303−0.8081+0.8350−0.8217
mar0.58120.5400+0.5815−0.5814−0.5751+0.5678+0.5812
pid0.7596+0.7706+0.7618+0.7613+0.7390+0.77720.7767
pks0.8738+0.8969+0.8679+0.8986+0.91280.9166−0.9055
pov0.6090+0.61170.5869+0.6021+0.62280.63860.6228
rng0.91790.9085+0.9374−0.92590.91690.9231−0.9161
sat0.9116+0.9139+0.91440.9136+0.91530.9044+0.9151
seg0.97770.97780.97750.9758+0.9822−0.97770.9783
smo0.6986−0.6511+0.6983−0.6938−0.662+0.6480+0.6835
snr0.7910+0.8216+0.8265+0.8119+0.84320.84160.8439
spa0.9499+0.9518+0.9548−0.9495+0.9537−0.95280.9527
spe0.7990+0.8032+0.8040+0.8008+0.7953+0.808+0.8155
tae0.6138+0.6333+0.6093+0.6222+0.6573−0.63420.6440
thy0.9965−0.9963−0.9811+0.9928+0.9953+0.99600.9961
trn0.8538+0.8596+0.8736−0.8604+0.8556+0.8604+0.8659
twn0.9627+0.9623+0.9672−0.96400.9587+0.96430.9657
veh0.7477+0.7527+0.7452+0.7528+0.76450.76450.7589
vol0.8625+0.90360.8893+0.90470.90410.9073−0.9035
vot0.9548+0.9548+0.9551+0.9511+0.95680.9517+0.9594
vow0.9379+0.9467+0.9733−0.9405+0.8936+0.9421+0.9551
wav0.8465+0.8466+0.85410.8475+0.8474+0.8462+0.8516
yea0.5933+0.6098+0.6062+0.5851+0.5808+0.5873+0.6201
zoo0.94730.94470.9393+0.95130.94470.94270.9507
W/T/L(40/8/4)(43/7/2)(29/12/11)(40/7/5)(30/13/9)(24/17/11) 
W: # of dataset HRF wins, T: # of dataset tied, L: # of dataset HRF loses.
Table 6. Summary of comparisons among methods.
Table 6. Summary of comparisons among methods.
 BaggingRFExtTreeRndPatchXGBCatBoostHRF
Bagging 35(22)33(25)31(12)24(19)37(30)45(40)
RF17(8) 30(19)22(11)29(23)32(22)48(43)
ExtTree20(12)22(14) 17(14)22(15)30(21)35(29)
RndPatch22(8)31(18)35(26) 23(18)37(27)43(40)
XGB28(22)24(21)30(24)29(18) 31(26)35(30)
CatBoost15(7)20(8)22(19)15(6)22(11) 33(24)
HRF7(4)4(2)17(11)9(5)18(9)19(11)
In each cell, the results are summarized as ‘ a ( b ) ’, where a indicates the number of datasets where the method in the column outperforms the method in the row, and b indicates the number of datasets where this difference is statistically significant according to the one-sided paired t-test.
Table 7. The dominance rank of the methods determined by the significant difference from the results in Table 6.
Table 7. The dominance rank of the methods determined by the significant difference from the results in Table 6.
MethodsWinsLossesDominance Rank
HRF20642164
CatBoost1377562
ExtTree12410519
RF85126−41
XGB95141−46
RndPatch66137−71
Bagging61148−87
Table 8. The estimated proportions of noise features.
Table 8. The estimated proportions of noise features.
Data Setp_noiseData Setp_noiseData Setp_noise
aba0.63ger0.65sat0.92
ail0.75gla0.44seg0.79
aus0.79hea0.69smo0.7
bal0.25hil0.6snr0.75
ban0.75imp0.78spa0.79
bcw0.89int0.8spe0.7
bld0.67ion0.85tae0.71
blo0.5iri0.75thy0.81
bod0.79led0.43trn0.4
bos0.71lib0.68twn0.6
bre0.89lit0.77veh0.67
cir0.7mam0.6vol0.71
cmc0.85mar0.75vot0.94
cre0.73pid0.57vow0.7
der0.82pks0.73wav0.71
dia0.75pov0.67yea0.5
dna0.9rng0.5zoo0.69
eco0.71    
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kim, Y.E.; Kim, S.Y.; Kim, H. Asymmetric Feature Weighting for Diversity-Enhanced Random Forests. Symmetry 2026, 18, 73. https://doi.org/10.3390/sym18010073

AMA Style

Kim YE, Kim SY, Kim H. Asymmetric Feature Weighting for Diversity-Enhanced Random Forests. Symmetry. 2026; 18(1):73. https://doi.org/10.3390/sym18010073

Chicago/Turabian Style

Kim, Ye Eun, Seoung Yun Kim, and Hyunjoong Kim. 2026. "Asymmetric Feature Weighting for Diversity-Enhanced Random Forests" Symmetry 18, no. 1: 73. https://doi.org/10.3390/sym18010073

APA Style

Kim, Y. E., Kim, S. Y., & Kim, H. (2026). Asymmetric Feature Weighting for Diversity-Enhanced Random Forests. Symmetry, 18(1), 73. https://doi.org/10.3390/sym18010073

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop