Next Article in Journal
Analysis of Price Dynamic Competition and Stability in Cross-Border E-Commerce Supply Chain Channels Empowered by Blockchain Technology
Previous Article in Journal
Advances in Quantum Computation in NISQ Era
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Aggregation in Ill-Conditioned Regression Models: A Comparison with Entropy-Based Methods

1
Center for Research and Development in Mathematics and Applications (CIDMA), Agueda School of Technology and Management, 3750-127 Agueda, Portugal
2
Department of Mathematics, University of Aveiro, 3810-193 Aveiro, Portugal
3
Department of Physics, University of Aveiro, 3810-193 Aveiro, Portugal
4
Institute for Nanostructures, Nanomodelling and Nanofabrication (i3N), 3810-193 Aveiro, Portugal
*
Author to whom correspondence should be addressed.
Entropy 2025, 27(10), 1075; https://doi.org/10.3390/e27101075
Submission received: 23 July 2025 / Revised: 2 October 2025 / Accepted: 10 October 2025 / Published: 16 October 2025

Abstract

Despite the advances on data analysis methodologies in the last decades, most of the traditional regression methods cannot be directly applied to large-scale data. Although aggregation methods are especially designed to deal with large-scale data, their performance may be strongly reduced in ill-conditioned problems (due to collinearity issues). This work compares the performance of a recent approach based on normalized entropy, a concept from information theory and info-metrics, with bagging and magging, two well-established aggregation methods in the literature, providing valuable insights for applications in regression analysis with large-scale data. While the results reveal a similar performance between methods in terms of prediction accuracy, the approach based on normalized entropy largely outperforms the other methods in terms of precision accuracy, even considering a smaller number of groups and observations per group, which represents an important advantage in inference problems with large-scale data. This work also alerts for the risk of using the OLS estimator, particularly under collinearity scenarios, knowing that data scientists frequently use linear models as a simplified view of the reality in big data analysis, and the OLS estimator is routinely used in practice. Beyond the promising findings of the simulation study, our estimation and aggregation strategies show strong potential for real-world applications in fields such as econometrics, genomics, environmental sciences, and machine learning, where data challenges such as noise and ill-conditioning are persistent.

1. Introduction

Large-scale data or big data usually refers to large and complex collections of datasets, which are difficult to process using database management tools or traditional data analysis methodologies. One of the biggest challenges associated with processing big data is retaining relevant statistical information while being computationally efficient. Most large-scale data are also inhomogeneous, meaning they are neither i.i.d. (independent and identically distributed) nor stationary observations from a statistical distribution, as well as noisy, dynamic and inter-related, leading to possible severe ill-conditioning problems.
One way to address the computational burden is through aggregation methods that can be easily implemented in parallel. These consist of the following steps: (i) select G groups from the large-scale data (these groups may be overlapping and not include the entirety of the observations); (ii) obtain the vector of estimates, β ^ g , using standard techniques (e.g., ordinary least squares, ridge, generalized maximum entropy) for each group g, g = 1 , 2 , , G ; and, finally, (iii) aggregate the ensemble of vectors of estimates into a single and final vector of estimates, β ^ a g g .
Costa and Macedo [1] introduced neagging, a new approach for aggregation based on normalized entropy, a concept from information theory and info-metrics [2,3,4,5], and its precision accuracy was compared only with bootstrap aggregating or mean aggregation (bagging) [6] through a simple simulation study. This work substantially improves the study of [1] by presenting results from a large simulation study with 48 different scenarios, where a two-step procedure was adopted: estimation is first performed using OLS or a maximum-entropy-based technique, and the resulting group estimates are then aggregated through bagging, magging (maximin aggregating [7], a recent and powerful aggregation method), or neagging. The purpose of this approach is to enhance the tractability of large-scale data analysis by decomposing the estimation of a potentially massive dataset into the estimation of multiple smaller subsets, which are then recombined into a final aggregated estimate. Precision and prediction accuracies of the combined two-step procedures are compared.
The remaining article is organized as follows: maximum entropy estimation and the aggregation methods discussed in the work are briefly presented in the next section; then follows the simulation study and the discussion of the results; finally, in the last section the most relevant conclusions are provided.

2. Methods

In this section are briefly presented two maximum entropy estimators, namely, the generalized maximum entropy (GME) estimator and the weighted generalized maximum entropy (W-GME) estimator, and the three aggregation methods discussed in this work, namely, the bootstrap aggregating or mean aggregation (bagging), the maximin aggregating (magging), and the normalized entropy aggregating (neagging).

2.1. Maximum Entropy Estimation

Info-metrics, by generalizing Jaynes’ maximum entropy (ME) principle [2,3]—which itself builds on Shannon’s concept of entropy [8]—relies fundamentally on the notions of information, uncertainty, and entropy. In this framework, information refers to the meaningful content, context, and interpretation of data; uncertainty arises from propositions or outcomes not known with certainty and is represented by probability distributions; and entropy quantifies the expected average information from observations, serving as a measure of uncertainty. The ME principle was taken by [2,3,4] as the basis for transforming the information in the data into a probabilistic distribution that reflects the uncertainty about individual outcomes, and was used by the same authors to develop analytical and empirical methods for recovering the unobservable parameters of the pure linear inverse problem,
y = X p ,
where y is a vector ( N × 1 ) of noisy observations, X is a non-invertible matrix ( N × K ) with N < K , and p is the vector ( K × 1 ) of unknown probabilities. From all the probability distributions that satisfy model (1), the ME principle allows us to pick an unambiguous estimate of p by choosing the probability distribution that maximizes Shannon´s entropy,
H ( p ) = k = 1 K p k ln p k = p ln p ,
subject to the model consistency restrictions, y = X p , and the additivity restriction, p 1 = 1 (note that p is the same as p T , representing the transpose of p ). Given both these constraints, the ME estimator is formally given by
argmax p p ln p .
The ME principle provides a tool to make the best prediction using only the available information. The maximization problem in (3) does not have a closed-form solution, which means that the solution must be found with numerical optimization procedures. Using the Lagrange multipliers method (Lagrangian function and first-order optimality conditions), it follows that
p ^ k = exp ( x k λ ^ ) k = 1 K exp ( x k λ ^ ) ,
where x k is a ( N × 1 ) vector corresponding to the kth column of X and λ ^ is a ( N × 1 ) vector of estimated Lagrange multipliers on the model consistency restrictions. Jaynes’ maximum entropy framework allows us to approach the inverse problem as an inference problem, using optimization to derive a probability distribution that best represents the information in the data and the uncertainty about possible outcomes.
Let us now turn our attention to the linear regression model that is usually represented as
y = X β + e ,
where, as before, y is the vector ( N × 1 ) of noisy observations, X is a known design matrix ( N × K ) of explanatory variables, β is the vector ( K × 1 ) of unknown parameters to be estimated, and e is a ( N × 1 ) vector of random disturbances (errors), usually assumed to have a conditional expected value of zero and representing spherical disturbances, i.e., E [ e | X ] = 0 and E [ ee | X ] = σ 2 I , where I is the ( N × N ) identity matrix and σ 2 is the error variance.
The ordinary least squares (OLS) estimator of β in the model (5), given by
β ^ : = ( X X ) 1 X y ,
is probably the most known and widely used estimator in linear regression. However, in the presence of collinearity (also known as ill-conditioning, referring to a near-linear relationship between two or more regressors), and with its increasing severity, the vector β ^ obtained by the OLS estimator can be expected to be farther from the vector β ; e.g., [9,10,11].
The ME principle can be extended to estimate the linear regression model in this context, and to deal with ill-posed problems, where collinearity is included, A. Golan and coauthors [4] proposed a reparameterization of the model in (5) as
y = XZp + Vw ,
where
β = Zp = z 1 0 0 0 z 2 0 0 0 z K p 1 p 2 p K ,
with Z a ( K × K M ) matrix of support spaces (closed and bounded intervals in which each parameter is restricted to belong) and p a ( K M × 1 ) vector of unknown probabilities to be estimated, and
e = Vw = v 1 0 0 0 v 2 0 0 0 v N w 1 w 2 w N ,
with V a ( N × N J ) matrix of support spaces (closed and bounded intervals in which each error is restricted to belong) and w a ( N J × 1 ) vector of unknown probabilities to be estimated. In this context, each β k , k = 1 , 2 , , K , and each e n , n = 1 , 2 , , N , are viewed as expected values of discrete random variables, with M 2 and J 2 as possible outcomes, within the lower and upper bounds of the corresponding support spaces. Thus, considering the linear regression model specified in (5) and assuming independence between p and w (additivity property of Shannon entropy), the generalized maximum entropy (GME) estimator is given by
argmax p , w p ln p w ln w ,
subject to the model constraints
y = XZp + Vw ,
and the additivity constraints for p and w , respectively,
1 K = ( I K 1 M ) p , 1 N = ( I N 1 J ) w ,
where ⊗ represents the Kronecker product, 1 is a column vector of ones with the specified dimension, and I is an identity matrix with the specified dimension. It is important to note that while the ME estimator was originally designed for pure linear inverse problems (with N < K ), the GME estimator broadens the scope of the method and remains valid regardless of the relative sizes of N and K, thereby encompassing underdetermined, overdetermined, and exactly identified problems. Using numerical optimization techniques, the GME estimator generates the optimal probability vectors p ^ and w ^ that are used to obtain point estimates of the parameters and the errors through the reparameterizations (8) and (9), respectively. For example, analogously to the solution of the ME estimator in (4), the formal solution of the GME estimator for p is given by
p ^ k m = exp z k m x k λ ^ m = 1 M exp z k m x k λ ^ ,
where x k is a ( N × 1 ) vector corresponding to the kth column of X and λ ^ is a ( N × 1 ) vector of estimated Lagrange multipliers on the model constraints (11). Additional details can be found in [4] (pp. 90–93).
Regarding the specification of the support spaces in Z , for the parameters of the model, the bounds are usually obtained based on theoretical constraints or information from previous research. Nevertheless, wide symmetric supports about zero, usually with five equally spaced points ( M = 5 ) between the lower and upper bounds, should be used whenever there is no a priori information available. The bounds for the support spaces in V , for the errors of the model, are usually obtained by the three-sigma rule [12], considering the standard deviation of the noisy observations (observed dependent/response variable) rounded up to the nearest integer, usually with three points ( J = 3 ) symmetric about zero. Additional details about support spaces can be found in [4,5,13], and an illustrative example of the matrix structure of the GME estimator is given in Appendix A, also discussing the reason to consider two different scenarios for supports in Section 3.
X. Wu [14] proposed the weighted generalized maximum entropy estimator with a data-driven weight, where the objective function in (10) is updated as
argmax p , w ( 1 γ ) p ln p γ w ln w ,
where γ ( 0 , 1 ) defines the weights assigned to each entropy component in the objective function, and it is selected by some kind of cross-validation through the minimization of a given loss function (e.g., the minimization of the sum of the squared prediction errors). Under different simulated scenarios, the author illustrates that the W-GME estimator provides superior performance than the GME estimator [14].

2.2. Aggregation Methods

The bagging method [6] simply averages, with uniform weights, over the ensemble of estimates from the groups, and the global estimate is given by
β ^ a g g : = g = 1 G w g β ^ g ,
where w g = 1 G for all g = 1 , 2 , , G . It is a simple method where the estimates are obtained from bootstrap (random with replacement) samples. However, it is not considered in this work the case of a single learning set and the need to take repeated bootstrap samples from it as in [6].
The magging method [7] builds an aggregate estimator as a convex combination of group-specific estimators, β ^ g . The weights w g are chosen not to minimize prediction error in an average sense like in bagging, but to minimize the l 2 -norm of the fitted values across groups, which can be interpreted as maximizing the “worst-case” explained variance across the different groups. This ensures that the final aggregate model captures the effects that are consistently present in all groups, rather than the effects that are strong in some groups but absent in others. Explicitly, the global estimate is given by
β ^ a g g : = g = 1 G w g β ^ g ,
where
w : = argmin w W g = 1 G w g y ^ g ,
with the constraint W = { w : min g   w g 0   and   g = 1 G w g = 1 } . As mentioned above, the weights are chosen as a convex combination to minimize the l 2 -norm of the vector of fitted values, y ^ g = X β ^ g , and this method can be easily implemented with quadratic programming. If the solution is not unique, it is considered the solution with lowest l 2 -norm of the weight vector among all solutions. The idea behind magging is that if an effect is common across all groups, then it cannot be “averaged away’’ by searching for a specific combination of the weights [7].
The neagging method, the approach for aggregation based on normalized entropy [1,15], consists of weighting the estimate obtained from the GME estimator [4,5,16] for each individual group according to the amount of information in that group, because if it is true that groups obtained from random sampling carry information about the whole dataset, they are of course not equally informative. To measure the information content in a system with K states, the normalized entropy, S ( p ^ ) can be used, as defined by
S ( p ^ ) : = k p k ^ ln p k ^ ln K ,
where S ( p ^ ) [ 0 , 1 ] , and ln ( K ) represents maximum uncertainty, which is the entropy level of a uniform distribution with K outcomes [4]. The normalized entropy aggregation scheme, neagging, is based on identifying the information content of a given group, g, through the calculation of the normalized entropy associated with the estimate of that group,
S ( p ^ ) g : = p ^ ln p ^ K ln M ,
where M is number of support points in the specific group g, and the denominator, K ln M , represents maximum entropy, necessary for normalization. The weights in the aggregation scheme are calculated such that w g 1 S ( p ^ ) g and g = 1 G w g = 1 . The global estimate is then given by
β ^ a g g : = g = 1 G w g β ^ g .
The resulting final estimates will be a weighted average of the ensemble of vectors with estimates based on the information content of each group, and this method is almost as simple as bagging, with the weights expected to be non-uniform as in magging.

3. Simulation Study

The simulation setup is designed to emulate high-dimensional regression problems with strong collinearity and different error distributions. Such settings are commonly found in econometric modeling, genomics, environmental studies, and machine learning applications. The synthetic datasets reflect these structures in a controlled way, allowing for a focused evaluation of different methods (estimation and aggregation) under ill-conditioning and data heterogeneity scenarios. All the code is developed in MATLAB (R2019b version) [17] by the authors.
The general context of the simulation is as follows: the two collinearity scenarios are intended to illustrate models with low and high collinearity found in practice; the vector of parameters illustrates scenarios “close” to zero, positive and negative, very common in practice; the different error distributions reflect variations in noise; the size of the models is still computationally feasible for the design of the simulations and allows for the possibility of sampling; the two parameter supports for GME and W-GME reflect reduced and high information by the user regarding the model parameters; the supports for the errors, as well as the number of points in the supports, are the usual ones in the literature. Finally, the number of groups and the number of observations per group try to reflect the purpose of aggregation (obtaining accurate information with minimum computational effort and observing a small part of the population under study).

3.1. Simulation Settings

A linear regression model is considered, with a number of observations of 30,000 and a number of explanatory variables of 10 (N = 30,000 and K = 10 ). Two X (30,000 × 10) matrices of explanatory variables are simulated, corresponding to matrices with two different condition numbers, representing two distinct collinearity scenarios: cond( X ) = 10, for cases of near absence of collinearity; and cond( X ) = 20,000, for cases where high collinearity is present in the data. The X matrices were initially constructed using standard normal distributions, and then the singular value decomposition was used to algebraically adjust the matrices to the desired condition numbers. The MATLAB functions available imposed a restriction on the size of the matrices that could be used, and the values N × K = 30,000 × 10 represented the maximum feasible under these conditions. Anyway, the size of the original dataset in the simulation study (whether thousands or millions) is not relevant, as we will always work on estimating parameters in models with small datasets extracted by sampling. The simulation framework assumes a linear relationship between explanatory variables and the dependent variable, with errors following known distributions. These assumptions were chosen deliberately in order to provide a controlled environment where the effects of ill-conditioning, choice of the estimators, and aggregation methods could be systematically isolated and evaluated. However, it is also important to note that data scientists dealing with huge amounts of data frequently use linear models as a simplified view of the reality and these are actually very satisfactory models quite often.
The number of parameters in the regression model will correspond to the number of explanatory variables and, consequently, the parameter vector β will be a ( 10 × 1 ) vector, defined as β = [ 2 , 1 , 3 , 5 , 5 , 3 , 4 , 2 , 1 , 4 ] , not containing the constant term. Furthermore, three distinct types of error are considered for the given datasets: errors modeled by a normal distribution with zero mean and unit standard deviation, e N ( 0 , 1 ) , to represent scenarios with relatively low noise; errors based on a t-Student distribution with three degrees of freedom, e t ( 3 ) , to represent scenarios with significant but moderate noise; and errors following a Cauchy distribution with location parameter zero and scale parameter two, e C ( 0 , 2 ) , to represent scenarios with high noise. In this way, three distinct vectors e (30,000 × 1 ) of random perturbations are obtained, resulting in six vectors y (30,000 × 1 ) of noisy observations, combining the two different condition numbers with the three vectors of random perturbations, to obtain a wide variety of noisy observation vectors with different characteristics.
The reparameterizations imposed by the GME and W-GME estimators are carried out by defining the matrix V , which contains the supports of the errors, with symmetric supports centered at zero, using the 3-sigma rule, where sigma is approximated by the empirical standard deviation of the noisy observations, and the support is always the same for each unknown perturbation. Consequently, the support vectors of the errors will take the form v n = [ 3 σ ^ , 0 , 3 σ ^ ] for all n = 1 , , 30,000, being equally spaced, with J = 3 points used to form these supports. Additionally, two distinct supports for the parameters are considered, both symmetric and centered at zero, so these vectors will also be equally spaced, considering M = 5 points to form these supports (the same for each of the unknown components of the parameter vector of the model). Thus, two distinct matrices Z are defined, containing the supports for the parameters: one with z k = [ 10 , 10 ] for all k = 1 , , 10 , to reflect scenarios where there is some prior knowledge about the range in which the parameters may be found, allowing for a reduction in the amplitude of the parameter supports; and another with z k = [ 200 , 200 ] for all k = 1 , , 10 , to reflect scenarios where the available prior knowledge is insufficient. As a clarification remark, the support vectors are actually given by z k = [ 10 , 5 , 0 , 5 , 10 ] and z k = [ 200 , 100 , 0 , 100 , 200 ] , but, to simplify notation, they will be mentioned henceforth as z k = [ 10 , 10 ] and z k = [ 200 , 200 ] . The objective is to understand (albeit preliminarily, as this is not the main aim of the work) how the amplitude of the parameter supports affects the results of the GME and W-GME estimators.
Completing with the necessary specifications regarding the aggregation methods, random sampling with replacement was performed considering two values for the number of groups, namely, G = 10 and G = 20 , with the aim of analyzing whether increasing the number of groups in each aggregation method causes significant differences in the results obtained, and two values for the number of observations per group, Obs = 50 and Obs = 100 , to assess whether the increase in observations in each group, that is, the increase in information available in each group, produces relevant changes in the performance of each of the aggregation methods.
In summary, a total of 48 scenarios will be analyzed, referring to the combination of the two matrices of explanatory variables (two condition numbers) and the three error vectors (three error distributions), generating a total of six different datasets, associated with the two supports for the parameters, the two values for the number of groups in the aggregation methods, and the two values for the number of observations per group, forming a total of eight variants of structures. Additionally, a key point of this simulation study will also be to study the variance of the results, as there is a random sampling process involved, to understand whether the results obtained are directly associated with the aggregation method used or if it is a consequence of a less informative sampling process. For this purpose, Monte Carlo experiments with 10 replicas for each scenario are conducted, with the objective of exploring the sampling behavior of the situations under study.
Once the datasets are obtained, the model coefficients are estimated using the previously mentioned aggregation techniques, bagging, magging, and neagging, varying the estimators used in each aggregation procedure, namely, the OLS, GME, and W-GME estimators. For each of the 48 Monte Carlo experiments related to the scenarios under analysis, the regression coefficient estimates of each considered aggregation method are calculated in each of the 10 replicas, using the different estimators discussed. These estimates will be defined by β ^ a g g r , expressing the regression coefficients estimated after aggregation. The final regression coefficient estimates, that is, those obtained after conducting the Monte Carlo experiment, relative to each of the aggregation procedures, using the different estimators considered in this study, are calculated by averaging the 10 β ^ a g g r values obtained from each replica, being referred to as β ^ a g g r ¯ , expressing the final estimated regression coefficients after aggregation (mean of the replicas). Henceforth, when presenting the results of this study, the general abbreviation ‘aggr’ will be replaced by the following abbreviations: bOLS to represent the bagging aggregation procedure with OLS coefficient estimates; bGME to represent the bagging procedure with GME estimates; bW-GME to represent the bagging procedure with W-GME estimates; mOLS to represent the magging procedure with OLS estimates; mGME to represent the magging procedure with GME estimates; mW-GME to represent the magging procedure with W-GME estimates; nGME to represent the neagging procedure with GME estimates; and nW-GME to represent the neagging procedure with W-GME estimates.

3.2. Evaluation Metrics Based on Prediction and Precision Errors

For each of the Monte Carlo experiments related to the scenarios under analysis, the prediction and precision errors associated with each of the β ^ a g g r obtained are calculated for all 10 replicas. The prediction error is derived from the Euclidean norm of the difference between the vector of predicted observations using the model (multiplication of the matrix of explanatory variables by the estimated regression coefficients after aggregation) and the simulated vector of noisy observations, that is, | | X β ^ a g g r y | | . The precision error is calculated through the Euclidean norm of the difference between the vector of estimated regression coefficients after aggregation and the original parameter vector considered in the simulation, that is, | | β ^ a g g r β | | . These two complementary metrics are used to assess the effectiveness of each aggregation strategy: precision error (a standard metric in regression analysis under simulation; reflects the overall stability of the parameter estimates across replications) and prediction error (a standard metric in regression analysis; reflects the practical predictive accuracy of the estimated model).
The prediction error 1 and precision error 1 are obtained by averaging the 10 prediction and precision errors obtained in each replica, thus obtaining | | y ^ y | | and | | β ^ β | | , respectively, and the prediction error 2 and precision error 2 are determined by implementing the formulas for prediction and precision errors, but using the final estimated regression coefficients after aggregation (mean of the replicas), β ^ a g g r ¯ , obtaining the values | | y ^ y | | * and | | β ^ β | | * , respectively.
The variances of the results are calculated using the usual variance formula, that is, the mean of the sum of the squares of the differences between each element and its mean. Therefore, the variance of β ^ aggr ¯ , the variance of the prediction error, and the variance of the precision error are
s β ^ a g g r ¯ 2 = 1 R i = 1 R ( β ^ a g g r i β ^ a g g r ¯ ) 2 ,
s | | y ^ y | | 2 = 1 R i = 1 R ( | | X β ^ a g g r y | | i | | y ^ y | | ) 2 ,
s | | β ^ β | | 2 = 1 R i = 1 R ( | | β ^ a g g r β | | i | | β ^ β | | ) 2 ,
respectively, with R corresponding to the number of replicas, 10. While formal hypothesis testing is not applied in this study, the use of replicas per scenario, variance estimates, and boxplot visualizations offer a preliminary assessment of the stability and dispersion of results across different techniques (estimation and aggregation).

4. Results and Discussion

The results presented in this work aim to highlight the general trends of this simulation study. To better understand and analyze the results obtained, summary tables of precision errors 1 and 2 for all the previously mentioned scenarios are presented in Table 1, Table 2 and Table 3. Data visualization was carried out using MATLAB for the boxplots and R (ggplot2 package) for the heatmaps.
Additionally, comparative graphs of precision errors between the different methods are presented, when the number of groups and the number of observations per group are modified, in order to evaluate the impact of these factors on the studied aggregation methods, with the respective estimators considered. Due to the extent of this simulation study, given the high number of presented scenarios and subsequent vectors of estimates and variances obtained (eight vectors per scenario, relative to each of the aggregation procedures with the different estimators considered), only the results of certain scenarios are highlighted. These scenarios aim to establish a representative set of the total scenarios considered in the study. The remaining are available upon request to the authors.
The values of the prediction errors 1 and 2 resulting from these selected scenarios, and the variances associated with these results, are found in Table 4, Table 5 and Table 6. Note that values presented as 0.00 do not represent a null value, but, rather, a value less than 0.01. Additionally, boxplots of the regression coefficient estimates and precision error for these scenarios are presented for a graphical analysis (among other aspects) of the variances relative to these results.
Analyzing the precision errors obtained in the scenarios e N ( 0 , 1 ) , Table 1, in situations of low collinearity (cond( X ) = 10), it is observed that | | β ^ β | | is generally lower when one of the estimators based on the ME principle is applied. Although the OLS estimator usually performs well in data with low collinearity, the estimates obtained by procedures using the GME and W-GME estimators stand out compared to the estimates obtained by the bOLS and mOLS procedures. The bGME, bW-GME, mGME, and mW-GME are the methods that distinguish themselves by their good performance. However, the performance of the bagging method with the GME and W-GME estimators tends to worsen with the increase in the width of the coefficient supports, highlighting the better performance of magging with the use of these estimators. Additionally, the performance of mW-GME generally surpasses that of mGME in scenarios where z k = [ 200 , 200 ] . The nGME also presents one of the best performances, but only in the situation where the support for the parameters is narrower ( z k = [ 10 , 10 ] ).
Now, analyzing a scenario with high collinearity (cond( X ) = 20,000), a significant increase in | | β ^ β | | is observed in aggregation methods using the OLS estimator, while the methodologies that stood out previously in the scenario with low collinearity continue to behave similarly. This indicates that the presence of collinearity does not seem to affect the results provided by the previously mentioned aggregation methods, namely, those using estimators based on the ME principle. As the performance of the OLS estimator is generally affected by the degree of collinearity in the data, and the GME estimator (and, by extension, the W-GME estimator) is suitable for handling data affected by collinearity, these results are not unexpected. Again, the magging aggregation method with the GME and W-GME estimators stands out for its performance, and the behaviors observed for low collinearity scenarios are also observed in these cases. However, with the increase in G and Obs, the performance of bGME and bW-GME improves once more.
Note that, when evaluating the quality of the obtained estimates (precision error 2), the scenarios with a wider range of parameter supports, a larger number of groups, and a greater number of observations exhibit the most suitable estimates for the regression coefficients of the presented problem, with the use of bGME, bW-GME, and nGME, regardless of the degree of collinearity present (but more visibly with higher collinearity).
When e t ( 3 ) , the performance of bGME and bW-GME decreases compared to the previous case. From Table 2, it can be seen that the methods that consistently present the lowest | | β ^ β | | are mGME and mW-GME. The bagging and neagging aggregation methods generally show a decrease in their performance compared to the previous table, suggesting that these methods are more affected by the presence of larger perturbations in the data. The effects caused by collinearity and the existence of more adverse perturbations are noticeable in the aggregation methods that apply the OLS estimator.
Finally, analyzing the case where e C ( 0 , 2 ) , Table 3, the conclusions are similar to the case e t ( 3 ) . The bagging and neagging aggregation methods show poor performance, especially when using the OLS estimator, and the magging aggregation method is recognized for its performance when using the GME and W-GME estimators. However, mW-GME reveals a predominantly dominant behavior in terms of precision error compared to mGME (with similar performance or insignificant differences in cases where it does not stand out), particularly in scenarios with z k = [ 200 , 200 ] .
Subsequently, it is necessary to analyze to what extent the increase in the number of groups (G) and observations per group (Obs) can affect the performance of the various aggregation methods using the considered estimators. Comparing the same support, z k = [ 10 , 10 ] , with cond( X ) = 10 and cond( X ) = 20,000, it is observed (illustrative graphical representions highlighting this observation can be requested from the authors) that bOLS demonstrates worse performance when Obs is smaller, and | | β ^ β | | is lower when this number increases. The increase in G seems to improve the performance of bOLS, except in the case e C ( 0 , 2 ) , which shows an atypical effect on the precision error of bOLS, possibly due to the pronounced presence of noise.
These results are expected due to sampling and inferential statistical theory. The mOLS method seems to follow a similar behavior, where both the increase in G and the increase in Obs lead to better performance, even in cases where e C ( 0 , 2 ) .
However, the estimators based on the ME principle do not appear to follow this behavior, remaining approximately constant with the increase in G and the addition of more observations per group, being more unequivocal in the bagging and magging aggregation methods. The invariability of | | β ^ β | | values concerning the increase in the number of groups/observations is a significant advantage in using the bagging and magging methods that use the GME and W-GME estimators. As there is no need for a larger number of groups or larger datasets to obtain suitable estimates for the problem in terms of precision error, it is possible to reduce the computational burden of implementing these aggregation procedures.
In summary, as observed in Table 1, Table 2 and Table 3, the best precision error results are obtained through bGME, bW-GME, mGME, and mW-GME in scenarios of errors modeled by a normal distribution, and by mGME and mW-GME in scenarios modeled by a t-Student or Cauchy distribution.
Analyzing the same support z k = [ 200 , 200 ] , with cond( X ) = 10 and cond( X ) = 20,000, the conclusions for the aggregation methods employing the OLS estimator remain substantially the same as before (illustrative graphical representations highlighting this observation can be requested from the authors) for all considered error distributions. However, with the increase in the parameter supports’ range (and despite scenarios where e N ( 0 , 1 ) , in which | | β ^ β | | remains approximately constant for bGME, bW-GME, mGME, and mW-GME), when e t ( 3 ) and e C ( 0 , 2 ) , all aggregation procedures with estimators based on the ME principle show a decrease in performance as Obs increases (contrary to the results obtained for the narrower support). Except for the neagging method in the case of e C ( 0 , 2 ) , this decrease appears to be more pronounced with the increase in the number of observations and remains unchanged with the increase in the number of groups. In the case of the neagging method, this effect might indicate that aggregating estimates with the knowledge of the proportion of the information content of various groups is more advantageous compared to aggregating with the proportion of the information content of larger groups, as groups with many observations do not necessarily imply a higher state of knowledge or a lower state of uncertainty.
Lastly, considering the comparison of situations with the same condition number, and naturally excluding methods that used the OLS estimator, it can be said that the increase in the parameter supports’ range, from z k = [ 10 , 10 ] to z k = [ 200 , 200 ] , caused a significant decrease in the performance of all aggregation procedures, except for mGME and mW-GME (excluding cases with lower G and higher Obs). This analysis is in agreement with the discussion of Table 1, Table 2 and Table 3.
One of the most characteristic aspects of this simulation study was the similar performance of prediction errors for all analyzed scenarios, with differences only occurring with changes in the distribution of random disturbances. The most pronounced difference in the values of | | y ^ y | | was associated with bOLS, which exhibits the highest prediction error value and associated variance. However, | | y ^ y | | is similar for all other aggregation methodologies and their estimators used, even when changing the level of collinearity present in the explanatory variables, the supports of the unknown parameters, and the number of groups and observations per group. The only exception is in the neagging aggregation methodologies when the supports’ range is broader, producing higher prediction errors and greater associated variance, except when e C ( 0 , 2 ) . However, the magging aggregation method consistently shows low prediction error and associated variance across all analyzed scenarios.
In this initial scenario, the estimates provided by bGME, bW-GME, mGME, mW-GME, and nGME in Table 4 show a high level of shrinkage towards the center of their supports and reduced variance in the results—the GME estimator can be considered a shrinkage estimator. As the support limits are narrow, they exert significant pressure on the estimates, as decreasing the support limits diminishes the impact of the data and increases the impact of the supports. Despite this, due to the prior information provided to the method, they yield better results than the conventional OLS estimator in the context of aggregation. The analysis of the boxplots associated with these estimates shows very little variation, indicating that these procedures are more stable than the other procedures analyzed, as evidenced in Figure 1. Additionally, this behavior is identified in Figure 2, as the widths of the box plot boxes for the precision errors of these procedures are much smaller than those of other methods, indicating reduced variation in the results (please note the different scales inside some figures and between figures throughout this work).
Given this, it can be said that the bagging and magging aggregation methods exhibit more stable estimates for this scenario, but only when applied with the GME and W-GME estimators, as well as the neagging method when applied with the GME.
When comparing the same scenario of normal errors, but with a higher condition number and an increase in the number of groups, in the number of observations, and in the amplitude of the support for the errors (cond( X ) =20,000, z k = [ 200 , 200 ] , G = 20 and Obs = 100), in Table 4, the estimates of the methods previously discussed (namely, bGME, bW-GME, mGME, mW-GME, and nGME) show a low precision error, but only mGME and mW-GME present estimates with less variation compared to the other procedures, as observed in Figure 3. Additionally, the same is observed in Figure 4, where a reduced variance corresponding to these methods in relation to precision error is noted. Thus, the methods highlighted in these scenarios are mGME and mW-GME, with the lowest values of | | β ^ β | | and associated variances, indicating more stable estimates.
In the scenario e t ( 3 ) , with a lower condition number, G, and Obs, but z k = [ 200 , 200 ] , in Table 5, the standout procedure is the magging aggregation methodology using the W-GME estimator due to the reduced | | β ^ β | | and associated variances (see Figure 5). As previously discussed, in this scenario, a decrease in prediction performance of bOLS, nGME, and nW-GME is already visible, along with a general increase in | | y ^ y | | in the remaining methods, compared to the scenarios with e N ( 0 , 1 ) .
With the reduction in the amplitude of the support to z k = [ 10 , 10 ] , and the increase in the condition number, in G, and in Obs, the aggregation methods with estimators based on the ME principle regain prominence, as can be seen from Table 5, both in terms of precision and reduced variance. In particular, the magging method exhibits the best results in terms of | | β ^ β | | and reduced variation in the estimates (see Figure 6). Another evident characteristic is the practically constant prediction error value corresponding to all methods (although higher than in the e N ( 0 , 1 ) scenarios).
Now, analyzing the situation of e C ( 0 , 2 ) , with values of G = 20 and Obs = 50 and z k = [ 200 , 200 ] , the estimates with lower precision error and low variance values come from mW-GME, as can be confirmed in Table 6, further supported by the analysis of Figure 7. However, despite showing low variance, the estimates are excessively centered around zero. Due to the more pronounced presence of noise in this scenario, the standard deviation of the random disturbances, estimated by the standard deviation of the noisy observations, was eventually very high, which may have resulted in obtaining wider error supports. Additionally, the supports for the parameters also have a large amplitude, increasing the impact of the data.
To present a comparison with the previous scenario, consider now the situation where z k = [ 10 , 10 ] , G decreases to 10, and Obs increases to 100, as described in Table 6. Again, the methodologies mGME and mW-GME show good performance, with reduced associated variances and corresponding precision errors with lower variation values, as confirmed in Figure 8.
To highlight how different simulated conditions affect performance and to summarize the most relevant relationships, two heatmaps were created. For illustrative purposes, the simulated condition with 10 groups and 50 observations per group (G = 10, Obs = 50) was selected. An inverse min–max transformation was applied to the precision error values (type 2 error was chosen as it represents the precision error of the final aggregated estimate), resulting in a normalized performance score, ranging from 0 (worst) to 1 (best). Figure 9 presents normalized performance scores, with normalization applied independently in two complementary ways: by scenario and by aggregation/estimation method. Panel (a) allows a comparative assessment of methods within each scenario, but not across scenarios. Conversely, in panel (b), the scores enable comparison of scenarios within each method, but not across methods.
As a final technical remark, the method designed for determining the parameter γ is a data-driven method, but different from that suggested in the work of [14], which was the least squares cross-validation (LSCV) method. Due to computational complexity, LSCV is not suitable for big data problems, as it implies thecalculation of N estimates of β using the W-GME estimator, for each value of the γ parameter to be tested, for each sample of size N. A new method was used in this work for determining this parameter, based on the Holdout methodology, which firstly involves dividing the entire dataset into two mutually exclusive subsets (the training set, T R S , and the test set, T E S ). For each value of the γ parameter to be tested, the β is then estimated, in the training set, using the W-GME estimator, and the prediction error is calculated using the test data, as follows:
S ^ ( γ ) = | | y T E S X T E S β ^ T R S ( γ ) | | .
This procedure is executed for any set of possible values of γ that one wishes to test, and, in the end, the value of γ that minimizes the prediction error, S ^ ( γ ) , is selected. Although this method aims to determine a reasonable estimate for the parameter γ , and not necessarily the best estimate, it is believed to be sufficient to obtain adequate results, with a much lower computational cost than with LSCV. Using this approach, all values in the range [ 0 , 1 ] with a spacing of 0.01 were tested for the parameter γ .
It was possible to observe the absence of any kind of pattern in the analyzed scenarios, indicating that the weighting parameter γ was being selected based on the data. As in the aggregation context, it is necessary to calculate a parameter γ for each group, it was possible to observe the value of γ adjusting to the data subsets according to the implemented method (it is not possible to present a parameter γ value for each scenario. In the aggregation context, an estimate of the parameter is calculated per group, using the W-GME estimator. Thus, the value of the parameter γ refers to each set of observations (group). Graphical representations available upon request from the authors). As noted earlier, the combination of the magging aggregation procedure with the estimation through W-GME, with this new method of selecting the parameter γ , results in the best outcomes in terms of precision error and low associated variances, among all the aggregation procedures considered with the different estimators discussed. From this, it can be concluded that the new method, although simple, works effectively in determining a reasonable value for the parameter γ , allowing the testing of a variety of values, as many as desired.

5. Conclusions

The effectiveness of aggregation methods in solving big data problems is, indeed, remarkable. The parameters of ill-posed linear regression models (particularly, ill-conditioned) in a big data context can be estimated stably through the info-metrics approach. The principle of maximum entropy developed by [2,3] is the theoretical basis for solving ill-posed problems, but the info-metrics approach developed by [4,5] allowed its generalization to more complex ill-posed problems, which are very common in various scientific fields.
The objective of this work was to understand which methodologies are most suitable for big data problems in linear regression models affected by collinearity. The first major conclusion was that the performance of aggregation methods critically depends on the estimator used in the groups obtained by sampling to obtain the respective regression coefficient estimates. The most significant differences highlighted by the simulation work were the diversity of estimates obtained by each aggregation procedure (with the respective applied estimators) and their corresponding precision errors. The prediction error shows significant differences only when the distribution associated with the error component is modified. On the other hand, when the supports for the regression coefficients have a smaller range, assuming that the true value of β is contained within these supports, the choice of the number of groups and the number of observations per group is indifferent in the use of the bagging and magging aggregation methods, using the GME and W-GME estimation methods. This is because the results are approximately similar, presenting reduced precision errors and associated variances, indicating stable estimates. To reduce computational load, a low number of groups and observations per group is sufficient to obtain consistent estimates in the previously described scenario. This is undoubtedly another important result of this work.
The excellent performance of the magging aggregation method using the W-GME estimator was perhaps the most relevant discovery in this research work. When there is no prior information about the supports of unknown parameters and the error distribution, the magging aggregation method with the W-GME estimator shows superior performance in all studied scenarios compared to other aggregation procedures with various estimators used. In particular, the W-GME estimator generally provides better results in terms of precision error (when combined with the magging aggregation method), especially when the supports are wider. This result is quite promising, as the GME estimator, which usually performs well with data affected by collinearity, behaves worse in this circumstance of wider supports (as expected). If the W-GME estimator, being one of its extensions, can handle this problem, it becomes a very attractive estimation method.
Beyond the promising results in the simulation study, the proposed strategies for estimation and aggregation have strong potential for application in diverse real-world domains. These include econometrics, genomics, environmental sciences, and machine learning, where data are often noisy and ill-conditioned. The good performance of GME and W-GME estimators under collinearity and different distributional contexts for the errors, combined with an appropriate aggregation procedure, make them attractive for inference purposes under big data scenarios.
Given the small number of groups and observations per group needed to obtain good results, the computational cost is not a relevant factor, although it is important to keep in mind that the GME and W-GME estimators are slightly slower (constrained nonlinear optimization). Of course, if necessary, parallel computing can be easily implemented, making processing times even shorter.
This work also alerts us to the risk of using the OLS estimator, knowing that data scientists frequently use linear models as a simplified view of the reality in big data analysis, and the OLS estimator is routinely used in practice. Under low collinearity scenarios (condition number of 10), it was verified that the traditional OLS estimator with bagging performs well; however, this strategy tends to become very unstable under high collinearity scenarios (condition number of 20,000), where magging with W-GME stands out in terms of higher performance.
It is important to note some limitations and new avenues of research uncovered by this work. Further research needs to include a broader set of aggregation procedures and estimation methods, which could enhance the comparative framework and potentially offer a more comprehensive view of the context-specific advantages or limitations of each technique (estimation and aggregation), which may not be yet completely revealed [18,19,20]. While the design adopted in this work enabled a systematic and controlled comparison of different methodologies under ill-posed linear regression models, future work will aim to extend the simulation study to account for more realistic scenarios, including nonlinear relationships, heteroscedasticity, skewed distributions for the explanatory variables, and other forms of model misspecification, as well as to test the methods’ performance on real-world datasets.
In summary, as previously mentioned, the identification of the excellent performance of the magging aggregation method, suitable for non-homogeneous data circumstances, using the W-GME estimator, which uses weights in the objective function of the optimization problem to seek a better balance between precision and prediction, was a relevant discovery that could contribute to the analysis of large volumes of information, which is an urgent need in many fields of human activities. Although the entropy-based framework of neagging provides an interpretable aggregation mechanism, its behavior may vary depending on group structure, data quality, and the stability of individual estimators. A deeper exploration of these factors will be fundamental to broadening its applicability. We hope that this work contributes not only by benchmarking methods under controlled conditions, but also by encouraging further research into stable and information-driven aggregation techniques that can meet the challenges posed by modern data environments.

Author Contributions

All the authors contributed to the methodology, software, and writing—original draft preparation. A.S. and T.F. contributed to the validation and visualization. A.H.T., M.C., P.M., and R.A.d.C. contributed to the conceptualization and writing—review and editing. M.C., P.M., and R.A.d.C. contributed to the supervision. All authors have read and agreed to the published version of the manuscript.

Funding

This work is partially supported by the Center for Research and Development in Mathematics and Applications (CIDMA) under the Portuguese Foundation for Science and Technology (FCT, https://ror.org/00snfqn58 (accessed on 23 July 2025)) Multi-Annual Financing Program for R&D Units, grants UID/4106/2025 and UID/PRR/4106/2025. R.d.C. acknowledges the FCT Grant No. CEECIND/04697/2017.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The computational structure defined in the manuscript allows the simulation to be replicated. A general code will soon be made public on CRAN, in a user-friendly format. Additional details are available upon request from the authors.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
GMEGeneralized maximum entropy
W-GMEWeighted generalized maximum entropy
OLSOrdinary least squares
bOLSBagging aggregation procedure with OLS coefficient estimates
bGMEBagging procedure with GME estimates
bW-GMEBagging procedure with W-GME estimates
mOLSMagging procedure with OLS estimates
mGMEMagging procedure with GME estimates
mW-GMEMagging procedure with W-GME estimates
nGMENeagging procedure with GME estimates
nW-GMENeagging procedure with W-GME estimates
GNumber of groups
ObsNumber of observations per group

Appendix A

To illustrate the algebraic structure of the GME estimator, and considering the same supports used in Section 3, suppose a simple linear regression model, with only ten observations, constructed with an explanatory variable following a U ( 5 , 10 ) and errors following a N ( 0 , 1 ) . The coefficients are β 0 = 0.5 and β 1 = 2.5 , and the response is obtained accordingly to (5). The values of the explanatory and the response variables are rounded to one decimal.
Considering the support spaces as [ 200 , 200 ] for all the parameters (with M = 5 ) and [ 11 , 11 ] for all the errors (with J = 3 ) using the three-sigma rule, the GME estimator is given by (10), subject to the model constraints in (11),
14.8 17.5 22.2 16.4 14.9 23.1 17.1 18.3 13.1 13.3 = 1 5.7 1 7.2 1 8.1 1 6.5 1 5.8 1 8.5 1 6.7 1 7.2 1 5.3 1 5.5 × 200 100 0 100 200 0 0 0 0 0 0 0 0 0 0 200 100 0 100 200 ×
× p 1 , 1 p 1 , 2 p 1 , 3 p 1 , 4 p 1 , 5 p 2 , 1 p 2 , 2 p 2 , 3 p 2 , 4 p 2 , 5 + 11 0 11 0 0 0 0 0 0 0 0 0 11 0 11 0 0 0 0 0 0 0 0 0 11 0 11 × w 1 , 1 w 1 , 2 w 1 , 3 w 2 , 1 w 2 , 2 w 2 , 3 w 10 , 1 w 10 , 2 w 10 , 3 ,
the additivity constraints for p in (12),
1 1 = 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 1 1 1 1 1 × p 1 , 1 p 1 , 2 p 1 , 3 p 1 , 4 p 1 , 5 p 2 , 1 p 2 , 2 p 2 , 3 p 2 , 4 p 2 , 5 ,
and the additivity constraints for w in (12),
1 1 1 1 1 1 1 1 1 1 = 1 1 1 0 0 0 0 0 0 0 0 0 1 1 1 0 0 0 0 0 0 0 0 0 1 1 1 × w 1 , 1 w 1 , 2 w 1 , 3 w 2 , 1 w 2 , 2 w 2 , 3 w 10 , 1 w 10 , 2 w 10 , 3 .
Considering the support spaces as [ 10 , 10 ] for all the parameters (with M = 5 ), only the model constraints in (11) need to be adapted as
14.8 17.5 22.2 16.4 14.9 23.1 17.1 18.3 13.1 13.3 = 1 5.7 1 7.2 1 8.1 1 6.5 1 5.8 1 8.5 1 6.7 1 7.2 1 5.3 1 5.5 × 10 5 0 5 10 0 0 0 0 0 0 0 0 0 0 10 5 0 5 10 × p 1 , 1 p 1 , 2 p 1 , 3 p 1 , 4 p 1 , 5 p 2 , 1 p 2 , 2 p 2 , 3 p 2 , 4 p 2 , 5 +
+ 11 0 11 0 0 0 0 0 0 0 0 0 11 0 11 0 0 0 0 0 0 0 0 0 11 0 11 × w 1 , 1 w 1 , 2 w 1 , 3 w 2 , 1 w 2 , 2 w 2 , 3 w 10 , 1 w 10 , 2 w 10 , 3 .
Using, for example, a nonlinear programming solver (https://www.mathworks.com/help/optim/ug/fmincon.html (accessed on 23 July 2025)) from MATLAB [17], the results are as follows (the products Z p ^ are computed with the maximum available precision in MATLAB):
β ^ 0 β ^ 1 = 200 100 0 100 200 0 0 0 0 0 0 0 0 0 0 200 100 0 100 200 × 0.2063 0.2031 0.1999 0.1968 0.1937 0.1939 0.1969 0.1999 0.2030 0.2061
3.1417 3.0395
and
β ^ 0 β ^ 1 = 10 5 0 5 10 0 0 0 0 0 0 0 0 0 0 10 5 0 5 10 × 0.2031 0.2015 0.1999 0.1984 0.1968 0.1096 0.1429 0.1865 0.2433 0.3174
0.0786 2.5808 .
It is important to note that the results from OLS, defined in (6),
β ^ 0 β ^ 1 3.2026 3.0485 ,
are similar to those from GME with wider support spaces for the parameters, as expected. However, when prior information exists and the supports are narrower, the precision of the estimates substantially increases. This is the reason to consider two different scenarios in Section 3: one where it is assumed some prior knowledge about the range in which the parameters may be found—support spaces as [ 10 , 10 ] for all the parameters—and another where the available prior knowledge is considered insufficient—support spaces as [ 200 , 200 ] for all the parameters.

References

  1. Costa, M.C.; Macedo, P. Normalized Entropy Aggregation for Inhomogeneous Large-Scale Data. In Proceedings of the Theory and Applications of Time Series Analysis, Granada, Spain, 19–21 September 2018; Valenzuela, O., Rojas, F., Pomares, H., Rojas, I., Eds.; Springer: Cham, Switzerland, 2019; pp. 19–29. [Google Scholar]
  2. Jaynes, E.T. Information theory and statistical mechanics. Phys. Rev. 1957, 106, 620–630. [Google Scholar]
  3. Jaynes, E.T. Information theory and statistical mechanics. II. Phys. Rev. 1957, 108, 171–190. [Google Scholar]
  4. Golan, A.; Judge, G.; Miller, D. Maximum Entropy Econometrics: Robust Estimation with Limited Data; Wiley: Chichester, UK, 1996. [Google Scholar]
  5. Golan, A. Foundations of Info-Metrics: Modeling, Inference, and Imperfect Information; Oxford University Press: New York, NY, USA, 2018. [Google Scholar]
  6. Breiman, L. Bagging Predictors. Mach. Learn. 1996, 24, 123–140. [Google Scholar]
  7. Bühlmann, P.; Meinshausen, N. Magging: Maximin Aggregation for Inhomogeneous Large-Scale Data. Proc. IEEE 2016, 104, 126–135. [Google Scholar]
  8. Shannon, C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef]
  9. Hoerl, A.E.; Kennard, R.W. Ridge Regression: Biased Estimation for Nonorthogonal Problems. Technometrics 1970, 12, 55–67. [Google Scholar] [CrossRef]
  10. Hoerl, A.E.; Kennard, R.W. Ridge Regression: Applications to Nonorthogonal Problems. Technometrics 1970, 12, 69–82. [Google Scholar] [CrossRef]
  11. Belsley, D.A.; Kuh, E.; Welsch, R.E. Regression Diagnostics–Identifying Influential Data and Sources of Collinearity; Wiley: Hoboken, NJ, USA, 2004. [Google Scholar]
  12. Pukelsheim, F. The Three Sigma Rule. Am. Stat. 1994, 48, 88–91. [Google Scholar] [CrossRef]
  13. Macedo, P. A two-stage maximum entropy approach for time series regression. Commun. Stat.-Simul. Comput. 2024, 53, 518–528. [Google Scholar] [CrossRef]
  14. Wu, X. A Weighted Generalized Maximum Entropy Estimator with a Data-driven Weight. Entropy 2009, 11, 917–930. [Google Scholar] [CrossRef]
  15. Costa, M.C.; Macedo, P.; Cruz, J.P. Neagging: An Aggregation Procedure Based On Normalized Entropy. AIP Conf. Proc. 2022, 2425, 190003. [Google Scholar] [CrossRef]
  16. Mittelhammer, R.; Cardell, N.S.; Marsh, T.L. The Data-Constrained Generalized Maximum Entropy Estimator of the GLM: Asymptotic Theory and Inference. Entropy 2013, 15, 1756–1775. [Google Scholar] [CrossRef]
  17. The MathWorks Inc. MATLAB Version: 9.7.0 (R2019b); The MathWorks Inc.: Natick, MA, USA, 2019. Available online: https://www.mathworks.com (accessed on 23 July 2025).
  18. Chen, Y.; Yang, Y.; Pei, S.; Chen, Y.; Du, J. A simple rapid sample-based clustering for large-scale data. Eng. Appl. Artif. Intell. 2024, 133, 108551. [Google Scholar] [CrossRef]
  19. Yoo, J.E. Penalized Regression in Large-Scale Data Analysis. In Machine Learning in Educational Sciences: Approaches, Applications and Advances; Khine, M.S., Ed.; Springer Nature: Singapore, 2024; pp. 71–91. [Google Scholar]
  20. Breiman, L. Stacked Regressions. Mach. Learn. 1996, 24, 49–64. [Google Scholar] [CrossRef]
Figure 1. Box plots of the estimates of the 10 linear regression coefficients. Scenario: e N ( 0 , 1 ) , cond( X ) = 10, z k = [ 10 , 10 ] , G = 10, Obs = 50. Note the different scales inside the figure.
Figure 1. Box plots of the estimates of the 10 linear regression coefficients. Scenario: e N ( 0 , 1 ) , cond( X ) = 10, z k = [ 10 , 10 ] , G = 10, Obs = 50. Note the different scales inside the figure.
Entropy 27 01075 g001
Figure 2. Box plots of the precision errors. Scenario: e N ( 0 , 1 ) , cond( X ) = 10, z k = [ 10 , 10 ] , G = 10, Obs = 50. (a,b) are zoomed-in images of the highlighted box plots.
Figure 2. Box plots of the precision errors. Scenario: e N ( 0 , 1 ) , cond( X ) = 10, z k = [ 10 , 10 ] , G = 10, Obs = 50. (a,b) are zoomed-in images of the highlighted box plots.
Entropy 27 01075 g002
Figure 3. Box plots of the estimates of the 10 linear regression coefficients. Scenario: e N ( 0 , 1 ) , cond( X ) = 20,000, z k = [ 200 , 200 ] , G = 20, Obs = 100. Note the different scales inside the figure.
Figure 3. Box plots of the estimates of the 10 linear regression coefficients. Scenario: e N ( 0 , 1 ) , cond( X ) = 20,000, z k = [ 200 , 200 ] , G = 20, Obs = 100. Note the different scales inside the figure.
Entropy 27 01075 g003
Figure 4. Box plots of the precision errors. Scenario: e N ( 0 , 1 ) , cond( X ) = 20,000, z k = [ 200 , 200 ] , G = 20, Obs = 100. Note that the y-axis scale of the main plot is of order 10 4 . (a,b) are zoomed-in images of the highlighted box plots.
Figure 4. Box plots of the precision errors. Scenario: e N ( 0 , 1 ) , cond( X ) = 20,000, z k = [ 200 , 200 ] , G = 20, Obs = 100. Note that the y-axis scale of the main plot is of order 10 4 . (a,b) are zoomed-in images of the highlighted box plots.
Entropy 27 01075 g004
Figure 5. Box plots of the precision errors. Scenario: e t ( 3 ) , cond( X ) = 10, z k = [ 200 , 200 ] , G = 10, Obs = 50. (a,b) are zoomed-in images of the highlighted box plots.
Figure 5. Box plots of the precision errors. Scenario: e t ( 3 ) , cond( X ) = 10, z k = [ 200 , 200 ] , G = 10, Obs = 50. (a,b) are zoomed-in images of the highlighted box plots.
Entropy 27 01075 g005
Figure 6. Box plots of the precision errors. Scenario: e t ( 3 ) , cond( X ) = 20,000, z k = [ 10 , 10 ] , G = 20 , Obs = 100 . Note that the y-axis scale of the main plot is of order 10 4 . (a,b) are zoomed-in images of the highlighted box plots.
Figure 6. Box plots of the precision errors. Scenario: e t ( 3 ) , cond( X ) = 20,000, z k = [ 10 , 10 ] , G = 20 , Obs = 100 . Note that the y-axis scale of the main plot is of order 10 4 . (a,b) are zoomed-in images of the highlighted box plots.
Entropy 27 01075 g006
Figure 7. Boxplots of the precision errors. Scenario: e C ( 0 , 2 ) , cond( X ) = 10, z k = [ 200 , 200 ] , G = 20, Obs = 50. Note that the y-axis scale of the main plot is of order 10 4 .
Figure 7. Boxplots of the precision errors. Scenario: e C ( 0 , 2 ) , cond( X ) = 10, z k = [ 200 , 200 ] , G = 20, Obs = 50. Note that the y-axis scale of the main plot is of order 10 4 .
Entropy 27 01075 g007
Figure 8. Boxplots of the precision errors. Scenario: e C ( 0 , 2 ) , cond( X ) = 20,000, z k = [ 10 , 10 ] , G = 10, Obs = 100. Note that the y-axis scale of the main plot is of order 10 6 .
Figure 8. Boxplots of the precision errors. Scenario: e C ( 0 , 2 ) , cond( X ) = 20,000, z k = [ 10 , 10 ] , G = 10, Obs = 100. Note that the y-axis scale of the main plot is of order 10 6 .
Entropy 27 01075 g008
Figure 9. Heatmap of the normalized performance score (scaled from 0 to 1) based on precision error 2, under the condition of the smallest number of groups and of the number of observations per group (G = 10, Obs = 50): (a) Precision errors were normalized independently across scenarios. The score allows comparative assessment of methods within each scenario (row), but not across scenarios. (b) Precision errors were normalized independently across methods, and enables comparison of scenarios within each method (column). Each simulated scenario is identified on the horizontal axis by the following: error distribution—N (Normal), t (t-Student) or C (Cauchy); collinearity strength—low (cond( X ) = 10) or high (cond( X ) = 20,000); and range of parameter supports—small ( z k = [ 10 , 10 ] ) or large ( z k = [ 200 , 200 ] ).
Figure 9. Heatmap of the normalized performance score (scaled from 0 to 1) based on precision error 2, under the condition of the smallest number of groups and of the number of observations per group (G = 10, Obs = 50): (a) Precision errors were normalized independently across scenarios. The score allows comparative assessment of methods within each scenario (row), but not across scenarios. (b) Precision errors were normalized independently across methods, and enables comparison of scenarios within each method (column). Each simulated scenario is identified on the horizontal axis by the following: error distribution—N (Normal), t (t-Student) or C (Cauchy); collinearity strength—low (cond( X ) = 10) or high (cond( X ) = 20,000); and range of parameter supports—small ( z k = [ 10 , 10 ] ) or large ( z k = [ 200 , 200 ] ).
Entropy 27 01075 g009
Table 1. Summary table of precision errors 1 and 2 for scenarios related to the data matrix with e N ( 0 , 1 ) . In each scenario, the lowest precision error is highlighted in bold.
Table 1. Summary table of precision errors 1 and 2 for scenarios related to the data matrix with e N ( 0 , 1 ) . In each scenario, the lowest precision error is highlighted in bold.
cond(X) z k G/ObsType ofBaggingMaggingNeagging
Error OLS GME W-GME OLS GME W-GME GME W-GME
10[−10,10]10/50186.2910.4610.7345.6010.4810.4810.4416.99
229.5110.4510.6213.8410.4810.4810.4412.90
10/100134.2310.3310.1824.8510.4810.4913.7514.50
215.1510.259.8913.5510.4810.499.728.40
20/50152.0910.4110.3913.6910.4810.4810.4117.53
214.3810.4110.348.8910.4810.4810.4012.65
20/100124.9110.229.6412.4010.4510.4713.9211.61
29.6710.209.548.4610.4510.4710.436.42
[−200,200]10/501116.4322.7521.1654.8414.6911.2225.4470.89
226.8711.6310.1731.2811.2610.6610.5926.23
10/100139.1517.3115.7925.8215.7610.3822.7340.36
213.848.747.4914.799.959.0911.4218.20
20/50168.2617.3118.3519.5011.4410.5119.3899.55
29.748.948.8513.8310.7610.508.5836.65
20/100129.8312.8512.2313.5613.1110.2316.7828.06
212.216.927.138.409.0710.038.0313.27
20,000[−10,10]10/501108,807.1410.4510.5353,822.6610.4710.4910.4213.76
243,621.6610.4510.4520,524.7910.4710.4910.4211.11
10/100141,645.1810.2110.2023,219.2910.3910.4613.5514.92
26285.3910.1410.041885.4210.3910.469.6710.48
20/501134,468.1910.4410.3611,333.2210.4810.4810.4113.03
230,674.0110.4410.341133.5210.4810.4810.409.95
20/100126,088.7910.219.915521.0610.4710.4813.5612.73
212,642.3410.209.851508.4410.4710.4810.428.04
[−200,200]10/501160,028.5223.2021.7494,346.6715.8410.7626.5468.17
26329.6011.419.6111,141.3711.1110.0811.6629.92
10/100146,393.1516.2815.7029,834.9811.599.9521.2548.95
25738.786.596.7920,050.435.728.718.6630.20
20/50191,114.5316.6815.508984.2110.5210.5321.2747.21
244,962.438.308.468975.9910.3810.529.4718.12
20/100137,064.0610.6010.7710,508.849.3810.0213.4744.37
25656.433.693.969093.499.029.923.7630.13
Table 2. Summary table of precision errors for scenarios related to the data matrix with e t ( 3 ) . In each scenario, the lowest precision error is highlighted in bold.
Table 2. Summary table of precision errors for scenarios related to the data matrix with e t ( 3 ) . In each scenario, the lowest precision error is highlighted in bold.
cond(X) z k G/ObsType ofBaggingMaggingNeagging
Error OLS GME W-GME OLS GME W-GME GME W-GME
10[−10,10]10/501202.8511.1911.1769.8410.4710.4822.7322.46
244.8710.7110.6723.1710.4710.4813.3913.50
10/100160.4412.2612.2234.7610.3710.4314.6514.60
242.949.869.7919.4010.3610.3510.2710.25
20/501130.9711.3511.3117.5010.4810.4822.9521.78
227.9111.0410.9710.4010.4810.4814.4414.24
20/100141.9512.5312.5213.1410.5210.5313.5513.53
212.1311.0611.0710.9610.5110.5311.2211.22
[−200,200]10/501179.8531.0629.8482.4216.1910.3392.55139.51
243.8713.6816.6637.8110.5010.0362.9177.26
10/100157.6387.1786.2128.6923.3026.74341.40343.79
210.1536.1336.8910.6210.7513.46130.56130.94
20/501124.3021.4823.1511.2010.8210.5693.96119.50
228.8013.2212.3810.4610.3310.5155.0843.32
20/100139.7751.4951.4511.6110.3912.10219.09216.45
212.5123.4423.6110.2210.0811.2486.3783.99
20,000[−10,10]10/501193,579.8111.1110.9176,289.4710.4710.4821.0221.30
213,066.0710.6210.319848.5110.4710.4811.0610.53
10/100195,677.6613.7413.7536,441.8510.5010.5215.2915.24
224,643.9911.2511.267486.2310.4810.5111.6211.64
20/501245,249.8410.8410.9410,325.3910.4910.4922.7920.49
241,755.9310.5110.601449.4510.4910.4913.2513.10
20/100155,489.5711.8911.873785.9910.5410.5712.8312.81
222,566.0910.2510.221618.9910.4910.5110.2110.20
[−200,200]10/501308,013.7234.3033.81133,724.1216.1210.97114.29147.73
24035.9814.8913.115534.5910.5410.1865.2360.69
10/100169,083.7277.5776.3233,072.8519.0615.02258.82259.92
23721.4434.3334.2311,179.6011.139.81113.40113.81
20/501135,002.5322.1121.252172.4510.0410.44102.45127.05
277,503.4710.7910.951299.899.8810.4351.6347.99
20/100145,676.8555.2356.3910,954.6510.8112.48240.71246.11
2523.3928.4330.17551.4910.1010.95119.62123.56
Table 3. Summary table of precision errors for scenarios related to the data matrix with e C ( 0 , 2 ) . In each scenario, the lowest precision error is highlighted in bold.
Table 3. Summary table of precision errors for scenarios related to the data matrix with e C ( 0 , 2 ) . In each scenario, the lowest precision error is highlighted in bold.
cond(X) z k G/ObsType ofBaggingMaggingNeagging
Error OLS GME W-GME OLS GME W-GME GME W-GME
10[−10,10]10/5017807.1814.1014.07492.1710.5010.5216.5616.55
22280.3311.4511.42143.1510.5010.4911.8511.85
10/10015561.9613.8413.84267.4710.8810.8813.8413.84
23782.2610.7510.7542.829.759.7510.7410.74
20/5013246.6512.1812.1646.1510.4810.4913.6613.65
21194.6710.6910.6915.4610.4810.4910.7310.73
20/100144,481.8613.1013.1039.4010.4610.4613.1013.10
242,915.8011.3911.3921.0710.3710.3711.3911.39
[−200,200]10/5015087.95131.86135.29289.2416.0814.32295.25293.38
21435.8535.0235.3553.4510.7410.7274.9673.93
10/100195,540.28225.51225.53309.14138.16138.17225.06225.10
293,551.8684.7584.81202.7344.3644.3684.7084.89
20/5016505.5378.9278.7952.4611.0810.52194.14192.51
24648.7225.8625.4334.4810.7410.5066.5165.59
20/100155,171.34134.45134.3931.1619.9519.95134.91134.82
252,819.7755.5055.3411.4210.8210.8156.3056.13
20,000[−10,10]10/50118,425,545.0513.0713.12816,608.7710.4710.3715.1615.13
214,091,527.339.979.9773,850.7510.4710.2810.2510.23
10/10013,960,459.1213.5613.56544,287.1910.9410.9413.5613.56
21,139,209.5410.8610.86140,661.7210.3810.3810.8610.86
20/501498,229,663.9611.4911.4653,777.2010.4910.4913.2613.15
2493,730,593.8710.0710.063794.0210.4910.4910.079.97
20/1001123,549,997.7412.9412.9438,200.4210.6410.6412.9512.95
2121,549,224.3511.2311.23929.5910.5210.5211.2411.24
[−200,200]10/5016,984,823.10116.09115.91487,157.7421.4511.75265.34260.49
2676,483.3930.1030.7031,685.3911.9110.1475.3673.29
10/10017,771,232.94196.80197.01417,367.82106.57107.76198.21197.33
23,216,059.6369.8568.6731,103.9146.8545.2370.5669.52
20/501597,840,024.3388.5689.9954,361.4512.9613.21199.86198.30
2591,403,182.8529.0829.8554,354.2011.2811.5362.5361.92
20/1001116,093,776.13138.41138.70134,750.2429.6229.79139.90140.18
2108,344,985.8445.6545.6973,639.6013.0813.0047.2247.23
Table 4. Prediction errors 1 and 2, and variance of the prediction error 1 for two scenarios related to the data matrix with e N ( 0 , 1 ) . Scenario A: cond( X ) = 10, z k = [ 10 , 10 ] , G = 10, Obs = 50. Scenario B: cond( X ) = 20,000, z k = [ 200 , 200 ] , G = 20, Obs = 100.
Table 4. Prediction errors 1 and 2, and variance of the prediction error 1 for two scenarios related to the data matrix with e N ( 0 , 1 ) . Scenario A: cond( X ) = 10, z k = [ 10 , 10 ] , G = 10, Obs = 50. Scenario B: cond( X ) = 20,000, z k = [ 200 , 200 ] , G = 20, Obs = 100.
BaggingMaggingNeagging
OLS GME W-GME OLS GME W-GME GME W-GME
Scenario A
| | y ^ y | | 184.56173.14173.15176.31173.14173.14173.14173.91
| | y ^ y | | * 173.62173.14173.14173.28173.14173.14173.14173.24
s | | y ^ y | | 2 50.040.000.006.310.000.000.001.13
Scenario B
| | y ^ y | | 173.36173.19173.13173.01173.02173.05173.41174.48
| | y ^ y | | * 172.87172.86172.85172.95173.00173.05172.87172.96
s | | y ^ y | | 2 0.040.030.010.000.000.000.065.94
Table 5. Prediction errors 1 and 2, and variance of the prediction error 1 for two scenarios related to the data matrix with e t ( 3 ) . Scenario A: cond( X ) = 10, Support = [ 200 , 200 ] , G = 10, Obs = 50. Scenario B: cond( X ) = 20,000, Support = [ 10 , 10 ] , G = 20, Obs = 100.
Table 5. Prediction errors 1 and 2, and variance of the prediction error 1 for two scenarios related to the data matrix with e t ( 3 ) . Scenario A: cond( X ) = 10, Support = [ 200 , 200 ] , G = 10, Obs = 50. Scenario B: cond( X ) = 20,000, Support = [ 10 , 10 ] , G = 20, Obs = 100.
BaggingMaggingNeagging
OLS GME W-GME OLS GME W-GME GME W-GME
Scenario A
| | y ^ y | | 320.08304.14303.90306.55302.57302.33341.86353.92
| | y ^ y | | * 302.70302.33302.42302.88302.30302.32308.16311.13
s | | y ^ y | | 2 121.466.675.3910.310.040.0012,041.6113,560.87
Scenario B
| | y ^ y | | 302.94302.42302.42302.36302.36302.36302.48302.48
| | y ^ y | | * 302.34302.34302.33302.34302.36302.36302.33302.33
s | | y ^ y | | 2 0.170.000.010.000.000.000.010.01
Table 6. Prediction errors 1 and 2, and variance of the prediction error 1, for two scenarios related to the data matrix with e C ( 0 , 2 ) . Scenario A: cond( X ) = 10, z k = [ 200 , 200 ] , G = 20, Obs = 50. Scenario B: cond( X ) = 20,000, z k = [ 10 , 10 ] , G = 10, Obs = 100.
Table 6. Prediction errors 1 and 2, and variance of the prediction error 1, for two scenarios related to the data matrix with e C ( 0 , 2 ) . Scenario A: cond( X ) = 10, z k = [ 200 , 200 ] , G = 20, Obs = 50. Scenario B: cond( X ) = 20,000, z k = [ 10 , 10 ] , G = 10, Obs = 100.
BaggingMaggingNeagging
OLS GME W-GME OLS GME W-GME GME W-GME
Scenario A
| | y ^ y | | 804,318.62804,271.11804,271.09804,271.06804,271.01804,271.01804,271.24804,271.23
| | y ^ y | | * 804,294.99804,271.10804,271.09804,271.06804,271.01804,271.01804,271.21804,271.20
s | | y ^ y | | 2 10,370.590.280.280.020.000.001.541.54
Scenario B
| | y ^ y | | 804,273.47804,270.97804,270.97804,270.72804,270.95804,270.95804,270.97804,270.97
| | y ^ y | | * 804,272.27804,270.97804,270.97804,270.70804,270.95804,270.95804,270.97804,270.97
s | | y ^ y | | 2 32.920.000.000.200.000.000.000.00
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tavares, A.H.; Silva, A.; Freitas, T.; Costa, M.; Macedo, P.; da Costa, R.A. Aggregation in Ill-Conditioned Regression Models: A Comparison with Entropy-Based Methods. Entropy 2025, 27, 1075. https://doi.org/10.3390/e27101075

AMA Style

Tavares AH, Silva A, Freitas T, Costa M, Macedo P, da Costa RA. Aggregation in Ill-Conditioned Regression Models: A Comparison with Entropy-Based Methods. Entropy. 2025; 27(10):1075. https://doi.org/10.3390/e27101075

Chicago/Turabian Style

Tavares, Ana Helena, Ana Silva, Tiago Freitas, Maria Costa, Pedro Macedo, and Rui A. da Costa. 2025. "Aggregation in Ill-Conditioned Regression Models: A Comparison with Entropy-Based Methods" Entropy 27, no. 10: 1075. https://doi.org/10.3390/e27101075

APA Style

Tavares, A. H., Silva, A., Freitas, T., Costa, M., Macedo, P., & da Costa, R. A. (2025). Aggregation in Ill-Conditioned Regression Models: A Comparison with Entropy-Based Methods. Entropy, 27(10), 1075. https://doi.org/10.3390/e27101075

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop