Next Article in Journal
Determinants of the Financing Mechanisms of Decentralization in Togo
Next Article in Special Issue
Who Is Leading in Communication Tone? Wavelet Analysis of the Fed and the ECB
Previous Article in Journal
Triple-Entry Accounting and Other Secure Methods to Preserve User Privacy and Mitigate Financial Risks in AI-Empowered Lifelong Education
Previous Article in Special Issue
Price Gaps and Volatility: Do Weekend Gaps Tend to Close?
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Finite Mixture at Quantiles and Expectiles

Department of Agricultural Sciences, University of Naples Federico II, 80055 Napoli, Italy
J. Risk Financial Manag. 2025, 18(4), 177; https://doi.org/10.3390/jrfm18040177
Submission received: 3 February 2025 / Revised: 17 March 2025 / Accepted: 25 March 2025 / Published: 27 March 2025
(This article belongs to the Special Issue Machine Learning Based Risk Management in Finance and Insurance)

Abstract

:
Finite mixture regression identifies homogeneous groups within a sample and computes the regression coefficients in each group. Groups and group coefficients are jointly estimated using an iterative approach. This work extends the finite mixture estimator to the tails of the distribution, by incorporating quantiles and expectiles and relaxing the constraint of constant group probability adopted in previous analysis. The probability of each group depends on the selected location: an observation can be allocated in the best-performing group if we look at low values of the dependent variable, while at higher values it may be assigned to the poorly performing class. We explore two case studies: school data from a PISA math proficiency test and asset returns from the Center for Research in Security Prices. In these real data examples, group classifications change based on the selected location of the dependent variable, and this has an impact on the regression estimates due to the joint computation of class probabilities and class regressions coefficients. A Monte Carlo experiment is conducted to compare the performances of the discussed estimators with results of previous research.

1. Introduction

The work extends the finite mixture regression estimator (FM), which is generally computed at the conditional mean of the dependent variable. The goal is to estimate it in the tails of the conditional distribution, focusing on both low and high values of the dependent variable. This approach provides a powerful tool for uncovering tail behaviors, such as skewness and heterogeneity in response.
The FM regression approach (McLachlan & Peel, 2004) assumes that the population comprises many homogeneous subgroups. For each subgroup, a probability density function describes the data generating process, with each cluster characterized by a set of regression parameters specific to that group. The impact of the predictors varies among classes and these differences help determine the class membership. The data are assumed to be drawn from a population divided into clusters sharing similar characteristics. FM identifies homogeneous groups within the sample while computing the regression coefficients for each group. The class-specific coefficients are estimated without any prior knowledge or assumption of the clustering. FM is a model-based clustering approach designed to account for between-groups heterogeneity, where the population of interest is generated by homogeneous sub-populations. The resulting FM regressions are more flexible than the standard regressions that typically pool the subsamples. FM is commonly used as a flexible method to account for heteroskedasticity (Compiani & Kitamura, 2016). Neglecting group heterogeneity leads to misleading results since (i) the estimated coefficients would average the true values of each group without mirroring any of them, and (ii) the estimated error variance would be larger due to group heterogeneity, thus causing flawed inference (Van Horn et al., 2013).
The standard FM model computes class assignment and class-specific regressions in each group at the conditional mean of the dependent variable. Due to data heterogeneity, the sample may form several distinct clusters, and FM adaptively models multiple regressions, each one responsible for one subset/cluster of the data. These models have been widely used in market segmentation studies or patients’ disease progression subtyping (Städler et al. (2010), Khalili (2010), Khalili and Chen (2007)). A market segmentation approach is implemented by Betrail and Callavet (2008) for fruit and vegetable consumption, as well as by Caudill and Mixon (2016). Liang et al. (2018) extend the FM approach to improve robustness. Alfò et al. (2017) extend FM to quantile and M-quantile (Breckling & Chambers, 1988) models in a random coefficient setting. In this work, we extend FM by computing it away from the mean following the approach in Furno (2023) and Furno and Caracciolo (2024). They impose a constraint on the class probabilities fixing the group probability at the mean estimated values and shifting only the class regressions to the quantiles. In contrast, here, we relax this constraint, and show in a few case studies and in Monte Carlo simulations that relaxing the constraint provides additional information and more accurate results. By implementing tail estimators like expectiles (Newey & Powell, 1987) and quantiles (Koenker, 2005), we define the FM-expectile and the FM-quantile. These estimators differ from one another—expectiles are pure location devices used to move away from the mean regression. The advantages of expectiles over quantile regressions are (i) in their computational ease and (ii) in preventing regression crossing. The latter can occur in quantile regressions and are hard to explain. However, expectiles lack the robustness of quantiles, just as OLS is less robust than the median regression.
Compared to previous research (Furno, 2023; Furno & Caracciolo, 2024), the approach here considered allows for the flexibility of class membership to change at different locations, relaxing the constraint of stable class composition across locations. The probability of each group depends on the selected location—an observation can be in the best-performing group if we look at low values of the dependent variable, but the same observation can have a low probability of belonging to the best class at the right tail. The case studies here analyzed show that classes do change depending on the selected location of the dependent variable—whether we focus on low-, medium-, or high-performing observations—and this influences the class regression estimates as well. Previous research focused on the tail behavior of the sole group regressions while keeping group probability constant. This approach excludes by assumption any heterogeneity in the probability of belonging to a group. Here we relax this constraint. By allowing both group probability and group regression to change across locations, we avoid unnecessary constraints and gain a deeper understanding of the data. The probability does change across quantiles, and the novelty is in modeling the changing group probability together with the changing explanatory power of the covariates at differing locations, by implementing expectiles or quantiles estimators. The changing probability has an impact on the class regression estimates as well.
A Monte Carlo experiment compares the performance of the changing probability estimators defined in this study with that of the existing constant probability estimators. The relative bias is generally smaller in the current approach, yielding an improved performance of the FM tail estimators once the constraint of constant group probability is removed.
We examine two case studies: school data from the OECD-PISA math proficiency test and asset returns form the Center for Research in Security Prices. Both analyses highlight the utility of the proposed changing probability approach.
The school data set is considered to compare our findings with the Furno (2023) analysis, using the same data set of students’ scores in the PISA math test. There, the group probability is estimated once and for all at the mean, and kept constant when moving to the tails. In contrast, by lifting the constraint of group probability invariance, we show that class membership changes across quartiles. For instance, when examining the right tail to model the high-proficiency students, their math test scores are split into excellent versus just-above-average, and the difference turns out to be significant but rather small. In the left tail, looking at low-proficiency students, the groups collect just-below-average versus very poor performance, and their difference is nearly three times larger than the difference between the high-scoring groups. There is no reason to assume that the low-proficiency groups should have the same proportions/probabilities as the high-proficiency groups. Keeping the groups invariant is an approximation that we are now able to relax.
With financial data, we analyze asset performance both at the conditional mean and in the tails, at low, medium, and high returns. FM selects, without any ad hoc procedure, two classes of homogeneous assets defining two differing portfolios based on the response to the explanatory variables. This example shows once again the relevance of allowing composition to change across locations. At low returns, the groups show different proportions with respect to groups at high returns. At the lower tail the group difference is not statistically significant; although each group responds in its own way to the explanatory variables, the overall group difference is not statistically relevant, and the two portfolios exhibit comparable performance. However, at the mean and top quartile, one group significantly improves upon the other. Class 1 assets turn out to be preferable at and above the mean, while class 2 collects more risky assets. This highlights the validity of grouping data in more homogenous classes not just at the mean, but also—as in this case—at the upper tail, without imposing the same group composition at differing locations of the conditional distribution of the dependent variable. The group composition does change in moving from the left to the right tail, and the probability of belonging to the best portfolio increases as the location goes up at higher returns.
In summary, in the student proficiency case, the two groups of highly performing students show a minimal difference between one another, while a significant disparity exists among the groups at and below the mean for the lower-scoring students. In contrast, in the financial case, the two groups of lower-performing assets are quite similar, whereas the difference between groups becomes more pronounced at the top quartile. These examples show the presence of differing processes operating away from the mean. Uncovering differing behaviors enhances the understanding of the variables being investigated and can be a valuable insight for policymakers.
Relaxing the constant group probability constraint impacts not only the class probability, but also the estimated coefficients within each class. Indeed, the FM components, class probability and class regressions, are jointly estimated through iterations. Alternative approaches typically address these components separately, one at the time, first estimating the class composition and then separately computing the within-class regressions, possibly inducing econometric issues impairing the validity of the results.
Section 2 and Section 3 describe the FM estimator at the mean and in the tails, respectively. Section 4 and Section 5 report the Monte Carlo experiments and their results. Section 6 discusses two case studies, one on student proficiency and the other on portfolio choice. The final section concludes the analysis, while Appendix A reports the codes used to compute FM in the tails using both expectiles and quantiles.

2. Methods: The Finite Mixture Estimator

In the linear regression model, yi is the dependent variable, xi collects the ith observation of the k explanatory variables, ui is the error term and β is the vector of regression coefficients. The standard finite mixture regression estimator relates the dependent variable yi to xi within each subgroup while a latent variable zi denotes the subgroup to which each unit belongs. For the ith unit in the kth group, with k = 1, …, G, the linear regression model is
yi = xiβk + ui
The latent variable zi defines the probability of a generic unit belonging to the kth component, and it defines the πk probabilities.
π(zi) = exp(zi)/Ʃk=1,G exp(zi)
For each observation, the probability of inclusion in a group, πk, is the function of the unobservable latent variable zi. The latter captures population heterogeneity and assumes values 0 or 1 to define the exclusion or the inclusion of the ith observation in the kth group.
The likelihood combines the conditional likelihood of each class weighted by the associated probability.
k = 1 , G π k i = 1 ,   n f ( y i | x i , β k )
Unfortunately, we do not know which group an observation belongs to, and the observed sample is incomplete since the latent variable zi is not observable. However, zi can be approximated by a subset of group-specific explanatory variables xiq, with q p, and this subset is used to model the prior probability of component membership. A multinomial logit model computes πk as
π k = exp ( g ( x iq ) ) / Ʃ k exp ( g ( x iq ) )
where g(xiq) is the function relating the probability of being in the kth class to the characteristics of the xiq data set.1 It is a model-based clustering approach where the mixture proportions are defined by a logistic model, and the mixture components are linear regressions. The outcome variable distribution depends on both the covariate xi and the latent cluster membership variable zi, as approximated by xiq.
The estimation relies on an iterative process. The Expectation-Maximization (EM) algorithm (Dempster et al., 1977) computes the probability of belonging to a given group by implementing Equation (4), which defines the probability weights. Next, the weighted regression parameters are estimated in (3), and the iterations update weights and regression coefficients until convergence is reached.

3. Method: The FM Estimator in the Tails

The purpose is to implement the finite mixture not only on average, at the conditional mean, but also in the tails of both components of the FM estimator. The FM group analysis reduces the heterogeneity of the data, selecting more homogeneous subsets. The tail analysis within FM considers the impacts of the explanatory variables away from the mean, at low/high values of the dependent variable within each class. In Furno (2023) and in Furno and Caracciolo (2024), the heterogeneity is modeled by splitting the sample into homogeneous sub-groups at the conditional mean. Once the clusters are defined, quantile regressions are estimated in the tails within each class, and the grouping does not vary across quantiles. This approach excludes by assumption any heterogeneity in the probability of belonging to a group. The probability may change as well according to the selected location. For instance, in one of the following case studies on students’ performance, the probability of being in the best group is higher in the left tail, at low scores, while it is lower in the right tail of the distribution. The probability undergoes changes, and the novelty is in modeling the changing group probability together with the changing explanatory power of the covariates at the differing locations, i.e., moving both probability and regression estimates away from the mean implementing expectiles or quantiles estimators.

3.1. FM-Expectiles

To move away from the average Equation (4), we consider expectiles (Newey & Powell, 1987) that provide a shifting weight to move the logistic model at various locations of the conditional distribution of the dependent variable. Equation (4) is modified to include an asymmetric weighting system wi that moves the equation up or down toward the tails,
π k = P ( w i z i = 1 | X ) = w i   exp ( x i β k ) { Σ exp ( x i β k ) }
where the asymmetric weighting system used to define location is w i = { θ   i f   u i > 0 1 θ   e l s e w h e r e , θ is the chosen location, and u is the regression error. For instance, to compute the θ   = 25th expectile, w i   assigns weights 0.75 to those observations below the regression to attract the estimated equation toward the lower tail, and it assigns weights of 0.25 to the observations above it.2
The regression model within each group is computed away from the mean as well, and in the regression k π k f ( y i | x i , β k ) , π k   is   approximated   by   its   estimated   values   π k ^   i n   ( 5 ) ,     k π k ^ f ( y i | x i , β k ) , and the standard linear regression OLS estimator is replaced by expectiles, with the following objective function
y i > x i β k θ ( y i x i β k ) 2 + y i < x i β k ( 1 θ ) ( y i x i β k ) 2
The FM-expectile estimator iterates between Equations (5) and (6) until convergence, at the given location θ , thus computing both class probability and class regression coefficients in the tail. The location θ away from the mean affects both terms of the finite mixture estimator, the logit and the regression model. Equations (5) and (6) provide the FM-expectile estimator.3

3.2. FM-Quantile

Besides expectiles, it is possible to implement finite mixtures at the quantiles, which provide more robust results. Instead of Equation (6), the regression step is computed as
y i > x i β k θ | y i x i β k | + y i < x i β k ( 1 θ ) | y i x i β k |
Koenker (2005) replaced the L2 norm with the L1 norm, and the weights generated by this equation pre-multiply the data when implementing the FM routine.4 The class probability equation in (4) at the quantiles becomes
Q y ( θ ) = [ exp   ( x i β ) y max + y min   ] / [ 1 + exp   ( x i β ) ]
to define the logistic quantile regression for outcomes that are bounded within ymax and ymin, where ymax and ymin are not the maximum and the minimum but the bounds of the interval of the outcome variable (Orsini & Bottai, 2011).5 The FM-quantile iterates between (7) and (8) until convergence.

4. Simulations

To better analyze the FM performance in the tails, the same simulation scheme as in Furno (2023) and in Alfò et al. (2017) is implemented.6 The model is a random coefficients equation, defined as
yik = (β1 + b1i) + (β2 + b2i) x2ik + β3 x3ik + εik,      i = 1, …, n      k = 1, …, G
with β3 being the sole non-random term. The error term is in turn a standard normal or a Student t with 3 degrees of freedom, t3; the β vector assumes values (β1, β2, β3)′ = (100, 2, 1)′; b1i and b2i are individual specific random parameters distributed in turn as a standard normal or a Student t with 3 degrees of freedom; x2ik is defined as the sum of two independent standard normal variables; x3ik is a binomial B(100;0.5); the sample size is n = 100 and there are 1000 iterations for each experiment.
The first set of experiments consider the standard normal distribution for b1i, b2i and the error term. The second set of experiments considers Student t distributions for the errors εik. The second set of experiments looks at the performance in cases of heavy tail distributions.
The estimator considered is the finite mixture model at the quantiles, together with its variants at the expectiles. The FM-quantile and FM-expectile estimators are computed at locations θ = 0.25 and 0.75, respectively at the first and third quartiles.
To compute the two finite mixture estimators of Equations (5)–(8), the data set is divided into k = 2 sub-groups; there is an estimated β vector for each sub-group, and the results are reported in Table 1. The relative bias and the average Akaike are reported together with their dispersion in the 1000 iterations.
Next, the standard normal error distribution is replaced by a 10% contaminated normal, with a contaminating distribution N(50, 100), while the Student t error distribution is replaced with a 10% contaminated Student t with a t6 central distribution and a t3 contaminating distribution. This set of experiments looks at the performances of the finite mixture estimators when there are two distributions generating the sample. This set of experiments is particularly suitable for the finite mixture estimators, with the central distribution describing one group and the contaminating distribution describing the other group of tail observations. Bartolucci and Scaccia (2005) specifically suggest implementing finite mixture regressions to model mixtures of normal errors. Of the two contaminating schemes selected, the contaminated normal defines a bimodal distribution with very different variances, while the Student t contamination defines a unimodal distribution.

5. Simulation Results

Table 1 presents the estimates computed in two homogeneous subgroups with two estimated β vectors, one for each group. The estimated coefficients do change, both across quartiles and between groups. The relative bias is generally small throughout the table, in both the FM-quantile and FM-expectile estimators, and when the errors are normal or follow a Student t distribution. Comparing these results with those reported in Furno (2023), Table 2, the constant term and the β3 coefficients present much smaller relative biases; the β2 relative bias is slightly larger, although comparable in many experiments, and its dispersion is generally smaller; the Akaike values are now notably smaller. Figure 1 reports the empirical distributions of the slope coefficients of Equation (9), in the experiment with normal distributions in the top graphs and in the case of Student t distributions. The β2 + b2i distributions are more dispersed than the β3 distributions since the latter is not a random coefficient. The Student t experiments are more dispersed than the standard normal ones. In all the experiments of this table, at the first quartile, group 1 overestimates the true value and group 2 underestimates it; the opposite occurs at the top quartile, as can be seen in Figure 1.
In Table 2, the errors are generated by contaminated distributions. Here, the classes mirror the contamination. The sample is divided in two sub-groups yielding two vectors of estimated coefficients: one vector for the group generated by the central distribution and the other for the data generated by the contaminating distribution. The relative bias in this table is sizably smaller throughout than in Table 4 in Furno (2023) with the sole exception of the β2 expectile results, which are slightly larger. Figure 2 reports the empirical distributions of the slope coefficients from the experiments with contaminated distributions. In this figure, the experiments with standard normal distributions exhibit the greatest dispersion, and the FM-expectile estimates are more dispersed than the FM-quantile results.
In all the experiments, the distributions of the non-random coefficient b3 = β3 are less dispersed than the b2 = β2 + b2i distributions.

6. Case Studies

6.1. School Data

To look at the difference between this approach and that in the previous research by Furno (2023), the same data set is analyzed. The math scores of 15-year-old Italian students taking the OECD-PISA test on math proficiency in 2009 are considered. The math score (math in the table) is explained by school characteristics to assess the link between performance and school structures. We select only a small number of explanatory variables—field (academic and technical), school size (schlsize in the table), student–teacher ratio (stratio), and teacher shortage (tcshort)—to streamline the discussion of the results. The field is defined by the dummy variables academic and technical to separate students enrolled in vocational fields; school size collects the number of students in the school; student–teacher ratio is the ratio between number of students and teachers, and tcshort is an index of staff shortage. The sample comprises n = 26,240 observations and Table 3 reports the summary statistics. Table 4 collects the FM-quantile results at 0.25 and 0.75, together with the standard FM results at the center, at the conditional mean 0.50. The quantile weights all equal to 0.5 to define the median regressions, do not modify the FM results at the conditional mean.7
Students are grouped in two classes, with one group improving upon the other in math proficiency. The first line of the table reports the estimated group differences. Looking at the high-proficiency students, their math scores are split into excellent versus just-above-average, and the difference turns out to be significant but rather small, with class 2 improving upon class 1 by 0.2014, with z = 1.87. In the left tail, looking at low-proficiency students, the groups show just-below-average versus very poor performance, and their difference is almost three times as large as the above difference between the high-scoring students’ groups. Group 2 shows a worse performance x with respect to group 1 by −0.5899 (z = −6.80) at the first quartile; and group 2 significantly worsen with respect to class 1 by −1.39 (z = −2.47) at the mean. For the top math scores, these results are reversed: group 2 performs better than group 1, although by a small amount. The opposite occurs at and below the mean, with group 1 outperforming group 2 by a larger margin. The last two rows of this table report the latent class probability, i.e., the groups’ compositions’. These probabilities do change across locations, assuming an inverse u-shaped pattern in class 1, which is higher at the mean, coupled with a u-shaped pattern in class 2. At the mean, the class probabilities are 80% and 20%, just as in Furno (2023), but away from the mean the probabilities between the two classes are more balanced: respectively, 65% and 35% in the left tail that turns to 45% and 55% at the upper tail. The groups’ proportions do change across locations, and are more balanced in the tails than at the mean. This shows the relevance of shifting both components in FM, both the logit—defining the class probabilities—and the class regressions. In contrast to previous work, here, the location changes in the two FM components, shifting both the logit model and the regression model at the same location of the conditional distribution of the dependent variables.
The changing group probability impacts the final FM results. Group difference is statistically significant throughout, assuming great values at and below the mean, while declining at the top quartile: the worse performance of class 2 at and below the mean is reversed at the top quartile, with class 2 collecting the best-performing students.
In the table, school size is not significant for group 2 at and above the mean. Between groups, divergence due to curricula is quite sizable; the academic and technical tracks improve with respect to vocational schools, particularly for group 2. The widest group discrepancy occurs at the mean.

6.2. Financial Data

Data from for the Center for Research in Security Prices in year 2022 are here analyzed. Table 5 reports the summary statistics of the variables in the model. The yearly average of monthly total return (return in the table) in our model is related to its lagged value (return1)—its recent past; the yearly average of monthly market capitalization (capit)—its market value; earnings before interest (earning)— profitability; the yearly average of monthly return on the Standard and Poor 500 index (spreturn)—a benchmark of market performance, and book-to-market value (bookmkt) defined as the difference between total assets (asset) minus total liability (liability) divided by the yearly average of monthly market capitalization, bookmkt = (asset−liability)/capit. The analysis focuses on the ordinary shares, NS type, in the CRSP data set. In addition, the observations with very low market capitalization, those below the 5th percentile, are excluded from the sample. The usual approach in cross-section studies consists in grouping similar assets in a first stage to define portfolios. Here, the selection of assets into homogeneous classes/portfolios is model-based and performed by the FM estimator, avoiding hand-made ad hoc selections. Table 5 collects the summary statistics.
Table 6 collects the estimated regression coefficients computed, respectively, by OLS, median (Koenker, 2005) and robust regression (Huber, 1981). The three estimators present comparable values, not statistically different from one another, with book-to-market not statistically significant throughout. The sole exception is given by the capitalization coefficient, the market value, which is not significant in OLS and is significant elsewhere. This suggests the need for further scrutiny, and possibly some trimming. Instead of defining a subjective trimming criterion, the latter is defined by the robust regression estimator. The latter excludes 58 observations from the sample, anomalous values that are either too small or too large with respect to the bulk of data. The robust regression induces a 1.9% trimming, and the final sample size is n = 2912. Figure 3 compares the box plot of returns before and after trimming, and the impact of trimming is quite evident in the graphs.
Table 7 reports the expectile regression estimates at locations 0.25, 0.50 and 0.75 for low/mean/high returns, respectively. In this table, most coefficients are statistically significant and present a pattern across the various locations. Book-to-market and earnings increase their impact in moving to the 0.75th location, while spreturn and lagged return both decrease at and above the mean. Market capitalization shows a statistically significant inverse U pattern.
Recalling that expectiles are just a location device, and that they are not robust and may still be attracted by high and low observations, Table 8 reports the quantile results, which are generally more robust. The table shows a non-significant impact of book-to-market, while the other estimated coefficients are significant and generally decline across quartiles. The spreturn coefficient is smaller than 1 and can be interpreted as low dependence on market conditions, while a negative constant term turning to positive at the top quartile signals returns smaller than a safe asset but at 0.75. Earning and market capitalization have a small and more stable impact across quartiles.
Table 7 and Table 8 show that the estimates do change across expectiles and quantiles, with estimates statistically different across locations. They display a generally diminishing impact in moving from the left to the right tail.
Next, we look for the presence of more homogeneous subsets in the data by implementing FM, since Table 6, Table 7 and Table 8 do not consider grouping the data into classes, analyzing all the observations together. Furthermore, by saving the weights generated in the expectile/quantile regressions of Equations (6) or (7), it is possible to move FM to the tails. The tail results are obtained by weighing all the variables by the location terms generated when computing the regression away from the conditional mean.8 Thus, besides analyzing the sample in more homogeneous subsets, as in the standard FM approach, the location weights allow for moving both logit and regression components to the tails, yielding the FM-quantile and FM-expectile estimators.
Table 9 collects FM-quantile results at the 0.25 and 0.75 locations, to be compared with the standard FM results at 0.50 reported in the last columns of this table.9
We select two classes of homogeneous assets, defining two differing portfolios depending on their response to the explanatory variables, and we analyze them at various locations, looking at low, medium and high returns. If needed, it is possible to further refine the analysis by increasing the number of classes. Classes are model-based and are not forced to remain constant across locations, and this is the novelty of our approach. At low returns, we sort poor from just-below-average assets. Analogously, at high returns, FM-quantile sorts excellent from just-above-average assets. The top line of the table reports the total difference between classes/portfolios. While the class difference is not significant at the lower tail, 0.239 (z = 1.16), class 2 collects the worse-performing assets at and above the mean, with an increasing difference equal to −0.361 (z = −3.74) at the mean as computed by the standard FM, and to −0.679 (z = −5.18) at 0.75 in the FM-quantile case. Class 1 is characterized by a better performance at θ = 0.50 and 0.75, where the class difference is statistically significant and increases at the top quartile. The significance of the total difference between the two groups shows that analyzing data all together disregards relevant group discrepancies.
At the bottom of the table, in the last two rows, the probability of belonging to each class is reported. At θ = 0.25, class 2 shows higher probability, 56%. At the other locations the group probabilities are instead higher in class 1, reaching 59% at the mean and 66% at the top quartile for the best-performing group. In this table, the group probability is a function of the same explanatory variables of the regression model. Later, we will introduce other variables to define the class probability as differing from the regression explanatory variables, their lagged values.
So far, there are two factors confirming the validity of a group analysis at various locations, as follows: differing group probabilities across location, which show the importance of shifting the logit together with the regression component in FM; significant between-class differences at and above the mean.
Moving to the estimated coefficients, the standard FM results are reported on the right-hand side of this table. On average, the best-performing class 1 assets are characterized by a positive impact of lagged returns and earnings, negative market capitalization, and a negative constant. The spreturn and book-to-market estimates are not significant in this group. Class 2 shows a large positive spreturn, signaling a high response to market performance, and a positive constant term suggesting returns higher than a safe asset. These are typical features of riskier shares. All the other coefficients do not statistically differ from zero.
At the lower tail, the two groups behave similarly, and their overall difference is not significant. However, their estimated coefficients do differ. Class 1 is positively related to lagged returns and earnings, while class 2 is driven by earnings and by negative market capitalization.
At the top quartile, once again, class 1 collects the best-performing assets. The negative impact of capitalization is statistically significant, together with the positive impact of lagged returns and earnings. Class 2 shows positive book-to-market and spreturn values, while lagged returns and earnings have a negative impact.
Finally, the introduction of lagged book-to-market, lagged spreturn, and lagged market capitalization in the definition of the group probability, the xiq set of variables in Equation (4), can further refine these findings.10 The impacts of these variables on the group probabilities in the logit model are reported in Table 10. The constant term, here representing the difference in class 2 with respect to class 1, is negative at and above the mean, while it is not significant at the lower quartile, confirming the findings in the top row of Table 9. When selecting lagged variables to explain FM class probability, these variables are all significant at the mean. Lagged capitalization is the sole significant factor at the first quartile. Past book-to-market and past capitalization are relevant at the top quartile. Lagged spreturn is not significant in defining class probability in the tails.

7. Conclusions

A finite mixture model is considered to divide data into homogeneous subsets without ad hoc intervention, employing a model-based clustering approach. FM assumes that the population consists of homogeneous subgroups, with a probability density function defining groups. Each of them is characterized by group-specific sets of regressions. The model involves two iteratively estimated components: one defining the group probability and the other computing the regression within each class. The impact of predictors varies between classes, and these differences determine the class membership.
The class analysis of FM is extended to the tails to define the FM-quantile and the FM-expectile estimators. The novelty of this proposal lies in allowing both class probability and class-specific regressions to vary across quantiles or expectiles. Previous research defines the FM-quantile by constraining the groups to have the same average class probabilities at all locations. By relaxing this constraint, both class probability and class regressions coefficients shift at the selected quantile or expectile. In the case studies analyzed, class probabilities do vary across locations, and the class composition has an impact on the class regression estimates as well, due to the iterative Expectation-Maximization algorithm.
A Monte Carlo experiment compares results with those from Furno’s (2023) constant probability analysis. The relative biases and the average Akaike Information Criterion in the changing probability experiments are notably smaller than in the previous research based on constant probability.
The case studies analyze data on students’ proficiency in a math test and data on assets returns and portfolio selection. In both studies, the class composition changes across locations. In the students’ performance case, on average, most students are gathered in the best group—80%. In the tails, the best group remains the majority, but the class probabilities are more balanced, accounting for 64% in the left tail and 55% in the right tail. At higher scores, the two groups show only slight differences, but the gap between groups widens at and below the mean, particularly among the low-scoring students.
In the asset portfolios analysis, class composition does change as well. For the best-performing class 1, the probability increases from 44% in the left tail to 59% at the mean and 66% in the right tail, showing a rising trend across quartiles. Changing the class proportions does change the class regressions estimates since the two FM components, class probability and class regressions, are jointly estimated in the FM approach. In this case study, the group difference is not significant at low returns, while it gets larger and more significant at, and more so above, the mean. The class 2 portfolio turns out to collect more risky assets.
In the student proficiency case, the two groups of high-performing students show minimal differences, while a significant group disparity occurs at and below the mean. In the financial case, the groups of lower-performing assets are quite similar, while the difference between groups becomes more pronounced at the top quartile. Uncovering these differing behaviors enhances the understanding of the variables being investigated and can offer valuable insights for policymakers.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data will be available on request.

Conflicts of Interest

The author declares no conflicts of interest.

Appendix A. Stata Codes

reg y x1 x2
predict resols, resid
*FM at the mean
fmm 2: regress y x1 x2
*Akaike and log likelihood
estat ic
*probability to belong to a group
estat lcprob
*expectile regression and definition of weights*first quartile regression weights
reg y x1 x2 +qreg y x1 x2, q(0.25) +++
gen eweight = 1predict qres, resid
scalar tau = 0.25gen qweight = 0.25 if qres > 0 & qres!=.
gen wy = eweight*yreplace qweight = 0.75 if qres < 0 & qres!=.
gen wx1 = eweight*x1
gen wx2 = eweight*x2
forvalues i = 1(1)8 {
 predict resid, resid
 replace eweight = 2*tau if resid >= 0
 replace eweight = 2*(1-tau) if resid < 0
 replace wy = eweight*y
 replace wx1 = eweight*x1
 replace wx2 = eweight*x2
reg wy wx1 wx2 ++
drop resid
}
*expectile FM*quantile FM
fmm 2: regress y x1 x2 [pw = eweight]fmm 2: regress y x1 x2 [pw = qweight]
estat icestat ic
estat lcprobestat lcprob
+ can be replaced by logit d x1 x2
++ can be replaced by logit wd wx1 wx2+++ can be replaced by lqreg qd qx1 qx2
where d is a location dummy assuming value equal to 1 at the chosen expectile/quantile, zero otherwise.
To select the variables defining the group probabilities the FM code becomes
fmm 2, lcprob (z1 z2): regress y x1 x2 [pw = weight]
where z1 and z2 are the explanatory variables defining the group probability. They can or cannot differ from the variables of the regression model. The weight within square brackets is qweight or eweight depending on the selected estimator, FM-quantile or FM-expectile.

Notes

1
The variables defining the group probability could differ from the variables defining the regression model. In the case studies, the explanatory variables of π k and y i coincide. However, the financial model has been re-estimated by setting lagged variables as predictors of group probability in xiq.
2
The stata code to compute the weights in the expectile logistic regression of Equation (5) is logit eweight-depvar eweight-indepvar within the iterative loop that computes the location expectile weights; eweight are defined by w i = θ or w i = 1 − θ as in (5) and they modify the regression variables into eweight-depvar and eweight-indepvar, as in (6) (see the Appendix A).
3
In the expectile case, the regression code to compute (6) is reg eweight-depvar eweight-indepvar within the iterative loop that computes the location weights (see the Appendix A).
4
The stata code in the quantile regression case of Equation (7) at the θ quantile is qreg depvar indepvar, q( θ ), and the residuals allow us to define the quantile weights (see the Appendix A).
5
In the quantile case, the stata code at the θ quantile is lqreg depvar indepvar, q(θ) (Orsini & Bottai, 2011), and the residuals allow us to define the weights (see the Appendix A).
6
The simulations and all the empirical analyses are implemented in Stata, version 15.
7
We compute only FM-quantile for comparability’s sake.
8
The weights can be determined starting from the quantile/expectile logit model as well.
9
We present the FM-quantile due to its robustness. The FM-expectile results are available on request.
10
The full set of results is available on request.

References

  1. Alfò, M., Salvati, N., & Ranalli, M. G. (2017). Finite mixtures of quantile and M-quantile regression models. Statistics and Computing, 27, 547–570. [Google Scholar] [CrossRef]
  2. Bartolucci, F., & Scaccia, L. (2005). The use of mixtures for dealing with non-normal regression errors. Computational Statistics and Data Analysis, 48, 821–834. [Google Scholar] [CrossRef]
  3. Betrail, P., & Callavet, F. (2008). Fruit and vegetable consumption: A segmentation approach. American Journal of Agricultural Economics, 90, 827–842. [Google Scholar]
  4. Breckling, J., & Chambers, R. (1988). M-quantiles. Biometrika, 75, 761–771. [Google Scholar] [CrossRef]
  5. Caudill, S. B., & Mixon, F. G., Jr. (2016). Estimating class-specific parametric models using finite mixtures: An application to a hedonic model of wine prices. Journal of Applied Statistics, 43(7), 1253–1261. [Google Scholar] [CrossRef]
  6. Compiani, C., & Kitamura, Y. (2016). Using mixtures in econometric models: A brief review and some new results. The Econometrics Journal, 19, C95–C127. [Google Scholar] [CrossRef]
  7. Dempster, P., Laird, N., & Rubin, D. (1977). Maximum likelihood for incomplete data via the EM algorithm. Journal of the Royal Statistical Society (B), 39, 1–38. [Google Scholar] [CrossRef]
  8. Furno, M. (2023). Computing finite mixture estimators in the tails. Journal of Classification, 40, 267–297. [Google Scholar] [CrossRef]
  9. Furno, M., & Caracciolo, F. (2024). Finite mixture model for the tails of distribution: Monte Carlo experiment and empirical applications. Statistical Analysis and Data Mining, 17, 1–15. [Google Scholar] [CrossRef]
  10. Huber, P. (1981). Robust statistics. Wiley. [Google Scholar]
  11. Khalili, A. (2010). New estimation and feature selection methods in mixture-of-experts models. The Canadian Journal of Statistics, 38, 519–539. [Google Scholar] [CrossRef]
  12. Khalili, A., & Chen, J. (2007). Variable selection in finite mixture of regression models. Journal of the American Statistical Association, 102, 1025–1038. [Google Scholar] [CrossRef]
  13. Koenker, R. (2005). Quantile regression. Cambridge University Press. [Google Scholar]
  14. Liang, J., Chen, K., Lin, M., Zhang, C., & Wang, F. (2018). Robust finite mixture regression for heterogeneous targets. Data Mining and Knowledge Discovery, 32, 1509–1560. [Google Scholar] [CrossRef]
  15. McLachlan, G., & Peel, D. (2004). Finite mixture models. Wiley Series in Probability and Statistics. John Wiley & Sons. [Google Scholar]
  16. Newey, W., & Powell, J. (1987). Asymmetric least squares estimation and testing. Econometrica, 55, 819–847. [Google Scholar] [CrossRef]
  17. Orsini, N., & Bottai, M. (2011). Logistic quantile regression in stata. The Stata Journal, 11, 327–344. [Google Scholar] [CrossRef]
  18. Städler, N., Bühlmann, P., & van de Geer, S. (2010). L1-penalization for mixture regression models. Test, 19, 209–256. [Google Scholar] [CrossRef]
  19. Van Horn, L., Smith, J., Fagan, A., Jaki, T., Feaster, D., Hawkins, D., & Howe, G. (2013). Not quite normal: Consequences of violating the assumption of normality in regression mixture models. Structural Equation Modeling, 19, 227–249. [Google Scholar] [CrossRef]
Figure 1. Empirical distributions in 1000 iterations of the slope regression coefficients within each group. The b2 distributions, with b2 = β2 + b2i, are more dispersed than the distributions of the non-random b3, where b3 = β3, and the Student t experiments are more dispersed than the standard normal ones.
Figure 1. Empirical distributions in 1000 iterations of the slope regression coefficients within each group. The b2 distributions, with b2 = β2 + b2i, are more dispersed than the distributions of the non-random b3, where b3 = β3, and the Student t experiments are more dispersed than the standard normal ones.
Jrfm 18 00177 g001aJrfm 18 00177 g001b
Figure 2. Empirical distributions in 1000 iterations of the regression slope coefficients within each group. The contaminated normal distributions are highly dispersed, significantly more than in the Student t experiments. The b2 distributions, with b2 = β2 + b2i, are more dispersed than the distributions of the non-random b3, where b3 = β3.
Figure 2. Empirical distributions in 1000 iterations of the regression slope coefficients within each group. The contaminated normal distributions are highly dispersed, significantly more than in the Student t experiments. The b2 distributions, with b2 = β2 + b2i, are more dispersed than the distributions of the non-random b3, where b3 = β3.
Jrfm 18 00177 g002aJrfm 18 00177 g002b
Figure 3. Total returns box plot before and after 1.9% trimming driven by the robust regression.
Figure 3. Total returns box plot before and after 1.9% trimming driven by the robust regression.
Jrfm 18 00177 g003
Table 1. Standard normal and Student t, t(3); two groups; n = 100; 1000 replicates.
Table 1. Standard normal and Student t, t(3); two groups; n = 100; 1000 replicates.
Standard Normal Errors
quartiles0.250.75
group1group2group1group2
FM-quantile
relative bias std. d.relative bias std. d.relative bias std. d.relative bias std. d.
beta1−0.01260.0630.00380.0610.00660.0550.01200.062
beta20.11780.461−0.12580.432−0.08470.4350.08370.457
beta30.00210.125−0.01110.119−0.01010.109−0.00060.124
Akaike167.3  7.90 171.2  7.86
FM-expectile
beta1−0.01460.0630.00410.0610.00630.0520.00850.059
beta20.08710.437−0.06840.407−0.09690.4110.08070.419
beta30.00580.128−0.01240.122−0.00930.1050.00580.117
Akaike344.4  16.6 343.6  17.1
Student t distributions
quartiles0.250.75
group1group2group1group2
FM-quantile
relative bias std. d.relative bias std. d.relative bias std. d.relative bias std. d.
beta1−0.01970.1640.00410.1330.01370.150−0.00890.261
beta20.07390.594−0.05990.571−0.07370.5200.05840.693
beta30.01000.333−0.00980.271−0.02590.3080.04980.515
Akaike180.1   9.21 184.4   9.61
FM-expectile
beta1−0.02250.1480.01240.1390.00780.1510.01480.310
beta20.08730.560−0.07490.551−0.07860.4900.05631.05
beta30.01610.296−0.02710.278−0.01200.2960.00020.649
Akaike368.7  20.0 369.4  18.5
Table 2. Contaminated error distributions; two groups; n = 100; 1000 replicates.
Table 2. Contaminated error distributions; two groups; n = 100; 1000 replicates.
10% Normal Contamination, N(0, 1) in the Center and N(50, 100) in the Tails
quartiles0.250.75
group1group2group1group2
FM-quantile
relative bias std. d.relative bias std. d.relative bias std. d.relative bias std. d.
beta10.00070.1620.01010.0760.00520.0610.00750.067
beta2−0.14220.8030.08650.457−0.09190.4470.07520.503
beta30.00090.2900.00240.151−0.00670.1230.01010.175
Akaike171.8  8.32 171.5  8.64
FM-expectile
beta1−0.30553.700.62463.950.42831.950.18664.06
beta20.83589.41−1.11213.8−0.73385.071.28813.3
beta30.52147.35−0.73677.85−0.64633.620.90268.07
Akaike502.07  83.4083.40406.7  42.642.6
10% contaminated student t distributions, t(6) in the Center and t(3) in the Tails;
quartiles0.250.75
group1group2group1group2
FM-quantile
relative bias std. d.relative bias std. drelative bias std. d.relative bias std. d.
beta10.00340.0570.01030.0600.00850.0570.01180.063
beta2−0.09990.4410.09880.436−0.08810.4320.08630.437
beta3−0.00320.1130.00230.120−0.01330.114−0.00020.123
Akaike171.6  7.94 171.8  8.08
FM-expectile
beta1−0.01620.1580.00240.093−0.00040.1120.00280.071
beta20.04170.578−0.08180.464−0.08590.4030.07000.397
beta3−0.00310.318−0.01920.185−0.00720.2310.01170.139
Akaike351.7 19.0 354.4 18.4
Table 3. Summary statistics, n = 26,240.
Table 3. Summary statistics, n = 26,240.
MeanStd. Dev.
math 493.768 84.7263
academic0.46548670.4988168
technical0.32815040.4695487
stratio 9.121897 2.767687
schlsize 682.1716 371.274
Table 4. Finite mixture at the quantiles, n = 26,240.
Table 4. Finite mixture at the quantiles, n = 26,240.
Location0.250.50 = Mean0.75
Coef.zCoef.zCoef.z
ΔClass 2−0.5899−6.80−1.397−2.470.20141.87
Class 1
mathCoef.zCoef.zCoef.z
academic64.4582417.7547.841082.9168.3273215.92
technical51.7872720.6631.220542.3154.0614318.37
stratio3.7892556.423.1690644.053.5588214.98
schlsize0.01878856.090.02783236.200.02090335.72
constant314.187166.01396.915370.34358.431253.76
Class 2
mathCoef.zCoef.zCoef.z
academic86.7191823.99204.055213.3486.3981927.73
technical63.8478617.24175.26969.4563.2456518.47
stratio2.6063434.892.2501821.502.5792555.98
schlsize0.00908632.07−0.0287043−1.220.00622381.59
constant445.105170.89375.938933.63484.1338107.46
log likelihoodAkaikelog likelihoodAkaikelog likelihoodAkaike
−308,776.3617,578.7−820,835.81,641,698−308,039.4616,104.9
Latent class probability
Class 10.643350.801750.44982
Class 20.356650.198250.55018
Table 5. Summary statistics, year 2022, n = 2970.
Table 5. Summary statistics, year 2022, n = 2970.
MeanStd. Dev.
return−0.028730.07589
asset total12,578.51111,076.1
liability total9851.449100,613.1
capit1.02 × 1077.23 × 107
bookmkt0.0006620.001517
earning829.3994983.197
spreturn−0.0157480.00923
Table 6. Estimated coefficients, n = 2970.
Table 6. Estimated coefficients, n = 2970.
Robust RegressionOLSMedian
returnCoef.tCoef.tCoef.t
bookmkt0.05990.100.953271.050.477560.71
capit −5.42 × 10−11−2.47−4.84 × 10−11−1.51−6.50 × 10−11−2.77
return1 0.055234.410.068913.770.060974.55
earning1.31 × 10−65.621.36 × 10−62.921.55 × 10−64.55
spreturn0.577305.620.784185.240.561975.12
constant−0.01267−6.42−0.01567−5.45−0.00890−4.22
R2/Pseudo R2 0.01740.0077
Note: In italics, the non-significant estimated coefficients.
Table 7. 1.9% trimmed expectile regression estimates, n = 2912.
Table 7. 1.9% trimmed expectile regression estimates, n = 2912.
0.250.50 = OLS0.75
returnCoef.tCoef.tCoef.t
bookmkt −5.31796−7.650.2143770.315.653518.88
capit−7.72 × 10−11−2.09−5.01 × 10−11−2.09−7.90 × 10−11−4.35
return10.19267913.260.066954.77−0.06558−5.63
earning1.27 × 10−63.681.29 × 10−63.711.57 × 10−65.18
spreturn3.6555437.570.598634.87−1.08261−15.2
constant0.009986.23−0.015938−6.93−0.025996−19.7
R20.36880.02060.1240
Note: In italics, the non-significant estimated coefficients.
Table 8. 1.9% trimmed quantile regression estimated coefficients, n = 2912.
Table 8. 1.9% trimmed quantile regression estimated coefficients, n = 2912.
0.250.50 = Median0.75
returnCoef.tCoef.tCoef.t
bookmkt−0.17182−0.14−0.55413−0.820.323760.58
capit−4.14 × 10−11−0.95−6.34 × 10−11−2.73−6.90 × 10−11−3.59
return10.0951653.750.0637084.690.0231322.06
earning1.40 × 10−62.221.49 × 10−64.441.26 × 10−64.52
spreturn0.769733.460.5400494.540.566495.75
constant−0.04149−9.97−0.00806−3.620.0149738.12
pseudo R20.01370.00890.0054
Note: In italics, the non-significant estimated coefficients.
Table 9. Finite mixture quantile estimates, 2 classes, n = 2912.
Table 9. Finite mixture quantile estimates, 2 classes, n = 2912.
Location0.250.750.50
Coef.zCoef.zCoef.z
ΔClass 20.2391.16−0.6797−5.18−0.3612−3.74
Class 1
returnCoef.zCoef.zCoef.z
bookmkt−0.673230.140.219480.200.010120.01
capit3.60 × 10−101.43−1.34 × 10−10−3.36−2.08 × 10−10−2.46
return10.1033844.080.082713.330.0871883.87
earning0.0000423.412.79 × 10−63.029.77 × 10−65.78
spreturn0.8089551.440.448761.250.177170.99
constant−0.0783777.650.005887−0.97−0.036031−9.99
Class 2
returnCoef.zCoef.zCoef.z
bookmkt0.6501010.390.331532.081.153210.77
capit−1.73 × 10−11−1.85−1.33 × 10−12−1.18−1.46 × 10−11−1.35
return10.0064520.19−0.04154−2.710.0083190.62
earning3.52 × 10−71.69−3.24 × 10−7−2.281.58 × 10−71.01
spreturn0.682570.890.6317016.02.0888813.4
constant0.0123160.980.02082518.20.023718.49
Akaikelog likelihoodAkaikelog likelihoodAkaikelog likelihood
−2987.491506.75−3303.871664.93−8863.634444.81
Latent class probabilities
Class 10.4400.6640.589
Class 20.5600.3360.411
Note: In italics, the non-significant estimated coefficients.
Table 10. Difference of class 2 with respect to class 1, as explained by the lagged variables.
Table 10. Difference of class 2 with respect to class 1, as explained by the lagged variables.
0.250.750.5
Δπ2Coef.zCoef.zCoef.z
bookmkt−1133.4931.4194.250792.52181.1312.52
capit−14.85 × 10−82.651.48 × 10−82.703.74 × 10−84.01
spreturn−13.065580.35−6.94009−0.65−17.3828−1.94
constant−0.21103−0.79−0.904068−4.37−0.8076291−4.07
Note: In italics, the non-significant estimated coefficients.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Furno, M. Finite Mixture at Quantiles and Expectiles. J. Risk Financial Manag. 2025, 18, 177. https://doi.org/10.3390/jrfm18040177

AMA Style

Furno M. Finite Mixture at Quantiles and Expectiles. Journal of Risk and Financial Management. 2025; 18(4):177. https://doi.org/10.3390/jrfm18040177

Chicago/Turabian Style

Furno, Marilena. 2025. "Finite Mixture at Quantiles and Expectiles" Journal of Risk and Financial Management 18, no. 4: 177. https://doi.org/10.3390/jrfm18040177

APA Style

Furno, M. (2025). Finite Mixture at Quantiles and Expectiles. Journal of Risk and Financial Management, 18(4), 177. https://doi.org/10.3390/jrfm18040177

Article Metrics

Back to TopTop