You are currently viewing a new version of our website. To view the old version click .
Mathematics
  • Article
  • Open Access

4 December 2025

Bayesian Estimation of R-Vine Copula with Gaussian-Mixture GARCH Margins: An MCMC and Machine Learning Comparison

and
1
Assurance Audit and Risk Group, Health New Zealand, Auckland 1023, New Zealand
2
Department of Mathematical Sciences, Auckland University of Technology, Auckland 1010, New Zealand
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Contemporary Bayesian Analysis: Methods and Applications

Abstract

This study proposes Bayesian estimation of multivariate regular vine (R-vine) copula models with generalized autoregressive conditional heteroskedasticity (GARCH) margins modeled by Gaussian-mixture distributions. The Bayesian estimation approach includes Markov chain Monte Carlo and variational Bayes with data augmentation. Although R-vines typically involve computationally intensive procedures limiting their practical use, we address this challenge through parallel computing techniques. To demonstrate our approach, we employ thirteen bivariate copula families within an R-vine pair-copula construction, applied to a large number of marginal distributions. The margins are modeled as exponential-type GARCH processes with intertemporal capital asset pricing specifications, using a mixture of Gaussian and generalized Pareto distributions. Results from an empirical study involving 100 financial returns confirm the effectiveness of our approach.

1. Introduction

Over the past few decades, following the work of [1], who demonstrated that copula functions can capture non-linear comovement between variables, research on copula-based dependence structure models has grown rapidly. One of the key advantages of copulas, as multivariate distribution tools, is their ability to link different marginal distributions. As a result, copulas have been widely applied across various fields, including hydrology [2], energy [3], agricultural and forest meteorology [4], oceanology [5], computer science [6], insurance [7], and more widely in economics and finance [8,9].
A substantial body of literature demonstrates that copula models, particularly due to their ability to capture tail asymmetry, are highly effective for modeling dependence in quantitative risk management. Tail modeling plays a crucial role in risk assessment, where common measures include Value-at-Risk (VaR) and Conditional Value-at-Risk (CVaR), also known as expected shortfall. For formal definitions of risk measures, see [10]. Among multivariate dependence models, standard copulas are widely used. These typically fall into two main families: elliptical and Archimedean copulas (see [11,12,13,14] for details). However, standard multivariate copulas come with certain limitations—most notably, they rely on a single parameter to govern all pairwise tail dependencies. This becomes problematic in high-dimensional settings, where such a simplification can be unrealistic. This limitation was identified by [15], who introduced a more flexible class of models known as regular vine (R-vine) copulas. Two notable subclasses within the vine copula framework are the canonical vine (C-vine) and the drawable vine (D-vine) copulas [16]. Further discussion on vine copulas can be found in Section 2, and additional details are provided in Chapter 3 of [17].
It consists of two copula families: one is elliptical copulas—see Wichitaksorn et al. [11], Liang and Rosen [13] for recent studies—and the other is Archimedean copulas; see Michaelides et al. [14] and Górecki et al. [12] for recent studies.
In the existing literature, research on vine copulas—particularly R-vines—remains relatively limited. This is likely due to their structural complexity and the intensive computational effort required. Specifically, R-vines allow for a significantly larger number of possible tree structures compared to C-vines and D-vines, especially as the dimensionality increases. Despite these challenges, R-vines offer considerable flexibility as dependence models. Their wide range of possible structures makes them highly adaptable, and importantly, they overcome the parameter limitation of standard multivariate copulas. This added flexibility is particularly advantageous for analyzing tail dependence in high-dimensional financial data.
To our knowledge, there are few studies in the financial literature on R-vines with high dimensions. For instance, ref. [18] studied up to 398 dimensions on S&P500 constituents and ref. [19] studied 2131 dimensions, while other R-vine studies have examined the lower dimensions (less than 25). See, among others, the studies of R-vines in lower dimensions in [20,21,22,23,24,25]. For studies on copula models, see, among others, [26,27,28,29]. Hence, the study of R-vines in high dimensions is still scarce and needs more investigation. This observation is the starting point of our study. The model most closely related to ours is that of [30], which studies R-vines in application to 16-dimensional financial indices using the quasi-maximum likelihood estimation (QMLE) method.
Given standardized residuals, vine copula models can be applied across five different classes: mixed R-vines, mixed C-vines, mixed D-vines, all-t R-vines, and the standard multivariate Gaussian copula. This study utilizes seven bivariate copula families. Based on the test proposed by [31] and likelihood comparisons, previous research has identified mixed R-vines as the best-performing models—an approach we adopt in this study.
Our first contribution is to extend this analysis to 100 dimensions of financial data using constituents of the S&P500, which, to our knowledge, has not been explored in this context. The second contribution is the expansion of the copula family set from seven to thirteen bivariate copulas, including two-parameter BB families and additional Archimedean copulas. This broader selection allows for a more accurate capture of tail asymmetries observed in financial data. Further details are provided in Section 4. Our third contribution lies in the modeling of the marginal distributions. We employ the intertemporal capital asset pricing model (ICAPM), originally proposed by [32], with an asymmetric GARCH model featuring a mixture innovation. This approach enhances the filtration of volatility clustering and yields deeper insights into risk premia across individual financial returns. Unlike [30], who used an ARMA-GARCH model with a single Student-t innovation, we incorporate a mixture of one Gaussian and two heavy-tailed distributions, offering a more flexible representation of return dynamics.
Another key contribution of this study lies in the estimation methodology. In the existing literature, quasi-maximum likelihood estimation (QMLE) is widely used for copula models; see, for example, [33,34,35]. However, QMLE has notable limitations—it is highly sensitive to initial parameter values and often suffers from convergence issues [36]. To address these shortcomings, we employ more efficient estimation techniques, specifically Bayesian Markov chain Monte Carlo (MCMC) and variational Bayes (considered as machine learning) methods, to estimate our proposed R-vine dependence models. These methods are also compatible with the two main subclasses of vine copulas: C-vines and D-vines. While Bayesian MCMC and machine learning approaches are computationally intensive, we mitigate the associated costs by incorporating parallel computing. Additionally, we use QMLE estimates as initial values to improve convergence speed in the Bayesian MCMC and machine learning algorithms.
Our study explores these advanced methods in two distinct ways. First, we apply the random walk Metropolis–Hastings (rw-MH) algorithm for Bayesian MCMC estimation. Second, we implement the variational Bayes method—both with and without data augmentation (VBDA)—as proposed by [37], to estimate the multivariate dependence structure. Furthermore, we propose an alternative, more efficient variational approximation to reduce computational time, the details of which are presented in the simulation study in the Supplementary Material. For related surveys on MLE, Bayesian, and machine learning methods, see [38,39,40,41], among others.
The remainder of the paper is structured as follows. Section 2 introduces the graph-theoretical framework for high-dimensional regular vine (R-vine) copulas, including the two special subclasses—C-vines and D-vines—as well as the marginal models. Section 3 outlines the Bayesian inference and machine learning algorithms used to estimate general R-vine dependence models. Section 4 presents empirical results based on 100-dimensional financial data. Section 5 offers concluding remarks. Details on the simulation study and data generation algorithms are provided in the Supplementary Material.

2. Graph Theory

2.1. Regular Vines in n-by-n Matrix Representation

This section provides the theoretical background on regular vine (R-vine) copulas, along with two important subclasses: canonical vine (C-vine) and drawable vine (D-vine) copulas, all represented in an n-dimensional graphical model framework. In graph theory, an R-vine is represented as a sequence of tree structures composed of nodes N and edges E. An R-vine ( V ) in n-dimensions is defined as a nested set of n 1 trees. The first tree ( T = 1 ) consists of n nodes and n 1 edges, each representing a pair of variables. Each edge in T = 1 becomes a node in the second tree ( T = 2 ), which then has n 1 nodes and n 2 edges. The construction follows the proximity condition, meaning that connected nodes in T = 2 must share a common node from T = 1 .
This nesting continues: in the third tree ( T = 3 ), there are n 2 nodes and n 3 edges, and again, any two connected nodes must share a common node from T = 2 . This process continues until the final tree ( T n 1 ), which has two nodes and a single edge. These two nodes must share all remaining nodes from the previous tree. For further definitions and examples—including complete unions, conditioning sets, and conditional sets of an edge—refer to [30]. Additional theoretical details and properties of vine copulas can be found in [17,42,43,44].
A convenient approach for developing an algorithm in a vine graphical model is to store the trees in an n-by-n matrix. The authors of [30] proposed using a lower triangular matrix M to represent vines as a tree structure in n dimensions; see Figure 1.
Figure 1. An illustrative example of vine copula graphical model in five dimensions.
Definition 1
(Matrix Constraint Set). Let M = ( m i , k ) be a lower triangular n-by-n matrix where i , k = 1 , , n . The i t h constraint set for M is
CM ( i ) = { ( { m k , k , m i , k } , D ) | i = k + 1 , , n , D = { m i + 1 , k , , m n , k } }
for i = n 1 , , 1 . Otherwise, D = (if i = n ). Hence, D is the conditioning set and { m k , k , m i , k } is the conditioned set.
In other words, every pair variable within the constraint set is built up from a diagonal element m k , k plus another element in the same column m i , k conditioning on set of D = { m i + 1 , k , , m n , k } . Otherwise, D = when i equal to n. To illustrate Definition 1, the 5-by-5 matrix M of Figure 1 can be constructed as
M R = 1     5 5   4 4 2   3 2 4 3 2 3 3 4 4 ,   M C = 5     4 4   3 3 3   2 2 2 2 1 1 1 1 1 ,   M D = 5     1 4   2 1 3   3 2 1 2 4 3 2 1 1
Once a 5-by-5 matrix is constructed, it becomes straightforward to generalize the construction to an M matrix for any dimension n. For different constructions of upper triangular matrices representing vines in graphical theory, see [17] (Chapter 3).

2.2. Multivariate Regular Vines Density

This section introduces the matrix M within a numerical framework for computing multivariate regular vine copula densities, including both multivariate canonical vine (C-vine) and drawable vine (D-vine) structures in n-dimensions. In the previous section, we generated an n-by-n matrix M satisfying a proximity condition. Building on this, the n-dimensional multivariate vine copula density is now formulated using two additional n-by-n matrices: the bivariate copula type matrix T = ( t i , k ) and its associated parameter matrix P = ( p i , k ) where i = n , , 2 and k = n 1 , , 1 .
We begin by expressing the vine copula density as a product of pairwise copula functions. According to Sklar’s theorem, a copula C is a multivariate distribution function that links a joint distribution to its marginals F 1 ( x 1 ) , , F n ( x n ) [1]. Thus, the multivariate copula distribution can be written as:
F ( x 1 , , x n )   =   C { F 1 ( x 1 ) , , F n ( x n ) }
for n-dimensional data. After that, vine copula density in n-dimensions can be expressed in graph theory. In product indexing, the C-vine density is
j = 1 n 1 i = 1 n j c j , j + i | 1 , , j 1 F ( x j | x 1 , , x j 1 ) , F ( x j + i | x 1 , , x j 1 ) .
The D-vine density is
j = 1 n 1 i = 1 n j c i , i + j | i + 1 , , i + j 1 F ( x i | x i + 1 , , x i + j 1 ) , F ( x i + j | x x + 1 , , x i + j 1 ) .
The R-vine density is
i = 1 n 1 e E i c C e , a , C e , b | D e ( F C e , a | D e ( X C e , a | X D e ) , F C e , b | D e ( X C e , b | X D e ) ) .
In a D-vine, each node in a given tree T i can be connected to only one edge. In contrast, within a C-vine, a specific node in tree T i must have n j edges. The D-vine structure allows for more independence assumptions compared to the C-vine. However, the C-vine can be advantageous when there is prior knowledge of a particular variable that strongly influences the dataset. The R-vine represents the most general form, offering significantly greater flexibility in modeling complex dependency structures.
Figure 1 provides an example of a five-dimensional regular vine copula density. The complete R-vine copula density depicted in this figure can be expressed as follows:
f ( x 1 , x 2 , x 3 , x 4 , x 5 ) = c 1 , 2 · c 2 , 3 · c 3 , 4 · c 3 , 5   ( unconditional   pairs )   ×   c 1 , 3 | 2 · c 2 , 4 | 3 · c 2 , 5 | 3 · c 1 , 4 | 23 · c 4 , 5 | 23 · c 1 , 5 | 234   ( conditional   pairs )
where c u , v | k stands for c u , v | k ( F u | k ( x u | x k ) , F v | k ( x v | x k ) ) . All definitions and further details can be found, among others, in [15] for C-vines and D-vines, and in [30] for R-vines. Next, we present R-vine density using matrix representation M * , T * , and P * in Figure 2, which can be used for its special cases.
Figure 2. Vine copula specification in n-by-n matrix representation including T * , P * and M * .
In Figure 2, the red label corresponds to the last tree, n 1 . The conditioned set is { m 1 , 1 , m 2 , 1 } M * , and the conditioning set D includes all elements m i , 1 ,   i = 3 , , n . The associated bivariate copula type and parameter are t 2 , 1 and p 2 , 1 , respectively.
The blue label represents the first tree, T 1 . Here, the conditioned set is { m 2 , 2 , m n , 2 } M * , and the conditioning set D is ∅. The corresponding bivariate copula type and parameter are t n , 2 and p n , 2 , respectively.
From a graph-theoretical perspective, the regular vine copula density can therefore be described as:
f 1 , , n = k = 1 n f k k = n 1 1 i = n k + 1 c m k , k , m i , k | m * ( F m k , k | m * , F m i , k | m * ) ,
where m * is m i + 1 , k , , m n , k M * with bivariate copula types T * , bivariate copula parameters P * , and the corresponding densities in Equations (1)–(3). For the R-vine density algorithm, see Algorithm 2.1 of [30].

2.3. Marginal Model and Distribution

This section presents our proposed marginal model, which combines elements of the ICAPM and an asymmetric GARCH model with a mixture distribution. This hybrid approach is designed to capture key stylized facts of financial returns, such as volatility clustering and time-varying heteroskedasticity. The model offers several advantages, including the incorporation of risk premium information and the ability to account for asymmetric shocks and fat-tailed behavior in financial data.
Our marginal model is similar to that of [45], who employed Gaussian innovations; however, we adopt a more flexible mixture innovation to better reflect the complexity of financial return distributions. Furthermore, we extend this framework to the multivariate setting in order to analyze dependence structures using regular vine copulas.
The ICAPM we use in this study follows [32], which is a linear-in-variance function of financial returns and is given by
y t = λ 1 + λ 2 h t   + ε t ,
where y t is the financial returns at time t = { 1 , , T } , λ 1 is a risk-free rate, λ 2 is a risk premium, h t is a conditional volatility, ε t = h t z t is an innovation with zero mean and the conditional variance (or volatility) h t , and z t is the independent and identically distributed (iid) random variable with zero mean and unit variance from a mixture distribution. To deal with the heteroskedastic returns, we use an exponential-type GARCH model or EGARCH(p,q,r) given by
log h t = ω + i = 1 p γ i z t i + j = 1 q δ j | z | t j + k = 1 r ν k log h t k .
where ω is a constant, γ i measures the direct innovation effects, δ j is the asymmetry parameter for the leverage effects, and ν k is the volatility persistence parameter. For details on the EGARCH model, see [46], as well as the surveys on asymmetric GARCH models, including EGARCH, in [47,48]. Regarding the distribution of innovations in the marginal model, we employ a mixture distribution originally proposed by [49] for use with the ARMA-GARCH model and standard copulas. In contrast, our study applies this mixture distribution within an ICAPM-EGARCH framework combined with regular vine copulas.
The mixture distribution consists of three components: one Gaussian distribution and two generalized Pareto distributions (GPDs) derived from extreme value theory (EVT). The GPD components are specifically designed to capture asymmetric fat tails—one of the most significant features observed in financial return data [50].
The mixture distribution has a flexible property. It can be represented by either a finite set of distribution functions (cdf), P k x t , or density functions (pdf), p k x t , k   = 1 , , K . The mixture distribution and its density are a weighted sum such that F x t = k = 1 K w k P k x t and f x t = k = 1 K w k p k x t , respectively. w > 0 is a weight and w = 1 . If the mixing weight w is pre-specified, it can also represent the quantile level of the mixture distribution. Given the three components, the mixture pdf is given by
p x t = 1 β L 1 + ξ L x t μ L β L 1 ξ L 1 , x t Φ 1 α L ϕ x t , Φ 1 α L < x t < Φ 1 α R 1 β R 1 + ξ R x t μ R β R 1 ξ R 1 , x t   Φ 1 α R ,
where α L and α R are a pre-specified quantile, μ = ( , ) is a location parameter, β = ( 0 , ) is a scale parameter, ξ = ( , ) is a shape parameter, and L and R denote the left and right tails, respectively. The support of x t G P D ( μ , β , ξ ) is x t 0 when ξ = [ 0 , ) and x t = [ μ , μ β / ξ ] when ξ = ( , 0 ) , and ξ = [ 0 , 1 / 2 ] represents a fat tail and exists at least up to the second moment. For more details on the moment property, see [51,52].
The GPDs, in a sense of generalization, consist of an ordinary Pareto distribution, an exponential distribution, and a short-tailed, Pareto type II distribution. Further details of the GPD used in the peak-over-threshold method are found in ([10] Chapter 5). Note that ϕ · and Φ 1 · are the standard Gaussian pdf and its inverse cdf, respectively.

3. Bayesian Inference and Machine Learning in the Graphical Model

This section outlines the parameter estimation techniques used in our study, which include both Bayesian and machine learning approaches. For the Bayesian framework here, we refer to the Markov chain Monte Carlo (MCMC), while the variational Bayes is considered a machine learning method. Though both methods are computationally demanding and mathematically intricate, they are well-suited for cases where the posterior distribution of the marginal models is intractable—as is the case in our analysis.
Specifically, we apply the random walk Metropolis–Hastings (rw-MH) MCMC algorithm proposed by [53], along with the variational Bayes method developed by [37], both with and without data augmentation. These methods are used to estimate model parameters and to analyze the R-vine copula dependence structure.
To support and enhance the efficiency of these estimation procedures, we also incorporate plug-in parallel computing and the quasi-maximum likelihood estimation (QMLE) method. These additions help reduce computational time while preserving estimation accuracy, thereby facilitating the integration of Bayesian inference and machine learning techniques in a computationally feasible manner.

3.1. Full Likelihood Function

It is important to first discuss a likelihood function since this is a connection between Bayesian MCMC and machine learning methods. Given the observed n-dimensional data, the full log-likelihood function of our model is
Λ ( θ | y k , i , t ) = k = 1 n l o g ( f k ) k = n 1 1 i = n k + 1 l o g ( c m k , k , m i , k | m * ( F m k , k | m * , F m i , k | m * ) )   = k = 1 n t = 1 T l o g [ q k L p L G P D x k , t + ( 1 q k L q k R ) p G x k , t + q k R p R G P D x k , t ]       k = n 1 1 i = n k + 1 t = 1 T l o g ( c t m k , k , m i , k | m * ( F t m k , k | m * , F t m i , k | m * ) ) ,
where
p L G P D x k , t = 1 h t k 1 ξ L , k 1 2 ξ L , k 1 / 2 1 + ξ L , k y k , t λ 1 k λ 2 k h t k h t k 1 ξ L , k 1 2 ξ L , k 1 / 2 1 ξ L , k 1 , p G x k , t = 1 2 π h t k 1 / 2 e x p y k , t λ 1 k λ 2 k h t k 2 2 h t k , p R G P D x k , t = 1 h t k 1 ξ R , k 1 2 ξ R , k 1 / 2 1 + ξ R , k y k , t λ 1 k λ 2 k h t k h t k 1 ξ R , k 1 2 ξ R , k 1 / 2 1 ξ R , k 1 ,
which is compatible with Equation (4). c t is a unique bivariate copula pdf. F t is a mixture cdf which is defined in Section 2.3. x t is a standardized residual given by Equation (5). The parameter θ in parameter space Θ is θ = { λ 1 k , λ 2 k , ξ L , k , ξ R , k , ω k , γ k , δ k , ν k , P * } . In n dimensions, there are at least n ( n + 15 ) 2 parameters that need to be estimated. The number of parameters grows proportionally to the number of dimensions. From our R-vine copula-based models, there are at least 606 parameters in 30 dimensions and 6094 parameters in 100 dimensions, respectively. Notably, the trade-off between flexibility and parsimony needs to be taken into account. This leads to a research area of regularization or shrinkage method where the most common shrinkage methods are, for instance, lasso and ridge—see [54,55,56], among others. However, we leave this for further research.

3.2. Bayesian MCMC Estimation

Due to the intractable nature of the likelihood, a Bayesian approach is a good candidate for the estimation. Hence, this section discusses the rw-MH algorithm, enhanced with the QMLE and parallel computing to improve computational efficiency. Given the large number of parameters in vine copula models, we adopt the Inference for Margins (IFM) approach [57], which enables the parallelization of marginal model estimation and significantly accelerates the overall computation. Regarding the choice of prior distributions, we use an informative prior, especially the uniform distribution, in the estimation with the hyperparameter values following previous studies.
In summary, Algorithm 1 outlines the procedure for obtaining the parameter estimates P ^ * for the R-vine copula model and θ ^ for the marginal models. The process begins by constructing the maximum matrix M from M * , where each element m i , k in M represents the maximum value among all entries in the k-th column from row i to row n.
The algorithm proceeds in several parallelized steps. The first loop performs data standardization in parallel. In the second loop, the algorithm computes the independent pairwise copulas in the first tree, T 1 , of the R-vine model—also in parallel. The third loop handles the dependent pairs in the subsequent trees T i , for i = 2 , , n 1 .
Algorithm 1 Parallel algorithm of regular vine copulas model using random walk chain Metropolis–Hastings sampling MCMC and maximum likelihood estimation.
Require: 
y t
  1:
Calculate M i , k = m i , k | i = n , k = 1 , , n . m n , k = max { m i , k , , m n , k } for all k = n , , 1 . m is spanning tree maximisation.
  2:
parfor  i g 1 , n  do        ▹ estimate margin parameters from Equations (5)–(7), do parallel
  3:
      Perform θ ^ i g Q M L E , M a r g i n s arg max θ Θ l o g ( f i g ( θ y i g ) ) .
  4:
      Perform θ ^ i g r m M H , M a r g i n s | ( θ ^ i g Q M L E , M a r g i n s ,   y i g ) via call Algorithm 2
  5:
      Calculate z t , i g via standardized y t , i g | θ ^ M a r g i n s , i g r m M H .
  6:
      Calculate u i g = F i g ( z t , i g ) U n i f ( 0 , 1 ) .
  7:
end parfor
  8:
Allocate V d i r e c t = ( v i , k d i r e c t ) where i = n and k = 1 , , n .
  9:
Allocate ( v n , 1 d i r e c t , , v n , n d i r e c t ) = ( u 1 , , u n ) .
10:
parfor  k 1 , n 1  do                                                                    ▹ first tree, do parallel
11:
      Perform p ^ n , k Q M L E , B I C Q M L E arg max θ Θ l o g ( c ( z n , k ( 2 ) , z n , k ( 1 ) ; t n , k ) ) using Equation (8) for copula candidate.
12:
      Find p ^ n , k Q M L E , t n , k by min { B I C Q M L E }
13:
      Set z n , k ( 1 ) = v n , k d i r e c t and z n , k ( 2 ) = v i , ( n m n , k + 1 ) d i r e c t .
14:
      Perform p ^ n , k r m M H | ( z n , k ( 1 ) ,   z n , k ( 2 ) ,   p ^ n , k Q M L E ) using Equation (8) via call Algorithm 2
15:
      Calculate v n 1 , k d i r e c t = h ( z n , k ( 1 ) , z n , k ( 2 ) | t n , k , p ^ n , k r m M H )
16:
      Calculate v n 1 , k i n d i r e c t = h ( z n , k ( 2 ) , z n , k ( 1 ) | t n , k , p ^ n , k r m M H )
17:
end parfor
18:
for  k n 2 , 1  do                            ▹ column iteration, iteration for tree n 1 to tree 2
19:
      for  i n 1 , k + 1  do                                                                         ▹ row iteration
20:
            Calculate M i , k = m i , k . m i , k = max { m i , k , , m n , k } .
21:
            Set z i , k ( 1 ) = v i , k d i r e c t .
22:
            if  m i , k = m i , k  then
23:
                z i , k ( 2 ) = v i , ( n m i , k + 1 ) d i r e c t
24:
            else
25:
                z i , k ( 2 ) = v i , ( n m i , k + 1 ) i n d i r e c t
26:
            end if
27:
            Perform p ^ i , k Q M L E , B I C Q M L E arg max θ Θ l o g ( c ( z i , k ( 2 ) , z i , k ( 1 ) ; t i , k ) ) using Equation (8) for copula candidate.
28:
            Find p ^ i , k Q M L E , t i , k by min { B I C Q M L E }
29:
            Perform p ^ i , k r m M H | ( z i , k ( 2 ) , z i , k ( 1 ) , t i , k , p ^ i , k Q M L E ) using Equation (8) via call Algorithm 2
30:
            Calculate v i 1 , k d i r e c t = h ( z i , k ( 1 ) , z i , k ( 2 ) | t i , k , p ^ i , k r m M H )
31:
            Calculate v i 1 , k i n d i r e c t = h ( z i , k ( 2 ) , z i , k ( 1 ) | t i , k , p ^ i , k r m M H )
32:
      end for
33:
end for
34:
return  P ^ * , T ^ * , M ^ * , θ ^
To ensure better convergence, the algorithm initially employs the QMLE to provide starting values for the rw-MH sampler, described in Algorithm 2, which functions as a subroutine within Algorithm 1.
The function c ( u , v ; t , p ) denotes a specific bivariate copula density, parameterized by copula type t and parameter p . The terms v i , k d i r e c t and v i , k i n d i r e c t refer to recursive conditional cumulative distribution functions (cdfs), often denoted by h ( u , v | t , p ) , which are fundamental components in the construction of conditional copulas in vine structures.
v i , k d i r e c t = C ( v m k , k | u m i + 1 , k , u m i + 2 , k , , u m n , k ) , v i , k i n d i r e c t = C ( v m i + 1 , k | u m k , k , u m i + 2 , k , , u m n , k ) ,
for all n i > 1 . Denote that h ( u , v | t , p ) = C 2 | 1 ( v | u ; θ , δ ) . C ( u , v ; θ , δ ) = G ( x , y ; θ , δ ) where G ( x , y ; θ , δ ) is a bivariate survival function and x , y are monotonically decreasing transforms. Therefore, the conditional cdf is C 2 | 1 ( v | u ; θ , δ ) = G x . x u . See [30] for further discussion.
Algorithm 2 presents the rw-MH sampling method used as part of the MCMC procedure in Algorithm 1. In general, the performance of the Metropolis–Hastings algorithms is sensitive to the choice of initial values ( θ , Ω ) where Ω is a covariance matrix. The first stage of the rw-MH algorithm involves a burn-in period, during which early samples of θ are discarded. A longer burn-in period increases computational time, and its length largely depends on how close the initial values are to the target distribution.
To address this, our approach initializes the algorithm with parameter estimates obtained via the QMLE. This helps achieve a reasonable acceptance rate more quickly— approximately 50% in our implementation.
The proposal step involves an increment random variable z, which follows a Gaussian distribution, z N ( 0 , c Ω ) , where c is a constant scaling factor. This constant is tuned to ensure an appropriate acceptance rate for the sampler. For further details on the MH algorithm, see [53] (Exercise 17.5 and Exercise 11.18).
Algorithm 2 Random walk Metropolis–Hastings algorithm
Require: 
Λ ( θ | y ) , θ ^ Q M L E , Σ ^ Q M L E , c , n b u r n , n r e p .
  1:
Calculate Σ * = c × Σ ^ Q M L E .
  2:
Set θ 1 = θ ^ Q M L E .
  3:
 
  4:
for  i 2 , n b u r n + n r e p   do
  5:
      Draw a candidate θ * = θ i 1 + z
  6:
      Draw an acceptance probability
α ( θ i 1 , θ * ) = m i n Λ ( θ * | y ) Λ ( θ i 1 | y ) , 1
  7:
      Set θ i = θ * with probability α ( θ i 1 , θ * ) . Otherwise, set θ i = θ i 1 .
  8:
end for
  9:
 
10:
Calculate posterior means θ ^ r w M H from the draws θ and whatever posterior inference of interest are needed.
11:
return  θ ^ r w M H

3.3. Variational Bayes Estimation

This section introduces an alternative algorithm based on the variational Bayes (VB) method, which integrates Bayesian inference with machine learning to estimate n-dimensional general regular vine copula models. Specifically, we apply the variational Bayes with (and without) data augmentation (VBDA) approach originally developed for a drawable vine structure by [37,58].
In this study, we extend the VBDA method to accommodate the more general class of regular vines by incorporating an additional stopping criterion. This enhancement is designed to reduce the number of iterative learning steps, thereby improving computational efficiency.
We explore three versions of the variational Bayes estimation method: one without data augmentation (VBDA0), and two with data augmentation (VBDA1 and VBDA2). All three methods rely on latent variables during estimation. To address convergence challenges and reduce computational time, we also incorporate plug-in QMLE estimates and leverage parallel computing techniques.
The augmented posterior density is
p ( θ , u | y ) = f ( y , u , θ ) f ( y ) = c ( u | θ ) p ( θ ) t = 1 T I ( a t u t < b t ) / f ( y ) .
where p ( θ ) is the prior density. f ( y ) is the marginal likelihood function. c ( u | θ ) is the copula density. I is the indicator function. VB estimation approximates p ( θ , u | y ) by a tractable density q λ ( θ , u ) in a so-called variational approximation. This estimation corresponds to minimizing the Kullback–Leibler (KL) divergence
K L ( q λ ( θ , u ) | | p ( θ , u | y ) ) = l o g q λ ( θ , u ) p ( θ , u | y ) q λ ( θ , u )   d θ   d u
or maximizing the lower bound of the logarithm of the marginal likelihood (or the evidence lower bound (ELBO)), given by
L ( λ ) = l o g p ( θ ) f ( y , u | θ ) q λ ( θ , u ) q λ ( θ , u )   d θ   d u ,
where θ and u are independent. For further details and examples of VB methods, see, among others, [59,60]. Also, see [61,62], among others, for reviews of VB inference in machine learning.
The density approximation of q λ can be written as q λ ( θ , u ) = q λ a ( θ ) q λ b ( u ) . The specification of q λ a and q λ b is crucial. q λ a has parameter λ a while q λ b has parameter λ b with the support on [ a 1 , b 1 ) × × [ a T , b T ) . This study selects the choice of u approximation as follows:
V B D A 0 : q λ b ( u ) = 0 , V B D A 1 : q λ b ( u ) = t = 1 T 1 b t a t I ( a t u t < b t ) , V B D A 2 : q λ b ( u ) = t = 1 T ϕ ( x t ; μ t , Ω t ) ( b t a t ) ϕ ( x t ) .
VBDA2 nests VBDA1, which covers a wide range of data dependencies with an effective mean field approximation. VBDA1 is expected to be more accurate for data with low dependence. λ b = { μ 1 , , μ T , l o g   ω 1 , , l o g   ω T } , Ω t = d i a g ( ω 1 , , ω T ) , and ϕ is the standard normal density.
To approximate θ , we adopt a commonly used variational distribution q λ a , specifically the multivariate normal (Gaussian) distribution with mean M and covariance matrix Σ , denoted as N ( M , Σ ) . In the context of R-vine models, a large number of parameter estimates are required. The use of a tractable multivariate normal distribution makes it relatively straightforward to compute the gradient of l o g   q λ a ( θ ) .
Moreover, the structure of Σ contributes to improving the accuracy of the gradient estimates. In particular, when Σ is expressed using its factorized form Σ = B B + D 2 , the gradient of this factorization has been derived and shown to be computationally efficient [63].
Algorithm 3 shows the VBDA algorithm with stochastic gradient ascent optimization. Note that, in Algorithm 1, Algorithm 2 is replaced by Algorithm 3 in order to perform VBDA method in the R-vine model. This approach is tractable posterior distribution.
Algorithm 3 Variational Bayes data augmentation with control variates and ADADELTA learning rate
Require: 
Λ ( θ | y ) , D A , λ ( 0 ) , μ Q M L E , S , n V B , c , t w , P , ρ , ϵ .
  1:
Set k = 1 .
  2:
Initialize λ | μ Q M L E .
  3:
Generate ( θ s k , u s k ) q λ k θ , u for s = 1 , , S where
  4:
q λ b ( u ) could be i.  D A 0 ii.  D A 1 or iii.  D A 2 and q λ a ( θ ) N ( M θ , Σ θ ) .
  5:
Calculate ς i ( k ) = ( ς 1 ( k ) , , ς m ( k ) ) with
ς i ( k ) = c o v l o g   h ( θ , u ) l o g   q λ ( θ , u ) × λ i l o g   q λ ( θ , u ) , λ i l o g   q λ ( θ , u ) v a r ( λ i l o g   q λ ( θ , u ) )
  6:
Set k = k + 1 .
  7:
while stopping rule is not satisfied do
  8:
      Generate ( θ s k , u s k ) q λ k θ , u for s = 1 , , S .
  9:
      Calculate λ L λ k ^ = g λ 1 k , , g λ m k
g λ i k = 1 S s = 1 S log h ( θ s k , u s k ) log q λ ( θ s k , u s k ) ς i ( k 1 )     × λ i log q λ ( θ s k , u s k )
10:
    Calculate ς i ( k ) = ( ς 1 ( k ) , , ς m ( k ) ) .
11:
    Perform Δ λ ( k ) the ADADELTA method.
12:
    Set λ ( k + 1 ) = λ ( k ) + Δ λ ( k )
13:
    Calculate L B ^ ( λ ( k + 1 ) ) = 1 S s = 1 S log h ( θ s k + 1 , u s k + 1 ) log q λ ( θ s k + 1 , u s k + 1 )
14:
    if  k > c  then
15:
         Calculate L B ¯ k t w = 1 / t w i = 1 t w L B ^ ( λ ( k i + 1 ) )
16:
         if  L B ¯ ( λ ( k t w ) ) > m a x ( L B ¯ )  then
17:
               p a t i e n c e = 0 .
18:
              Collect λ ( * ) λ ( k + 1 ) .
19:
         else
20:
               p a t i e n c e = p a t i e n c e + 1 .
21:
         end if
22:
         if  p a t i e n c e > P  then
23:
              stop.
24:
         end if
25:
      end if
26:
      Set k = k + 1 .
27:
end while
28:
return  λ ( * )
A brief overview of the procedure is as follows:
Three VB estimators for regular vine dependence models are introduced in Line 4, where the specification of q ( u ) can be one of DA0, DA1, or DA2. To initiate the algorithm, an initial value for the variational parameter, λ ( 0 ) , is required, typically initialized using QMLE estimates μ Q M L E .
In Line 12, the lower bound (LB) is updated iteratively as follows: λ ( k + 1 ) = λ ( k ) + ρ ( k ) λ L ( λ ( k ) ) ^ . Here, ρ ( k ) is the learning rate, and λ L ( λ ( k ) ) ^ is an unbiased estimate of the gradient of the lower bound L , computed using the ADADELTA optimization method (Line 11). For further details on adaptive learning rate techniques like ADADELTA, see [64].
The gradient of the lower bound, λ L ( λ ) , is expressed as the expectation:
λ L ( λ ) = E q ( λ l o g   q λ ( θ , u ) { l o g   h ( θ , u ) l o g   q λ ( θ , u ) } )
where p ( θ , u | y ) f ( y , u | θ ) p ( θ ) = h ( θ , u ) represents the unnormalized posterior distribution.
To reduce the variance of this unbiased gradient estimator, a control variate vector ς is employed. Consequently, the unbiased estimator of the gradient is given by λ L ( λ ( k ) ) ^ = ( g λ 1 , , g λ m ) where g λ = 1 S s = 1 S l o g   h ( θ s , u s ) l o g   q λ ( θ s , u s ) λ l o g   q λ ( θ s , u s ) where g λ = 1 S s = 1 S l o g   h ( θ s , u s ) l o g   q λ ( θ s , u s ) λ l o g   q λ ( θ s , u s ) and m is the number of elements in λ , and SS is the number of Monte Carlo samples.
For a more detailed discussion of the VBDA algorithm, refer to [30] (Section 3).
The stopping rule for the algorithm terminates the iterative procedure if
(i.)
There is a fixed number of stochastic gradient ascent steps, n V B , or
(ii.)
The moving average L B ^ is not improved, L B ¯ k t w = 1 / t w i = 1 t w L B ^ ( λ ( k i + 1 ) ) , which is referred to as the p a t i e n c e parameter and p a t i e n c e = p a t i e n c e + 1 in Line 20 and the whole procedure stops when p a t i e n c e > maximum patience P in Line 22–23, and
(iii.)
All of the above conditions will be checked when the k t h iteration is greater than a constant c.
Finally, λ ( * ) is collected when L B ¯ k t w is greater than m a x ( L B ¯ ) [60]. Note that the simulation study in Bayesian inference and machine learning including the rw-MH algorithm and the VBDA algorithm can be found in the Supplementary Material. The results show that, in general, the VB methods perform better than the others.

4. Empirical Study

4.1. Data

This study analyzes financial market time series data with up to 100 dimensions. For the training set, we selected the daily adjusted closing prices of the top 100 Nasdaq-listed stocks by market capitalization, covering the period from 2 January 2009 to 20 September 2021, where we intentionally avoided the impact of the 2008 global financial crisis. Only stocks listed before 2009 were included, resulting in a total of 3201 daily observations. The test set spans from 21 September 2021 to 21 December 2021, comprising 65 daily observations. All data were obtained from Yahoo Finance using MATLAB 2022a’s data feed function. The experiments were implemented in MATLAB and conducted on a laptop equipped with an 11th Gen Intel(R) Core(TM) i7-1185G7 processor (3.00 GHz) and 32 GB of RAM.
After completing data cleaning, we obtained 100 stock indices for use in our empirical study. A full list of these indices is provided in Appendix C. To evaluate the performance of the proposed models described in Section 2, we conducted two empirical experiments using two datasets: one consisting of all 100 time series and another with only the last 30 stock indices. The goal of these experiments was to assess the scalability and effectiveness of the proposed model as the dimensionality increases.
The descriptive statistics of the dataset are summarized as follows. The top five business sectors represented in our analysis are technology (22 firms), health care (18 firms), finance (11 firms), consumer services (11 firms), and capital goods (8 firms). The total market capitalization of the selected stocks is USD 30,814 billion, accounting for approximately 46% of the total market capitalization on the Nasdaq exchange.
All stocks exhibit a positive mean return, with return values ranging from −39.02% to 52.29%. The standard deviation ranges from 1.06% to 3.67%, while skewness values range between −1.04 and 1.56. Kurtosis spans from 5.95 to 47.26. These statistics indicate that the return distributions are asymmetric and exhibit heavy tails.
Additionally, the Jarque–Bera test statistics for all returns strongly reject the null hypothesis of normality at the 1% significance level, confirming that the return distributions deviate significantly from normality.

4.2. Regular Vine Distribution Selection

Multivariate copula models can be fitted to n-dimensional time series data using a two-stage sequential approach [30]. In the first stage, a univariate marginal model is selected and fitted separately to each time series to obtain standardized residuals, which are then transformed into marginally uniform variables. In this study, we apply our proposed marginal model, detailed in Section 2.3, and compare its performance to a widely used benchmark: the ARMA(1,1)-GARCH(1,1) model with Student-t innovations.
In the second stage, the dependence structure among the transformed residuals is modeled using a multivariate copula. It is well established in the literature that different pairs of financial variables exhibit varying degrees of asymmetry and tail dependence. Traditional multivariate copulas are generally unable to capture such heterogeneous dependence structures. Vine copulas, however, are specifically designed to address this limitation.
While D-vine and C-vine copulas are effective in modeling complex dependencies, they are constrained by their predefined path structures. To overcome this limitation, we investigate the use of regular vine (R-vine) copulas—a more flexible and general class that encompasses both D-vines and C-vines as special cases. R-vines offer significantly greater modeling capacity due to their extensive variety of possible tree structures and pair-copula combinations.
In fact, for high-dimensional datasets, the number of possible R-vine structures grows rapidly, with the total number given by n ! 2 2 n 2 2 [65], underscoring their flexibility and complexity compared to standard copula models and their vine subclasses.
The following tasks are the sequential steps of fitting an R-vine copula with marginal specification:
(a)
Standardize the residuals of the returns using the univariate marginal model.
(b)
Select the R-vine path structure of the tree, M = ( N , E ) , given by maximized empirical K e n d a l l s   τ via
b.1
Formulation of empirical K e n d a l l s   τ of all possible pairs variable, τ ^ i , k ,   1 i < k n ;
b.2
Performing the spanning tree maximization (the sum of absolute empirical K e n d a l l s   τ )
m a x e = { i , k }   i n   s p a n n i n g   t r e e | τ ^ i , k | ;
b.3
For each edge, selection of bivariate copula family t i , k and its parameter(s) p i , k from all possible bivariate copula candidates via Bayesian information criterion (BIC) value.
(c)
Iterate Step (b) from the first tree T 1 (unconditional tree, τ ^ i , k ) to the last tree T n 1 (conditional tree, τ ^ i , k | D where D is the proximity condition).
The methodology described above yields a complete specification of a regular vine copula, including M , T , P . The QMLE is employed in Step b.3 to estimate all relevant parameters. This full specification is then used as the foundation for our Bayesian inference and machine learning procedures.
It is important to note that, for the first tree in the vine structure, parallel computation is feasible due to the independence of the path structure in R-vines. The method proceeds as follows:
  • Step 1 (Tree 1): A parallel computation technique is applied to estimate τ ^ 1 , k for k = 1 , , n 1 .
  • Step 2: A FOR loop is used to compute τ ^ i , k where i < k , i = n 1 , , 2 ,   k = 1 , , n 2 .
For more details on the maximum spanning tree algorithm used in this context, see [30].
Our study employs a range of bivariate copula functions to capture the diverse (a)symmetrical dependence structures present in financial time series data. While the core families considered include the Gaussian, Student-t, Gumbel, and Frank copulas—consistent with the approach of [30]—we extend the analysis by incorporating a total of 13 copula families. The complete list of candidate bivariate copula families used in this study is provided below:
  • Gaussian/Normal copula presents symmetric and no tail dependence.
  • Student-t copula presents symmetric and upper and lower tail dependence.
  • Frank copula presents symmetric and tail independence.
  • Clayton copula presents asymmetric and lower tail dependence.
  • B5/Joe, Gumbel, and Galambos copulas present asymmetric and different upper and lower tail dependence.
  • BB1, BB3, BB4, BB5, BB7 and BB10 copulas present asymmetric and different upper and lower tail dependence.
Details of the bivariate copula distributions and their key dependence structure properties are summarized in Appendix A. These include the cumulative distribution functions, lower and upper tail dependence orders, and rank-based dependence measures. For additional discussion on copula properties, see, among others, [17] (Chapter 4) and [66] (Appendix).
We also conducted Ljung–Box tests on the standardized residuals from all proposed models. The results confirm the independence of the residuals. Full details of these tests are provided in Appendix B.

4.3. Empirical Results

In this section, we present the empirical results obtained from our proposed Bayesian models and estimation methods for R-vine copulas. A comprehensive set of 13 bivariate copula families is employed within the regular vine structure to capture complex dependence patterns.
For the univariate margins, we apply the mixture distribution defined in Equation (7) to the ICAPM-EGARCH models, which are designed to capture key features of financial time series data such as volatility clustering and heavy tails.
As a benchmark, we use the traditional QMLE method and compare its performance with our proposed approaches, including the rw-MH algorithm and the VBDA methods. These methods are implemented in the R-vine copula framework for dimensions of up to 100. In total, the empirical study involves the estimation of up to 4950 bivariate copula functions and 6138 parameters.
The experimental results, summarized in Table 1 and Table 2, indicate that the VBDA1 method delivers the best performance in terms of marginal likelihood on the training set. On the test set, the variational Bayes approaches—particularly VBDA1—demonstrate substantial improvements in forecasting accuracy as the dimensionality increases, making them strong candidates for high-dimensional modeling. Specifically, as both the number of dimensions and the forecast horizon grow, VBDA1 provides more accurate variability estimates. In contrast, the QMLE method performs relatively poor in higher dimensions and longer forecast horizons. However, in the case of 30 dimensions with a three-month forecast window, QMLE outperforms the other methods.
Table 1. Performance comparison of 30 dimensions.
Table 2. Performance comparison of 100 dimensions.
For 100-dimensional data, VBDA1 achieves the best forecasting performance based on the mean absolute deviation (MAD). Meanwhile, both the rw-MH and QMLE methods yield the best performance in terms of root mean square error (RMSE) for the three-month test set, while rw-MH shows the strongest performance in one-month forecasts at the 100-dimensional level.
Figure 3 illustrates a representative example of the empirical results for a normalized R-vine copula contour plot and rank-based dependence measures. These results are based on the univariate ICAPM-EGARCH–mixture marginal model using the VBDA2 estimation method in a 30-dimensional setting.
Figure 3. Estimation results of normalized mixed R-vine copula contour plot, Kendall’s τ ^ , and Spearman’s ρ s ^ for 10 stock indices in the 30-dimensional cases using R-vine–ICAPM–EGARCH–Mixture model and VBDA2 method. Notes: Due to space, only 10 representative stocks are shown. Contour plots are shown in the lower triangular matrix while [Kendall’s τ ^ , Spearman’s ρ s ^ ] are shown in the upper triangular matrix.
Due to space, we present a normalized 10-dimensional subset of the R-vine contour plot for selected stock indices. The stock names appear along the diagonal of the matrix. The lower triangular part of the matrix displays the R-vine copula contours, where each row corresponds to the i t h tree ( i = 1 , , 9 ) derived from the full 30-dimensional dataset. The upper triangular matrix shows the empirical Kendall’s τ ^ k and Spearman’s ρ s ^ k , where the k t h column ( k = 10 , , 2 ) corresponds to tree T ^ i for i = 1 , , 9 ) .
Similarly, Figure 4 provides an empirical example of the normalized 10-dimensional R-vine contour plots and rank-based dependence measures from the 100-dimensional analysis, using the VBDA1 method. Complete empirical results—including Kendall’s τ ^ , Spearman’s ρ s ^ , and the estimated vine matrices M ^ and T ^ for both figures—are available in the Supplementary Materials.
Figure 4. Estimation results of normalized mixed R-vine copula contour plot, Kendall’s τ ^ , and Spearman’s ρ s ^ for 10 stock indices in the 100-dimensional cases using R-vine–ICAPM–EGARCH–Mixture model and VBDA1 method. Notes: Due to space, only 10 representative stocks are shown. Contour plots are shown in the lower triangular matrix while [Kendall’s τ ^ , Spearman’s ρ s ^ ] are shown in the upper triangular matrix.
Table 1 summarizes the model performance for the 30-dimensional case. Based on the training dataset, the proposed VBDA1–R-vine–ICAPM–EGARCH–mixture model delivers the best performance in terms of marginal likelihood. The next best performers are the VBDA2 and VBDA0 methods, respectively. Among all estimation methods, VBDA2 is the most computationally intensive, which explains its longer runtime for R-vine copula estimation. VBDA1, while still relatively complex, requires less computing time than VBDA2 and converges more efficiently than simpler methods such as VBDA0 due to its favorable convergence properties.
An illustration of the convergence behavior of the evidence lower bound (ELBO) for all VBDA methods is shown in Figure 5, using the stock index pair (DE, CAT) as an example. Additionally, Figure 6 presents the first tree structure from the spanning tree maximization algorithm for the VBDA2–R-vine–ICAPM–EGARCH–mixture model. In contrast, when evaluating by Akaike information criterion (AIC), Bayesian information criterion (BIC), and log-likelihood values, the QMLE–R-vine–ARMA–GARCH–Student-t model achieves the best performance. The second-best performance under these criteria is the rw-MH-based R-vine–ICAPM–EGARCH–mixture model, which achieves a favorable acceptance rate in the MH algorithm.
Figure 5. An example of ELBO convergence across VBDA methods for a pair of stocks (DE, CAT), using a Student-t copula in the 30-dimensional case. Note: The ELBO values of VBDA0 are plotted against the right y-axis while those of VBDA1 and VBDA2 are plotted against the left y-axis.
Figure 6. An example of Tree 1 in the 30-dimensional case of an R-vine copula using VBDA2 within the R-vine–ICAPM–EGARCH–Mixture model.
For the test dataset, forecasting performance was evaluated over one-month and three-month horizons using the MAD and RMSE metrics. In terms of forecast errors, the QMLE and rw-MH methods show comparable predictive performance and emerge as the best-performing models overall. Interestingly, within the variational Bayes framework, VBDA1 demonstrates improved performance—especially as the forecast horizon increases—outperforming the other VBDA variants based on MAD values. This trend is further supported by the experimental results.
Table 2 summarizes the performance of the proposed models in the 100-dimensional setting, using both training and test datasets. Overall, as the number of dimensions increases to 100, the performance of variational Bayes methods improves noticeably.
Among all approaches, the Bayesian inference and machine learning methods—particularly the Variational Bayes with Data Augmentation Type 1 (VBDA1)—outperform the traditional QMLE method, especially in terms of the mean absolute deviation (MAD) for the three-month forecast horizon. The VBDA1–R-vine–ICAPM–EGARCH–mixture model achieves the lowest three-month MAD, accurate to three decimal places, indicating superior forecasting accuracy.
These findings suggest that variational Bayes—specifically, the VBDA1 algorithm—performs effectively in high-dimensional settings. Both the training and test results confirm that the VBDA1–R-vine–ICAPM–EGARCH–mixture model surpasses all competing models in terms of marginal likelihood and three-month MAD forecast error.
Based on the model comparisons, the rw-MH algorithm combined with the R-vine-ICAPM-EGARCH-Mixture model emerges as the most outstanding, particularly due to its reasonable acceptance rate. Variational Bayes inference also performs exceptionally well, with the VBDA0-R-vine-ICAPM-EGARCH-Mixture model ranking as the second most preferable according to the AIC and BIC values. While the traditional QMLE method shows a decline in performance on the training data, it still remains a viable alternative for the test data.
Table 3 and Table 4 summarize the copula families and their parameter counts across models and methods for 30- and 100-dimensional settings, as selected by Algorithm 1. In total, there are 435 pair-copula functions in 30 dimensions and 4950 in 100 dimensions. In the 30-dimensional case, the best-performing method—VBDA1—estimates 638 parameters, which is the second-lowest among all methods. Conversely, in the 100-dimensional case, VBDA1 estimates 6138 parameters—the highest. This indicates that better model performance does not necessarily correspond to a smaller number of parameters. Among the selected bivariate copula families, the elliptical copulas—Student-t and Gaussian—fit the empirical data best, followed by Archimedean families. The Student-t, Gaussian, and Frank copulas together account for at least 68% of all copulas used, implying that approximately 32% of the data exhibit asymmetric or tail dependence characteristics. In 30 dimensions, the most frequently selected families are Student-t and Frank, while in 100 dimensions, the elliptical copulas dominate. Notably, the BB6 copula was not selected in the 30-dimensional case, and the BB10 copula was not used in the 100-dimensional case.
Table 3. Summary of copula families and their parameter numbers across models and methods for 30 dimensions.
Table 4. Summary of copula families and their parameter numbers across models and methods for 100 dimensions.
In this study, we apply the ICAPM–EGARCH–mixture with R-vine copula, which we believe is a useful model. However, this might be subject to the model misspecification, which can be a limitation of this study. The computing performance is another possible limitation, which can be improved, but we leave for future research. A practical implementation from the practitioner’s viewpoint poses an interesting challenge to be validated in the future.

5. Concluding Remarks

This study introduces the variational Bayes methods—with and without data augmentation types 1 and 2—for Bayesian inference and machine learning, alongside the rw-MH Bayesian MCMC approach and the competing QMLE method, applied to regular vine copula models in graph theory for n-dimensional data. Standard multivariate copulas often face limitations due to having a single control parameter, a problem that worsens as dimensionality increases. Although C-vine and D-vine copulas offer more flexibility by allowing different bivariate copula choices and tree constructions, they impose restrictions on the tree path structure. To overcome this, we employ the more general R-vine copula, which eliminates the tree path structure constraint and provides richer insights into the relationships within financial data. This study considers 13 candidate bivariate copula families for the R-vine model, though the algorithm easily accommodates additional families. Furthermore, our proposed model is fully compatible with the special subclasses of C-vine and D-vine copulas and can be applied to n-dimensional data analysis.
Furthermore, this study is the first to introduce a mixture distribution into the R-vine model, enhancing its flexibility. The mixture distribution combines a Gaussian distribution with two generalized Pareto distributions within a univariate marginal intertemporal capital asset pricing model and an asymmetric GARCH framework, capturing additional risk premium stylized facts of specific financial securities. Additionally, the study presents a parallel R-vine with margins algorithm for Bayesian inference and machine learning, a bivariate copula selection algorithm within the R-vine framework, and an R-vine with margins data simulation algorithm. Importantly, our proposed algorithm can compute the joint distribution for arbitrary R-vine copulas.
In conclusion, our study produces the R-vine tree path dependence structure, pair-copula types, and parameter estimates in n-dimensions. The univariate marginal model captures individual stock risk premiums and risk-free rates. Our proposed model, particularly the VBDA1–R-vine–ICAPM–EGARCH–mixture model, outperforms competing approaches. Moreover, we demonstrate that the traditional QMLE estimation method significantly deteriorates as dimensionality increases.
For future research, there are numerous avenues to explore depending on specific interests. For instance, the model could be applied to higher-frequency data or extended with alternative copula families such as factor copulas or rotated copulas. The copula selection criteria could also be modified—using Spearman’s ρ s or other relevant measures tailored to particular financial contexts instead of Kendall’s τ . Additionally, the framework can be adapted to other domains, such as high-dimensional portfolio optimization and higher-dimensional datasets. From a computational perspective, further development could involve implementing the algorithm on high-performance computing platforms to reduce computational time in Bayesian inference and machine learning.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/math13233886/s1.

Author Contributions

Conceptualization, R.K. and N.W.; methodology, R.K.; software, R.K.; validation, R.K. and N.W.; formal analysis, R.K.; data curation, R.K.; writing—original draft preparation, R.K.; writing—review and editing, N.W.; visualization, R.K.; supervision, N.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Acknowledgments

We thank the reviewers for their useful comments and feedback that help improve the manuscript.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Some Parametric Copula Families and Their Properties

This appendix presents the dependence structure of all copula families used in this study including one- and two-parameter copulas. The elliptical copulas are the Gaussian (normal) copula and Student-t copula. The bivariate copula distributions of Gaussian and Student-t, respectively, are
C Σ G a   ( u ) = Φ Σ   ( Φ 1   ( u 1   ) , Φ 1   ( u 2   ) ) C v , Σ t u = t v , Σ t v 1 u 1 , t v 1 u 2
where Φ 1 is an inverse cumulative Gaussian distribution function. Σ is a 2 × 2 correlation matrix; u U n i f 0 , 1 is a continuous standard uniform random variable; t 1 is an inverse cumulative Student-t distribution function; t v , Σ is the cdf of bivariate t v , Σ distribution where v > 2 degree of freedom. The copula is denoted as C u , v = φ φ 1 u + φ 1 v . It is worth mentioning that C ( u , v ; θ , δ ) = G ( x , y ; θ , δ ) where G ( x , y ; θ , δ ) is a bivariate survival function and x , y are monotonically decreasing transforms. Therefore, the conditional cdf and copula pdf are:
C 2 | 1 ( v | u ; θ , δ ) = G x . x u c ( u , v ; θ , δ ) = 2 G x y . x u . y v .
The bivariate cdf copula families in this study including one and two parameters are presented in Table A1 while their tail dependence measures are shown in Table A2.
Table A1. Some parametric copula distributions.
Table A1. Some parametric copula distributions.
CopulaBivariate Function C ( u , v ) Parameter θ , δ
One-parameter
Gumbel exp [ ( ( log u ) θ + ( log v ) θ ) 1 / θ ] θ 1
Clayton ( u θ + v θ 1 ) 1 / θ θ 1 , θ 0
Frank 1 θ ln 1 + ( e θ u 1 ) ( e θ v 1 ) e θ 1 < θ < , θ 0
Joe/B5 1 [ ( 1 u ) θ + ( 1 v ) θ ( 1 u ) θ ( 1 v ) θ ] 1 / θ θ 1
Galambos u v exp { [ ( l o g   u ) θ + ( l o g   v ) θ ] 1 / θ } 0 θ <
Two-parameters
BB1 ( 1 + [ ( u θ 1 ) δ + ( v θ 1 ) δ ] 1 / δ ) 1 / θ θ > 0 , δ 1
BB3 e x p { [ δ 1 l o g ( e δ u ˜ θ + e δ v ˜ θ 1 ) ] 1 / θ } , u ˜ = l o g ( u ) , v ˜ = l o g ( v ) θ 1 , δ > 0
BB4 ( u θ + v θ 1 [ ( u θ 1 ) δ + ( v θ 1 ) δ ] 1 / δ ) 1 / θ θ 0 , δ > 0
BB5 e x p { [ x θ + y θ ( x θ δ + y θ δ ) 1 / δ ] 1 / θ } , x = l o g ( u ) , y = l o g ( v ) θ 1 , δ > 0
BB7 1 ( 1 [ ( 1 u ¯ θ ) δ + ( 1 v ¯ θ ) δ 1 ] 1 / δ ) 1 / θ , u ¯ = 1 u , v ¯ = 1 v θ 1 , δ > 0
BB10 u v [ 1 δ ( 1 u θ ) ( 1 v θ ) ] 1 / θ θ > 0 , 0 δ 1
Table A2. Tail dependence measures of some parametric copulas.
Table A2. Tail dependence measures of some parametric copulas.
Copula K L or λ L K U or λ U
One-parameter
Gaussian/Normal K L = 2 / ( 1 + ρ ) K U = K L
Gumbel (extreme value) K L = 2 1 / θ K U = 2 2 1 / θ
Clayton K L = 2 1 / θ K U = 2
Frank K L = 2 K U = 2
Joe/B5 K L = 2 λ U = 2 2 1 / θ
Galambos (extreme value) K L = 2 2 1 / θ λ U = 2 1 / θ
Two-parameters
Student-t λ L = λ U λ U = 2 T ν + 1 ( ( ν + 1 ) ( 1 ρ ) 1 + ρ )
BB1 λ L = 2 1 / θ δ λ U = 2 2 1 / δ
BB3 λ L = 1 λ U = 2 2 1 / θ
BB4 λ L = ( 2 2 1 / δ ) 1 / θ λ U = 2 1 / δ
BB5 (extreme value) K L = 2 λ U λ U = 2 ( 2 2 1 / δ ) 1 / θ
BB7 λ L = 2 1 / δ λ = 2 2 1 / θ
BB10 K L = 2 K U = 2
A rank-based measurement of the dependence measure of the copula families can be computed either from a closed-form solution (if available) or numerical approximation via a Monte Carlo simulation through two-variable integrals as in Equation (A1) for Kendall’s τ and Equation (A2) for Spearman ρ s , respectively.
τ ^ ( C ) = 1 4 0 1 0 1 C 1 | 2 ( u | v ) C 2 | 1 ( v | u ) d u d v
ρ s ^ ( C ) = 3 12 0 1 0 1 u C 2 | 1 ( v | u ) d u d v
Note that a closed-form Kendall’s τ ^ and Spearman ρ s ^ of Gaussian copula, t copula, Gumbel copula, Frank copula, and Clayton copula are used. Otherwise, we use Monte Carlo approximation.

Appendix B. Independence Test Results of the Standardized Residuals

This appendix discusses the Ljung–Box independence testing of the marginal model residuals of this study including ICAPM-EGARCH with mixture innovation and the competing ARMA-GARCH with Student-t innovation with the estimation method of rw-MH, variational Bayes and QMLE. For the 100 stocks, standardized residuals of all the proposed models indicated independence where α = 0.01 meant failure to reject the null hypothesis that the residuals are independently distributed, except for the stock abbreviations ‘V’, ‘MA’, ‘TXN’, ‘QCOM’, ‘SCHW’ and ‘DEO’ in ICAPM-EGARCH(1,1)-Mixture-rw-MH model. For the ICAPM-EGARCH(1,1)-Mixture-VBDA0 model, the stock abbreviations are ‘V’, ‘TXN’, ‘QCOM’ and ‘DEO’. For the ICAPM-EGARCH(1,1)-Mixture-VBDA1 model, the stock abbreviations are ‘V’, ‘MA’, ‘TXN’, ‘QCOM’ and ‘DEO’. For ICAPM-EGARCH(1,1)-Mixture-VBDA2 model, the stock abbreviations are ‘V’, ‘TXN’, ‘QCOM’ and ‘DEO’. For ICAPM-EGARCH(1,1)-Mixture-QMLE model, the stock abbreviations are ‘V’, ‘MA’, ‘TXN’, ‘QCOM’, ‘SCHW’ and ‘DEO’. For ARMA(1,1)-GARCH(1,1)-t-QMLE model, the stock abbreviations are ‘QCOM’and ‘HON’.

Appendix C. List of 100 Nasdaq Stocks Used in the Study (The Last 30 Stocks Used in the 30-Dimensional Analysis)

NoSymbolNameSector
1AAPLApple Inc. Common StockTechnology
2MSFTMicrosoft Corporation Common StockTechnology
3GOOGAlphabet Inc. Class C Capital StockTechnology
4GOOGLAlphabet Inc. Class A Common StockTechnology
5AMZNAmazon.com Inc. Common StockConsumer Services
6MCHPMicrochip Technology Incorporated Common StockTechnology
7TSMTaiwan Semiconductor Manufacturing Company Ltd.Technology
8NVDANVIDIA Corporation Common StockTechnology
9VVisa Inc.Miscellaneous
10JPMJP Morgan Chase & Co. Common StockFinance
11IDXXIDEXX Laboratories Inc. Common StockHealth Care
12JNJJohnson & Johnson Common StockHealth Care
13WMTWalmart Inc. Common StockConsumer Services
14UNHUnited Health Group Incorporated Common Stock (DE)Health Care
15ADIAnalog Devices Inc. Common StockTechnology
16HDHome Depot Inc. (The) Common StockConsumer Services
17ASMLASML Holding N.V. New York Registry SharesTechnology
18PGProcter & Gamble Company (The) Common StockConsumer Non-Durables
19BACBank of America Corporation Common StockFinance
20MAMastercard Incorporated Common StockMiscellaneous
21KLACKLA Corporation Common StockCapital Goods
22DISWalt Disney Company (The) Common StockConsumer Services
23ADBEAdobe Inc. Common StockTechnology
24CMCSAComcast Corporation Class A Common StockConsumer Services
25NFLXNetflix Inc. Common StockConsumer Services
26CRMSalesforce.com Inc Common StockTechnology
27PFEPfizer Inc. Common StockHealth Care
28TMToyota Motor Corporation Common StockCapital Goods
29NKENike Inc. Common StockConsumer Non-Durables
30CSCOCisco Systems Inc. Common Stock (DE)Technology
31ORCLOracle Corporation Common StockTechnology
32KOCoca-Cola Company (The) Common StockConsumer Non-Durables
33TMOThermo Fisher Scientific Inc Common StockCapital Goods
34DHRDanaher Corporation Common StockHealth Care
35NVONovo Nordisk A/S Common StockHealth Care
36XOMExxon Mobil Corporation Common StockEnergy
37VZVerizon Communications Inc. Common StockPublic Utilities
38LLYEli Lilly and Company Common StockHealth Care
39ABTAbbott Laboratories Common StockHealth Care
40INTCIntel Corporation Common StockTechnology
41NTESNetEase Inc. American Depositary SharesMiscellaneous
42PEPPepsiCo Inc. Common StockConsumer Non-Durables
43ACNAccenture plc Class A Ordinary Shares (Ireland)Technology
44COSTCostco Wholesale Corporation Common StockConsumer Services
45TAT&T Inc.Consumer Services
46WFCWells Fargo & Company Common StockFinance
47NVSNovartis AG Common StockHealth Care
48CVXChevron Corporation Common StockEnergy
49MRKMerck & Company Inc. Common Stock (new)Health Care
50AZNAstraZeneca PLC American Depositary SharesHealth Care
51MSMorgan Stanley Common StockFinance
52MCDMcDonald’s Corporation Common StockConsumer Services
53TXNTexas Instruments Incorporated Common StockTechnology
54MDTMedtronic plc. Ordinary SharesHealth Care
55UPSUnited Parcel Service Inc. Common StockTransportation
56SAPSAP SE ADSTechnology
57NEENextEra Energy Inc. Common StockPublic Utilities
58PMPhilip Morris International Inc Common StockConsumer Non-Durables
59TMUST-Mobile US Inc. Common StockPublic Utilities
60LINLinde plc Ordinary ShareBasic Industries
61INTUIntuit Inc. Common StockTechnology
62QCOMQUALCOMM Incorporated Common StockTechnology
63HONHoneywell International Inc. Common Stock
64LOWLowe’s Companies Inc. Common StockConsumer Services
65ULUnilever PLC Common StockConsumer Non-Durables
66RYRoyal Bank Of Canada Common Stock
67ILMNIllumina Inc. Common StockHealth Care
68BHPBHP Group Limited American Depositary Shares (Each representing two Ordinary Shares)Basic Industries
69CCitigroup Inc. Common StockFinance
70SONYSony Group Corporation American Depositary SharesConsumer Non-Durables
71HDBHDFC Bank Limited Common StockFinance
72BMYBristol-Myers Squibb Company Common StockHealth Care
73AMTAmerican Tower Corporation (REIT) Common StockFinance
74SBUXStarbucks Corporation Common Stock
75BLKBlackRock Inc. Common StockFinance
76SCHWCharles Schwab Corporation (The) Common StockFinance
77UNPUnion Pacific Corporation Common StockTransportation
78BBLBHP Group PlcSponsored ADREnergy
79AXPAmerican Express Company Common StockFinance
80GSGoldman Sachs Group Inc. (The) Common StockFinance
81RTXRaytheon Technologies Corporation Common StockCapital Goods
82AMDAdvanced Micro Devices Inc. Common StockTechnology
83BABoeing Company (The) Common StockCapital Goods
84AMATApplied Materials Inc. Common StockCapital Goods
85AMGNAmgen Inc. Common StockHealth Care
86ISRGIntuitive Surgical Inc. Common StockHealth Care
87IBMInternational Business Machines Corporation Common StockTechnology
88SNYSanofi ADSHealth Care
89TGTTarget Corporation Common StockConsumer Services
90TDToronto Dominion Bank (The) Common Stock
91TTETotalEnergies SEEnergy
92ELEstee Lauder Companies Inc. (The) Common StockConsumer Non-Durables
93CVSCVS Health Corporation Common StockHealth Care
94GEGeneral Electric Company Common StockConsumer Durables
95DEODiageo plc Common StockConsumer Non-Durables
96SPGIS&P Global Inc. Common StockTechnology
97RIORio Tinto Plc Common StockBasic Industries
98JBHTJ.B. Hunt Transport Services Inc. Common StockTransportation
99DEDeere & Company Common StockCapital Goods
100CATCaterpillar Inc. Common StockCapital Goods

References

  1. Sklar, M. Fonctions de repartition an dimensions et leurs marges. Publ. L’Institut Stat. L’Universit’E Paris 1959, 8, 229–231. [Google Scholar]
  2. Achite, M.; Bazrafshan, O.; Pakdaman, Z.; Simsek, O.; Caloiero, T. Multivariate uncertainty analysis of severity-duration-magnitude frequency curves using Khoudraji copula and bootstrap method. Hydrol. Sci. J. 2025. [Google Scholar] [CrossRef]
  3. Deng, T.; Zeng, Z.; Xu, J. A copula-based statistical method of estimating production profile from temperature measurement for perforated horizontal wells. Pet. Sci. Technol. 2024, 42, 4589–4609. [Google Scholar] [CrossRef]
  4. Zhou, C.; van Nooijen, R.; Kolechkina, A.; Gargouri, E.; Slama, F.; van de Giesen, N. The uncertainty associated with the use of copulas in multivariate analysis. Hydrol. Sci. J. 2023, 68, 2169–2188. [Google Scholar] [CrossRef]
  5. Aghatise, O.; Khan, F.; Ahmed, S. Reliability assessment of marine structures considering multidimensional dependency of the variables. Ocean. Eng. 2021, 230, 109021. [Google Scholar] [CrossRef]
  6. Chen, P.; Zhang, J. Copula-based multivariate control charts for monitoring multiple dependent Weibull processes. Commun. Stat.-Simul. Comput. 2025, 0, 1–22. [Google Scholar] [CrossRef]
  7. Farkas, S.; Lopez, O. Semiparametric copula models applied to the decomposition of claim amounts. Scand. Actuar. J. 2024, 2024, 1065–1092. [Google Scholar] [CrossRef]
  8. Wang, Q.; Wu, X. A Nonparametric Bayesian Estimator of Copula Density with Applications to Financial Market. J. Bus. Econ. Stat. 2025, 43, 1021–1033. [Google Scholar] [CrossRef]
  9. Kim, J.M.; Ha, I.D.; Kim, S. Copula deep learning control chart for multivariate zero inflated count response variables. Statistics 2024, 58, 749–769. [Google Scholar] [CrossRef]
  10. McNeil, A.J.; Frey, R.; Embrechts, P. Quantitative Risk Management: Concepts, Techniques and Tools; Princeton Series in Finance; Princeton University Press: Princeton, NJ, USA, 2015. [Google Scholar]
  11. Wichitaksorn, N.; Gerlach, R.; Choy, S.T.B. Efficient MCMC estimation of some elliptical copula regression models through scale mixtures of normals. Appl. Stoch. Model. Bus. Ind. 2019, 35, 808–822. [Google Scholar] [CrossRef]
  12. Górecki, J.; Hofert, M.; Okhrin, O. Outer power transformations of hierarchical Archimedean copulas: Construction, sampling and estimation. Comput. Stat. Data Anal. 2021, 155, 107109. [Google Scholar] [CrossRef]
  13. Liang, P.; Rosen, O. Flexible Bayesian estimation of elliptical copulas. J. Comput. Graph. Stat. 2025, 0, 1–18. [Google Scholar] [CrossRef]
  14. Michaelides, M.; Cossette, H.; Pigeon, M. Simulations of Bivariate Archimedean Copulas from Their Nonparametric Generators for Loss Reserving under Flexible Censoring. N. Am. Actuar. J. 2025, 0, 1–28. [Google Scholar] [CrossRef]
  15. Aas, K.; Czado, C.; Frigessi, A.; Bakken, H. Pair-copula constructions of multiples dependence. Insur. Math. Econ. 2009, 44, 182–198. [Google Scholar] [CrossRef]
  16. Kurowicka, D.; Cooke, R.M. Distribution-free continuous Bayesian belief nets. In Modern Statistical and Mathematical Methods in Reliability; World Scientific: Singapore, 2005; Volume 10, pp. 309–322. [Google Scholar]
  17. Joe, H. Dependence Modelling with Copulas; Monographs on Statistics and Applied Probability: 134; CRC Press: Boca Raton, FL, USA, 2015. [Google Scholar]
  18. Müller, D.; Czado, C. Representing sparse Gaussian DAGs as sparse R-vines Allowing for Non-Gaussian Dependence. J. Comput. Graph. Stat. 2018, 27, 334–344. [Google Scholar] [CrossRef]
  19. Müller, D.; Czado, C. Dependence modelling in ultra high dimensions with vine copulas and the Graphical Lasso. Comput. Stat. Data Anal. 2019, 137, 211–232. [Google Scholar] [CrossRef]
  20. Brechmann, E.C.; Czado, C.; Aas, K. Truncated regular vines in high dimensions with application to financial data. Can. J. Stat. 2012, 40, 68–85. [Google Scholar] [CrossRef]
  21. Maya, R.A.L.; Gomez-Gonzalez, J.E.; Velandia, L.F.M. Latin American exchange rate dependencies: A regular vine copula approach. Contemp. Econ. Policy 2015, 33, 535–549. [Google Scholar] [CrossRef]
  22. Li, H.; Liu, Z.; Wang, S. Vines climbing higher: Risk management for commodity futures markets using a regular vine copula approach. Int. J. Financ. Econ. 2022, 27, 2438–2457. [Google Scholar] [CrossRef]
  23. Xiao, Q.; Yan, M.; Zhang, D. Commodity market financialization, herding and signals: An asymmetric GARCH R-vine copula approach. Int. Rev. Financ. Anal. 2023, 89, 102743. [Google Scholar] [CrossRef]
  24. Čeryová, B.; Árendáš, P. Vine copula approach to the intra-sectoral dependence analysis in the technology industry. Financ. Res. Lett. 2024, 60, 104889. [Google Scholar] [CrossRef]
  25. Hamza, T.; Ben Haj Hamida, H.; Mili, M.; Sami, M. High inflation during Russia–Ukraine war and financial market interaction: Evidence from C-Vine Copula and SETAR models. Res. Int. Bus. Financ. 2024, 70, 102384. [Google Scholar] [CrossRef]
  26. Czado, C. Pair-Copula Constructions of Multivariate Copulas. In Copula Theory and Its Applications; Jaworski, P., Durante, F., Härdle, W.K., Rychlik, T., Eds.; Springer: Berlin/Heidelberg, Germany, 2010; pp. 93–109. [Google Scholar]
  27. Joe, H.; Li, H.; Nikoloulopoulos, A.K. Tail dependence functions and vine copulas. J. Multivar. Anal. 2010, 101, 252–270. [Google Scholar] [CrossRef]
  28. Patton, A.J. A review of copula models for economic time series. J. Multivar. Anal. 2012, 110, 4–18. [Google Scholar] [CrossRef]
  29. Fan, Y.; Patton, A.J. Copulas in econometrics. Annu. Rev. Econ. 2014, 6, 179–200. [Google Scholar] [CrossRef]
  30. Dißmann, J.; Brechmann, E.C.; Czado, C.; Kurowicka, D. Selecting and estimating regular vine copulae and application to financial returns. Comput. Stat. Data Anal. 2013, 59, 52–69. [Google Scholar] [CrossRef]
  31. Vuong, Q.H. Likelihood ratio tests for model selection and non-nested Hypotheses. Econometrica 1989, 57, 307–333. [Google Scholar] [CrossRef]
  32. Merton, R.C. An intertemporal capital asset pricing model. Econometrica 1973, 41, 867–887. [Google Scholar] [CrossRef]
  33. Sun, F.; Fu, F.; Liao, H.; Xu, D. Analysis of multivariate dependent accelerated degradation data using a random-effect general Wiener process and D-vine Copula. Reliab. Eng. Syst. Saf. 2020, 204, 107168. [Google Scholar] [CrossRef]
  34. Nagler, T.; Bumann, C.; Czado, C. Model selection in sparse high-dimensional vine copula models with an application to portfolio risk. J. Multivar. Anal. 2019, 172, 180–192. [Google Scholar] [CrossRef]
  35. Chang, B.; Joe, H. Prediction based on conditional distributions of vine copulas. Comput. Stat. Data Anal. 2019, 139, 45–63. [Google Scholar] [CrossRef]
  36. Mantel, N.; Myers, M. Problems of convergence of maximum likelihood iterative procedures in multiparameter situations. J. Am. Stat. Assoc. 1971, 66, 484–491. [Google Scholar] [CrossRef]
  37. Loaiza-Maya, R.; Smith, M.S. Variational Bayes estimation of discrete-margined copula models with application to time series. J. Comput. Graph. Stat. 2019, 28, 523–539. [Google Scholar] [CrossRef]
  38. Chang, C.L.; McAleer, M.; Wong, W.K. Big data, computational science, economics, finance, marketing, management, and psychology: Connections. J. Risk Financ. Manag. 2018, 11, 15. [Google Scholar] [CrossRef]
  39. Oh, D.H.; Patton, A.J. High-dimensional copula-based distributions with mixed frequency data. J. Econom. 2016, 193, 349–366. [Google Scholar] [CrossRef]
  40. Pareek, M.K.; Thakkar, P. Surveying stock market portfolio optimization techniques. In Proceedings of the 2015 5th Nirma University International Conference on Engineering (NUiCONE), Ahmedabad, India, 26–28 November 2015; pp. 1–5. [Google Scholar]
  41. Kolm, P.N.; Tütüncü, R.; Fabozzi, F.J. 60 Years of portfolio optimization: Practical challenges and current trends. Eur. J. Oper. Res. 2014, 234, 356–371. [Google Scholar] [CrossRef]
  42. Bedford, T.; Cooke, R.M. Probability density decomposition for conditionally dependent random variables modeled by vines. Ann. Math. Artif. Intell. 2001, 32, 245–268. [Google Scholar] [CrossRef]
  43. Bedford, T.; Cooke, R.M. Vines: A new graphical model for dependent random variables. Ann. Stat. 2002, 30, 1031–1068. [Google Scholar] [CrossRef]
  44. Kurowicka, D.; Joe, H. (Eds.) Dependence Modeling: Vine Copula Handbook; World Scientific: Singapore, 2010. [Google Scholar]
  45. Hafner, C.M.; Kyriakopoulou, D. Exponential-Type GARCH models with linear-in-variance risk premium. J. Bus. Econ. Stat. 2021, 39, 589–603. [Google Scholar] [CrossRef]
  46. Nelson, D.B. Conditional heteroskedasticity in asset returns: A new approach. Econometrica 1991, 59, 347–370. [Google Scholar] [CrossRef]
  47. Hentschel, L. All in the family nesting symmetric and asymmetric GARCH models. J. Financ. Econ. 1995, 39, 71–104. [Google Scholar] [CrossRef]
  48. Jose Rodriguez, M.; Ruiz, E. Revisiting several popular GARCH models with leverage effect: Differences and similarities. J. Financ. Econom. 2012, 10, 637–668. [Google Scholar] [CrossRef]
  49. Sahamkhadam, M.; Stephan, A.; Östermark, R. Portfolio optimization based on GARCH-EVT-Copula forecasting models. Int. J. Forecast. 2018, 34, 497–506. [Google Scholar] [CrossRef]
  50. Engle, R. Risk and volatility: Econometric models and financial practice. Am. Econ. Rev. 2004, 94, 405–420. [Google Scholar] [CrossRef]
  51. Hosking, J.R.M.; Wallis, J.R. Parameter and Quantile Estimation for the generalized Pareto Distribution. Technometrics 1987, 29, 339–349. [Google Scholar] [CrossRef]
  52. Singh, V.P.; Guo, H. Parameter estimation for 3-parameter generalized Pareto distribution by the principle of maximum entropy (POME). Hydrol. Sci. J. 1995, 40, 165–181. [Google Scholar] [CrossRef]
  53. Koop, G.; Poirier, D.J.; Tobias, J.L. Bayesian Econometric Methods; Econometric Exercises; Cambridge University Press: Cambridge, UK, 2007. [Google Scholar]
  54. Scherr, S.; Zhou, J. Automatically Identifying Relevant Variables for Linear Regression with the Lasso Method: A Methodological Primer for its Application with R and a Performance Contrast Simulation with Alternative Selection Strategies. Commun. Methods Meas. 2020, 14, 204–211. [Google Scholar] [CrossRef]
  55. Nezamdoust, S.; Eskandari, F. Ridge Shrinkage Estimators in Finite Mixture of Generalized Estimating Equations. J. Math. Model. Financ. 2022, 2, 91–106. [Google Scholar]
  56. Liang, C.; Xu, Y.; Chen, Z.; Li, X. Forecasting China’s stock market volatility with shrinkage method: Can Adaptive Lasso select stronger predictors from numerous predictors? Int. J. Financ. Econ. 2022, 28, 3689–3699. [Google Scholar] [CrossRef]
  57. Al Janabi, M.A.M.; Arreola Hernandez, J.; Berger, T.; Nguyen, D.K. Multivariate dependence and portfolio optimization algorithms under illiquid market scenarios. Eur. J. Oper. Res. 2017, 259, 1121–1131. [Google Scholar] [CrossRef]
  58. Yu, W.; Bondell, H.D. Variational Bayes for Fast and Accurate Empirical Likelihood Inference. J. Am. Stat. Assoc. 2024, 119, 1089–1101. [Google Scholar] [CrossRef]
  59. Gunawan, D.; Tran, M.N.; Suzuki, K.; Dick, J.; Kohn, R. Computationally efficient Bayesian estimation of high-dimensional Archimedean copulas with discrete and mixed margins. Stat. Comput. 2019, 29, 933–946. [Google Scholar] [CrossRef]
  60. Tran, M.N.; Nguyen, T.N.; Dao, V.H. A practical tutorial on variational Bayes. arXiv 2021, arXiv:2103.01327. [Google Scholar] [CrossRef]
  61. Blei, D.M.; Kucukelbir, A.; McAuliffe, J.D. Variational inference: A review for statisticians. J. Am. Stat. Assoc. 2017, 112, 859–877. [Google Scholar] [CrossRef]
  62. Smith, M.S.; Loaiza-Maya, R.; Nott, D.J. High-dimensional copula variational approximation through transformation. J. Comput. Graph. Stat. 2020, 29, 729–743. [Google Scholar] [CrossRef]
  63. Ong, V.M.H.; Nott, D.J.; Smith, M.S. Gaussian variational approximation with a factor covariance structure. J. Comput. Graph. Stat. 2018, 27, 465–478. [Google Scholar] [CrossRef]
  64. Zeiler, M.D. ADADELTA: An Adaptive Learning Rate Method. arXiv 2012, arXiv:1212.5701. [Google Scholar] [CrossRef]
  65. Morales Napoles, O.; Cooke, R.M.; Kurowicka, D. About the Number of Vines and Regular Vines on n Nodes (Submitted for Publication). 2010. Available online: https://resolver.tudelft.nl/uuid:912abf55-8112-48d2-9cca-323f7f6aecc7 (accessed on 17 November 2025).
  66. Nikoloulopoulos, A.K.; Joe, H.; Li, H. Vine copulas with asymmetric tail dependence and applications to financial return data. Comput. Stat. Data Anal. 2012, 56, 3659–3673. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.