Next Article in Journal / Special Issue
Bayesian Nonparametric Measurement of Factor Betas and Clustering with Application to Hedge Fund Returns
Previous Article in Journal / Special Issue
Parallelization Experience with Four Canonical Econometric Models Using ParMitISEM
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Evolutionary Sequential Monte Carlo Samplers for Change-Point Models

Department of Economics, Laval University, 2216 Pavillon J.-A.-DeSève, QC G1V 0A6, Canada
Econometrics 2016, 4(1), 12; https://doi.org/10.3390/econometrics4010012
Submission received: 24 August 2015 / Revised: 27 December 2015 / Accepted: 28 January 2016 / Published: 8 March 2016
(This article belongs to the Special Issue Computational Complexity in Bayesian Econometric Analysis)

Abstract

:
Sequential Monte Carlo (SMC) methods are widely used for non-linear filtering purposes. However, the SMC scope encompasses wider applications such as estimating static model parameters so much that it is becoming a serious alternative to Markov-Chain Monte-Carlo (MCMC) methods. Not only do SMC algorithms draw posterior distributions of static or dynamic parameters but additionally they provide an estimate of the marginal likelihood. The tempered and time (TNT) algorithm, developed in this paper, combines (off-line) tempered SMC inference with on-line SMC inference for drawing realizations from many sequential posterior distributions without experiencing a particle degeneracy problem. Furthermore, it introduces a new MCMC rejuvenation step that is generic, automated and well-suited for multi-modal distributions. As this update relies on the wide heuristic optimization literature, numerous extensions are readily available. The algorithm is notably appropriate for estimating change-point models. As an example, we compare several change-point GARCH models through their marginal log-likelihoods over time.

1. Introduction

Sequential Monte Carlo (SMC) algorithm is a simulation-based procedure used in Bayesian frameworks for drawing distributions. Its core idea relies on an iterated application of the importance sampling technique to a sequence of distributions converging to the distribution of interest 1. For many years, on-line inference was the most relevant applications of SMC algorithms. Indeed, one powerful advantage of sequential filtering consists in being able to update the distributions of the model parameters in light of new coming data (hence the term on-line) allowing for an important time saving compared to off-line methods such as the popular Markov-Chain Monte-Carlo (MCMC) procedure that requires a new estimation based on all the data at each new observation entering in the system. Other SMC features making it very promising are an intuitive implementation based on the importance sampling technique ([1,2,3]) and a direct computation of the marginal likelihood (i.e., the normalizing constant of the targeted distribution, see, e.g., [4]).
Recently, the SMC algorithms have been applied to infer static parameters, field in which the MCMC algorithm excels. Neal [5] provides a relevant improvement in this direction by building a SMC algorithm, named annealed importance sampling (AIS), that sequentially evolves from the prior distribution to the posterior distribution using a tempered function, which basically consists in gradually introducing the likelihood information into the sequence of distributions by means of an increasing function. To preclude particles degeneracies, he uses an MCMC kernel at each SMC iteration. Few years later, [6] proposes an Iterated Batch Importance Sampling (IBIS) SMC algorithm, a special case of the Re-sample Move (RM) algorithm of [7], which sequentially evolves over time and adapts the posterior distribution using the previous approximate distribution. Again, an MCMC move (and a re-sampling step) is used for diversifying the particles. The SMC sampler (see [8]) unifies, among others, these SMC algorithms in a theoretical framework. It is shown that the methods of [5,7] arise as special cases with a specific choice of the ’backward kernel function’ introduced in their paper. These researches have been followed by empirical works (see [9,10,11]) where it is demonstrated that the SMC mixing properties often dominate the MCMC approach based on a single Markov-chain. Nowadays papers are devoted to build self-adapting SMC samplers by automatically tuning the MCMC kernel (e.g., [12]), by marginalizing the state vector (in a state space specification) using the particle MCMC framework (e.g., [13,14]), to construct efficient SMC samplers for parallel computations (see [15]) or to simulate from complex multi-modal posterior distributions (e.g., [16]).
In this paper, we document a generic SMC inference for change-point models that can additionally be updated through time. For example, in a model comparison context the standard methodology consists in repeating estimations of the parameters given an evolving number of observations. In circumstances where the Bayesian estimation is highly demanding as it is usually the case for complex models and where the number of available observations is huge, this iterative methodology can be too intensive. Change-point (CP) Generalized Autoregressive Conditional Heteroskedastic (GARCH) processes may require several hours for one inference (e.g., [17]). A recursive forecast exercise on many observations is therefore out of reach. Our first contribution is a new SMC algorithm, called tempered and time (TNT), which exhibits the AIS, the IBIS and the RM samplers as special cases. It innovates by switching over tempered and time domains for estimating posterior distributions. For instance, it firstly iterates from the prior to the posterior distributions by means of a sequence of tempered posterior distributions. It then updates in the time dimension the slightly different posterior distributions by sequentially adding new observations, each SMC step providing all the forecast summary statistics relevant for comparing models. The TNT algorithm combines the tempered approach of [5] with the IBIS algorithm of [6] if the model parameters are static or with the RM method of [7] if their support evolves with the SMC updates. Since all these methods are built on the same SMC steps (re-weighting, re-sampling and re-juvenating) and the same SMC theory, the combination is achieved without efforts.
The proposed methodology exhibits several advantages over SMC algorithms that directly iterate on the time domain ([6,7]). In fact, these algorithms may experience high particle discrepancies. Although the problem is more acute for models where the parameter space evolves through time, it remains an issue for models with static parameters at the very first SMC steps. To quote [6] (p. 546) :
Note that the particle system may degenerate strongly in the very early stages, when the evolving target distribution changes the most[...].
The combination of tempered and time SMC algorithms allows to limit this particle discrepancy observed at the early stage since the first posterior distribution of interest is estimated by taking into account more than a few observations. One advantage of using a sequence of tempered distributions to converge to the posterior distribution consists in the number of SMC steps that can be used. Compared to the SMC algorithms that directly iterate on time domain where the sequence of distributions is obviously defined by the number of data, the tempered approach allows for choosing this sequence of distribution and for targeting the posterior distribution of interest by using as many bridging distributions as needed.
Many SMC algorithms rely on MCMC kernels to rejuvenate the particles. The TNT sampler is no exception. We contribute by proposing several new generic MCMC kernels based on the heuristic optimization literature. These kernels are well appropriated in the SMC context as they build their updates on the particles interactions. We start by emphasizing that the DiffeRential Evolution Adaptive Metropolis (DREAM, see [18]), the walk move (see [19]) and the stretch one (see [20]) separately introduced in the statistic literature as generic Metropolis-Hastings proposals are in fact standard mutation rules of the Differential Evolution (DE) optimization. From this observation, we propose seven new MCMC updates based on the heuristic literature and emphasize that many other extensions are possible. The proposed MCMC kernel is adapted for continuous parameters. Consequently, discrete parameters such as the break parameters of change-point models cannot directly be inferred from our algorithm. To solve this issue, we transform the break parameters into continuous ones which make them identifiable up to a discrete value. To illustrate the potential of the TNT sampler, we compare several CP-GARCH models differing by their number of regimes on the S&P 500 daily percentage returns.
The paper is organized as follows. Section 2 presents the SMC algorithm as well as its theoretical derivation. Section 3 introduces the different Metropolis-Hastings proposals which compose what we call the Evolutionary MCMC. We then detail a simulation exercise on the CP-GARCH process in Section 4. Eventually we study the CP-GARCH performance on the S&P 500 daily percentage returns in Section 5. Section 6 concludes.

2. Off-line and On-Line Inferences

We first theoretically and practically introduce the tempered and time (TNT) framework. To ease the discussion, let us consider a standard state space model:
y t = f ( θ , s t , ω t )
s t = g ( θ , s t - 1 , v t )
where s t is a random variable driven by a Markov chain and the functions f(-) and g(-) are deterministic given their arguments. The observation y t belongs to the set y 1 : T = { y 1 , . . . , y T } with T denoting the sample size and is assumed to be independent conditional to the state s t and θ with distribution f ( y t | θ , s t ) . The innovations ω t and v t are mutually independent and stand for the noise of the observation/state equations. The model parameters included in θ do not evolve over time (i.e., they are static). Let us denote the set of parameters at time t by x t = { θ , s 1 : t } defined on the measurable space E t .
We are interested in estimating many posterior distributions starting from π ( x τ | y 1 : τ ) , where τ < < T , until T. The SMC algorithm approximates these posterior distributions with a large (dependent) collection of M weighted random samples { W t i , x t i } i = 1 M where W t i > 0 and i = 1 M W t i = 1 such that as M , the empirical distribution converges to the posterior distribution of interest, meaning that for any π ( x t | y 1 : t ) -integrable function g : E t :
i = 1 M W t i g ( x t i ) E t g ( x t ) π ( x t | y 1 : t ) d x t almost surely .
The TNT method combines an enhanced Annealed Importance sampling 2 (AIS, see [5]) with the Re-sample Move (RM) SMC inference of [7] 3. To build the TNT algorithm, we rely on the theoretical paper of [8] that unifies the two SMC methods into one SMC framework called “SMC sampler”. The TNT algorithm first estimates an initial posterior distribution, namely π ( x τ | y 1 : τ ) , by an enhanced AIS (E-AIS) algorithm and then switches from the tempered domain to the time domain and sequentially updates the approximated distributions from π ( x τ | y 1 : τ ) to π ( x T | y 1 : T ) by adding one by one the new observations. We now begin by mathematically deriving the validity of the SMC algorithms under the two different domains and by showing that they are particular cases of the SMC sampler. The practical algorithm steps are given afterward (see Subsection 2.3).

2.1. E-AIS : The Tempered Domain

The first phase, carried out by an E-AIS, creates a sequence of probability measures { π n } n = 0 p that are defined on the measurable spaces { E n , ξ n } , where E n = E n + 1 = E x τ , n { 0 , 1 , . . . , p } is a counter and does not refer to ’real time’, p denotes the number of posterior distribution estimations and π p coincides with the first posterior distribution of interest π ( x τ | y 1 : τ ) . The sequence distribution, used as bridge distributions, is defined as π n ( x n | y 1 : τ ) = γ ( y 1 : τ | x n ) ϕ ( n ) f ( x n ) / Z n where Z n = E γ ( y 1 : τ | x n ) ϕ ( n ) f ( x n ) d x n denotes the normalizing constant, γ ( y 1 : τ | x n ) and f ( x n ) respectively are the likelihood function and the prior density of the model. Through an increasing function ϕ ( n ) respecting the bound conditions ϕ ( 0 ) = 0 and ϕ ( p ) = 1 , the E-AIS artificially builds a sequence of distributions that converges to the posterior distribution of interest.
Remark 1: The random variables { x n } n = 0 p exhibit the same support E which is also shared by x τ . Furthermore, the random variable x τ coincides with x p since ϕ ( p ) = 1 .
The E-AIS is merely a sequential importance sampling technique where the draws of a proposal distribution η n combined with an MCMC kernel are used to approximate the next posterior distribution π n + 1 , the difficulty lying in specifying the sequential proposal distribution. Del Moral, Doucet, and Jasra [8] theoretically develop a coherent framework for choosing a generic sequence of proposal distributions.
In the SMC sampler framework, we augment the support of the posterior distribution ensuring that the targeted posterior distribution marginally arises :
π ˜ n ( x 1 : n ) = π n ( x n ) k = 2 n L k ( x k - 1 | x k ) , = γ n ( x n ) Z n k = 2 n L k ( x k - 1 | x k ) ,
where γ n ( x n ) = γ ( y 1 : τ | x n ) ϕ ( n ) f ( x n ) , Z n = E γ ( y 1 : τ | x n ) ϕ ( n ) f ( x n ) d x n is the normalizing constant, and L k ( x k - 1 | x k ) is a backward MCMC kernel such that E L k ( x k - 1 | x k ) d x k - 1 = 1 .
By defining a sequence of proposal distributions as
η n ( x 1 : n ) = f ( x 1 ) k = 2 n K k ( x k | x k - 1 ) ,
where K k ( x k | x k - 1 ) is an MCMC kernel with stationary distribution π k such that it verifies π k ( x k ) = E K k ( x k | x k - 1 ) π k ( x k - 1 ) d x k - 1 , we derive a recursive equation of the importance weight:
w n ( x 1 : n ) = γ n ( x n ) k = 2 n L k ( x k - 1 | x k ) Z n f ( x 1 ) k = 2 n K k ( x k | x k - 1 ) , = w n - 1 ( x 1 : n - 1 ) Z n - 1 γ n ( x n ) L n ( x n - 1 | x n ) Z n γ n - 1 ( x n - 1 ) K n ( x n | x n - 1 ) .
For a smooth increasing tempered function ϕ ( n ) , we can argue that π n - 1 will be close to π n . W therefore define the backward kernel by detailed balance argument as
L n ( x n - 1 | x n ) = π n ( x n - 1 ) π n ( x n ) K n ( x n | x n - 1 ) .
It gives the following weights :
w n ( x 1 : n ) = w n - 1 ( x 1 : n - 1 ) Z n - 1 γ n ( x n - 1 ) Z n γ n - 1 ( x n - 1 ) , w n - 1 ( x 1 : n - 1 ) γ ( y 1 : τ | x n - 1 ) ϕ n - ϕ n - 1 .
The normalizing constant Z n is approximated as
Z n Z n - 1 i = 1 M W n - 1 i γ n ( x n - 1 i ) γ n - 1 ( x n - 1 i ) ,
where W n - 1 i = w n - 1 ( x 1 : n - 1 i ) / j = 1 M w n - 1 ( x 1 : n - 1 j ) , i.e., the normalized weight.
The E-AIS requires to tune many parameters : an increasing function ϕ ( n ) , the MCMC kernels with the invariant distributions π n ( . ) , a number of particles M, of iterations p, of MCMC steps J. Adjusting these parameters can be difficult. Some guidance are given in [16] for DSGE models. For example, they propose a quadratic tempered function ϕ ( n ) . It slowly increases for small values of n and the step becomes larger and larger as n tends to p. In this paper, the TNT algorithm generically adapts the different user-defined parameters and belongs to the class of adaptive SMC algorithms. It automatically adjusts the tempered function with respect to an efficiency measure as it was proposed by [10]. By doing so, we preclude the difficult choice of the function ϕ ( n ) and the number of iteration p. The number of MCMC steps J will be controlled by the acceptance rate exhibited by the MCMC kernels. The choice of MCMC kernels and the number of particles are discussed later (see Section 3).

2.2. The Re-Sample Move Algorithm : The Time Domain

Once we have a set of particles that approximates the first posterior distribution of interest π ( x τ | y 1 : τ ) , a second phase takes place. Firstly, let us assume that the support of x t does not evolve over time (i.e., x t E t ). In this context, the SMC sampler framework shortly reviewed here for the tempered domain still applies. Let us define the following distributions:
π t ( x t ) = π ( x t | y 1 : t ) , π ˜ t ( x 1 : t ) = π t ( x t ) k = 2 t L k ( x k - 1 | x k ) , η t ( x 1 : t ) = f ( x 1 ) k = 2 t K k ( x k | x k - 1 ) , L k ( x k - 1 | x k ) = π k ( x k - 1 ) π k ( x k ) K k ( x k | x k - 1 ) .
Then the weight equation of the SMC sampler is equal to
w t ( x 1 : t ) = w t - 1 ( x 1 : t - 1 ) π t ( x t - 1 ) π t - 1 ( x t - 1 ) ,
which is exactly the weight equation of the IBIS algorithm (see [6], step 1, p. 543).
Let now consider the more difficult case where a subset of the support of x t evolves with t such as x t = { x t - 1 , s t } = { θ , s 1 : t - 1 , s t } (see the state space model Equations (1) and (2)) meaning that t [ 1 , T ] , x t E t and E t - 1 E t . The previous method cannot directly be applied (due to the backward kernel) but with another choice of the kernel functions, the SMC sampler also operates. Let us define the following distribution :
π t ( x t ) = π ( x t | y 1 : t ) ,
π ˜ t ( x 1 : t , x 2 : t * ) = π t ( x t ) k = 2 t L k ( x k - 1 , x k * | x k ) ,
η t ( x 1 : t , x 2 : t * ) = f ( x 1 ) k = 2 t K ˜ k ( x k , x k * | x k - 1 ) ,
K ˜ k ( x k , x k * | x k - 1 ) = q ˜ k ( x k * | x k - 1 ) K k ( x k | x k * ) ,
L k ( x k - 1 , x k * | x k ) = q k ( x k - 1 | x k * ) K k ( x k * | x k ) ,
K k ( x k * | x k ) = π k ( x k * ) K k ( x k | x k * ) π k ( x k ) by detailed balance argument .
To deal with the time-varying dimension of x t , we augment the support of the artificial sequence of distributions by several new random variables (see x 2 : t * in Equation (7)) while ensuring that the posterior distribution of interest π t ( x t ) marginally arises. Sampling from the proposal distribution η t ( x 1 : t , x 2 : t * ) is achieved by drawing from the prior distribution and then by sequentially sampling from the distributions K ˜ k ( x k , x k * | x k - 1 ) , which are composed by a user-defined distribution q ˜ k ( x k * | x k - 1 ) and an MCMC kernel exhibiting π k ( x k ) as the invariant distribution.
Under this framework, the weight equation of the SMC sampler becomes
w t ( x 1 : t , x 2 : t * ) = w t - 1 ( x 1 : t - 1 , x 2 : t - 1 * ) π t ( x t * ) q t ( x t - 1 | x t * ) π t - 1 ( x t - 1 ) q ˜ t ( x t * | x t - 1 ) .
By setting the distributions q t ( x t - 1 | x t * ) = δ { x t - 1 = θ * , s 1 : t - 1 * } , where δ i denotes the probability measure concentrated at i, and q ˜ t ( x t * | x t - 1 ) = ν t ( s t * | x t - 1 ) δ { θ * , s 1 : t - 1 * = x t - 1 } , we recover the weight equation of [7] (see Equation (20), p. 135)
w t ( x 1 : t , x 2 : t * ) = w t - 1 ( x 1 : t - 1 , x 2 : t - 1 * ) π t ( x t - 1 , s t * ) π t - 1 ( x t - 1 ) ν t ( s t * | x t - 1 ) .
Like in [7], only the distribution ν t ( . ) has to be specified. For example, it can be set either to the prior distribution or the full conditional posterior distribution (if the latter exhibits a closed form).
Remark 2: The division π ( x t | y 1 : t ) π ( x t - 1 | y 1 : t - 1 ) appearing in the weight Equations ((5) and (13)) can be reduced to Z t - 1 Z t γ ( y t | x t , y 1 : t - 1 ) f ( s t | x t - 1 ) which highly limits the computational cost of the weights.

2.3. The TNT Algorithm

The algorithm initializes the M particles using the prior distributions, sets each initial weight { W 0 i } i = 1 M to W 0 i = 1 M and then iterates from n = 1 , , p , p + 1 , . . . , p + ( T - τ ) + 1 as follows
  • Correction step: i [ 1 , M ] , Re-weight each particle with respect to the nth posterior distribution
    If in tempered domain ( n p ) :
    w ˜ n i = γ ( y 1 : τ | x n - 1 i ) ϕ ( n ) - ϕ ( n - 1 )
    If in time domain ( n > p ) and the parameter space does not evolve over time (i.e., E n - 1 = E n ) :
    w ˜ n i = γ ( y 1 : τ + n - p | x n - 1 i ) γ ( y 1 : τ + n - p - 1 | x n - 1 i ) = γ ( y τ + n - p | x n - 1 i , y 1 : τ + n - p - 1 ) ( see remark 2 )
    If in time domain ( n > p ) and the parameter space increases (i.e., E n - 1 E n ) :
    Set x n i = { x n - 1 i , s n i } with s n i ν n ( . | x n - 1 i ) .
    w ˜ n i = γ ( y 1 : τ + n - p | x n i ) f ( x n i ) γ ( y 1 : τ + n - p - 1 | x n - 1 i ) f ( x n - 1 i ) ν n ( s n i | x n - 1 i )
    Compute the unnormalized weights : W ˜ n i = w ˜ n i W n - 1 i .
    Normalize the weights : W n i = W ˜ n i j = 1 M W ˜ n j
  • Re-sampling step: Compute the Effective Sample Size (ESS) as
    E S S = 1 i = 1 M ( W n i ) 2
    If E S S < κ where κ is a user-defined threshold then re-sample the particles and reset the weight uniformly.
  • Mutation step: i [ 1 , M ] , run J steps of an MCMC kernel with an invariant distribution π n ( x n | y 1 : τ ) for n p and π ( x τ + n - p | y 1 : τ + n - p ) for n > p .
Remark 3: According to the algorithm derivation, note that the mutation step is not required at each SMC iteration.
When the parameter space does not change over time (i.e., tempered or time domains with E n - 1 = E n ), the algorithm reduces to the SMC sampler with a specific choice of the backward kernel (see Equation (3), more discussions in [8]) that implies that π n - 1 ( . ) must be close to π n ( . ) for non-degenerating estimations. The backward kernel is introduced for avoiding the choice of an importance distribution at each iteration of the SMC sampler. This specific choice of the backward kernel does not work for model where the parameter space increases with the sequence of posterior distributions (hence the use of a second weighting scheme when n > p , see Equation (12)) but the algorithm also reduces to a SMC sampler with another backward kernel choice (see Equation (10)). In the empirical applications, we first estimate an off-line posterior distribution with fixed parameters and then by just switching the weight equation, we sequentially update the posterior distributions by adding new observations. This two phases preclude the particle degeneration that may occur at the early stage of the SMC algorithms that directly iterate on time such as the IBIS and the RM algorithms. The tempered function ϕ ( n ) allows for converging to the first targeted posterior distribution as slowly as we want. Indeed, as we are not constraint by the time domain, we can sequentially iterate as much as needed to get rid of the degeneracy problem. The choice of the tempered function ϕ ( n ) is therefore relevant. In the spirit of a black-box algorithm as the IBIS one is, the Section 2.4 shows how the TNT algorithm automatically adapts the tempered function at each SMC iteration.
During the second phase (i.e., updating the posterior distribution through time), one may observe high particle discrepancies especially when the space of the parameters evolves over time 4. In that case, one can run an entire E-AIS on the data y 1 : τ + n - p when a degeneracy issue is detected (i.e., the ESS falls below a user-defined value κ 1 < κ ). The adaptation of the tempered function (discussed in the next section) makes the E-AIS faster than usual since it reduces the number of iteration p at its minimum given the ESS threshold κ. Controlling for the degeneracy issue is therefore automated and a minimal number of effective sample size is ensured at each SMC iteration.

2.4. Adaptation of the Tempered Function

Previous works on the SMC sampler usually provide a tempered function ϕ ( n ) obtained by several empirical trials 5, making these functions model-dependent. Jasra et al. [10] innovate by proposing a generic choice of ϕ ( n ) that only requires a few more codes. The E-AIS correction step (see Equation (14)) of iteration n is modified as follows
  • Find ϕ ¯ n such that
    1 i = 1 M ( W n i ) 2 = 0 . 95 ESS n - 1 ,
    where ESS n - 1 refers to the Effective Sample Size of the previous SMC iteration, W n i = W ˜ n i j = 1 M W ˜ n j is the normalized weights and the unnormalized W ˜ n i depends on ϕ ¯ n as
    w ˜ n i = γ ( y 1 : τ | x n - 1 i ) ϕ ¯ n γ ( y 1 : τ | x n - 1 i ) ϕ ¯ n - 1 .
  • Compute the normalized weights { W n i } i = 1 M under the value of ϕ ¯ n
Roughly speaking, we find the value ϕ ¯ n that makes the ESS criterion close to the previous one in order to keep the artificial sequence of distributions very similar as required by the choice of the backward kernel (3).
Because the tempered function is adapted on the fly using the SMC history, the usual SMC asymptotic results do not apply. Del Moral et al. [21] and Beskos et al. [22] provide asymptotic results by assuming that the adapted tempered function converges to the optimal one (if it exists).

3. Choice of MCMC Kernels

The MCMC kernel is the most computational demanding step of the algorithm and it determines the posterior support exploration, making its choice very relevant. Chopin [6] emphasizes that the IBIS algorithm is designed to be a true “black box” (i.e., whose the sequential steps are not model-dependent), reducing the task of the practitioner to only supply the likelihood function and the prior densities. For this purpose, a natural choice of the MCMC kernel is the Metropolis-Hastings with an independent proposal distribution whose summary statistics are derived from the particles of the previous SMC step and from the weight of the current step. The IBIS algorithm uses an independent Normal proposal. It is worth noting that this “black box” structure is still applicable in this framework that combines SMC iterations on tempered and time domains.
Nevertheless the independent Metropolis-Hastings kernel may perform poorly at the early stage of the algorithm if the posterior distribution is well behaved and at any time otherwise. We rather suggest using a new adaptative Metropolis algorithm of random walk (RW) type that is generic, fully automated, suited for multi-modal distributions and that dominates most of the other RW alternatives in terms of sampling efficiencies. The algorithm is inspired from the heuristic Differential Evolution (DE) optimization literature (for a review, see [23]).
The DE algorithms have been designed to solve optimization problems without requiring derivatives of the objective function. The algorithms are initiated by randomly generating a set of parameter values. Afterward, relying on a mutation rule and a cross-over (CR) probability, these parameters are updated in order to explore the space and to converge to the global optimum. The mutation equation is usually linear with respect to the parameters and the CR probability determines the number of parameters that changes at each iteration. The first DE algorithm dates back to [24]. Nowadays, numerous alternatives based on this principle have been designed and many of them display a different mutation rule. Considering a set of parameters { x t i } i = 1 M lying in d , the standard algorithm operates by sequentially updating each parameter given the other ones. For a specific parameter x t j , the mutation equation to obtain a new value x ˜ t are typically chosen among the following ones
x ˜ t = x t j + F ( g = 1 2 x t r 1 ( g ) - h = 1 2 x t r 2 ( h ) ) ,
x ˜ t = x t j + F ( x t j - x t best ) + F ( x t r 1 - x t r 2 ) ,
x ˜ t = x t best + F ( g = 1 2 x t r 1 ( g ) - h = 1 2 x t r 2 ( h ) ) ,
where i r 1 ( g ) , r 2 ( h ) ; r 1 ( . ) and r 2 ( . ) stand for random integers uniformly distributed on the support [ 1 , M ] - i and it is required that r 1 ( g ) r 2 ( h ) when g = h , F is a fixed parameter and x t best denotes the parameter related to the highest objective function in the swamp. Then, for each element of the new vector x ˜ t , the CR step consists in replacing its value with the one of x t j according to a fixed probability.
The DE algorithm is appealing in an MCMC context as it has been built up to explore and find the global optimum of complex objective functions. However, the DE method has to be adapted if one wants to draw realizations from a complex distribution. To employ the mutation Equations (17)–(19) into an MCMC algorithm, we need to insure that the detailed balance is preserved, that the Markov-chain is ergodic with a unique stationary distribution and that this distribution is the targeted one. To do so, we slightly modify the mutation equations as follows
x ˜ t = x t j + F ( δ , d ) ( g = 1 δ x r 1 ( g ) - h = 1 δ x r 2 ( h ) ) + ζ ,
x ˜ t = x t j + Z W ( x t j - g = 1 δ x t r 1 ( g ) δ ) ,
x ˜ t = g = 1 δ x t r 1 ( g ) δ + Z Stretch ( x t j - g = 1 δ x t r 1 ( g ) δ ) ,
in which ζ N ( 0 , η x 2 I ) ; δ U [ 1 , 3 ] , F ( δ , d ) is a fixed parameter, Z RW and Z Stretch are random variables driven by two different distributions defined below.
These three update rules (20)–(22) are valid in an MCMC context and have been separately proposed in the literature. The first Equation (20) refers to the DiffeRential Evolution Adaptive Metropolis (DREAM) proposal distribution of [18] and is the MCMC analog of the DE mutation (17). In their paper, it is shown that the proposal distribution is symmetric and so that the acceptance ratio is independent of the proposal density. Also, they fix η x to a very small value (such as 1e-4) and F ( δ , d ) to 2 . 38 / 2 δ d because it constitutes the asymptotic optimal choice for a multivariate Normal posterior distribution as demonstrated in [25].6 Since the posterior distribution is rarely a Normal one, we prefer adapting F ( δ , d ) from one SMC iteration to another so that the scale parameter is fixed during the entire MCMC moves of each SMC step. The adapting procedure is detailed below. Importantly, [18] provide empirical evidence that the DREAM equation dominates most of the other RW alternatives (including the optimal scaling and the adaptive ones) in terms of sampling efficiencies.
The second Equation (21) is an adapted version of the walk move of [19] and can be thought as the MCMC equivalence of the mutation (18).7 When the density g W ( . ) of Z W verifies g W ( - z / ( 1 + z ) ) = ( 1 + z ) g W ( z ) , it can be shown that the proposal parameter x ˜ t is accepted with a probability given by
min { | 1 + Z W | d - 1 π ( x ˜ t | y 1 : t ) π ( x t j | y 1 : t ) , 1 } .
As in their paper, we set the density to g W ( z ) 1 / 1 + z if z [ - a W 1 + a W , a W ] (with a + ) and zero otherwise. The cumulative density function, its inverse and the first two moments of the distributions are given by
F Z W ( x ) = 1 - ( a W + 1 ) 1 / 2 - ( x + 1 ) 1 / 2 ( a W + 1 ) 1 / 2 - ( a W + 1 ) - 1 / 2 , F Z W - 1 ( u ) = - 1 + [ ( a W + 1 ) - 1 / 2 + u ( ( a W + 1 ) 1 / 2 - ( a W + 1 ) - 1 / 2 ) ] 2 , E ( Z W ) = a W 2 3 ( a W + 1 ) , V ( Z W ) = a W 2 4 a W 2 + 15 a W + 15 45 ( a W + 1 ) 2 .
In the seminal paper of the walk move, the parameter a W is set to 2. However, we rather suggest solving the equation V ( Z W ) = 2 . 38 / 2 δ d in order to obtain the optimal value of a W . Note that Equation (21) is slightly different from the standard walk move of [19] in the sense that the random parameter δ can be greater than one and also because only one realization from Z W is generated in order to update an entire new vector x t j . The latter modification is motivated by the success of the novel MCMC algorithm based on deterministic proposals (see the Transformation-based MCMC of [26]) and by the DREAM update which also exhibits one (fixed) parameter F ( δ , d ) to propose the entire new vector.
Lastly, the third Equation (22) corresponds to the stretch move proposed in [19] and improved by [20]. The probability of accepting the proposal x ˜ t is
min { | Z S | d - 1 π ( x ˜ t | y 1 : t ) π ( x t j | y 1 : t ) , 1 } ,
when the density g S ( . ) of Z S verifies z g S ( z ) = g S ( z - 1 ) . We adopt the same density function as in their paper which is given by g S ( z ) = 1 / z for z [ 1 / a S , a S ] , a S + and zero otherwise. The corresponding cumulative density function, its inverse, the expectation and the variance are analytically tractable and are given by
F Z S ( x ) = a S x - 1 a - 1 , F Z S - 1 ( u ) = ( u ( a S - 1 ) + 1 ) 2 a S , E ( Z S ) = a s + a S - 1 + 1 3 , V ( Z S ) = ( a S - 1 ) 2 ( 4 a S 2 + 7 a S + 4 ) 45 a S .
In the standard stretch move, the parameter a S is set to 2.5. Like the DREAM algorithm, the stretch move has been proven to be a powerful generic MCMC approach to generate complex posterior distributions. The method is becoming very popular in astrophysics (see references in [20]).
Once it is recognized that all these updates are also involved in the DE optimization problems, incorporating many other techniques from the latter becomes straightforward. To highlight the potential, we extend the DREAM, the walk and the stretch moves by proposing new update equations that are derived from the trigonometric move, the standard DE mutation and the firefly optimization.
In the DE literature, [27] suggest using a trigonometric mutation equation based on three random parameters x t r 1 , x t r 2 , x t r 3 and their corresponding posterior density values γ ( y 1 : t | x t r i ) f ( x t r i ) with i = 1 , 2 , 3 . From these quantities, the new parameter is given by
x t trigo = i = 1 3 x t r i / 3 + ( p 2 - p 1 ) ( x t r 1 - x t r 2 ) + ( p 3 - p 2 ) ( x t r 2 - x t r 3 ) + ( p 1 - p 3 ) ( x t r 3 - x t r 1 ) ,
in which p i γ ( y 1 : t | x t r i ) f ( x t r i ) for i [ 1 , 3 ] are probabilities such that i = 1 3 p i = 1 . Similarly, we can extend the DREAM, the walk and the stretch moves using the trigonometric parameter as follows
DREAM : x ˜ t = x t j + Z Dir F ( δ = 1 , d ) ( x t trigo - x t q ) + ζ ,
Walk move : x ˜ t = x t j + Z RW ( x t j - x t trigo ) ,
Stretch move : x ˜ t = x t trigo + Z Stretch ( x t j - x t trigo ) ,
where Z Dir = 1 with probability 0 . 5 and -1 otherwise. Note that due to the random variable Z Dir , the DREAM proposal (25) is still symmetric and therefore the acceptance ratio remains identical to the standard RW one.
The last two extensions are adaptations only for the stretch and the walk moves (as for the DREAM one, it does not change the initial proposal distribution). The next proposal comes from another heuristic optimization technique. The firefly (FF) algorithm, initially introduced in [28], updates the parameters by combining the attractiveness and the distance of the particles. For our purpose, we define the FF update as
x t FF = x t r 1 + F F F ( x t r 1 - x t r 2 ) ,
where F F F is a chosen constant and r 1 , r 2 are taken without replacement in the M - 1 remaining particles. The two new moves based on the FF equation are given by
Walk move : x ˜ t = x t j + Z RW ( x t j - x t FF ) , = x t j + Z RW ( x t j - x t r 1 ) + Z RW F F F ( x t r 1 - x t r 2 ) ,
Stretch move : x ˜ t = x t FF + Z Stretch ( x t j - x t FF ) , = x t r 1 + Z Stretch ( x t j - x t r 1 ) + ( 1 + Z Stretch ) F F F ( x t r 1 - x t r 2 ) .
We set F F F of the walk move to 2 . 38 E ( Z RW ) 2 d and the constant F F F of the stretch move is fixed to E ( Z Stretch ) E ( Z Stretch ) + 1 .
Regarding the last new updates, one can notice that the standard Differential Evolution mutation can also be used to improve the proposal distribution of the stretch and the walk moves. In particular, we consider the move of the DE optimization given by
x t DE = x t r 1 + F D E ( x t r 2 - x t r 3 ) ,
in which F D E is a fixed constant and r 1 , r 2 , r 3 are taken without replacement in the M - 1 remaining particles. Inserting this update into the stretch and the walk moves delivers new proposal distributions as follows
Walk move : x ˜ t = x t j + Z RW ( x t j - x t DE ) , = x t j + Z RW ( x t j - x t r 1 ) + Z RW F D E ( x t r 2 - x t r 3 ) ,
Stretch move : x ˜ t = x t DE + Z Stretch ( x t j - x t DE ) , = x t r 1 + Z Stretch ( x t j - x t r 1 ) + ( 1 + Z Stretch ) F D E [ ( x t r 2 - x t r 3 ) ] .
Similarly to the Firefly proposal, we fix F D E of the walk move to 2 . 38 E ( Z RW ) 2 d and the constant F D E of the stretch move is set to E ( Z Stretch ) E ( Z Stretch ) + 1 .
The standard DREAM, the walk and the stretch moves are typically used in an MCMC context. However, when the parameter dimension d is large, many parallel chains must be run because, as all these updates are based on linear transformations, they can only generate subspaces spanned by their current positions. To remedy this issue in the MCMC scheme, [18] have introduced the CR probability. Once the proposal parameter has been generated, each element is randomly kept or set back to the previous value according to some fixed probability p C R . Eventually, the standard MH acceptance step takes place. In contrast, these multiple chains arise naturally in SMC frameworks since the rejuvenate step consists in updating all the particles by some MCMC iterations. However, the CR probability has the additional advantage of generating many other moves of the parameters. For this reason, we also include the CR step into our MCMC kernel.
In order to test all the new move strategies, Table 1 documents the average autocorrelation times over the multivariate random realizations (computed by batch means, see [29]), obtained from each update rule. The dimension of each distribution from which the realizations are sampled is set to 5 and we consider Normal distributions with low and high correlations as well as a student distribution with a degree of freedom equal to 5. From this short analysis, the DREAM update is the most efficient in terms of mixing. We also observe that the additional moves perform better than the standard ones for the walk and the stretch moves.
As the posterior distribution can take many different shapes, a specific MCMC kernel which may work in bags of situations can fail for some ill posterior distributions (see for example the anisotropic density in [20] or the twisted gaussian distribution in [18]). An appealing automatic approach is to use several kernels which can behave differently depending on the posterior distribution. To do so, we suggest to incorporate all the generic moves in combination with a fixed CR probability into the MCMC rejuvenation step of the SMC algorithm. In practice, at each MCMC iteration, the proposal distribution is chosen among the different update Equations ((20)–(22), (25)–(31)) according to a multinomial probability p kernel . Then, some of the new elements of the updated vector are set back to their current MCMC value according to the CR probability. The proposal is then accepted with probability that is defined either by the standard RW Metropolis ratio, by (23) or by (24) depending on the selected mutation rule.8 By assessing the efficiency of each update equation with the Mahalanobis distance, one can monitor which proposal leads to the best exploration of the support and can appropriately and automatically adjust the probability p kernel at the end of the rejuvenation step. More precisely, once a proposed parameter is accepted, we add the Mahalanobis distance between the previous and the accepted parameters to the distance already achieved by the selected move. At the end of each rejuvenation step, the probabilities p kernel are reset proportionally to the distance performances of all the moves.
Two relevant issues should be discussed. First, the MCMC kernel makes interacting the particles, which rules out the desirable parallel property of the SMC. To keep this advantage, we apply the kernel on subsets of particles instead of on all the particles and we perform paralelization between the subsets. Secondly and more importantly, the SMC theory derived in Section 2 does not allow for particle interactions. Proposition 1 ensures that the TNT sampler also works under a DREAM-type MCMC kernel.
Proposition 1. 
Consider a SMC sampler with a given number of particles M and the MCMC kernels given by the proposal distribution (20) or (25). Then, it yields a standard SMC sampler with particle weights given by the Equation (4).
Proof. 
Adapting the proof for the walk and the stretch moves is straightforward as the stationary distribution of the Markov-chain also factorizes into a product of the targeted distribution.

Adaptation of the Scale Parameters F ( δ , d ) , a W and a S

Since the chosen backward MCMC kernel in the algorithm derivation implies that the consecutive distributions approximated by the TNT sampler are very similar, we can analyze the mixing properties of the previous MCMC kernel to adapt the scale parameters F ( δ , d ) , a W and a S . Atchadé and Rosenthal [31] present a simple recursive algorithm in order to achieve a specified acceptance rate in an MCMC framework. Considering one scale parameter (either F ( δ , d ) , a W or a S ) generically denoted by c n - 1 , at the end of the n - 1 SMC step, we adapt the parameter as follows :
c n - 1 = p ( c n - 2 + α n - 1 - α targeted ( n - 1 ) 0 . 6 )
where the function p ( . ) is such that p ( c ) = c if c [ A 0 , + ] and p ( c ) = A 0 if c < A 0 , the parameter α n - 1 stands for the acceptance rate of the MCMC kernel of the n - 1 SMC step and α targeted is a user-defined acceptance rate. The function p ( . ) prevents from negative values of the recursive equation and if the optimal scale parameter lies in the compact set [ A 0 , + ] , the equation will converge to it (in an MCMC context). In the empirical exercise, we fix the variable A 0 to 1e-8 for the DREAM-type move and to 1.01 for the other updates. The rate α targeted is set to 1 3 implying that every three MCMC iterations, all the particles have been approximately rejuvenated. It is worth emphasizing that the denominator ( n - 1 ) 0 . 6 has been chosen as proposed in [31] but its value, which ensures the ergodicity property in an MCMC context, is not relevant in our SMC framework since at each rejuvenation step, the scale parameter c n is fixed for the entire MCMC step. The validity of this adaptation can be theoretically justified by [22].
When the parameter space evolves over time, the MCMC kernel can become model dependent since sampling the state vector using a filtering method is often the most efficient technique in terms of mixing. In special cases where the forward-backward algorithm ([32]) or the Kalman filter ([33]) operate, the state variables can be filtered out. By doing so, we come back to the framework with static parameter space. For non linear state space model, recent works of [13,14] rely on the particle MCMC framework of [34] for integrating out the state vector. We believe that switching from the tempered domain to the time one as well as employing the evolutionary MCMC kernel presented above could even more increase the efficiency of these sophisticated SMC samplers. For example, the particle discrepancies of the early stage inherent to the IBIS algorithm is present in all the empirical simulations of [13] whereas with the TNT sampler, we can ensure a minimum ESS value during the entire procedure.

4. Simulations

We first illustrate the TNT algorithm through a simulation exercise before presenting results on the empirical data. As the TNT algorithm is now completely defined, we start by spelling out the values set for the different parameters to be tuned. The threshold κ is recommended to be high as the evolutionary MCMC updates crucially depend on the diversification of the particles. For that reason we set it to 0.75 M. The second threshold κ 1 that triggers a new run of the simulated annealing algorithm is chosen as 0.1 M and the number of particles is set to M = 2000 . We fix the acceptance rate of the MCMC move to 1/3 and the number of MCMC iterations is set to J = 90 . This number should insure that each particle has moved away from its current position as it approximately implies 30 accepted draws. For all the simulations of the paper, Table 2 summaries these choices.
Our benchmark model for testing the algorithm is a change-point Generalized Autoregressive Conditional Heteroskedastick (CP-GARCH) process that is defined as follows
y t = μ i + ϵ t with ϵ t | y 1 : t - 1 N ( 0 , σ t 2 ) ,
σ t 2 = ω i + α i ϵ t - 1 2 + β i σ t - 1 2 for t τ i - 1 + 1 , τ i and t > 1 ,
where τ 1 = 0 , τ K = T and τ i with i [ 2 , K ] denotes the observation when the break i occurs. The number K of break points are fixed before the estimation and occur sequentially (i.e., τ i - 1 < τ i < τ i + 1   i [ 2 , K - 1 ] ). Stationarity conditions are imposed within each regime by assuming | α i + β i | < 1 . The Table 3 documents the prior distributions of the model parameters.
We innovate by assuming that the regime durations d 1 = τ 1 and d i = τ i - τ i - 1 i [ 2 , K - 1 ] are continuous and are driven by exponential distributions. The duration parameters are therefore identifiable up to a discrete value since they indicate at which observation the process switches from one set of parameters to another. However it brings an obvious advantage as it makes possible to use the Metropolis update developed in Section 3 for the duration parameters too. Consequently, we are able to update in one block all the model parameters. The TNT algorithm of the CP-GARCH models is available on the author’s website.
In this section, we test our algorithm on a simulated series and a financial time series. In addition to that, we found relevant to compare our results with the algorithm of [35] which allows for an online detection of the breaks in the GARCH parameters. A proper comparison of the two approaches is detailed in Appendix C.
We generate 4000 observations from the data generating process (DGP) of Table 4. The DGP exhibits four breaks in the volatility dynamic and tries to mimic the turbulent and quiet periods observed in a financial index. Figure 1 shows a simulated series and the corresponding volatility over time.
We use the marginal log-likelihood (MLL) for selecting the number of regimes by estimating several CP-GARCH models differing by their number of regimes (see [36]). As the TNT algorithm is both an off-line and an on-line method, we start by estimating the posterior distribution with 3000 observations (i.e., τ = 3000 ) and then we add one by one the remaining observations. For each model, we obtain 1001 estimated posterior distributions (from π ( x τ | y 1 : τ ) to π ( x T | y 1 : T ) ) and their respective 1001 MLLs. By so doing, the evolution of the best model over time can be observed. A sharp decrease in the MLL value means that the model cannot easily capture the new observation. According to the DGP Table 4, the model exhibiting three regimes should at least dominate over the first 170 observations and then the model with four regimes should gradually take the lead.
Figure 2 shows the log-Bayes factors (log-BFs) of CP-GARCH models with respect to the standard GARCH process (i.e., K = 1 ).9 The best model over give or take the first 300 observations is the one exhibiting three regimes. Afterward, it is gradually dominated by the process with four regimes (in red). The on-line algorithm has been able to detect the coming break and according to the MLLs, around 150 observations are needed to identify it.
Table 5 documents the posterior means of the parameters of the model exhibiting the highest MLL at the end of the simulation (i.e., with a number of regimes equal to 4) as well as their standard deviations. We observe that the values are close to the true ones which indicates an accurate estimation of the model. The breaks are also precisely inferred. At least for this particular DGP, the TNT algorithm is able to draw the posterior distribution of the CP-GARCH models and correctly updates the distribution in the light of new observations.
Eventually, one can have a look to the varying probabilities associated with each evolutionary update function. These probabilities are computed at each SMC iteration and are proportional to the Mahalanobis distances of the accepted draws. Table 6 documents the values for several SMC iterations. We observe that the probabilities highly vary over the SMC iterations. Moreover, the stretch move and the DREAM algorithm slightly dominate the walk update.

5. Empirical Application

As emphasized in the simulated exercise, the TNT algorithm allows to compare complex models through the marginal likelihoods. We examine the performances of the CP-GARCH models over time on the S&P 500 daily percentage returns spanning from February 08, 1999 to June 24, 2015 (4000 observations). We estimate the models with a number of regimes varying from 1 to 5 using the TNT algorithm and we fix the value τ = 3000 which controls the change from the tempered to the time domain.
To begin with, Table 7 documents the MLLs of the CP-GARCH models with different number of regimes when all the observations have been included. The best model exhibits four regimes.
Table 8 provides the posterior means and the standard deviations of the best CP-GARCH model. Not surprisingly, the break dates occur after the dot-com bubble and at the beginning of the financial crisis. To link the results with the crisis event, Freddie Mac company announced that it will no longer buy the most risky subprime mortgages and mortgage-related securities in 17 February 2007. This date sometimes refers to the beginning of the collapse of the financial system.
We now turn to the recursive estimations of the CP-GARCH models. For the 1001 estimated posterior distributions (from π ( x τ | y 1 : τ ) to π ( x T | y 1 : T ) ), Figure 3 shows the log-BFs of the CP-GARCH models with respect to the fixed parameter GARCH model (i.e., K = 1 ). The CP-GARCH model with four regimes dominates over the entire period. The process with five regimes fits similarly the data but is over-parametrized compared to the same model with four regimes. The difference between the two models comes from the penalization of this over-parametrization through the prior distributions. For interested readers, Appendix B provides additional results on the CP-GARCH models such as the filtered volatility of the preferred process and a detailed comparison with other CP-GARCH models exhibiting breaks in the intercept ω or with student innovations.
To end this study, Table 9 delivers the probabilities associated with each evolutionary update function for several SMC iterations. As in the simulated exercise, the stretch move and the DREAM algorithm slightly dominate the walk one. We also observe that the trigonometric move exhibits good mixing properties since its associated probabilities are high, especially for the DREAM-type update.

6. Conclusions

We develop an off-line and on-line SMC algorithm (called TNT) well-suited for situations where a relevant number of similar distributions has to be estimated. The method encompasses the off-line AIS of [5], the on-line IBIS algorithm of [6] and the RM method of [7] that all arise as special cases in the SMC sampler theory (see [8]). The TNT algorithm benefits from the conjugacy of the tempered and the time domains to avoid particle degeneracies observed in the on-line methods. More importantly, we introduce a new adaptive MCMC kernel based on the evolutionary optimization literature which consists in 10 different moves based on the interactions of the particles. These MCMC updates are selected according to some probabilities that are adjusted over the SMC iterations. Furthermore, the scale parameter of these updates are also automated thanks to the method of [31]. It makes the TNT algorithm fully generic and one needs only to plug the likelihood function, the prior distributions and the number of particles to use it.
The TNT sampler combines on-line and off-line estimations and is consequently suited for comparing complex models. Through a simulated exercise, the paper highlights that the algorithm is able to detect structural breaks of a CP-GARCH model on the fly. Eventually, an empirical application on the S&P 500 daily percentage log-returns shows that no break in the volatility of the GARCH model had arisen from 7 January 2011 to 24 June 2015. In fact, the MLL clearly indicates evidence in favor of a CP-GARCH model exhibiting four regimes in which the breaks occur at the end of the dot-com bubble as well as at the beginning of the financial crisis.
We believe that the TNT algorithm could be adapted to recent SMC algorithms such as [12,13] since they propose advanced SMC samplers based on the IBIS and the E-AIS samplers. Another avenue of research could be an application on change-point stochastic volatility models. Indeed, the evolutionary MCMC kernel is potentially able to update the volatility parameters of these models without filtering them.

Acknowledgments

The author would like to thank Rafael Wouters for his advices on an earlier version of the paper and is grateful to Nicolas Chopin who provided his comments that helped to improve the quality of the paper. He also thanks the referees of the Econometrics journal for their precious comments. Research has been supported by the National Bank of Belgium and by the contract ”Investissement d’Avenir” ANR-11-IDEX-0003/Labex Ecodec/ANR-11-LABX-0047 granted by the Centre de Recherche en Economie et Statistique (CREST). Arnaud Dufays is also a CREST associate Research Fellow. The views expressed in this paper are the author’s ones and do not necessarily reflect those of the National Bank of Belgium. The scientific responsibility is assumed by the author.

Author Contributions

The author contributes entirely to the work presented in this paper.

Conflicts of Interest

The author declares no conflict of interest.

Appendix

A. Proof of Proposition 1

Using the notation x 1 : n 1 : M = { x 1 1 , . . . , x 1 M , x 2 1 , . . . , x n M } which stands for NxM random variables and assuming that x i j E i , j as in the E-AIS method (tempered domain) or the IBIS one (time domain), we consider the augmented posterior distribution :
π ˜ n ( x 1 : n 1 : M ) = [ i = 1 M π n ( x n i ) ] k = 2 n L k ( x k - 1 1 | x k 1 : M ) q = 2 M L k ( x k - 1 q | x k - 1 1 : q - 1 , x k q : M )
If the backward kernels L k ( . | . ) denote proper distributions, the product of the distribution of interest marginally arises:
π ˜ n ( x n 1 : M ) = [ i = 1 M π n ( x n i ) ] k = 2 n L k ( x k - 1 1 | x k 1 : M ) q = 2 M L k ( x k - 1 q | x k - 1 1 : q - 1 , x k q : M ) d x 1 : n - 1 1 : M = [ i = 1 M π n ( x n i ) ]
The SMC sampler with DREAM MCMC kernels leads to a proposal distribution of the form :
η n ( x n 1 : M ) = [ i = 1 M f ( x 1 i ) ] k = 2 n K k ( x k 1 | x k - 1 1 : M ) q = 2 M K k ( x k q | x k 1 : q - 1 , x k - 1 q : M )
where K k ( . | . ) denotes the DREAM subkernel with invariant distribution π k ( . ) . Sampling one draw from this proposal distribution is achieved by firstly drawing M realizations from the prior distribution and then applying the DREAM algorithm (N-1)xM times. As proven in [18,38], the DREAM algorithm leads to the detailed balance equation :
[ i = 1 M π k ( x k - 1 i ) ] K k ( x k 1 | x k - 1 1 : M ) q = 2 M K k ( x k q | x k 1 : q - 1 , x k - 1 q : M ) = [ i = 1 M π k ( x k i ) ] K k ( x k - 1 1 | x k 1 : M ) q = 2 M K k ( x k - 1 q | x k - 1 1 : q - 1 , x k q : M )
Using this relation, we specify the backward kernel as
L k ( x k - 1 1 | x k 1 : M ) q = 2 M L k ( x k - 1 q | x k - 1 1 : q - 1 , x k q : M ) = [ i = 1 M π k ( x k - 1 i ) ] K k ( x k 1 | x k - 1 1 : M ) q = 2 M K k ( x k q | x k 1 : q - 1 , x k - 1 q : M ) [ i = 1 M π k ( x k i ) ]
The sequential importance sampling procedure generates weights given by
w ˜ n ( x 1 : n 1 : M ) π ˜ n ( x 1 : n 1 : M ) η n ( x n 1 : M ) = w ˜ n - 1 ( x 1 : n - 1 1 : M ) [ i = 1 M π n ( x n - 1 i ) ] [ i = 1 M π n - 1 ( x n - 1 i ) ] = i = 1 M w n ( x 1 : n i )
resulting in a product of independent weights exactly equal to the product of SMC sampler weights (see Equation (4)).

B. Additional Estimation Results

From a financial econometric point of view, volatility modeling is very important. In this appendix, we go deeper in our analysis of the CP-GARCH model. To begin with, Appendix B.1 provides more details on the CP-GARCH results obtained in the empirical section. We then propose two extensions of the process by allowing for partial breaks in the intercept in Appendix B.2 and by relaxing the Normal assumption of the innovation (see Appendix B.3).

B.1. Filtered Variance of the CP-GARCH Model

Figure B1 displays the S&P 500 daily log-returns and different quantiles of the filtered variance posterior distribution given by the best model. The volatility tracks very well the magnitude of the returns. Interestingly, since the different quantiles are almost undistinguishable, we conclude that the uncertainty on the filtered variance is very small.
In the results documented in the empirical exercise, the closeness of the last two break dates may appear suspicious. Figure B2 highlights the presence of a huge extreme value occurring in 2 February 2007. The process based on the GARCH parameters derived from the quiet period preceding the financial crisis (from 2003 to 2007) cannot produce such a negative return (that amounts to –3.98). In fact, the parameters have been estimated on a period where the second extreme value reaches 2.1 which is almost half of the outlier magnitude. Consequently, a new regime is created just to handle this high return. In this specific case, the flexibility of the CP-GARCH process turns out to be negative as creating a regime for an outlier is unappealing and counter-productive. To avoid the detection of outliers, one can for instance increase the persistence in the transition probabilities λ since it will penalize very short regimes.

B.2. A CP-GARCH Process with Partial Breaks

Instead of considering breaks in all parameters, one can be interested in a more parsimonious model where only the intercept ω evolves over time. In fact, as the latter parameter is related to the long-run volatility of a regime, the partial-break model can highlight if it is the volatility persistence that is changing over time or only the long-term variance.
We start by a simulation exercise. A series of 4000 observations exhibiting breaks only in the ω parameter has been generated from the DGP documented in Table B1.
We carry out a In-sample simulation by fixing τ = 4000 (no online estimation) and so by estimating only the posterior distribution of the model parameters given all the observations. To begin with, Table B2 delivers the MLLs for the partial CP-GARCH models exhibiting different number of regimes. In this specific case, the criterion selects the true number of regimes.
Table B3 provides summary statistics of the targeted distribution. The break dates are well detected and the highest standard deviation is related to the first break which is in agreement with the DGP. In fact, the difference in the long run variance is the smallest when moving from the first to the second regime. We also observe that the persistence parameters as well as the intercepts are quite accurately estimated.
We now turn to the empirical series and study the S&P 500 daily log returns. As before, we select the best model according to the MLL. Table B4 documents the criterion for the CP-GARCH and the partial CP-GARCH processes given several numbers of breaks. Interestingly, if only the intercept varies over time, the preferred model solely exhibits one break. Additionally, the model where all the parameters vary is preferred which gives evidence in favor of some breaks in the volatility persistence. The break of the partial CP-GARCH model corresponds to the end of the dot com bubble since the posterior mean of the break date is 2 February 2003 with a standard deviation of 103 days.

B.3. A CP-GARCH Process with Student Innovations

As it is well known that financial returns exhibit fat tails, we now study the t-CP-GARCH model, i.e., a CP-GARCH process with t-student innovations. To begin with, a series of 4000 observations from the DGP displayed in Table B5 has been simulated. Figure B3 shows a simulated series and the corresponding conditional volatility over time.10
For each regime i, the degree of freedom d o f i is an additional parameter to be estimated. We constraint the parameter d o f i [ 2 , 100 ] and use the non-linear transformation d o f ˜ i = log ( d o f i - 2 100 - d o f i ) in order to map it on the real line. For each transformed parameter, we choose as a prior distribution a Normal distribution with mean equal to 0 and a variance amounting to 2. As in the previous exercises, we fix τ = 4000 and simulate the final posterior distribution. Table B6 documents the MLLs for the t-CP-GARCH process with different numbers of regimes. The criterion selects the right specification.
Table B7 gives the standard summary statistics of the posterior distribution. We observe that all the parameters are quite accurately estimated. Obviously, as long as the student distribution gets closer to the Normal one, the likelihood function becomes flatter with respect to the degree of freedom. Consequently, it makes difficult to obtain a precise estimation of it.
We now apply the t-CP-GARCH process to the empirical series. Table B8 provides the MLLs for many different numbers of regimes. There is a clear evidence in favour of the t-CP-GARCH process compared to the standard CP-GARCH one. Moreover, the best model does not exhibit any break. It emphasizes that the Normal innovations is clearly rejected and that the breaks were mainly detected to capture extreme values.

C. Comparison with the Online SMC Algorithm

He and Maheu [35] provide an online algorithm able to detect structural breaks on the fly in a process which exhibits the path dependence issue. The paper strongly relies on [39] which develop an auxiliary particle filter (hereafter APF, see [40]) to estimate the static parameters. The sequence of targeted distributions in a standard APF evolves over the time domain and consists in the posterior distributions given a growing number of observations. As a consequence, the number of iteration equals to the sample size.
We propose a formal comparison of the two approaches. For the sake of completeness, the next section briefly reviews the method of [35] and discusses the important differences with the TNT sampler. Eventually, in Section C.2, we apply the algorithm on the two financial time series of the paper to empirically compare the two approaches.

C.1. The APF Algorithm for CP-GARCH Models

Following [35], the CP-GARCH model is specified as follows11
y t = μ s t + ϵ t with ϵ t | y 1 : t - 1 N ( 0 , σ t 2 ) , σ t 2 = ω s t + α s t ϵ t - 1 2 + β s t σ t - 1 2 ,
where s t is a discrete latent variable taking values in [1,K] that is driven by a Markov-chain. The K by K transition matrix is given by
P = p 1 1 - p 1 0 . . . 0 0 p 2 1 - p 2 . . . 0 . . . 0 0 0 0 . . . 1
As already mentioned, the CP-GARCH model exhibits the path dependence issue which makes the standard inference relying on the forward-backward filter not appropriate. To solve the problem, [35] use an APF algorithm (see [40]) in combination with an artificial sequence for the static parameters (see [39]). The static parameters are assumed to evolve over time according to a shrinkage random walk process. As this process requires unbounded support for all the model parameters, they map the bounded parameters on the real line using non-linear transformations. To be precise, if a parameter r belongs to [a,b] (with 0 < a < b ), they apply the mapping r ˜ = log r - a b - r . For notational convenience, we gather all the model parameters into the set θ = { μ 1 , ω 1 , α 1 , β 1 , p 1 , . . . , μ K , ω K , α K , β K , p K } and their associated mapping into θ ˜ . The key idea of [35] consists in including the lagged conditional variances into the set of the SMC particles (and the previous error terms as we have included mean parameters { μ i } i = 1 K in the CP-GARCH specification). Let define the filtration F t - 1 = { θ ˜ t - 1 i , s t - 1 i , σ t - 1 i , ϵ t - 1 i , W t - 1 i } i = 1 N generated by the SMC at time t - 1 . By Bayes’ theorem, we have
π ( θ ˜ t , s t | y 1 : t , F t - 1 ) f ( y t | y 1 : t - 1 , F t - 1 , s t , θ ˜ t ) f ( s t | y 1 : t - 1 , F t - 1 , θ ˜ t ) f ( θ ˜ t | y 1 : t - 1 , F t - 1 )
In Equation (C1), the static parameters θ are now time-varying and the distribution f ( θ ˜ t | y 1 : t - 1 , F t - 1 ) is approximated by a mixture as follows
f ( θ ˜ t | y 1 : t - 1 , F t - 1 ) i = 1 N W t - 1 i N ( θ ˜ μ , t - 1 , ( 1 - a 2 ) V t - 1 ) with θ ˜ μ , t - 1 = a θ ˜ t - 1 + ( 1 - a ) θ ˜ ¯ t - 1
where θ ˜ ¯ t - 1 = i = 1 N W t - 1 i θ ˜ t - 1 i , V t - 1 = i = 1 N W t - 1 i ( θ ˜ t - 1 i - θ ˜ ¯ t - 1 ) ( θ ˜ t - 1 i - θ ˜ ¯ t - 1 ) , a = ( 3 δ - 1 ) / ( 2 δ ) and δ is a discount factor belonging to [ 0 , 1 ] , typically set around 0.95-0.99. Note also that as the filtration includes the previous conditional variances and error terms, the path dependence issue has been solved since the density function f ( y t | y 1 : t - 1 , F t - 1 , s t , θ ˜ t ) is a computable Normal density. The APF algorithm of [35] is briefly detailed in Algorithm 1.
The APF algorithm exposed in Algorithm 1 is distinct in many ways from the TNT sampler. We believe that none of the two approaches dominate and that the algorithmic choice should depend on the needs of the user. In fact, if she is interested in smooth estimates of the states and in an exact posterior simulation (besides the monte carlo error), then she should use the TNT sampler. On the contrary, if she needs a simulation of all the posterior distributions (given only the very first observation to the entire sample) and if she wants a fast estimation, then she should definitely choose the APF approach. Below we discuss in deep details the noticeable differences of the methods.
  • The complexity of the APF is O(NT) which is by far faster than the TNT sampler. However, in order to avoid particles’ discrepancies, the number of particles needs to be very high. He and Maheu [35] recommend N = 300 . 000 .
  • The algorithm of [35] provides an online detection but does not deliver smooth estimates of the states while the TNT sampler gives the smooth probabilities of the states and an online detection when the algorithm switches to the time domain.
  • The APF algorithm relies on an approximation of the distribution f ( θ ˜ t | y 1 : t - 1 , F t - 1 ) and on an artificial process for the static parameters which depends on a user-defined parameter δ. On the contrary, the TNT sampler is not based on any approximations and can directly estimate static parameters.
  • The TNT sampler updates the particles according to an ergodic MCMC kernel which converges to the targeted distribution. By contrast, the APF algorithm is sensible to outliers (as discussed in [40]) and if the number of particles is too low, the APF method may not well approximate the final posterior distribution.
  • The algorithm of [35] exhibits the appealing advantage of estimating the number of regimes. In fact, one can fix the number of regimes K to a very high value and counts the number of detected regimes in the output. Oppositely, we have to fix the number of regimes before estimating the CP-GARCH model. However, it is more a model issue than an algorithmic one as a CP-GARCH model with an undetermined number of regimes (which relies on hierarchical Dirichlet processes, see [41]) can be estimated by the TNT sampler.
  • As mentioned in [35], the Algorithm 1 requires a prior on the transition probability that is close to one for obtaining sensible results since the estimation of the states is sensitive to the outliers. In the same spirit, the other priors need to be informative in order to improve the detection of new regimes (and to limit the number of particles). The TNT sampler allows to freely choose the prior distributions of the model parameters.
  • As long as the probability of being in a state is not equal to zero, the algorithm of [35] can get back to an already visited regime. This feature is highlighted in the simulation exercise below.
  • Since the TNT sampler relies on an MCMC kernel to update the particles, the detections of breaks (exacerbated in the second phase when observations are introduced one by one) highly depend on the mixing of the MCMC kernel and therefore on the number J of MCMC iterations. On the contrary, the online detection of the APF algorithm will depend on the number of particles and how the prior distributions are diffuse. Therefore, for online estimations, the two algorithms exhibit pros and cons which makes their performance depending on the problem at hand.
Algorithm 1 APF algorithm for the CP-GARCH(1,1) model
  • for i = 1 to N do
  •     Sample s 1 i f ( s 1 )
  •     Sample θ ˜ 1 i f ( θ ˜ 1 ) where f ( θ ˜ 1 ) denotes the prior density of θ ˜
  •     Compute the un-normalized weight : w 1 i = f ( y 1 | s 1 i , θ ˜ 1 i )
  •     Save the conditional standard error σ 1 i and the error term ϵ 1 i
  • end for
  • For i=1,...,N, compute normalized auxiliary weights : W 1 i = w 1 i / r = 1 N w 1 r
  • for t = 2 to T do
  •     Prediction step :
  •     for i = 1 to N do
  •         Compute θ ˜ μ , t i and the mode s μ , t i of f ( s t | s t - 1 i , θ ˜ t - 1 i )
  •         Compute the un-normalized auxiliary weight : g t i = f ( y t | y 1 : t - 1 , s μ , t i , θ ˜ μ , t i , σ t - 1 i ) W t - 1 i
  •     end for
  •     For i=1,...,N, compute normalized auxiliary weights : G t i = g t i / r = 1 N g t r
  •     Resample step on the auxiliary weights :
  •     For i=1,...,N, assign a particle to each r i by stratified sampling
  •     (see [42])
  •     Update step :
  •     for i = 1 to N do
  •         Sample θ ˜ t i N ( θ ˜ μ , t - 1 r i , ( 1 - a 2 ) V t - 1 )
  •         Sample s t i f ( s t | s t - 1 r i , θ ˜ t i )
  •         Compute un-normalized weight : w t i = f ( y t | y 1 : t - 1 , s t i , θ ˜ t i , σ t - 1 r i ) f ( y t | y 1 : t - 1 , s μ , t r i , θ ˜ μ , t r i , σ t - 1 r i )
  •         Save the conditional standard error σ t i and the error term ϵ t i
  •     end for
  •     For i=1,...,N, compute normalized auxiliary weights : W t i = w t i / r = 1 N w t r
  • end for

C.2. Comparison of the He and Maheu Algorithm with the TNT Sampler

We estimate the CP-GARCH model with the algorithm of [35]. To begin with, Table C1 summarizes the prior distributions of [35]. In fact, the algorithm requires transition probabilities close to one as well as quite informative priors for the other parameters to correctly perform. Regarding the tuning parameters, we follow the recommendations of [35] and fix δ to 0.99, N to 300 . 000 and the number K of regimes to 6.

C.2.1. Results on the Simulated Series

We start by estimating the model on the same simulated series as in the paper. The DGP is given in Table C2.
As the method cannot provide smooth probabilities of the states, Figure C1 displays the filtered distribution (i.e., f ( s t | y 1 : t ) ). All the breaks have been correctly detected. However, the number of detected regimes only amounts to two. In fact, the probability of being in regime 1 was not zero when the third break has been detected and since the third and the first regimes exhibit similar parameters (see DGP Table C2), the weights of the particles related to the first states were again influential. To compare with the TNT estimation, Figure C2 documents the smooth probabilities (i.e., f ( s t | y 1 : T ) ). The breaks are sharply identified as the estimation uses future observations to confirm the presence of a change.

C.2.2. Results on the S&P 500 Daily Percentage Returns

We now compare the two algorithms on the empirical time series used in the paper. Figure C3 documents the filtered probabilities of the states provided by the He and Maheu’s algorithm. The graphic shows the detection of at least five regimes which means one additional break compared to the preferred model obtained by the TNT sampler. To compare the break detection, Figure C4 documents the smooth probabilities of the TNT sampler for the CP-GARCH model with five regimes (the second best model according to the MLL). The two approaches give very similar results. Interestingly, all the breaks obtained by the TNT sampler are also identified by the APF algorithm. Two differences are worth discussing.
  • The algorithm of [35] detects two breaks during the financial crisis while only one regime covers the period for the TNT sampler. Nevertheless the presence of two regimes in this spell is quite uncertain by judging the probabilities of the third regime (in red) which remain above 20% over the entire period.
  • Compared to the online detection of [35], the uncertainties around the break dates are very small in Figure C4 as the smooth probabilities take the future observations into account.

References

  1. J. Geweke. “Bayesian Inference in Econometric Models Using Monte Carlo Integration.” Econometrica 57 (1989): 1317–1339. [Google Scholar] [CrossRef]
  2. A.F.M. Smith, and A.E. Gelfand. “Bayesian Statistics without Tears: A Sampling-Resampling Perspective.” Am. Stat. 46 (1992): 84–88. [Google Scholar]
  3. N. Gordon, D. Salmond, and A.F.M. Smith. “Novel approach to nonlinear/non-Gaussian Bayesian state estimation.” IEE Proc. F Radar Signal Process. 140 (1993): 107–113. [Google Scholar] [CrossRef]
  4. S. Chib, F. Nardari, and N. Shephard. “Markov chain Monte Carlo methods for stochastic volatility models.” J. Econom. 108 (2002): 281–316. [Google Scholar] [CrossRef]
  5. R.M. Neal. “Annealed Importance Sampling.” Stat. Comput. 11 (1998): 125–139. [Google Scholar] [CrossRef]
  6. N. Chopin. “A Sequential Particle Filter Method for Static Models.” Biometrika 89 (2002): 539–551. [Google Scholar] [CrossRef]
  7. W.R. Gilks, and C. Berzuini. “Following a moving target—Monte Carlo inference for dynamic Bayesian models.” J. R. Stat. Soc. B 63 (2001): 127–146. [Google Scholar] [CrossRef]
  8. P. Del Moral, A. Doucet, and A. Jasra. “Sequential Monte Carlo samplers.” J. R. Stat. Soc. B 68 (2006): 411–436. [Google Scholar] [CrossRef]
  9. A. Jasra, D.A. Stephens, and C.C. Holmes. “On population-based simulation for static inference.” Stat. Comput. 17 (2007): 263–279. [Google Scholar] [CrossRef]
  10. A. Jasra, D.A. Stephens, A. Doucet, and T. Tsagaris. “Inference for Lévy-Driven Stochastic Volatility Models via Adaptive Sequential Monte Carlo.” Scand. J. Stat. 38 (2011): 1–22. [Google Scholar] [CrossRef]
  11. E. Jeremiah, S. Sisson, L. Marshall, R. Mehrotra, and A. Sharma. “Bayesian calibration and uncertainty analysis of hydrological models: A comparison of adaptive Metropolis and sequential Monte Carlo samplers.” Water Resour. Res. 47 (2011). [Google Scholar] [CrossRef]
  12. P. Fearnhead, and B. Taylor. “An adaptive Sequential Monte Carlo Sampler.” Bayesian Anal. 8 (2013): 411–438. [Google Scholar] [CrossRef]
  13. A. Fulop, and J. Li. “Efficient learning via simulation: A marginalized resample-move approach.” J. Econom. 176 (2013): 146–161. [Google Scholar] [CrossRef]
  14. N. Chopin, P.E. Jacob, and O. Papaspiliopoulos. “SMC2: An efficient algorithm for sequential analysis of state space models.” J. R. Stat. Soc. B 75 (2013): 397–426. [Google Scholar] [CrossRef]
  15. G. Durham, and J. Geweke. “Adaptive Sequential Posterior Simulators for Massively Parallel Computing Environments.” In Bayesian Model Comparison (Advances in Econometrics). Edited by I. Jeliazkov and D.J. Poirier. Bingley, UK: Emerald Group Publishing Limited, 2014, Volume 34, pp. 1–44. [Google Scholar]
  16. E. Herbst, and F. Schorfheide. “Sequential Monte Carlo Sampling for DSGE Models.” J. Appl. Econom. 29 (2014): 1073–1098. [Google Scholar] [CrossRef]
  17. L. Bauwens, A. Dufays, and J. Rombouts. “Marginal Likelihood for Markov Switching and Change-point GARCH Models.” J. Econom. 178 (2013): 508–522. [Google Scholar] [CrossRef]
  18. J.A. Vrugt, C.J.F. ter Braak, C.G.H. Diks, B.A. Robinson, J.M. Hyman, and D. Higdon. “Accelerating Markov Chain Monte Carlo Simulation by Differential Evolution with Self-Adaptative Randomized Subspace Sampling.” Int. J. Nonlinear Sci. Numer. Simul. 10 (2009): 271–288. [Google Scholar] [CrossRef]
  19. J.A. Christen, and C. Fox. “A general purpose sampling algorithm for continuous distributions (the t-walk).” Bayesian Anal. 5 (2010): 263–281. [Google Scholar] [CrossRef]
  20. D. Foreman-Mackey, D.W. Hogg, D. Lang, and J. Goodman. “Emcee: The MCMC Hammer.” PASP 125 (2013): 306–312. [Google Scholar] [CrossRef]
  21. P. Del Moral, A. Doucet, and A. Jasra. “On adaptive resampling strategies for sequential Monte Carlo methods.” Bernoulli 18 (2012): 252–278. [Google Scholar] [CrossRef]
  22. A. Beskos, A. Jasra, and A. Thiery. “On the Convergence of Adaptive Sequential Monte Carlo Methods.” 2013. Available online: http://arxiv.org/pdf/1306.6462v2.pdf (accessed on 16th February 2016).
  23. S. Das, and P. Suganthan. “Differential Evolution: A Survey of the State-of-the-Art.” IEEE Trans Evolut. Comput. 15 (2011): 4–31. [Google Scholar] [CrossRef]
  24. R. Storn, and K. Price. “Differential Evolution—A Simple and Efficient Heuristic for Global Optimization over Continuous Spaces.” J. Glob. Optimiz. 11 (1997): 341–359. [Google Scholar] [CrossRef]
  25. G.O. Roberts, and J.S. Rosenthal. “Optimal scaling for various Metropolis-Hastings algorithms.” Stat. Sci. 16 (2001): 351–367. [Google Scholar] [CrossRef]
  26. S. Dutta, and S. Bhattacharya. “Markov chain Monte Carlo based on deterministic transformations.” Stat. Methodol. 16 (2014): 100–116. [Google Scholar] [CrossRef]
  27. H.-Y. Fan, and J. Lampinen. “A Trigonometric Mutation Operation to Differential Evolution.” J. Glob. Optim. 27 (2003): 105–129. [Google Scholar] [CrossRef]
  28. X.-S. Yang. “Firefly Algorithms for Multimodal Optimization.” In Stochastic Algorithms: Foundations and Applications. Edited by O. Watanabe and T. Zeugmann. Berlin, Germany: Springer, 2009, pp. 169–178. [Google Scholar]
  29. C.J. Geyer. “Practical Markov Chain Monte Carlo.” Stat. Sci. 7 (1992): 473–511. [Google Scholar] [CrossRef]
  30. J. Geweke. Contemporary Bayesian Econometrics and Statistics. Wiley Series in Probability and Statistics; Hoboken, NJ, USA: John Wiley and Sons, Inc., 2005. [Google Scholar]
  31. Y. Atchadé, and J. Rosenthal. “On adaptive Markov chain Monte Carlo algorithms.” Bernoulli 11 (2005): 815–828. [Google Scholar] [CrossRef]
  32. L.R. Rabiner. “A tutorial on hidden Markov models and selected applications in speech recognition.” Proc. IEEE 77 (1989): 257–286. [Google Scholar] [CrossRef]
  33. R.E. Kalman. “A New Approach to Linear Filtering and Prediction Problems.” Trans. ASME J. Basic Eng. 82 (1960): 35–45. [Google Scholar] [CrossRef]
  34. C. Andrieu, A. Doucet, and R. Holenstein. “Particle Markov Chain Monte Carlo Methods.” J. R. Stat. Soc. B 72 (2010): 269–342. [Google Scholar] [CrossRef]
  35. Z. He, and J. Maheu. “Real Time Detection of Structural Breaks in GARCH Models.” Comput. Stat. Data Anal. 54 (2010): 2628–2640. [Google Scholar] [CrossRef]
  36. S. Chib. “Estimation and comparison of multiple change-point models.” J. Econom. 86 (1998): 221–241. [Google Scholar] [CrossRef]
  37. R. Kass, and A. Raftery. “Bayes Factors.” J. Am. Stat. Assoc. 90 (1995): 773–795. [Google Scholar] [CrossRef]
  38. L. Bauwens, A. Dufays, and B. De Backer. “Estimating and forecasting structural breaks in financial time series.” J. Empir. Financ., 2011. [Google Scholar] [CrossRef]
  39. J. Liu, and M. West. “Combined Parameter and State Estimation in Simulation-Based Filtering.” In Sequential Monte Carlo Methods in Practice. Edited by A. Doucet, N. de Freitas and N. Gordon. New York, USA: Springer, 2001, pp. 197–224. [Google Scholar]
  40. M.K. Pitt, and N. Shephard. “Filtering via Simulation: Auxiliary Particle Filters.” J. Am. Stat. Assoc. 94 (1999): 590–599. [Google Scholar] [CrossRef]
  41. S. Ko, T. Chong, and P. Ghosh. “Dirichlet Process Hidden Markov Multiple Change-point Model.” Bayesian Anal. 2 (2015): 275–296. [Google Scholar] [CrossRef]
  42. J. Carpenter, P. Clifford, and P. Fearnhead. “Improved particle filter for nonlinear problems.” IEE Proc. Radar Sonar Navig. 146 (1999): 2–7. [Google Scholar] [CrossRef]
  • 1.whereas the sequence does not have to necessarily evolve over the time domain
  • 2.enhanced in the sense that the AIS incorporates a re-sampling step
  • 3.The IBIS algorithm being a particular case
  • 4.Theorem 1 in [6] ensures that with a sufficiently large number of particles M, any relative precision of the importance sampling can be obtained if the number of observations already covered is large enough in the IBIS context
  • 5.a piecewise cooling linear function for [8] and a quadratic function for [16]
  • 6.in the sense of M
  • 7.The best particle is replaced by an average over the particles g = 1 δ x t r 1 ( g ) δ which further diversifies the proposed parameters.
  • 8.Although only the chosen mixture enters in the MH acceptance ratio, the MCMC algorithm is still valid. For further explanations, see [30], section Transition Mixtures.
  • 9.We remind that the log-BF is computed as the difference of the MLL of two models. Following the informal rule of [37], if the logarithm of the Bayes factor exceeds 3, we have strong evidence in favor of the model with the highest value.
  • 10.taking into account the degree of freedom of the student distribution.
  • 11.To compare with the specification of the paper, we have included the mean parameters { μ i } i = 1 K .
Figure 1. Simulated series from the data generating process (DGP) exhibited in Table 4 and its corresponding volatility over time. (a) Simulated series; (b) Volatility over time.
Figure 1. Simulated series from the data generating process (DGP) exhibited in Table 4 and its corresponding volatility over time. (a) Simulated series; (b) Volatility over time.
Econometrics 04 00012 g001
Figure 2. Log-BF over time of the CP-GARCH models in relation to the GARCH one. The log-BF of the CP-GARCH model with two, three, four and five regimes are depicted in yellow, blue, red and cyan respectively. A positive value provides evidence in favor of the considered model compared to the GARCH one.
Figure 2. Log-BF over time of the CP-GARCH models in relation to the GARCH one. The log-BF of the CP-GARCH model with two, three, four and five regimes are depicted in yellow, blue, red and cyan respectively. A positive value provides evidence in favor of the considered model compared to the GARCH one.
Econometrics 04 00012 g002
Figure 3. S&P 500 daily log-returns — log-BFs over time of the volatility models in relation to the GARCH one. The log-BFs of the CP-GARCH models with two, three, four and five regimes are depicted in yellow, blue, red and cyan respectively. A positive value provides evidence in favor of the considered model compared to the GARCH one.
Figure 3. S&P 500 daily log-returns — log-BFs over time of the volatility models in relation to the GARCH one. The log-BFs of the CP-GARCH models with two, three, four and five regimes are depicted in yellow, blue, red and cyan respectively. A positive value provides evidence in favor of the considered model compared to the GARCH one.
Econometrics 04 00012 g003
Figure B1. S&P 500 daily log-returns in relation with the 5% (blue), 50% (red) and the 95% (yellow) quantiles of the filtered volatility posterior distribution provided by the CP-GARCH model with 4 regimes.
Figure B1. S&P 500 daily log-returns in relation with the 5% (blue), 50% (red) and the 95% (yellow) quantiles of the filtered volatility posterior distribution provided by the CP-GARCH model with 4 regimes.
Econometrics 04 00012 g004
Figure B2. S&P 500 daily log-returns over the quiet period preceding the financial crisis in relation with the median of the filtered volatility posterior distribution provided by the CP-GARCH model with four regimes.
Figure B2. S&P 500 daily log-returns over the quiet period preceding the financial crisis in relation with the median of the filtered volatility posterior distribution provided by the CP-GARCH model with four regimes.
Econometrics 04 00012 g005
Figure B3. Simulated series from the DGP Table B5 and its corresponding volatility over time. (a) Simulated series; (b) Volatility over time.
Figure B3. Simulated series from the DGP Table B5 and its corresponding volatility over time. (a) Simulated series; (b) Volatility over time.
Econometrics 04 00012 g006
Figure C1. Simulated series—filtered probabilities of the states given by the He and Maheu’s algorithm. The probabilities of the first state are displayed in blue, the second in green, the third in red and the last three regimes have virtually zero probabilities over the entire sample.
Figure C1. Simulated series—filtered probabilities of the states given by the He and Maheu’s algorithm. The probabilities of the first state are displayed in blue, the second in green, the third in red and the last three regimes have virtually zero probabilities over the entire sample.
Econometrics 04 00012 g007
Figure C2. Simulated series—smooth probabilities of the states given by the TNT algorithm. The probabilities of the first state are displayed in blue, the second in green, the third in red and the last one in turquoise.
Figure C2. Simulated series—smooth probabilities of the states given by the TNT algorithm. The probabilities of the first state are displayed in blue, the second in green, the third in red and the last one in turquoise.
Econometrics 04 00012 g008
Figure C3. S&P 500 daily log-returns—filtered probabilities of the states given by the He and Maheu’s algorithm. The probabilities of the first state are given in blue, the second in green, the third in red and the fourth in turquoise, the fifth in purple and the last one in pale green.
Figure C3. S&P 500 daily log-returns—filtered probabilities of the states given by the He and Maheu’s algorithm. The probabilities of the first state are given in blue, the second in green, the third in red and the fourth in turquoise, the fifth in purple and the last one in pale green.
Econometrics 04 00012 g009
Figure C4. S&P 500 daily log-returns - smooth probabilities of the states given by the TNT algorithm for the CP-GARCH model exhibiting five regimes. The probabilities of the first state are displayed in blue, the second in green, the third in red, the fourth in turquoise and the fifth in purple.
Figure C4. S&P 500 daily log-returns - smooth probabilities of the states given by the TNT algorithm for the CP-GARCH model exhibiting five regimes. The probabilities of the first state are displayed in blue, the second in green, the third in red, the fourth in turquoise and the fifth in purple.
Econometrics 04 00012 g010
Table 1. Average of the autocorrelation times over the five dimensions for multiple update moves and different distributions.
Table 1. Average of the autocorrelation times over the five dimensions for multiple update moves and different distributions.
Stretch MoveWalk MoveDREAM
Move5-Dimension Normal Distribution
with correlation of 0.5 and variances set to unity
Standard84.99106.1913.79
Trigo56.6258.5920.36
FF52.2151.75
DE44.8566.81
5-Dimension Normal Distribution
with correlation of 0.999 and variances set to unity
Standard92.9496.6334.93
Trigo63.1482.8323.91
FF59.3138.54
DE70.0761.45
5-Dimension Student Distribution
with correlation of 0.999, df = 5 and variances set to unity
Standard104.5975.9123.11
Trigo54.6665.3819.83
FF38.1735.21
DE37.0135.53
Table 2. Tuned parameters for the TNT algorithm.
Table 2. Tuned parameters for the TNT algorithm.
ParametersTNT Algorithm
Nb. Particles M 2 . 000
Threshold ESS κ 0 . 75 M
Sec. threshold ESS κ 1 0 . 1 M
Acc. rate α targeted 1 / 3
Nb. MCMC J90
Table 3. Prior distributions of the CP parameters. The distribution N ( a , b ) denotes the Normal distribution with expectation a and variance b and U[a,b] stands for the uniform distribution with lower bound a and upper bound b. The exponential distribution with parameter λ is expressed as Exp(λ)(with density function : f ( τ | λ ) = λ e - λ τ ) and the gamma distribution is denoted by Gamma ( a , b ) in which a is the shape parameter and b the scale one (with density function f ( λ | a , b ) = b a Γ ( a ) λ a - 1 e - b λ ).
Table 3. Prior distributions of the CP parameters. The distribution N ( a , b ) denotes the Normal distribution with expectation a and variance b and U[a,b] stands for the uniform distribution with lower bound a and upper bound b. The exponential distribution with parameter λ is expressed as Exp(λ)(with density function : f ( τ | λ ) = λ e - λ τ ) and the gamma distribution is denoted by Gamma ( a , b ) in which a is the shape parameter and b the scale one (with density function f ( λ | a , b ) = b a Γ ( a ) λ a - 1 e - b λ ).
Mean Parameter i [ 1 , K ] :
μ i ∼ N(0,1)
GARCH Parameters i [ 1 , K ] :
ω i U [ 0 , 1 ] α i | β i U [ 0 , 1 - β i ] β i U [ 0 . 2 , 1 ]
Break Parameters i [ 2 , K - 1 ] :
d 1 = τ 1 ∼ Exp ( λ ) d i = τ i - τ i - 1 ∼ Exp ( λ ) λ∼ Gamma ( 1 , T )
Table 4. Data generating process of the Change-point-Generalized Autoregressive Conditional Heteroskedastic (CP-GARCH) model.
Table 4. Data generating process of the Change-point-Generalized Autoregressive Conditional Heteroskedastic (CP-GARCH) model.
μωαβτ
Regime 100.10.10.851250
Regime 200.30.030.952230
Regime 300.250.20.703170
Regime 400.40.050.9
Table 5. Posterior means of the parameters of the CP-GARCH model with four regimes and their corresponding standard deviations.
Table 5. Posterior means of the parameters of the CP-GARCH model with four regimes and their corresponding standard deviations.
μωαβτ
Regime 1–0.020.110.110.851253.7
(0.04)(0.04)(0.03)(0.04)(9.91)
Regime 20.150.670.050.912238.4
(0.13)(0.21)(0.02)(0.03)(17.42)
Regime 30.010.250.160.733169.4
(0.04)(0.08)(0.03)(0.05)(11.33)
Regime 40.060.610.070.85
(0.1)(0.22)(0.02)(0.04)
Table 6. Probabilities (proportional to the Mahalabonis distance) of choosing a specific type of Metropolis-Hastings move. Mean stands for the average over the SMC iterations.
Table 6. Probabilities (proportional to the Mahalabonis distance) of choosing a specific type of Metropolis-Hastings move. Mean stands for the average over the SMC iterations.
SMC IterationStretch MoveWalk MoveDREAM Move
TrigoDEFireflyStandardTrigoDEFireflyStandardTrigoStandard
1th0.10.10.10.10.10.10.10.10.10.1
10th0.130.090.080.070.100.060.070.080.220.11
100th0.140.110.130.130.110.040.060.160.090.04
last0.110.120.110.110.080.050.060.100.170.09
Mean0.130.120.120.110.100.050.060.140.120.06
Table 7. S&P 500 daily log-returns—marginal log-likelihoods (MLLs) of the CP-GARCH models given different number of regimes. The highest value is bolded.
Table 7. S&P 500 daily log-returns—marginal log-likelihoods (MLLs) of the CP-GARCH models given different number of regimes. The highest value is bolded.
#Regimes12345
MLL–5732.6–5730.86–5731.1–5727.87–5729.1
Table 8. S&P 500 daily log-returns - posterior means of the parameters of the CP-GARCH model with four regimes and their corresponding standard deviations.
Table 8. S&P 500 daily log-returns - posterior means of the parameters of the CP-GARCH model with four regimes and their corresponding standard deviations.
μωαβτ
Regime 1–0.010.160.110.8125 March 2003
(0.05)(0.06)(0.03)(0.05)(74.22)
Regime 20.060.020.050.9114 February 2007
(0.02)(0.03)(0.02)(0.05)(62.33)
Regime 3–0.330.660.130.5209 March 2007
(0.34)(0.22)(0.15)(0.20)(32.63)
Regime 40.080.020.120.86
(0.02)(0.01)(0.01)(0.01)
Table 9. S&P 500 daily log-returns—probabilities (proportional to the Mahalabonis distance) of choosing a specific type of Metropolis-Hastings move. Mean stands for the average over all the Sequential Monte Carlo (SMC) iterations. The highest probability is bolded.
Table 9. S&P 500 daily log-returns—probabilities (proportional to the Mahalabonis distance) of choosing a specific type of Metropolis-Hastings move. Mean stands for the average over all the Sequential Monte Carlo (SMC) iterations. The highest probability is bolded.
SMC IterationStretch MoveWalk MoveDREAM Move
TrigoDEFireflyStandardTrigoDEFireflyStandardTrigoStandard
1th0.10.10.10.10.10.10.10.10.10.1
10th0.120.090.080.060.100.060.070.070.240.12
100th0.130.130.120.120.080.050.050.110.150.07
last0.120.120.120.110.080.050.060.100.170.09
Mean0.120.120.110.110.090.050.050.100.160.08
Table B1. Data generating process of the CP-GARCH model exhibiting breaks in ω. The long run variance of each regime is given in the row ω 1 - α - β .
Table B1. Data generating process of the CP-GARCH model exhibiting breaks in ω. The long run variance of each regime is given in the row ω 1 - α - β .
μωαβτ ω 1 - α - β
Regime 100.10.10.8512102
Regime 200.320606
Regime 300.0530301
Regime 400.48
Table B2. Simulated series from the DGP Table B1—MLLs of the partial CP-GARCH models given different number of regimes. The highest value is bolded.
Table B2. Simulated series from the DGP Table B1—MLLs of the partial CP-GARCH models given different number of regimes. The highest value is bolded.
# Regimes12345
MLL–7746.57–7748.83–7749.35–7738.38–7738.68
Table B3. Simulated series from the DGP Table B1—posterior means of the parameters of the partial CP-GARCH model with four regimes and their corresponding standard deviations.
Table B3. Simulated series from the DGP Table B1—posterior means of the parameters of the partial CP-GARCH model with four regimes and their corresponding standard deviations.
μωαβτ
Regime 10.030.160.110.811197.78
(0.04)(0.02)(0.02)(0.04)(58.29)
Regime 2–0.050.522062.17
(0.09)(0.17)(22.83)
Regime 30.030.083033.17
(0.04)(0.02)(22.55)
Regime 40.040.7
(0.1)(0.20)
Table B4. S&P 500 daily log-returns — MLLs of the CP-GARCH models given different number of regimes for two distinct specifications. The partial CP-GARCH model denotes a CP-GARCH process where only the intercept is allowed to exhibit breaks. The standard CP-GARCH model exhibits breaks in all its parameters. The highest value is bolded.
Table B4. S&P 500 daily log-returns — MLLs of the CP-GARCH models given different number of regimes for two distinct specifications. The partial CP-GARCH model denotes a CP-GARCH process where only the intercept is allowed to exhibit breaks. The standard CP-GARCH model exhibits breaks in all its parameters. The highest value is bolded.
# Regimes12345
Partial CP-GARCH Model
MLL–5733.71–5732.81–5737.97–5735.85–5734.47
Standard CP-GARCH Model
MLL–5732.6–5730.86–5731.1–5727.87–5729.1
Table B5. Data generating process of the t-CP-GARCH model. The degree of freedom is denoted by dof.
Table B5. Data generating process of the t-CP-GARCH model. The degree of freedom is denoted by dof.
μωαβτ
Regime 100.10.10.85151080
Regime 200.30.030.95402390
Regime 300.250.20.7053280
Regime 400.40.050.910
Table B6. Simulated series from the DGP Table B5—MLLs of the t-CP-GARCH models given different number of regimes. The highest value is bolded.
Table B6. Simulated series from the DGP Table B5—MLLs of the t-CP-GARCH models given different number of regimes. The highest value is bolded.
# Regimes12345
MLL–10132.12–10124.63–10106.30–10095.56–10098.63
Table B7. Simulated series from the DGP Table B5—posterior means of the parameters of the t-CP-GARCH model with four regimes and their corresponding standard deviations.
Table B7. Simulated series from the DGP Table B5—posterior means of the parameters of the t-CP-GARCH model with four regimes and their corresponding standard deviations.
μωαβτdof
Regime 10.030.160.160.81094.7437.25
(0.05)(0.05)(0.03)(0.03)(31.65)(22.39)
Regime 20.060.270.040.942373.951.42
(0.11)(0.12)(0.01)(0.01)(14.1)(22.74)
Regime 30.170.250.20.693306.045.14
(0.07)(0.08)(0.03)(0.04)(22.71)(0.89)
Regime 40.130.620.030.91 42.78
(0.13)(0.28)(0.02)(0.04) (24.12)
Table B8. S&P 500 daily log-returns—MLLs of the CP-GARCH models given different number of regimes for two distinct specifications. The t-CP-GARCH model denotes a CP-GARCH process with t-student innovations. The innovations of the standard CP-GARCH model are driven by a Normal distribution. The highest value is bolded.
Table B8. S&P 500 daily log-returns—MLLs of the CP-GARCH models given different number of regimes for two distinct specifications. The t-CP-GARCH model denotes a CP-GARCH process with t-student innovations. The innovations of the standard CP-GARCH model are driven by a Normal distribution. The highest value is bolded.
# Regimes12345
t-CP-GARCH model
MLL–5673.98–5675.91–5678.11–5682.85–5686.78
Standard CP-GARCH model
MLL–5732.6–5730.86–5731.1–5727.87–5729.1
Table C1. Prior distributions of the CP-GARCH parameters. The distribution N ( a , b ) denotes the Normal distribution with expectation a and variance b and Gam(a,b) stands for the gamma distribution with a being the shape parameter and b the scale one. The beta distribution is expressed as Beta(a,b) with shape parameters a and b.
Table C1. Prior distributions of the CP-GARCH parameters. The distribution N ( a , b ) denotes the Normal distribution with expectation a and variance b and Gam(a,b) stands for the gamma distribution with a being the shape parameter and b the scale one. The beta distribution is expressed as Beta(a,b) with shape parameters a and b.
Mean parameter i [ 1 , K ] :
μ i ∼ N(0,0.1)
GARCH parameters i [ 1 , K ] :
ω i G a m [ 1 , 0 . 2 ] α i B e t a [ 1 , 8 ] β i B e t a [ 4 , 1 ]
Transition parameters i [ 1 , K - 1 ] :
p ˜ i N ( 10 , 1 )
Table C2. Data generating process of the CP-GARCH model.
Table C2. Data generating process of the CP-GARCH model.
μωαβτ
Regime 100.10.10.851250
Regime 200.30.030.952230
Regime 300.250.20.703170
Regime 400.40.050.9

Share and Cite

MDPI and ACS Style

Dufays, A. Evolutionary Sequential Monte Carlo Samplers for Change-Point Models. Econometrics 2016, 4, 12. https://doi.org/10.3390/econometrics4010012

AMA Style

Dufays A. Evolutionary Sequential Monte Carlo Samplers for Change-Point Models. Econometrics. 2016; 4(1):12. https://doi.org/10.3390/econometrics4010012

Chicago/Turabian Style

Dufays, Arnaud. 2016. "Evolutionary Sequential Monte Carlo Samplers for Change-Point Models" Econometrics 4, no. 1: 12. https://doi.org/10.3390/econometrics4010012

APA Style

Dufays, A. (2016). Evolutionary Sequential Monte Carlo Samplers for Change-Point Models. Econometrics, 4(1), 12. https://doi.org/10.3390/econometrics4010012

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop