variance clustering improved dynamic conditional correlation mgarch estimators

21
Computational Statistics and Data Analysis ( ) Contents lists available at SciVerse ScienceDirect Computational Statistics and Data Analysis journal homepage: www.elsevier.com/locate/csda Variance clustering improved dynamic conditional correlation MGARCH estimators Gian Piero Aielli, Massimiliano Caporin Department of Economics and Management Marco Fanno, University of Padova, Via del Santo 33, Padova, Italy article info Article history: Received 4 May 2012 Received in revised form 30 January 2013 Accepted 30 January 2013 Available online xxxx Keywords: Dynamic conditional correlations Time series clustering Multivariate GARCH Composite likelihood abstract It is well-known that the estimated GARCH dynamics exhibit common patterns. Starting from this fact the Dynamic Conditional Correlation (DCC) model is extended by allowing for a clustering structure of the univariate GARCH parameters. The model can be estimated in two steps, the first devoted to the clustering structure, and the second focusing on dynamic parameters. Differently from the traditional two-step DCC estimation, large system feasibility of the joint estimation of the whole set of model’s dynamic parameters is achieved. A new approach to the clustering of GARCH processes is also introduced. Such an approach embeds the asymptotic properties of the univariate quasi-maximum- likelihood GARCH estimators into a Gaussian mixture clustering algorithm. Unlike other GARCH clustering techniques, the proposed method provides a natural estimator of the number of clusters. © 2013 Published by Elsevier B.V. 1. Introduction The Dynamic Conditional Correlation GARCH (DCC) model (Engle, 2002) has recently become one of the most popular tools for the estimation of multivariate asset volatility dynamics (for applications see, e.g., Cappiello et al. (2006), Billio et al. (2006), Billio and Caporin (2009), Franses and Hafner (2009) and Pesaran and Pesaran (2007); some theoretical results are in Engle and Sheppard (2001), McAleer et al. (2008) and Aielli (forthcoming)). The basic idea of the DCC modelling approach is to obtain first the variance process via simple univariate specifications, and, then to build a model for the correlation process from some appropriate function of the univariate variance standardized returns. As a direct advantage of such a modelling strategy, we obtain a two-step estimation procedure that is feasible with large systems: in the first step, the univariate variance processes are estimated one at a time, and, then, in the second step, the correlation process is estimated from the estimated standardized returns provided by the first step. Under appropriate specifications of the correlation process such as the cDCC specification adopted in this paper (Aielli, forthcoming), the two-step estimation procedure is shown to have desirable theoretical and empirical properties. Most proposed extensions of the DCC model have been developed with the aim of providing more flexibility to the model of the correlation process. Two examples are the Asymmetric DCC model by Cappiello et al. (2006), and the Flexible DCC model by Billio et al. (2006); see also Billio and Caporin (2009), Engle and Kelly (2012) and Franses and Hafner (2009) for other applications. In particular, with the Flexible DCC model a rather rich parametrization of the correlation process is obtained by relying on a priori knowledge of the presence of some asset partition. The focus of this paper is different: we aim at improving the flexibility of the DCC modelling approach by focusing attention on the ‘‘variance side’’, rather than on the ‘‘correlation side’’, of the DCC model. Our final interest is to avoid, or reduce, the loss of efficiency (if any) that An appendix containing additional tables and figures can be found as a supplementary material of the electronic version of the paper. Corresponding author. Tel.: +39 049 8274258; fax: +39 049 8274211. E-mail addresses: [email protected], [email protected] (M. Caporin). 0167-9473/$ – see front matter © 2013 Published by Elsevier B.V. doi:10.1016/j.csda.2013.01.029

Upload: massimiliano

Post on 01-Jan-2017

214 views

Category:

Documents


2 download

TRANSCRIPT

Page 1: Variance clustering improved dynamic conditional correlation MGARCH estimators

Computational Statistics and Data Analysis ( ) –

Contents lists available at SciVerse ScienceDirect

Computational Statistics and Data Analysis

journal homepage: www.elsevier.com/locate/csda

Variance clustering improved dynamic conditionalcorrelation MGARCH estimators✩

Gian Piero Aielli, Massimiliano Caporin ∗

Department of Economics and Management Marco Fanno, University of Padova, Via del Santo 33, Padova, Italy

a r t i c l e i n f o

Article history:Received 4 May 2012Received in revised form 30 January 2013Accepted 30 January 2013Available online xxxx

Keywords:Dynamic conditional correlationsTime series clusteringMultivariate GARCHComposite likelihood

a b s t r a c t

It is well-known that the estimated GARCH dynamics exhibit common patterns. Startingfrom this fact the Dynamic Conditional Correlation (DCC) model is extended by allowingfor a clustering structure of the univariate GARCH parameters. Themodel can be estimatedin two steps, the first devoted to the clustering structure, and the second focusing ondynamic parameters. Differently from the traditional two-step DCC estimation, largesystem feasibility of the joint estimation of the whole set of model’s dynamic parametersis achieved. A new approach to the clustering of GARCH processes is also introduced.Such an approach embeds the asymptotic properties of the univariate quasi-maximum-likelihood GARCH estimators into a Gaussian mixture clustering algorithm. Unlike otherGARCH clustering techniques, the proposed method provides a natural estimator of thenumber of clusters.

© 2013 Published by Elsevier B.V.

1. Introduction

The Dynamic Conditional Correlation GARCH (DCC) model (Engle, 2002) has recently become one of the most populartools for the estimation of multivariate asset volatility dynamics (for applications see, e.g., Cappiello et al. (2006), Billio et al.(2006), Billio and Caporin (2009), Franses and Hafner (2009) and Pesaran and Pesaran (2007); some theoretical results are inEngle and Sheppard (2001), McAleer et al. (2008) and Aielli (forthcoming)). The basic idea of the DCCmodelling approach isto obtain first the variance process via simple univariate specifications, and, then to build amodel for the correlation processfrom some appropriate function of the univariate variance standardized returns. As a direct advantage of such a modellingstrategy, we obtain a two-step estimation procedure that is feasible with large systems: in the first step, the univariatevariance processes are estimated one at a time, and, then, in the second step, the correlation process is estimated from theestimated standardized returns provided by the first step. Under appropriate specifications of the correlation process suchas the cDCC specification adopted in this paper (Aielli, forthcoming), the two-step estimation procedure is shown to havedesirable theoretical and empirical properties.

Most proposed extensions of the DCCmodel have been developedwith the aim of providingmore flexibility to themodelof the correlation process. Two examples are the Asymmetric DCC model by Cappiello et al. (2006), and the Flexible DCCmodel by Billio et al. (2006); see also Billio and Caporin (2009), Engle and Kelly (2012) and Franses and Hafner (2009)for other applications. In particular, with the Flexible DCC model a rather rich parametrization of the correlation processis obtained by relying on a priori knowledge of the presence of some asset partition. The focus of this paper is different:we aim at improving the flexibility of the DCC modelling approach by focusing attention on the ‘‘variance side’’, ratherthan on the ‘‘correlation side’’, of the DCC model. Our final interest is to avoid, or reduce, the loss of efficiency (if any) that

✩ An appendix containing additional tables and figures can be found as a supplementary material of the electronic version of the paper.∗ Corresponding author. Tel.: +39 049 8274258; fax: +39 049 8274211.

E-mail addresses:[email protected], [email protected] (M. Caporin).

0167-9473/$ – see front matter© 2013 Published by Elsevier B.V.doi:10.1016/j.csda.2013.01.029

Page 2: Variance clustering improved dynamic conditional correlation MGARCH estimators

2 G.P. Aielli, M. Caporin / Computational Statistics and Data Analysis ( ) –

potentially affects the traditional two-step DCC estimator; in fact, in that case, the variance processes are estimated one ata time and independently from the information embedded in the correlation dynamics. With this aim in mind, we take ourcue from the well-known fact that the univariate estimates of GARCH dynamic parameters are often similar. We then allowthe specification of the cDCC model to take advantage of this empirical finding by assuming that there can be assets in theselected asset sample that share commonGARCHdynamics.We call this assumption variance clustering (VC) assumption, andthe resultingmodel, VC-cDCCmodel. If the number of clusters is equal to the number of assets, we obtained the cDCCmodelas a special case of the VC-cDCC model. Given an estimate of the underlying asset variance clustering structure, the jointestimation of the whole set of model dynamic parameters – including the variance dynamic parameters and the correlationdynamic parameters – becomes possible.

Under a fully Bayesian framework, Bauwens and Rombouts (2007) recently addressed the estimation of the underlyingGARCH dynamic clustering structure. The refined clustering approach proposed by the authors makes heavy use of MonteCarlo Markov Chain (MCMC) methods. As a result, their approach is computationally intensive, in particular with a largenumber of assets. A mixed frequentist-Bayesian approach is proposed by Brownlees (2009), in which the parameters ofa set of GARCH processes are modelled as functions of observed regressors and unobserved idiosyncratic shocks. On thecontrary, in our paper we adopt a fully frequentist approach, with the explicit aim of providing a fast and friendly clusteringmethodology, in accordance with the motivations behind the DCC estimation paradigm, as originally proposed by Engle(2002). Our clustering algorithm constitutes a new contribution to the literature on the clustering of GARCH processes, or,in more general terms, on the clustering of financial time series (see Liao (2005) for a survey on time series clustering). Weknow from Francq and Zakoïan (2010) and therein cited references that the asymptotic distribution of the univariate Quasi-maximum Likelihood (QML) GARCH estimator is Gaussian. Therefore, a possible choice is to consider a Gaussian mixture(GM) clustering framework. Unlike Bauwens and Rombouts (2007), who specify a GM model for the observed series as aconsequence of the assumption of Gaussian conditionally distributed returns, our GM model is specific to the sample ofthe univariate QML estimates of the GARCH dynamic parameters. We thus change the focus of the GM clustering from thereturn process to the GARCH estimators. As a result, our approach is more flexible since we are working on a distribution-free GARCH framework where asymptotic properties of the QML univariate GARCH estimators are valid anyway. We alsopoint out that, to estimate the peculiar mixture component covariance matrix implied by our GM model specification, weresort to an appropriate combination of standard univariate QML estimation outputs. Furthermore – and differently fromother ad hoc, non-Bayesian GARCH clustering techniques (e.g., Otranto (2008, 2010)) – our clustering methodology logicallyleads to a BIC-based selection procedure for the identification of the number of clusters.

In order to deal with many assets, however, the reduction of the parameter dimensionality allowed by the varianceclustering structure generally is not enough to allow the joint estimation of the variance/correlation dynamic parameters.The infeasibility is due to the fact that the dimensionality of the variance/correlation intercept parameters is O(N2),where N is the number of series. Following Aielli (forthcoming), we successfully solve this problem by resorting toan ad hoc generalized profile estimator (Severini, 1998). We thus treat the variance/correlation intercept parametersand the clustering structure as nuisance parameters, and the variance/correlation dynamic parameters as parametersof interest. We then replace the nuisance parameters in the model pseudo-likelihood with an estimator that can beeasily computed conditionally on the parameters of interest. Thanks to such an estimation device, the joint estimationof the variance/correlation dynamic parameters becomes feasible (we call it the joint VC-cDCC estimator). The joint VC-cDCC estimator is shown to be superior to the traditional two-step cDCC estimator on the basis of both simulations andapplications to real data.

For the joint VC-cDCC estimator to be feasible, it is required that the number of clusters is small, which, fortunately, isprecisely what happens when dealing with large systems. An estimator of the VC-cDCC model that is feasible irrespectiveof the number of clusters is, however, also suggested in this paper (the sequential VC-cDCC estimator). With the sequentialestimator, the asset variance dynamics are still estimated jointly, but separately from the correlation process. As objectivefunctions for the variance dynamic parameters we adopt univariate composite likelihoods (Pakel et al., 2011). A specialcase of the sequential estimator is the traditional two-step cDCC estimator. The sequential estimator is extremely rapidto compute, and, surprisingly, does not cause any sensible loss of efficiency with respect to the joint estimator. From atheoretical perspective, this finding seems to suggest that the dynamic variance parameter and the dynamic correlationparameter are practically orthogonal (with respect to the adopted pseudo-likelihood).

The rest of the paper is organized as follows: Section 2 introduces the VC-cDCC model; Section 3 describes the VC-cDCCestimatorwhile Section 4 contains theMonte Carlo simulations; Section 5 reports some applications to real data, and, finally,Section 6 concludes.

2. The VC-cDCC model

Before introducing the VC-cDCCmodel we review the DCCmodelling approach. Denote by yt = [y1,t , . . . , yN,t ]′ theN×1vector of the asset returns at time t , and assume that

Et−1[yt ] = 0, Et−1[yty′

t ] = Ht ,

where Et [·] is the conditional expectation on yt , yt−1, . . . . The asset conditional covariance matrix can be written as

Ht = D1/2t RtD

1/2t , (1)

Page 3: Variance clustering improved dynamic conditional correlation MGARCH estimators

G.P. Aielli, M. Caporin / Computational Statistics and Data Analysis ( ) – 3

where Rt = [ρij,t ] is the asset conditional correlation matrix, and Dt = diag(h1,t , . . . , hN,t) is the diagonal matrix of theasset conditional variances. By construction, Rt is the conditional covariance matrix of the asset standardized returns, thatis, Et−1[εtε

′t ] = Rt , where εt =

ε1,t , . . . , εN,t

′, and εi,t = yi,t/

hi,t . Engle (2002), following Bollerlsev (1990), suggests

modelling the right hand side of (1) rather than Ht directly. The resulting model is called the DCC model. In this work, weadopt the cDCC specification proposed by Aielli (forthcoming), and then later used in empirical applications by Engle et al.(2009), Engle and Kelly (2012), and Hafner and Reznikova (2012).

2.1. The cDCC model

The cDCC model assumes that the elements of Dt follow GARCH(P,Q ) processes, Bollerlsev (1986),

hi,t = wi +

Pp=1

ai,py2i,t−p +

Qq=1

bi,qhi,t−q, (2)

where the GARCH parameters, wi, ai,p, and bi,q, are non-negative to ensure positivity of hi,t . We further assume weaklystationarity of yi,t , that is, the parameters satisfy

Pp=1 ai,p +

Qq=1 bi,q < 1. We note that, under stationarity, wi can be

reparametrized as

wi =

1 −

Pp=1

ai,p −

Qq=1

bi,q

τi, τi ≥ 0, (3)

where it can be easily proved that

τi = E[y2i,t ], (4)(τi is the stationary second order moment of yi,t ). We then define the cDCC correlation process as

ρij,t =(1 − αij − βij)sij,t−1 + αijεi,t−1εj,t−1 + βijρij,t−1

{(1 − αii − βii)sii,t−1 + αiiε2i,t−1 + βiiρii,t−1}{(1 − αjj − βjj)sjj,t−1 + αjjε

2j,t−1 + βjjρjj,t−1}

, (5)

wheresij,t = sij/

√qi,tqj,t , (6)

and

qi,t = (1 − αii − βii)+ αiiε2i,t−1qi,t−1 + βiiqi,t−1. (7)

We collect the parameters of the correlation process in the matrices A = [αij], B = [βij] and S = [sij]. The representation in(5) of ρij,t highlights that the cDCCmodel combines into a correlation-like ratio, a sort of GARCH device for both the relevantinnovations and the past correlations (the numerator and the terms in braces at the denominators of Eq. (5), respectively).We also note that the denominator of sij,t is an ad hoc correction required for tractability. Compared with the traditional –and equivalent – representation of ρij,t in terms of rescaled elements of the correlation driving process, Qt (see Engle (2002)and Aielli (forthcoming)), the representation of ρij,t provided here ismore immediate and clearly highlights the constructionprocess of the correlation dynamic. For Rt to be a proper correlation matrix we must impose that A, B and (ιι′ − A − B)⊙ Sare positive semi-definite, where ⊙ denotes the element-wise (Hadamard) matrix product, and ι denotes the N × 1 unitvector. To achieve identification, S must have unit diagonal elements sii = 1, i = 1, 2, . . . ,N . If αij = α and βij = β fori ≤ j = 1, 2, . . . ,N , the model is called Scalar cDCC. In the last case, Rt is a proper correlation matrix provided that S ispositive semi-definite, α ≥ 0, β ≥ 0 and α + β ≤ 1. Setting α = β = 0 yields the Constant Conditional Correlation (CCC)model of Bollerlsev (1990), where Rt = S at each point in time.

It can be shown that the cDCC model satisfiesS = E[{Q ∗

t εt}{Q∗

t εt}′], (8)

where Q ∗t = diag(

√q1,t , . . . ,

√qN,t). This property plays a crucial role in the construction of reasonable large system

estimators (see Section 3.2). For further details on the cDCC model properties, see Aielli (forthcoming).

2.2. The VC-cDCC model

As we state in the introduction, we assume that the N assets are grouped into K clusters. Assets within a given clustershare a common GARCH dynamic, whereas assets belonging to different clusters can exhibit different GARCH dynamics.Our purpose is to combine the cDCC dynamic with a restriction on the possible GARCH dynamics followed by the N assets.We stress that our modelling approach is different from several contributions of the MGARCH literature that focused oncorrelation dynamic parameters. If we introduce variance clustering into the cDCCmodel we obtain the Variance ClusteringcDCC model (VC-cDCC). Denote as Ci = [ai1, ai2, . . . , aiP , bi1, bi2, . . . , biQ ]

′, the (P + Q ) × 1 vector collecting the GARCHdynamic parameters of the i-th asset. We assume that, for any pair (Ci, Cj),

Ci = Cj if πi = πj, (9)

Page 4: Variance clustering improved dynamic conditional correlation MGARCH estimators

4 G.P. Aielli, M. Caporin / Computational Statistics and Data Analysis ( ) –

Fig. 1. Underlying variance clustering structure. On the left: plot of estimates of univariate GARCH(1, 1) dynamic parameters from N = 100 series ofreal data. The dataset is the equity dataset described in Section 5.1. On the right: plot of estimates of univariate GARCH(1, 1) dynamic parameters fromsimulated data. The DGP underlying the simulated series provides for K = 4 distinct GARCH dynamics, namely, µ1 = (0.13, 0.56) for N1 = 13 assets,µ2 = (0.11, 0.73) for N2 = 39 assets,µ3 = (0.03, 0.96) for N3 = 42 assets, andµ4 = (0.14, 0) for N4 = 6 assets. The simulated correlation process is setas constant and equal to sij = s = 0.3 for all asset pairs.

where πi ∈ {1, 2, . . . , k, . . . , K}, i = 1, 2, . . . , k, . . . ,N . The partition vector, π = [π1, π2, . . . , πN ]′, arranges the N

assets into the K clusters. The k-th cluster is defined as the subset of assets {yit : πi = k}. The size of the k-th cluster isNk =

Ni=1 I(πi = k),where I(·) is the indicator function. It holds that

Kk=1 Nk = N .We denote byµk the vector of GARCH

dynamic parameters common to the assets included in the k-th cluster. By definition, it holds that Ci = µπi , i = 1, 2, . . . ,N .The set of VC-cDCC GARCH dynamic parameters is thus given by µi, i = 1, 2, . . . , K . If πi = πj for i < j = 2, . . . ,N , that is,no common GARCH dynamics, we get the cDCC model, which is therefore a special case of the VC-cDCC model.

The VC-cDCC model assumes that a set of assets is clustered with respect to the dynamic evolution of their GARCHvariances, as summarized by the univariate GARCH parameters (intercept excluded). One possible objection to such anassumption is the lack of evident clustering patterns across estimated GARCH parameters. For instance, the first panel ofFig. 1 reports the scatter plot of the univariate QML estimates of GARCH(1, 1) parameters on 100 stock returns series. Thegraph does not show any clear cluster. The second panel of Fig. 1 is very similar to the first one, but derives from 100simulated series clustered into four groups. This result is due to the estimation error of the GARCH parameters, and thuson their dispersion around the true data generating process values. Therefore, the estimation error could be responsiblefor the apparent absence of clusters within a set of GARCH(1, 1) estimated parameters. Note that the clusters we mightidentify are purely data-driven or dynamic-driven, without any specific link to the economic motivation that might lead tothe presence of clusters. There might be different reasons, starting from similarities between the activities of the companiesrelated to the assets up to the existence of common features. Those aspects are not investigates in the present paper thataims at introducing the clustering approach and the estimation algorithm. Furthermore, we know that in large samples, theestimation error of the univariate QML GARCH estimator is approximately Gaussian. This fact provides a strong motivationfor constructing an ad hoc algorithm for the clustering of GARCH processes based on Gaussian mixture (GM) models. Thealgorithm we propose in the next section, called the GARCH-GM clustering algorithm, provides an estimate of the groupsand thus of the partition vector π .

3. VC-cDCC estimation

We assume that P and Q are known (typically P = Q = 1), and we arrange the unknown VC-cDCC parameters as(π, τ , S, C, φ), where τ = [τ1, τ2, . . . , τN ]

′ is the vector of the asset unconditional variances (see 4), C = [C1 : C2 : · · · : CN ]

is the (P + Q ) × N matrix of the GARCH dynamic parameters (if the GARCH orders P and Q differ across the N assets, themissing elements are replaced by zeros in the C matrix), and φ is the vector stacking the distinct elements of the correlationdynamic parameter matrices, (A, B). The notation we adopt for the representation of the entire set of model parameters isredundant. In fact, the parameter matrix C contains replicated columns if the selected assets have a clustering structure.Nevertheless, to simplify the notation and the description of the estimation step, we maintain the matrix C , while we stressthat the VC-cDCC GARCH dynamic parameters are those in the vectors µi, i = 1, 2, . . . , K (one vector for each cluster). Wealso note that in the limiting case of the cDCC model, all columns of C are different given the assets do not show evidenceof a clustering structure.

Unless N is very small, the joint QML estimation of the VC-cDCC variance and correlation parameters is infeasible. Theproblem is in fact evenmore complex, given that the partition vector must also be estimated. As a feasible though no longer

Page 5: Variance clustering improved dynamic conditional correlation MGARCH estimators

G.P. Aielli, M. Caporin / Computational Statistics and Data Analysis ( ) – 5

efficient estimation strategy, we consider a two-step approach, called VC-cDCC estimator for short. The first step focuseson the estimation of the number of clusters and of the partition vector, and the second step on the joint estimation ofthe dynamic parameters. We stress that, thanks to variance clustering, even for large N , we are able to jointly estimate thewhole set ofmodel dynamic parameters (variance and correlation dynamic parameters). This is a relevant advantage againstthe traditional cDCC estimator for large systems, where the variance process and the correlation process are separatelyestimated. In the following sections, we address the estimation of the partition vector and of the other model parameters.

3.1. The GARCH-GM clustering algorithm

Denote as (τi, Ci) the univariate QML estimate of (τi, Ci). We assume the following:1. the return processes are independent;2. the assets are drawn at random from the considered market, with discrete probability density function

Pr(Ci = µk) = ωk,

k = 1, 2, . . . , K ,K

k=1 ωk = 1;3. the estimation error of the dynamic parameter, Ci − Ci, satisfies

√T (Ci − Ci)

A∼ N(0,Σ(Ci)). (10)

Assumptions 1–2 (also adopted in the clustering approach of Bauwens and Rombouts (2007)) are typically violated infinancial applications: assets are normally correlated and, for instance, portfolios are not randomly created. In Section 4.1wepresent a simulation study showing that the performances of our clustering algorithm turn out to be largely unaffected byviolations of assumptions 1 and 2. We stress that assumption 2 also implies a discrete GARCH dynamic distribution withinthe market; this is not restrictive in that, for sufficiently large K , we can approximate any continuous distribution on the(P + Q )-simplex of R(P+Q+1). Assumption 3 requires that the asymptotic distribution of Ci is Gaussian and centred on Ci,which holds under general conditions (see Francq and Zakoïan (2010), and therein cited references). Moreover, it requiresthat the asymptotic covariancematrix of Ci does not depend on τi. We do not provide a rigorous proof of this, but we confineourself to suggest a heuristic argument in favour of it. Note that the generating process of yi,t is yi,t = εi,t

hi,t . Setting

hi,0 = τi, it can be easily seen that, for fixed Ci and εi,0, εi,1, . . . , εi,T , the series yi,t andhi,t depend on changes in τi only by

a (common) scale factor; their dynamic properties, as governed by Ci, do not change. From an estimation perspective, thisimplies that, for fixed Ci and εi,0, εi,1, . . . , εi,T , the estimator of the dynamic parameter, Ci, does not depend on τi (of course,the estimator τi does depend on it); one can easily verify this fact via simulation. Therefore, for fixed Ci, the distribution ofCi, and its covariance matrix as a consequence, do not depend on τi. Such a property holds for any fixed T ; then, it shouldalso hold asymptotically, which, under asymptotic normality, would implies the correctness of (10). We might have put uson the safe side allowing for a very mild dependence of (10) on τi and possibly on other moments of the iid noise. However,we prefer to keep the notation of (10) for simplicity.

Under assumptions 1–3, for large T the Ci’s are approximately iidwith GM distributionK

k=1

ωk × f (Ci;µk,Σ(µk)/T ), (11)

i = 1, 2, . . . ,N , where the component covariance matrix is a function of the component mean. If the parameters K , µk, fori = 1, 2, . . . , K , and the explicit form ofΣ(·)were known, the partition vector, π , could be estimated setting

πi = argmaxk∈{1,2,...,K}

f (Ci;µk,Σ(µk)/T ), (12)

i = 1, 2, . . . ,N . Here, πi, is the index of the Gaussian component with the largest contribution to the GM likelihood whenevaluated at Ci (see, among others, Xu andWunsch (2009), and therein cited references). We notice that, under consistencyof the Ci’s (see Eq. (10)), for T −→ ∞ the dispersion of the k-th component decreases around µk; therefore, the estimatedpartition vector, π , is consistent. We also point out that the mixture probabilities ωk are not used for the estimation of thepartition vector.

Since the parameters K ,µk, for i = 1, 2, . . . , K , and the explicit form ofΣ(·) are unknown, we replace them in (12) withappropriate estimators. We denote as Σi the block corresponding to Ci in the QML estimator of the covariance matrix of(τi, Ci) (see Bollerslev and Wooldridge (1992)). As an estimator ofΣ(·)we suggest the following:

Σ(x) =

Ni=1

ψi(x)× Σi, ψi(x) = ∥x − Ci∥−1 N

i=1

∥x − Ci∥−1, (13)

where ∥ · ∥ is the Euclidean norm. Note that all the inputs needed to compute Σ(x) are provided by most codes forunivariate GARCH estimation. Furthermore, we observe that ψi(x) ≥ 0 and

Ni=1 ψi(x) = 1, so that Σ(x) is a weighted

Page 6: Variance clustering improved dynamic conditional correlation MGARCH estimators

6 G.P. Aielli, M. Caporin / Computational Statistics and Data Analysis ( ) –

mean of Σ1, . . . , ΣN ; therefore, by construction, Σ(x) is strictly positive definite uniformly on the (P + Q )-simplex ofR(P+Q+1). Thanks to this property, the GM log-likelihood remains well-conditioned irrespective of critical quantities, suchas the minimum cluster size. With other specifications of the component covariance matrix, such as the heteroskedasticGM clustering (see below), keeping a well-conditioned GM log-likelihood is often a problem. Finally, we note that underconsistency of (Ci, Σi) it should be possible to verify that for sufficiently large T and in a neighbourhood of the truecomponent mean, Σ(·) is a good approximation ofΣ(·).

To estimate the number of clusters, we adopt a BIC-based selection procedure. For K = 1, 2, . . . , we compute the QMLestimates of (11) conditionally toΣ(·) = Σ(·), denoted as

{(ωk, µk), k = 1, 2, . . . , K}. (14)

Note that the estimation is performed using the EM algorithm. Then, we set

K = min {K = 1, 2, . . . : BICK < BICK+1} , (15)

where

BICK = −2Ni=1

logK

k=1

ωk × f (Ci; µk, Σ(µk)/T )+ {(K − 1)+ K(P + Q )} × logN. (16)

If BICK is infeasible (for example, for K > N , when the GM log-likelihood is ill-conditioned), we set BICK = ∞. Recallingthat there are no parameters to take into account for the component covariance matrices, the number of parameters inthe BIC formula, (K − 1) + K(P + Q ), is equal to the number of free mixture weights plus the number of parameters inthe component means. We notice that the definition adopted for K provides a stopping rule that activates when BICK+1 isreached. If BICK is a convex function of K , then K is a global minimizer of BICK . Furthermore, for small K (as it is typicallythe case; see Section 5.3) this stopping rule also induces a computational advantage since it strongly reduces the estimationtime.

In order to compute π , we also need an estimator of the component means, µk, k = 1, 2, . . . , K , for K = K . As a naturalchoice, we adopt the component means provided by the QML estimates (14) for K = K . We notice that such estimators areprovided as a by-product of the BIC-selection procedure (14)–(16). Finally, we observe that the estimates ofωk, the mixtureprobabilities, which obtained together with the estimates of the component means µk, are never used in the clusteringalgorithm.

We summarize here the clustering procedure just described, which we label GARCH-GM clustering algorithm.

GARCH-GM clustering algorithm

i. Compute (τi, Ci) and Σi, i = 1, 2, . . . ,N;ii. set K = 0 and BIC0 = +∞;iii. set K = K + 1 and compute the maximizer, {(ωk, µk), k = 1, 2, . . . , K}, of

Ni=1

logK

k=1

ωk × f (Ci;µk, Σ(µk)/T ); (17)

iv. set BICK as in (16); if BICK ≥ BICK−1, set K = K − 1; else, go to (iii);v. for K = K , µk = µk, k = 1, 2, . . . , K , andΣ(·) = Σ(·), compute π as in (12).

We now consider some graphical examples comparing the performances of the GARCH-GM clustering algorithm withthose of other simpler GM clustering techniques, namely, heteroskedastic (HE) GM clustering and homoskedastic (HO)GM clustering. HE-GM clustering is computed replacing Σ(µk) in (17) with a free, positive semi-definite, component-specific parameter matrix Σk, k = 1, 2, . . . , K . The number of parameters entering the BIC formula in this case is(K −1)+K(P+Q )+K(P+Q )(P+Q +1)/2. In contrast, HO-GM clustering is computed replacing Σ(µk) in (17) with a free,positive semi-definite parameter matrixΣ , common to all components. The number of distinct parameters entering the BICformula in this case is (K−1)+K(P+Q )+(P+Q )(P+Q+1)/2. Tomaximize the univariate GARCHquasi-log-likelihood,weadopt a grid-based choice for the starting values. This procedure reduces the risk of choosing as optimal parameters thoseassociated with a local maxima; see Paolella (2010) for a discussion on the local maxima in GARCH QML estimation. Theplot relative to the proposed clustering methods is reported in Fig. 2, while additional graphs are included in the Appendix.We start from the GARCH-GM clustering (see Fig. 2). As expected, for small sample sizes (T = 250), the dispersion of theunivariate QML estimators is too large for the asset GARCH dynamics to be correctly partitioned. The estimated number ofclusters is incorrect (the smallest BIC is for K = 3, whereas the true number of clusters is K = 2). On the other hand, in thecases of T = 1250 and T = 2500, the number of clusters is correctly detected, and the asset GARCH dynamics are accuratelypartitioned. By comparing the shape and the dispersion of the component contour plots with the related clusters, the choiceof Σ(µk) as component covariance matrix estimator seems appropriate.

Page 7: Variance clustering improved dynamic conditional correlation MGARCH estimators

G.P. Aielli, M. Caporin / Computational Statistics and Data Analysis ( ) – 7

Fig. 2. Univariate QML GARCH(1, 1) dynamic parameter estimates from simulated data (N = 100) and related GARCH-GM component contour plots. TheDGP variance parameters are set as in Table 1—large dataset, with constant conditional correlations set as sij = s = 0.3 for all asset pairs. The true numberof clusters is K = 2. The component contour plots are drawn in correspondence to the Gaussian confidence sets of level 0.50, 0.75, 0.90. The lower triangleof each plot is the admissible parameter space, {(a, b) : a ≥ 0, b ≥ 0, a + b < 1}.

Moving to the HE-GM clustering, we note that the component covariancematrix is correctly specified but, unfortunately,for this clustering technique to be feasible, it is required that K is not too large with respect to N , and that the clustersizes are not too small. Fig. A.1 in the Appendix illustrates the behaviour of the HE-GM clustering relying on the samesimulated dataset used with the GARCH-GM clustering. In spite of the correct specification of the component covariancematrix, some odd behaviours can arise. In fact, there are positively correlated clusters whose shape is opposite with respectto the stochastic properties of the GARCH(1, 1) QML estimators. However, for T = 2500 the number of clusters is correctlydetected (the smallest BIC is for K = 2), and the asset GARCH dynamics are accurately partitioned.

Finally, in the HO-GM clustering, we observe that, thanks to homoskedasticity, this procedure remains feasible even forsmall cluster sizes. However, the misspecification of the component covariance matrix can be substantial (see Fig. A.2 in the

Page 8: Variance clustering improved dynamic conditional correlation MGARCH estimators

8 G.P. Aielli, M. Caporin / Computational Statistics and Data Analysis ( ) –

Appendix), as shown by the presence of positively correlated clusters, and by the fact that many observations fall outsidethe 90% component contour plots. Even for large T the number of clusters is overestimated.

As a further issue, we are interested in evaluating the consistency of the estimated partition vector for T −→ ∞ given aset of selected assets. In this respect, we note that, under consistency of the univariate GARCH QML estimators (see Francqand Zakoïan (2010)), for increasing T , the Cis concentrate around their true values with increasing probability. Therefore,for a fixed set of assets (given a choice of N) and for sufficiently large T , any ‘‘well-constructed’’ clustering algorithm shouldbe capable of correctly detecting the true underlying GARCH dynamic partition.

3.2. Joint dynamic parameters estimation

The second step of the VC-cDCC estimator focuses on the joint estimation of both the GARCH and correlation dynamicparameters. Even if the partition vector were known, unless N is very small, the dimension of the parameter space wouldbe too large for the joint feasible QML estimator of (τ , S, C, φ). This limitation depends mainly on the need to estimate theelements in the parameter matrix S (that enter in the intercept of the dynamic correlations). In fact, S contains N(N − 1)/2parameters, and is thus of order O(N2). On the other hand, the remaining model parameters include N unconditionalvariances (the elements of τ ), plus (P + Q )K dynamic GARCH parameters (the elements of the distinct rows of C), plus2 parameters (the elements of φ) in the case of Scalar VC-cDCC parametrization. Following Aielli (forthcoming), we restorefeasibility by resorting to a profile estimator. In VC-cDCC estimation, we treat the dynamic parameters, (C, φ), as parametersof interest and the parameters (π, τ , S) as nuisance parameters. Denote the pseudo-likelihood for the joint estimation of(π, τ , S, C, φ) as

L(π, τ , S, C, φ), (18)

and suppose the availability of estimators of π , τ , and S, conditionally to (C, φ), denoted as π(C,φ), τ(C,φ), and S(C,φ),respectively. The profile pseudo-likelihood,

L(π(C,φ), τ(C,φ), S(C,φ), C, φ), (19)is a function of (C, φ) only, where C is subject to Ci = Cj if πi(C, φ) = πj(C, φ), i, j = 1, 2, . . . ,N , i = j. The estimator of(C, φ) is defined as a maximizer of the profile pseudo-likelihood,

(C, φ) = argmax(C,φ)

L(π(C,φ), τ(C,φ), S(C,φ), C, φ). (20)

The (unconditional) estimators of π , τ , and S are defined, respectively, as π = π(C,φ), τ = τ(C,φ), and S = S(C,φ). They arethe values of the conditional estimators of π , τ , and S, at the end of the maximization of the profile pseudo-likelihood.

The feasibility of the estimation procedure just described depends on the number of parameters of the profile pseudo-likelihood, (20), which is (P + Q )K + dim(φ), and on the feasibility of the conditional estimators, π(C,φ), τ(C,φ), and S(C,φ).For large N , as in our focus, and for common values of P and Q , such as P = Q = 1, the estimated number of clusters,K , turns out to be typically small (e.g., K ≤ 5 with N = 100—see Section 5). Hence, for suitably restricted φ, as in thecase of scalar correlation dynamics, the profile pseudo-likelihood typically is a function of a small number of parameters.As feasible conditional estimators of π and τ we adopt, respectively, the GARCH-GM estimated partition vector, π , andτ = [τ1, . . . , τN ]

′, that is, the vector of the estimators of the τi’s as provided by the univariate GARCH QML estimators. As afeasible conditional estimator of S, we suggest the following specification, which is constructed relying on the property ofthe cDCC model given in (8). For fixed (C, φ), set

S(C,φ) = T−1T

t=1

{Q ∗

t εt}{Q∗

t εt}′, (21)

wherei. ε∗

t = [ε∗

1,t , . . . , ε∗

N,t ]′;

ii. ε∗

1,t = yi,t/h1,t ;

iii. hi,t = (1 −P

p=1 aip −Q

q=1 bi,q)τi +P

p=1 ai,py2i,t−p +

Qq=1 bi,qhi,t−q;

iv. Q ∗t = diag(

q1,t , . . . ,

qN,t);

v. qi,t = (1 − αii − βii)+ αiiε2it−1qi,t−1 + βiiqi,t−1.

The estimator S(C,φ) is defined as the sample counterpart of S evaluated at (C, φ) conditionally to π = π and τ = τ , see (8).Since S(C,φ) depends on (C, φ), it must be recomputed at each evaluation of the profile pseudo-likelihood. Note, however,that the recursions required by the evaluation of S(C,φ) are not computationally complex. Furthermore, if Q ∗

t εt is secondmoment ergodic, if evaluated at the true value of (τ , C, φ) the estimator S(C,φ) is consistent (see Aielli (forthcoming)).

A natural choice for the pseudo-likelihood (18) would be the quasi-log-likelihood. In that case, the profile pseudo-likelihood (19) were called a generalized profile quasi-log-likelihood (Severini, 1998). Following Engle et al. (2009), in place

Page 9: Variance clustering improved dynamic conditional correlation MGARCH estimators

G.P. Aielli, M. Caporin / Computational Statistics and Data Analysis ( ) – 9

Table 1DGP variance parameter values.

Large dataset (N = 100) Small dataset (N = 10)

K = 2 K = 2N1 = 66,N2 = 34 N1 = 9,N2 = 1τi = 1, i = 1, . . . ,N τi = 1 i = 1, . . . ,N(ai, bi) = (0.05, 0.90), i = 1, . . . ,N1 (ai, bi) = (0.06, 0.93), i = 1, . . . ,N1(ai, bi) = (0.18, 0.54), i = N1 + 1, . . . ,N1 + N2 (ai, bi) = (0.12, 0.85), i = N1 + 1 = N1 + N2

of the quasi-log-likelihood we adopt a bivariate composite quasi-log-likelihood, namely,

L(π, τ , S, C, φ) =

Ni=2

Li,i−1(πi, πi−1, τi, τi−1, S, C, φ), (22)

where Li,j(πi, πj, τi, τj, S, C, φ) is the quasi-log-likelihood of the submodel of (yi,t , yj,t). With respect to the full quasi-log-likelihood, the bivariate composite quasi-log-likelihood allows a sensible reduction of the bias of the estimated parametersof interest, caused by the large number of nuisance parameters. Moreover, with respect to the full quasi-log-likelihood, thebivariate composite quasi-log-likelihood results in a dramatic reduction of CPU time, especially for large N .

We finally consider a variant of the estimation procedure just described, which turns out to be feasible irrespective ofK . It can be obtained treating (π, τ , S, C) as nuisance parameters and φ as a parameter of interest. The pseudo-likelihoodis the bivariate composite quasi-log-likelihood in (22), and the conditional (and unconditional) estimators of π and τ , are,respectively, π and τ . As a conditional estimator of C , we set

C = [γπ1 : γπ2 : · · · : γπN ], (23)

where

γk = argmaxµ

Ni=1

I(πi = k)× Li(τi, γ ), k = 1, . . . , K , (24)

where Li(τi, Ci) denotes the univariate quasi-log-likelihood of yi,t . Finally, φ is computed as

φ = argmaxφ

L(π, τ , Sφ, C, φ), (25)

where Sφ = S(C,φ), where S(C,φ) is defined as in (21). Since C does not depend on φ, it needs not to be recomputed during the

computation of φ, and it coincideswith the unconditional estimator of C . The unconditional estimator of S is S = Sφ = S(C,φ).Themaindifference between the estimator definedby (20) (hereafter, joint VC-cDCCestimator) and the estimator defined

by (23)–(25) (hereafter, sequentialVC-cDCC estimator), is that, with the joint estimator, all dynamic parameters of themodelare estimated jointly, whereas, with the sequential estimator, the dynamic variance parameters and the dynamic correlationparameters are estimated in subsequent steps. The calculations required by the sequential estimator are much faster thanthose required by the joint estimator. In particular, the computation of C reduces to K small size maximizations, wherethe k-th objective function is the sum of the univariate quasi-log-likelihoods of the assets belonging to the k-th cluster(also called univariate composite likelihood—see Pakel et al. (2011)). In the case of scalar correlation dynamics, the profileobjective function in (25) turns out to be a function of two variables only, irrespective of N . For K ≥ N (that is, possiblyno common GARCH dynamics) the sequential estimator coincides with the traditional large system cDCC estimator (Aielli,forthcoming). The targeted version of the joint and sequential estimators is obtained replacing the τi’s with the asset samplevariances (Engle and Mezrich, 1996; Caporin and McAleer, 2012).

Finally, we note that, from a computational point of view, both the joint and the sequential estimators benefit from theavailability of a good starting value for the parameters in C , that is, the set of the component means provided as an outputof the GARCH-GM clustering (see Section 3.1).

4. Simulation studies

4.1. Clustering estimation performances

In this section we focus on the clustering simulation performances of the GARCH-GM clustering algorithm describedin Section 3.1. For the simulation experiments, we consider two data generating processes, one for large datasets, whereN = 100, and another one for small datasets, where N = 10. The variance processes are assumed GARCH(1, 1) withparameters set as in Table 1. These parameter settings are coherent with the estimation output provided by the real datasetsemployed in this paper (see Section 5.1). For appropriate specifications of Rt , we then generate the variance standardized

Page 10: Variance clustering improved dynamic conditional correlation MGARCH estimators

10 G.P. Aielli, M. Caporin / Computational Statistics and Data Analysis ( ) –

returns as εt = R1/2t zt , where R1/2

t is the Cholesky square root of Rt , and where the elements of zt are iid standardizedStudent’s t-distributedwith ν = 9 degrees of freedom.We preferred the Student density to be consistent with the empiricalevidences of leptokurtosis in financial returns series. Furthermore, to limit the computation time, we run the BIC-selectionprocedure only for K = 1, 2, 3, 5, 8, 13 on the large datasets, and only for K = 1, 2, 3, 5, 8 on the small datasets. Tointroduce dependence across the simulated assets, we assume constant conditional correlations, which we set as

sij = s = 0, 0.3, 0.6, 0.9, i < j = 2, . . . ,N.Forcing s to vary in a range of admissible values allows us to check whether the GARCH-GM clustering performances areaffected by violations of the assumption of independence across the return processes. For each parameter constellation, wegenerate M = 500 series of length T = 1250. We compare the performances of the GARCH-GM clustering algorithm withthose of the HE-GM clustering and the HO-GM clustering.

We evaluate the performances of such clustering algorithms by comparing the estimated number of clusters, K , with thetrue number of clusters, K . Furthermore, we consider a measure of incorrect partition, which we define as

Ψ =1

N(N + 1)/2

i<j=2,...,N

|pij − pij|,

wherepij = I(πi = πj);pij = I(πi = πj).

(26)

A smaller Ψ is associated to a better fit of π toπ . It holds 0 ≤ Ψ ≤ 1,where Ψ = 0 if and only if π = π . The actualmaximumof Ψ is less than one and it depends on π . We note that Ψ does not require that the compared partitions need have the samenumber of clusters. See Xu andWunsch (2009) for other approaches andmethods to evaluate the performances of clusteringalgorithms.

By analysing the simulation results of the large dataset experiment (see Fig. 3), we note that GARCH-GM clustering andHE-GM clustering have similar performances. However, there is a slight preference for the GARCH-GM clustering in terms ofpercentage of incorrect allocations. In contrast, the HE-GM clustering is slightly preferred in terms of estimated number ofclusters. Note that, as long as the asset cross-correlation increases, the performances of both clustering algorithms improve.For s = 0.9, the GARCH-GM clustering globally outperforms HE-GM clustering in terms of both incorrect partition andestimated number of clusters. The performances of HO-GM clustering in terms of incorrect allocations are satisfactory, andthey are almost unaffected by changes in the asset cross-correlation. In terms of the estimated number of clusters, however,HO-GM clustering exhibits a marked tendency to overestimate the true value.

Moving to the small dataset experiment (see Fig. 4), GARCH-GM clustering strongly outperforms HO-GM clustering interms of both incorrect allocations and estimated number of clusters. The HO-GM clustering confirms the previous results,that is, the overestimation of the true number of clusters. Because of the large number of parameters to estimatewith respectto the number of observations, HE-GM clustering is infeasible in this experiment.

In summary, for reasons of flexibility and the performances in both the simulation experiments, GARCH-GM clusteringseems to be globally the best method among the three considered clustering techniques. In the case of large datasets withapproximately balanced clusters, HE-GM clustering can be a valid alternative to the GARCH-GM clustering. However, westress that a-priori the number of clusters and their sizes are not known, an element favouring the use of the GARCH-GMclustering method.

4.2. Variance/correlation estimation performances

In this section we focus on the simulation performances of the joint dynamic parameter estimation as provided by theVC-cDCC estimators described in Section 3.2. In summary, we find that the joint VC-cDCC estimator and the sequential VC-cDCC estimator both perform similarly in practice and better than the cDCC estimator. The targeted VC-cDCC estimatorsare recommendable with respect to the corresponding non-targeted versions, especially when dealing with large systems.The fact that the joint estimator and the sequential estimator are empirically equivalent means that there is no sensibleloss of efficiency due to estimating the model’s dynamic parameters in two steps. We stress that the sequential estimator isextremely rapid to compute and is feasible irrespective of K .

The DGP’s of the simulation experiment are set as in Section 4.1, apart from the correlation process, which is scalardynamic with parameters

α = 0.02; β = 0.97; sij = s = 0.3 i < j = 2, . . . ,N.These correlation settings are coherent with typical cDCC estimation outputs. Again, for each parameters constellation, wegenerate M = 500 series of length T = 1250. To measure the performances of the estimated variance process, we adoptthe following mean absolute errors

VARMAE =1N

Ni=1

hi,T+1 − hi,T+1

(27)

Page 11: Variance clustering improved dynamic conditional correlation MGARCH estimators

G.P. Aielli, M. Caporin / Computational Statistics and Data Analysis ( ) – 11

a b

c d

e f

g h

Fig. 3. Simulation-based (N = 100) percentage of incorrect allocations, Ψ , and estimated number of clusters, K , arranged in ascending order. Straightlines for the GARCH-GM clustering; dashed lines for the HE-GM clustering, and dash–dotted lines for the HO-GM clustering. The true number of clustersis K = 2.

and

CORRMAE =1

N(N − 1)/2

i<j=2,...,N

ρij,T+1 − ρij,T+1 , (28)

where hi,T+1 and ρij,T+1 are the out-of-sample estimates of, hi,T+1 and ρij,T+1, respectively, based on y1, y2, . . . , yT . We alsocompare the density plots of the parameter estimators.

Fig. 5 refers to the large dataset experiment. All plots show that the joint VC-cDCC estimator and the correspondingsequential VC-cDCC estimators provide the same empirical performances in practice. On the contrary, the performances of

Page 12: Variance clustering improved dynamic conditional correlation MGARCH estimators

12 G.P. Aielli, M. Caporin / Computational Statistics and Data Analysis ( ) –

a b

c d

e f

g h

Fig. 4. Simulation-based (N = 10) percentage of incorrect allocations, Ψ , and estimated number of clusters, K , arranged in ascending order. Straight linesfor the GARCH-GM clustering; dash–dotted lines for the HO-GM clustering. The true number of clusters is K = 2. Due to the large number of parameters,the HE-GM clustering is infeasible for N = 10.

the targeted VC-cDCC estimators and of the corresponding non-targeted versions can be different, with a preference for thetargeted estimator. Consider the variance results in plot (1): the VC-cDCC out-of-sample predictions, both targeted and non-targeted, perform better than the cDCC variance prediction (the VARMAE density of the VC-cDCC estimators is more shiftedtowards zero than the corresponding cDCC density). As expected, thanks to variance clustering, the VC-cDCC parameterestimators are more concentrated around the true values than the corresponding cDCC estimators (see plots [3], [5] and [7]in Fig. 5). Because of the estimation error in the partition vector, the behaviour of the VC-cDCC estimator can be less regularthan the one of the cDCC estimator, as shown by the bimodality of the VC-cDCC estimators of bi (see plot [7]). The targetedestimators are generally well centred around the true values, whereas the non-targeted versions can exhibit rather large

Page 13: Variance clustering improved dynamic conditional correlation MGARCH estimators

G.P. Aielli, M. Caporin / Computational Statistics and Data Analysis ( ) – 13

a b

c d

e f

g h

Fig. 5. Variance/correlation estimation performances with N = 100: cDCC estimator in straight dotted lines; targeted joint VC-cDCC estimator in straightlines; targeted sequential VC-cDCC estimator in dashed lines; non-targeted joint VC-cDCC estimator in dashed–dotted lines; non-targeted sequential VC-cDCC estimator in dotted lines.

relative biases (see plots [3] and [7]). Focusing now on the correlation results, there is no apparent increase of predictionefficiency due to the variance clustering (see plot [2]). As for the correlation parameters, the targeted VC-cDCC density plotsand the corresponding cDCC density plots are well centred on the true values and they are almost identical in practice (seeplots [4], [6] and [8]). The non-targeted VC-cDCC estimators instead exhibit a rather large relative bias in the estimation ofα (see plot [6]).

The simulation results for small N are reported in Fig. A.3 in the Appendix. They confirm the fact that there is littledifference between the joint VC-cDCC estimation and the sequential VC-cDCC estimation. The best variance predictionshere are provided not only by the non-targeted VC-cDCC estimators, but also the targeted VC-cDCC estimators outperformthe cDCC estimator (see plot [1]). As expected, in the case of the second cluster, which includes one asset only, the estimationerror in the partition vectormakes the VC-cDCC variance parameter estimators performancesworse than the correspondingcDCC estimator (see the right quadrants in plots [5] and [7]). As noted in the case of the large dataset, the correlationperformances appear to be unaffected by allowing for variance clustering (see plots [2], [4], [6] and [8]).

Page 14: Variance clustering improved dynamic conditional correlation MGARCH estimators

14 G.P. Aielli, M. Caporin / Computational Statistics and Data Analysis ( ) –

5. Applications to real data

In this section, we compare the empirical performances of the VC-cDCC estimator and those of the cDCC estimator onreal data. In summary, according to the considered performance criteria, we find that the targeted VC-cDCC estimatorsalways perform either equal or better than the cDCC estimator. As anticipated by the simulation evidence, the differencebetween the joint VC-cDCC estimators and the corresponding sequential versions is negligible. In contrast to the simulationresults, the estimation of the correlation process is markedly improved by allowing for variance clustering. This somewhatsurprising outcome depends, in our opinion, on the rolling estimation strategy adopted.We believe that the targeted versionis less sensible to large changes in the assets unconditional variances. Furthermore, within the simulations the unconditionalvariances remained fixed, while in the real data case, the unconditional variances might change over time, with a possibleimpact on the model ranking. Note this section is purely illustrative of the application of the VC-cDCC model, and limitedattention is given to the economic interpretation behind the presence of clusters.

5.1. Datasets and compared estimators

We consider a large dataset of N = 100 randomly selected equities from the S&P 1500 Index. These equities belongeither to the industrial or to the consumer good sectors, and their daily total returns are available from January 2001 toDecember 2007. Therefore, this dataset has an expectation on the number of clusters, which is equal to 2. The sample periodthus includes 1770 daily returns. For the same period, we consider also a small dataset of equity indices: the nine S&P 500SPDR’s sector indices and the S&P 500 index itself (see Table A.10 in the Appendix for the detailed list of the consideredassets). On the two datasets, we estimate both cDCC and VC-cDCCmodels. For the secondmodel, we consider both the jointestimators of the number of clusters and of the other model parameters, denoted as cDCCVC, as well as the estimators ofmodel parameters conditionally on a a priori defined number of clusters denoted as cDCCK . Estimating models with a fixednumber of clusters allows a deeper understanding of the behaviour of the VC-cDCC estimator. Another interesting questionis whether the data driven partition of the VC-cDCC estimator outperforms some relevant a priori partition criterion. Toshed some light on this aspect, we also estimate on the large dataset a VC-cDCC model with a priori fixed partition vectordenoted as cDCC∗. The a priori partition is given by the sector partition into industrial equities (53 assets) and consumergood equities (47 assets). For all of these estimators we also consider the related targeted version (see Section 3.2) denotedby a superscript ‘‘Trg ’’. Thus, for example, the targeted VC-cDCC estimator is denoted as cDCCTrg

VC . We evaluate all estimatorsassuming GARCH(1, 1) variance processes and scalar correlation dynamics.

5.2. Performance criteria

We compare the several model estimators by means of out-of-sample prediction criteria. In particular, we considerpairwise comparison tests of equal predictive ability (see Diebold and Mariano (1995) and Patton and Sheppard (2009),for applications in MGARCH models) and regression-based specification tests, such as the Engle–Colacito regression (Engleand Colacito, 2006), the dynamic quantile test (Engle and Manganelli, 2004), and the LM test of ARCH effects on portfolioreturns. We consider as portfolio specifications the equally weighted portfolio, denoted as EW, the minimum varianceportfolio with short selling, denoted as MV, and the minimum variance portfolio without short selling, denoted as MV∗.The weight vector of the EW portfolio iswt = w = ι/N . The weight vector of the MV portfolio iswt = {Ht}

−1ι/(ι′{Ht}−1ι),

and, then it is an explicit function of the conditional covariance matrix. Finally, the weight vector of the MV∗ portfolio is anon-closed function of the conditional covariance matrix, and it must be numerically computed. During the computationof the portfolio weights, Ht is replaced by an estimate. Since the EW weights do not depend on Ht , they are free fromestimation errors. In the following, we report a description of the adopted comparison criteria. The symbol H(m)t denotesthe out-of-sample estimate of Ht , which is computed applying the m-th estimator on a rolling window of t de-meanedobserved returns, yt−1, yt−2, . . . , yt−t , where yt−j = zt−j − zt−1, and zt−j is the vector of the observed return at time t− j andzt−1 = t−1t

j=1 zt−j. We adopt a rolling window of length t = 1250. The number of estimated forecasts for each estimatoris then T = T − t = 520, roughly corresponding to two years of data.

5.2.1. Pairwise comparison tests of equal predictive abilityDenote as d(m)t the loss due to predicting Ht with H(m)t , and set

d(m) = T−1T

t=t+1

d(m)t .

The null hypothesis

H0: Ed(m) − d(n)

= 0

Page 15: Variance clustering improved dynamic conditional correlation MGARCH estimators

G.P. Aielli, M. Caporin / Computational Statistics and Data Analysis ( ) – 15

is a null of equal predictive ability for the estimators H(m)t and H(n)t . Under appropriate conditions (see Diebold andMariano,1995), a test statistic for H0 is

T (d(m) − d(n))VAR[

T (d(m) − d(n))]∼ N(0, 1),

where VAR[

T (dm − dn)] is a heteroskedasticity and autocorrelation consistent estimate of the variance of√

T (dm − dn).Negative values of the test statistic provide evidence in favour of H(m)t against H(n)t . As possible specifications of d(m)t weconsider score-based losses as in Amisano and Giacomini (2007), and MSE-based losses as in Diebold and Mariano (1995)and Patton and Sheppard (2009). Notably, theMSE loss is robust to the use of a noisy proxy as shown in Patton and Sheppard(2009) and Laurent et al. (2011, 2013).

Amisano–Giacomini losses. Given our dynamic correlation setting, we decide to compute two separate score-based losses,one for the asset variance process and another for the asset correlation process, both computed under Gaussianity. The lossfor the variance process is

d(m)t = VARSCORE(m)t =

Ni=1

log h(m)i,t + y2i,t/h

(m)i,t

, (29)

where h(m)it is the out-of-sample estimate of hit . The loss for the correlation process is

d(m)t = CORRSCORE(m)t = log |R(m)t | + ε(m)′t {R(m)t }

−1ε(m)t , (30)

where ε(m)t =

y1,t/

h(m)1,t , . . . , yN,t/

h(m)N,t

. The loss for the correlation process depends on the estimated variances

through ε(m)t . We also consider a score-based loss for EW portfolio returns, defined as

d(m)t = EWSCORE(m)t = log(w′H(m)t w)+w′yt

2/(wH(m)t w), (31)

where w is the vector of the EW portfolio weights. The definition of the losses for MV portfolio returns and MV∗ portfolioreturns, denoted as MVSCORE and MVSCORE∗, respectively, is equivalent, with the caveat that in these cases, the vector ofthe portfolio weights is the appropriate function of H(m)t . Given any of such score-based losses, under Gaussianity d(m) isminimum in large samples if and only if H(m)t = Ht .

Diebold–Mariano losses. The Amisano–Giacomini score-based loss accounts for misspecifications in either the model of Ht ,or the asset conditional distribution, or both. A loss function that accounts for misspecification in the model of Ht only is theDiebold–Mariano MSE-based loss. Again we consider a loss for the variance process and a loss for the correlation process.The variance loss is

d(m)t = VARMSE(m)t =1N

Ni=1

y2i,t − h(m)i,t

2, (32)

and the correlation loss is

d(m)t = CORRMSE(m)t =1

N(N − 1)/2

i<j=2,...,N

ε(m)i,t ε

(m)j,t − ρ

(m)ij,t

2, (33)

where ρ(m)ij,t is the out-of-sample estimate of ρij,t . The correlation loss depends on the estimated variances through ε(m)t . Theportfolio MSE-based loss is defined as

d(m)t = EWMSE(m)t =

w′

t yt2

− w′

t H(m)t wt

2. (34)

The definition of the losses for MV portfolio returns and MV∗ portfolio returns, denoted as MVMSE and MVMSE∗,respectively, is analogue. Given any of such MSE-based losses, d(m) is minimum in large samples if and only if H(m)t = Ht .

The previously described loss functions requires as an input the true and unknown variances or correlations. Those havebeen replaced by quantities measured using the observed returns. Such a choice is clearly suboptimal since the proxy weare using is contaminated by a noise. However, as discussed in Patton and Sheppard (2009), the previous loss functions arerobust to the noise of the volatility proxy used. Alternatively, other approaches could have been considered, such as the useof realized variances as in Laurent et al. (2011), but that require the availability of high frequency data.

Page 16: Variance clustering improved dynamic conditional correlation MGARCH estimators

16 G.P. Aielli, M. Caporin / Computational Statistics and Data Analysis ( ) –

Fig. 6. Univariate QML estimates of GARCH(1, 1) dynamic parameters obtained from real data. The contour plots are relative to the components of theestimated GM-GARCHmodel (K = 3 on the left; K = 2 on the right). The component contour plots are drawn in correspondence to the Gaussian confidencesets of level 0.50, 0.75, 0.90.

5.2.2. Regression-based specification testsEngle–Colacito regression. Consider the regression model {(w′

t yt)2/(w′

t H(m)t wt)} − 1 = λ+ ξt ,where ξt is an innovation

term. By construction, λ is zero if H(m)t = Ht . Hence, the significance of a test of λ = 0, where HAC-robust standard errorsare required, provides evidence of misspecification.

Dynamic quantile test. The α × 100% Value-at-Risk (VaR) at time t for a given portfolio, w′tyt , is denoted as VaRt(α)

and is defined as the α × 100%-quantile of the conditional distribution of w′tyt . Under Gaussianity it holds VaRt(α) =

Φ−1(α)w′

tHtwt , where Φ(·) is the standard Gaussian distribution function. Define HITt = 1 if w′t yt < VaRt(α) and

HITt = 0 otherwise, where VaRt(α) is the estimated VaR. The dynamic quantile test is an F-test of the hypothesis that allcoefficients as well as the intercept are zero in a regression of {HITt −α} on its past, the current VaR and any other variables.In this paper, fifteen lags and the current VaR are used. Rejecting the null provides evidence of model misspecification.

Portfolio ARCH effect test. If the model is correctly specified, the series of the square standardized portfolio returns,(w′

tyt)2/(w′

tHtwt), is serially uncorrelated, or w′tyt exhibits no ARCH effects. Rejecting the null of no serial correlation

of (w′t yt)

2/(w′t H

(m)t wt), then show evidence of model misspecification. We fit the test using fifteen lags on the squared

residuals.

5.3. Results

5.3.1. Large datasetFig. 6 reports the univariate estimates of theGARCH(1, 1) dynamic parameters computed on thewhole sample period. The

figure also shows the contour plots of the estimated best GMmodel. The estimated number of clusters is K = 3. Accordingly,most of the rolling window estimates of K oscillate between three and five with the large dataset (see Fig. 7). In order tolimit the calculations, the BIC-selection procedure is run only for K = 1, 2, 3, 5, 8, 13 rather than for K = 1, . . . ,N , whichdoes not cause any notable loss of generality, as shown by the fact that the rolling window estimate of K is almost alwaysless than 13 (see Fig. 7).

Table 2 refers to the Amisano–Giacomini tests computed for the joint VC-cDCC estimator. Our particular focus is on thecomparison of the VC-cDCC estimators with data driven partition, denoted as cDCCVC and cDCCTrg

VC , with the traditional cDCCestimator. We first discuss the results relative to the targeted VC-cDCC estimators (see the left column of Table 2). TheVARSCORE table shows that allowing for variance clustering does not provide sensible improvements with respect to thecDCC estimator. The test statistic comparing cDCCTrg

VC and cDCC is equal to −.61, which is in favour of cDCCTrgVC (the sign is

negative) but is not significant at the 5% level. The CORRSCORE table shows different results. The correlation predictions aresensibly improved by allowing for variance clustering (the last column of the table is all negative and strongly significant).Among the considered VC-cDCC estimators, cDCCTrg

VC provides the second best performance against cDCC, with a test statisticequal to −6.69 (the best performance is provided by cDCC3 with a test statistic equal to −6.72). The good performance ofthe a priori partitioned estimator, cDCCTrg

∗, is a fictitious effect. Note, in fact, that the overall best estimator is the one-

cluster estimator, cDCCTrg1 , as shown in the CORRSCORE table (the elements in the first row are all negative and significant),

suggesting that the a priori sector partition basically acts as a random partition. Since there are only two clusters in the a

Page 17: Variance clustering improved dynamic conditional correlation MGARCH estimators

G.P. Aielli, M. Caporin / Computational Statistics and Data Analysis ( ) – 17

Fig. 7. Rolling window estimates of the number of clusters on real data.

priori partition, the loss of efficiency of cDCCTrg∗

with respect to the well-specified estimator cDCCTrg1 is small. This conjecture

is supported by the fact that the sign of most pairwise comparison tests between cDCCTrg∗

and cDCCTrg1 is in favour of cDCCTrg

1 .As for the results from the portfolio-based pairwise comparisons (see EVSCORE, MVSCORE and MVSCORE∗ in Table 2), thetest to cDCCTrg

VC and cDCC, though not significant, is in favour of cDCCTrgVC . The results relative to the non-targeted VC-cDCC

estimators (see the right column of Table 2) indicate that the overall message is that the performances are worse thanthose of the corresponding targeted versions. These results are in accordance with the simulation evidence discussed in theprevious sections. Table A.1 in the Appendix reports the Amisano–Giacomini pairwise comparison tests computed for thesequential VC-cDCC estimators. In accordance with the simulation evidence, the pattern of the results is very similar to thatobtained with the joint estimators.

The results from the Diebold–Mariano pairwise comparison tests (see Table 3 for an example) are analogous to – thoughless decisive than – those from the Amisano–Giacomini tests. In particular, it is confirmed that the performance of thesequential estimators are very similar to that of the corresponding joint estimators, and that the targeted estimatorsperform better than the corresponding non-targeted estimators. Provided that the targeting is applied, the predictionof the correlation process is significantly improved by allowing for variance clustering (see the last column of the leftCORRSCORE table). As desired, cDCCTrg

VC significantly outperforms cDCC. Note that the a priori partitioned estimator, cDCCTrg∗

,is significantly better than cDCC, but it is also significantly worse than the one-cluster estimator, cDCCTrg

1 . This differencesupports the conjecture that the good performances of the sector partition are fictitious. As for the results from theregression-based tests (Table 4 includes some examples), note that in all tests the performances of the targeted VC-cDCCestimators are either equal or better than those of the cDCC estimator.

In both targeted and non-targeted cases, the overall message from the pairwise comparison tests is that the estimatorswith fixed K ≤ 3 provide the best performances (most of the dark grey cells are concentrated on the rows related to suchestimators). Since the rollingwindow estimates of the VC-cDCC estimator aremostly larger than three (see Fig. 6), they seemto suggest that the performances of theVC-cDCCestimator could be sensibly improvedby correcting it for the overestimationof the number of clusters already shown via simulation (see Fig. 3).

5.3.2. Small datasetThe univariate estimates of the GARCH(1, 1) dynamic parameters for the whole sample period are plotted in Fig. 6. The

estimated number of clusters is K = 2, and, accordingly, most of the rolling window estimates of K are equal to two (seeFig. 7). In order to speed up the calculations, the BIC-selection procedure is run for K = 1, 2, 3, 5, 8. With respect to thelarge dataset, the superiority of the targeted estimators with respect to the corresponding non-targeted versions appears tobe less marked, which is in accordance with the simulation results reported in Section 4.2. According to all the consideredcomparison criteria, including the regression-based tests, all the considered VC-cDCC estimators, either joint or sequential,as well as targeted or non-targeted, perform equal to or better than the cDCC estimator. In particular, in the ten consideredpairwise comparison tests, cDCCTrg

VC significantly beats cDCC three times in the sequential version and four times in the jointversion; cDCCVC significantly beats cDCC five times in the sequential version and six times in the joint version.

As in the case of the large datasets, the estimators with a fixed, small number of clusters perform generally better thanthe remaining estimators. Furthermore, most statistically significant comparisons are related to the estimators with fixedK ≤ 2, which seems to confirm that the VC-cDCC estimator can be improved by correcting it for the overestimation of thenumber of clusters.

Page 18: Variance clustering improved dynamic conditional correlation MGARCH estimators

18 G.P. Aielli, M. Caporin / Computational Statistics and Data Analysis ( ) –

Table 2Amisano–Giacomini pairwise comparison tests for N = 100 and with joint dynamic parameter estimation. White cells denote equal predictive ability at5% level; light (dark) grey cells denote rejection of the null of equal predictive ability at 5% level, and suggest preference for the column (row) model.

cDCCTrg1 cDCCTrg

2 cDCCTrg3 cDCCTrg

5 cDCCTrg∗

cDCCTrgVC cDCC cDCC1 cDCC2 cDCC3 cDCC5 cDCC∗ cDCCVC cDCC

VARSCORE

cDCCTrg1 – 0.77 0.54 0.49 −1.42 −0.38 −0.53 cDCC1 – 0.76 0.29 0.03 −1.28 −1.05 −0.02

cDCCTrg2 – – 0.13 −0.04 −1.61 −1.29 −1.06 cDCC2 – – −0.34 −0.45 −1.70 −1.98 −0.48

cDCCTrg3 – – – −0.30 −1.03 −2.54 −1.50 cDCC3 – – – −0.37 −1.10 −2.71 −0.37

cDCCTrg5 – – – – −1.04 −2.75 −1.39 cDCC5 – – – – −0.56 −2.17 −0.11

cDCCTrg∗

– – – – – −0.02 −0.35 cDCC∗ – – – – – −0.41 0.51

cDCCTrgVC – – – – – – −0.61 cDCCVC – – – – – – 1.75

cDCC – – – – – – – cDCC – – – – – – –

CORRSCORE

cDCCTrg1 – −2.45 −2.89 −3.96 −3.15 −3.96 −5.84 cDCC1 – 2.36 1.48 1.49 −1.05 0.87 0.27

cDCCTrg2 – – −2.32 −4.20 0.87 −4.19 −6.53 cDCC2 – – −0.69 0.24 −3.37 −0.91 −1.21

cDCCTrg3 – – – −1.39 2.24 −2.74 −6.72 cDCC3 – – – 0.92 −2.38 −0.71 −1.10

cDCCTrg5 – – – – 3.37 −1.81 −6.66 cDCC5 – – – – −2.01 −2.57 −3.07

cDCCTrg∗

– – – – – −3.48 −6.00 cDCC∗ – – – – – 1.49 0.70

cDCCTrgVC – – – – – – −6.69 cDCCVC – – – – – – −1.08

cDCC – – – – – – – cDCC – – – – – – –

EWSCORE

cDCCTrg1 – 1.64 1.70 1.04 −0.88 1.05 0.26 cDCC1 – 0.82 0.18 −1.14 −0.61 −1.25 −0.69

cDCCTrg2 – – 1.36 0.06 −2.73 0.06 −0.70 cDCC2 – – −1.01 −2.01 −1.90 −2.10 −1.59

cDCCTrg3 – – – −2.04 −2.06 −2.06 −2.12 cDCC3 – – – −2.17 −0.71 −2.33 −1.58

cDCCTrg5 – – – – −1.48 −0.01 −1.23 cDCC5 – – – – 0.76 −0.67 1.65

cDCCTrg∗

– – – – – 1.50 0.52 cDCC∗ – – – – – −0.85 −0.33

cDCCTrgVC – – – – – – −1.21 cDCCVC – – – – – – 1.81

cDCC – – – – – – – cDCC – – – – – – –

MVSCORE

cDCCTrg1 – 0.49 −0.70 0.39 0.10 −0.13 −0.20 cDCC1 – −1.84 −2.65 −1.86 −0.91 −2.15 −1.13

cDCCTrg2 – – −1.18 0.05 −0.43 −0.58 −0.45 cDCC2 – – −1.66 −0.73 1.07 −1.13 −0.18

cDCCTrg3 – – – 1.28 0.71 0.75 0.18 cDCC3 – – – 1.19 2.42 0.73 1.22

cDCCTrg5 – – – – −0.35 −0.87 −0.61 cDCC5 – – – – 1.47 −0.69 0.51

cDCCTrg∗

– – – – – −0.14 −0.21 cDCC∗ – – – – – −1.84 −0.80

cDCCTrgVC – – – – – – −0.20 cDCCVC – – – – – – 0.82

cDCC – – – – – – – cDCC – – – – – – –

MVSCORE∗

cDCCTrg1 – −0.64 −0.38 −0.01 −1.44 −0.73 −0.68 cDCC1 – −2.05 −1.42 −0.81 −0.22 −1.57 −0.95

cDCCTrg2 – – 0.24 0.63 0.15 −0.15 −0.46 cDCC2 – – 0.54 1.16 1.90 0.31 0.19

cDCCTrg3 – – – 0.54 −0.11 −0.54 −0.60 cDCC3 – – – 0.96 1.42 −0.30 −0.16

cDCCTrg5 – – – – −0.42 −1.44 −0.86 cDCC5 – – – – 0.75 −1.35 −0.68

cDCCTrg∗

– – – – – −0.27 −0.46 cDCC∗ – – – – – −1.55 −0.95

cDCCTrgVC – – – – – – −0.40 cDCCVC – – – – – – 0.03

cDCC – – – – – – – cDCC – – – – – – –

6. Conclusions

Wepropose an estimationmethodology to improve the performances of the traditional cDCC estimator for large systemsby allowing for assets sharing the same dynamic variance parameters. The suggested estimator is fully frequentist, easy to

Page 19: Variance clustering improved dynamic conditional correlation MGARCH estimators

G.P. Aielli, M. Caporin / Computational Statistics and Data Analysis ( ) – 19

Table 3Diebold–Mariano pairwise comparison tests withN = 100 and joint dynamic parameter estimation.White cells denote equal predictive ability at 5% level;light (dark) grey cells denote rejection of the null of equal predictive ability at 5% level, and suggest preference for the column (row) model.

cDCCTrg1 cDCCTrg

2 cDCCTrg3 cDCCTrg

5 cDCCTrg∗

cDCCTrgVC cDCC cDCC1 cDCC2 cDCC3 cDCC5 cDCC∗ cDCCVC cDCC

VARMSE

cDCCTrg1 – −0.98 −0.65 −1.64 0.81 −1.81 −1.52 cDCC1 – −2.40 −1.35 −2.00 −1.78 −2.20 −1.49

cDCCTrg2 – – 0.02 −1.53 1.20 −1.75 −1.15 cDCC2 – – −0.12 −1.38 1.67 −1.63 −0.55

cDCCTrg3 – – – −2.28 0.84 −2.73 −1.20 cDCC3 – – – −2.63 0.98 −3.04 −0.76

cDCCTrg5 – – – – 1.63 −0.91 0.20 cDCC5 – – – – 1.70 −0.64 1.02

cDCCTrg∗

– – – – – −1.79 −1.72 cDCC∗ – – – – – −1.89 −1.18

cDCCTrgVC – – – – – – 0.47 cDCCVC – – – – – – 1.17

cDCC – – – – – – – cDCC – – – – – – –

CORRMSE

cDCCTrg1 – −0.84 −0.84 −1.46 −2.38 −1.71 −2.92 cDCC1 – 1.97 1.36 0.95 −1.05 0.47 0.25

cDCCTrg2 – – −0.61 −1.85 −0.53 −2.12 −3.62 cDCC2 – – −0.16 0.05 −2.74 −1.10 −0.92

cDCCTrg3 – – – −1.27 0.18 −2.98 −4.33 cDCC3 – – – 0.19 −2.13 −1.83 −1.25

cDCCTrg5 – – – – 0.80 −1.84 −3.90 cDCC5 – – – – −1.38 −1.68 −2.80

cDCCTrg∗

– – – – – −1.19 −2.91 cDCC∗ – – – – – 1.09 0.69

cDCCTrgVC – – – – – – −3.78 cDCCVC – – – – – – −0.25

cDCC – – – – – – – cDCC – – – – – – –

EWMSE

cDCCTrg1 – 1.47 1.63 0.93 −0.95 0.94 0.22 cDCC1 – 0.22 −0.28 −1.52 −1.11 −1.64 −1.04

cDCCTrg2 – – 1.42 −0.02 −2.33 0.01 −0.81 cDCC2 – – −0.86 −1.85 −1.59 −1.91 −1.37

cDCCTrg3 – – – −2.29 −1.95 −2.29 −2.42 cDCC3 – – – −2.19 −0.77 −2.30 −1.46

cDCCTrg5 – – – – −1.32 0.15 −1.35 cDCC5 – – – – 0.85 −0.36 2.09

cDCCTrg∗

– – – – – 1.34 0.48 cDCC∗ – – – – – −0.92 −0.33

cDCCTrgVC – – – – – – −1.35 cDCCVC – – – – – – 2.03

cDCC – – – – – – – cDCC – – – – – – –

MVMSE

cDCCTrg1 – −0.07 −1.01 −0.12 −1.06 −0.40 −0.02 cDCC1 – −1.48 −2.34 −1.45 −1.03 −2.15 −1.39

cDCCTrg2 – – −0.89 −0.08 −0.17 −0.32 0.02 cDCC2 – – −1.24 −0.37 0.48 −1.16 −0.62

cDCCTrg3 – – – 0.77 0.86 0.79 0.61 cDCC3 – – – 1.03 1.97 0.62 0.51

cDCCTrg5 – – – – −0.05 −0.25 0.07 cDCC5 – – – – 0.73 −0.59 −0.36

cDCCTrg∗

– – – – – −0.21 0.10 cDCC∗ – – – – – −1.84 −0.78

cDCCTrgVC – – – – – – 0.29 cDCCVC – – – – – – 0.17

cDCC – – – – – – – cDCC – – – – – – –

MVMSE∗

cDCCTrg1 – −1.02 0.03 0.19 −1.42 −0.67 −0.15 cDCC1 – −2.32 −1.58 −1.32 −0.89 −2.15 −1.77

cDCCTrg2 – – 0.78 0.85 0.56 0.45 0.84 cDCC2 – – 0.11 0.38 1.29 −0.42 −0.77

cDCCTrg3 – – – 0.27 −0.55 −0.74 −0.17 cDCC3 – – – 0.65 1.18 −0.62 −1.04

cDCCTrg5 – – – – −0.62 −1.39 −0.32 cDCC5 – – – – 0.83 −1.16 −1.45

cDCCTrg∗

– – – – – −0.05 0.38 cDCC∗ – – – – – −1.80 −1.41

cDCCTrgVC – – – – – – 0.41 cDCCVC – – – – – – −0.59

cDCC – – – – – – – cDCC – – – – – – –

implement, and endowed with good statistical properties. We first cluster the assets into groups, and then we estimatethe model parameters subject to constant within-group variance dynamic parameters. The adopted clustering algorithmis new. It exploits the asymptotic properties of the univariate GARCH quasi-maximum-likelihood estimator, and thus it isdistribution-free. Furthermore, in contrast to other clustering methodologies, the algorithm leads to the estimation of thenumber of clusters. Thanks to the equality parameter constraints induced by the estimated variance clustering structure,

Page 20: Variance clustering improved dynamic conditional correlation MGARCH estimators

20 G.P. Aielli, M. Caporin / Computational Statistics and Data Analysis ( ) –

Table 4Regression-based specification tests with N = 100 and joint dynamic parameter estimation. White cells denote insignificance at 5% level, light grey cellsdenote significance at 5% level and dark grey cells denote significance at 1% level.

cDCCTrg1 cDCCTrg

2 cDCCTrg3 cDCCTrg

5 cDCCTrg∗

cDCCTrgVC cDCC cDCC1 cDCC2 cDCC3 cDCC5 cDCC∗ cDCCVC cDCC

Engle–Colacito regression

EW 1.90 1.97 1.95 2.03 1.97 2.01 2.27 EW 2.37 2.28 2.23 2.28 2.38 2.29 2.27

MV 3.16 3.22 3.52 3.47 3.15 3.49 4.26 MV 4.17 4.49 4.63 4.47 4.13 4.45 4.26

MV∗ 1.47 1.64 1.76 1.83 1.54 1.87 2.64 MV∗ 2.69 2.86 2.79 2.66 2.65 2.74 2.64

Portfolio arch effect test

EW 37.03 36.33 35.73 36.31 36.89 36.31 35.77 EW 33.53 36.64 35.98 36.67 35.04 36.06 35.77

MV 17.60 16.53 17.25 16.32 18.28 19.27 21.63 MV 24.69 19.85 17.49 18.35 25.87 22.32 21.63

MV∗ 32.00 36.92 26.16 26.08 32.64 27.51 31.62 MV∗ 29.18 31.86 26.70 28.07 30.54 28.48 31.62

1% dynamic quantile test

EW 1.76 1.80 1.80 1.84 1.77 1.80 1.84 EW 1.85 1.85 1.84 1.85 1.84 1.85 1.84

MV 0.91 0.88 1.03 0.75 0.92 1.24 2.02 MV 0.47 0.65 2.19 1.72 0.95 2.12 2.02

MV∗ 1.36 1.38 1.49 1.53 1.29 1.53 2.37 MV∗ 1.89 2.31 2.09 2.32 2.53 2.34 2.37

5% dynamic quantile test

EW 0.72 0.71 0.71 0.71 0.72 0.71 0.71 EW 1.11 0.71 0.97 0.97 1.10 0.97 0.71MV 1.42 1.71 0.87 0.87 1.19 1.21 2.36 MV 0.96 1.09 0.92 1.33 1.76 1.99 2.36

MV∗ 1.81 1.60 1.25 1.26 1.36 1.19 1.54 MV∗ 1.33 1.21 1.22 1.70 1.12 1.41 1.54

we are capable of jointly estimating the whole set of model dynamic parameters. Simulations and applications to real dataare included, which show that the suggested estimator allows some improvements in terms of efficiency with respect to thetraditional cDCC estimator, where the variance dynamics are estimated one at a time and separately from the correlationdynamics.We complete the paperwith an illustrative example.We did not tackled the issue of linking the identified clustersto some economicmotivation, beside the possible presence of clusters due to the extraction of the assets from two economicsectors. A more complete empirical evaluation of the VC-cDCC model should have also compared its performances to thoseof competing approaches outside the DCC model class. We cite the Factor GARCH of Engle et al. (1990), the OGARCH ofAlexander (2001), and the common features model of Engle and Marcucci (2006). A detailed empirical analysis pointingalso to the economic interpretation of the GARCH clusters is left to future researches.

Acknowledgements

We thank the Associate Editor and two anonymous referees for their stimulating and helpful comments. Moreover,we thank Christian Brownlees, Giorgio Calzolari, Eduardo Otranto, Marc Paolella, Michael McAleer, Nicola Sartori andthe participants to the Computational and Financial Econometrics 2011 conference in London for their comments andsuggestions.

Appendix. Supplementary data

Supplementary material related to this article can be found online at http://dx.doi.org/10.1016/j.csda.2013.01.029.

References

Aielli, G.P., 2013. Dynamic conditional correlation MGARCH models: on properties and estimation. Journal of Business and Economic Statistics(forthcoming).

Alexander, C.O., 2001. Orthogonal GARCH. In: Alexander, C.O. (Ed.), Mastering Risk, Vol. II. Prentice-Hall, pp. 21–38.Amisano, G., Giacomini, R., 2007. Comparing density forecasts via weighted likelihood ratio tests. Journal of Business and Economic Statistics 25, 177–190.Bauwens, L., Rombouts, J.V.K., 2007. Bayesian clustering of many garch models. Econometric Reviews 26, 365–386.Billio, M., Caporin, M., 2009. A generalised dynamic conditional correlation model for portfolio risk evaluation. Mathematics and Computers in Simulation

79, 2566–2578.Billio, M., Caporin, M., Gobbo, M., 2006. Flexible dynamic conditional correlation multivariate GARCH for asset allocation. Applied Financial Economics

Letters 2, 123–130.Bollerlsev, T., 1986. Generalized autoregressive conditional heteroskedasticity. Journal of Econometrics 31, 307–327.Bollerlsev, T., 1990. Modelling the coherence in short run nominal exchange rates: a multivariate generalized ARCH model. The Review of Economics and

Statistics 72, 498–505.Bollerslev, T., Wooldridge, J.M., 1992. Quasi maximum likelihood estimation and inference in dynamic models with time varying covariances. Econometric

Reviews 11, 143–172.Brownlees, C., 2009. On the relation between firm characteristics and volatility dynamics with an application to the 2007–2009 financial crisis. Available

at SSRN: http://ssrn.com/abstract=1695649.

Page 21: Variance clustering improved dynamic conditional correlation MGARCH estimators

G.P. Aielli, M. Caporin / Computational Statistics and Data Analysis ( ) – 21

Caporin, M., McAleer, M., 2012. Dowe really need both BEKK and DCC? A tale of twomultivariate GARCHmodels. Journal of Economic Surveys 26, 736–751.Cappiello, L., Engle, R.F., Sheppard, K., 2006. Asymmetric dynamics in the correlations of global equity and bond returns. Journal of Financial Econometrics

4, 537–572.Diebold, F.X., Mariano, R.S., 1995. Comparing predictive accuracy. Journal of Business and Economic Statistics 13, 253–263.Engle, R.F., 2002. Dynamic conditional correlation—a simple class of multivariate GARCHmodels. Journal of Business and Economic Statistics 20, 339–350.Engle, R.F., Colacito, R., 2006. Testing and valuing dynamic correlations for asset al location. Journal of Business and Economic Statistics 24, 238–253.Engle, R.F., Kelly, B.T., 2012. Dynamic equicorrelation. Journal of Business and Economic Statistics 30, 212–228.Engle, R.F., Manganelli, S., 2004. CAViaR: conditional Value at Risk by regression quantiles. Journal of Business and Economic Statistics 22, 367–381.Engle, R.F., Marcucci, J., 2006. A long-run pure variance common Features model for the common volatilities of the Dow Jones. Journal of Econometrics 132,

7–42.Engle, R.F., Mezrich, J., 1996. GARCH for groups. Risk 9, 36–40.Engle, R.F., Ng, V.K., Rothschild, M., 1990. Asset pricing with a factor ARCH covariance structure: empirical estimates for Treasury bills. Journal of

Econometrics 45, 213–238.Engle, R.F., Shephard, N., Sheppard, K., 2009. Fitting vast dimensional time-varying covariance models. Working Paper FIN-08-009, NYU Stern School of

Business, Department of Finance.Engle, R.F., Sheppard, K., 2001. Theoretical and empirical properties of dynamic conditional correlation multivariate GARCH. Working Paper FIN-01-027,

NYU Stern School of Business, Department of Finance.Francq, C., Zakoïan, J., 2010. GARCH Models, Structure, Statistical Inference and Financial Applications. Wiley, UK.Franses, P.H., Hafner, C.M., 2009. A generalised dynamic conditional correlation model: simulation and application to many assets. Econometric Reviews

28, 612–631.Hafner, C.M., Reznikova, O., 2012. On the estimation of dynamic conditional correlationmodels. Computational Statistics and Data Analysis 56, 3533–3545.Laurent, S., Rombouts, J.V.K., Violante, F., 2011. On the forecasting accuracy of multivariate GARCH models. Journal of Applied Econometrics 27, 934–955.Laurent, S., Rombouts, J.V.K., Violante, F., 2013. On loss functions and ranking forecasting performances of multivariate volatility models. Journal of

Econometrics 173, 1–10.Liao, T.W., 2005. Clustering of time series data—a survey. Pattern Recognition 38, 1857–1874.McAleer, M., Chan, F., Hoti, S., Lieberman, O., 2008. Generalized autoregressive conditional correlation. Econometric Theory 24, 1554–1583.Otranto, E., 2008. Clustering heteroskedastic time series by model-based procedures. Computational Statistics and Data Analysis 52, 4685–4698.Otranto, E., 2010. Identifying financial time series with similar dynamic conditional correlation. Computational Statistics and Data Analysis 54, 1–15.Pakel, C., Shephard, N., Sheppard, K., 2011. Nuisance parameters, composite likelihoods, and a panel of GARCH models. Statistica Sinica 21, 307–329.Paolella, M.S., 2010. ALRIGHT: asymmetric LaRge-Scale (I) GARCH with Hetero-Tails, Swiss Finance Institute Research Paper Series N. 10-27, available at

SSRN. http://ssrn.com/abstract=1628146.Patton, A.J., Sheppard, K., 2009. Evaluating volatility and correlation forecasts. In: Andersen, T.G., Davis, R.A., Kreiß, J.P., Mikosch, T. (Eds.), Handbook of

Financial Time Series. Springer.Pesaran,M.H., Pesaran, B., 2007.Modelling volatilities and conditional correlations in futuresmarketswith amultivariate t distribution. Unpublished paper:

Department of Economics, University of Cambridge.Severini, T.A., 1998. Likelihood functions for inference in the presence of a nuisance parameter. Biometrika 85, 507–522.Xu, R., Wunsch II, D., 2009. Clustering. In: IEEE Press Series on Computational Intelligence, IEEE Press and Wiley, NJ.