review of discrete and continuous methods

33
Electronic copy available at: http://ssrn.com/abstract=1373102 Review of Discrete and Continuous Processes in Finance Theory and Applications Attilio Meucci 1 [email protected] this version: December 04 2010 last version available at http://ssrn.com/abstract=1373102 Abstract We review the main processes used to model nancial variables. We emphasize the parallel between discrete-time processes, mainly used by econometricians for risk- and portfolio-management, and their continuous-time counterparts, mainly used by mathematicians to price derivatives. We highlight the relationship of such processes with the building blocks of stochastic dynamics and statistical inference, namely the invariants. Figures and practical examples support intu- ition. Fully documented code illustrating these processes in practice is available at MATLAB Central File Exchange under the author’s name. JEL Classication: C1, G11 Keywords: invariants, random walk, Levy processes, autocorrelation, ARMA, Ornstein-Uhlenbeck, long memory, fractional integration, fractional Brownian motion, volatility clustering, GARCH, stochastic volatility, subordination, real measure, risk-neutral measure, fat tails. 1 The author is grateful to Giuseppe Castellacci, Gianluca Fusai, Fabio Mercurio and Arne Staal for their helpful feedback 1

Upload: ryan-davis

Post on 14-Apr-2015

38 views

Category:

Documents


4 download

DESCRIPTION

Buy-side quantitative finance

TRANSCRIPT

Page 1: Review of Discrete and Continuous Methods

Electronic copy available at: http://ssrn.com/abstract=1373102

Review of Discrete and Continuous Processes in FinanceTheory and Applications

Attilio [email protected]

this version: December 04 2010last version available at http://ssrn.com/abstract=1373102

Abstract

We review the main processes used to model financial variables. We emphasizethe parallel between discrete-time processes, mainly used by econometricians forrisk- and portfolio-management, and their continuous-time counterparts, mainlyused by mathematicians to price derivatives. We highlight the relationship ofsuch processes with the building blocks of stochastic dynamics and statisticalinference, namely the invariants. Figures and practical examples support intu-ition. Fully documented code illustrating these processes in practice is availableat MATLAB Central File Exchange under the author’s name.

JEL Classification: C1, G11

Keywords: invariants, random walk, Levy processes, autocorrelation, ARMA,Ornstein-Uhlenbeck, long memory, fractional integration, fractional Brownianmotion, volatility clustering, GARCH, stochastic volatility, subordination, realmeasure, risk-neutral measure, fat tails.

1The author is grateful to Giuseppe Castellacci, Gianluca Fusai, Fabio Mercurio and ArneStaal for their helpful feedback

1

Page 2: Review of Discrete and Continuous Methods

Electronic copy available at: http://ssrn.com/abstract=1373102

Contents1 Introduction 3

I P: discrete-time processes 5

2 Random walk 52.1 Continuous invariants . . . . . . . . . . . . . . . . . . . . . . . . 72.2 Discrete invariants . . . . . . . . . . . . . . . . . . . . . . . . . . 72.3 Generalized representations . . . . . . . . . . . . . . . . . . . . . 8

2.3.1 Stochastic volatility . . . . . . . . . . . . . . . . . . . . . 82.3.2 Mixture models . . . . . . . . . . . . . . . . . . . . . . . . 92.3.3 Stochastic volatility as mixture models . . . . . . . . . . . 9

2.4 Heavy tails . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10

3 ARMA processes 11

4 Long memory 13

5 Volatility clustering 15

II Q: continuous-time processes 17

6 Levy processes 176.1 Diffusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 176.2 Jumps . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 196.3 Generalized representations . . . . . . . . . . . . . . . . . . . . . 196.4 Notable examples . . . . . . . . . . . . . . . . . . . . . . . . . . . 20

7 Autocorrelated processes 21

8 Long memory 22

9 Volatility clustering 249.1 Stochastic volatility . . . . . . . . . . . . . . . . . . . . . . . . . 249.2 Subordination . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25

References 27

Index 29

A Appendix 30A.1 The Ornstein-Uhlenbeck process . . . . . . . . . . . . . . . . . . 30A.2 Relationship between CIR and OU processes . . . . . . . . . . . 31A.3 The Levy-Khintchine representation of a Levy process . . . . . . 32A.4 Representation of compound Poisson process . . . . . . . . . . . 33

2

Page 3: Review of Discrete and Continuous Methods

Electronic copy available at: http://ssrn.com/abstract=1373102

1 IntroductionSince the financial markets evolve in a random fashion, the natural object tomodel its evolution are stochastic processes. Stochastic processes are thus essen-tial in risk management, portfolio management and trade optimization to modelthe evolution of the risk factors that determine the price of a set of securitiesat a given point in the future. We denote by Xt the value of one such factor Xat a generic time t, i.e. the stochastic process for X. This stochastic process isfully described by a multivariate function, the process cumulative distributionfunction, i.e. the joint probability of its realizations at any set of times

Ft1,t2,... (x1, x2, . . .) ≡ P {Xt1 ≤ x1,Xt2 ≤ x2, . . .} . (1)

For the above mentioned purposes of risk management and portfolio manage-ment , the monitoring times t1, t2, . . . are typically chosen on a discrete, equally-spaced, grid t, t+∆, t+2∆, . . .: for instance ∆ can be one day, or one week, orone month, etc.The probability measure "P" is the "real" probability that governs the evo-

lution of the process. In reality, such probability is not known and needs to beestimated using advanced multivariate statistics: indeed, estimation is the mainconcern of the real-measure P-world. We summarize this in Figure 1 below, seeMeucci (2011) for more details.

Goal: "model the future"Environment: real probability PProcesses: discrete-time seriesTools: multivariate statisticsChallenges: estimationBusiness: buy-side

Figure 1: Risk and portfolio management: summary

There exists a parallel, much developed, area of application of stochasticprocesses in finance: derivatives pricing. According to the fundamental theoremof pricing , the normalized price ePt of a security is arbitrage-free only if thereexists a fictitious stochastic process describing its future evolution such that itsexpected value is always constant and equals the current price:

ePt = En ePt+τo , τ ≥ 0. (2)

A process satisfying (2) is called a martingale: since a martingale does not re-ward risk, the fictitious probability of the pricing process is called risk-neutraland is typically denoted by "Q". Since some prices are liquid and readily observ-able, (2) becomes a powerful tool to assign a fair price to a non-liquid security,

3

Page 4: Review of Discrete and Continuous Methods

once a martingale has been calibrated to other traded securities. In other words,derivatives pricing is an extrapolation exercise which lives in the present. Wesummarize this in Figure 2 below, see Meucci (2011) for more details and com-pare with Figure 1.

Goal: "extrapolate the present"Environment: risk-neutral probabilityQ

Processes: continuous-time martingalesTools: Ito calculus, PDE’sChallenges: calibrationBusiness: sell-side

Figure 2: Derivatives pricing: summary

Interactions between the P-world and the Q-world occur frequently. Forinstance, the sensitivities of a security to a set of risk factors, the so-calledGreeks, are computed using Q models, but then they are applied in the P-worldfor hedging purposes. Similarly, the evolution through time of the Q parametersof a given pricing model can be used to predict the P-world distribution of thesame prices in the future and generate such statistics as the value at risk. Also,Q-models are often used in the P world to spot misalignment and thus identifyopportunities for statistical arbitrage.Notice that the martingales used in the Q-world of derivatives pricing repre-

sent a restrictive class of processes. A tractable and much more general class ofprocesses are the so-called semimartingales, on which a theory of integration canbe defined that makes sense for financial modeling. However, for the P-world ofrisk and portfolio management we do not need to restrict our attention to semi-martingales: any process that suitably models the behavior of financial timesseries is a-priori admissible. On the other hand, the discrete-time processes ofthe P-world are a restrictive class of the more general continuos time processesexplored by the Q-world.We refer to Meucci (2011) for more perspective on the relationships between

the P and the Q world. Here we present an overview of the main processes infinance, highlighting the relationships between the P-world, discrete-time modelsand their Q-world, continuous-time counterparts.We start in Part I with the P-world discrete-time processes. In Section 2

we introduce the random walk, which is the cumulative sum of invariants. Thisprocess represents the benchmark for buy-side modelling: often considered toosimplistic by Q quants and P econometricians not involved in risk and portfo-lio management, the random walk proves to be very hard to outperform whenestimation error is accounted for. In Section 3 we relax the hypothesis that theincrements of the process be invariant: we introduce autocorrelation, therebyobtaining ARMA processes. In Section 4 we account for the empirical observa-

4

Page 5: Review of Discrete and Continuous Methods

tions that at times autocorrelation, when present, decays very slowly with thenumber of lags: this behavior is suitably modeled by long memory-processes. InSection 5 we discuss volatility clustering: the scatter of the process, rather thanthe process itself, displays autocorrelation: GARCH and generalizations thereofcapture this feature.In Part II we discuss the Q-world continuous-time counterparts of the above

P-world models. In Section 6 we present Levy processes, the continuous-timeversion of the random walk. In Section 7 we model autocorrelation with theOrnstein-Uhlenbeck and related processes. In Section 8 we model long memoryby means of the fractional Brownian motion. In Section 9 we tackle volatilityclustering in two flavors: stochastic volatility and subordination.We summarize in the table below the processes covered and their key char-

acteristics

discrete time (P) continuous time (Q)base case random walk Levy processesautocorrelation ARMA Ornstein-Uhlenbecklong memory fractional integration fractional Brownian motion

volatility clustering GARCH½stochastic volatilitysubordination

Fully documented code that illustrates the theory and the empirical aspects ofthe models in this article is available at symmys.com.

Part I

P: discrete-time processesWe assume that we monitor a financial process at discrete, equally spaced timeintervals. Without loss of generality, we can assume this interval to have unitlength: therefore, denoting time by t, we assume that t ∈ Z.

2 Random walkThe random walk with drift is a process that cumulates invariants

Xt+1 = Xt + t+1, (3)

The invariants t, are independent and identically distributed (i.i.d.) shockswith any distribution. Since the distribution of the invariants is fully general,the random walk has a much richer structure than one would expect.Two very simple tests to spot invariance in a realized time series are discussed

in Meucci (2005).First, due to the Glivenko-Cantelli theorem, if t is an invariant the his-

togram of its time series represents its probability density function (pdf). There-fore, also any portion of its time series represents the pdf. As a consequence,

5

Page 6: Review of Discrete and Continuous Methods

the histogram of the first half of the time series of t and that of the secondhalf should look alike, see the top portion of Figure 3 and the discussion in theexamples below.Second, if t is an invariant, then t and t+1 are independent and identically

distributed. Therefore the mean-covariance ellipsoid of t and t+1 must be acircle. A scatter plot of t versus t+1 with the plot of the mean-covarianceellipsoid should convey this impression, see the bottom portion of Figure 3 andthe discussion in the examples below.

-0.05 0 0.05 -0.05 0 0 .05

-0 .05 0 0 .05-0.05

0

0.05

time series first half time series second half

across-time scatter-plot

1tε +

tεtε

frequ

ency

frequ

ency

Figure 3: Invariance check

Consider a security that trades at time t at the price Pt and consider thecompounded return

Ct ≡ lnµ

PtPt−1

¶. (4)

In the equity world the benchmark assumption is that the compounded returnsare invariants, i.e. Ct ≡ t. This amounts to saying that Xt ≡ lnPt evolvesaccording to a random walk:

ln (Pt+1) = ln (Pt) + t+1. (5)

The plots in Figure 3 refer to this case. Also refer to Figure 3.2 and relateddiscussion in Meucci (2005).As another standard example, denote by Z

(E)t the price at time t of a zero-

coupon bond that matures at time E. As discussed in Meucci (2005), the rawbond price cannot give rise to an invariant, because the approaching maturitydates breaks the time translation invariance of the analysis. Therefore, one

6

Page 7: Review of Discrete and Continuous Methods

introduces the yield to maturity , which is the annualized compounded returnover the life of the bond

Y(υ)t ≡ − 1

υln³Z(t+υ)t

´. (6)

The time series of the yield to maturity, as opposed to the yield of maturity, isinvariant under time translations and therefore we can hope to use it to generateinvariants. Notice however that this time series does not correspond to the priceof one single bond throughout, but rather to the price of a different bond ateach time. In the fixed-income world the benchmark assumption is that thechanges in yield to maturity are invariants

Y(υ)t+1 = Y

(υ)t +

(υ)t+1, (7)

which is clearly in the format (3). One can check this assumption with the simpletest described in Figure 3. Also refer to Figure 3.5 and related discussion inMeucci (2005).

2.1 Continuous invariants

As a base case, the invariants can have a normal distribution

t ∼ N¡µ, σ2

¢, (8)

where µ is the expectation and σ2 is the variance. Independent normal distrib-utions are closed under the sum. With slight abuse of notation we write:

N¡µ1, σ

21

¢+N

¡µ2, σ

22

¢ d= N

¡µ1 + µ2, σ

21 + σ22

¢. (9)

Since the sum of independent normal variables is normal, from normal invariantsfollows a normally distributed random walk, i.e. a Brownian motion, see Section6.1.Different continuous distributions for the invariants give rise to different

distributions for the random walk (3). For instance, one can model the invariantsusing stable distributions, Student t distributions, skewed distributions, etc.

2.2 Discrete invariants

The invariants can also take on a set a ≡ {a0, . . . , aK} of discrete values with re-spective probabilities p ≡ {p0, . . . , pK} that sum to one. This is the generalizedBernoulli distribution:

t ∼ Be (p;a) . (10)

If the invariant has a generalized Bernoulli distribution, the ensuing randomwalk also has a generalized Bernoulli distribution with different parameters ateach time step.

7

Page 8: Review of Discrete and Continuous Methods

When the number of values in a is large, a more parsimonious parameteri-zation becomes necessary. An important case is the Poisson distribution:

t ∼ Po (λ;∆) . (11)

This represents a special case of (10) where the variable takes values on aninfinite, equally spaced grid ak ≡ k∆, also denoted as ∆N, with probabilities

pk ≡ P { t = k∆} ≡ λke−λ

k!. (12)

It follows that λ represents both the expectation and the variance of the invariant(11), as one can prove using the interchangeability of summation and derivationin the power series expansion of the exponential function.Independent Poisson distributions are closed under the sum. With slight

abuse of notation we write:

Po (λ1;∆) + Po (λ2;∆)d= Po (λ1 + λ2;∆) . (13)

Therefore the random walk (3) ensuing from a Poisson invariant is Poissondistributed on the same grid.

2.3 Generalized representations

Since any function of invariants is an invariant, we can create invariants byaggregation.

2.3.1 Stochastic volatility

A flexible formulation in this direction is the stochastic volatility formulation:

td= µt + σtZt. (14)

In this expression µt is a fully general invariant, σt is a positive invariant andZt is a general invariant with unit scatter and null location: these invariantsrepresent respectively the location, the scatter and the shape of t. specification.

For instance, for specific choices of the building blocks we recover the Studentt distribution:

µt ≡ µν/σ2t ∼ χ2νZt ∼ N

¡0, σ2

¢⎫⎬⎭⇒ t ∼ St

¡ν, µ, σ2

¢. (15)

We stress that the term stochastic volatility is used mostly when σt in (14)is not an invariant, see Section 5, the case where σt is an invariant is a specialcase, which we proceed to discuss further.

8

Page 9: Review of Discrete and Continuous Methods

2.3.2 Mixture models

Another flexible aggregate representation is the mixture model

td= (1−Bt)Yt +BtZt (16)

where Yt and Zt are generic invariants and Bt are invariants with a standardBernoulli distribution (10) with values in {0, 1}: the ensuing model for t isa mixture of Yt and Zt respectively, with a given probability: it is useful, forinstance, to model the regime switch between a regular market environment Ytand a market crash model Zt. More general mixture models select among Sstates

td=

SXs=1

B(s)t Y

(s)t , (17)

where Bt is a multiple-selection process with values in the canonical basis ofRS .

2.3.3 Stochastic volatility as mixture models

Stochastic volatility models can be seen as special cases of multi-mixture modelswith a continuum of components. Indeed, consider a generic stochastic volatilitymodel (14), which we report here

td= µt + σtZt. (18)

Denote by fµ,σ the joint pdf of the location and scatter invariants µt and σtrespectively. Define the following multi-mixture model

t ≡Z Z

B(µ,σ)t Y

(µ,σ)t dµdσ, (19)

whereY(µ,σ)t

d= µ+ σZt (20)

andB(µ,σ)t is a continuous multiple-selection process with values in the indicators

of infinitesimal intervals of R×R+ such that

PnB(µ,σ)t = I[dµ×dσ]

o= fµ,σ (µ, σ) dµdσ. (21)

Then the multi-mixture model (19) is distributed as the stochastic volatilitymodel (18).

From (15) the Student t distribution is a special instance of the stochasticvolatility representation (14) and thus it is also a multi-mixture distribution witha continuum of components: Zt in (20) is standard normal, and fµ,σ ≡ fµfσ,where fµ is a Dirac delta and fσ is the density of the square root of the inversechi-square, suitably rescaled.

9

Page 10: Review of Discrete and Continuous Methods

2.4 Heavy tails

Heavy-tailed distributions are those whose pdf decreases away from the locationof the distribution slower than exponentially. Suitable models for heavy tailscan be found among the continuous distributions, the discrete distributions, orthe generalized representations discussed above.

mar

gin

al d

istr

ibut

ion

DE

LL

marginal distr ib ution IBM

joint distribution IBM -D ELL

normal fit

norma l f it

Figure 4: Heavy tails in empirical distribution of compounded stock returns

Stable distributions are continuous models popular for their tractability, seee.g. Rachev (2003) and Meucci (2005): if the invariants t are stable, the randomwalk Xt has the same distribution as t, modulo rescaling.Another popular parametric model for heavy tails is the Student t distri-

bution (15), which for ν > 2 has finite variance. For ν ≡ 1 the Student tdistribution is stable and it is known as Cauchy distribution.In Figure 4 we display the empirical distribution of the daily compounded

returns (4) of two technology stocks. It is apparent that a normal fit is unsuitableto describe the heavy-body, heavy tail empirical behavior of these returns, referto MATLAB Central File Exchange under the author’s name for the code.

The heavy tails of a random walk tend to disappear with aggregation. Moreprecisely, consider the aggregate invariant

et,τ ≡ t + t−1 · · ·+ t−τ+1. (22)

If the one-step dynamics is the random walk (3), the τ -step dynamics is also arandom walk:

Xt+τ = Xt +et+τ,τ , (23)

10

Page 11: Review of Discrete and Continuous Methods

However, the central limit theorem prescribes that, if the invariant t has fi-nite variance, the aggregate invariant et,τ becomes normal for large enough anaggregation size.

In particular It follows that stable distributions must have infinite variance,or else they would violate the central limit theorem.

Therefore, random walks for large enough an aggregation size behave likediscrete samples from the Brownian motion, see Section 6.1.From (22) it follows that the variance of the steps of a random walk is

a linear function of the time interval of the step. This is the square root rulepropagation of risk: risk, defined as the standard deviation of the step, increasesas the square root of the time interval:

Sd {et,τ} ∝ √τ . (24)

For stable distributions, where the variance is not defined, one can use otherdefinitions of risk, such as the inter-quantile range. In particular, for the Cauchydistribution one can check that the propagation law of risk does not grow asthe square root of the time interval, but instead it grows linearly with the timeinterval.

3 ARMA processesThe steps of a random walk display no dependence across time, because they arethe invariants. This feature makes random walks not stationary: if Xt evolvesas in (3), its distribution never stabilizes. To model stationarity we consider theone-lag autoregressive process:

Xt+1 = aXt + t+1, (25)

where t are the invariants. The limit a ≡ 1 is the random walk (3). If |a| < 1 thedistribution of Xt eventually stabilizes and thus the process becomes stationary.The autocorrelation of this process reads:

Cor {Xt,Xt−τ} = e(ln a)τ , (26)

see Appendix A.1.

For instance, interest rates cannot evolve as random walks at any scale. If(7) held true exactly then for any aggregation size τ the following would holdtrue

Y(υ)t+τ = Y

(υ)t +e(υ)t+τ,τ , (27)

where the aggregate invariant e(υ)t+τ is defined as in (22). However, rates cannotdiffuse indefinitely. Therefore, for aggregation size τ of the order of a month orlarger mean-reverting effects must become apparent. We see this in Figure 5where we plot the fitted value of a as a function of the aggregation size for thetime series of the five-year par swap rate.

11

Page 12: Review of Discrete and Continuous Methods

0 5 10 1 5 20 2 50.9 6

0 .96 5

0.9 7

0 .97 5

0.9 8

0 .98 5

0.9 9

0 .99 5

1

1 .00 5

tim e s te p (da ys )

regr

essi

on c

oeff

icie

nt

Figure 5: AR(1) fit of five-year swap rate

We can generalize (25) by adding more lags of the process and of the in-variant. The so-called autoregressive moving average process, or ARMA(p, q)process is defined as

pYj=1

(1− ajL)Xt = Dt +

qYj=1

(1− bjL) t, (28)

where L denotes the lag operator LXt ≡ Xt−1, and Dt is a deterministic com-ponent that we have added in such a way that we can assume the location oft to be zero. It can be proved that the process Xt stabilizes in the long run,and therefore it is stationary, only if all the a’s in (28) lie inside the unit circle,see Hamilton (1994). The fast, exponential decay (26) of the autocorrelation iscommon to all ARMA(p, q) processes with finite p and q.It is always possible to switch from an ARMA(p, q), to an ARMA(0,∞) or

an ARMA(∞, 0) representation. Indeed, the following identity holds

(1− γL)−1 ≡∞Xk=0

(γL)k , (29)

which can be checked by applying the operator (1− γL) to the right hand side.Therefore, we can remove iteratively all the lags from the MA portion of anARMA(p, q) process, making it ARMA(∞, 0), as long as q < ∞. Similarly, wecan remove all the lags from the AR portion of an ARMA(p, q) process, makingit ARMA(0,∞), as long as p <∞.Although the converse is not true, the ARMA(0,∞) is extremely impor-

tant because of Wold’s theorem, which states that any stationary process canbe expressed as ARMA(0,∞). The infinite parameters of such a process are

12

Page 13: Review of Discrete and Continuous Methods

impossible to estimate, but a parsimonious ARMA(p, q) should in most casesprovide a good approximation.

4 Long memory

time s te p ( days)

time s te p ( days)

auto

corr

elat

ion

auto

corr

elat

ion

lag

lag

1yr swap

5yr swap

Figure 6: Autocorrelation decay properties of swap rate changes

Long memory is a phenomenon whereby the autocorrelation decays slowerthan exponentially, see e.g. Beran (1994).

For instance, in Figure 6 we display the autocorrelation decay of the five-year and the one year swap rates changes. The five year autocorrelation quicklydecays to zero. This is consistent with (27) and comments thereafter. Onthe other hand, the one-year autocorrelation consistently displays a small, yetnon-zero pattern. This pattern is persistent, or else the non-linear aggregationproperties of the sample variance displayed in Figure 7 would not be justified,refer to MATLAB Central File Exchange under the author’s name for the code.

Wold’s theorem states that any stationary process can be represented as anARMA process. However, a parsimonious, finite-lag ARMA specification givesrise to an exponential decay in the autocorrelation and thus it does not yieldlong memory. On the other hand, a generic infinite-lag ARMA specificationdoes not make sense from an estimation perspective because of the excessivenumber of parameters to estimate. Therefore, in order to model long-memory,it becomes necessary to impose a parametric structure on an infinite-lag ARMAprocess.

13

Page 14: Review of Discrete and Continuous Methods

1yr swap

t ime ste p ( days)

5yr swap

sam

ple

varia

nce

sam

ple

varia

nce

time step (da ys)

i.i .d. pr ocess

actua l process

estimation error

i. i.d. process

actua l process

estimation err or

Figure 7: Variance aggregation properties of swap rate changes

This can be achieved by generalizing the random walk (3) into a fractionallyintegrated process. First we twist the invariants as follows

et = (1− L)−d t, (30)

where 0 ≤ d < 1/2; then we define

Xt+1 = Xt +et+1, (31)

If d ≡ 0 in (30) we obtain the standard random walk. If 0 < d < 1/2 we obtain ahighly structured integrated ARMA(0,∞) series. Indeed, the following identityholds

(1− L)−d =∞Xk=0

Γ (k + d)

Γ (k + 1)Γ (d)Lk. (32)

In particular, the autocorrelation of the shocks decays according to a power law

Cor {et,et−τ} ≈ Γ (1− d)

Γ (d)τ2d−1, (33)

see Baillie (1996). Therefore the fractional process (30) suitably describes thelong memory phenomenon. Notice that the autocorrelation decay is reflected ina non-linear, power-law increase of the variance of Xt+τ −Xt as a function ofτ .

14

Page 15: Review of Discrete and Continuous Methods

For instance, the non-linear aggregation properties of the sample varianceof the five-year swap rate changes displayed in Figure 7 are consistent with avalue d ≈ 0, whereas for the one-year swap rate changes we have d ≈ 0.1, referto MATLAB Central File Exchange under the author’s name.

Further generalization of the fractional process (30)-(31) include adding sev-eral fractional lags to both the process and the invariant, similarly to (28).

5 Volatility clusteringVolatility clustering is the phenomenon whereby the scatter of a financial vari-able displays evident autocorrelation, whereas the variable itself does not. Thisis the case among others for the daily returns of most stocks, see e.g. Ghoulmie,Cont, and Nadal (2005) and Cont (2005).

1tε −

2tε

21tε −

locat ion-dispersion ellipsoid

M a r 6 0 S e p 6 5 M a r 7 1 A u g 7 6 F e b 8 2 A u g 8 7 J a n 9 3 J u l 9 8 J a n 0 4 J u l 0 9- 0 . 1 5

- 0 . 1

- 0 . 0 5

0

0 . 0 5

0 . 1

0 . 1 5

0 . 2

( )1ln /t t tP Pε −≡ : daily compounded return of IBM stock

Figure 8: Autocorrelation properties of stocks log-returns

For instance, in the left portion of Figure 8 we display the scatter plot ofthe daily returns of the IBM stock versus their lagged values one day before.The location-dispersion ellipsoid is a circle and thus the log-returns are notautocorrelated.On the other hand, in the right portion of Figure 8 we display the scatter

plot of the squared daily returns of the IBM stock versus their lagged values oneday before. The location-dispersion ellipsoid now is an ellipse tilted across the

15

Page 16: Review of Discrete and Continuous Methods

positive quadrant: the square log-returns, which proxy volatility, are positivelyautocorrelated, refer to MATLAB Central File Exchange under the author’sname.

The generalized representation (14) of the invariants suggests to model suchvolatility clustering in terms of stochastic volatility:

td= µt + σtZt. (34)

In this expression, the location invariant is typically constant µt ≡ µ and Zt is ageneric standardized shape invariant. However, unlike in (14), the scatter termσt displays a non-invariant dynamics. The dynamics of σt can be specified interms of an ARMA(p, q) process, a long memory process, or any other modelthat is not i.i.d. across time.In particular, if the process for σ2t is ARMA(p, q), with the same set of

invariants Zt as in (34), then the process t, which is no longer an invariant, iscalled a generalized autoregressive conditional heteroskedastic, or GARCH(p, q),process. Among the most popular and parsimonious such specifications is theGARCH(1, 1) model:

σ2t = σ2 + aσ2t−1 + bZ2t−1, (35)

which is stationary if 0 < a+ b < 1.

ttX

t

subord in ated proces s

subordina tor

tTtT

.

0 1 2 3 4 5 6 7 8 9 1 0-0 .4

-0 .2

0

0 .2

0 .4s h oc k s

0 1 2 3 4 5 6 7 8 9 1 00

0 .1

0 .2

0 .3

0 .4vo lat ili t y

G ARCH proces s

G ARC H v ola tilit y

tX

2tσ

t

t

Figure 9: GARCH model for volatility clustering

In Figure 9 we display a sample path from the GARCH process, along witha sample path of the volatility (64) that generates it, refer to MATLAB CentralFile Exchange under the author’s name.

16

Page 17: Review of Discrete and Continuous Methods

If the process for σ2t is driven by sources of randomness other than the invari-ants Zt that appear in (34) then we have the most common type of stochasticvolatility models, where the variance is not conditionally deterministic.

Part II

Q: continuous-time processesHere we discuss stochastic processes in continuous time: therefore we assumet ∈ R. In order to support intuition, each section mirrors its discrete-timecounterpart in Part I.

6 Levy processesA Levy process Xt is the continuous-time generalization of the random walkwith drift. Here we highlights the main features; for more on this subject referto Schoutens (2003) and Cont and Tankov (2008).A Levy process is by definition a process such that its increments over any

time interval are invariants, i.e. i.i.d. random variables. Therefore, any invariantdiscussed in Section 2 gives rise to a Levy process, as long as the distributionof the invariant is infinitely divisible. This technical condition is importantbecause it ensures that each invariant can be represented in turns as the sumof invariants. More precisely, for an invariant is infinitely divisible if for anyinteger K the following holds

t ≡ Xt −Xt−1 =³Xt −Xt− 1

K

´+ · · ·+

³Xt−1+ 1

K−Xt−1

´d= Z1 + · · ·+ ZK , (36)

where the Zk’s are i.i.d. This requirements is important to guarantee that wecan consider time intervals of any size in the definition of the invariant.Levy processes are fundamentally of two kinds: Brownian diffusion and Pois-

son jumps. As it turns out, any Levy process can be generated with thesefundamental bricks.

6.1 Diffusion

In order to construct a stochastic diffusion, we need its building blocks, i.e.the increments over any time interval, to be infinitely divisible and to span acontinuum of potential values. In order for the model to be tractable, we needto be able to parametrize the distribution of these increments. Therefore, it isnatural to turn to continuous stable distributions, and in particular to the onestable distribution that displays finite variance: the normal distribution

17

Page 18: Review of Discrete and Continuous Methods

0 0. 1 0 .2 0 .3 0 .4 0 .5 0 . 6 0. 7 0 . 8 0 .9 1- 0. 2

- 0. 1

0

0. 1

0. 2

0. 3

0. 4

0. 5

0. 6

0. 7M er to n ju m p -d iffu s io n

0 0 .1 0 .2 0 .3 0 . 4 0 .5 0 . 6 0. 7 0. 8 0 .9 1- 0. 4

- 0. 3

- 0. 2

- 0. 1

0

0. 1

0. 2

0. 3M er to n ju m p -d iff us io n

0 0. 1 0 .2 0 .3 0 .4 0 .5 0 . 6 0. 7 0 . 8 0 .9 10

2

4

6

8

1 0

1 2

1 4M er to n ju m p -d iffu s io n

0 0 .1 0 .2 0 .3 0 . 4 0 .5 0 . 6 0. 7 0. 8 0 .9 1- 0. 6

- 0. 4

- 0. 2

0

0. 2

0. 4

0. 6

0. 8

1M er to n ju m p -d iff us io n

Brownian motion Poisson process

compound Poisson process Merton jump-diffusion

Figure 10: Sample paths from Levy processes

The Brownian motion with drift Bµ,σ2

t is a continuous, diffusive processwhose invariants are normally distributed with expectation and variance pro-portional to the time lag between the increments:

t,τ ≡ Bµ,σ2

t −Bµ,σ2

t−τ ∼ N¡µτ, σ2τ

¢. (37)

This formulation generalizes (8). Since the sum of independent normal variablesis normal, the increments of the Brownian motion are infinitely divisible as in(36), where each term Zk is normally distributed. Therefore, the Brownianmotion is properly defined and it is normally distributed at any time.

The benchmark process in the Q -measure quantitative finance is the Black-Scholes-Merton geometric Brownian motion, which is used to model, amongmany others financial variables, the price Pt of stocks. This is obtained bymodeling the variable Xt ≡ lnPt as a Brownian motion.

lnPtd≡ Bµ,σ2

t . (38)

Notice that the random walk (5) is much more general, in that the invariants,i.e. the compounded returns, need not be normally distributed. In Figure 10we plot a few paths from the Brownian motion, refer to MATLAB Central FileExchange under the author’s name for the code.

18

Page 19: Review of Discrete and Continuous Methods

6.2 Jumps

In order to construct a stochastic jump, we need its building blocks, i.e. theincrements over any time interval, to be infinitely divisible and to span a discreteset of potential values. In order for the model to be tractable, we need to beable to parametrize the distribution of these increments. From (13) it followsthat the natural choice in this case is the Poisson distribution.The Poisson process P∆,λt jumps by positive multiples of a base (positive or

negative) step ∆ according to a Poisson distribution:

t,τ ≡ P∆,λt − P∆,λt−τ ∼ Po (λτ ;∆) . (39)

Since from (13) the sum of independent Poisson variables has a Poisson distri-bution, the increments of the Poisson process are infinitely divisible as in (36),where each term Zk is Poisson distributed. Therefore, the Poisson process isproperly defined and it is Poisson distributed at any non-integer time.

In 10 we plot a few paths from the standard Poisson process, which takesvalues on the integer grid, refer to MATLAB Central File Exchange under theauthor’s name for the code.

As the time step decreases so does the probability of a jump. Indeed from(39) and (12) we obtain

P { t,τ = 0} ≈ 1− λτ (40)

P { t,τ = ∆} ≈ λτ (41)

P { t,τ > ∆} ≈ 0. (42)

Since E { t,τ} = λτ , the parameter λ plays the role of the intensity of the jumps.

6.3 Generalized representations

The generic Levy process is a sum of the above processes. From (9) and (37),the weighted sum of independent Brownian motions is a Brownian motion:

KXk=1

Bµk,σ

2k

t,kd= Bµ,σ2

t , (43)

where µ ≡PK

k=1 µk and σ2 ≡PK

k=1 σ2k. Also, from (13) and (39), the sum of

independent Poisson processes with the same base step ∆ is a Poisson processwith that base step:

KXk=1

P∆,λktd= P∆,λt , (44)

where λ ≡PK

k=1 λk. Therefore we can construct more general Levy processesfrom our building blocks by considering a continuum of terms in the sum asfollows

Xt = Bµ,σ2

t +

Z +∞

−∞P∆,λ(∆)t d∆, (45)

19

Page 20: Review of Discrete and Continuous Methods

where the intensity λ (∆) determines the relative weight of the discrete Poissonjumps on the grid ∆N.As it turns out, this representation is exhaustive: any Levy process, stem-

ming from slicing as in (36) an arbitrarily distributed infinitely divisible in-variant, can be represented as (45)2 . As we show in Appendix A.3, from (45)follows the Levy-Khintchine representation of Levy processes in terms of theircharacteristic function:

ln¡E©eiωXt

ª¢= iωµt− 1

2σ2ω2t+ t

Z +∞

−∞

¡eiω∆ − 1

¢λ (∆) d∆. (46)

An alternative representation of the Levy processes follows from a result byMonroe (1978), according to which Levy processes can be written as

Xtd= Bµ,σ2

Tt, (47)

where Bt is a Brownian motion and Tt is another Levy process that neverdecreases, called subordinator . In practice, Tt is a stochastic time that indicatesthe activity of the financial markets. The specific case Tt ≡ t recovers theBrownian motion.

6.4 Notable examples

In addition to (geometric) Brownian motion and Poisson processes, another im-portant class of Levy processes are the α-stable processes by Mandelbrot (1963),see also Samorodnitsky and Taqqu (1994), Rachev (2003), whose distributionis closed under the sum. However, in order not to contradict the central limittheorem these processes have infinite variance.Another important subclass are the compound Poisson processes, which gen-

eralize Poisson processes by allowing jumps to take on random values, insteadof values in a fixed grid ∆N. The compound Poisson processes can be expressedas

Xt = Bµ,σ2

t +

PλtX

n=1

Zn, (48)

where Pλt is a Poisson process with values in the unit-step grid N and Zn are

arbitrary i.i.d. variables. As we show in Appendix A.4, the parameterization(45) for jump-diffusion models (48) is given by

λ (∆) ≡ λfZ (∆) , (49)

were fZ is the pdf of Zn.

2This is a slightly abridged version that avoids the problem of infinite activity near thezero

20

Page 21: Review of Discrete and Continuous Methods

0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1-1

- 0.5

0

0.5double ex ponentia l

0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0. 8 0.9 1-1

-0.5

0

0.5normal-inv erse-G aus s ian

0 0.1 0.2 0.3 0.4 0.5 0. 6 0.7 0.8 0.9 1-0.6

-0.5

-0.4

-0.3

-0.2

-0.1

0

0.1

0.2variance gam ma

variance-gammaNIGKou

Figure 11: Sample paths from Levy processes

A notable example in the class of compound Poisson processes is the the-jump diffusion by Merton (1976), where Zn is normally distributed. We plotin Figure 10 a few paths sampled from this process, refer to MATLAB CentralFile Exchange under the author’s name.

Another tractable jump-diffusion process is the double-exponential by Kou(2002). Other notable parametric Levy processes include the Normal-Inverse-Gamma by Barndorff-Nielsen (1998), the variance-gamma process by Madanand Milne (1991) and the CGMY by Carr, Geman, Madan, and Yor (2003), seeFigure 11.

7 Autocorrelated processesThe continuous-time version of the AR(1) process (25) is the Ornstein-Uhlenbeckprocess. Its dynamics is defined by a stochastic differential equation

dXt = −θ (Xt − µ) dt+ dLt, (50)

where µ and σ > 0 are constants and L is a Levy process. This process isdefined for θ ≥ 0 and is stationary if θ > 0. This dynamics is easy to interpret:a change in the process is due to a deterministic component, which tends to pullthe process back towards the equilibrium asymptote µ at an exponential rate,and a random shock dLt that perturbs this exponential convergence.Of particular importance is the case where the Levy driver is a Brownian

motion:dXt = −θ (Xt − µ) dt+ σdBt. (51)

As we show in Appendix A.1 this dynamics can be integrated explicitly:

Xtd=¡1− e−θτ

¢µ+ e−θτXt−τ + ηt,τ , (52)

where ηt,τ are normally distributed invariants

ηt,τ ∼ Nµ0,σ2

¡1− e−2θτ

¢¶. (53)

21

Page 22: Review of Discrete and Continuous Methods

Then the long-term autocorrelation reads:

Cor {Xt,Xt+τ} = e−θτ . (54)

A comparison of (25) with (52) yields the correspondence between discrete andcontinuous time parameters.Notice that for small time steps τ the process (51) behaves approximately as

a Brownian motion. Indeed, a Taylor expansion of the terms in (52)-(53) yields

Xt ≈ Xt−τ + t,τ , (55)

where t,τ is in the form of the normal invariant in (37). On the other hand, asthe step τ increases, the distribution of the process stabilizes to its stationarybehavior

Xt ∼ Nµµ,

σ2

¶. (56)

The AR(1) process for the five year swap rate in Figure 5 can be seen as thediscrete-time sampling of an Ornstein-Uhlenbeck process, which in the contextof interest rate modeling is known as the Vasicek model , see Vasicek (1977).

Another useful autocorrelated process is obtained by applying Ito’s rule tothe square Yt ≡ X2

t of the Ornstein-Uhlenbeck process (51) for m ≡ 0. As weshow in Appendix A.2, this yields the CIR process by Cox, Ingersoll, and Ross(1985):

dYt = −eθ (Yt − em) dt+ eσ√Y tdBt, (57)

where the parameters eθ, em and eσ are simple functions of the parameters in (51).This process, is useful to model the evolution of positive random variables.

For instance, the CIR dynamics provides an alternative to the Vasicek modelfor interest rates. Furthermore, the CIR process can model stochastic volatility,see Section 9.1.

The same way as by adding lags to the one-lag autoregression (25) we obtainthe ARMA processes (28), so it is possible to add differentials to the Ornstein-Uhlenbeck process (50): the result are the so-called continuous autoregressivemoving average, or CARMA, processes.

8 Long memoryAs discussed in Section 4, some financial variables display long memory: theempirical autocorrelation displays a decay slower than the exponential pattern(54) prescribed by the Ornstein-Uhlenbeck or, in discrete time, by finite-lagsARMA processes.

22

Page 23: Review of Discrete and Continuous Methods

The continuous-time version of the fractional process with normal invariantsis the fractional Brownian motion. Similarly to (30)-(32), this process is definedas a structured average of past invariants3

Xt ≡ µt+ σ

Z t

−∞

(t− s)d

Γ (d+ 1)dBs, (58)

where 0 ≤ d < 1/2. The increments of this process

Xt = Xt−τ +et,τ , (59)

are normally distributed:

et,τ ∼ N ¡µτ, σ2τ2H¢ , (60)

whereH ≡ d+

1

2(61)

is the Hurst coefficient .

0 0 . 1 0 . 2 0 . 3 0 . 4 0 . 5 0 . 6 0 . 7 0 . 8 0 . 9 1- 2

- 1 . 5

- 1

- 0 . 5

0

0 . 5

1

1 . 5fB m w i t h H = 0 . 8 ( B m : H = 0 . 5 )

t

tX

Figure 12: Sample paths from a fractional Brownian motion

For d ≡ 0 we recover the regular Brownian motion, where the increments areinvariants as in (37). If 0 < d < 1/2, the increments are identically distributedand normal, but they are not independent, and therefore they do not representinvariants.

Figure 12 displays the persistent pattern of a few sample paths of fractionalBrownian motion with Hurst coefficient H = 0.8, refer to MATLAB CentralFile Exchange under the author’s name for the code. For a financial applicationof this process, refer to the one-year swap rate in Figure 7.

3The integral (58) is not defined. This is to be interpreted as a definition of the increments

Xt −Xs ≡ limτ→−∞

(t

τ

(t− u)d

Γ (d+ 1)dBu −

s

τ

(t− u)d

Γ (d+ 1)dBu).

23

Page 24: Review of Discrete and Continuous Methods

Unlike for the Ornstein-Uhlenbeck process (54), the differences of the frac-tional Brownian motion display a power law decay:

Cor {∆Xt,∆Xt+τ} ≈ d (2d+ 1) τ2d−1, (62)

see Dieker (2004), Comte and Renault (1996), Baillie (1996).

9 Volatility clusteringTo model volatility clustering in continuous time, we can proceed in two ways.

9.1 Stochastic volatility

To generate volatility clustering in discrete time we modified the stochasticvolatility representation of the invariants (14) by imposing autocorrelation onthe scatter as in (34).Similarly, in continuous time first we represent the dynamics of the cumula-

tive invariants, i.e. the generic Levy process (45) as follows:

Xt ≡ µt+ σZt. (63)

In this expression the first term is the location of Xt, which must grow linearly;Zt is a zero-location Levy process whose scatter at time t ≡ 1 is one; and σis a positive constant. Then we modify the (trivial) Levy process for σ into astationary process σt that displays autocorrelation: in this situation Xt is nolonger a Levy process and volatility clustering follows.The most popular such model is the Heston model ,see Heston (1993): the

normalized Levy process Zt in (63) is a Brownian motion; the scatter parameterfollows a CIR process(57) shocked by a different Brownian motion

dσ2t = −κ¡σ2t − σ2

¢dt+ λ

qσ2tdBt, (64)

where2κσ2 > λ2; (65)

the copula between Bt and Zt is normal at all times with constant correlationρ; and ρ < 0 due to the leverage effect : lower returns tend to correspond tohigher volatility. The Heston model is heavily used by Q-finance quants to priceoptions. Indeed, the original Black-Scholes-Merton model (38) is inconsistentwith the observed smiles, skews and smirks in the implied volatility profile.

In Figure 13 we display a sample path from the Heston stochastic volatilityprocess, along with a sample path of the volatility (64) that generates it, referto MATLAB Central File Exchange under the author’s name for the code

24

Page 25: Review of Discrete and Continuous Methods

0 2 4 6 8 10 1 2 1 4 16 18 2 0- 0.0 5

0

0.0 5

0 .1

0.1 5H e s to n p ro c e s s

0 2 4 6 8 10 1 2 1 4 16 18 2 00

0 .2

0 .4

0 .6

0 .8C IR vola t il ity

ttX

t

H eston pro ce ss

C IR v olat ilit y

2tσ

Figure 13: Heston stochastic volatility process

9.2 Subordination

An alternative way to model volatility clustering in continuous time is inspiredby the self-similarity of the Brownian motion:

Bσ2td= σBt. (66)

Consider the subordination (47) of a Levy processes:

Xtd= Bµ,σ2

Tt. (67)

If the subordinator, i.e. the stochastic time Tt increases faster than linearly,than the volatility increases: periods of large volatility correspond to periodswhere the pace of the market increases. In order to generate volatility clustering,i.e. autocorrelation in volatility, we simply relax the assumption that the paceof the market Tt is a Levy process.Since Tt must be increasing, it can be defined as the integral of a process Ts

which is positive at all times

Tt ≡Z t

0

Tsds. (68)

A natural choice for Ts is the CIR process (57) and a natural choice for theequilibrium value for Ts is one:

dTt = −eθ ³Tt − 1´ dt+ eσqTtdZt. (69)

25

Page 26: Review of Discrete and Continuous Methods

0 1 2 3 4 5 6 7 8 9 1 0-1

0

1

2

3

4C IR -s u bo rd ina te d p roc es s

0 1 2 3 4 5 6 7 8 9 1 00

5

1 0

1 5

ttX

t

subord in ated proces s

subordina tor

tTtT

.

Figure 14: CIR-driven subordinated Brownian motion

In order to further increase the flexibility of the subordination approach tostochastic volatility, one can subordinate a generic Levy process instead of theBrownian motion in (67).

In Figure 14 we display a sample path from the subordinated Brownianmotion, along with a sample path of the subordinator (68) that generates it,as well as the CIR process that generates the subordinator, refer to MATLABCentral File Exchange under the author’s name for the code. Notice that periodsof higher volatility correspond to period where the stochastic time elapses faster

26

Page 27: Review of Discrete and Continuous Methods

ReferencesBaillie, R.T., 1996, Long memory processes and fractional integration in econo-metrics, Journal of Econometrics 73, 5—59.

Barndorff-Nielsen, O.E., 1998, Processes of normal inverse Gaussian type, Fi-nance and Stochastics 2, 41—68.

Beran, J., 1994, Statistics for Long-Memory Processes (Chapman and Hall).

Carr, P., H. Geman, D. H. Madan, and M. Yor, 2003, Stochastic volatility forLevy processes, Mathematical Finance 13, 345—382.

Comte, F., and E. Renault, 1996, Long memory continuous time models, Journalof Econometrics 73, 101—149.

Cont, R., 2005, Volatility clustering in financial markets: Empirical facts andagent-based models, in A. Kirman, and G. Teyssiere, ed.: Long Memory inEconomics. Springer.

, and P. Tankov, 2008, Financial Modelling with Jump Processes (Chap-man and Hall/CRC) 2nd edn.

Cox, J. C., J. E. Ingersoll, and S. A. Ross, 1985, A theory of the term structureof interest rates, Econometrica 53, 385—407.

Dieker, T., 2004, Simulation of fractional Brownian motion, Master’s Thesis.

Ghoulmie, F., R. Cont, and J.P. Nadal, 2005, Heterogeneity and feedback inan agent-based market model, Journal of Physics: Condensed Matter 17,S1259—S1268.

Hamilton, J. D., 1994, Time Series Analysis (Princeton University Press).

Heston, S. L., 1993, Closed-form solution for options with stochastic volatilitywith applications to bond and currency options, The Review of FinancialStudies 6, 327—343.

Kou, S. G., 2002, A jump-diffusion model for option pricing, Management Sci-ence 48, 1086—1101.

Madan, D., and F. Milne, 1991, Option pricing with VG martingale components,Mathematical Finance 1, 39—56.

Mandelbrot, B., 1963, The variation of certain speculative prices, Journal ofBusiness 36, 394—419.

Merton, R. C., 1976, Option pricing when underlying stocks are discontinuous,Journal of Financial Economics 3, 125—144.

Meucci, A., 2005, Risk and Asset Allocation (Springer).

27

Page 28: Review of Discrete and Continuous Methods

, 2011, "p" versus "q": Differences and commonalities between the twoareas of quantitative finance, GARP Risk Professional - "The Quant Class-room by Attilio Meucci" February, 43—44 Available as "Quant Nugget 7" athttp://ssrn.com/abstract=1717163.

Monroe, I., 1978, Processes that can be imbedded in Brownian motion, TheAnnals of Probability 6, 42—56.

Rachev, S. T., 2003, Handbook of Heavy Tailed Distributions in Finance(Elsevier/North-Holland).

Samorodnitsky, G., and M. S. Taqqu, 1994, Stable Non-Gaussian RandomProcesses: Stochastic Models with Infinite Variance (Chapman and Hall).

Schoutens, W., 2003, Levy Processes in Finance (Wiley.).

Vasicek, O., 1977, An equilibrium characterisation of the term structure, Journalof Financial Economics 5, 177—188.

Vicente, R., C. M. de Toledo, Leite V. B. P., and N. Caticha, 2005, Underlyingdynamics of typical fluctuations of an emerging market price index: TheHeston model from minutes to months, arXiv:physics/0506101v1 [physics.soc-ph].

28

Page 29: Review of Discrete and Continuous Methods

IndexARMA process, 11autoregressive process, 11

Bernoulli distribution, 7Black-Scholes-Merton, 18Brownian motion, 7, 11, 18

CARMA process, 22Cauchy distribution, 10, 11central limit theorem, 11, 20CIR process

and Ornstein-Uhlenbeck process,22

Heston stochastic volatility, 24subordination, 25

compound Poisson processes, 20compounded return, 6, 7Cox-Ingersoll-Roll process, see see CIR

diffusion, 17distributions

Bernoulli, 7Cauchy, 10, 11infinitely divisible, 17normal, 7, 17Poisson, 8, 19stable, 10, 17Student t, 8, 9

fractional Brownian motion, 23fundamental theorem of pricing, 3

GARCH, 16geometric Brownian motion, 18Greeks, 4

heavy tails, 10Heston model, 24Hurst coefficient, 23

infinitely divisible, 17intensity, 19invariants, 5

jumps, 19

leverage effect, 24Levy process, 17Levy-Khintchine, 20Long memory

continuous time, 22discrete time, 13

martingale, 3mixture model, 9

normal distribution, 7, 17

Ornstein-Uhlenbeck process, 21

Poisson distribution, 8, 19Poisson process, 19

random walk, 5regime switch, 9risk-neutral, 3

semimartingale, 4square root rule, 11stable distribution, 10, 17stable processes, 20stochastic time, 20, 25stochastic volatility, 24

i.i.d., 8Student t distribution, 8, 9subordination, subordinator, 20, 25

Vasicek model, 22volatility clustering

continuous time, 24discrete time, 15

volatility smile, 24

Wold’s theorem, 12

29

Page 30: Review of Discrete and Continuous Methods

A AppendixIn this appendix we present proofs, results and details that can be skipped atfirst reading.

A.1 The Ornstein-Uhlenbeck process

The Ornstein-Uhlenbeck process is defined by the dynamics

dXt = −θ (Xt − µ) dt+ σdBt. (70)

We introduce the integrator

Yt ≡ eθt (Xt − µ) . (71)

Then from Ito’s lemma

dYt = θeθt (Xt − µ) dt+ eθt (−θ (Xt − µ) dt+ σdBt) (72)

= σeθtdBt

Integrating we obtain

Yt = Y0 + σ

Z t

0

eθsdBs (73)

which from (71) becomes

Xt =¡1− e−θt

¢µ+ e−θtX0 + σ

Z t

0

e−θ(t−s)dBs. (74)

ThereforeXt|x0 ∼ N

¡µt|x0, σ2t |x0

¢, (75)

whereµt|x0 ≡

¡1− e−θt

¢µ+ e−θtx0. (76)

To compute σ2t |x0 we use Ito’s isometry

σ2t |x0 = σ2Z t

0

e−2θ(t−s)ds = σ2Z t

0

e−2θudu (77)

= σ2e−2θu

−2θ

¯t0

=σ2

¡1− e−2θt

¢.

30

Page 31: Review of Discrete and Continuous Methods

Similarly for the autocovariance

σ2t,τ |x0 ≡ Cov {Xt,Xt+τ |x0} (78)

= E

½µσe−θt

Z t

0

eθsdBs

¶µσe−θ(t+τ)

Z t+τ

0

eθsdBs

¶¾= σ2e−2θte−θτ

Z t

0

e2θsds

= σ2e−2θte−θτe2θu

¯t0

=σ2

2θe−θτ

¡1− e−2θt

¢Therefore the autocorrelation reads

Cor {Xt,Xt+τ |x0} =σ2t,τ |x0p

σ2t |x0qσ2t−τ |x0

(79)

= e−θτ

s1− e−2θt

1− e−2θ(t+τ),

which in the limit t→∞ becomes the unconditional autocorrelation

Cor {Xt,Xt+τ} = e−θτ . (80)

A.2 Relationship between CIR and OU processes

Consider the OU process with null unconditional expectation

dXt = −θXtdt+ σdBt. (81)

Using Ito’s ruled¡X2t

¢= 2XtdXt + (dXt)

2 (82)

we obtain

dYt = 2pYt

³−θpYtdt+ σdBt

´+ σ2dt (83)

=¡−2θYt + σ2

¢dt+ 2

pYtσdBt,

which is a specific instance of the general CIR process (57).More in general, following Vicente, de Toledo, P., and Caticha (2005), we

consider several independent OU processes

dXj,t = −θXj,tdt+ σdBj,t, (84)

whose square evolves as follows:

d (Xj,t)2 = 2Xj,t (−θXj,tdt+ σdBj,t) + σ2dt (85)

= −2θX2j,tdt+ 2σXj,tdBj,t + σ2dt.

31

Page 32: Review of Discrete and Continuous Methods

Then vt ≡P

Jj=1X

2j,t evolves as a CIR process

dvt =¡Jσ2 − 2θvt

¢dt+ 2σ

√vtdBt, (86)

where in the last step we used the isometry

JXj=1

Xj,tdBj,td=

vuut JXj=1

X2j,tdBt. (87)

A.3 The Levy-Khintchine representation of a Levy process

First we compute the characteristic function of the Poisson process (39), whichfollows from (12):

φt(ω) ≡ E

©eiω t

ª=∞Xk=0

(λt)k

k!e−(λt)eiωk∆ (88)

= e−(λt)∞Xk=0

¡λteiω∆

¢kk!

= e−λteλteiω∆

= eλt(eiω∆−1).

Now the generic representation of a Levy process (45), which we report here:

Xt = Bµ,σ2

t +

Z +∞

−∞P∆,λ(∆)t d∆. (89)

Its characteristic function reads

E©eiωXt

ª= E

½eiω µt+σBt+

+∞−∞ P

∆,λ(∆)t d∆

¾(90)

Approximating the integral with sums we obtain

ln¡E©eiωXt

ª¢= iωµt− 1

2σ2tω2 + ln

µE

½eiω k P

∆k,λkt

¾¶= iωµt− 1

2σ2tω2 + ln

Yk

EneiωP

∆k,λkt

o(91)

= iωµt− 12σ2tω2 +

Xk

ln eλkt(eiω∆k−1)

= iωµt− 12σ2tω2 + t

Xk

λk¡eiω∆k − 1

¢= iωµt− 1

2σ2tω2 + t

Z +∞

−∞λ (∆)

¡eiω∆ − 1

¢d∆

32

Page 33: Review of Discrete and Continuous Methods

Reverting from the sum to the integral notation we obtain

ln¡E©eiωXt

ª¢=

µiωµ− 1

2σ2ω2 +

Z +∞

−∞λ (∆)

¡eiω∆ − 1

¢d∆

¶t, (92)

where the intensity λ (∆) determines the relative weight of the discrete Poissonjumps on the grid ∆N.

A.4 Representation of compound Poisson process

Consider the general representation (48) of a Poisson process. We do not needto worry about the diffusive portion, so we only consider the jump portion

Xt =

PλtX

n=1

Zn. (93)

The characteristic function reads

E©eiωXt

ª= EP

½EZ

½e

Pλtn=1 iωZn

¾¾(94)

= EP

nφZ (ω)

Pλt

o= EP

neiγP

λt

o¯φZ(ω)=e

= etλ(φZ(ω)−1),

where in the last step we used (88).Suppose now that in (92) we set µ ≡ σ2 ≡ 0 and λ (∆) ≡ λfZ (∆). Then

(92) becomes

ln¡E©eiωXt

ª¢= tλ

µZ +∞

−∞fZ (∆) e

iω∆d∆−Z +∞

−∞fZ (∆) d∆

¶(95)

= tλ (φZ (∆)− 1) ,

which is (94).

33