Multivariate Time Series Analysis and Its Applications

44
1 Multivariate Time Series Analysis and Its Applications [Tsay (2005), chapter 8] Insights: Price movements in one market can spread easily and instantly to another market [economic globalization and internet communication] Financial markets are more dependent on each other than ever before, and one must consider them jointly to better understand the dynamic structure of global finance o One market may lead the other market under some circumstances In multiple assets, the dynamic relationships between returns of the assets play an important role in decision making

Transcript of Multivariate Time Series Analysis and Its Applications

Page 1: Multivariate Time Series Analysis and Its Applications

1

Multivariate Time Series Analysis and Its Applications [Tsay (2005), chapter 8]

Insights:

• Price movements in one market can spread easily and instantly to another

market [economic globalization and internet communication]

• Financial markets are more dependent on each other than ever before, and

one must consider them jointly to better understand the dynamic structure of

global finance

o One market may lead the other market under some circumstances

• In multiple assets, the dynamic relationships between returns of the assets

play an important role in decision making

Page 2: Multivariate Time Series Analysis and Its Applications

2

Weak stationarity and cross-correlation matrices

k-dimensional time series:

rt is weakly stationary if its first and second moments are time-invariant. Let its

means vector and covariance matrix be,

Page 3: Multivariate Time Series Analysis and Its Applications

3

Cross-correlation matrices:

.

Correlation coefficient between rit and rjt (concurrent or contemporaneous):

It’s easy to see that ρij(0) = ρji(0), -1≤ ρij(0) ≤ 1, and ρij(0) = 1 for 1 ≤ i, j ≤ k.

So, ρ(0) is a symmetric matrix with unit diagonal elements.

Page 4: Multivariate Time Series Analysis and Its Applications

4

Lag-l cross-covariance matrix:

The (i,j)th element of is the covariance between rit and rj,t-l

Lag-l cross-correlation matrix:

Lag-l correlation coefficient between rit and rj,t-l:

is the lag-l autocorrelation coefficient of rit

Page 5: Multivariate Time Series Analysis and Its Applications

5

Linear dependence of a weakly stationary:

In general, we have,

Page 6: Multivariate Time Series Analysis and Its Applications

6

Sample cross-covariance matrix:

Sample cross-correlation matrix:

where, is a kxk diagonal matrix of the sample standard deviations of the

component series.

Fuller (1976): is consistent but is biased in a finite sample.

Page 7: Multivariate Time Series Analysis and Its Applications

7

The finite sample distribution of é rather complicated partly because of the

presence of conditional heteroscedasticity and high kurtosis.

Tsay (2005) recommends using a proper bootstrap resampling method to obtain

an approximate estimate of the distribution.

See examples 8.1 and 8.2 (Tsay, chapter 8, pp. 343-346)

Using simplifying notation of Tiao and Box (1981):

Page 8: Multivariate Time Series Analysis and Its Applications

8

Multivariate Portmanteau Tests:

Hosking (1980, 1981) and Li and McLeod (1981): The multivariate case of

univariate Ljung-Box statistic Q(m).

Where tr(A) is the trace of the matrix A.

Under null hypothesis and some regularity conditions, 22~)(mkk mQ χ

(asymptotically).

Page 9: Multivariate Time Series Analysis and Its Applications

9

Rewritten:

Note: if it rejects the null hypothesis, then we build a multivariate model for the

series to study the lead-lag relationships between the component series.

See application in pp. 348-349 (Tsay, 2005)

Page 10: Multivariate Time Series Analysis and Its Applications

10

Vector autoregressive models - VAR

VAR(1):

To simplifying, consider the bivariate case (k = 2):

Page 11: Multivariate Time Series Analysis and Its Applications

11

But, the reduced-form model doesn’t show explicitly the concurrent dependence

between the component series.

Make a Cholesky decomposition: , L a lower triangular matrix with

unit diagonal elements, G a diagonal matrix.

Since G is a diagonal matrix, the components of bt are uncorrelated.

Page 12: Multivariate Time Series Analysis and Its Applications

12

Structural form:

,

The kth equation of de model:

Because bkt is uncorrelated with bit, thus this equation shows explicitly the

concurrent linear dependence of rkt on rit.

See example 8.3, pp. 350 (Tsay, 2005)

Page 13: Multivariate Time Series Analysis and Its Applications

13

Stationarity condition and moments of a VAR(1) model:

Rewritten,

where

and so,

conclusions:

1)

2)

3)

� necessary and sufficient conditions for weak stationarity of rt.

Page 14: Multivariate Time Series Analysis and Its Applications

14

Furthermore,

And so,

Page 15: Multivariate Time Series Analysis and Its Applications

15

And the correlation,

Page 16: Multivariate Time Series Analysis and Its Applications

16

VAR(p):

Mean:

Rewritten:

Similarly,

Page 17: Multivariate Time Series Analysis and Its Applications

17

Yule-Walker equation for a CCM (cross-correlation matrix):

It’s possible to rewritten a VAR(p) as a VAR(1) and use the further conclusions.

Page 18: Multivariate Time Series Analysis and Its Applications

18

The results of a VAR(1) model can now be used to derive properties of the

VAR(p) model.

Page 19: Multivariate Time Series Analysis and Its Applications

19

Estimation of a VAR(p) model:

We need to choose the order specification.

Consider the consecutive VAR models estimated by the ordinary least squares:

VAR(1), VAR(2), …, VAR(i), …

For a VAR(i), the residuals

Page 20: Multivariate Time Series Analysis and Its Applications

20

Residual covariance matrix

To specify the order p, one can test the null hypothesis versus

sequentially for l = 1, 2, … using the test statistic

to testing a VAR(i) model versus a VAR(i-1) model.

is asymptotically a chi-squared distribution with k2 degrees of freedom

(see Tiao and Box (1981))

Alternatively, one can use an Akaike information criterion (AIC). In this case,

we have to use the ML estimator of Σ,

Page 21: Multivariate Time Series Analysis and Its Applications

21

The order p will be such that AIC(p) = min AIC(i)

Or using another information criterion,

See example 8.4, pp. 356-358

Page 22: Multivariate Time Series Analysis and Its Applications

22

Forecasting a VAR(p):

1-step ahead:

• forecast at the time origin h:

• forecast error:

• covariance matrix of the forecast error: ΣΣΣΣ

2-step ahead: substitute rh+1 by its forecast

• forecast at the time origin h:

• forecast error:

• covariance matrix of the forecast error:

If rt is weakly stationary, then the l-step ahead forecast rh(l) converges to its

mean µ as the horizon h increases and the covariance matrix of its forecast error

converges to the covariance matrix of rt.

Page 23: Multivariate Time Series Analysis and Its Applications

23

Impulse response function:

MA(∞) representation:

However, the elements of at are often correlated. We need to use a Cholesky

decomposition.

Page 24: Multivariate Time Series Analysis and Its Applications

24

is the impact of on the future observation

Problem: the result depends on the ordering of the components of rt.

Example:

Page 25: Multivariate Time Series Analysis and Its Applications

25

Page 26: Multivariate Time Series Analysis and Its Applications

26

Page 27: Multivariate Time Series Analysis and Its Applications

27

Page 28: Multivariate Time Series Analysis and Its Applications

28

VECTOR MOVING AVERAGE MODELS - VMA(q)

So, the cross-correlation matrix,

Page 29: Multivariate Time Series Analysis and Its Applications

29

Application: VMA(1) � bivariate MA(1) model

We have a finite-memory model, because rt only depends on the current and past

shocks.

The off-diagonal elements of show the dynamic dependence between the

component series.

Results:

Page 30: Multivariate Time Series Analysis and Its Applications

30

Procedure:

1. use the sample cross-correlation matrices to specify the order q for a

VMA(q) model,

2. estimate the specifies model by using either the conditional or exact

likelihood method (the exact method is preferred when the sample size is

not large)

3. the fitted model should be checked for adequacy applying the Qk(m)

statistics to the residual series.

Page 31: Multivariate Time Series Analysis and Its Applications

31

VECTOR ARMA MODELS - VARMA

Here, one of the issues is the identifiability problem.

Examples:

VMA(1) = VAR(1):

VARMA(1,1):

They are identical.

Page 32: Multivariate Time Series Analysis and Its Applications

32

This problem results in a situation similar to the exact multicollinearity in a

regression analysis.

Tio and Tsay (1989), Tsay (1991) proposed methods of structural specification

to overcome the identifiability problem.

Therefore, VAR and VMA models are sufficient in most financial applications

and when VARMA models are used, only lower order models are entertained.

Page 33: Multivariate Time Series Analysis and Its Applications

33

UNIT ROOT NONSTATIONARITY AND COINTEGRATION

Consider the bivariate ARMA(1,1) model:

This is not a weakly stationary model because the two eigenvalues of the AR

coefficient matrix are 0 and 1.

Rewriting,

Premultiplying by,

Page 34: Multivariate Time Series Analysis and Its Applications

34

We obtain,

So, each component is unit-root nonstationary and follows an ARIMA(0,1,1).

Considering the line transformation, L,

Page 35: Multivariate Time Series Analysis and Its Applications

35

Results:

1. y1t and y2t are uncoupled series with concurrent correlation equal to that

between b1t and b2t.

2. y1t follows ARIMA(0,1,1)

3. y2t is a white noise

Page 36: Multivariate Time Series Analysis and Its Applications

36

So, there is only a single unit root in the system, or, in other words, there is a

common trend of x1t and x2t = cointegration.

x1t and x2t is cointegrated if:

(a) both of them are unit-root nonstationary;

(b) they have a linear combination that is a unit-root stationary.

For a k-dimensional unit-root nonstationary time series, cointegration exists if

there are less than k unit roots in the system.

If there are h unit root series, 0<h<k and k-h is the number of cointegration

factors, i.e. the number of different linear combinations that are unit-root

stationary, called cointegration vectors.

At the example: and is a cointegrating vector.

Page 37: Multivariate Time Series Analysis and Its Applications

37

The idea of cointegration is highly relevant in financial study: the prices must

move in unison; otherwise there exists some arbitrage opportunity for investors.

Error-correction form:

Engle and Granger (1987) discuss an error-correction representation for a

cointegrated system.

At the example, let

This is a stationary model because both and are unit-root

stationary.

Page 38: Multivariate Time Series Analysis and Its Applications

38

For a cointegrated VARMA(p,q) model with m cointegrating factors (m<k):

are the cointegration vectors of xt

The model can be estimated by likelihood methods

Page 39: Multivariate Time Series Analysis and Its Applications

39

COINTEGRATED VAR MODELS

k-dimensional VAR(p):

Let the autoregressive polynomial of the

VAR.

Consider xt ~ I(1), i.e. (1-B) xt ~ I(0)

An error-correction model (ECM):

Where and is the error-correction term.

3 cases:

Page 40: Multivariate Time Series Analysis and Its Applications

40

Page 41: Multivariate Time Series Analysis and Its Applications

41

Specifications

Page 42: Multivariate Time Series Analysis and Its Applications

42

Page 43: Multivariate Time Series Analysis and Its Applications

43

Cointegration Test:

Let H(m) be the null hypothesis that the rank of is m.

Mathematically, the rank of is the number of nonzero eigenvalues of .

Trace cointegration test:

Page 44: Multivariate Time Series Analysis and Its Applications

44

Maximum eigenvalue test:

See example 8.6.5, pp.385.