Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter...

45
Multiple Regression II Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Learning, Inc. Chapter 4 Chapter 4 Multiple Regression Multiple Regression Analysis Analysis (Part 2) (Part 2) Terry Dielman Terry Dielman Applied Regression Applied Regression Analysis: Analysis: A Second Course in Business and A Second Course in Business and Economic Statistics, fourth Economic Statistics, fourth edition edition

description

Multiple Regression II 3 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc Full and Reduced Model Using Separate Regressions  Suppose we wanted to test a subset of the x variables for significance as a group.  We could do this by comparing two models.  The first (Full Model) has K variables in it.  The second (Reduced Model) contains only the L variables that are NOT in our group.

Transcript of Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter...

Page 1: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 11Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

Chapter 4Chapter 4Multiple Regression AnalysisMultiple Regression Analysis

(Part 2)(Part 2)

Terry DielmanTerry DielmanApplied Regression Analysis:Applied Regression Analysis:

A Second Course in Business and A Second Course in Business and Economic Statistics, fourth editionEconomic Statistics, fourth edition

Page 2: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 22Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

4.4 Comparing Two Regression Models4.4 Comparing Two Regression Models

So far we have looked at two types of So far we have looked at two types of hypothesis tests. One was about the hypothesis tests. One was about the overall fit:overall fit:HH00:: 11 = = 22 = …= = …= KK = 0 = 0

The other was about individual terms:The other was about individual terms:HH00:: jj = 0 = 0

HHaa:: jj ≠ 0 ≠ 0

Page 3: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 33Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

4.4.1 Full and Reduced Model Using 4.4.1 Full and Reduced Model Using Separate RegressionsSeparate Regressions

Suppose we wanted to test a subset of the Suppose we wanted to test a subset of the xx variables for significance as a group. variables for significance as a group.

We could do this by comparing two We could do this by comparing two models. models.

The first (Full Model) has The first (Full Model) has KK variables in it. variables in it.

The second (Reduced Model) contains only The second (Reduced Model) contains only the the L L variables that are NOT in our group.variables that are NOT in our group.

Page 4: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 44Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

The Two ModelsThe Two Models

For convenience, let's assume the For convenience, let's assume the group is the last (group is the last (K-LK-L) variables. The ) variables. The Full Model is:Full Model is:

The Reduced Model is just:The Reduced Model is just:

exxxxy KKLLLL 11110

exxy LL 110

Page 5: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 55Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

The Partial F TestThe Partial F Test

We test the group for significance with We test the group for significance with another F test. The hypothesis is:another F test. The hypothesis is:

HH00:: L+1L+1 = = L+2L+2 = …= = …= KK = 0 = 0HHaa: : At least one At least one ≠ 0 ≠ 0

The test is performed by seeing how The test is performed by seeing how much SSE changes between models.much SSE changes between models.

Page 6: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 66Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

The Partial F StatisticThe Partial F Statistic

Let SSELet SSEFF and SSE and SSERR denote the SSE in the denote the SSE in the full and reduced models.full and reduced models.

(SSE(SSERR – SSE – SSEFF) / () / (K – LK – L))F = F = SSESSEF F / (/ (n-K-1n-K-1))

The statistic has (The statistic has (K-LK-L) numerator and ) numerator and (n-(n-K-1)K-1) denominator d.f. denominator d.f.

Page 7: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 77Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

The "Group"The "Group" In many problems the group of In many problems the group of

variables has a natural definition.variables has a natural definition. In later chapters we look at groups In later chapters we look at groups

that provide curvature, measure that provide curvature, measure location and model seasonal location and model seasonal variation.variation.

Here we are just going to look at the Here we are just going to look at the effect of adding two new variables.effect of adding two new variables.

Page 8: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 88Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

Example 4.4 Meddicorp (yet again)Example 4.4 Meddicorp (yet again)

In addition to the variables for In addition to the variables for advertising and bonuses paid, we advertising and bonuses paid, we now consider variables for market now consider variables for market share and competition.share and competition.

xx33 = Meddicorp market share in each area= Meddicorp market share in each area

xx44 = largest competitor's sales in each area = largest competitor's sales in each area

Page 9: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 99Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

The New Regression ModelThe New Regression ModelThe regression equation isSALES = - 594 + 2.51 ADV + 1.91 BONUS + 2.65 MKTSHR - 0.121 COMPET

Predictor Coef SE Coef T PConstant -593.5 259.2 -2.29 0.033ADV 2.5131 0.3143 8.00 0.000BONUS 1.9059 0.7424 2.57 0.018MKTSHR 2.651 4.636 0.57 0.574COMPET -0.1207 0.3718 -0.32 0.749

S = 93.77 R-Sq = 85.9% R-Sq(adj) = 83.1%

Analysis of Variance

Source DF SS MS F PRegression 4 1073119 268280 30.51 0.000Residual Error 20 175855 8793Total 24 1248974

Page 10: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 1010Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

Did We Gain Anything?Did We Gain Anything? The old model had RThe old model had R22 = 85.5% so we = 85.5% so we

gained only .4%.gained only .4%. The t ratios for the two new variables The t ratios for the two new variables

are .57 and -.32.are .57 and -.32. It does not look like we have an It does not look like we have an

improvement, but we really need the improvement, but we really need the F test to be sure.F test to be sure.

Page 11: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 1111Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

The Formal TestThe Formal TestNumerator df =Numerator df = (K-L) = 4-2 = 2 (K-L) = 4-2 = 2Denominator df =Denominator df = (n-K-1) = 20 (n-K-1) = 20

At a 5% level, FAt a 5% level, F2,202,20 = 3.49 = 3.49

HH00:: MKTSHRMKTSHR = = COMPETCOMPET = 0 = 0HHaa: : At least one is At least one is ≠ 0≠ 0

Reject HReject H00 if F > 3.49 if F > 3.49

Page 12: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 1212Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

Things We NeedThings We Need

Full Model: Full Model: (K = 4)(K = 4)SSESSEF F = 175855= 175855(n-K-1(n-K-1) = 20) = 20

Reduced Model: Reduced Model: (L = 2)(L = 2)Analysis of Variance

Source DF SS MS F PRegression 2 1067797 533899 64.83 0.000Residual Error 22 181176 8235Total 24 1248974

SSER

Page 13: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 1313Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

ComputationsComputations

(SSE(SSERR – SSE – SSEFF) / () / (K – LK – L))F = F = SSESSEF F / (/ (n-K-1n-K-1))

(181176 – 175855)/ (4(181176 – 175855)/ (4 – 2 – 2)) = = 175855175855 / (25/ (25-4-1-4-1))

5321/5321/22 = = .3026= = .3026 87938793

Page 14: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 1414Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

4.4.2 Full and Reduced Model Comparisons 4.4.2 Full and Reduced Model Comparisons Using Conditional Sums of SquaresUsing Conditional Sums of Squares

In the standard ANOVA table, SSR shows In the standard ANOVA table, SSR shows the amount of variation explained by all the amount of variation explained by all variables together.variables together.

Alternate forms of the table break SSR Alternate forms of the table break SSR down into components.down into components.

For example, Minitab shows sequential For example, Minitab shows sequential SSR which shows how much SSR increases SSR which shows how much SSR increases as each new term is added.as each new term is added.

Page 15: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 1515Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

Sequential SSR for MeddicorpSequential SSR for MeddicorpS = 93.77 R-Sq = 85.9% R-Sq(adj) = 83.1%

Analysis of Variance

Source DF SS MS F PRegression 4 1073119 268280 30.51 0.000Residual Error 20 175855 8793Total 24 1248974

Source DF Seq SSADV 1 1012408BONUS 1 55389MKTSHR 1 4394COMPET 1 927

Page 16: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 1616Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

Meaning What?Meaning What?1.1. If ADV was added to the model first, SSR would If ADV was added to the model first, SSR would

rise from 0 to 1012408.rise from 0 to 1012408.

2.2. Addition of BONUS would yield a nice increase Addition of BONUS would yield a nice increase of 55389.of 55389.

3.3. If MKTSHR entered third, SSR would rise a paltry If MKTSHR entered third, SSR would rise a paltry 4394.4394.

4.4. Finally, if COMPET came in last, SSR would Finally, if COMPET came in last, SSR would barely budge by 927.barely budge by 927.

Page 17: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 1717Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

ImplicationsImplications This is another way of showing that once you This is another way of showing that once you

account for advertising and bonuses paid, you do account for advertising and bonuses paid, you do not get much more from the last two variables.not get much more from the last two variables.

The last two sequential SSR values add up to The last two sequential SSR values add up to 5321, which was the same as the (SSE5321, which was the same as the (SSERR – SSE – SSEFF) ) quantity computed in the partial F test.quantity computed in the partial F test.

Given that, it is not surprising to learn that the Given that, it is not surprising to learn that the partial F test can be stated in terms of sequential partial F test can be stated in terms of sequential sums of squares.sums of squares.

Page 18: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 1818Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

4.5 Prediction With a Multiple 4.5 Prediction With a Multiple Regression EquationRegression Equation

As in simple regression, we will look at As in simple regression, we will look at two types of computations:two types of computations:

1.1. Estimating the mean Estimating the mean yy that can occur that can occur at a set of at a set of xx values. values.

2.2. Predicting an individual value of Predicting an individual value of y y that that can occur at a set of x values.can occur at a set of x values.

Page 19: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 1919Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

4.5.1 Estimating the Conditional Mean 4.5.1 Estimating the Conditional Mean of of yy Given Given xx11, x, x22, ..., x, ..., xKK

This is our estimate of the point on our This is our estimate of the point on our regression surface that occurs at a regression surface that occurs at a specific set of specific set of xx values. values.

For two For two xx variables, we are estimating: variables, we are estimating:

22110,| 21xxxxy

Page 20: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 2020Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

ComputationsComputations

The point estimate is straightforward, The point estimate is straightforward, just plug in the just plug in the xx values. values.

The difficult part is computing a The difficult part is computing a standard error to use in a confidence standard error to use in a confidence interval. Thankfully, most computer interval. Thankfully, most computer programs can do that.programs can do that.

22110ˆ xbxbbym

Page 21: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 2121Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

4.5.2 Predicting an Individual Value 4.5.2 Predicting an Individual Value of of yy Given Given xx11, x, x22, ..., x, ..., xKK

Now the quantity we are trying to Now the quantity we are trying to estimate is:estimate is:

Our interval will have to account for Our interval will have to account for the extra term ( the extra term ( eei i ) in the equation, ) in the equation, thus will be wider than the interval thus will be wider than the interval for the mean.for the mean.

iiii exxy 22110

Page 22: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 2222Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

Prediction in MinitabPrediction in Minitab

Here we predict sales for a territory Here we predict sales for a territory with 500 units of advertising and 250 with 500 units of advertising and 250 units of bonusunits of bonus

Predicted Values for New Observations

New Obs Fit SE Fit 95.0% CI 95.0% PI1 1184.2 25.2 (1131.8, 1236.6) ( 988.8, 1379.5)

Values of Predictors for New Observations

New Obs ADV BONUS1 500 250

Page 23: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 2323Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

InterpretationsInterpretations

We are 95% sure that the We are 95% sure that the averageaverage sales in sales in territories with $50,000 advertising and territories with $50,000 advertising and $25,000 of bonuses will be between $25,000 of bonuses will be between $1,131,800 and $1,236,600.$1,131,800 and $1,236,600.

We are 95% sure that any We are 95% sure that any individualindividual territory with this level of advertising and territory with this level of advertising and bonuses will have between $988,800 and bonuses will have between $988,800 and $1,379,500 of sales$1,379,500 of sales

Page 24: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 2424Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

4.6 Multicollinearity: A Potential Problem4.6 Multicollinearity: A Potential Problem in Multiple Regression in Multiple Regression

In multiple regression, we like the In multiple regression, we like the xx variables to variables to be highly correlated with be highly correlated with yy because this implies because this implies good prediction ability.good prediction ability.

If the If the xx variables are highly correlated among variables are highly correlated among themselves, however, much of this prediction themselves, however, much of this prediction ability is redundant.ability is redundant.

Sometimes this redundancy is so severe that it Sometimes this redundancy is so severe that it causes some instability in the coefficient causes some instability in the coefficient estimation. When that happens we say estimation. When that happens we say multicollinearitymulticollinearity has occurred. has occurred.

Page 25: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 2525Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

4.6.1 Consequences of Multicollinearity4.6.1 Consequences of Multicollinearity

1.1. The standard errors of the The standard errors of the bbjj are larger are larger than they should be. This could cause all than they should be. This could cause all the the tt statistics to be near 0 even though statistics to be near 0 even though the the FF is large. is large.

2.2. It is hard to get good estimates of the It is hard to get good estimates of the jj. . The The bbjj may have the wrong sign. They may have the wrong sign. They may have large changes in value if may have large changes in value if another variable is dropped from or another variable is dropped from or added to the regression.added to the regression.

Page 26: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 2626Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

4.6.2 Detecting Multicollinearity4.6.2 Detecting Multicollinearity

Several methods appear in the Several methods appear in the literature. Some of these are:literature. Some of these are:

1.1. Examining pairwise correlationsExamining pairwise correlations2.2. Seeing large Seeing large FF but small but small t t ratiosratios3.3. Computing Variance Inflation FactorsComputing Variance Inflation Factors

Page 27: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 2727Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

Examining Pairwise Examining Pairwise CorrelationsCorrelations

If it is only a If it is only a collinearitycollinearity problem, you can detect problem, you can detect it by examining the correlations for pairs of it by examining the correlations for pairs of xx values.values.

How large the correlation needs to be before it How large the correlation needs to be before it suggests a problem is debatable. One rule of suggests a problem is debatable. One rule of thumb is .5, another is the maximum correlation thumb is .5, another is the maximum correlation between between yy and the various and the various xx values. values.

The major limitation of this is that it will not help The major limitation of this is that it will not help if there is a linear relationship involving several if there is a linear relationship involving several xx values, for example, values, for example, xx11 = 2= 2xx22 - .07 - .07xx33 + a small random error + a small random error

Page 28: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 2828Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

Large Large FF, Small , Small tt With a significant With a significant FF statistic you would expect to statistic you would expect to

see at least one significant predictor, but that see at least one significant predictor, but that may not happen if all the variables are fighting may not happen if all the variables are fighting each other for significance.each other for significance.

This method of detection may not work if there This method of detection may not work if there are, say, six good predictors but the are, say, six good predictors but the multicollinearity only involves four of them.multicollinearity only involves four of them.

This method also may not help identify what This method also may not help identify what variables are involved.variables are involved.

Page 29: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 2929Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

Variance Inflation FactorsVariance Inflation Factors This is probably the most reliable method This is probably the most reliable method

for detection because it both shows the for detection because it both shows the problem exists and what variables are problem exists and what variables are involved.involved.

We can compute a VIF for each variable. A We can compute a VIF for each variable. A high VIF is an indication that the variable's high VIF is an indication that the variable's standard error is "inflated" by its standard error is "inflated" by its relationship to the other relationship to the other xx variables. variables.

Page 30: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 3030Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

Auxiliary RegressionsAuxiliary Regressions

Suppose we regressed each Suppose we regressed each xx value, in value, in turn, on all of the other turn, on all of the other xx variables. variables.

Let RLet Rjj22 denote the model's R denote the model's R22 we get we get

when when xxjj was the "temporary was the "temporary y".y".

11The variable's VIF is VIFThe variable's VIF is VIFjj = = 1 - R1 - Rjj

22

Page 31: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 3131Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

VIFVIFjj and R and Rjj2 2

If If xxjj was totally was totally uncorrelated with uncorrelated with the other the other xx variables, its VIF variables, its VIF would be 1.would be 1.

This table shows This table shows some other some other values.values.

RRjj22 VIFVIFjj

0%0% 11

50%50% 22

80%80% 55

90%90% 1010

99%99% 100100

Page 32: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 3232Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

Auxiliary Regressions: A Lot of Work?Auxiliary Regressions: A Lot of Work?

If there were a large number of If there were a large number of xx variables variables in the model, obtaining the auxiliaries in the model, obtaining the auxiliaries would be tedious.would be tedious.

Most statistics package will compute the Most statistics package will compute the VIF statistics for you and report them with VIF statistics for you and report them with the coefficient output.the coefficient output.

You can then do the auxiliary regressions, You can then do the auxiliary regressions, if needed, for the variables with high VIF.if needed, for the variables with high VIF.

Page 33: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 3333Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

Using VIFsUsing VIFs A general rule is that any VIF > 10 is a problem.A general rule is that any VIF > 10 is a problem.

Another is that if the average VIF is considerably Another is that if the average VIF is considerably larger than 1, SSE may be inflated.larger than 1, SSE may be inflated.

The average VIF indicates how many times larger The average VIF indicates how many times larger SSE is due to multicollinearity than if the SSE is due to multicollinearity than if the predictors were uncorrelated.predictors were uncorrelated.

Freund and Wilson suggest comparing the VIF to Freund and Wilson suggest comparing the VIF to 1/(1-R1/(1-R22) for the main model. If the VIF are less ) for the main model. If the VIF are less than this, multicollinearity is not a problem.than this, multicollinearity is not a problem.

Page 34: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 3434Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

Our ExampleOur Example

Pairwise correlationsPairwise correlations

The maximum correlation among the x The maximum correlation among the x variables is .452 so if multicollinearity variables is .452 so if multicollinearity exists it is well hidden.exists it is well hidden.

Correlations: SALES, ADV, BONUS, MKTSHR, COMPET

SALES ADV BONUS MKTSHRADV 0.900BONUS 0.568 0.419MKTSHR 0.023 -0.020 -0.085COMPET 0.377 0.452 0.229 -0.287

Page 35: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 3535Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

VIFs in MinitabVIFs in Minitab

The regression equation isSALES = - 594 + 2.51 ADV + 1.91 BONUS + 2.65 MKTSHR - .121 COMPET

Predictor Coef SE Coef T P VIFConstant -593.5 259.2 -2.29 0.033ADV 2.5131 0.3143 8.00 0.000 1.5BONUS 1.9059 0.7424 2.57 0.018 1.2MKTSHR 2.651 4.636 0.57 0.574 1.1COMPET -0.1207 0.3718 -0.32 0.749 1.4

S = 93.77 R-Sq = 85.9% R-Sq(adj) = 83.1%

No Problem!

Page 36: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 3636Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

4.6.3 Correction for Multicollinearity4.6.3 Correction for Multicollinearity

One solution would be to leave out one or One solution would be to leave out one or more of the redundant predictors.more of the redundant predictors.

Another would be to use the variables Another would be to use the variables differently. If differently. If xx11 and and xx22 are collinear, you are collinear, you might try using might try using xx11 and the ratio and the ratio xx22// xx11 instead. instead.

Finally, there are specialized statistical Finally, there are specialized statistical procedures that can be used in place of procedures that can be used in place of ordinary least squares.ordinary least squares.

Page 37: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 3737Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

4.7 Lagged Variables as Explanatory 4.7 Lagged Variables as Explanatory Variables in Time-Series RegressionVariables in Time-Series Regression

When using time series data in a regression, the When using time series data in a regression, the relationship between relationship between y y andand x x may bemay be concurrentconcurrent or or xx may serve as a may serve as a leading indicatorleading indicator..

In the latter, a past value of In the latter, a past value of xx appears as a appears as a predictor, either with or without the current value predictor, either with or without the current value of of xx..

An example would be the relationship between An example would be the relationship between housing starts as housing starts as yy and interest rates as and interest rates as xx. When . When rates drop, it is several months before housing rates drop, it is several months before housing starts increase.starts increase.

Page 38: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 3838Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

Lagged VariablesLagged Variables

The effect of advertising on sales is often The effect of advertising on sales is often cumulative so it would not be surprising cumulative so it would not be surprising see it modeled as:see it modeled as:

Here Here xxtt is advertising in the current month is advertising in the current month and the lagged variables and the lagged variables xxt-1t-1 and and xxt-2t-2 represent advertising in the two previous represent advertising in the two previous months.months.

ttttt exxxy 231210

Page 39: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 3939Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

Potential PitfallsPotential Pitfalls If several lags of the same variable are If several lags of the same variable are

used, it could cause multicollinearity if used, it could cause multicollinearity if xxtt was highly was highly autocorrelatedautocorrelated (correlated with (correlated with its own past values).its own past values).

Lagging causes lost data. If Lagging causes lost data. If xxt-2t-2 is is included in the model, the first time it can included in the model, the first time it can be computed is at time period be computed is at time period t t = 3. We = 3. We lose any information in the first two lose any information in the first two observations.observations.

Page 40: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 4040Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

Lagged Lagged yy Values Values Sometimes a past value of Sometimes a past value of yy is used is used

as a predictor as well. A relationship as a predictor as well. A relationship of this type might be:of this type might be:

This implies that this month's sales This implies that this month's sales yytt are related to by two months of are related to by two months of advertising expense advertising expense xxt t andand xxt-1t-1 plus plus last month's sales last month's sales yyt-1t-1..

ttttt exxyy 132110

Page 41: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 4141Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

Example 4.6 Unemployment RateExample 4.6 Unemployment Rate

The file UNEMP4 contains the The file UNEMP4 contains the national unemployment rates national unemployment rates (seasonally-adjusted) from January (seasonally-adjusted) from January 1983 through December 2002.1983 through December 2002.

On the next few slides are a time On the next few slides are a time series plot of the data and regression series plot of the data and regression models employing first and second models employing first and second lags of the rates.lags of the rates.

Page 42: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 4242Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

10.5

9.5

8.5

7.5

6.5

5.5

4.5

3.5

Aug-1999Jun-1995Apr-1991Feb-1987

UN

EM

P

Date/Time

Time Series PlotTime Series PlotAutocorrelation is .97 at lag 1 and .94 at lag 2

Page 43: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 4343Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

Regression With First LagRegression With First LagThe regression equation isUNEMP = 0.153 + 0.971 Unemp1

239 cases used 1 cases contain missing values

Predictor Coef SE Coef T PConstant 0.15319 0.04460 3.44 0.001Unemp1 0.971495 0.007227 134.43 0.000

S = 0.1515 R-Sq = 98.7% R-Sq(adj) = 98.7%

Analysis of Variance

Source DF SS MS F PRegression 1 414.92 414.92 18070.47 0.000Residual Error 237 5.44 0.02Total 238 420.36

High R2 because of autocorrelation

Page 44: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 4444Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

Regression With Two LagsRegression With Two LagsThe regression equation isUNEMP = 0.168 + 0.890 Unemp1 + 0.0784 Unemp2

238 cases used 2 cases contain missing values

Predictor Coef SE Coef T P VIFConstant 0.16764 0.04565 3.67 0.000Unemp1 0.89032 0.06497 13.70 0.000 77.4Unemp2 0.07842 0.06353 1.23 0.218 77.4

S = 0.1514 R-Sq = 98.7% R-Sq(adj) = 98.6%

Analysis of Variance

Source DF SS MS F PRegression 2 395.55 197.77 8630.30 0.000Residual Error 235 5.39 0.02Total 237 400.93

Page 45: Multiple Regression II 1 Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Multiple Regression Analysis (Part 2) Terry Dielman.

Multiple Regression IIMultiple Regression II 4545Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc.

CommentsComments It does not appear that the second lag It does not appear that the second lag

term is needed. Its term is needed. Its tt statistic is 1.23. statistic is 1.23.

Because we got RBecause we got R22 = 98.7% from the = 98.7% from the model with just one term, there was not model with just one term, there was not much variation left for the second lag term much variation left for the second lag term to explain.to explain.

Note that the second model also had a lot Note that the second model also had a lot of multicollinearity.of multicollinearity.