Cross-Sectional Regressions in Event Studies · 2017. 2. 16. · 3 Christie shows [2] is...

54
Cross-Sectional Regressions in Event Studies Jim Musumeci Department of Finance, MOR 107 Bentley University Waltham, MA 02452 [email protected] 781.891.2235 Mark Peterson Department of Finance, Rehn 134A Gordon and Sharon Teel Professor of Finance Southern Illinois University Carbondale, IL 62901-4626 [email protected] 618.453.1426 Current Draft: September, 2015 The authors are grateful to Claude Cicchetti, Marcia Millon Cornett, Dhaval Dave, Otgo Erhemjamts, Atul Gupta, Kartik Raman, Len Rosenthal, Richard Sansing, and Aimee Smith for their helpful comments. The usual disclaimer applies.

Transcript of Cross-Sectional Regressions in Event Studies · 2017. 2. 16. · 3 Christie shows [2] is...

  • Cross-Sectional Regressions in Event Studies

    Jim Musumeci

    Department of Finance, MOR 107

    Bentley University

    Waltham, MA 02452

    [email protected]

    781.891.2235

    Mark Peterson

    Department of Finance, Rehn 134A

    Gordon and Sharon Teel Professor of Finance

    Southern Illinois University

    Carbondale, IL 62901-4626

    [email protected]

    618.453.1426

    Current Draft: September, 2015

    The authors are grateful to Claude Cicchetti, Marcia Millon Cornett, Dhaval Dave, Otgo

    Erhemjamts, Atul Gupta, Kartik Raman, Len Rosenthal, Richard Sansing, and Aimee Smith for

    their helpful comments. The usual disclaimer applies.

    mailto:[email protected]:[email protected]

  • Cross-Sectional Regressions in Event Studies

    Abstract

    Christie [1987] demonstrated that when regressing abnormal returns on various firm

    characteristics, the “correct deflator [for those firm characteristics] is the market value of

    equity at the beginning of the period.” Despite this, many researchers deflate a variable of

    interest by total assets, and then add leverage, a ratio of book equity to market equity (or its

    reciprocal), or other independent variables. We show that such a method regularly produces

    relationships that appear to be statistically significant, but which in fact are spurious and

    attributable only to a mathematical artifact, not to any causal effects.

  • Cross-Sectional Regressions in Event Studies

    “You keep using that method. I do not think it

    measures what you think it measures.”—with

    apologies to Inigo Montoya and William Goldman.

    I. Introduction

    Most modern event studies test not only whether average abnormal return is equal to zero,

    but also how abnormal returns are related to firm characteristics. For example, if the event

    under consideration pertains to the mortgage crisis, it is natural to conjecture that the firm’s

    response to the event is larger when it owns more mortgages. If V denotes the value of the

    presumably affected assets, TA the value of total assets, ME the market value of the firm’s

    equity, D the value of the firm’s debt, and ∆ denotes “change in,” and if everything else is held

    constant, then the fundamental accounting identity Total Assets = Total Liabilities +

    Stockholders’ Equity tells us to expect ∆V = ∆TA = ∆D + ∆ME. Later in the paper we consider the

    implications of risky debt, but for now we assume the firm’s debt is riskless, in which case ∆D =

    0 and so ∆V = ∆ME. It is often difficult to measure ∆V on a regular basis, and so we typically

    assume it is proportional to V, or equivalently that ∆ME = kV for some presumably unknown

    constant k. A typical null hypothesis would be that k = 0; the alternative might be k ≠ 0, k > 0, or

    k < 0, depending on context.

    A direct ordinary least squares (OLS) regression of ∆ME on V is inappropriate for at least a

    couple of reasons. First, a heteroscedasticity problem likely exists, with larger firms having

    larger error terms. While under this condition OLS will still produce unbiased estimates, these

    estimates will no longer have minimum variance within the set of linear estimates, and tests of

    statistical significance will be undependable. Second, both ∆ME and V are generally related to

    firm size, in which case we will find a significant relationship even if we mistakenly choose V to

    be some irrelevant variable that also tends to increase with firm size, i.e., almost any balance

    sheet or income statement item. To eliminate these problems, we typically scale ∆ME and V by

    some variable related to size before we estimate our regression. Christie [1987] showed that

    the dependent variable, abnormal return, is essentially a measure of ∆𝑀𝐸

    𝑀𝐸, and so consistency

    implies we need to scale V by ME as well. Normalizing V instead by Total Assets spreads the

    gain (or loss) from V equally across all the firm’s claimants, which is not what ∆𝑀𝐸

    𝑀𝐸 does. This

    necessarily produces extraneous noise and weaker tests unless there exists a universal constant

    c such that ME = c∙TA for each firm in the sample.

    Christie formally demonstrated that use of normalizing variables other than the market

    value of equity results in misspecification problems, but did not elaborate on the nature or

  • 2

    severity of the misspecification. We demonstrate that the commonly used variables 𝑉

    𝑇𝐴,

    𝐷

    𝑇𝐴, and

    𝐵𝐸

    𝑀𝐸 (or

    𝑀𝐸

    𝐵𝐸) will usually produce misleading statistical significance. Moreover, we show that

    some seemingly reasonable (and some deliberately unreasonable) choices of variables produce

    not only spuriously significant relations, but even contradictory ones.

    II. Modern Practice in Cross-Sectional Event-Study Analysis

    Despite Christie’s admonition, few modern event studies scale their independent variables

    by market equity. To identify the false inferences to which alternative methods can lead, we

    consider a hypothetical unexpected change in tax law that would lead to lower taxes associated

    with greater cash and short-term receivables (Compustat item DATA1; henceforth “cash”). A

    typical hypothesis might be that, ceteris paribus, ∆ME = ∆V = kV for some positive constant k,

    where V in this case denotes cash. We select a random sample of 150 firms1 from the 2012

    Compustat database and initially assume this relation holds exactly for all the firms in the

    sample. In practice, of course, error terms exist, but if an empirical method works poorly under

    perfect conditions with no error terms, then it cannot be expected to perform well when noise

    is present. Nevertheless, after the main points are established, we consider the more realistic

    case in which ∆ME = ∆V does not hold exactly, and we find similar results.

    To make the main point, we assume the firm’s gain in market value is 5% of the cash

    balance, and that this entire increase in value accrues to equityholders. Thus ∆ME = .05V =

    .05Cash, and so dividing both sides by ME gives us

    ∆𝑀𝐸

    𝑀𝐸 =

    .05𝑉

    𝑀𝐸 =

    .05𝐶𝑎𝑠ℎ

    𝑀𝐸. [1]

    Without loss of generality, we assume that the expected return of each firm conditioned on the

    market return is zero, in which case the abnormal return, AR, exactly equals .05𝐶𝑎𝑠ℎ

    𝑀𝐸. For this

    sample, parameter estimation for the cross-sectional regression Christie shows to be correct,

    AR = + 𝐶𝑎𝑠ℎ

    𝑀𝐸, would produce �̂� = 𝛼 = 0 and �̂� = 𝛽 = .05 by construction.

    A modern researcher unaware that the true state of nature is given by [1] might well

    estimate a cross-sectional regression with commonly used variables, specifically,

    AR = 𝛼 + 𝛽1𝐶𝑎𝑠ℎ

    𝑇𝐴+ 𝛽2

    𝐷

    𝑇𝐴+ 𝛽3

    𝐵𝐸

    𝑀𝐸+ 𝜀. [2]

    1 We excluded financials (SIC Code 6000-6900) and utilities (SIC code 4900-4999) and firms with negative book

    equity, leaving us with 144 firms. The latter exclusion was made because 𝐵𝐸

    𝑀𝐸 is a meaningless statistic when book

    equity is less than zero.

  • 3

    Christie shows [2] is misspecified, yet regressions like this make frequent appearances in the

    literature. One motivation to use them is that they appear to separate the event’s effects into

    component parts. The question we address here is largely empirical in nature: do such

    regressions really explain the components of abnormal return, and how should we interpret the

    results?

    We estimate [2]’s parameters for our sample and find [with t-values in brackets]

    AR = -.009 + .039𝐶𝑎𝑠ℎ

    𝑇𝐴+ .014

    𝐷

    𝑇𝐴+ .010

    𝐵𝐸

    𝑀𝐸. [3]

    [-3.46] [8.14] [3.45] [5.59] Adj R2 = .351

    The regression appears to identify several different firm characteristics that are strongly related

    to AR. Proponents of this technique would presumably infer from the significant t-values that

    firms’ abnormal returns are increasing in each of 𝐶𝑎𝑠ℎ

    𝑇𝐴,

    𝐷

    𝑇𝐴, and

    𝐵𝐸

    𝑀𝐸.

    If no firms in the sample had outstanding preferred stock, then 𝐷

    𝑇𝐴 = 1 −

    𝐵𝐸

    𝑇𝐴. Because

    𝐷

    𝑇𝐴 and

    𝐵𝐸

    𝑇𝐴 would have identical variances and a correlation of -1 with each other, and have correlations

    with other variables that are identical except for sign, substituting 𝐵𝐸

    𝑇𝐴 for

    𝐷

    𝑇𝐴 would change only

    the value and t-statistic of the intercept �̂� and the signs (but not the magnitudes2) of �̂�2 and,

    most importantly, the t-statistic of �̂�2. In general, however, some firms will have outstanding

    preferred stock, and so 𝐵𝐸

    𝑇𝐴 and

    𝐷

    𝑇𝐴 will have a strong negative correlation, but not a perfect one.

    This is sufficient for their tests of significance to be very similar:

    AR = . 004 + .043𝐶𝑎𝑠ℎ

    𝑇𝐴 − .019

    𝐵𝐸

    𝑇𝐴+ .011

    𝐵𝐸

    𝑀𝐸. [4]

    [2.18] [9.02] [-4.83] [6.23] Adj R2 = .396

    Compared with the results of [3], [4]’s larger R2 and t-values for the coefficients seem to

    suggest it is a more powerful and therefore presumably a better test.

    We next check what happens if we alter the form of the model by substituting for each

    independent variable its logarithm instead. Estimating the parameters of that regression gives

    us

    2 If r2 is substituted for r1 as an independent variable, a perfect correlation is necessary and sufficient to ensure that the magnitude of the slope’s t-statistic remains unchanged. In general, however, the coefficient itself may change (for example, changing the units of measure of any independent variable from dollars to thousands of dollars in any regression will result in identical t-values, but coefficients that are different by a factor of 1000). What guarantees in this case that the magnitude of the coefficient also remains unchanged is that the absolute

    value of the coefficient of 𝐷𝑒𝑏𝑡

    𝑇𝐴 in

    𝐵𝐸

    𝑇𝐴= 𝑎 + 𝑏

    𝐷𝑒𝑏𝑡

    𝑇𝐴 is 1.

  • 4

    AR = . 019 + .004 ln (𝐶𝑎𝑠ℎ

    𝑇𝐴) − .006 ln (

    𝐵𝐸

    𝑇𝐴) + .007ln (

    𝐵𝐸

    𝑀𝐸). [5]

    [12.92] [9.88] [-6.62] [7.60] Adj R2 = .449

    Compared with [4], the larger t-values and R2 of [5] would seem to suggest it is a more powerful

    test, and therefore a better one.

    Finally, to see if we can improve on this some more, we take the log of the abnormal

    returns as well.3 This gives us the following result:

    ln (𝐴𝑅) = −2.996 + 1.00 ln (𝐶𝑎𝑠ℎ

    𝑇𝐴) − 1.00ln (

    𝐵𝐸

    𝑇𝐴) + 1.00ln (

    𝐵𝐸

    𝑀𝐸). [6]

    [∞] [∞] [–∞] [∞] Adj R2 = 1.0

    To see how such extreme statistics are possible, we exponentiate e by each side of [6], giving us

    AR = ∆𝑀𝐸

    𝑀𝐸= 𝑒−2.996 (

    𝐶𝑎𝑠ℎ

    𝑇𝐴) (

    𝐵𝐸

    𝑇𝐴)

    −1(

    𝐵𝐸

    𝑀𝐸), or

    ∆𝑀𝐸

    𝑀𝐸= .05 (

    𝐶𝑎𝑠ℎ

    𝑀𝐸), [7]

    which is precisely equation [1]. The regression results from [3] appeared to suggest Total

    Assets and Book Equity were factors that contribute to an explanation of abnormal returns, but

    they are absent from [7]. While it is not necessarily incorrect to essentially add and then

    subtract the effects of Total Assets and Book Equity (as regression [6] basically does), Occam’s

    Razor implies it is inappropriate (and somewhat misleading) to do so. The results of [3]—[6] are

    summarized in Table 1.

    Assuming [7] (or, equivalently, [1]) holds, the coefficients in regression [2] will typically

    show up as significant. The basic principles in play here are that if two candidates for an

    independent variable have a perfect correlation, use of either will produce identical magnitudes

    for that variable’s t-statistic and for the overall R2, and when two variables have a large

    correlation with each other, either will produce very similar t-statistics and R2 values.4 In our

    specific context, all that is required is that the proportion of the firm that is financed with

    preferred stock is relatively constant across firms, and that the correlation between each of the

    independent variables and its logarithm is fairly large. Even this second condition is not

    3 This requires that all the abnormal returns be positive, but by assumption ∆ME = kV > 0, so this condition is met. Similarly, if ∆ME = kV < 0, we can apply the same analysis to the variables -∆ME = -kV > 0. The more realistic case in which AR may have different signs for different firms is addressed after we establish the basic intuition for this deterministic special case. 4 The inverse is not necessarily true, i.e., if two variables have a low correlation with each other, the t-statistics are not necessarily substantially different when one of these variables is substituted for the other. For example, if W1

    and W2 are independent and identically distributed and Y = W1 + W2 + , then the regressions 𝑌 = �̂�1 + �̂�1𝑊1 + 𝜀

    and 𝑌 = �̂�2 + �̂�2𝑊2 + 𝜀 will on average produce similar estimates and similar t-statistics for their respective parameters, despite the fact that W1 and W2 are uncorrelated.

  • 5

    necessary for the t-statistics to be misleading; because the correlation between a variable and

    its logarithm is always positive, the stated t-statistics will always be distorted away from zero.

    It is when this correlation is large, however, that the distortion will be particularly egregious.

    Christie showed that a regression like [7] is the correct specification, and regression [6] is

    mathematically identical to [7]. Working backwards from [6] to [3] simply substitutes for each

    variable another variable that has a high magnitude of correlation with it. Thus the parameters

    of equation [3] all show up as significant not because all the independent variables provide

    explanatory power, but as a mathematical consequence of the true state of nature given by [7].

    While the first and third independent variables, (𝑉

    𝑇𝐴) and (

    𝐵𝐸

    𝑀𝐸), are at least related to the

    correct variable, (𝑉

    𝑀𝐸), the middle independent variable, (

    𝐵𝐸

    𝑇𝐴), is not.5 Any rejection of the null

    hypothesis that its coefficient is equal to zero, then, is a form of Type I error. This component

    of Type I error can be quite large and is in addition to whatever level is chosen for the

    significance level . To see in greater detail why an irrelevant independent variable can have a

    statistically significant coefficient in a multiple regression, we first consider two effects of

    multicollinearity.

    III. Multicollinearity

    The most commonly appreciated aspect of multicollinearity is that it can obscure a true

    relationship between the dependent variable and one or more independent variables, leading

    to less powerful tests. For example, consider a dependent variable Y, independent variables W1

    and W2, and “building blocks” Zi, where the Zi are independent of each other. If

    Y = Z1 + Z2

    W1 = Z1 + Z3

    and W2 = Z1 + Z4,

    then simple regressions of Y on either W1 or W2 are likely to reveal significant coefficients

    provided the sample is sufficiently large or the variances of Z2, Z3, and Z4 are sufficiently small

    relative to that of Z1. A multiple regression of Y on W1 and W2 is problematic, though, because

    while an F-test will reveal that at least one of W1 or W2 is related to Y, it is difficult to ascertain

    which of W1 or W2 is the culprit.6

    5 If (

    𝐵𝐸

    𝑇𝐴) happens to be correlated with (

    𝑉

    𝑀𝐸), it may appear to be significant in a simple regression. However, by

    construction, its significance would disappear in a multivariable regression with both variables. 6 In the belief that orthogonalization solves this multicollinearity problem, some researchers first orthogonalize one of the independent variables, say W1, by regressing it on another, W2, with which it is highly correlated, and

  • 6

    A less-appreciated consequence of multicollinearity is that it may give the appearance of a

    significant relationship, even when none exists. For example, consider

    Y = Z1 + Z2

    W1 = Z1 + Z3

    and W2 = + Z3.

    Now a simple regression of Y on W1 will typically produce a significant coefficient as before, but

    Y is independent of W2 and the results of that simple regression will reveal this. However, W1

    and W2 have a positive correlation because of the common component Z3, and in a multivariate

    regression featuring W1 and W2 as independent variables, W2 will have the effect of “cleaning

    up” the noise induced by the presence of Z3 as a component of W1.7 Specifically, we will find

    that for the parameters of the sample regression line

    �̂�𝑖 = �̂� + �̂�1𝑊1,𝑖 + �̂�2𝑊2,𝑖 + 𝑒𝑖

    �̂�1 → 1 and �̂�2 → -1 as sample size increases. However, W2 is not in any way related to Y, and

    its coefficient appears to be significant not because of any direct relationship with Y, but

    because it counteracts the Z3 noise term in W1.

    The havoc that can be created by the introduction of superfluous independent variables is

    reminiscent of one version of Griliches’ Law: “Any cross-sectional regression with more than

    five variables produces garbage.”8 It is for this reason that estimating a series of simple

    regressions as a complement to any multiple regression is desirable. When a multiple

    regression suggests an independent variable is significant, it is important to know whether the

    significance is due to a direct effect on the dependent variable, or to a reduction of noise in one

    or more other independent variables; simple regressions are not a panacea,9 but they can give

    us some indication of this. Additionally, it is a good idea to have a solid model a priori that is

    then replace W1 in a multiple regression with the residuals e1 from the regression 𝑊1 = �̂� + �̂�𝑊2 + 𝑒1. Mitchell [1991], however, shows that this does not solve the multicollinearity problem at all. The coefficient and t-statistic for e1 in the ensuing multiple regression are identical to what they would have been for W1, while those for W2 are the same as they would have been in a simple regression without W1 or e1. 7 We are hardly the first to make this observation that superfluous independent variables may spuriously appear to be significant if they have this “cleaning up” effect, or even to use that expression. Griliches and Wallace [1965] note the same possibility in their footnote 7. 8 While this is the version attributed to Griliches on p. 28 of McCloskey’s The Writing of Economics, the only version we have been able to track down is “any time series regression containing more than four independent variables results in garbage” in Griliches’ comments on p. 335 of Intriligator and Kendrick [1974]. 9 For example, if Y = Z1 + Z2 + Z3, W1 = Z1, and W2 = Z2, and if the variance of Z1 is large relative to those of the other

    Zi, then W2 does have some explanatory power (through Z2) that might be revealed in a multiple regression, but

    might not be apparent in a simple regression. The reason is that, if the variance Z1 is sufficiently large, then Y =

    W2 + will have a great deal of noise (due to the presence of Z1 in the error term),while Y = 1W12W2

    + provides a more powerful test (because Z1 will be removed from the error term).

  • 7

    used as a basis for including an independent variable, rather than to add variables because they

    might produce better results.

    IV. A Reconsideration of Cross-Sectional Regressions

    There is nothing special about the choice of Total Assets and Book Equity as scaling

    variables in the Section II. We can obtain similar results for most strictly positive firm

    attributes. In this section we examine portfolios with a variety of possible Compustat values in

    addition to Book Equity and Total Assets. For now, we continue to assume that abnormal

    returns are deterministic, specifically, AR = .05𝐶𝑎𝑠ℎ

    𝑀𝐸. For each such pair of Compustat values,

    which we designate X1 and X2, we estimate parameters of the sequence of regressions

    AR = 𝛼 + 𝛽𝑋1

    𝑋2 + 𝜀 [8]

    AR = 𝛼 + 𝛽ln (𝑋1

    𝑋2) + 𝜀 [8a]

    AR = 𝛼 + 𝛽1𝐶𝑎𝑠ℎ

    𝑋1+ 𝛽2

    𝑋1

    𝑋2+ 𝛽3

    𝑋2

    𝑀𝐸 + 𝜀 [9]

    AR = 𝛼 + 𝛽1𝐶𝑎𝑠ℎ

    𝑋1+ 𝛽2ln (

    𝑋1

    𝑋2) + 𝛽3

    𝑋2

    𝑀𝐸 + 𝜀 [10]

    AR = 𝛼 + 𝛽1ln (𝐶𝑎𝑠ℎ

    𝑋1) + 𝛽2ln (

    𝑋1

    𝑋2) + 𝛽3ln (

    𝑋2

    𝑀𝐸) + 𝜀 [11]

    In an effort to find ratios 𝑋1

    𝑋2 that would be unrelated to our portfolios’ dependent variable,

    .05𝐶𝑎𝑠ℎ

    𝑀𝐸, we formed ratios with numerators and denominators taken from thirteen Compustat

    database variables from 1995 to 2014.10 This gave us 156 ratios, and for our entire dataset we

    found the correlation between .05𝐶𝑎𝑠ℎ

    𝑀𝐸 and each ratio

    𝑋1

    𝑋2. We selected the eight ratios that had

    the lowest correlations within the entire database with the expectations that they would also

    have the lowest correlations within the portfolios of 150 firm-years, and so in a simple

    regression might be expected to reject a null hypothesis of zero slope for 𝑋1

    𝑋2 (or its log) at a

    10 The variables (with Compustat Data Item numbers) are Total Receivables (DATA2), Total Inventories (DATA3), Total Current Assets (DATA4), Total Current Liabilities (DATA5), Total Assets (DATA6), Gross PP&E (DATA7), Net Sales (DATA12), Operating Income Before Depreciation (DATA13), Depreciation and Amortization (DATA14), Interest Expense (DATA15), PP&E Capital Expenditures (DATA30), Cost of Goods Sold (DATA41), Accounts Payable (DATA70). In all cases we trimmed from any regression any observation that had a negative component of a ratio.

  • 8

    frequency equal to the significance level.11 Thus we would expect any statistical significance of 𝑋1

    𝑋2 in a multiple regression to be solely attributable to the “cleaning up” effect described in

    Section III.

    Table 2 shows the results of 1000 portfolios of 150 firm-years each. We analyzed each

    portfolio using actual ratios described above (the eight having the lowest correlation with

    Cash/ME, plus X1 = BE, X2 = TA and X1 = TA, X2 = BE12). Each panel shows the progression for

    one ratio from simple regression to a multivariate regression featuring logs of all three

    independent variables.

    For example, Panel A of Table 2 features X1 = Operating Income before Depreciation =

    OIBDP and X2 = Capital Expenditures = CapEx, and is fairly typical. In the simple regression of

    the first row, we reject the null hypothesis that the coefficient of 𝑋1

    𝑋2 is zero 11.2% of the time.

    This is substantially greater than the 5% we were expecting based on the fact that we chose this

    ratio because of its low correlation with 𝐶𝑎𝑠ℎ

    𝑀𝐸 for the full sample from which the portfolios were

    drawn. It is not clear why this is so much greater than 5%, but may be due to the fact that

    while such tests are asymptotically well-specified, they may not be well-specified in samples of

    only 150. Whether we use 𝑂𝐼𝐵𝐷𝑃

    𝐶𝑎𝑝𝐸𝑥 or ln(

    𝑂𝐼𝐵𝐷𝑃

    𝐶𝑎𝑝𝐸𝑥) does not make a large difference in overall

    rejection rates of H0: = 0, except that 𝑂𝐼𝐵𝐷𝑃

    𝐶𝑎𝑝𝐸𝑥’s overall rejection rate of 11.2% featured 11.1%

    with a positive and significant t-statistic, and only .1% with a negative and significant t-statistic,

    while the overall rejection rate of 11.9% for ln(𝑂𝐼𝐵𝐷𝑃

    𝐶𝑎𝑝𝐸𝑥) was more symmetrically distributed, with

    a 5.8% frequency of positive rejections and a 6.1% frequency of negative rejections. Because

    the distortion in the coefficient of 𝑋1

    𝑋2 (in this case,

    𝑂𝐼𝐵𝐷𝑃

    𝐶𝑎𝑝𝐸𝑥) or its log described in Sections II and

    III is positive, we focus mainly on the rejections of H0 due to positive and significant t-statistics

    when we consider the multiple regressions. In the first multiple regression [9] featuring 𝐶𝑎𝑠ℎ

    𝑂𝐼𝐵𝐷𝑃,

    𝑂𝐼𝐵𝐷𝑃

    𝐶𝑎𝑝𝐸𝑥, and

    𝐶𝑎𝑝𝐸𝑥

    𝑀𝐸 as independent variables, we find the positive and significant rejection rate for

    the coefficient of 𝑂𝐼𝐵𝐷𝑃

    𝐶𝑎𝑝𝐸𝑥 has more than doubled (to 23.7% from 11.1%) when compared with

    that of the simple regression [8]. When we proceed to an identical regression [10] except 𝑂𝐼𝐵𝐷𝑃

    𝐶𝑎𝑝𝐸𝑥

    is replaced by ln(𝑂𝐼𝐵𝐷𝑃

    𝐶𝑎𝑝𝐸𝑥), this rejection rate again more than doubles, from 23.7% to 61.0%.

    Finally, when we take the log of each of the three independent variables [11], this rejection rate

    increases from 61.0% to 99.3%, or nearly all the time. Thus, in a fashion analogous to the

    11 Alternatively, we can think of these as the eight ratios with the largest p-values. The smallest p-value of this set of eight was depreciation/total inventory, with a p-value of .8259 and a correlation of only -0.00111. 12 Note that in Section II, we always used TA in the denominator to conform to common usage. Here, to be consistent with the other eight pairs of ratios, we let it appear in numerator or denominator, depending on whether it is X1 or X2.

  • 9

    progression from regression [4] to [5] in section II, and consistent with the framework of

    Section III, we find that a variable that was selected because it had no apparent effect on the

    dependent variable appears to be significant in a multiple regression (with logs) almost all the

    time. As before, this is not because the variable is conveying information on its own, but rather

    because it is “cleaning up” the error created by weak choices for the other two independent

    variables.

    Panel B features the same two variables, but in the opposite order—X1 = Capital

    Expenditures and X2 = Operating Income before Depreciation—and contains another surprising

    result. All five rows feature rejection rates that are fairly similar to those of Panel A. Why is

    this surprising? Because we would expect any dependent variable that is increasing in 𝑂𝐼𝐵𝐷𝑃

    𝐶𝑎𝑝𝐸𝑥

    (or its log) to be decreasing in its reciprocal, 𝐶𝑎𝑝𝐸𝑥

    𝑂𝐼𝐵𝐷𝑃 (or its log). However, that is not what the

    penultimate multivariate regressions of Panels A and B show. When using ln(𝑂𝐼𝐵𝐷𝑃

    𝐶𝑎𝑝𝐸𝑥) as the

    middle independent variable in Panel A, 61% of the time we find a positive and significant t-

    statistic, but for the same portfolios we find in Panel B that the dependent variable is also

    increasing in ln(𝐶𝑎𝑝𝐸𝑥

    𝑂𝐼𝐵𝐷𝑃) and has a positive and significant t-statistic 52.7% of the time. The last

    regressions of Panels A and B—featuring logs of all three ratios used as independent variables—

    are even more damning. Here we have a 99.3% rejection rate suggesting the coefficient of

    ln(𝑂𝐼𝐵𝐷𝑃

    𝐶𝑎𝑝𝐸𝑥) is positive, but also a 99.2% rejection rate suggesting the coefficient of ln(

    𝐶𝑎𝑝𝐸𝑥

    𝑂𝐼𝐵𝐷𝑃) is

    positive. Ln(𝑂𝐼𝐵𝐷𝑃

    𝐶𝑎𝑝𝐸𝑥) = - ln(

    𝐶𝑎𝑝𝐸𝑥

    𝑂𝐼𝐵𝐷𝑃), so clearly if abnormal returns are increasing in one of these

    variables, they must be decreasing in the other. What purports to be information regarding

    those two variables is really indicative of the fact that we made poor choices of independent

    variables in 𝐶𝑎𝑠ℎ

    𝑂𝐼𝐵𝐷𝑃 and

    𝐶𝑎𝑝𝐸𝑥

    𝑀𝐸, and also in

    𝐶𝑎𝑠ℎ

    𝐶𝑎𝑝𝐸𝑥 and

    𝑂𝐼𝐵𝐷𝑃

    𝑀𝐸. It also emphasizes that coefficients

    in multiple regressions have meaning only in context; multicollinearity creates a type of

    entanglement which implies the results do not have any generalizable interpretation, but

    instead have significance only in the context of the specific forms of the other variables.

    Panels C—H in Table 2 are fairly similar, and are summarized in Table 3. Panels I (X1 = Book

    Equity, X2 = Total Assets) and J (X1 = Total Assets, X2 = Book Equity) are a bit different and merit

    extra discussion. First of all we note that these are similar to Panels A and B in that together

    the multiple regressions [9]—[11] suggest the dependent variable is increasing in 𝐵𝑜𝑜𝑘 𝐸𝑞𝑢𝑖𝑡𝑦

    𝑇𝑜𝑡𝑎𝑙 𝐴𝑠𝑠𝑒𝑡𝑠

    and also in 𝑇𝑜𝑡𝑎𝑙 𝐴𝑠𝑠𝑒𝑡𝑠

    𝐵𝑜𝑜𝑘 𝐸𝑞𝑢𝑖𝑡𝑦. As discussed in the last paragraph, this is implausible. Second, the

    22.9% rejection rate for the coefficient of 𝐵𝑜𝑜𝑘 𝐸𝑞𝑢𝑖𝑡𝑦

    𝑇𝑜𝑡𝑎𝑙 𝐴𝑠𝑠𝑒𝑡𝑠 in the simple regression [8], for example,

    suggests that indeed there might be a weak relationship between 𝐶𝑎𝑠ℎ

    𝑀𝑎𝑟𝑘𝑒𝑡 𝐸𝑞𝑢𝑖𝑡𝑦 and

    𝐵𝑜𝑜𝑘 𝐸𝑞𝑢𝑖𝑡𝑦

    𝑇𝑜𝑡𝑎𝑙 𝐴𝑠𝑠𝑒𝑡𝑠.

    The fact that this rejection rate increases to 94.3% in the multivariate regression [9] might seem

  • 10

    to suggest [9] is a more powerful test. Unfortunately, when the significance levels are

    substantially misaligned, as Panels A—H show them to be, it is impossible to draw any

    meaningful inference when the null is rejected. For example, suppose a colleague presents a

    new test that he alleges is substantially more powerful than the generally accepted method.

    The only problem, he acknowledges, is that the test is misspecified—at a significance level of

    = 5%, it actually rejects true null hypotheses at a rate of about 19%. Unbeknownst to you, your

    colleague’s method is simply to calculate the t-statistic the standard way and multiply it by 1.5

    before comparing it with the critical value. This test is obviously more powerful (compared

    with the standard test, it will reject more frequently when the null is false), but the problem is

    that the misspecification (rejecting too frequently when the null is true) makes the test results

    unreliable. So it is with the multiple regressions involving Book Equity and Total Assets in

    Panels I and J. Because the results of Panels A—H demonstrate substantial misspecification, we

    cannot draw any meaningful inference from any results of Panels I and J except that the

    multivariate regressions reject much more frequently than the simple regressions as the

    discussion in Sections II and III demonstrates they would.

    Panel A of Table 3 summarizes the frequency with which the coefficient of 𝑋1

    𝑋2 [or ln(

    𝑋1

    𝑋2)] is

    found to have a significant positive t-statistic for the sequence of regressions [8]—[11].

    Generally we find that the rejection rate is increasing as we move from a simple regression [8]

    to a multiple regression which uses the logs of each of the three independent variables [11].

    These changes in rejection rates are reported in Panel B of Table 3. For example, for all ten

    ratios considered, the average rejection rate for simple regressions (Panel A’s 10.1%) increases

    by 18.0% to 28.1% when we consider a multivariate regression with all three (unlogged)

    variables. This rejection rate increases by another 18.9% when we use the log of 𝑋1

    𝑋2 (but not of

    the other two variables), and increases by an additional 51.7% (to an average rejection rate of

    98.7%) when we take logs of each of the three independent variables.

    Not all pairs of X1 and X2, however, produce increases at the same rates. For example, Table

    3’s Panel B shows that when we move from the simple, unlogged regression [8] to the

    multivariate regression [9] (with no logs of any of the three independent variables), the

    rejection rate for X1 = Book Equity, X2 = Total Assets shows the biggest increase in positive

    rejections at 71.4% (from 22.9% to 94.3%). In contrast, X1 = Accounts Payable, X2 = Total

    Receivables features an increase in rejection rates of only 4.2%. What accounts for this

    difference? Basically, when any independent variable is replaced with another independent

    variable with which it has a perfect correlation, the rejection rates for the slope coefficient will

  • 11

    be identical.13 Continuity implies that using 𝑋1

    𝑋2 or ln(

    𝑋1

    𝑋2) will produce very similar results when

    their correlation is large. We have already seen that the rejection rates when we use logs of all

    three independent variables are consistently near 100%, and average almost 50% when we take

    the log only of 𝑋1

    𝑋2. Because the correlation between

    𝐵𝐸

    𝑇𝐴 and its log is quite large (0.892), almost

    all the bump in rejection rates of H0: 2 = 0 in moving from [8] to [10] occurs in the first stage,

    from [8] to [9]. The same is not true of 𝐴𝑐𝑐𝑜𝑢𝑛𝑡𝑠 𝑃𝑎𝑦𝑎𝑏𝑙𝑒

    𝑇𝑜𝑡𝑎𝑙 𝑅𝑒𝑐𝑒𝑖𝑣𝑎𝑏𝑙𝑒𝑠, which has a correlation with its log of

    only 0.121. For this variable, the total increase in rejection rates of 29.6% from [8] to [10] does

    not come mostly in the move from [8] to [9], but rather in the move from [9] to [10], with its

    increase of 25.4%. Consistent with this, we find for the ten ratios the correlation between

    changes in positive rejection rates one the one hand and the correlation between 𝑋1

    𝑋2 and its log

    on the other is 0.768. Similarly, if the correlation between 𝑋1

    𝑋2 and its log is a driving factor here,

    we should also expect to find that variables with a low correlation get a bigger bump in

    rejection rates as they move from [9] to [10], and indeed we find the correlation between

    changes in rejection rates and the correlation between 𝑋1

    𝑋2 and its log is negative (-0.392).

    The other results in Table 3 are also disturbing. We selected the (X1, X2) pairs based on their

    low correlations with the dependent variable, and so any method suggesting they are

    significant—as do the multivariate regressions [9]-[11]—is seriously flawed. However, if the

    method is flawed, then it can (and does) produce erroneous results even when it is used with

    seemingly meaningful variables such as the firm’s total assets or book equity. Nevertheless,

    this method is commonly seen.

    How does multicollinearity cause these results? For the relationship in question, for

    example, Christie showed that the correct independent variable is 𝑉

    𝑀𝐸, and when we instead

    used 𝑉

    𝑋1 in regression [9] we not only left out the relevant term ME, but also introduced some

    noise in the form of X1. When we added the third variable, 𝑋2

    𝑀𝐸, we restored the relevant term

    ME, but we introduced more noise in the form of X2. Finally, when we added the term 𝑋1

    𝑋2 we

    mitigated both noise terms at once. Thus the term 𝑋1

    𝑋2 serves the same role in regressions [9]—

    [11] that W2 serves in the second example of section III. It does not in any way directly explain

    the dependent variable, but it does reduce the noise that was introduced by the erroneous

    inclusion of X1 and X2 in the first and third independent variables of [9]—[11].

    13 Two variables W1 and W2 will have a perfect correlation if and only if W1,i = a + bW2,i for every observation i, and so substituting one for the other will change the slope coefficient itself by a factor of b; and will result in identical t-statistics for the slopes. The substitution will leave the intercept unchanged if and only if a = 0.

  • 12

    Panels C and D focus on rejections of H0: = 0 due to any significant t-statistic, whether

    positive or negative. The results are quite similar to those of Panels A and B.

    Tables 2 and 3 raise an interesting issue about the correct choice between W and ln(W) as

    an independent variable in regressions. Casual observation suggests many finance researchers

    believe that the reason for using ln(W) is that it will make a skewed distribution of W more

    symmetric. It is true that the distribution of ln(W) is likely to be more symmetric than that of W

    if W is skewed, but symmetry of the independent variable’s distribution is not one of the OLS

    assumptions; the only distributional assumptions made of the independent variable are that it

    has a positive sample variance and a population variance that is finite (e.g., Kmenta (1997), p.

    208). Apart from this, the distributional assumptions required for OLS all pertain to the

    disturbance term 𝜀, not to the distribution of the independent variable itself (e.g., see Kmenta

    or Kennedy (2008), p. 41). Of course, it is possible that the distributions of the independent

    variable and the error term are related, but this is not necessarily so. For example, consider an

    institution that always has a substantial long position in options. Its return relative for the

    options division will be heavily skewed, as will its total return relative if options constitute most

    of its trading. However, there is no reason to suspect the disturbance term in a regression of

    total return relative on the option division’s return relative would be anything but normally

    distributed, and taking the log of the independent variable because it is skewed would be

    inappropriate.

    Econometrics texts are not silent on the issue of log transformations; however, the main

    application has nothing to do with the distribution of an independent variable, but with the

    functional form. If it is additive, then OLS is used; logs are generally recommended only if the

    functional form is multiplicative, e.g., a Cobb-Douglas production function. Because our

    dependent variable is a multiplicative function of the independent variables, it should be no

    surprise that the t-statistics improve (in our case, misleadingly so) as we move from [9] to [10]

    to [11].

    As a final example before proceeding, consider the DuPont formula, ROE = Return on Equity

    = Profit Margin*Total Asset Turnover*Equity Multiplier. As it is, the formula is a tautology, but

    suppose there is some measurement error so that it is only an approximation. If we wanted to

    see the relative contributions of the right-hand side ratios on Return on Equity for a sample

    with positive net incomes, it would be inappropriate to estimate parameters of the regression

    ROE = + 1*Profit Margin + 2*Total Asset Turnover + 3*Equity Multiplier +

    because the relationship is multiplicative, not additive. Instead, it would be more accurate to

    estimate the parameters of

    ln(ROE) = +1*ln(Profit Margin)+2*ln(Total Asset Turnover)+3*ln(Equity Multiplier) + .

  • 13

    The choice of logs has nothing to do with skewness of Profit Margin, Total Asset Turnover, or

    the Equity Multiplier. It is driven by the correct choice of a model, not on the distribution of the

    sample observations. Similarly, even if the context were different, the effect of leverage (more

    accurately, of 1 – leverage, or 𝐵𝐸

    𝑇𝐴) on ROE is known to be multiplicative, so if ROE is the

    dependent variable, ln(𝐵𝐸

    𝑇𝐴) is a better choice for an independent variable in a multivariate

    regression than is simply 𝐵𝐸

    𝑇𝐴.

    Tables 2 and 3 consider a deterministic abnormal return, specifically, AR = ∆𝑀𝐸

    𝑀𝐸 =

    .05𝐶𝑎𝑠ℎ

    𝑀𝐸. In

    practice, of course, abnormal returns are noisy. Consequently, we extend the noiseless

    equation [1] to a more practical example in which an error term is present, specifically,

    AR = ∆𝑀𝐸

    𝑀𝐸 =

    .05𝐶𝑎𝑠ℎ

    𝑀𝐸+ 𝛿. [12]

    Now the left side of [12] can be negative, and we can no longer take the log of both sides as

    we did to get from equation [2] to equation [7]. Nevertheless, the same intuition we developed

    in the deterministic case will apply here. For example, consider any regression (simple or

    multiple) and what happens if for every observation a constant is added to the dependent

    variable. This will affect only the intercept, and will leave the slope(s) unchanged.14 With that

    in mind, we can replace the dependent variable with its return relative, or one plus the

    abnormal return. Because the problems we have identified pertain only to slopes, and because

    the slopes remain unchanged, the difficulties when all abnormal returns are positive will persist

    when some are negative as well.

    The extent of the misspecification problem in the presence of a stochastic error term is an

    empirical issue, and we address it through simulations. To avoid unnecessary complications,

    we simulate 𝛿 as normally distributed with a mean of zero and a standard deviation of 2.5%.15

    We find results that are not as strong as those in Table 2, but which nevertheless exemplify the

    main problem, namely, that variables not involving V or Market Equity may spuriously appear

    to be significant. The correct regression AR = �̂� + �̂�𝐶𝑎𝑠ℎ

    𝑀𝐸+ 𝑒 is very powerful. We found the

    average t-statistic for �̂� to be 7.39, and we rejected H0: = 0 in all but one of the 1000

    14 The same is true if we add any constant(s) to one or more independent variables, but we do not use this fact here. 15 We chose these values because they are approximately the average parameters estimated by Brown and Warner [1985]. A main alternative would be to simply let 𝜀 be the stock’s market-model residual that day, but this would necessarily lead to heteroscedasticity and require the use of weighted least squares or generalized least squares, as pointed out by Karafiath et al [1991]. To avoid this additional complication, we opted for a homoscedastic error term, which allows us to use ordinary least squares.

  • 14

    simulated portfolios. Table 4 reports estimates the parameters of each of regressions [8]—[11],

    and Table 5 shows the differences in the rejection rates of H0: = 0 (simple) and H0: 2 = 0

    (multivariate).

    The results are fairly similar to those of Tables 2 and 3, but not as dramatic. Still, when

    ln(X1/X2) is used, Table 5’s Panel A shows the average rejection rate of H0: 2 = 0 in favor of a

    positive 2 is 17.6%, substantially larger than the 3.2% rejection rate for the simple regression

    [8a] featuring ln(X1/X2) as the only independent variable. Even worse, when all three ratios are

    logged, the rejection rate of H0: 2 = 0 in favor of a positive 2 is 64.6%. The results of Table 5’s

    Panels C and D are quite similar; the middle independent variable [𝑋1

    𝑋2 or ln(

    𝑋1

    𝑋2)] appears to be

    significant substantially more often than the 5% significance level or even regression [8a]’s

    7.7%, again suggesting that its role in the multivariate regression is primarily “cleaning up” the

    noise created by using as normalizing variables something other than what Christie shows to be

    the correct normalizer, market equity.

    V. Other Issues

    Because it is well-established that leverage is usually associated with an increase in the

    dispersion of returns, the reason it provides no marginal explanatory power for ∆𝑀𝐸

    𝑀𝐸 above and

    beyond what is provided by 𝑉

    𝑀𝐸 merits a bit of extra explanation. Basically, the familiar result

    that additional debt increases equityholders’ risk is based on the assumption that such debt is

    used to increase the firm’s assets to scale. However, if the increase in debt is not associated

    with a proportional increase in risky assets, it is important to use 𝑉

    𝑀𝐸 rather than

    𝑉

    𝑇𝐴.

    As a thought experiment, we consider two cases for a firm that starts as 100% equity and

    doubles its size by borrowing; in the first case, the firm increases its assets to scale, while in the

    second it invests the proceeds from the debt in a riskless asset. In the first case, the firm’s

    equity is indeed more risky. In the second case, however, the issuance of more debt and the

    purchase of riskless assets form a perfect hedge; the firm’s equityholders are in no more risky a

    position than before. Use of 𝑉

    𝑀𝐸 as the independent variable captures this; in the first case,

    equity is more risky because V has doubled, while in the second case, the risk to equityholders

    has not changed because V has not changed. Use of 𝐷

    𝑇𝐴 (or, equivalently,

    𝐵𝐸

    𝑇𝐴), however, is

    misleading because it suggests equityholders’ risk has increased in both cases due to the

    increase in debt. Thus 𝑉

    𝑀𝐸 works correctly in both cases, while

    𝐷

    𝑇𝐴 erroneously implies

    equityholders always bear more risk, even if V remains unchanged. In either case, leverage

  • 15

    adds no explanatory power for equityholders’ returns above and beyond that which was

    provided by 𝑉

    𝑀𝐸.

    The analysis above (and indeed throughout the paper) has been based on the assumption of

    riskless debt, but we now briefly consider the impact of risky debt. If debt is risky, then ∆V = kV

    = ∆ME + ∆D, or equivalently, ∆𝑀𝐸

    𝑀𝐸=

    𝑘𝑉

    𝑀𝐸−

    ∆𝐷

    𝑀𝐸. Now if we estimate the parameters of the

    regression

    AR = 𝛼 + 𝛽1𝑉

    𝑀𝐸+ 𝛽2

    𝐷

    𝑀𝐸+ 𝜀, [13]

    we would expect 𝛽1 = k to have a sign opposite that of 𝛽2. If k > 0, for example, and the firm’s

    debt is risky, then debtholders share some part of the gain ∆V. The more debt there is to share

    that gain, the smaller the gain that accrues to the equityholders, and thus E(𝛽2) < 0.16 This is

    superficially the opposite of the result found in [3], where 𝐷

    𝑇𝐴 had a positive coefficient, but [3]

    assumed riskless debt and used the erroneous variables 𝑉

    𝑇𝐴 and

    𝐷

    𝑇𝐴, while here we are explicitly

    allowing risky debt and using the correct variables 𝑉

    𝑀𝐸 and

    𝐷

    𝑀𝐸. The result here is that, if debt is

    risky, then it mitigates gains and losses to shareholders. This is similar to a consideration of

    risky debt in the Miller and Miller Capital Structure Proposition; risky debt itself does not create

    extra risk, but rather shares (and thus mitigates) risk that would otherwise accrue to

    equityholders. More importantly here, the terms of [13] are normalized by market equity for

    the same reasons as discussed in Christie [1987] and in the introduction.

    While not explicitly stated, we have assumed so far that V is an entry from a market value

    balance sheet. As a consequence, it could also be some capitalized value from an income

    statement, and such variables as sales, cost of goods sold, or net income are also candidates for

    V. In practice, however, market value balance sheets are generally academic fictions. Only

    book values are typically available, and we simply make the assumption that they are closely

    related to analogous market-value balance sheet entries. Provided any discrepancies between

    relevant book-value entries and market-value entries are proportional across firms, this

    changes the slope coefficients themselves but does not alter the t-statistics in any way. For

    example, suppose the true dependence of abnormal returns on the variable of interest is

    c(𝑉𝑀𝑎𝑟𝑘𝑒𝑡 𝑉𝑎𝑙𝑢𝑒

    𝑀𝐸) for some constant c, but all we can measure is

    𝑉𝐵𝑜𝑜𝑘 𝑉𝑎𝑙𝑢𝑒

    𝑀𝐸. If for all firms book

    values are overstated (or understated) by the same proportion d, then the coefficient we

    measure will be equal to 𝑐

    𝑑 instead of c, but the t-values will be identical.

    16 A result of 𝛽2 = 0 would indicate the debt is riskless, at least as far as changes in V are concerned.

  • 16

    VI. Extensions

    Apart from use of Book Equity and Total Assets as X1 and X2, Tables 2—5 may appear to be

    contrived, and indeed to some extent they are. However, the problems they identify can occur

    even in more common settings. Specifically, is not strictly necessary for the regression to

    include X1 and X2 in multiple places. Suppose, for example, there existed another balance sheet

    item, A, whose value was a constant proportion of, say, total assets, for every firm, or Ai = cTAi.

    In this case, substitution of Ai for TAi (or vice versa) as a component of any variable would result

    in beta coefficients that were different by a factor of c, but which would have identical t-

    statistics. If there is a slight perturbation so that Ai = cTAi + i (where i has a small variance

    relative to the dispersion of TAi), then substitution of Ai for TAi (or vice versa) in any ratio will

    produce very similar t-statistics. Since Ai = cTAi + i is essentially a regression of Ai on TAi that is

    constrained to go through the origin, one measure of how slight the perturbations i really are

    is the R2 of this constrained regression. While the R2 of a regression with an intercept is

    uniquely defined, there are no fewer than eight17 common ways of measuring the R2 of a

    regression constrained to go through the origin. We choose the one SAS uses, 1 −∑(𝑦𝑖− �̂�𝑖)

    2

    ∑ 𝑦𝑖2 ,

    and dub it the constrained R2 for the remainder of this paper.18 We note in passing that when it

    comes to the entire ratio that serves as an independent variable, that a substitute have a very

    large linear correlation is sufficient to produce similar t-statistics for the slope coefficients.

    However, when we are looking at either the numerator or denominator of an independent

    variable, a high linear correlation is insufficient to produce comparable results; instead, the

    constrained correlation must be very large to ensure similar t-statistics. The reason is that a

    high linear correlation between X1 and X2 implies that there exist a and b such that X2 ≅ a + bX1,

    but if a ≠ 0 and if X2 appears in a ratio in lieu of X1, there will not be a cancelling out effect

    because the constant term a will not cancel out. However, if the constrained correlation is

    large, there will be a cancelling out effect because X2 ≅ bX1 is ensured.19

    17 E.g., see Kvalseth (1985). 18 In mathematics, Y = X is termed a linear transformation and Y = X an affine transformation, and thus it seems natural to call their measures of R2 linear and affine, respectively. However, the use of the expression “linear correlation” to describe the affine transformation is already widespread, so we resort to this alternative. 19 For example, suppose X ~U[5, 10]. The correlation between X and the simple translation 6.02*1023 + X is 1, and

    yet, because 6.02*1023 dwarfs X, 1

    6.02∗1023+𝑋 will have approximately the same value for any values of X that are

    orders of magnitude smaller, and so will not produce the desired “cancelling out” effect with any X that may

    appear in a numerator. A high constrained R2, however, will necessarily produce a cancelling out effect. A similar

    result would occur if X in a numerator is replaced with 6.02*1023 + X.

  • 17

    For the full set of observations, one of the largest constrained correlations was that

    between Operating Income Before Depreciation and Total Assets, at 0.9741. This suggests that

    we should obtain similar results when the value for Total Assets is substituted for one of the

    values of Operating Income Before Depreciation in Panel A of Tables 2 and 4. We made this

    substitution and report the results in Table 6.

    As Table 6 suggests, the problem of misleading significance is somewhat reduced but

    nevertheless persists when Total Assets replace Operating Income Before Depreciation in either

    the first or third variable (because the middle variables, 𝑂𝐼𝐵𝐷𝑃

    𝐶𝑎𝑝𝐸𝑥 or

    𝐶𝑎𝑝𝐸𝑥

    𝑂𝐼𝐵𝐷𝑃 are the ones producing

    misleading significant slope coefficients, we left them intact). Panels A and B show the results

    for the deterministic AR = ∆𝑀𝐸

    𝑀𝐸 =

    .05𝐶𝑎𝑠ℎ

    𝑀𝐸, while C and D show them for the stochastic AR =

    ∆𝑀𝐸

    𝑀𝐸 =

    .05𝐶𝑎𝑠ℎ

    𝑀𝐸+ 𝛿. In the first three rows Panel A (with no substitutions), for example, frequency of

    positive and significant rejections that the coefficient of the middle term (𝑂𝐼𝐵𝐷𝑃

    𝐶𝑎𝑝𝐸𝑥 or its logarithm)

    equals zero range from 23.7% to 61.0% to 99.3% (as we proceed from taking no logs to taking

    the log of the middle independent variable only to taking logs of all three independent

    variables, i.e., regressions [9]—[11]). When we substitute Total Assets for the first independent

    variable’s denominator of OIBDP in the last three rows, these rejection rates are significantly

    smaller, proceeding from 11.3% to 22.2% to 73.8%. Still, all are larger than the 5.8% rejection

    rate we actually found in Table 2, Panel A’s simple regression [8a], AR = 𝛼 + 𝛽ln (𝑂𝐼𝐵𝐷𝑃

    𝐶𝑎𝑝𝐸𝑥) + 𝜀. A

    similar result is found in Panel B of Table 6, which uses 𝐶𝑎𝑝𝐸𝑥

    𝑂𝐼𝐵𝐷𝑃 (or its logarithm) as the middle

    variable. The original frequencies of positive and significant t-values for [9]—[11] of 24.1%,

    52.7%, and 99.2% drop to 14.2%, 33.7%, and 84.9% when Total Assets are substituted for

    Operating Income Before Depreciation in the numerator of the third independent variable.

    Again, while the rejection rates are somewhat lower after the substitution is made, all are

    substantially larger than the 6.1% we found for the actual simple regression [8a] in Panel B of

    Table 2. When the abnormal return AR is not deterministic and instead includes an error term,

    Panel C of Table 6 shows similar results: for 𝑂𝐼𝐵𝐷𝑃

    𝐶𝑎𝑝𝐸𝑥 (or its logarithm) the original positive and

    significant rejection rates of 9.5%, 20.4%, and 50.3% become 5.2%, 9.9%, and 24.0% after the

    substitution is made. All exceed the 3.4% rejection rate for Table 4, Panel A’s simple regression

    [8a], the last two by substantial amounts. Panel D of Table 6 shows similar results.

    In all cases, Table 6 suggests the problem of distorted t-statistics stemming from the fact

    that an irrelevant independent variable may “clean up” poor measures of other (relevant)

    independent variables can remain even when the irrelevant component appears in only one

    ratio. Provided they have a high constrained correlation, one variable may be substituted for

    another in either a numerator or a denominator and produce deceptive significance levels. This

    can be a substantial problem for regressions with a large number of independent variables, as

    any set of three or more independent variables can combine in such a way that one (or more)

  • 18

    of them appear to be significant even though their only role is cleaning up noise created by

    poor choices of other independent variables.20

    VII. Conclusions

    Christie [1987] showed that market equity [ME] is the correct scaling variable for any cross-

    sectional regression of abnormal returns on firm characteristics, but many researchers scale by

    such variables as Total Assets or Book Equity instead. We demonstrate that the apparent

    significance of the resulting variables can be a mathematical artifact that is unrelated to their

    true significance. Not only can true (and sometimes clearly so) null hypotheses frequently be

    rejected, but depending on the initial setup, they can be rejected in favor of contradictory

    alternatives. Moreover, because the effect can occur whenever two variables have a high

    constrained R2, detection of false inferences is quite difficult. One (imperfect) step towards

    confirming coefficients in a multivariate regression are truly what we purport them to be is to

    test them in simple regressions as well. These results also highlight the importance of having a

    specific model and using it to determine the exact variables and their appropriate form rather

    than let these choices be made by the data. Finally, we can conclude it is inappropriate to add

    independent variables based on the assumptions that they might matter, and if they don’t they

    will not cause any damage. They may well cause damage by creating (or perhaps resolving)

    needless noise that makes them and other irrelevant variables appear to be significant.

    Estimates of multivariate regressions have meaning only in the context of all the independent

    variables selected, and adding, deleting, or changing variables can completely alter the meaning

    of other variables’ coefficients.

    Moreover, while our framework is based on event-study cross-sectional regressions, this

    condition was only invoked because in this setting Christie has identified the correct functional

    form. The general principle could be shown to apply to other cross-sectional regressions as

    well, but a proof would require that we knew the correct functional form. While such a correct

    functional form may exist, we will rarely know what it is; nevertheless, its existence implies that

    estimates of other multiple regressions will be subject to the same problem we have identified

    for event-study cross-sectional regressions. Thus this paper provides evidence in support of

    Griliches’ Law: “Any cross-sectional regression with more than five variables produces

    garbage.” The more independent variables that appear in a regression, the greater the chance

    that some subset of them will combine to make an irrelevant variable appear to be significant

    because it is “cleaning up” the noise created by an incorrect choice of other variables.

    20 For example, a regression with 10 independent variables will have 210 − 11 − (

    102

    ) = 968 subsets of three or

    more independent variables, any of which can be plagued by this problem.

  • 19

    References

    Brown, S., and J. Warner, 1985, Using daily stock returns: The case of event studies,” Journal of

    Financial Economics, 14(1): 3—31.

    Christie, A. A., 1987, On Cross-Sectional Analysis in Accounting Research, Journal of Accounting

    and Economics, 9(3):231—258.

    Goldman, W., 1973, The Princess Bride, Harcourt Brace Jovanovich (San Diego).

    Griliches, Z, and N. Wallace, 1965, The Determinants of Investment Reinvestigated,

    International Economic Review, 6(3): 311—329.

    Kendrick, D., and M. Intriligator, 1974, Frontiers of quantitative economics: Papers invited for

    presentation at the Econometric Society Winter Meetings, New York, 1969 [and] Toronto, 1972.

    Vol. 2. North-Holland Pub. Co.

    Karafiath, I., R. Mynatt, and K. Smith, 1991, The Brazilian default announcement and the

    contagion effect hypothesis, Journal of Banking and Finance, 15(3): 699—716.

    Kennedy, P., 2008, A Guide to Econometrics, 6th Edition, Blackwell Publishing.

    Kmenta, J., 1997, Elements of Econometrics, Second Edition, The University of Michigan Press.

    Kvalseth, T., 1985, Cautionary Note about R2, The American Statistician, 39(4): 279-285.

    McCloskey, D., 1987, The Writing of Economics, MacMillan Publishing Co., New York.

    Mitchell, D., 1991, Invariance of Results Under a Common Orthogonalization, Journal of

    Economics and Business, 43: 193-196.

  • 20

    Table 1—Summary of successive regression forms from equations (3)—(6) and related multivariate

    regressions.

    Multivariate Regression Constant

    (t-statistic) Coefficient (t-statistic) of Cash/TA

    or ln(Cash/TA)

    Coefficient (t-statistic)

    of D/TA, BE/TA, or ln(BE/TA)

    Coefficient (t-statistic) of BE/ME

    or ln(BE/ME)

    [3]

    .05Cash/ME = + 1(Cash/TA) +

    2(D/TA) + 3(BE/ME)

    -.0090 (-3.46)

    .0386 (8.14)

    .0143 (3.45)

    .0098 (5.59)

    [4]

    .05Cash/ME = + 1(Cash/TA) +

    2(BE/TA) + 3(BE/ME)

    .0040 (2.18)

    .0431 (9.02)

    -.0186 (-4.83)

    .0107 (6.23)

    [5]

    .05Cash/ME = + 1ln(Cash/TA) +

    2ln(BE/TA) + 3ln(BE/ME)

    .0185

    (12.92)

    .0040 (9.88)

    -.0061 (-6.62)

    .0068 (7.60)

    [6]

    Ln(.05Cash/ME) = + 1ln(Cash/TA) +

    2ln(BE/TA) + 3ln(BE/ME)

    -2.9957 (−∞)

    1.000 (∞)

    -1.000 (−∞)

    1.000 (∞)

    In the multivariate regressions, t-values for the three variable’s coefficients improve as we move from

    using 𝐷𝑒𝑏𝑡

    𝑇𝐴 [equation (3)] to

    𝐵𝐸

    𝑇𝐴 [equation (4)], and then improve more as when we take natural

    logarithms of the three independent variables [equation (5)], and finally become infinite when we take

    the logarithm of the dependent variable as well [equation (6)]. The multivariate t-statistics also

    generally improve [except when we take logarithms to move from (4) to (5)], but not as dramatically.

  • 21

    Table 2: Simulated Portfolios for various choices of X1 and X2 in sequence of regressions from

    .𝟎𝟓(𝑪𝒂𝒔𝒉)

    𝑴𝑬 = 𝜶 + 𝜷

    𝑿𝟏

    𝑿𝟐 to

    .𝟎𝟓(𝑪𝒂𝒔𝒉)

    𝑴𝑬 = 𝜶 + 𝜷𝟏𝒍𝒏(

    𝑽

    𝑿𝟏) + 𝜷𝟐𝒍𝒏(

    𝑿𝟏

    𝑿𝟐) + 𝜷𝟑𝒍𝒏(

    𝑿𝟐

    𝑴𝑬).

    Panel A: X1 = Operating Income Before Depreciation, X2 = Capital Expenditures

    Independent Variables

    𝑪𝒂𝒔𝒉

    𝑿𝟏

    (or its log)

    𝑿𝟏𝑿𝟐

    (or its log)

    𝑿𝟐𝑴𝑬

    (or its log)

    𝑿𝟏

    𝑿𝟐 only

    [8]

    (0.50) [0.112] {0.516}

    𝒍𝒏(𝑿𝟏

    𝑿𝟐) only

    [8a]

    (-0.10) [0.119] {0.437}

    All three

    (unlogged) [9]

    (4.18) [0.829] {0.053}

    (1.36) [0.237] {0.367}

    (4.15) [0.604] {0.176}

    All three

    (log of 𝑿𝟏

    𝑿𝟐

    only) [10]

    (4.76) [0.882] {0.028}

    (2.48) [0.610] {0.138}

    (4.88) [0.704] {0.100}

    All three logged

    [11]

    (10.87) [1.000] {0.000}

    (6.16) [0.993] {0.001}

    (7.78) [1.000] {0.000}

    (average t)

    [total rejection rate]

    {average p-value}

  • 22

    Table 2, Panel B: X1 = Capital Expenditures, X2 = Operating Income Before Depreciation

    Independent Variables

    𝑪𝒂𝒔𝒉

    𝑿𝟏

    (or its log)

    𝑿𝟏𝑿𝟐

    (or its log)

    𝑿𝟐𝑴𝑬

    (or its log)

    𝑿𝟏

    𝑿𝟐 only

    [8]

    (0.51) [0.104] {0.500}

    𝒍𝒏(𝑿𝟏

    𝑿𝟐) only

    [8a]

    (0.10) [0.119] {0.437}

    All three

    (unlogged) [9]

    (4.48) [0.885] {0.039}

    (1.31) [0.242] {0.362}

    (5.83) [0.754] {0.086}

    All three

    (log of 𝑿𝟏

    𝑿𝟐

    only) [10]

    (4.86) [0.898] {0.031}

    (2.03) [0.531] {0.186}

    (5.96) [0.775] {0.082}

    All three logged

    [11]

    (10.87) [1.000] {0.000}

    (7.09) [0.992] {0.002}

    (7.78) [1.000] {0.000}

    (average t)

    [total rejection rate]

    {average p-value}

  • 23

    Table 2, Panel C: X1 = Accounts Payable, X2 = Total Receivables

    Independent Variables

    𝑪𝒂𝒔𝒉

    𝑿𝟏

    (or its log)

    𝑿𝟏𝑿𝟐

    (or its log)

    𝑿𝟐𝑴𝑬

    (or its log)

    𝑿𝟏

    𝑿𝟐 only

    [8]

    (0.32) [0.082] {0.575}

    𝒍𝒏(𝑿𝟏

    𝑿𝟐) only

    [8a]

    (0.28) [0.099] {0.453}

    All three

    (unlogged) [9]

    (3.79) [0.799] {0.059}

    (0.77) [0.127] {0.519}

    (4.62) [0.722] {0.105}

    All three

    (log of 𝑿𝟏

    𝑿𝟐

    only) [10]

    (4.08) [0.839] {0.044}

    (1.63) [0.377] {0.243}

    (4.77) [0.748] {0.091}

    All three logged

    [11]

    (11.14) [1.000] {0.000}

    (7.48) [1.000] {0.000}

    (9.30) [1.000] {0.000}

    (average t)

    [total rejection rate]

    {average p-value}

  • 24

    Table 2, Panel D: X1 = Depreciation and Amortization, X2 = Total Inventory

    Independent Variables

    𝑪𝒂𝒔𝒉

    𝑿𝟏

    (or its log)

    𝑿𝟏𝑿𝟐

    (or its log)

    𝑿𝟐𝑴𝑬

    (or its log)

    𝑿𝟏

    𝑿𝟐 only

    [8]

    (0.07) [0.061] {0.542}

    𝒍𝒏(𝑿𝟏

    𝑿𝟐) only

    [8a]

    (0.33) [0.100] {0.460}

    All three

    (unlogged) [9]

    (3.76) [0.807] {0.053}

    (0.63) [0.115] {0.531}

    (3.49) [0.488] {0.217}

    All three

    (log of 𝑿𝟏

    𝑿𝟐

    only) [10]

    (4.09) [0.873] {0.033}

    (2.00) [0.482] {0.172}

    (4.09) [0.626] {0.129}

    All three logged

    [11]

    (10.40) [1.000] {0.000}

    (8.09) [1.000] {0.000}

    (8.73) [1.000] {0.000}

    (average t)

    [total rejection rate]

    {average p-value}

  • 25

    Table 2, Panel E: X1 = Inventory, X2 = Cost of Goods Sold

    Independent Variables

    𝑪𝒂𝒔𝒉

    𝑿𝟏

    (or its log)

    𝑿𝟏𝑿𝟐

    (or its log)

    𝑿𝟐𝑴𝑬

    (or its log)

    𝑿𝟏

    𝑿𝟐 only

    [8]

    (-0.30) [0.077] {0.475}

    𝒍𝒏(𝑿𝟏

    𝑿𝟐) only

    [8a]

    (-0.30) [0.099] {0.452}

    All three

    (unlogged) [9]

    (2.47) [0.550] {0.154}

    (0.75) [0.142] {0.490}

    (4.43) [0.619] {0.158}

    All three

    (log of 𝑿𝟏

    𝑿𝟐

    only) [10]

    (2.58) [0.596] {0.138}

    (0.97) [0.200] {0.360}

    (4.48) [0.630] {0.151}

    All three logged

    [11]

    (10.44) [1.000] {0.000}

    (6.51) [0.994] {0.001}

    (8.96) [1.000] {0.000}

    (average t)

    [total rejection rate]

    {average p-value}

  • 26

    Table 2, Panel F: X1 = Interest Expense, X2 = Cost of Goods Sold

    Independent Variables

    𝑪𝒂𝒔𝒉

    𝑿𝟏

    (or its log)

    𝑿𝟏𝑿𝟐

    (or its log)

    𝑿𝟐𝑴𝑬

    (or its log)

    𝑿𝟏

    𝑿𝟐 only

    [8]

    (0.40) [0.119] {0.484}

    𝒍𝒏(𝑿𝟏

    𝑿𝟐) only

    [8a]

    (-0.14) [0.096] {0.449}

    All three

    (unlogged) [9]

    (1.48) [0.271] {0.313}

    (1.01) [0.205] {0.443}

    (4.62) [0.639] {0.141}

    All three

    (log of 𝑿𝟏

    𝑿𝟐

    only) [10]

    (1.53) [0.285] {0.283}

    (0.620) [0.166] {0.398}

    (4.57) [0.626] {0.148}

    All three logged

    [11]

    (10.14) [1.000] {0.000}

    (8.16) [1.000] {0.000}

    (8.97) [1.000] {0.000}

    (average t)

    [total rejection rate]

    |negative and significant rejection rate|

    {average p-value}

  • 27

    Table 2, Panel G: X1 = Capital Expenditures, X2 = Accounts Payable

    Independent Variables

    𝑪𝒂𝒔𝒉

    𝑿𝟏

    (or its log)

    𝑿𝟏𝑿𝟐

    (or its log)

    𝑿𝟐𝑴𝑬

    (or its log)

    𝑿𝟏

    𝑿𝟐 only

    [8]

    (-0.27) [0.067] {0.470}

    𝒍𝒏(𝑿𝟏

    𝑿𝟐) only

    [8a]

    (-0.90) [0.178] {0.366}

    All three

    (unlogged) [9]

    (3.30) [0.707] {0.102}

    (0.46) [0.093] {0.525}

    (4.27) [0.621] {0.159}

    All three

    (log of 𝑿𝟏

    𝑿𝟐

    only) [10]

    (3.44) [0.737] {0.082}

    (1.06) [0.259] {0.345}

    (4.43) [0.644] {0.142}

    All three logged

    [11]

    (11.36) [1.000] {0.000}

    (7.55) [0.999] {0.000}

    (9.19) [1.000] {0.000}

    (average t)

    [total rejection rate]

    {average p-value}

  • 28

    Table 2, Panel H: X1 = Capital Expenditures, X2 = Interest Expense

    Independent Variables

    𝑪𝒂𝒔𝒉

    𝑿𝟏

    (or its log)

    𝑿𝟏𝑿𝟐

    (or its log)

    𝑿𝟐𝑴𝑬

    (or its log)

    𝑿𝟏

    𝑿𝟐 only

    [8]

    (0.19) [0.042] {0.657}

    𝒍𝒏(𝑿𝟏

    𝑿𝟐) only

    [8a]

    (-0.50) [0.107] {0.432}

    All three

    (unlogged) [9]

    (3.45) [0.728] {0.090}

    (0.60) [0.079] {0.576}

    (5.58) [0.724] {0.107}

    All three

    (log of 𝑿𝟏

    𝑿𝟐

    only) [10]

    (3.69) [0.789] {0.066}

    (1.43) [0.302] |{0.282}

    (5.85) [0.769] {0.081}

    All three logged

    [11]

    (10.41) [1.000] {0.000}

    (7.19) [0.999] {0.000}

    (8.53) [1.000] {0.000}

    (average t)

    [total rejection rate]

    {average p-value}

  • 29

    Table 2, Panel I: X1 = Book Equity, X2 = Total Assets

    Independent Variables

    𝑪𝒂𝒔𝒉

    𝑿𝟏

    (or its log)

    𝑿𝟏𝑿𝟐

    (or its log)

    𝑿𝟐𝑴𝑬

    (or its log)

    𝑿𝟏

    𝑿𝟐 only

    [8]

    (0.94) [0.253] {0.326}

    𝒍𝒏(𝑿𝟏

    𝑿𝟐) only

    [8a]

    (-0.03) [0.228] {0.349}

    All three

    (unlogged) [9]

    (6.38) [0.920] {0.024}

    (4.26) [0.943] {0.015}

    (7.74) [0.988] {0.003}

    All three

    (log of 𝑿𝟏

    𝑿𝟐

    only) [10]

    (7.17) [0.973] {0.007}

    (4.54) [0.946] {0.014}

    (7.99) [0.990] {0.002}

    All three logged

    [11]

    (10.88) [1.000] {0.000}

    (4.53) [0.909] {0.029}

    (9.41) [1.000] {0.000}

    (average t)

    [total rejection rate]

    {average p-value}

  • 30

    Table 2, Panel J: X1 = Total Assets, X2 = Book Equity

    Independent Variables

    𝑪𝒂𝒔𝒉

    𝑿𝟏

    (or its log)

    𝑿𝟏𝑿𝟐

    (or its log)

    𝑿𝟐𝑴𝑬

    (or its log)

    𝑿𝟏

    𝑿𝟐 only

    [8]

    (0.85) [0.188] {0.442}

    𝒍𝒏(𝑿𝟏

    𝑿𝟐) only

    [8a]

    (0.03) [0.228] {0.349}

    All three

    (unlogged) [9]

    (9.54) [1.000] {0.000}

    (3.30) [0.643] {0.112}

    (9.27) [0.999] {0.000}

    All three

    (log of 𝑿𝟏

    𝑿𝟐

    only) [10]

    (9.95) [1.000] {0.000}

    (3.87) [0.872] |{0.034}

    (9.44) [0.999] {0.000}

    All three logged

    [11]

    (10.88) [1.000] {0.000}

    (5.72) [0.988] {0.002}

    (9.41) [1.000] {0.000}

    (average t)

    [total rejection rate]

    {average p-value}

  • Table 3: Summary of Rejection Rates of the Coefficient of X1/X2, Deterministic ARs

    Panel A: Only positive rejections (t-statistic > 1.96)

    A B C D E F G H I J average

    X1 Oper. Inc.

    before Deprec.

    Capital Expend.

    A/P Deprec. Total Inventory

    Interest Expense

    Capital Expend.

    Capital Expend.

    Book Equity

    Total Assets

    X2 Capital Expend.

    Oper. Inc.

    before Deprec.

    Total Receivables

    Total Inventory

    Cost of Goods Sold

    Cost of Goods Sold

    A/P Interest Expense

    Total Assets

    Book Equity

    Simple, not logged [8] 0.111 0.103 0.081 0.061 0.045 0.116 0.044 0.042 0.229 0.179 0.101

    Simple, logged [8a] 0.058 0.061 0.079 0.075 0.018 0.034 0.009 0.013 0.097 0.131 0.058

    Multivariate, none logged

    [9] 0.237 0.241 0.123 0.115 0.136 0.205 0.092 0.079 0.943 0.643 0.281 Multivariate,

    only X1/X2 logged [10] 0.61 0.527 0.377 0.482 0.194 0.142 0.249 0.301 0.945 0.872 0.470

    Multivariate, all logged

    [11] 0.993 0.992 1 1 0.994 1 0.999 0.999 0.906 0.988 0.987

  • 32

    Table 3, Panel B: Increases in rejection rates: Only positive rejections (t-statistic > 1.96)

    A B C D E F G H I J average

    X1 Oper. Inc.

    before Deprec.

    Capital Expend.

    A/P Deprec. Total Inventory

    Interest Expense

    Capital Expend.

    Capital Expend.

    Book Equity

    Total Assets

    X2 Capital Expend.

    Oper. Inc.

    before Deprec.

    Total Receivables

    Total Inventory

    Cost of Goods Sold

    Cost of Goods Sold

    A/P Interest Expense

    Total Assets

    Book Equity

    Multivariate, none logged

    minus Simple, not

    logged 0.126 0.138 0.042 0.054 0.091 0.089 0.048 0.037 0.714 0.464 0.180 Multivariate,

    only X1/X2 logged minus

    Multivariate, none logged 0.373 0.286 0.254 0.367 0.058 -0.063 0.157 0.222 0.002 0.229 0.189 Multivariate,

    all logged minus

    Multivariate, only X1/X2

    logged 0.383 0.465 0.623 0.518 0.8 0.858 0.75 0.698 -0.039 0.116 0.517 Correlation

    between X1/X2 and ln(X1/X2) 0.065 0.071 0.121 0.253 0.173 0.069 0.169 0.162 0.892 0.11

  • 33

    Table 3, Panel C: All rejections (|t-statistic| > 1.96)

    A B C D E F G H I J average

    X1 Oper. Inc.

    before Deprec.

    Capital Expend.

    A/P Deprec. Total Inventory

    Interest Expense

    Capital Expend.

    Capital Expend.

    Book Equity

    Total Assets

    X2 Capital Expend.

    Oper. Inc.

    before Deprec.

    Total Receivables

    Total Inventory

    Cost of Goods Sold

    Cost of Goods Sold

    A/P Interest Expense

    Total Assets

    Book Equity

    Simple, not logged [8] 0.112 0.104 0.082 0.061 0.077 0.119 0.067 0.042 0.253 0.188 0.111

    Simple, logged [8a] 0.119 0.119 0.099 0.1 0.099 0.096 0.178 0.107 0.228 0.228 0.137

    Multivariate, none logged

    [9] 0.237 0.242 0.127 0.115 0.142 0.205 0.093 0.079 0.943 0.643 0.283

    Multivariate, only X1/X2 logged [10] 0.61 0.531 0.377 0.482 0.2 0.166 0.259 0.302 0.946 0.872 0.475

    Multivariate, all logged

    [11] 0.993 0.992 1 1 0.994 1 0.999 0.999 0.909 0.988 0.987

  • 34

    Table 3, Panel D: Increases in rejection rates (all |t-statistic| > 1.96)

    A B C D E F G H I J average

    X1 Oper. Inc.

    before Deprec.

    Capital Expend.

    A/P Deprec. Total Inventory

    Interest Expense

    Capital Expend.

    Capital Expend.

    Book Equity

    Total Assets

    X2 Capital Expend.

    Oper. Inc.

    before Deprec.

    Total Receivables

    Total Inventory

    Cost of Goods Sold

    Cost of Goods Sold

    A/P Interest Expense

    Total Assets

    Book Equity

    Multivariate, none logged

    minus Simple, not

    logged 0.125 0.138 0.045 0.054 0.065 0.086 0.026 0.037 0.69 0.455 0.172 Multivariate,

    only X1/X2 logged minus

    Multivariate, none logged 0.373 0.289 0.25 0.367 0.058 -0.039 0.166 0.223 0.003 0.229 0.192 Multivariate,

    all logged minus

    Multivariate, only X1/X2

    logged 0.383 0.461 0.623 0.518 0.794 0.834 0.74 0.697 -0.037 0.116 0.513

    Correlation between X1/X2 and ln(X1/X2) 0.065 0.071 0.121 0.253 0.173 0.069 0.169 0.162 0.892 0.11

  • Table 4: Simulated Portfolios for various choices of X1 and X2 in sequence of regressions from

    .𝟎𝟓(𝑪𝒂𝒔𝒉)

    𝑴𝑬+ 𝜹 = 𝜶 + 𝜷

    𝑿𝟏

    𝑿𝟐+ 𝜺 to

    .𝟎𝟓(𝑪𝒂𝒔𝒉)

    𝑴𝑬+ 𝜹 = 𝜶 + 𝜷𝟏𝒍𝒏 (

    𝑽

    𝑿𝟏) + 𝜷𝟐𝒍𝒏 (

    𝑿𝟏

    𝑿𝟐) + 𝜷𝟑𝒍𝒏 (

    𝑿𝟐

    𝑴𝑬) + 𝜺.

    Panel A: X1 = Operating Income Before Depreciation, X2 = Capital Expenditures

    Independent Variables

    𝑪𝒂𝒔𝒉

    𝑿𝟏

    (or its log)

    𝑿𝟏𝑿𝟐

    (or its log)

    𝑿𝟐𝑴𝑬

    (or its log)

    𝑿𝟏

    𝑿𝟐 only

    [8]

    (0.21) [0.076] {0.501}

    𝒍𝒏(𝑿𝟏

    𝑿𝟐) only

    [8a]

    (-0.02) [0.070] {0.488}

    All three

    (unlogged) [9]

    (1.47) [0.326] {0.269}

    (0.54) [0.102] {0.453}

    (1.72) [0.314] {0.323}

    All three

    (log of 𝑿𝟏

    𝑿𝟐

    only) [10]

    (1.67) [0.398] {0.234}

    (0.99) [0.209] {0.360}

    (1.96) [0.352] {0.286}

    All three logged

    [11]

    (3.30) [0.870] {0.028}

    (2.02) [0.503] {0.186}

    (2.56) [0.619] {0.121}

    (average t)

    [total rejection rate]

    {average p-value}

  • 36

    Table 4, Panel B: X1 = Capital Expenditures, X2 = Operating Income Before Depreciation

    Independent Variables

    𝑪𝒂𝒔𝒉

    𝑿𝟏

    (or its log)

    𝑿𝟏𝑿𝟐

    (or its log)

    𝑿𝟐𝑴𝑬

    (or its log)

    𝑿𝟏

    𝑿𝟐 only

    [8]

    (0.19) [0.074] {0.494}

    𝒍𝒏(𝑿𝟏

    𝑿𝟐) only

    [8a]

    (0.02) [0.070] {0.488}

    All three

    (unlogged) [9]

    (1.58) [0.322] {0.242}

    (0.45) [0.088] {0.452}

    (2.33) [0.419] {0.254}

    All three

    (log of 𝑿𝟏

    𝑿𝟐

    only) [10]

    (1.68) [0.382] {0.224}

    (0.68) [0.132] {0.413}

    (2.35) [0.423] {0.249}

    All three logged

    [11]

    (3.30) [0.870] {0.029}

    (2.15) [0.548] {0.139}

    (2.56) [0.619] {0.121}

    (average t)

    [total rejection rate]

    {average p-value}

  • 37

    Table 4, Panel C: X1 = Accounts Payable, X2 = Total Receivables

    Independent Variables

    𝑪𝒂𝒔𝒉

    𝑿𝟏

    (or its log)

    𝑿𝟏𝑿𝟐

    (or its log)

    𝑿𝟐𝑴𝑬

    (or its log)

    𝑿𝟏

    𝑿𝟐 only

    [8]

    (0.13) [0.063] {0.484}

    𝒍𝒏(𝑿𝟏

    𝑿𝟐) only

    [8a]

    (0.15) [0.081] {0.480}

    All three

    (unlogged) [9]

    (1.73) [0.392] {0.239}

    (0.34) [0.090] {0.468}

    (2.32) [0.403] {0.271}

    All three

    (log of 𝑿𝟏

    𝑿𝟐

    only) [10]

    (1.86) [0.436] {0.212}

    (0.78) [0.152] {0.411}

    (2.39) [0.408] {0.259}

    All three logged

    [11]

    (4.29) [0.975] {0.006}

    (2.92) [0.769] {0.057}

    (3.67) [0.896] {0.027}

    (average t)

    [total rejection rate]

    {average p-value}

  • 38

    Table 4, Panel D: X1 = Depreciation and Amortization, X2 = Total Inventory

    Independent Variables

    𝑪𝒂𝒔𝒉

    𝑿𝟏

    (or its log)

    𝑿𝟏𝑿𝟐

    (or its log)

    𝑿𝟐𝑴𝑬

    (or its log)

    𝑿𝟏

    𝑿𝟐 only

    [8]

    (0.05) [0.046] {0.534}

    𝒍𝒏(𝑿𝟏

    𝑿𝟐) only

    [8a]

    (0.15) [0.063] {0.486}

    All three

    (unlogged) [9]

    (1.54) [0.338] {0.254}

    (0.31) [0.054] {0.501}

    (1.70) [0.272] {0.338}

    All three

    (log of 𝑿𝟏

    𝑿𝟐

    only) [10]

    (1.67) [0.379] {0.229}

    (0.91) [0.175] {0.386}

    (1.93) [0.330] {0.304}

    All three logged

    [11]

    (3.66) [0.924] {0.017}

    (2.97) [0.787] {0.064}

    (3.26) [0.818] {0.049}

    (average t)

    [total rejection rate]

    {average p-value}

  • 39

    Table 4, Panel E: X1 = Inventory, X2 = Cost of Goods Sold

    Independent Variables

    𝑪𝒂𝒔𝒉

    𝑿𝟏

    (or its log)

    𝑿𝟏𝑿𝟐

    (or its log)

    𝑿𝟐𝑴𝑬

    (or its log)

    𝑿𝟏

    𝑿𝟐 only

    [8]

    (-0.19) [0.070] {0.491}

    𝒍𝒏(𝑿𝟏

    𝑿𝟐) only

    [8a]

    (-0.20) [0.062] {0.481}

    All three

    (unlogged) [9]

    (1.08) [0.211] {0.364}

    (0.29) [0.080] {0.482}

    (2.16) [0.351] {0.293}

    All three

    (log of 𝑿𝟏

    𝑿𝟐

    only) [10]

    (1.11) [0.224] {0.358}

    (0.39) [0.074] {0.471}

    (2.18) [0.353] {0.286}

    All three logged

    [11]

    (3.70) [0.931] {0.016}

    (2.27) [0.605] {0.121}

    (3.28) [0.838] {0.047}

    (average t)

    [total rejection rate]

    {average p-value}

  • 40

    Table 4, Panel F: X1 = Interest Expense, X2 = Cost of Goods Sold

    Independent Variables

    𝑪𝒂𝒔𝒉

    𝑿𝟏

    (or its log)

    𝑿𝟏𝑿𝟐

    (or its log)

    𝑿𝟐𝑴𝑬

    (or its log)

    𝑿𝟏

    𝑿𝟐 only

    [8]

    (0.23) [0.077] {0.477}

    𝒍𝒏(𝑿𝟏

    𝑿𝟐) only

    [8a]

    (-0.01) [0.064] {0.477}

    All three

    (unlogged) [9]

    (0.64) [0.099] {0.438}

    (0.49) [0.111] {0.455}

    (2.32) [0.386] {0.278}

    All three

    (log of 𝑿𝟏

    𝑿𝟐

    only) [10]

    (0.68) [0.115] {0.425}

    (0.32) [0.099] {0.452}

    (2.30) [0.379] {0.281}

    All three logged

    [11]

    (3.83) [0.944] {0.016}

    (3.14) [0.824] {0.054}

    (3.48) [0.859] {0.038}

    (average t)

    [total rejection rate]

    |negative and significant rejection rate|

    {average p-value}

  • 41

    Table 4, Panel G: X1 = Capital Expenditures, X2 = Accounts Payable

    Independent Variables

    𝑪𝒂𝒔𝒉

    𝑿𝟏

    (or its log)

    𝑿𝟏𝑿𝟐

    (or its log)

    𝑿𝟐𝑴𝑬

    (or its log)

    𝑿𝟏

    𝑿𝟐 only

    [8]

    (-0.17) [0.056] {0.492}

    𝒍𝒏(𝑿𝟏

    𝑿𝟐) only

    [8a]

    (-0.46) [0.093] {0.449}

    All three

    (unlogged) [9]

    (1.54) [0.365] {0.269}

    (0.20) [0.069] {0.500}

    (2.108) [0.379] {0.276}

    All three

    (log of 𝑿𝟏

    𝑿𝟐

    only) [10]

    (1.61) [0.389] {0.251}

    (0.53) [0.101] {0.438}

    (2.25) [0.379] {0.269}

    All three logged

    [11]

    (4.39) [0.979] {0.005}

    (2.97) [0.783] {0.054}

    (3.68) [0.897] {0.027}

    (average t)

    [total rejection rate]

    {average p-value}

  • 42

    Table 4, Panel H: X1 = Capital Expenditures, X2 = Interest Expense

    Independent Variables

    𝑪𝒂𝒔𝒉

    𝑿𝟏

    (or its log)

    𝑿𝟏𝑿𝟐

    (or its log)

    𝑿𝟐𝑴𝑬

    (or its log)

    𝑿𝟏

    𝑿𝟐 only

    [8]

    (0.04) [0.035] {0.522}

    𝒍𝒏(𝑿𝟏

    𝑿𝟐) only

    [8a]

    (-0.30) [0.071] {0.471}

    All three

    (unlogged) [9]

    (1.38) [0.311] {0.291}

    (0.21) [0.048] {0.500}

    (2.54) [0.432] {0.249}

    All three

    (log of 𝑿𝟏

    𝑿𝟐

    only) [10]

    (1.47) [0.326] {0.269}

    (0.59) [0.098] {0.444}

    (2.62) [0.451] {0.229}

    All three logged

    [11]

    (3.86) [0.947] {0.013}

    (2.74) [0.737] {0.071}

    (3.30) [0.830] {0.045}

    (average t)

    [total rejection rate]

    {average p-value}

  • 43

    Table 4, Panel I: X1 = Book Equity, X2 = Total Assets

    Independent Variables

    𝑪𝒂𝒔𝒉

    𝑿𝟏

    (or its log)