multiple regression with inequality constraints ... regression with...multiple regression with...

14
Multiple Regression with Inequality Constraints: Pretesting Bias, Hypothesis Testing and Efficiency Author(s): Michael C. Lovell and Edward Prescott Source: Journal of the American Statistical Association, Vol. 65, No. 330 (Jun., 1970), pp. 913- 925 Published by: American Statistical Association Stable URL: http://www.jstor.org/stable/2284597 Accessed: 25/06/2010 19:21 Your use of the JSTOR archive indicates your acceptance of JSTOR's Terms and Conditions of Use, available at http://www.jstor.org/page/info/about/policies/terms.jsp. JSTOR's Terms and Conditions of Use provides, in part, that unless you have obtained prior permission, you may not download an entire issue of a journal or multiple copies of articles, and you may use content in the JSTOR archive only for your personal, non-commercial use. Please contact the publisher regarding any further use of this work. Publisher contact information may be obtained at http://www.jstor.org/action/showPublisher?publisherCode=astata. Each copy of any part of a JSTOR transmission must contain the same copyright notice that appears on the screen or printed page of such transmission. JSTOR is a not-for-profit service that helps scholars, researchers, and students discover, use, and build upon a wide range of content in a trusted digital archive. We use information technology and tools to increase productivity and facilitate new forms of scholarship. For more information about JSTOR, please contact [email protected]. American Statistical Association is collaborating with JSTOR to digitize, preserve and extend access to Journal of the American Statistical Association. http://www.jstor.org

Upload: others

Post on 13-Dec-2020

26 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Multiple Regression with Inequality Constraints ... Regression with...Multiple Regression with Inequality Constraints: Pretesting Bias, Hypothesis Testing and Efficiency ... tistical

Multiple Regression with Inequality Constraints: Pretesting Bias, Hypothesis Testing andEfficiencyAuthor(s): Michael C. Lovell and Edward PrescottSource: Journal of the American Statistical Association, Vol. 65, No. 330 (Jun., 1970), pp. 913-925Published by: American Statistical AssociationStable URL: http://www.jstor.org/stable/2284597Accessed: 25/06/2010 19:21

Your use of the JSTOR archive indicates your acceptance of JSTOR's Terms and Conditions of Use, available athttp://www.jstor.org/page/info/about/policies/terms.jsp. JSTOR's Terms and Conditions of Use provides, in part, that unlessyou have obtained prior permission, you may not download an entire issue of a journal or multiple copies of articles, and youmay use content in the JSTOR archive only for your personal, non-commercial use.

Please contact the publisher regarding any further use of this work. Publisher contact information may be obtained athttp://www.jstor.org/action/showPublisher?publisherCode=astata.

Each copy of any part of a JSTOR transmission must contain the same copyright notice that appears on the screen or printedpage of such transmission.

JSTOR is a not-for-profit service that helps scholars, researchers, and students discover, use, and build upon a wide range ofcontent in a trusted digital archive. We use information technology and tools to increase productivity and facilitate new formsof scholarship. For more information about JSTOR, please contact [email protected].

American Statistical Association is collaborating with JSTOR to digitize, preserve and extend access to Journalof the American Statistical Association.

http://www.jstor.org

Page 2: Multiple Regression with Inequality Constraints ... Regression with...Multiple Regression with Inequality Constraints: Pretesting Bias, Hypothesis Testing and Efficiency ... tistical

? Journal of the American Statistical Association June 1970, Volume 65, Number 330

Theory and Methods Section

Multiple Regression with Inequality Constraints: Pretesting Bias, Hypothesis Testing and Efficiency

MICHAEL C. LOVELL and EDWARD PRESCOTT*

This article analyzes, within the context of the standard multiple regression model, the problem of handling inequality constraints specifying the signs of cer- tain regression coefficients. It is common econometric practice when regression coefficients are encountered with incorrect sign to delete the variables in question and reestimate the equation. This article shows that this procedure causes bias and can lead to inefficient parameter estimates. Furthermore, we show that grossly exaggerated statements concerning significance levels are likely to be made when other regression coefficients in the model are tested with the final regression obtained after deleting variables with incorrect sign.

1. INTRODUCTION Tossing a fair coin is not necessarily a fair game-the artful player with a

naive opponent can raise his probability of winning to 5/8 by suggesting "two out of three" whenever he fails to win on the first toss. The artful econometri- cian often fits several models experimentally to the same body of data in an attempt to find a model that looks right in terms of such criteria as the con- formity of the signs of regression coefficients to a priori notions, their signifi- cance, the Durbin-Watson statistic and the magnitude of R2. We shall consider the implications of a particular aspect of this art with regard to problems of parameter estimation and hypothesis testing.

Suppose that an econometrician is interested in the following model,

yt =- lXlt + ?/2X2t + + kXkt + Et, t =1,**,T (1.1) where xit denotes observation t on the ith independent variable, and (t is an independently distributed random variable with finite variance r-2 and zero expected value.' Now suppose that economic theory suggests as part of the maintained hypothesis the condition 31 ? 0. We wish to evaluate the following two-step procedure for using the a priori information on the sign of 1, in esti- mating the fi:

For the first step, fit by least squares the regression

yt = b,x,t + * * * + bkXkt + Ef. (1.2)

If bi!0, use the bi as estimators of the /3i. Else, if bi <0, use th e second-step estimators bi* =0 and, for i> I, the coefficients bj* obtained from the regression:

* Michael C. Lovell is professor of economics, Wesleyan University. Edward Prescott is on leave from the Department of Economics, University of Pennsylvania, as a Brookings Economic Policy Fellow, Department of Labor. The authors are indebted to Robert Suimmers, John W. Pratt and Arnold Zellner for helpful advice and comments on a preliminary draft. Computations were executed on the University of Pennsylvania comnputer. This article was supported by the Graduate School of Industrial Administration, Carnegie-Mellon University, Wesleyan University, the Wharton School of Finance and Commerce and the National Science Foundation.

1 The intercept may be introduced, in accordance wNith the standard convention, by making one of the explauna- tory variables identically equal to unity for all t.

913

Page 3: Multiple Regression with Inequality Constraints ... Regression with...Multiple Regression with Inequality Constraints: Pretesting Bias, Hypothesis Testing and Efficiency ... tistical

914 Journal of the American Statistical Association, June 1970

yt - b2*x2t + * * * + bkxk + e t*. (1.3)

Thus, the two-step estimator, call it bi, of #j is

{bi* otherwise. (1.4)

This two-step procedure seems eminently reasonable.2 It constitutes a sta- tistical application of quadratic programming.3 It is probably the approach most likely to be used in practice when regression coefficients with incorrect sign are encountered.4 Note that bi will constitute the best linear unbiased esti- mator of f3s if the Et are independently distributed with constant variance, but this does not imply that it is preferable to the non-linear estimator bi. Indeed, bj is consistent.5 Further, it is the maximum likelihood estimator of j3i if the disturbance Et, in addition to the other restrictions, is normally distributed.

For finite samples, is the two-step procedure necessarily preferable to the use of the first-step estimators, the bi that are derived without regard for the sign constraint? A simple contrived example which we present in Section 2 will con- firm that our intuition is correct if it suggests that the two-step estimators, the bi, may be biased. However, in evaluating any estimator, bias is not necessarily the most critical property to consider. We must also consider the efficiency of the two-step estimator; an appropriate measure is provided by the mean- square-error-E[(bj-fj)2]. Our numerical example will reveal that contrary to intuition, estimators derived by ignoring inequality constraints may be more efficient (have a smaller expected square error) than the two-step esti- mators bi. Since it is common econometric practice to report two-step estimators satisfying inequality constraints, it is of obvious interest to determine whether this approach, in practice, is likely to constitute an inefficient estimation procedure. In subsequent sections of this article, we shall explore this issue, and in addition, the question of bias. Our analysis will build in part on the work of Zellner [9], who was concerned with the distribution of b1 in the case of simple regression.6

It is common practice, for purposes of hypothesis testing, to report t statistics for whichever computer run yields the regression equation satisfying the sign

2 The problem is not totally unrelated to the commonly used procedure of deleting variables that have insignifi- cant regression coefficients. The analysis of Bancroft [1] and Larson and Bancroft [3] suggests that pretesting in this way can yield highly inefficient estimates of the remaining parameters.

3 There exists a considerable literature, reviewed by [2], concerning the computational problem of miniimizing the sum of squared deviations subject to inequality constraints. This article focuses on problems of estimation and hypothesis testing rather than computational aspects of "sample fitting."

4 It is not the only procedure. Some practicing econometricians, considering a wrong sign indicative of a mis- specification, would change the structure rather than drop the variable. Additional variables might be introduced and/or included variables transformed in the hope that the sign will straighten out. It is sometimes decided on the basis of the sample data to drop the sign constraint from the "maintained hypothesis," and some investigators have interpreted incorrect signs as indicative of a problem of multicollinearity that should be handled by resorting to factor analysis. While the task of evaluating these alternatives to the two-step estimation procedure deserves high priority, it lies beyond the scope of this article. Note, however, that even when the model is specified correctly, one may obtain a bi with negative sign quite frequently; e.g., in the extreme case when ,m =0, the negative sign may be encountered 50 percent of the time. The strategy of modifying the miiodel whenever bi is negative may frequently lead the investigator to abandon correctly specified models.

6 The asymptotic properties of a class of estimators of which bi is a member are developed in [4, Chapter 91. 6 Zellner's contribution is summarized in [4, p. 317]. Much of the argument of this article is readily generalized

to the case in which the switching rule hinges on whether a linear comibination of the ,i is suibject to an inequality constraint.

Page 4: Multiple Regression with Inequality Constraints ... Regression with...Multiple Regression with Inequality Constraints: Pretesting Bias, Hypothesis Testing and Efficiency ... tistical

Regression with Inequality Constraints 915

constraints imposed by theory on the model. That is, for i> 1 the hypothesis testing analogue of the two-step estimation procedure at the a-significance level is provided by the following decision rule:

Accept Ho: z = 0, i > 1 if

b> 0 and ti< t?ta

or (1.5)

bi < 0 and ti* j < t.* else reject Ho.

Here the statistic ti is computed by standard formula along with regression Equation (1.2), which contains the variable xit; ti* is computed from the second regression, which excludes the first explanatory variable; further, ta and ta* the coefficients defining the widths of the acceptance regions, are determined for desired significance level a from standard tables in the customary fashion. We demonstrate in Section 5 that the use of this two-step procedure in testing hypotheses about a model's parameters is quite likely to exaggerate significance levels; i.e., the probability of committing a Type I error, of rejecting the null hypothesis when it is true, may be much larger than the claimed significance level.

2. A CONTRIVED EXAMPLE

Here is a simple, contrived example designed to illustrate that estimates pro- vided by the two-step procedure, the bt, do not necessarily dominate the bi estimates obtained by ignoring the inequality constraint. The example involves only two observations on each of two explanatory variables. We suppose that the econometrician correctly assumes that his data is generated by the following model:

Yt -i 1 + 32X2t + Et

31 > 0 (2.1) E(Et) E(Etx2t) 2 0.

Table 1 reveals the possible sample outcomes that would be generated if the actual structure involves ,/=.5, 02 =1, and P(Et= 1) = P(Et= -1) = 0.5, while the explanatory variable takes on the values x21 =-2 and x22 =1.7 The first four columns of data on the table reveal, for each of the four possible outcomes of the experiment, the values of the dependent variables and the various esti- mators. Note that E(b1) =5/87 E(b1) =/1 = .5 and E(b2) -=41/40 E(b2) =32= 1; thus, the two-step estimators are biased'

On the figure the four Si lines indicate the regression outcome computed from each of the four equally likely samples. Because the disturbance is dis- tributed independently of xit, their average slope is unity, although of course some lines are flatter and some are steeper. Only the first sample regression line, which has a slope of unity, has an inadmissible negative intercept. The dotted

7 We have a special case of the problem posed in Section 1, with Xll =x12 =1, in order that 61 will constitute the intercept.

Page 5: Multiple Regression with Inequality Constraints ... Regression with...Multiple Regression with Inequality Constraints: Pretesting Bias, Hypothesis Testing and Efficiency ... tistical

916 Journal of the American Statistical Association, June 1970

Table 1. ILLUSTRATIVE EXAMPLE

Four possible samples; P(Si) =1 /4 Statistic Expected value

Si S2 ka 54

Disturbance el -1 -1 1 1 E(ei) =0 O2 -1 1 -1 1 e2)=0

Dependent yi -5/2 -5/2 -1/2 -1/2 E(yi) =-3/2 variable Y2 1/2 5/2 1/2 5/2 E(y2) =3/2

bi -1/2 5/6 1/6 3/2 E(b) =1/2 =,1 0 5/6 1/6 3/2 E(,l) =5/8 09,

Estimate b2 1 5/3 1/3 1 E(bN) = 1 =.82 b2* 11/10 15/10 3/10 7/10 E(b2*) =9/10 0,82 b2 11/10 5/3 1/3 1 E(b2) =41/40 i2

(bi -,61)2 1 1/9 1/9 1 E(bi -,6)2=MSE(bi) =5/9 MSE (tl -#1)2 1/4 1/9 1/9 1 E(i-,1) 2=MSE(1l) =53!144 compu- (b2-,02)2 0 4/9 4/9 0 E(b0a-62)2= MSE(bW) =2/9 tation (b2*- 2)2 1/100 1/4 49/100 9/100 E(b2* -2)2=MSE(b2*) =21/10

(f,9 2 _162) 2 1/100 4/9 4/9 0 E(b2-%2)2=MSE(b2) =2/9+1/400

S1* line indicates the second-step regression computed with the intercept ex- cluded; line 81* is steeper than unity because point x21 is further than x22 from the origin. Consequently, the invocation of the sign constraint on the intercept serves to replace, with probability of 1/4, the line S1, which has a slope equal to 2, with the steeper line S1*. Since an error is introduced whenever Sample 1 is

drawn, but none of the other sample outcomes are changed, it is clear that in this instance the consequence of the two-step procedure is inefficiency as well as bias.

From the point of view of efficiency, we note that for this example b1 is preferable to b1 as an estimator of ,1 because it has the smaller mean square error (MSE); that is, MSIE(bi)=E[(b1-,a1)2I=5/9 while with the two-step procedure we have MSE(b1) =53/144. The two-step procedure thus serves to cut by about one-third the expected loss in estimating 11, if the statistician is penalized in proportion to the square of his estimation error. For the other parameter, however, we have M\ISE(b2)= 2/9 while M\SE(b2) =2/9 + 1/400. In this particular case, the unfortunate effect of utilizing the two-step procedure is to increase the mean-square-error.8

We explore in subsequent sections the pros and cons of the two-step estima- tion procedure. Section 3 derives a precise expression for the bias involved in the two-step procedure and shows how the econometrician can place bounds on the extent of the bias involved in practical applications. Then, Section 4 demonstrates that a quite natural restriction will insure that even for small samples the two-step estimation procedure yields more efficient estimators than those obtained by ignoring the inequality constraints. Section 5 is concerned with problems of hypothesis testing.

8 Rothenberg [71 independently constructed an example in which he also showed that MSE (12) can exceed MSE(b2). He conjectures that it the disturbance is normal and 6 is constrained to a convex set, the constrained least squares estimator cannot have larger expected quadratic loss than the unconstrained estimator. In Section 4, we prove the validity of a restrictive case for this conjecture.

Page 6: Multiple Regression with Inequality Constraints ... Regression with...Multiple Regression with Inequality Constraints: Pretesting Bias, Hypothesis Testing and Efficiency ... tistical

Regression with Inequality Constraints 917

GRAPH FOR ILLUSTRATIVE EXAMPLE

sz 22

3~~~~~~~s

/ E ~~(y)

2 2~~~~~~~~~ 1 i x~~~~~~~~~~~~~~~~~~

A/X53~~~~~~~

-3

3. BIAS The bias in the two-step estimation procedure is most easily analyzed with

the aid of matrix notation. If we let y = col(yt), xj= col(xjt), E = col(et), X2 = [x2, * , Xk], and 52 = col($2, . . ) where col indicates a column vec- tor, we can rewrite (1.1) as

y = X1f1 + X252 + E- (3.1)

We now transform the problem into orthogonal form by defining the vector *=-X1-X2(X2'X2)-tX2'x1=x -X2B1.2, where B1.2 is the vector of k-1 coeffi-

Page 7: Multiple Regression with Inequality Constraints ... Regression with...Multiple Regression with Inequality Constraints: Pretesting Bias, Hypothesis Testing and Efficiency ... tistical

918 Journal of the American Statistical Association, June 1970

cients from the auxiliary regression of the first explanatory variable on the remaining. We then note that xl=xl*+X1B1.2, and obtain on substitution into (3.1):

y = X1*01 + X2(02 + B1.2j31) + E (3.2)

Since X2'xl*=O, application of least squares to (3.2) yields the estimates

b= (xi*/xj*)-xl*'y (3.3)

B2 + B1.2b, = (X2'X2)-lX2'y. (3.4)

Now the vector on the right of (3.4) is the set of regression coefficients of the second-step regression involving the deletion of the first variable, which we denote as B2*-col(b2*, bk*). Thus we may rewrite (3.4) as:

B2 = B2 + B1.2b4. (3.5)

Further, our two-step estimator is

b = max(0, b1) = b- min(0, b1)

and the k -1 component column vector

B2 col(b2, ... , bk) = B2 + min(O, bl)B1.2. (3.6)

To analyze the bias, note that E(b) = -1 and E(B2) = 92; therefore:

E , - 5 = E[min(O, b1)] (3.7)

Now we note that E [min(0, b1) ] <0, unless b < 0 is impossible (b' = b1). Except in this uninteresting case, bi has positive bias. Furthermore,

E(bi)- At = E[min(O, bi)Jbi.j. (3.7')

Here b1.i denotes the relevant component of vector B1.2; i.e. B1.2 = col(b1.2, * *, bl.k). Thus, the remaining coefficients will be biased unless either b1 is unbiased or the ith explanatory variable enters with zero coefficient in the auxiliary re- gression of the first explanatory variable on the remaining k -1 explanatory variables.

If the disturbance term e in equation (1.1) is normally distributed, tables in [9, p. 10] which are reproduced in [4, p. 317] can be conveniently used to com- pute the bias as a function of the unknown parameter i31, the variance u,2 of the disturbances, and the explanatory variables.9 When i31 is not known, a bound on the bias is implied by the knowledge that this parameter is non- negative, for we note that when 1 = 0 the bias is maximum and then

E(b1)- = - E[min(O, b1)] = -4crb1, (3.8)

and for i> 1,

E(bi) - f3i = bl.iO'b1/v2wr .40blb .i. (3.9)

9 The following formula obtained by integration, may be used to compute the bias:

E(bi) - pi = [sF(s) -f b,bii

where a =f81/Tlb and f and F are the normal density and distribution functions respectively. This is the expression used by Zellner [9] to compute his table; Raiffa and Schlaifer [6, p. 3561 tabulate the expression in brackets.

Page 8: Multiple Regression with Inequality Constraints ... Regression with...Multiple Regression with Inequality Constraints: Pretesting Bias, Hypothesis Testing and Efficiency ... tistical

Regression with Inequality Constraints 919

4. EFFICIENCY Let us now compare estimator bi with the two-step estimator bi in terms of

the mean-square-error criterion. Our numerical example illustrated that it is possible for bt to have a larger mean square error than bi. But we shall demon- strate that bi will necessarily have the smaller mean square error if we require, as an additional restriction, that the disturbance e be normally distributed.

In order to compute the difference in mean square error, we note

(bi - )2- (i - )2 (bi - i)(bi + i - 2/i). (4.1)

If biO0, bi=bi and this expression is 0. If bi<O, b-i =bi* =-bi+b,.ib, so we obtain

MSE(bi) - MSE(bi) = E[(bi - )2 - (b6 _ -)2]

= P[bi < O]Ei -b,.ib1(2bi* a 2E(bi*) (4.2)

- b.ibi + 2b1.j03) I bi < 0].

Because b, and bi* are independently distributed,'0 taking the expectation yields

MSE(bi) - MSE(b) -P(b, < O)b21.iE[(b12- 2b,01) 1 bi < 0]. (4.3)

This expression is necessarily non-negative; for the term inside the expectations operator cannot be negative because of the condition on bi. We have thus established that the two-step estimation procedure is more efficient than the unconstrained estimator, bi. The relative gain in mean square error is

MSE(bi) - MSE(bi)

MSE(bi) = P(bi < O)b21.iE[(b12- 2b,01) I bi < O]/MSEbi

(0 12 - 2-rti = pl2 j -2_- exp[- (t, - 7)2/2]dt, (4.4)

= 2ii exp(-T /2) + [1 - r] fT 2 exp( x2/2)dx]

= P2,l[Tf(r) + (1 - T2)F(-T)],

where t, = b,/o-b1, r ,B1/0b, MSE(bi) = (T2bi, pl1-- bbl,.ib,/OTb1, and f and F are

the normal density and distribution functions. Consideration of (4.4) reveals that the reduction in the mean square error

obtained by using the two-step estimator under conditions of normality is proportional to p2lk, which can be calculated from the matrix of exogenous variables." Unless xit and xit display a high degree of colinearity, p21 will be small which implies the gain in efficiency will be small. Given p2lj, the gain in efficiency is maximized when /,1 =0; indeed, for 1, = 0 we find that the relative

15 Both be* and bi are linear combinations of the disturbance terms and therefore are jointly normal. Since uncorrelated jointly normal variates are independent, it suffices to show the covariance between bi* and bi is 0. From (3.3) and (3.5) we derive

E{ [bi - #I] [B E$)j' E{ [(Xl*'Xl*)-IXs'ef XI'X2) '1X']'m 0

once we remember x4VX2 =0. 11 Let C =[X'X] 1; then Pai =culil(uciil/2.

Page 9: Multiple Regression with Inequality Constraints ... Regression with...Multiple Regression with Inequality Constraints: Pretesting Bias, Hypothesis Testing and Efficiency ... tistical

920 Journal of the American Statistical Association, June 1 970

gain in mean square error is .5p2li. As i3l increases the efficiency gain decreases; thus, for 31/0b, = 1, the difference is approximately p2il1O.

5. HYPOTHESIS TESTING

The true significance level of a test constructed at the .05 "significance level" by the two-step hypothesis testing procedure defined by equation (1.5) is revealed in Table 2. Clearly, the two-step hypothesis testing procedure may lead to exaggerated claims of significance. Inspection of the table reveals that the true probability of incorrectly rejectiing the null hypothesis depends on both pli and I1/Tb,. The entries in Table 2 were calculated under the assumption that the disturbance e in equation (1.1) is normally distributed. Although the probabilities reported are for the case in which the investigator appropriately incorporates knowledge of cTe2 in applying the two-step testing procedure, es- sentially the same results are obtainied in the case of unknown variance with as few as six degrees of freedom.12

Table 2. PROBABILITY OF INCORRECTLY REJECTING Ho O5=0 WITH THE TWO-STEP TEST AT THE FIVE PERCENT

SIGNIFICANCE LEVEL

pit! f3~~~~~~~~~1/ofb1 0 .25 .50 1.00 1.50 2.0

0 .05 .05 .05 .05 .05 .05 .50 .05 .05 .05 .05 .05 .05 .75 .05 .05 .05 .06 .05 .05 .90 .05 .06 .08 .11 .09 .06 .99 .05 .19 .31 .18 .09 .05

Table 2 reveals that the two-step hypothesis testing procedure exaggerates the significance level only when I Pil4 is large and, at the same time, f1/-bl is of moderate magnitude. To see why this is so, recall first that under the null hypothesis

P ( |t > ta) =aot.

If, in addition, Pii = 0, then also t= ti* = ?i.13 which explains why entries in the top row of Table 2 are all 0.05. Table 2 reveals that this effect persists even for pli = 0.75.

Trouble can arise, however, with the larger values of I Pl4 frequently en- countered when dealing with highly collinear economic time series. To appre- ciate the circumstances in which the two-step procedure overstates the true significance level, we note that the probability of rejecting Ho: fi=0 with the two-step hypothesis testing procedure is

P( t > te) -P(|t*| > ta* bi < O) +P(| ti! > t 'n bi? 0). (5.1)

12 When the investigator knows a, he may use the normal distribution to determine t,a, which defines the width of the acceptance region; further, a2bi =Cjij2e and ti =bi/oa. When a- is unknown the investigator muist substitute its unbiased estimator and utilize the t distribution. The appendix analyzes the additional complications for the two-step hypothesis testing procedure in the case of unknown variance.

13 When.pii =0, bi =bi* =b.i, but this implies ti =ti* = ti only in the case of known oa. The result is only approxi- mately true when a- is not known. See the appendix for further explanation.

Page 10: Multiple Regression with Inequality Constraints ... Regression with...Multiple Regression with Inequality Constraints: Pretesting Bias, Hypothesis Testing and Efficiency ... tistical

Regression with Inequality Constraints 921

From Equation (3.5) we observe that bi* will be subject to substantial bias when Pli and fi both depart considerably from zero; in these circumstances, P(I ti* I > ta*) will exceed the claimed significance level by a substantial margin, and it is not surprising, in light of (5.1), that the two-step hypothesis testing procedure may incorrectly reject the null hypothesis with excessive frequency.

Even when the problem of collinearity is severe, the two-step test may in- correctly reject the null hypothesis with the claimed probability. In particular, when 0B = fi = 0, from (3.5) the mean of bi* is zero. It is not surprising that under these conditions the two-step test has rejection probability a.14 This explains why all the entries in the first column of Table 2 equal 0.05 regardless of the degree of multicollinearity. At the other extreme, as 1/o0-b, becomes large, P(b1<O) approaches zero and the two-step test becomes equivalent to the use of ti; thus, the actual probability of rejecting the null hypothesis approaches the claimed significance level as 3l/-b1 becomes sufficiently larger.

It is necessary to concede that Table 2 cannot be used in practical applica- tions to determine the true significance level generated by the two-step hypoth- esis testing procedure. An investigator can compute Pli from the matrix X of observations on the explanatory variables, but this reveals only the relevant row of Table 2. The true significance level of the two-step hypothesis testing procedure cannot be determined because only the sign rather than the magni- tude of ,13 is known. This table does reveal that the true significance level cannot depart too much from the stated level when the data are not very collinear. But the bottom rows of this table show that the range of indeterminancy may be quite large when dealing with highly correlated economic time series. Al- though an investigator willing to invoke a subjective probability distribution on the parameter i1 can evaluate the significance of the two-step testing pro- cedure, he would want to formulate the problem from the begiining in Bayesian rather than classical terms.

Since it is impossible in practical applications to determine the true signifi- cance level of the two-step hypothesis testing procedure, it is essential to con- sider the possibility of always using the statistic ti provided by regression equa- tion (1.2), which includes xlt. An investigator who uses ti for his test, regardless of the sign of b1, is in effect ignoring the a priori information that j1 is non-nega- tive, but at least he knows the true significance level of his test. That the costs of ignoring a priori information may not be high in this particular application is suggested by the comparison of alternative test procedures presented in Table 3. For each test Table 3 reveals how the probability of rejecting the null hypothesis varies with fl/0-bl and 13i/0bi when pli = 0.9. The upper part of Table 3 reveals the probability of rejecting Ho: fi=O for the two-step testing pro- cedure at a claimed significance level of a= 0.05. The last two rows report the probabilities of rejecting the null hypothesis using the t, statistic provided by regression (1.2) regardless of the sign of b1 at the 0.05 and the 0.11 significance levels; both these tests ignore the a priori information that 013>O. We note that the two-step hypothesis testing procedure has a higher probability of rejecting the null hypothesis when it is false than the simpler five percent significance level test that uses the results of regression (1.2) regardless of the sign of b1. This does not mean that the two-step procedure dominates; the true significance

14 A rigorous argument for this result can be found in the appendix.

Page 11: Multiple Regression with Inequality Constraints ... Regression with...Multiple Regression with Inequality Constraints: Pretesting Bias, Hypothesis Testing and Efficiency ... tistical

922 Journal of the American Statistical Association, June 1 970

Table 3. ALTERNATIVE POWER FUNCTIONS: PROBABILITY OF REJECTING Ho: -0 WHEN pli =.90

Probability v l/Tbi

-3.0 -2.0 -1.0 0.0 1.0 2.0 S.0

0 .85 .58 .32 .05 .48 .93 1.00 .25 .85 .55 .34 .06 .35 .87 1.00

P(I|t tI >?.05) .50 .85 .53 .29 .08 .24 .79 .99 1.00 .85 .52 .21 .11 .18 .63 .96 2.00 .85 .52 .17 .06 .18 .52 .87

P(ItiI >t.05) - .85 .52 .17 .05 .17 .52 .85

P(ti| >t.11) - .92 .65 .28 .11 .28 .65 .92

level of that test is not 0.05. For example, when 31/rb= 1.0, the probability of incorrectly rejecting Ho with the two-step test is eleven percent rather than five percent. A more meaningful comparison is facilitated with the test on the bot- tom row of Table 3 for a = -.11; it has approximately the same significance level as the two-step test. Observe that neither test dominates the other. It is unfor- tunate that in practical applications the very nature of the problem is such as to deny the investigator the information required to determine which pro- cedure is more powerful. The test that ignores the sign condition does have the very important advantage that tables are available that can be used to determine the acceptance region yielding the claimed significance level.

Additional information on the relative merits of the two alternative testing procedures is presented in Table 4. Because only a moderate degree of multi- collinearity is involved, the two-step test is significant at the claimed level of five percent. The two-step test is not uniformly more powerful. Indeed, neither test dominates the other. Clearly, it is not inappropriate to use the two-step hypothesis testing procedure when I pi4i is of small or moderate magnitude.

6. CONCLUSION

Our analysis suggests that for purposes of least square parameter estimation it is appropriate practice to delete a variable with incorrect sign. Although this customary procedure generates bias, it leads to more efficient parameter esti-

Table 4. POWER FUNCTIONS WHEN p1X=.50

fli/O-b Probability fl/crb1

-3.0 -2.0 -1.0 0.0 1.0 2.0 3.0

0 .85 .49 .14 .05 ,24 .66 .94 .25 .85 .50 .15 .05 .22 .62 .93

P(fhj >t.05) .50 .85 .51 .16 .05 .20 .59 .91 1.00 .85 .52 .17 .05 .18 .55 .88 2.00 .85 .52 .17 .05 .17 .52 .86

P(Itil >t.05) - .85 .52 .17 .05 ,17 .52 .85

Page 12: Multiple Regression with Inequality Constraints ... Regression with...Multiple Regression with Inequality Constraints: Pretesting Bias, Hypothesis Testing and Efficiency ... tistical

Regression with Inequality Constraints 923

mates when the error term is normally distributed as well as independent of the explanatory variables. The customary procedure for hypothesis testing when sign constraints are present is inappropriate when dealing with highly collinear economic tiine series, for it can lead to substantially exaggerated claims of sig- nificance. A valid test, which may be more powerful than the customary pro- cedure, is to always report the t statistics calculated from the initial regression regardless of the signs of the parameter estimates.

APPENDIX: HYPOTHESIS TESTING

In order to simplify the analysis, we normalize by measuring Yt, x)t, and xIt in units such that e= bj* = =1. Nonetheless, the results of this appendix are completely general for the arguments are presented in terms of parameters which are invariant to the required transformation. We assume e, the distur- bance term, has unknown variance and, of course, that it is normally distributed.

Let S be the sum of the squared residuals for regression (1.2), which includes xit. It has a chi-square distribution with T - k degrees of freedom and is inde- pendent of both bi and bi. Further, since b,* is a linear combination of b1 and bi, S is independent of bi* as well. Using (3.5), we find

bi/(Tbi = (1 + pli2) l12bc* + plibi. (A. 1)

Consideration of (3.2) suggests the vector of residuals for regression (1.2) is

e = Y - X2B2*- xl*bl, (A.2)

and for regression (1.3), which excludes xit,

e* = y - XR2B2* (A.3)

Since S=e'e, S* the sum of the squared residuals for the regression which ex- cludes xit is

S* = e*'e* = (e + xi*xl*bi)'(e + xi*bi) (A.4)

-S + blxl*'xl*bl + 2bixi*'e. But,

le= x1*(y - X2B2*- xi*bi) = x*'y -xlxl*bi- 0 (A.5)

as xl*1X2=0, xi$'xl *j 1 and bi (xl*'xi)-lxl*'y. Thus, we obtain the result,

S* = S + b12.

Given the above discussion, the relevant t statistics for the two-step test can be represented as

t bi/ab p(1l-pi2) l12bi* + plibi t= =-(A.6)

[S/(T - c)]112 [S/(T - k) ]112 and

t.* _ bj* __ bi . (A.7) [S*/(T - k + I)J 1/2 [(S + b12)/(T - k + 1) ]1/2

The random variables S, b1, and bi* are independently distributed. S has a chi- square distribution with T - k degrees of freedom, while bi and bj* are normally

Page 13: Multiple Regression with Inequality Constraints ... Regression with...Multiple Regression with Inequality Constraints: Pretesting Bias, Hypothesis Testing and Efficiency ... tistical

924 Journal of the American Statistical Association, June 1970

distributed with means 1l/0b, and (fli/0bi-Plif1/Jb1)/(1- p21t) 1/2 respectively and variances 1.

When = i= 0, the likelihood of (ti, b1) equals the likelihood of (- ti, - b1). This implies P(| tI > tnbi,0) =p(I tij >t )/2= a/2. For tVi and bi, the almost identical argument implies P(| ti*> t*b1 <0) = a/2 as well. Substituting into (5.1), we find the probability of incorrectly rejecting Ho when ,1 = 0 is a, the claimed significance level for the two-step test.

That the two-step testing procedure is approximately valid when Pli=O is easily demonstrated once we note that this condition implies that b1 and bi are independently distributed; since S is also independently distributed, we have under Ho

P( I ti I > tG n bi > 0) P(bi > O)P( I ti I > t,) = P(bi > O)a.

Furthermore,

P( I ti* I > ta n bi < O) -P(b, < O)P( I ti* I > t,* I bi < O) P(b1 < O)P( I ti* I > ta*) = P(b1 < O)a,

provided there are more than a few degrees of freedom. Hence, we have upon substitution into (5.1) that P(| ti > t) a under the null hypothesis when Pli = 0.

To see how the entries in Tables 2, 3, and 4 were calculated, we first define

p(bi, S) p( jPt *I > ta I bil S) for bi > 0 (A.8)

Inspection of (A.6) and (A.7) reveals that this function can be evaluated using the normal distribution function. Next we observe

P(I t > t?) = E[p(b1, S)]. (A.9)

This expression was evaluated using numerical integration for several of the parameters Pli, lf1/0,b and fi/b biwhen T - k- =6 and 18 as well as when ?e is known. In this latter case S will not be an argument of p(, ) as t, = bi//bi and ti* = bi*. Note that all these parameters are invariant to the units in which the variables are measured. The difference in power between the two-step test and the simple test which uses only ti and ignores the sign constraint is approx- imately the same when T - k- =6 and 18 and when the variance of the distur- bance term is known. We were not surprised to find the power is greatest for known variance and least for T - k = 6 and unknown variance. When T - k = 6 the significance level of the test is not exaggerated by as great a margin as when there are more degrees of freedomn.

REFERENCES

[1] Bancroft, T. A. "On Biases in Estimation Due to the Use of Preliminary Tests of Significance," The Annals of Mathematical Statistics, 15 (1944), 190-204.

[2] Judge, G. G. and Takayama, T., "Inequality Restrictions in Regression Analysis," Journal of the American Statistical Association, 61 (1966), 166-81.

Page 14: Multiple Regression with Inequality Constraints ... Regression with...Multiple Regression with Inequality Constraints: Pretesting Bias, Hypothesis Testing and Efficiency ... tistical

Regression with Inequality Constraints 925

[3] Larson, Harold J. and Bancroft, T. A., "Sequential Model Building for Prediction in Regression Analysis," The Annals of Mathematical Statistics, 34 (1963), 462-79.

[41 Malinvaud, E., Statistical Methods of Econometrics, Chicago: Rand McNally, 1966. [5] Park, Soo Bin, "Two-Stage LeastSquares Estimation with Inequality Constraints oa

the Coefficients," Unpublished paper presented at the Evanston, Illinois, meetings of the Econometric Society, 1968.

[6] Raiff a, H. and Schlaifer, R. Applied Statistical Decision T'heory, Cambridge, Mass.: Division of Research, Harvard Business School, 1961.

[71 Rothenberg, Thomas, "Estimation with Inequality Restrictions," Heverlee, Belgium. for Operations Research and Econometrics, Discussion Paper No. 6825, 1968.

[8] Zellner, A., "Linear Regression with Inequality Constraints on the Coefficients: An Application of Quadratic Programming and Linear Decision Rules," Report 6109, Rotterdam: International Center for Management Science, 1961.

[9] - and Goldberger, A. "On Pure and Mixed Statistical Estimation in Economics," International Economtic Review, 2 (1961), 65-78.