Download attachment

4/9/2009
QM 6203: Econometrics I
Mohd‐Pisal Zainal, Ph.D
Department of Banking
INCEIF
Basic Econometrics
Chapter 8
MULTIPLE REGRESSION ANALYSIS: The Problem of Inference
1
4/9/2009
Chapter 8
MULTIPLE REGRESSION ANALYSIS:
The Problem of Inference
8-3. Hypothesis testing in multiple regression:
•Testing hypotheses about an individual partial regression coefficient
•Testing the overall significance of the estimated multiple regression model, that is, finding out if all the partial slope coefficients are simultaneously equal to zero
•Testing that two or more coefficients are equal to one another
•Testing that the partial regression coefficients satisfy certain restrictions
•Testing the stability of the estimated regression model over time or in different cross‐sectional units
•Testing the functional form of regression models
Chapter 8
MULTIPLE REGRESSION ANALYSIS:
The Problem of Inference
8-4. Hypothesis testing about individual partial regression coefficients
With the assumption that u i ~ N(0,σ2) we can use t‐test to test a hypothesis about any individual partial regression coefficient.
H0: β2 = 0
H1: β
: β2 # 0
If the computed t value > critical t value at the chosen level of significance, we may reject the null hypothesis; otherwise, we may not reject it
2
4/9/2009
Chapter 8
MULTIPLE REGRESSION ANALYSIS:
The Problem of Inference
8‐5. Testing the overall significance of a multiple regression: The F‐Test
i
Th F T t
For Yi = β1 + β2X2i + β3X3i + ........+ βkXki + ui
• To test the hypothesis H0: β2 =β3 =....= βk= 0 (all slope coefficients are simultaneously zero) versus H1: Not at all slope coefficients are simultaneously zero, compute
•F=(ESS/df)/(RSS/df)=(ESS/(k‐1))/(RSS/(n‐k)) (8.5.7) (k = total number of parameters to be estimated including intercept)
• If F > F critical = Fα(k‐1,n‐k), reject H0
• Otherwise you do not reject it
Chapter 8
MULTIPLE REGRESSION ANALYSIS:
The Problem of Inference
8‐5. Testing the overall significance of a multiple regression
z Alternatively, if the p‐value of F obtained from (8.5.7) is sufficiently low, one can reject H0
z An important relationship between R2 and F:
F=(ESS/(k‐1))/(RSS/(n‐k)) or
R2/(k‐1)
/(k 1)
F = ‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐
(1‐R2) / (n‐k) ( see prove on page 249)
(8.5.1)
3
4/9/2009
Chapter 8
MULTIPLE REGRESSION ANALYSIS:
The Problem of Inference
8‐5. Testing the overall significance of a multiple regression in terms of R
in terms of R2
For Yi = β1 + β2X2i + β3X3i + ........+ βkXki + ui
z To test the hypothesis H0: β2 = β3 = .....= βk = 0 (all slope coefficients are simultaneously zero) versus H1: Not at all slope coefficients are simultaneously zero, compute
z F = [R
F = [R2/(k‐1)] / [(1‐R
/(k‐1)] / [(1‐R2) / (n‐k)] (8.5.13) (k = total ) / (n‐k)] (8 5 13) (k = total
number of parameters to be estimated including intercept)
z If F > F critical = F α, (k‐1,n‐k), reject H0
Chapter 8
MULTIPLE REGRESSION ANALYSIS:
The Problem of Inference
8‐5. Testing the overall significance of a multiple regression
zAlternatively, if the p‐value of F obtained from (8.5.13) is sufficiently low, one can reject H0
The “Incremental” or “Marginal” contribution of
an explanatory variable:
Let βX is the new ((additional)) term in the right
g
hand of a regression. Under the usual
assumption of the normality of ui and the HO: β
= 0, it can be shown that the following F ratio
will follow the F distribution with respectively
degree of freedom
4
4/9/2009
Chapter 8
MULTIPLE REGRESSION ANALYSIS:
The Problem of Inference
8‐5. Testing the overall significance of a multiple regression
[R2new ‐ R2old] / Df1
F com = ‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐
(8.5.18)
2
[1 ‐ R new] / Df2
Where Df1 = number of new regressors
Df2 = n – number of parameters in the new model
new model
R2new is standing for coefficient of determination of the new regression (by adding βX);
R2old is standing for coefficient of determination of the old regression (before adding βX).
Chapter 8
MULTIPLE REGRESSION ANALYSIS:
The Problem of Inference
8‐5. Testing the overall significance of a multiple regression
g
Decision Rule:
If F com > F α, Df1 , Df2 one can reject the Ho that β = 0 and conclude that the addition of X to the model significantly increases ESS and hence the R2 value
f || f
ff
™ When to Add a New Variable? If |t| of coefficient of X > 1 (or F= t 2 of that variable exceeds 1)
™ When to Add a Group of Variables? If adding a group of variables to the model will give F value greater than 1;
5
4/9/2009
Chapter 8
MULTIPLE REGRESSION ANALYSIS:
The Problem of Inference
8‐6. Testing the equality of two regression coefficients
Yi = β1 + β2X2i + β3X3i + β4X4i + ui
(8.6.1)
Test the hypotheses:
(8.6.2)
H0: β3 = β4 or β3 - β4 = 0
H1: β3 # β4 or β3 - β4 # 0
Under the classical assumption it can be shown:
t = [(β^3 - β^4) – (β3 - β4)] / se(β^3 - β^4)
follows the t distribution with (n-4) df because (8.6.1) is a fourvariable model or
or, more generally,
generally with (n
(n-k)
k) df.
df where k is the
total number of parameters estimated, including intercept
term.
se(β^3 - β^4) = √ [var((β^3) + var( β^4) – 2cov(β^3, β^4)]
(see appendix)
(8.6.4)
Chapter 8
MULTIPLE REGRESSION ANALYSIS:
The Problem of Inference
t = (β^3 - β^4) / √ [var((β^3) + var( β^4) – 2cov(β^3, β^4)]
(8.6.5)
Steps for testing:
1. Estimate β^3 and β^4
2. Compute se(β^3 - β^4) through (8.6.4)
3. Obtain t‐ ratio from (8.6.5) with H0: β3 = β4
4. If t‐computed > t‐critical at designated level of significance for given df, then reject H0. Otherwise do not reject it. Alternatively, if the p‐value of t statistic from (8.6.5) is reasonable low, one can reject H0.
•
Example 8.2: The cubic cost function revisited
6
4/9/2009
Chapter 8
MULTIPLE REGRESSION ANALYSIS:
The Problem of Inference
8‐7. Restricted least square: Testing linear
equality restrictions
Yi = β1X β22i X β33i eui (7.10.1) and (8.7.1)
Y = output
X2 = labor input
X3 = capital input
In the log‐form:
l Yi = β
lnY
β0 + β
+ β2lnX
l X2i + β
+ β3lnX
l X3i + u
+ i (8.7.2)
(8 7 2)
with the constant return to scale: β2 + β3 = 1 (8.7.3)
Chapter 8
MULTIPLE REGRESSION ANALYSIS:
The Problem of Inference
8‐7. Restricted least square: Testing linear
equality restrictions
How to test (8.7.3)
z The t Test approach (unrestricted): test of the hypothesis H0: β2 + β3 = 1 can be conducted by t‐ test:
t = [(β^2 + β^3) – (β2 + β3)] / se(β^2 - β^3)
(8.7.4)
z The F Test approach (restricted least square ‐RLS): Using, say, β2 = 1‐β3 and substitute it into (8.7.2) we get: ln(Yi /X2i) = β0 + β3 ln(X3i /X2i) + ui (8.7.8). Where (Yi /X2i) is output/labor ratio, and (X3i / X2i) is capital/labor ratio
7
4/9/2009
Chapter 8
MULTIPLE REGRESSION ANALYSIS:
The Problem of Inference
8‐7. Restricted least square: Testing linear equality restrictions
Σu^2UR=RSSUR of unrestricted regression (8.7.2) of unrestricted regression (8 7 2)
2
and Σ u^ R = RSSR of restricted regression (8.7.7), m = number of linear restrictions,
k = number of parameters in the unrestricted regression,
n = number of observations.
2
2
R UR and R R are R2 values obtained from unrestricted and restricted regressions respectively. Then
F=[(RSSR – RSSUR)/m]/[RSSUR/(n‐k)] = = [(R2UR – R2R) / m] / [1 – R2UR / (n‐k)] (8.7.10)
follows F distribution with m, (n‐k) df.
Decision rule: If F > F m, n‐k , reject H0: β2 + β3 = 1
Chapter 8
MULTIPLE REGRESSION ANALYSIS:
The Problem of Inference
8‐7. Restricted least square: Testing linear equality restrictions
2
•
Note: R UR ≥ R2R
(8.7.11)
2
2
(8.7.12)
•
and Σ u^ UR ≤ Σ u^ R
z Example 8.3: The Cobb‐Douglas Production function for Taiwanese Agricultural Sector, 1958‐1972. (pages 259‐260). Data in Table 7.3 (p g
(page 216)
)
z General F Testing (page 260)
z Example 8.4: The demand for chicken in the US, 1960‐1982. Data in exercise 7.23 (page 228)
8
4/9/2009
Chapter 8
MULTIPLE REGRESSION ANALYSIS:
The Problem of Inference
8‐8. Comparing two regressions: Testing for structural y
g
stability of regression models
Table 8.8: Personal savings and income data, UK, 1946‐
1963 (millions of pounds)
Savings function:
z Reconstruction period:
Y t = α1+ α2X t + U1t (t = 1,2,...,n1)
z Post‐Reconstruction period:
Y t = β1 + β2X t + U2t (t = 1,2,...,n2)
Where Y is personal savings, X is personal income, the us
are disturbance terms in the two equations and n1, n2 are the number of observations in the two period Chapter 8
MULTIPLE REGRESSION ANALYSIS:
The Problem of Inference
8‐8. Comparing two regressions: Testing for structural stability of regression models
regression models
+ The structural change may mean that the two intercept are different, or the two slopes are different, or both are different, or any other suitable combination of the parameters. If there is no structural change we can combine all the n1, n2 and just estimate one savings function as:
Y t = λ1 + λ2X t + Ut (t = 1,2,...,n1, 1,....n2). (8.8.3)
How do we find out whether there is a structural change in the savings‐income relationship between the two period? A popular test sa
ings income relationship bet een the t o period? A pop lar test
is Chow‐Test, it is simply the F Test discussed earlier
HO: αi = βi ∀i Vs H1: ∃i that αi # βi
9
4/9/2009
Chapter 8
MULTIPLE REGRESSION ANALYSIS:
The Problem of Inference
8‐8. Comparing two regressions: Testing for structural stability of regression models
regression models
+ The assumptions underlying the Chow test
u1t and u2t ~ N(0,s2), two error terms are normally distributed
with the same variance
u1t and u2t are independently distributed
Step 1: Estimate (8.8.3), get RSS, say, S1 with df = (n1+n2 – k);
k is number of parameters estimated )
Step 2: Estimate (8.8.1)
(8 8 1) and (8.8.2)
(8 8 2) individually and get their
RSS, say, S2 and S3 , with df = (n1 – k) and (n2-k) respectively.
Call S4 = S2+S3; with df = (n1+n2 – 2k)
Step 3: S5 = S1 – S4;
Chapter 8
MULTIPLE REGRESSION ANALYSIS:
The Problem of Inference
8‐8. Comparing two regressions: Testing for structural stability of regression models
Step 4: Given the assumptions of the Chow Test, it can be
show that
(8.8.4)
F = [S5 / k] / [S4 / (n1+n2 – 2k)]
follows the F distribution with Df = (k, n1+n2 – 2k)
Decision Rule: If F computed by (8.8.4) > F- critical at the
chosen level of significance a => reject the hypothesis that the
regression (8.8.1) and (8.8.2) are the same, or reject the
hypothesis of structural stability; One can use p-value of the F
obtained from (8.8.4) to reject H0 if p-value low reasonably.
+ Apply for the data in Table 8.8
10
4/9/2009
Chapter 8
MULTIPLE REGRESSION ANALYSIS:
The Problem of Inference
8-9. Testing the functional form of
regression:
Choosing between linear and log‐linear regression models: MWD Test (MacKinnon, White and Davidson)
H0: Linear Model : Linear Model Y is a linear function of Y is a linear function of
regressors, the Xs;
H1: Log‐linear Model Y is a linear function of logs of regressors, the lnXs;
Chapter 8
MULTIPLE REGRESSION ANALYSIS:
The Problem of Inference
8-9. Testing the functional form of regression:
Step 1: Estimate the linear model and obtain the
estimated Y values. Call them Yf (i.e.,Y^). Take lnYf.
Step 2: Estimate the log-linear model and obtain the
estimated lnY values, call them lnf (i.e., ln^Y )
Step 3: Obtain Z1 = (lnYf – lnf)
Step 4: Regress Y on Xs and Z1. Reject H0 if the
coefficient of Z1 is statistically significant, by the usual t test
Step 5: Obtain Z2 = antilog of (lnf – Yf)
Step 6: Regress lnY on lnXs and Z2. Reject H1 if the
coefficient of Z2 is statistically significant, by the usual ttest
11
4/9/2009
Chapter 8
MULTIPLE REGRESSION ANALYSIS:
The Problem of Inference
Example
p 8.5: The demand for Roses (p
(page
g 266267). Data in exercise 7.20 (page 225)
8-10. Prediction with multiple regression
Follow the section 5-10 and the illustration in pages
267-268 by using data set in the Table 8.1 (page
241)
8 11 The troika of hypothesis tests: The likelihood
8-11.
ratio (LR), Wald (W) and Lagarange Multiplier (LM)
Tests
8‐12. Summary and Conclusions
12