Professional Documents
Culture Documents
&
Multiple Regression
(Partial and overall significance test)
^
2
0 Xi
M. Balclar, ECON 503, EMU
5.3
Regression through the origin
The estimated model:
~ ~
Y = 2 X i
~ ~
or Y = 2 X i + u i
Applied OLS method:
^2
~ X iYi ~
and Var 2 = ( )
2 = X2i
Xi 2
~
and ~
2 =
u 2
N -1
M. Balclar, ECON 503, EMU
5.4
Some feature of interceptless model
1. u
~
i need not be zero
can occasionly turn out to be negative.
2. R2 may not be appropriate as a summary
statistics (goodness of fit).
Y = 1 + 2 X+ u i Y = 2 X + ui
^ xy XY
2 = ~
=
x2 2 X2
^
^ 2
Var (2 ) =
~ ^
2
Var 2 = ( )
x2 X2
^ 2 =
^2 ~2
u
~
2
=
u
n2 n 1
=
[ (X X)(Y Y)]
2
( XY ) 2
2
R
(X X) (Y Y)
2 2
raw R 2 =
( xy ) 2 X2 Y2
or R =
2
x y 2 2
M. Balclar, ECON 503, EMU
5.6
Y
False SFR:
^ ^ X
Y = 2
^
1
X
10 50
M. Balclar, ECON 503, EMU
5.7
Y
False SFR:
^
Y = ^2X
5
^
1
X
10 50
M. Balclar, ECON 503, EMU
5.8
Y
True and best SFR:
^
Y = ^ X
2
False SFR:
^ ^
Y =1+ ^2X False SFR:
^ ^ ^X
Y = 1+ 2
^
1
X
10 50
M. Balclar, ECON 503, EMU
5.9
Example 1: Capital Asset Pricing Model (CAPM) security is
expect risk premium=expected market risk premium
( ER i
rf ) = 2 ( ER m rf )
Example 1:(cont.)
2
1
ERm f
e
International interest rate differentials equal exchange rate
forward premium.
e
i.e., (i i ) = 2 (
* FN
)
e
i i*
Covered interest
parity line
2
1
FN e
e
M. Balclar, ECON 503, EMU
5.12
Example 2:(Cont.)
in regression: E (1 ) = 0
F e
(i i ) = 1 + 2(
*
) + ui
e
If covered interest parity holds, 1 is expected to be zero.
^=
Formal report: Y 1.0899 X R2=0.714 N=10
(5.689) SEE=19.54
M. Balclar, ECON 503, EMU
5.14
H0: 1 = 0
1.279 - 0
7.668
^=
Y 1.2797 + 1.0691X R2=0.715 N=10
(0.166) (4.486) SEE=20.69
Multiple Regression
Y = 1 + 2 X2 + 3 X3 ++ k Xk + u
u = Y - 1 - 2 X2 - 3 X3
OLS is to minimize the SSR( ^u2)
min. S = min. u2 = min. (Y - 1 - 2X2 - 3X3)2
RSS ^ ^ ^
^ =2 ( Y - - X - 3X3)(-1) = 0
1 1 2 2
RSS ^ ^ ^
^ =2 ( Y - - X - 3X3)(-X2) = 0
2 1 2 2
RSS ^ ^ ^
^ =2 ( Y - - 2 2 3X3)(-X3) = 0
X -
3 1
^ X + ^ X X + ^ X 2 = X Y
1 3 2 2 3 3 3 3
^
(XX) = XY Matrix notation
n X2 Y
X2 X22 X2Y
X3 X2X3 X3Y (yx3)(x22) - (yx2)(x2x3)
^ =
3 =
n X2 X3 (x22)(x32) - (x2x3)2
X2 X22 X2X3
X3 X2X3 X32
_ ^_ ^_
^
1 = Y - 2 X2 - 3 X3
M. Balclar, ECON 503, EMU
Alternative Notation
k
Summation Form Yi = j X ji + ui , X 1i = 1, i
j =1
Matrix Form Y = X + u
where
Y1 1 X 21 " X k1 1 u1
Y = # , X = # # % # , = # , u = #
Yn 1 X 2 n " N kn k un
M. Balclar, ECON 503, EMU 5.19
5.20
or in matrix form:
(XX) ^ = XY
3x3 3x1 3x1
^ = (XX)-1
==> (XY)
3x1 3x3 3x1
^
u 2
^ ^u2 (XX)-1
Var-cov() = and ^
u2 =
n-3
Variance-Covariance matrix
^ ^ ^ ^ ^
Var(1) Cov(1, 2) Cov(1, 3)
^ ^ ^ ^) ^ ,^
Var-cov() = Cov (2 ,1) Var( 2 Cov( 2 3)
^ ,
Cov ( ^ ) Cov(
^ , ^) Var(^)
3 1 3 2 3
= ^u2(XX)-1
M. Balclar, ECON 503, EMU
5.21
-1
n X2 X3
=^
u2 X2 X22 X2X3
X3 X3X2 X32
k=3
# of independent variables
( including the constant term)
X3 = 2X 2
r23 = 1 ???
r23 = cor ( X 2 , X 3 ) =
cov( X 2 , X 3 )
=
x x 2 3
sd ( X 2 )sd ( X 3 ) x x 2
2 EMU
M. Balclar, ECON 503,
2
3
5.23
The meaning of partial regression coefficients
Y = 1 + 2X2 + 3X3 + u (suppose this is a true model)
5. ^^
uY=0 random sample
M. Balclar, ECON 503, EMU
Properties of multiple OLS estimators 5.25
^
6. As X2 and X3 are closely related ==> var(2) and var(^3)
become large and infinite. Therefore the true values of 2
and 3 are difficult to know.
All the assumptions in the two-variable regression case
are also applied to the multiple variable regression.
But one addition assumption is
No exact linear relationship among the independent variables.
k-1
(No perfect collinearity, i.e., Xk iXj 2X2 + + k-1Xk-1)
i=2
Y Y
Note that this is
3 5
1 2 not perfect
4
X2 X3 multicollinearity
X2 X3
Y = 1 + 2 X2 + 3 X3 + u
Y TSS
^Y n-1
^u
C
X2
X3
X4
H0 : 2 = 0 Y
= 2 = 0?
X2
H1 : 2 0
^
2 - 0 0.726
t= = = 14.906
Se (^ )2 0.048
^
Answer : Yes, 2 is statistically significant and is
significantly different from zero.
M. Balclar, ECON 503, EMU
5.38
1. Individual partial coefficient test (cont.)
2 holding X2 constant: Whether X3 has the effect on Y?
Y
H0 : 3 = 0 = 3 = 0?
X3
H1 : 3 0
^
3 - 0 2.736-0
t= = = 3.226
Se (^ ) 0.848
3
^
Answer: Yes, 3 is statistically significant and is
significantly different from zero.
M. Balclar, ECON 503, EMU
5.39
2. Testing overall significance of the multiple regression
3-variable case: Y = 1 + 2 X2 + 3 X3 + u
H0 : 2 = 0, 3 = 0, (all variable are zero effect)
H1 : 2 0 or 3 0 (At least one variable is not zero,
At least one variable has the effect)
(1 )
0 Fc M. Balclar, ECON 503, EMU
F
Analysis of Variance: Since y=^y + ^u 5.41
^2
^2 + u
==> y2 = y
TSS = ESS + RSS
ANOVA TABLE
(SS) (MSS)
Source of variation Sum of Square df Mean sum of Sq.
Due to regression(ESS) y^2 k-1 y^2
k-1
Due to residuals(RSS) u^2 n-k u^2 ^2
n-k = u
H0 : 2 = = k = 0
if F > Fc,k-1,n-k ==> reject Ho
H1 : 2 k 0
M. Balclar, ECON 503, EMU
5.42
^ ^
Three- y = 2x2 + 3x3 + u^
variable ^ ^
y2 = 2 x2 y + 3 x3 y + u^2
case
TSS = ESS + RSS
ANOVA TABLE
Source of variation SS df(k=3) MSS
ESS ^ x y + ^ x y
3-1 ESS/3-1
2 2 3 3
RSS ^2
u n-3 RSS/n-3
(n-k)
TSS y2 n-1
R2 / (k-1) (k-1) F
F = Reverse : R2 =
(1-R2) / n-k (k-1)F + (n-k)
M. Balclar, ECON 503, EMU
Overall significance test: 5.44
H 0 : 2 = 3 = 4 = 0
H1 : at least one coefficient
is not zero.
2 0 , or 3 0 , or 4 0
R 2 / k-1
F* = 2
=
(1-R ) / n- k
0.9710 / 3
=
(1-0.9710) /16
= 179.13
Fc(0.05, 4-1, 20-4) = 3.24
k-1 n-k
Since F* > Fc ==> reject H0.
M. Balclar, ECON 503, EMU
Example:Gujarati(2003)-Table6.4, pp.185) 5.45
H0 : 2 = 3 = 0
ESS / k-1 R 2 / k-1 0.707665 / 2
*
F = = 2
=
RSS/(n- k) (1-R ) / n- k (1-0.707665)/ 61
F* = 73.832
k-1 n-k
Since F* > Fc
==> reject H0.
M. Balclar, ECON 503, EMU
Construct the ANOVA Table (8.4) .(Information from EVIEWS) 5.46
Source of
variation
SS Df MSS
2 2 2 2
Due to R (y ) k-1 R (y )/(k-1)
regression =(0.707665)(75.97807)2x63
(ESS) = 257362.2121 =2 =128681.1061
2 2 2 2 2
Due to (1- R )(y ) or ( u ) n-k (1- R )(y )/(n-k)
Residuals =(0.292335)(75.97807)2x63
(RSS) = 106315.8165 =61 = 1742.8822
2
Total (y ) n-1
(TSS) =(75.97807)2x63
= 363678.0286 =63
Consider Y, X2, X3
Assume cor(X2,X3) 0
What is the cor(Y,X2) net of effect of X3 on
X2. This is known as the partial correlation
between Y and X2, denated r123.
Conceptually, it is similar to the partial
regression coefficient.
Partial Correlation Coefficient
Partial Correlation Coefficient