Professional Documents
Culture Documents
4 Prediction
Once our regression model has been estimated, we can use the estimated model to predict or
forecast some future value of the dependent variable, Y , that is associated with a given level of
the regression, X.
There are two types of prediction:
1. Prediction of the conditional mean value of Y that is associated with a given level of X, that
lies on the population regression line.
The first and second types are known as mean prediction and individual prediction respectively.
In what follows, we will assume that X0 is a given level of X and Y0 is the corresponding true
level of Y .
Example 6.11. For Yi = β0 + β1 Xi + Ui , let Ŷi = 2.5 + 0.1Xi . Given that X0 = 20, what is
E(Y | X0 )?
Solution:
Using the equation of the fitted line,
Note that Ŷ0 is the point predictor of E(Y | Xi ) and is BLUE. Given that Ŷ0 is an estimator,
we can conduct hypothesis tests for this estimator. Recall that to construct a test statistic for
hypothesis testing we need to find variance of the relevant estimator. In this case we need to find
var(Ŷ0 ).
Before we compute var(Ŷ0 ), let us show that Ŷ0 is an unbiased estimator of E(Y0 | X0 ).
For unbiasedness:
Let
Y0 = β0 + β1 X0 + Ui (6.4)
35
taking conditional expectation of (6.4) at X0 , yields
E(Y0 | X0 ) = β0 + β1 X0
36
Ŷ0 = β̂0 + β̂1 X0
=⇒ var(Ŷ0 ) = var(β̂0 ) + var(β̂1 X0 ) + 2Cov(β̂0 , β̂1 X0 )
= var(β̂0 ) + var(β̂1 )X0 2 + 2X0 Cov(β̂0 , β̂1 )
P 2 · ¸
σ2 Xi 2 σ2 −σ 2 X̄
= P + X0 P + 2X0 P
n (Xi − X̄)2 (Xi − X̄) (Xi − X̄)2
·P 2 ¸
σ2 Xi
= P + X0 2 − 2X0 X̄
(Xi − X̄) 2 n
2
·P 2 ¸
σ Xi 2 2 2
= P − X̄ + X0 − 2X0 X̄ + X̄
(Xi − X̄)2 n
·P 2 ¸
σ2 Xi 2 2
= P − X̄ + (X0 − X̄)
(Xi − X̄)2 n
·P 2 ¸
σ2 Xi − nX̄ 2 2
= P + (X0 − X̄)
(Xi − X̄)2 n
·P ¸
σ2 (Xi − X̄)2 2
= P + (X0 − X̄)
(Xi − X̄)2 n
· ¸
2 1 (X0 − X̄)2
=⇒ var(Ŷ0 ) = σ +P as required.
n (Xi − X̄)2
H0 : Y0 = β0 + β1 X0
q
Ŷ0 − (β0 + βX0 )
we use tŶ0 = , whereSE(Ŷ0 ) = var(Ŷ0 ).
SE(Ŷ0 )
37
6.4.2 Individual Prediction
Y0 = β0 + β1 X0 + U0 .
T hen Y0 − Ŷ0 = (β0 − β̂0 ) + (β1 − β̂1 )X0 + U0 (6.6)
E(Y0 − Ŷ0 ) = E(β0 − β̂0 ) + E(β1 − β̂1 )X0 + E(U0 )
= 0
Note that: mean prediction has smaller variance than individual prediction.
H0 : Y0 − Ŷ0 = 0 or H0 : Y0 = Ŷ0
q
Y0 − Ŷ0
t(Y0 −Ŷ0 ) = where SE(Y0 − Ŷ0 ) = var(Y0 − Ŷ0 ).
SE(Y0 − Ŷ0 )
(iii) var(Y0 − Ŷ0 ) and var(Ŷ0 ) are minimized when X0 = X̄0 . This is because the point (X̄, Ȳ )
lies on the sample regression line; thus, at (X̄, Ȳ ), the corresponding residual is zero. What
is the implication of this for points far away from the sample mean X̄?
38
6.5 Interpreting Regression Results under different functional forms
Consider the functional forms:
M odel 1 : Yi = β0 + β1 Xi + ui (6.7)
M odel 2 : Yi = β0 + β1 ln Xi + ui (6.8)
M odel 3 : ln Yi = β0 + β1 Xi + ui (6.9)
M odel 4 : ln Yi = β0 + β1 ln Xi + ui (6.10)
39