Professional Documents
Culture Documents
1. Some terminology
2. If the data does not reject a null hypothesis at, say, the 5% level,
THAT DOES NOT MEAN THE NULL HYPOTHESIS IS TRUE. It
only means that we can’t be more than 95% sure it is false.
A Comment:
1. The univariate normal has only two parameters, the mean and the variance.
So, if a univariate normal distribution’s mean and variance are independent of
X, then that distribution is identical for all X.
Further, recall once again that the sampling error of the OLS coefficient
vector is given by:
b β ( XX) 1 Xε Aε
The above property is the basis for most of the hypothesis testing in
everyday economic research.
3. Tests Concerning Individual Regression Coefficients
bk k
zk follows a standard normal distribution.
( XX) kk
2 1
Thus, suppose we calculated zk and it was 3. That means that, under our
null hypothesis, the chances of drawing the particular ε vector that
yielded our estimate of b were very low. In such a case we would reject
the null hypothesis.
More concretely, if zk < -1.96 or zk > 1.96, we can reject the null
hypothesis with 95% confidence (because Φ(-1.96) =.025 and Φ(1.96) =
.975, where Φ is the standard normal cdf.):
.
4
.3
f .2
.1 .025
.025
0
-3 -1.96 0 1.96 3
z
Of course, there’s only one problem with the above
procedure:
The calculations use σ2, when in fact we don’t know σ2.
SSR ee 2
Solution: Use an unbiased estimate, s2 , in place of σ .
nK nK
bk k
Now, t k 2 follows a t distribution with n-K degrees of
s ( XX) kk
1
freedom. The t distribution looks like the normal, but with fatter tails
(the tails get fatter as n-K shrinks):
t distribution with 5 degrees of freedom:
.4
.3
.2
.1
.025 .025
0
-3 -2.57 0 2.57 3
z
f t5
Summary of t tests for individual regression coefficients:
bk k
1. Estimate the regression and calculate t k for your null
s ( XX) kk
2 1
hypothesis (e.g. βk 0 ).
bk k
t* t * , or t * SE (bk ) bk k tSE (bk ) * , or:
SE (bk )
Rough rule: reject all parameter values that differ from the estimated
coefficient by more than two standard errors.
Yet another approach: p-values.
Then we can:
bk k
1) Compute the t-ratio t k , as before.
s ( XX) kk
2 1
Formally, p is implicitly defined by: Prob (- |tk| < t < |tk|) = 1-p
Note: when Stata and other statistical packages report p-values for
estimated coefficients, these refer to the specific null hypothesis that
βk 0 .
4. Tests Concerning Multiple Regression Coefficients
β r
H0 : R
#r K # r 1
K 1
Examples include:
F
(Rb r) R(XX) R 1
1
(Rb r)/# r
2
s
Estimate the regression and calculate the F statistic for your null
hypothesis, using the above formula.
3. Compare your F statistic to the critical value. If F > F*, you reject the
null at the level of significance you’ve selected.
Applications of F tests:
b2
β1
b1
Suppose you were able to estimate the “restricted” model, i.e. minimize
the sum of squared residuals while imposing the restriction Rβ = r.
(This is often much easier to do than it appears).
( SSR R SSRU ) /# r
F
SSRU /( n K )
Intuitively, this is a measure of how much your SSR (i.e. the regression’s
‘unexplained variance’) rises when you impose the restriction, adjusted
for the number of restrictions imposed.