Professional Documents
Culture Documents
Summation Operator
x
i
= x
1
+ x
2
+ + x
n
For any constant c, c = nc
For any constants a and b,
(ax
i
+by
i
) a x
i
+b y
i
i1
n
i1
n
i1
n
Statistics
If x~N(,
2
), then
Expectation:
If c is a constant, then E(c) = c
If {a
1
, a
2
, , a
n
} are constants and {x
1
, x
2
, , x
n
}
are random variables, then
z
x
~ N(0,1)
E(a
1
x
1
+a
2
x
2
+... +a
n
x
n
) a
1
E(x
1
)+a
2
E(x
2
)+... +a
n
E(x
n
)
Statistics (Continued)
Variance and Covariance:
If c is a constant, then Var(c) = 0
Cov(x, y) E[(x
x
)(y
y
)]
E{[x E(x)][y E(y)]}
E(xy) E(x)E(y)
E[(x
x
)y]
E[x(y
y
)]
Statistics (Continued)
If x and y are independent, then Cov(x,y) = 0
because E(xy) = E(x)E(y) when they are
independent
However, zero covariance between x and y
does not necessarily imply that x and y are
independent
Statistics (Continued)
Try the following. If you can prove this, then
you are all set for properties of expectation,
variance, and covariance!
Var(ax +by) a
2
Var(x)+b
2
Var(y)+2abCov(x, y)
Statistics (Continued)
Correlation coefficient:
If x and y are independent, then
x,y
= 0, but zero
correlation itself does not necessarily imply
independence
[ 1,1]
Ordinary Least Squares
Least squares refers to the minimizing the sum
of squared residuals
Trivia: why not
Cannot be minimized because solution can be
either positive or negative
min u
i
2
i1
n
min u
i
i1
n
?
Ordinary Least Squares (Continued)
Estimators:
First order conditions or method of moments
Hint: the point is always on the OLS
regression line (SRF line)
(x, y)
y
0
+
1
x
0
y
1
x
Statistical Properties of OLS
An estimator is unbiased if its expected value
(or mean of its sampling distribution) equals
the population value
Statistical Properties of OLS
(Continued)
Gauss-Markov assumptions:
iid independently (i.e., each random variable is
mutually independent of every other random variable)
and identically distributed (i.e., each random variable
has the same probability distribution as every other
random variable)
u
i
~ iid(0,
u
2
)
E(u | x) E(u)
0
Statistical Properties of OLS
(Continued)
E(u|x) = E(u) = 0 is also referred to as the zero
conditional mean assumption; the expectation
of u is zero given any values of x
Cov(u
i
, u
j
) = 0 for all i != j; also referred to as
no serial correlation (if you choose to go on to
time series, especially financial econometrics)
for all i = 1, 2, , n
Also referred to as the constant
variance/homoskedasticity assumption
Var(u
i
)
u
2
Unbiasedness of
1
1
(x
i
x)y
i
i1
n
(x
i
x)
2
i1
n
E(
1
) E
(x
i
x)y
i
i1
n
(x
i
x)
2
i1
n
E
(x
i
x)(
0
+
i1
n
1
x
i
+u
i
)
(x
i
x)
2
i1
n
Unbiasedness of
1
(Continued)
E
0
(x
i
x) +
i1
n
1
x
i
(x
i
x) +
i1
n
u
i
(x
i
x)
i1
n
(x
i
x)
2
i1
n
E
1
+
u
i
(x
i
x)
i1
n
(x
i
x)
2
i1
n
E(
1
) + E
u
i
(x
i
x)
i1
n
(x
i
x)
2
i1
n
Unbiasedness of
1
(Continued)
1
+
E[u
i
(x
i
x)
i1
n
]
(x
i
x)
2
i1
n
1
The numerator is essentially
Cov(x,u) (E(u) = 0), which is
assumed to be zero per the
zero conditional mean
assumption (E(u|x) = E(u) =
0)
Unbiasedness of
0
0
y
1
x
0
+
1
x +u
1
x
0
+(
1
1
)x +u
E(
0
) E[
0
+(
1
1
)x +u]
E(
0
) + E[(
1
1
)x]+ E(u)
0
+ E[(
1
1
)x]
0
If
1
is
unbiased, then
the second
term is equal to
zero
Gauss-Markov Theorem
Under the Gauss-Markov assumptions, the OLS
estimator(s) is/are the BLUE(s)
Best
Linear*
Unbiased*
Estimator
The theorem only applies to comparisons
between unbiased estimators
Measures of Goodness of Fit
Total sum of squares (SST/TSS):
Total sample variation (spread) in the
dependent variable about its sample average
Explained sum of squares (SSE/ESS):
Total sample variation of the fitted values in a
(multiple) regression model
(y
i
y)
2
i1
n
y
i
)
(
y
i
y)
2
i1
n
u
i
2
i1
n
SSE
SST
1
SSR
SST
Measures of Goodness of Fit
(Continued)
100*R
2
is the percentage of the sample variation
in the dependent variable that is explained by the
independent variable(s)
Note that R-squared only measures the linear
relationship between dependent and
independent variables
R
2
can be easily inflated by increasing the size of
the sample (and the number of independent
variables [other issues will also arise, as you will
see later in the course (?)])
Units of Measurement
Change in the unit of measurement of the
independent variable:
Now, lets change x for , where the latter is x/10,
then
Previously, for every increase in x, the fitted value
increases by 1.5
Now, for every increase in , the fitted value increases
by 15, but each independent variable has been
changed to be 1/10 of their original value, so the
relationship remains the same
y
i
2+1.5x
i
x
i
y
i
2+15x
i
x
i
Units of Measurement (Continued)
This can be proved very simply (your professor
presented a more formal proof, but this is for
intuition):
The converse is also true (i.e., change x to ,
valued at 10x divide by 10)
y
i
2+15x
i
2 +10(1.5)(
x
i
10
)
2 +1.5x
i
x
i
1
Units of Measurement (Continued)
Change in the unit of measurement of the
dependent variable:
Now, lets change to , where the latter is 1/10
the value, then
Do we still have the same relationship?
Of course!
The converse applies also in this case
y
i
2+1.5x
i
y
i
y
i
y
i
0.2+0.15x
i
Units of Measurement (Continued)
Change in unit of measurement in both
variables:
Now, we will combine the two changes ( and )
y
i
2+1.5x
i
x
i
y
i
y
i
10
2
10
+10
1
10
(1.5)
x
i
10
y
i
0.2+10
1
10
(1.5)x
i
y
i
0.2+1.5x
i
Statistical Inference
Terms:
Null hypothesis one takes this hypothesis as
true and requires the data to provide substantial
evidence that suggests otherwise
H
0
Alternative hypothesis hypothesis against
which the null hypothesis is tested
H
1
/H
a
Statistical Inference (Continued)
Never accept the null hypothesis only do
not reject
Type I error rejection of H
0
when it is true
Type II error failure to reject H
0
when it is
false
Significance level probability of type I error in
hypothesis testing
alpha
Statistical Inference (Continued)
One-sided alternative <-> one-sided test
H
0
: <
0,
H
1
: >
0
H
0
: >
0,
H
1
: <
0
Two-sided alternative <-> two-sided test
H
0
: =
0,
H
1
: !=
0
Statistical Inference (Continued)
t-statistic for
0
and
1
:
Why n 2?
Degree of freedom (df)
# of observations # of estimated parameters (i.e.,
estimators)
Decreases as you add more independent variables (later in the
course) (i.e., intercept + 2 independent variables n 3)
t
0
0
se(
0
)
~ t
n 2
t
1
1
se(
1
)
~ t
n 2
Statistical Inference (Continued)
Side note:
Var(
0
)
u
2
n
1
x
i
2
i1
n
(x
i
x)
2
i1
n
1
CI [
1
c
2
se(
1
),
1
+c
2
se(
1
)]
[ 0.3671 1.984(0.1941), 0.3671+1.984(0.1941)]
[ 0.3671 0.3850944, 0.3671+0.3850944]
[ 0.7521944, 0.0179944]
STATA!
Disclaimer: pure R user, never used STATA
However, I am able to read the outputs and have
constructed questions around them
STATA Output
STATA Output
SSE/ESS SSR/RSS SST/TSS any two can be used to solve for R-squared
STATA Output
0
1
se(
i
), i 0,1
STATA Output
t
i
se(
i
)
, i 0,1
Significance test (H
0
:
i
= 0)
STATA Output
p-value of significance test