Professional Documents
Culture Documents
• Univariate Data
– Descriptive Statistics
– Type of questions Addressed
• Alternate Notations !
• Theory says : Y = f(x)
– Mathematical form :
y = mx + b
– Statistics Literature
Yi a bX i ei
– ei is the error (disturbance term)
Why do we include a Disturbance term?
• Predicting complete system behavior in research is almost impossible.
(Unlike the models in Maths/Physics..laboratory experiments)
• a = the intercept
– the point where the line crosses the Y-axis.
– (the value of the dependent variable when all of the
independent variables = 0)
• b = the slope
– the increase in the dependent variable per unit change
in the independent variable (also known as the 'rise
over the run')
Simple Regression
x
Ordinary Least Squares
• The most common method used to fit a line to the data is known as
OLS (ordinary least squares).
• What we actually do is take each distance and square it (i.e. take the
area of each of the squares in the diagram) and minimise the total sum
of the squares (hence least squares).
yi
û i
ŷi
xi x
How OLS Works
• But what was ût ? It was the difference between the actual point and
the line, yt - ŷt .
• So minimising y ˆ
y
t t is equivalent to minimising
2
t
ˆ
u 2
• Linear in the parameters means that the parameters are not multiplied
together, divided, squared or cubed, etc.
Yt e X t e ut ln Yt ln X t ut
• Then let yt= ln Yt and xt=ln Xt
yt xt ut
The Assumptions Underlying the
Classical Linear Regression Model (CLRM)
• The model which we have used is known as the classical linear regression model.
• We observe data for xt, but since yt also depends on ut, we must be specific about
how the ut are generated.
• We usually make the following set of assumptions about the ut’s (the
unobservable error terms):
• Technical Notation Interpretation
1. E(ut) = 0 The errors have zero mean
2. Var (ut) = 2 The variance of the errors is constant and finite
over all values of xt
3. Cov (ui,uj)=0 The errors are statistically independent of
one another
4. Cov (ut,xt)=0 No relationship between the error and
corresponding x variate
Expressing Multiple Linear Regression Model
• We used the t-test to test single hypotheses, i.e. hypotheses involving only
one coefficient. But what if we want to test more than one coefficient
simultaneously?
RRSS URSS T k
test statistic
URSS m
where URSS = RSS from unrestricted regression
RRSS = RSS from restricted regression
m = number of restrictions
T = number of observations
k = number of regressors in unrestricted regression
including a constant in the unrestricted regression (or the total number
of parameters to be estimated).
Goodness of Fit Statistics
• We would like some measure of how well our regression model actually fits
the data.
• We have goodness of fit statistics to test this: i.e. how well the sample
regression function fits the data.
• The most common goodness of fit statistic is known as R2. One way to define
R2 is to say that it is the square of the correlation coefficient between y and y$ .
• For another explanation, recall that what we are interested in doing is
explaining the variability of y about its mean value, y , i.e. the total sum of
squares, TSS:
TSS yt y
2
• We can split the TSS into two parts, the part which we have explained (known
as the explained sum of squares, ESS) and the part which we did not explain
using the model (the RSS).
Defining R2
ty y 2
tˆ
y y 2
t
ˆ
u 2
t t t
• Our goodness of fit statistic is
ESS
R2
TSS
• R2 must always lie between zero and one. To understand this, consider two
extremes
RSS = TSS i.e. ESS = 0 so R2 = ESS/TSS = 0
ESS = TSS i.e. RSS = 0 so R2 = ESS/TSS = 1
The Limit Cases: R2 = 0 and R2 = 1
yt
yt
xt xt
Problems with R2 as a Goodness of Fit Measure
• There are a number of them:
T 1
R 1
2
(1 R 2 )
T k
• So, if we add an extra regressor, k increases and unless R2 increases by
a more than offsetting amount, R 2 will actually fall.
Regression Examples
Production Function Analysis
Assumptions in Regressions