Professional Documents
Culture Documents
INTRODUCTION
The classical linear regression model is a statistical model that describes a
data generation process.
SPECIFICATION
The specification of the classical linear regression model is defined by the
following set of assumptions.
Assumptions
1. The functional form is linear in parameters.
Yt = 1Xt1 + 2Xt2 + + kXtk + t
2. The error term has mean zero.
E(t) = 0 for t = 1, 2, , T
3. The error term has constant variance.
Var(t) = E(t2) = 2 for t = 1, 2, , T
4. The errors are uncorrelated.
Cov(t,s) = E(t s) = 0 for all t s
5. The error term has a normal distribution.
t ~ N for t = 1, 2, , T
6. The error term is uncorrelated with each explanatory variable.
Cov(t,Xti) = E(t Xti) = 0 for t = 1, 2, , T and i = 1, 2, , K
7. The explanatory variables are nonrandom variables.
Classical Linear Regression Model Concisely Stated
The sample of T multivariate observations (Yt, Xt1, Xt2, , Xtk) are generated
by a process described as follows.
Yt = 1Xt1 + 2Xt2 + + kXtk + t
t ~ N(0, 2)
for t = 1, 2, , T
or alternatively,
Yt ~ N( 1Xt1 + 2Xt2 + + kXtk , 2) for t = 1, 2, , T
Classical Linear Regression Model in Matrix Format
The sample of T multivariate observations (Yt, Xt1, Xt2, , Xtk) are generated
by a process described by the following system of T equations.
Observation 1
Y1 = 1X11 + 2X12 + + kX1k + 1
Observation 2
Y2 = 1X21 + 2X22 + + kX2k + 2
Observation T
YT = 1XT1 + 2XT2 + + kXTk + T
Note the following. 1) There is one equation for each multivariate
observation. 2) The parameters are constants, and therefore have the same
value for each multivariate observation. 3) The system of T equations can
be written equivalently in matrix format as follows.
y = X +
y is a Tx1 column vector of observations on the dependent variable. X is a
TxK matrix of observations on the K-1 explanatory variables X2, X3, Xk. The
first column of the matrix X is a column of 1s representing the constant
(intercept) term. The matrix X is called the data matrix or the design matrix.
is a Kx1 column vector of parameters 1, 2 k. is a Tx1 column vector
of disturbances (errors).
Assumptions in Matrix Format
1. The functional form is linear in parameters.
y = X +
2. The mean vector of disturbances is a Tx1 null vector.
E() = 0
3. The disturbances are spherical. (The variance-covariance matrix of
disturbances is a TxT diagonal
matrix).
Cov() = E( T) = 2I
Where superscript T denotes transpose and I is a TxT identity matrix.
4. The disturbance vector has a multivariate normal distribution.
~N
5. The disturbance vector is uncorrelated with the data matrix.
Cov (,X) = 0
6. The data matrix is a nonstochastic matrix.
Classical Linear Regression Model Concisely Stated in Matrix Format
The sample of T multivariate observations (Yt, Xt1, Xt2, , Xtk) are generated
by a process described as follows.
or alternatively
y = X + ,
~ N(0, 2I)
y ~ N(X, 2I)
ESTIMATION
For the classical linear regression model, there are K+1 parameters to
estimate: K regression coefficients 1, 2 k, and the error variance
(conditional variance of Y) 2.
Choosing an Estimator for 1, 2 k
To obtain estimates of the parameters, you need to choose an estimator. To
choose an estimator, you choose an estimation procedure. You then apply
the estimation procedure to your statistical model. This yields an estimator.
In econometrics, the estimation procedures used most often are:
1. Least squares estimation procedure
2. Maximum likelihood estimation procedure
Least Squares Estimation Procedure
When you apply the least squares estimation procedure to the classical
linear regression model you get the ordinarily least squares (OLS) estimator.
The least squares estimation procedure tells you to choose as your estimates
of the unknown parameters those values that minimize the residual sum of
squares function for the sample of data. For the classical linear regression
model, the residual sum of squares function is
RSS(1^, 2^ k^) = (Yt - 1^ - 2^ X12 - - k^ X1k)2
Or in matrix format,
RSS( ^) = (y - X ^)T(y - X ^)
The first-order necessary conditions for a minimum are
XTX ^ = XTy
These are called the normal equations. If the inverse of the KxK matrix XTX
exists, then you can find the
solution vector ^. The solution vector is given by
^ = (XTX)-1XTy
where ^ is a Kx1 column vector of estimates for the K-parameters of the
model. This formula is the OLS estimator. It is a rule that tells you how to
use the sample of data to obtain estimates of the population parameters.
2^
RSS
=
=
Tk
T
Tk
That is, the mean vector of the OLS estimator is equal to the true values of
the population parameters being estimated. This tells us that for the
classical linear regression model the OLS estimator is an unbiased estimator.
4. Variance-Covariance Matrix of Estimates
The variance-covariance matrix of estimates gives the variances and
covariances of the sampling distributions of the estimators of the K
parameters. To derive the variance-covariance matrix of estimates, you
need to make four assumptions:
1. The error term has mean zero.
2. The error term is uncorrelated with each explanatory variable
3. The error term has constant variance.
4. The errors are uncorrelated.
If these four assumptions are satisfied, then it can be shown that the
variance-covariance matrix of estimates is
Cov( ^) = 2(XTX)-1
For the classical linear regression model, it can be shown that the elements
in the variance-covariance matrix of OLS estimates is less than or equal to
the corresponding elements in the variance-covariance matrix for any
alternative linear unbiased estimator; therefore, for the classical linear
regression model the OLS estimator is an efficient estimator.
5. Sampling Distribution of the OLS Estimator Written Concisely
^ ~ N(, 2(XTX)-1)
The OLS estimator has a multivariate normal distribution with mean vector
and variance-covariance matrix 2(XTX)-1.
Summary of Small Sample Properties
Gauss-Markov Theorem - For the classical linear regression model, the OLS
estimator is the best linear unbiased estimator (BLUE) of the population
parameters.
Summary of Large Sample Properties
For the classical linear regression model, the OLS estimator is asymptotically
unbiased, consistent, and asymptotically efficient.
Estimating the Variance-Covariance Matrix of Estimates
t-test
F-test
Likelihood ratio test
Wald test
Lagrange multiplier test
You must choose the appropriate test to test the hypothesis in which you are
interested.
GOODNESS-OF-FIT
If our objective is to use the explanatory variable(s) to predict the dependent
variable, then we should measure the goodness of fit of the model. Goodness-of-fit
refers to how well the model fits the sample data. The better the model fits the
data, the higher the predictive validity of the model, and therefore the better values
of X should predict values of Y. The statistical measure that is used most often to
measure the accuracy of a classical linear regression model is he R-squared (R2)
statistic.
R-Squared Statistic
The coefficient of determination measures the proportion of the variation in the
dependent variable that is explained by the variation in the explanatory variables.
It can take any value between 0 and 1. If the R 2 statistic is equal to zero, then the
explanatory variables explain none of the variation in the dependent variable. If the
R2 is equal to one, then the explanatory variables explain all of the variation in the
dependent variable. The R2 statistic is a measure of goodness of fit. This is because
it measures how well the sample regression line fits the data. If the R 2 is equal to
one, then all of the data points lie on the sample regression line. If the R 2 is equal
to zero, then the data points are highly scattered around the regression line, which
is a horizontal line. The higher the R 2 statistic, the better the explanatory variables
explain the dependent variable, and using this criterion the better the model.