Professional Documents
Culture Documents
1 3 k
X 2i X 1i X 3i ... X ki (3)
2 2 2
1 3 k 1
X 2i X 1i X 3i ... X ki vi (4)
2 2 2 2
The numerical example:
X2 X3 X3*
10 50 52
15 75 75
18 90 97
24 120 129
30 150 152
Multi-collinearity refers only to linear
relationship among the X variables. It does
not rule out nonlinear relationships among
the X variables.
For example:
Yi = 0 + 1Xi + 2 Xi2 + 3Xi3 + ui (5)
Where:
Y = total cost of production
X = output
Model (5) does not violate the assumption of no
multicollinearity.
If multicollinearity is perfect, the regression
coefficients of X variables are indetermine
and the standard error are infinite
If multicollinearity is less than perfect, the
regression coefficients, although determine,
possess large standard errors (in the relation
to the coefficients themselves), which
means the coefficients cannot be estimated
with greater precision or accuracy
Assume: X3i = X2i , where is a nonzero
constant. Substituting (6) into (7)
yi 2 x2i 3 x3i ui (6)
y x x y x x
2
x
i 2i 3i i 3i 2 i 3i
(7)
2
x x x x
2
2i
2
3i 2 i 3i
2
y x x y x x
2 2 2
2 i 2i 2i i 2i 2i
(8)
x x x
2
2i
2 2
2i
2 2 2
2i
0
0
2 is indeterminate, 3 is also indeterminate.
Recall the meaning of 2 , the rate of change in the
average value of Y as X2 change by a unit, holding
X3 constant. If X3 and X2 are perfectly collinear,
there is no way X3 can be kept constant. As X2
changes, so does X3 by the factor .
There is no way of disentangling the separate
influences of X2 and X3 from the given sample. For
practical purpose X2 and X3 are indistinguishable.
This problem is most damaging since the entire
intent is to separate the partial effects of each X
upon the dependent variable.
To see differently:
Substituting X3i = X2i into (6)
y i 2 x 2i 3 (x 2i ) u i
( ) x u (10)
2 3 2i i
x 2i u i
where
( 2 3 ) (11)
Applying the usual OLS, we get
2 3
x y2i 2i
(12)
x 2
2i
2 i 2i 2i i i 2i i i 2i
(16)
x x v x
2
2i
2 2
2i
2
i
2 2
2i
2
where x v 0
2i i
No reason to believe a priori that 2 and 3
cannot be estimated
If vi is sufficiently small, very closed to
zero, (15) will indicate almost perfect
collinearity and we shall be back to the
indeterminate case of (9)
Theoretical Consequences of
Multicollinearity
Multicollinerity violates no regression
assumptions. If multicollinearity is very
high, the OLS estimators still retain the
property of BLUE
Unbiased, consistent estimates will occur,
but their standard errors will be correctly
estimated
It hard to get coefficient estimates with
small standard error
The collinearity does not destroy the
property of minimum variance; that is , they
are efficient.
Having small number of observations also
has that effect. What should I do if I dont
have many observation? No statistical
answer can be given. Multicollinearity is
essentially a sample phenomenon
The consumption-income example
2
var( 3 ) (18)
x 2
3i (1 r )
2
23
r23 2
cov( 2 , 3 ) (19)
(1 r232 ) 2 i 3i
x 2
x 2
Wider Confidence Interval
Because of the large SE, the confidence
intervals for the relevant population
parameters tend to be larger
In the case of high multicollinearity, the
sample data may be compatible with a
diverse set of hypotheses. Hence, the
probability of accepting a false hypothesis
increases
r23 95% confidence int erval for 2
2
0,00 2 1,96
2ix 2
2
0,50 2 1,96 1,33
2i
x 2
2
0,95 2 1,96 10,26
2i
x 2
2
0,995 2 1,96 100
2i
x 2
2
0,999 2 1,96 500
2i
x 2
The signals of multicollinearity
2
var( 2 ) VIF (21)
x 2i2
2
var( 3 ) VIF (22)
x 2
3i
2
var( j ) VIF j (23)
x 2
j
1
TOL j (24)
VIF j
r23 VIF TOL
0,00 1,00 1,00
0,50 1,33 0,75
0,70 1,96 0,51
0,80 2,78 0,36
0,90 5,26 0,19
0,95 10,26 0,10
0,995 100,25 0,01
0,999 500,25 0,00
If the VIF of a variable exceeds 10,
the variable is said to highly
collinear
If the TOL of a variable less than
0.10, the variable is said to highly
collinear
Eigenvalues and conditional index
The condition number k:
Max eigenvalue
k
Min eigenvalue
Yt 1 X 2t ut
1 2 3 (29)
X 3t X 3t X 3t X 3t
The first-difference or ratio transformations are not
without problems. The error term vt in (27) may not
satisfy one of the assumptions of the classical
linear regression model (the disturbances are
serially uncorrelated). If the original disturbance
term ut is serially uncorrelated, the error term v
obtained previously will in most cases be serially
correlated.
There is a loss of one observation due to the
differencing procedure, and therefore the degree of
freedom are reduced by one. In a small sample, this
could be a factor one would wish at least to take
into consideration
The first-diffrencing procedure may not be
appropriate in cross-sectional data where there
is no logical ordering of the observations
In the ratio model: the error term (ut/X3t) will
be heteroscedastic, if the original error term ut
is homoscedastic.
Again the remedy may be worse than the
disease of collinearity
Additional or new data
Sometimes simply increasing the size of the
sample may attenuate the collinearity
problem
The variance of i will decrease, thus
decreasing the standard error, which will
enable us to estimate i more precisely.
Reducing multicollinearity in polynomial
regrresions
In the polynomial regression models when
the explanatory variable(s) are expressed in
the deviation form (deviation in the mean
value), multicollinearity is substantially
reduced.
But even then the problem may persist, in
which case one may want to consider
technique such as orthogonal polynomials
Other methods of remedying multicol-
linearity
Multivariate statistical techniques such as
factor analysis; principal components or
ridge regression are often employed to
solve the problem of multicollinearity