Professional Documents
Culture Documents
Correlation
r=
n XY X Y
n X X
2
n Y Y
2
H0 : = 0
H1 : 0
= 0.05
t = r n-2
1 - r2
1/2
t = 0.9361
12-2
1 - (0.9361)2
1/2
= 8.414
and the degrees of freedom = 12-2 = 10. From the
t distribution table , the critical value of t for a two-tailed test and
= 0.05 is 2.228.
Partial Correlation
A partial correlation coefficient measures the
association between two variables after controlling for,
or adjusting for, the effects of one or more additional
variables.
rx y . z =
rx y - (rx z ) (ry z )
1 - rx2z 1 - ry2z
Regression Analysis
Regression analysis examines associative relationships
between a metric dependent variable and one or more
independent variables in the following ways:
Regression Analysis
Population
Y intercept
Dependent
(Response)
Variable
Random
Error
Yi x
Xii i
Population
Regression yx
Line
(Expected mean E(Y)
Independent
(Explanatory)
Variable
Yt a bX t et
Yt a bX
t
et Yt Yt
t 1
t 1
t 1
2
2
)2
(
Y
Y
)
(
Y
bX
t t t t
t
(X
t 1
X )(Yt Y )
a Y bX
2
(
X
X
)
t
t 1
OR
b=
n XY X Y
n X X
2
Xt
1
2
3
4
5
6
7
8
9
10
10
9
11
12
11
12
13
13
14
15
120
n 10
Yt
44
40
42
46
48
52
54
58
56
60
500
X t 120
Yt 500
t 1
X
t 1
X t 120
12
n
10
t 1
Yt 500
50
10
t 1 n
Xt X
Yt Y
-2
-3
-1
0
-1
0
1
1
2
3
( X t X )(Yt Y )
( X t X )2
12
30
8
0
2
0
4
8
12
30
106
4
9
1
0
1
0
1
1
4
9
30
-6
-10
-8
-4
-2
2
4
8
6
10
(X
t 1
X )2 30
X )(Yt Y ) 106
(X
t 1
n 10
n
X
t 1
(X
t 1
t 1
120
Y 500
t 1
X ) 30
X )(Yt Y ) 106
(X
t 1
X t 120
12
n
10
Yt 500
50
10
t 1 n
106
b
3.533
30
a 50 (3.533)(12) 7.60
Y = 7.60+3.533X
t= b
SEb
Tests of Significance
Example Calculation
Yt
et Yt Yt
et2 (Yt Yt )2
( X t X )2
44
42.90
1.10
1.2100
40
39.37
0.63
0.3969
11
42
46.43
-4.43
19.6249
12
46
49.96
-3.96
15.6816
11
48
46.43
1.57
2.4649
12
52
49.96
2.04
4.1616
13
54
53.49
0.51
0.2601
13
58
53.49
4.51
20.3401
14
56
57.02
-1.02
1.0404
10
15
60
60.55
-0.55
0.3025
65.4830
30
Time
Xt
Yt
10
e (Yt Yt )2 65.4830
t 1
2
t
t 1
(X
t 1
(Y Y )
( n k ) ( X X )
2
X ) 30
2
sb
65.4830
0.52
(10 2)(30)
Tests of Significance
Example Calculation
n
t 1
t 1
2
2
(
Y
Y
)
t t t 65.4830
2
(
X
X
)
30
t
t 1
(Yt Y )
65.4830
sb
0.52
2
( n k ) ( X t X )
(10 2)(30)
Tests of Significance
Calculation of the t Statistic
b 3.53
t
6.79
sb 0.52
Y = 7.60+3.533 X
Hence we can say that b is a significant
regression coefficient which infers that
X is a significant explanatory variable
for Y.
Test of Significance of R2
Decomposition of Sum of Squares
Total Variation = Explained Variation + Unexplained Variation
2
2
(Yt Y ) (Y Y ) (Yt Yt )
2
Test of Significance
Coefficient of Determination
R2
(
Y
Y
)
Explained Variation
2
TotalVariation
(
Y
Y
)
t
373.84
R
0.85
440.00
2
Coefficient of Correlation
r R2 withthe sign of b
1 r 1
r 0.85 0.92
H1: R2
0.05
Under the validity of
ANOVA Table
Source
Sum of Squares
D.F.
Mean Square
SSR
k 1
Regression
SSR
k-1
MSR
Error
SSE
n-k
MSE
Total
SST
n-1
If
F Fk 1, n k is accepted,
SSE
nk
F
F
MSR
MSE
n-1
Total
Sample
Variability
SSR
Explained
Variability
k-1
SSE
Unexplained
Variability
n-k
S S y = (Y i - Y )2
i =1
S S reg = (Y i - Y )
i =1
SS
= (Y i - Y i )
i =1
Multiple Regression
The general form of the multiple regression model
is as follows:
Population slopes
Random
Error
Independent (Explanatory)
variables for sample model
Residual
R 2 1 (1 R 2 )
(n 1)
(n k )
R 2 /(k 1)
F
(1 R 2 ) /(n k )
H0: 1 = 2 = 3 = . . . = k = 0
The overall test can be conducted by using an F statistic:
F=
R 2 /K-1
(1 - R 2 )/(n- k)
Stepwise Regression
The purpose of stepwise regression is to select, from a large
number of predictor variables, a small subset of variables that
account for most of the variation in the dependent or criterion
variable. In this procedure, the predictor variables enter or are
removed from the regression equation one at a time. There are
several approaches to stepwise regression.
Multicollinearity
Multicollinearity
Original
Variable
Code
1
2
3
4
D1
D2
D3
1
0
0
0
0
1
0
0
0
0
1
0
SPSS Windows
The CORRELATE program computes Pearson product moment correlations
and partial correlations with significance levels. Univariate statistics,
covariance, and cross-product deviations may also be requested.
Significance levels are included in the output. To select these procedures
using SPSS for Windows click:
Analyze>Correlate>Bivariate
Analyze>Correlate>Partial
Scatterplots can be obtained by clicking:
Graphs>Scatter >Simple>Define
REGRESSION calculates bivariate and multiple regression equations,
associated statistics, and plots. It allows for an easy examination of
residuals. This procedure can be run by clicking:
Analyze>Regression Linear