Professional Documents
Culture Documents
Slide 1
Compu
ters II
Types of multiple regression
Slide 2
Compu
ters II
Stepwise multiple regression
Slide 3
Compu
ters II
Differences in statistical outputs
Slide 4
Compu
ters II
Differences in solving stepwise regression problems
Slide 5
Compu
ters II
Slide 6
Compu
ters II
Slide 7
Compu
ters II
Slide 8
Compu
ters II
Slide 9
ters II
Slide
10
ters II
Slide
11
ters II
Slide
12
Fifth, click on
the Continue
button to close
the dialog box.
ters II
Slide
13
ters II
Slide
14
Second, click
on the Continue
button to close
the dialog box.
ters II
Slide
15
Click on the OK
button to
request the
regression
output.
ters II
Slide
16
ters II
Slide
17
Descriptive Statistics
RINCOM98
HRS1
WRKSLF
PRESTG80
Mean
13.94
41.22
1.88
45.96
Std. Deviation
5.287
12.776
.331
14.174
N
145
145
145
145
ters II
Slide
18
ters II
Slide
19
Fourth, click on
the OK button to
produce the output.
Third, mark the checkboxes
for the transformations that
we want to test in evaluating
the assumption.
ters II
Slide
20
Mean
95% Confidence
Interval for Mean
Lower Bound
Upper Bound
5% Trimmed Mean
Median
Variance
Std. Deviation
Minimum
Maximum
Range
Interquartile Range
Skewness
Kurtosis
Statistic
13.35
12.52
Std. Error
.419
14.18
13.54
15.00
29.535
5.435
1
23
22
8.00
-.686
-.253
.187
.373
True is the
correct answer.
ters II
Slide
21
ters II
Slide
22
Lower Bound
Upper Bound
Statistic
40.99
39.10
5% Trimmed Mean
Median
Variance
Std. Deviation
Minimum
Maximum
Range
Interquartile Range
Skewness
Kurtosis
Std. Error
.958
42.88
41.21
40.00
161.491
12.708
4
80
76
10.00
-.324
.935
.183
.364
True is the
correct answer.
ters II
Slide
23
ters II
Slide
24
Mean
95% Confidence
Interval for Mean
5% Trimmed Mean
Median
Variance
Std. Deviation
Minimum
Maximum
Range
Interquartile Range
Skewness
Kurtosis
Lower Bound
Upper Bound
Statistic
44.17
42.45
Std. Error
.873
45.89
43.82
43.00
194.196
13.935
17
86
69
18.00
.401
-.630
.153
.304
True is the
correct answer.
ters II
Slide
25
ters II
Slide
26
Second, click on
the OK button to
produce the output.
ters II
Slide
27
RESPONDENTS INCOME
NUMBER OF HOURS
WORKED LAST WEEK
Logarithm of Reflected
Values of HRS1 [LG10(
81-HRS1)]
Pearson Correlation
Sig. (2-tailed)
N
Pearson Correlation
Sig. (2-tailed)
N
Pearson Correlation
Sig. (2-tailed)
N
Logarithm of
Square Root
Inverse of
NUMBER OF
Reflected
of Reflected
Reflected
HOURS
Values of
Values of
Values of
RESPONDEN
WORKED
HRS1 [LG10( HRS1 [SQRT(
HRS1 [-1/(
TS INCOME
LAST WEEK
81-HRS1)]
81-HRS1)]
81-HRS1)]
1
.337**
-.231**
-.303**
-.059
.
.000
.005
.000
.475
168
149
149
149
149
.337**
1
-.871**
-.981**
-.361**
True is the correct answer.
.000
.
.000
.000
.000
"number of176
149
176The correlation
176 between 176
hours
worked
in
the
past
week" and.743**
-.231**
-.871**
1
.946**
"income" was statistically significant
.005
.000(r=.337, p<0.001).
.
.000
.000
A linear
149
Square Root of Reflected
Values of HRS1 [SQRT(
81-HRS1)]
Inverse of Reflected
Values of HRS1 [-1/(
81-HRS1)]
Pearson Correlation
Sig. (2-tailed)
N
Pearson Correlation
Sig. (2-tailed)
N
-.303**
.000
149
-.059
.475
149
176variables. 176
-.981**
.000
176
-.361**
.000
176
.946**
.000
176
.743**
.000
176
176
176
1
.
176
.502**
.000
176
.502**
.000
176
1
.
176
ters II
Slide
28
ters II
Linearity test: respondents income
and occupational prestige score
Slide
29
Correlations
RESPONDENTS INCOME
RS OCCUPATIONAL
PRESTIGE SCORE
(1980)
Logarithm of PRESTG80
[LG10(PRESTG80)]
Square Root of
PRESTG80
[SQRT(PRESTG80)]
Inverse of PRESTG80
[-1/(PRESTG80)]
Pearson Correlation
Sig. (2-tailed)
N
Pearson Correlation
Sig. (2-tailed)
RS
OCCUPA
TIONAL
PRESTIG
RESPONDEN E SCORE
TS INCOME
(1980)
1
.440**
.
.000
168
168
.440**
1
.000
168
Logarithm of
Square Root
Inverse of
PRESTG80
of PRESTG80 PRESTG80
[LG10(PRES
[SQRT(PRES
[-1/(PREST
TG80)]
TG80)]
G80)]
.436**
.440**
.414**
.000
.000
.000
168 correct answer.
168
168
True is the
.985**
.996**
.936**
Pearson Correlation
Sig. (2-tailed)
N
.436**
.000
168
255
255
255
255
Pearson Correlation
Sig. (2-tailed)
N
Pearson Correlation
Sig. (2-tailed)
N
.440**
.000
168
.414**
.000
168
.996**
.000
255
.936**
.000
255
.996**
.000
255
.982**
.000
255
1
.
255
.962**
.000
255
.962**
.000
255
1
.
255
ters II
Slide
30
ters II
Slide
31
Second, click on
the OK button to
produce the output.
ters II
Slide
32
ters II
Slide
33
ters II
Slide
34
Having decided to
use the baseline
model for the
interpretation of this
analysis, the SPSS
regression output
was re-created.
To run the baseline
model, select Regression
| Linear from the
Analyze model.
ters II
Slide
35
ters II
Slide
36
ters II
Slide
37
Sixth, click on
the Continue
button to close
the dialog box.
ters II
Slide
38
Click on the
Save button to
save the
standardized
residuals to the
data editor.
ters II
Slide
39
Click on the
Continue
button to close
the dialog box.
ters II
Slide
40
Click on the OK
button to
request the
regression
output.
ters II
Slide
41
ters II
Slide
42
ters II
Slide
43
ters II
Slide
44
Model
1
2
Std. Error
Multiple regression Adjusted
assumes that
theof
R
R Square
R Square
the Estimate
errors are
independent and there is no
.424a
.180
.174
4.804
serial correlation.
Errors are the
b
.507
.257
.247
4.588
R Square
Change
.180
.077
F Change
31.350
14.788
df1
1
1
df2
143
142
Sig. F Change
.000
.000
Durbin-W
atson
1.866
ters II
Slide
45
Multicollinearity - question
ters II
Slide
46
ters II
Slide
47
ters II
Slide
48
Regression
Residual
Total
Regression
Residual
Total
Sum of
Squares
723.647
3300.795
4024.441
1034.982
2989.460
4024.441
df
1
143
144
2
142
144
Mean Square
723.647
23.082
F
31.350
Sig.
.000a
517.491
21.053
24.581
.000b
ters II
Slide
49
ters II
Slide
50
ters II
Slide
51
Model
1
2
(Constant)
PRESTG80
(Constant)
PRESTG80
HRS1
Unstandardized
Coefficients
B
Std. Error
6.669
1.358
.158
.028
2.862
1.632
.135
.028
.118
.031
Standardized
Coefficients
Beta
.424
.363
.285
t
4.911
5.599
1.754
4.898
3.846
Sig.
.000
.000
.082
.000
.000
Collinearity Statistics
Tolerance
VIF
1.000
1.000
.954
.954
1.049
1.049
ters II
Slide
52
ters II
Slide
53
In the table
of variables
entered and
removed,
"number of
hours worked
in the past
week" [hrs1]
was added to
the
regression
equation in
model 2.
ters II
Slide
54
Model
1
2
(Constant)
PRESTG80
(Constant)
PRESTG80
HRS1
Unstandardized
Coefficients
B
Std. Error
6.669
1.358
.158
.028
2.862
1.632
.135
.028
.118
.031
Standardized
Coefficients
Beta
.424
.363
.285
t
4.911
5.599
1.754
4.898
3.846
Sig.
.000
.000
.082
.000
.000
Collinearity Statistics
Tolerance
VIF
1.000
1.000
.954
.954
1.049
1.049
ters II
Slide
55
ters II
Slide
56
ters II
Slide
57
ters II
Slide
58
In the table of
variables
entered and
removed,
"occupational
prestige
score"
[prestg80]
was added to
the regression
equation in
model 1.
The increase in R Square as a
result of including this variable
was .180 which was statistically
significant, F(1, 143) = 31.350,
p<0.001.
ters II
Relationship of the third independent variable and
the dependent variable evidence and answer 2
Slide
59
Coefficientsa
Model
1
2
(Constant)
PRESTG80
(Constant)
PRESTG80
HRS1
Unstandardized
Coefficients
B
Std. Error
6.669
1.358
.158
.028
2.862
1.632
.135
.028
.118
.031
Standardized
Coefficients
Beta
.424
.363
.285
t
4.911
5.599
1.754
4.898
3.846
Sig.
.000
.000
.082
.000
.000
Collinearity Statistics
Tolerance
VIF
1.000
1.000
.954
.954
1.049
1.049
ters II
Slide
60
ters II
Slide
61
Validation analysis:
set the random number seed
ters II
Slide
62
ters II
Slide
63
Validation analysis:
compute the split variable
ters II
Slide
64
ters II
Slide
65
ters II
Slide
66
ters II
Slide
67
First, scroll
down the list of
variables and
highlight the
variable split.
ters II
Slide
68
Click on the
Rule button
to enter a
value for split.
ters II
Slide
69
ters II
Slide
70
Click on the OK
button to
request the
output.
ters II
Slide
71
ters II
Slide
72
ters II
Slide
73
SW388R7
Data Analysis &
Computers II
Slide 74
ters II
Slide
75
Is the dependent
variable metric and the
independent variables
metric or dichotomous?
Examine all independent
variables controls as
well as predictors
No
Incorrect
application of
a statistic
Yes
Ordinal variables included
in the relationship?
No
True
Yes
ters II
Slide
76
Ratio of cases to
independent variables at
least 5 to 1?
No
Inappropriate
application of
a statistic
Yes
Ratio of cases to
independent variables at
preferred sample size of at
least 50 to 1?
Yes
True
No
ters II
Slide
77
Yes
True
If more than one
transformation
satisfies normality,
use one with
smallest skew
No
False
Yes
Use transformation
in revised model,
no caution needed
No
Use untransformed
variable in analysis,
add caution to
interpretation for
violation of normality
ters II
Slide
78
Probability of Pearson
correlation (r) <=
level of significance?
If independent variable
was transformed to
satisfy normality, skip
check for linearity.
No
No
Yes
Yes
Use transformation
in revised model
True
Weak
relationship.
No caution
needed
ters II
Slide
79
Probability of Levene
statistic <= level of
significance?
No
True
Yes
False
ters II
Slide
80
Yes
False
No
True
ters II
Slide
81
Yes
Pick revised regression with
transformations and omitting
outliers for interpretation
True
No
Pick baseline regression with
untransformed variables and all
cases for interpretation
False
ters II
Slide
82
Residuals are
independent,
Durbin-Watson between
1.5 and 2.5?
Yes
True
No
False
NOTE: caution
for violation of
assumption of
independence of
errors
ters II
Slide
83
Yes
True
No
False
NOTE: halt the
analysis if it is not
okay to simply
exclude the
variable from the
analysis.
ters II
Slide
84
No
False
Yes
No
False
Yes
Small sample, ordinal
variables, or violation of
assumption in the
relationship?
No
True
Yes
ters II
Slide
85
No
False
Yes
Small sample, ordinal
variables, or violation of
assumption in the
relationship?
No
True
Yes
ters II
Slide
86
No
False
Yes
Significance of R2 change
for variable <= level of
significance?
Yes
No
False
ters II
Slide
87
Direction of relationship
between included variables
and DV interpreted oorrectly?
No
False
Yes
Small sample, ordinal
variables, or violation of
assumption in the
relationship?
No
True
Yes
ters II
Slide
88
Yes
No
False
ters II
Slide
89
No
False
Yes
Shrinkage in R (R for
training sample - R for
validation sample) < 2%?
Yes
True
No
False
ters II
Slide
90
Homework Problems
Multiple Regression Stepwise Problems - 1
ters II
Slide
91
Homework Problems
Multiple Regression Stepwise Problems - 2
The statement of the stepwise
regression problem identifies the
dependent variable and the
independent variables from which we
will extract a parsimonious subset.
ters II
Slide
92
Homework Problems
Multiple Regression Stepwise Problems - 3
ters II
Slide
93
Homework Problems
Multiple Regression Stepwise Problems - 4
ters II
Slide
94
Homework Problems
Multiple Regression Stepwise Problems - 5
The assumption of normality requires that
each metric variable be tested. If the
variable is not normal, transformations
should be examined to see if we can
improve the distribution of the variable. If
transformations are unsuccessful, a
caution is added to any true findings.
ters II
Slide
95
Homework Problems
Multiple Regression Stepwise Problems - 6
The assumption of linearity is examined
for any metric independent variables that
were not transformed for the assumption
of normality.
ters II
Slide
96
Homework Problems
Multiple Regression Stepwise Problems - 7
After incorporating any
transformations, we look for outliers
using standard residuals as the
criterion.
ters II
Slide
97
Homework Problems
Multiple Regression Stepwise Problems - 8
We compare the results of the regression
without transformations and exclusion of
outliers to the model with transformations
and excluding outliers to determine
whether we will base our interpretation on
the baseline or the revised analysis.
ters II
Slide
98
Homework Problems
Multiple Regression Stepwise Problems - 9
ters II
Slide
99
Homework Problems
Multiple Regression Stepwise Problems - 9
ters II
Slide
100
Homework Problems
Multiple Regression Stepwise Problems - 10
ters II
Slide
101
Homework Problems
Multiple Regression Stepwise Problems - 11
ters II
Slide
102
Homework Problems
Multiple Regression Stepwise Problems - 12
ters II
Slide
103
Homework Problems
Multiple Regression Stepwise Problems - 13