Professional Documents
Culture Documents
markusk@unr.edu
What it is
Guilty vs. not-guilty Passed vs. failed Long-term resident vs. short-term resident Just like OLS regression
.though substantially different in some respects
Regression technique
Prediction
Finding the most efficient model Finding the most comprehensive model Does X predict Y (or not)?
Hypothesis testing
Odds ratio
Left-hand 8 of 10 .80 Right-hand 6 of 10 .60 Left-hand 8 over 2 4 Right-hand 6 over 4 1.5 Odds (left hand) / Odds (right hand) Odds (right hand) / Odds (left hand)
Odds
Exp(B)
Odds ratio
Logit
Answer #1
Computer simulation shows that often you may draw using the right conclusions when using ANOVA instead of logistic regression (Linney, 1970) And how do you know that you are drawing the right conclusion this time? You only know that you have the right results if you have run the right analysis
Answer #2
Answer #3
You are violating ANOVA, regression assumptions that can lead to misleading conclusions
Data workup
Is it a categorical or continuous predictor? What do high/low values mean? What value do I want to use a reference category? (categorical predictor variables)
Data coding
Data workup
Check distributions
Univariate outliers (EXPLORE) Multivariate outliers: Mahalanobis distances Run OLS regression (!), choose option SAVE Mahalanobis; ignore results Data set now includes Mahalanobis value for each case Are there any Mahalanobis values that are greater than the 2 value for df = k, with k being the number of predictors/terms in your logistic regression model?
verdict responsible
LOGISTIC REGRESSION VARIABLES verdict /METHOD=ENTER responsible /CRITERIA=PIN(.05) POUT(.10) ITERATE(20) CUT(.5).
SPSS Output
Different components
Case processing summary Dependent variable coding Beginning model (Block 0) Statistics for overall model fit Classification tables Summary of model variables
SPSS Output
Constant-only model
Significant chi-square: Your model is better at predicting than constant-only model (Block 0)
Cases with predicted probability of >=.50 guilty Cases with predicted probability of < .50 not guilty Crude measure, because it ignores
Significance of estimates B, based on S.E. and Wald test Continuous predictors variable has df = 1 Positive B = Exp(B) > 1 Negative B = Exp(B) < 1
Account needs to be defined as categorical (sonamed button in logistic regression menu of SPSS)
LOGISTIC REGRESSION VARIABLES verdict /METHOD=ENTER Responsible Account /CONTRAST (Account)=Indicator /CRITERIA=PIN(0.05) POUT(0.10) ITERATE(20) CUT(0.5).
Block 0 identical
Model fit
Model variables
1-predictor model
2-predictors model
The hypothesis that the account variable s coefficients are 0 has to be rejected.
1-predictor model
2-predictors model
Matching actual category membership against predicted probabilities Step 1: Save predicted probabilities for models Step 2: Use ROC to plot actual category membership against predicted probabilities
STEP 1: SAVE predicted probabilities for models (first model: pred_1; second model pred_2)
LOGISTIC REGRESSION VARIABLES verdict /METHOD=ENTER Responsible /SAVE=PRED /CRITERIA=PIN(.05) POUT(.10) ITERATE(20) CUT(.5). LOGISTIC REGRESSION VARIABLES verdict /METHOD=ENTER Account Responsible /CONTRAST (Account)=Indicator(1) /SAVE=PRED /CRITERIA=PIN(.05) POUT(.10) ITERATE(20) CUT(.5).
Model 2
ROC PRE_2 BY verdict (1) /PLOT=CURVE(REFERENCE) /PRINT=SE /CRITERIA=CUTOFF(INCLUDE) TESTPOS(LARGE) DISTRIBUTION(FREE) CI(95) /MISSING=EXCLUDE.
1-predictor model
2-predictors model
Hierarchical regression; r2
Stepwise regression
Letting the program decide which variables to include in the model based on pre-defined statistical criteria Iterative, automated process Incompatible with theory-guided hypothesis testing, but Used in applied research
Letting the program decide which variables to include in the model based on pre-defined statistical criteria Iterative, automated process Incompatible with theory-guided hypothesis testing, but Used in applied research
Forward selection
Backward (de)selection
Pre-defined criteria:
Yes include the variable in forward selection No exclude the variable in backward (de)selection
LOGISTIC REGRESSION VARIABLES verdict /METHOD=FSTEP(LR) Honest8 Credible10 Believable11 Deceitful12 /CONTRAST (Account)=Indicator(1) /CRITERIA=PIN(0.05) POUT(0.10) ITERATE(20) CUT(0.5).
Model calibration
LOGISTIC REGRESSION VARIABLES verdict /METHOD=ENTER Account Responsible /CONTRAST (Account)=Indicator(1) /PRINT=GOODFIT /CRITERIA=PIN(0.05) POUT(0.10) ITERATE(20) CUT(0.5).
Model calibration
Interaction terms
Did the defendant show remorse? (yes, no) Gender of mock juror? (m, f)
LOGISTIC REGRESSION VARIABLES verdict /METHOD=ENTER remorse gender gender*remorse /CONTRAST (remorse)=Indicator(1) /CONTRAST (gender)=Indicator(1) /CRITERIA=PIN(.05) POUT(.10) ITERATE(20) CUT(.5). OR LOGISTIC REGRESSION VARIABLES verdict /METHOD=ENTER remorse gender gender*remorse /CRITERIA=PIN(.05) POUT(.10) ITERATE(20) CUT(.5).
Interaction terms
Did the defendant show remorse? (yes=1, no=2) Gender of mock juror? (m=1, f=2)
LOGISTIC REGRESSION VARIABLES verdict /METHOD=ENTER remorse gender gender*remorse /CONTRAST (remorse)=Indicator(1) /CONTRAST (gender)=Indicator(1) /CRITERIA=PIN(.05) POUT(.10) ITERATE(20) CUT(.5).
Interaction terms
Mediational modeling
What is mediation?
Definition: Mediation is when the effect of one variable A onto an outcome variable C is transmitted through another variable B. In other words, A has an indirect effect on C.
Full mediation: A B C
Partial mediation:
1) 2) 3)
that does not necessarily imply that there is now a significant indirect effect