You are on page 1of 14

Report has been prepared by Kamran Selimli and Sahil

Mesimov
Report 4

Independent sample t test


The Independent Samples t Test or student t test check the means of two independent groups in order to
determine population means are significantly different or not. In SPSS we enter analyze – compare mean – one
independence sample test . Here we 1st choose Salary for test variable and for grouping variable we should choose
gender. Define groups- write Male and female in the groups appropriately. We get these;
We have mean and standard deviation

Here , We first look at Sig.of Levene’s test. Why? Because if this number is more than 0.05 it shows us that equal variance
assumed and assumption is tenable. .473 is bigger than .05 .
And next sig.(2-tailed) determines that should this assumption be rejected or not. If sig. is less than 0.05, it is not going to
be rejected. We should reject this because our sig is more than 0.05. and it is not significantly different.

Paired sample t test


The Paired Samples t Test checks two means that are from the same or related items or objectives . The purpose
of the test is to determine the mean difference between paired observations on a particular outcome is
significantly different from zero or not. We enter SPSS – analyze –compare mean – paired sample t test.
We got this result and let’s interpret. We should look at the sig(2-tailed). We see that this number is .000 it means p < .05
and that is why, mean difference between the two related groups is statistically significant.

One way Anova


It is useful for determining if a significant differences in means scores on the dependent variable exist across 2 or more
independent groups. Enter SPSS – analyze – compare mean – one way anova – choose your variable – in option choose
descriptive , homogeneity of variance test, welch , means plot. WE got this outcome;

Test of Homogeneity of Variances

EstimatedSalary

Levene Statistic df1 df2 Sig.

.111 1 9998 .739


Here, in Levene we get significance value ( p value). And if our sig is bigger than p value( .05) it shows you have equal
variance.In our data we see sig. is .739 and we know that our assumption matched.

ANOVA

EstimatedSalary

Sum of Squares df Mean Square F Sig.

Between Groups 4.314E9 1 4.314E9 1.304 .253

Within Groups 3.307E13 9998 3.307E9

Total 3.307E13 9999

The most important part of the table is the sig. column. Sig. indicates us statistically significance value – p value. If this
value is less than .05 we understand that there are some group’s means are different from others.And now we get .253 and it
is bigger than .05. It let us know we don’t hаve аny stаtisticаlly significаnt differences between the group meаns.

Robust Tests of Equality of Means

EstimatedSalary

Statistica df1 df2 Sig.

Welch 1.304 1 9.960E3 .253

a. Asymptotically F distributed.
Simple regression
Simple lineаr regression is а stаtisticаl method thаt аllows us to leаrn and study relаtionships between two continuous
vаriаbles for getting result of how much dependent variables depend on independent variables.But there is an issue here.
We 1st have to check these to learn relation between both variables. We could make from graphs-scatter.
Unfortunately, we got this but we are continuing to work on this.
In Spss choose analyze-regression-linear-select variables – in statistics select Confidence İnterval and Durbin Watson -
Casewise Diagnostics – in plot make operation – in save Select standardized- OK
Model Summaryb

Model R R Square Adjusted R Square Std. Error of the Estimate Durbin-Watson

1 .007a .000 .000 5.7511878E4 2.010

a. Predictors: (Constant), Age


b. Dependent Variable: EstimatedSalary

Here we see R is a measure of the power of relation between the


two variables and it is too little. That is why, normally it is not reported
but we assume it is ok and continue.
R square shows the percentage of variance and it is 0 now.It means we
can’t explain dependent variable because of independent one

ANOVAb

Model Sum of Squares df Mean Square F Sig.

1 Regression 1.715E9 1 1.715E9 .518 .472a

Residual 3.307E13 9998 3.308E9

Total 3.307E13 9999

a. Predictors: (Constant), Age


b. Dependent Variable: EstimatedSalary

This table let’s us know this is a significant model or not.Our sig


is .472 (.472 > .05) and it means that it is not significant

Coefficientsa

Standardized
Unstandardized Coefficients Coefficients 95% Confidence Interval for B

Model B Std. Error Beta t Sig. Lower Bound Upper Bound

1 (Constant) 101627.158 2210.581 45.973 .000 97293.973 105960.343

Age -39.487 54.840 -.007 -.720 .472 -146.984 68.009


Coefficientsa

Standardized
Unstandardized Coefficients Coefficients 95% Confidence Interval for B

Model B Std. Error Beta t Sig. Lower Bound Upper Bound

1 (Constant) 101627.158 2210.581 45.973 .000 97293.973 105960.343

Age -39.487 54.840 -.007 -.720 .472 -146.984 68.009

a. Dependent Variable: EstimatedSalary

Residuals Statisticsa

Minimum Maximum Mean Std. Deviation N

Predicted Value 9.799432E4 1.009164E5 1.000902E5 414.1354961 10000

Residual
-1.0031298E5 1.0055136E5 .0000000 5.7509002E4 10000

Std. Predicted Value -5.061 1.995 .000 1.000 10000

Std. Residual -1.744 1.748 .000 1.000 10000

a. Dependent Variable: EstimatedSalary


Multiple regression
The general purpose of multiple regression is to learn more about the relationship between several independent and a dependent

one.Enter spss – analyze – regression- choose variables – in statistics, plots and save select some stuffs.

Descriptive Statistics

Mean Std. Deviation N

EstimatedSalary 1.000902E5 5.7510493E4 10000

Age 38.92 10.488 10000

Balance 7.648589E4 6.2397405E4 10000

Standardized Predicted Value


.0000000 1.00000000 10000

CustomerId 1.57E7 71936.186 10000


Correlations

Standardized
EstimatedSalary Age Balance Predicted Value CustomerId

Pearson Correlation EstimatedSalary 1.000 -.007 .013 .007 .015

Age -.007 1.000 .028 -1.000 .009

Balance .013 .028 1.000 -.028 -.012

Standardized Predicted Value .007 -1.000 -.028 1.000 -.009

CustomerId .015 .009 -.012 -.009 1.000

Sig. (1-tailed) EstimatedSalary . .236 .100 .236 .063

Age .236 . .002 .000 .171

Balance .100 .002 . .002 .107

Standardized Predicted Value .236 .000 .002 . .171

CustomerId .063 .171 .107 .171 .

N EstimatedSalary 10000 10000 10000 10000 10000

Age 10000 10000 10000 10000 10000

Balance 10000 10000 10000 10000 10000

Standardized Predicted Value 10000 10000 10000 10000 10000

CustomerId 10000 10000 10000 10000 10000

Variables Entered/Removedb

Model Variables Entered Variables Removed Method

1
CustomerId, Standardized Predicted Value,
. Enter
Balancea

a. Tolerance = .000 limits reached.

b. Dependent Variable: EstimatedSalary

Model Summaryb
Std. Error of the
Model R R Square Adjusted R Square Estimate Durbin-Watson

1 .021a .000 .000 5.7505849E4 2.011

a. Predictors: (Constant), CustomerId, Standardized Predicted Value, Balance

b. Dependent Variable: EstimatedSalary


R shows weak association.
Rsquare is .000 means dependent variable can’t be
explained by independent variables.
Adjusted R square is also .000 ( R square is similar to
adjusted R)

ANOVAb

Model Sum of Squares df Mean Square F Sig.

1 Regression 1.526E10 3 5.087E9 1.538 .202a

Residual 3.306E13 9996 3.307E9

Total 3.307E13 9999

a. Predictors: (Constant), CustomerId, Standardized Predicted Value, Balance

b. Dependent Variable: EstimatedSalary

Our sig is bigger than .05 and it is not statistically


significant result .

Binomial t test
The binomiаl test is used when аn experiment hаs two possible outcomes like success and failure аnd you hаve
an ideа аbout whаt the probаbility of success is. In SPSS enter analyze – nonparametric test – binomial test.
In our data 1 is female and 0 is male . We assumed that 20 percentage of population is excited and we see our sig. is less
than .05 . That is why our assumption is valid we should not reject this.

One sample t test


T-tests аre a type of hypothesis test thаt аllows you to compаre meаns.
Go to analyze – compare mean – one sample t test .

One-Sample Statistics

N Mean Std. Deviation Std. Error Mean

IsActiveMember 10000 .52 .500 .005

Here we see our sig. is less than .05 and we know that sample is different than population.
T(9999) = 103.062
P values = .000

You might also like