Professional Documents
Culture Documents
Goodness-of-fit-measures:
Multiple R:
Tells how strong the linear relationship is. R = square root of R 2. In this case
approaches to 1 (0,99), so the relationship is positive.
Adjusted R square:
Adjusted R-squared, incorporates the models degrees of freedom. It is
interpreted as the proportion of total variance that is explained by the model.
97,55% of total variance is explained by regression model.
R Square:
Coefficient of determination. This shows how well our model predicts the
quantity of population. 98,91% of the variation of the population quantity is
explained by the regression model.
Standard error:
The observed values of the quantity of population differ from the theoretical
(or estimated) values, on average, by 285,27 thousands of people.
Observations:
Number of observations used in the regression (n) = 19
The ANOVA (analysis of variance) table splits the sum of
squares into its components.
SS = Sum of Squares.
Regression MS = Regression SS / Regression degrees of freedom.
Residual MS = mean squared error (Residual SS / Residual degrees of
freedom).
F: Overall F test for the null hypothesis.
Significance F: The significance associated P-Value.
These are the categories we will examine: Regression, Residual, and Total.
The Total variance is partitioned into the variance which can be explained by
the independent variables (Regression) and the variance which is not
explained by the independent variables (Residual, sometimes called Error).
Regression Sum of Squares (SS) - the Mean Sum of Squares between the
groups. It quantifies the variability between the groups of interest.
Residual Sum of Squares (SS) - the Error Mean Sum of Squares. It
quantifies the variability within the groups of interest.
Total Sum of Squares (SS)
F F test
The F column, not surprisingly, contains the F-statistic. Because we want to
compare the "average" variability between the groups to the "average"
variability within the groups, we take the ratio of the Between Mean Sum of
Squares to the Error Mean Sum of Squares. That is, the F-statistic is
calculated as F = MSB/MSE.
Since the p-value (significance F) = 0,00000094 < .05 = , we conclude
that the regression model is a significantly good fit; i.e. there is only a
0.000094 % possibility of getting a correlation this high assuming that the
null hypothesis is true.
t Stat this column shows the 2-tailed p-values used in testing the null
hypothesis that the coefficient (parameter) is 0. Using an alpha of 0.05:
The coefficient for intercept is significantly different from 0 because
its p-value is 0,00000094, which is smaller than 0.05.
The coefficient for Number of hospitals established is significantly
different from 0 because its p-value is 0,042, which is smaller than
0.05.
The rest values are not statistically significant at the 0,05 level since the p-
values are greater than 0,05.
The p-values for all the coefficients with the exception of the coefficient for
Number of hospitals established are bigger than 0.05. This means that we
cannot reject the hypothesis that they are zero (and so can be eliminated
from the model). This is also confirmed from the fact that 0 lies in the
interval between the lower 95% and upper 95% (i.e. the 95% confidence
interval) for each of these coefficients.
95% Confidence Interval these are the 95% confidence intervals for the
coefficients. Confidence intervals help to put the estimate from the
coefficient into perspective by seeing how much the value could vary.