Professional Documents
Culture Documents
Table of Contents Chapter 1 Getting Started Downloading and preparing the source data base Creating and merging tall skinny files for your research database Chapter 2 Constructing Pupil and School/Classroom Variables Approaches to scaling variables Handling missing data Chapter 3 Using Plausible Values and Survey Design Weights The Function of Survey Design Weights The Function of Replicate Design Weights The Use of Plausible Values Chapter 4 Estimating Basic Hierarchical Linear Models The null HLM and variance components A two-level HLM with variables at level 1 Random or fixed effects? A two-level HLM with variables at levels 1 and 2 Intercepts and slopes as outcomes 15 11 5 2
Second, it included the code to create a PISA ID variable that can be used for matching all files at the student level. The code also calls for the data to be sorted by this ID variable, and a file handle is given for the PISA2003_pupil.sav data set: *** inserted pisaid here. compute countryx=country*10000000000. compute schlidx=schoolid*100000. compute pisaid=countryx+schlidx+stidstd. compute countryz=100000*country. compute schlid=countryz+schoolid. format pisaid (f11.0) schlid (f8.0). execute. sort cases by pisaid. save outfile="f:\pisa2003\data\pupil\pisa2003_pupil.sav". execute.
TIMSS 2003. The sources data bases for TIMSS 2003 should be available for download from the TIMSS web site at the following address in March 2005: http://timss.bc.edu/timss2003.html The data files referred to in this manual were those made available to TIMSS National Coordinators around the time of the press release in December 2004. The SPSS syntax file Join_M3.sps (available on the UIS website) can be used to merge the SPSS portable files into an international dataset. Just modify the list of countries, the datafile type (student=BSG, math teacher=BMT, science teacher=BST, school=BCG, etc), and the sort variables (depending on the corresponding data file type). As the add file command in SPSS is limited to 50 files, only 50 countries/jurisdictions can be processed at one time. The syntax file included on the website does not include the US state of Indiana, one of the benchmarking jurisdictions participating in TIMSS 2003.
Additional technical information about the structure of the data files, as well as how they were created and cleaned was included in the file: Cleaning_Docu_Version_3.0.pdf, which was distributed to TIMSS National Coordinators.
Int_Stui.TXT
Read_Stui.SPS
PISA_2003_Pupil .SAV
Weights.SPS Pared.SPS SES.SPS
W E I G H T S .SAV
P A R E D .SAV
S E S .SAV
JoinPupil.SPS
Pupil Master.SAV
Figure 1. Data structure for the use of tall skinny files.
analysis techniques generally strive to classify people into a small number of clusters based on their responses to a set of observed continuous variables. Finally, when both the latent and observed variables are categorical, a technique called latent class analysis is commonly used. Like IRT methods, the classification of subjects is based on the entire set of possible response patterns for the observed data. In practice, most researchers use fairly simple approaches for scaling data. For example, on a 6-item scale of self esteem, with 4 possible responses for each item, a common approach is to score the responses as 0, 1, 2, and 3, and then either determine the total score (ranging from 0 to 18) for each respondent, or the average score (ranging from 0 to 3). Although this is not the preferred approach, it usually yields a scale that one can work with, at least in the early stages of analysis. An approach suggested by Mosteller and Tukey is to presume that the underlying latent variable has a logit distribution (like a normal distribution but with fatter tails). Their approach treats the observed scaled scores (e.g., ranging from 0 to 18) as 19 ordinal categories. The scores are recoded into new scale values based on the assumption of a logit distribution and the frequency of each observed score. The new scale values preserve the original order, but the distance between scores varies, depending on the original distribution of scores. A variant of this approach is to presume the underlying latent variable is distributed as a uniform distribution, and rescale the data based on the observed frequency of each attained observed score. We have provided an SPSS syntax file for scaling variables using these two approaches. The second common approach is to conduct a factor analysis of the observed data. In this case, the researcher is presuming that the Likert responses (0, 1, 2, and 3) are measures on a continuous scale. Although this assumption can easily be proven to be false, the results provide an indication of whether there is more than one underlying latent construct. The most straightforward factor analytic technique, principal components analysis, yields a factor score, which is commonly used as the scaled score representing the latent variable. We have used this approach, for example, to construct our measure of socioeconomic status for PISA. IRT is the preferred approach for our 6-item Likert scale, as it considers a much wider range of response patterns. Fitting IRT models requires specialized software, and this can be a rather time-consuming process. On the longer term, we would like to see most of the major constructs we are using in PISA and TIMSS analysed with IRT techniques. That said, however, in practice the simpler techniques suggested above nearly always yield scaled scores that are highly correlated with scores derived with IRT (in our experience with r = 0.98 or higher). To make progress, therefore, many of the variables we use have been scaled using the simpler approaches. The most important variable for PISA and TIMSS, the mathematics scores, were scaled centrally with IRT techniques, and are available on the data downloaded from the web-sites.
which are larger because less information is utilized. Therefore the analyst runs the risk of a Type II error. Moreover, if data are MAR rather than MCAR, estimates can be biased. 2. Pairwise deletion. With pairwise deletion, regression analyses or factor analyses are based on a correlation matrix derived from the data available for each pair of variables. Statisticians tend not to like the approach, as the resulting correlation matrix tends to have some undesirable properties (e.g., sometimes it cannot be inverted). We tend to use it for preliminary exploratory analyses, because it can be done easily in SPSS for ordinary least squares regression. However, for logistic regression or various other non-liner models, it is not an option. This option used to be available for level 1 (e.g., pupil-level) data in HLM, but it was never an option for level 2 (e.g., school-level) data. The latest version of HLM does not make provision for pairwise deletion. 3. Mean substitution with dummy variable adjustment. This method is decribed in more detail in Cohen & Cohen (1985) and Cohen et al. (2003). The analyst creates dummy (flag) variable D that is equal to 1 if data are missing on X and equal to 0 otherwise. He or she then creates a new variable, X*, which is equal to X when data are data are not missing (i.e., when D = 0). When cases are missing on X, X* is set to some constant, c. Normally one sets c equal to the mean value of X. The regression model regresses Y on X*, D, and other independent variables. The coefficient for X* is interpreted as the relationship between Y and X for those subjects for whom data on X were available, while the coefficient for c, if mean substitution was used, is the missing effect, that is, the difference in Y values for those who had data on X and those who did not. One of the problems with this approach is that the estimate for the coefficient for X* pertains to subjects for whom data on X were available, and we what we would really like to know is the estimate for all subjects. Thus, the estimates can be biased, even under MCAR (Jones, 1996). Another problem is that when one has 6 or 7 variables, each with missing data, the analyst had 6 or 7 dummy variables to contend with. This set of variables tends to be collinear, and can therefore cause a regression model to become unstable or yield unrealistically large standard errors. That said, this approach is easy to use, and is quite practical when for dealing with variables that have a small amount of missing data (e.g., less than 10%). 4. Maximum Likelihood/EM Algorithm. This approach examines the relationships among variables intended for use in an analysis, and imputes values for the missing data using iterative approaches (Little & Rubin, 1987). It provides appropriate standard errors for the estimates for studies using large samples (n>200). The model assumes missingness is completely accounted for by variables in the model, and under this assumption, the estimates are the best possible. This procedure works for missing data for Xs and for Y. This technique requires specialized software, although many of the large commercial programs (e.g., SPSS, BMPD, MPLUS) now include modules for imputing missing data. Moreover, the best imputation techniques require the analyst to specify a particular regression model, and include Y in the imputation. The technique then estimates several plausible values for the missing values on X variables, and ideally one should analyze each of the several data sets and aggregate the results (this is known as multiple imputation) (See Allison 2002 for a complete discussion of options for handling missing data). This allows the uncertainty of the imputed values to be represented in the analysis, thus resulting in appropriate standard errors. While this is feasible it requires much more analysis, and is not practical in the exploratory phase of analysis. Another hurdle is that the procedures used to impute values on X do not account for the multilevel structure of most educational data. Multilevel procedures for imputing missing data are in their early stages of development.
Therefore, for this project, we advocate the use of the mean substitution/dummy variable to get us started. The syntax files used to make the long skinny files described earlier include variables with missing data set to the country mean alongside a dummy variable flag. The HLM analyses described in Chapter 4 demonstrate the use and interpretation of this approach. As the project proceeds, we plan to use more advanced methods to determine whether our principal findings are sensitive to the method used for handling missing data.
10
11
n 1 n 2 ( i ) . n i =1
In the PISA and TIMSS surveys, the sampling design calls for national centres to stratify their sampling frame of schools into strata that are likely to be related to school performance. These strata might be certain geographic regions, or they can be types of schools, such as public and private or vocational and academic. Then, schools and students are selected in a two-stage design, with schools being selected proportionately to their size. The strategy for estimating the Jackknife sampling variance is similar to that described above for a simple random sample, except that it entails the pairing of schools within strata, such that schools of 12
similar size are paired. In this case there are now J pairs of schools (j = 1 J), and the formula is:
2 Sampling Variance = sv = ( j ) 2 j =1 J
Rust (1995) provides further details, and shows how this approach takes both the stratification and the two-stage clustering into account. When conducting the analysis, we do not have to draw the random replicate samples within each strata. Instead, a set of weights are provided which do the sampling for us. These weights simply select some cases (by weighting by 1.0 for example) and not select others (by weighting by 0 for example). To get the estimate a particular statistic we must simply weight by the replicate weight for replicate sample 1, compute 1 , and then repeat this step for all of the J replicates. With these j estimates in hand we can estimate the sample variance using the formula above. We have provided SPSS programs to do this operation in a reasonably automated fashion. Other programs such as WESVAR or SUDAN are somewhat better suited for this, but this requires you to learn a new program, and so for this project we prefer to keep all of the operations in SPSS. The TIMSS study uses the jackknife as described above, while PISA uses a variant of this called Balanced Repeated Replication (BRR), modified with a procedure develop by Fay. The BRR approach selects one school for removal from each stratum, but then doubles the weights for all other schools. The Fay modification, which avoids problems associated with estimation in small strata, uses weights of -0.5 for the school selected for removal, and -1.5 for the schools selected for inclusion in each replicate. (Thus, schools are not really removed, but the effect is the same.) In PISA the BRR procedure has 80 replicate samples, and the formula for the sampling variance is:
2 Sampling Variance = sv =
2 80 2 ( i ) 80 i =1
The BRR procedure can also be accommodated with SPSS, although the syntax is somewhat awkward, especially for regression coefficients. We provide examples for estimating means and regression coefficients. The multilevel regression package, HLM (or the alternative, MLWin), explicitly take into account the hierarchical nature of the two-stage (pupils within schools) sample. However, the analyst must also use the overall design weight provided for PISA or TIMSS. With a weighted multilevel regression analysis, the estimates of the standard errors are quite close to those obtained using either the jackknife or BRR procedures. It would of course be feasible to use the replicate weights to run the desired HLM analysis multiple times, and then compute the standard errors using the formula above. However, the results are quite similar, and HLM is not very well set up for doing multiple runs.
13
14
15
Choose MATHPV1
To incorporate plausible values and use design weights select Other Settings and Estimation Settings
mathematics achievement as the dependent variable in the analysis, first select MATHPV1 as the outcome variable and then select Other Settings, then Estimation Settings, then Plausible values. Choose MATHPV1 from the pull down menu labelled Choose first variable from level 1 equation and then double click on MATHPV2, MATHPV3, MATHPV4, and MATHPV to move them from the Possible choices column to the Plausible Values column. Click OK. Now click on the Design Weights button and select TOTWGTN from the pull down menu under Level-1 Weight. As we have already normalized the weight to sum to the number of observations we can leave Normalize weight blank. Also click Generalize at level 1 because students are our unit of analysis.
16
Now we are ready to estimate this null or unconditional model, i.e., model with no predictors at either level 1 or level 2. The level 1 model (Yij= 0j +rij ) predicts math achievement within each school with just Chose TOTWGTN & one school-level parameter, the intercept 0j, which is Generalize at level 1 simply the mean outcome for school j. The level-2 model is 0j = 00 + u0j, where 00 represents the grand-mean outcome (here average math achievement in Russia) and u0j is the random effect associated with unit j. By substituting the level 2 model into the level 1 model, Yij= 00 + u0j + rij, it is easier to see how we are partitioning the variance. That is, an individual students score can be decomposed into the country mean score (00), how much the students school differs from the country mean score (u0j), and how much the individual students score differs from their schools average score (rij). It is the partioning of the residual or random component of the model into school/class and student level components that make this a multilevel model. Running the model. It is useful to give each model that you run a different name, so that you can easily recall it later if you want to make a modification. We also recommend that you assign the output file name a label that corresponds to your command file name. Click on Basic Settings to change the Output File Name. It might be Null Model.TXT. Click on File, then Save as, and then assign the Command File Name as, say, Null model.HLM. For some reason, HLM likes the command files and the data files to be in the same directory, so we recommend that you do so (this will usually be the default). Now click on Run on the menu to estimate the model. You will, hopefully, see the maximum likelihood estimation procedure iterating towards convergence. After this screen disappears, you can view the output by clicking on File and then View Output. The output appears in an ASCII text format (most likely in Notepad), which you can save under a different filename if you wish or copy pieces out of it into a spreadsheet. Interpreting the output. This first part of the output recalls what you have just estimated you should see evidence that the data were weighted in the analysis, that all 5 plausible values were used, as well as the specific model estimate (in this case a two level model with no predictors at either level). The most informative output for the null model, however, is found near the bottom of the file.
17
Estimate of grand mean achievement, 00. The standard error can be used to construct a confidence interval around this estimate.
Intraclass correlation. As noted above, the Null model partitions variation in the dependent variable (Yij ) into two components: between classes (Var(u0j) = 00) and within classes (Var(rij) = 2). In the output above we see estimates for 00=2133.09 and 2=3830.38. The proportion of the total variance that is between classes is called the Intra-class correlation ( = 00 / (00 + 2)) , which in this model is estimated to be = 2133.09/(2133.19 + 3830.83) = .36. Thus, in the Russian TIMSS 2003 data, 36% of the variation in 8th grade students math scores is between schools. Such a large between school variance component is an indication that the threat of alpha inflation that would result from assuming the data came from a simple random sample is considerable (by as much as a factor of 10, see Kreft & de Leeuw, 1998: 10) and that there is considerable variation that could be explained using school or class-level variables. To assess how much Russian achievement scores vary between schools we can construct a plausible value range around the school level means using the estimate of 00 (Raudenbush & Bryk, 2002: 71): 00 +/- 1.96(001/2) = 507.62 +/- 1.96*(3830.381/2) = (417.1, 598.1) This indicates that school mean achievement is likely to range from 417 to 598 across 95% of the schools in Russia that serve 8th grade students.
18
19
Here grand mean achievement (00), interpreted as the average of the school means on math achievement, is 507.5, similar to the estimate in the Null model. Since we interpret this coefficient when HPEC_CLM=0, i.e., if each student was at their class average level of parental education, we say that the intercept is unadjusted for parental education. That is, it is not adjusted for between class differences in the distribution of students parents education level. The average parental education-math achievement slope (10) is 10.51., meaning that we could expect a student whose parents have a one unit higher ISECD level (e.g., an upper secondary vs. a post secondary, sub bachelors degree) to have a 10.51 higher mathematics score. As the ratio of both of these coefficients to their corresponding standard errors is greater than 1.96, we consider them statistically significant at the .05 level (here the p-values are <.0001). A coefficient of -20.38 on the dummy variable flag M_HPEDij is an indication that students that skipped this question scored about 20 below average on mathematics achievement. Next we interpret the variance components:
Level 1 residual variance (2) has been reduced to 3710.02, compared to 3830.38 in our model
without any predictors (the random effects ANOVA or Null Model). We can calculate the proportion of variance explained by students highest parental education by comparing the 2 estimates of these two models (see left). Proportional variance explained at level 1 Here we see that adding students 2(random ANOVA)-2 (HPED_CLM) highest level of parental education as a 2(random ANOVA) predictor of math achievement reduced the within-school variance by only = 3830.38 - 3710.02 = 0.314 3.14%.
3830.38
We can also see that the variance in the intercept, Var(0j)= 00, is 2166.24 with a 2 of 2791.3, to be compared to a critical value of 2 with J-1=213 degrees of freedom. We infer that a large difference exists across the school means. The estimated variance in the parental education-math achievement slope, Var(1j)=11, is 48.9 with a of 2 of 242.4 with J-1=213 degrees of freedom, p<.081. Using conventional standards for detecting statistical significance (p<.05) we fail to reject the null hypothesis that 11=0 . As with the intercept, we can use this estimate of variance of the education-math achievement slope to construct a 95% plausible value range around the average parental education-math achievement slope (10) of 10.51. 10 +/- 1.96(111/2) = 10.51+/- 1.96*(48.901/2 ) = (-3.20, 24.23) 20
This indicates that education-math achievement slope is likely to range from (-3.20, 24.23) across 95% of the schools in Russia that serve 8th grade students. This suggests that we might try to model this variance. The HLM output also provides information on the covariance of the random effects, Cov(0j,1j) = 01. Above the Final estimation of fixed effects in the HLM output you will find the variance-covariance matrix (see the format below)also called the Tau matrix. The variances, Var(0j)= 00=2166.24 and Var(1j)=11= 48.9 are on the left to right diagonal and the covariance, Cov(0j,1j) = 01= -73.60 is on the off diagonal. In the table below the Tau matrix these estimates are reported as correlations. The correlation between 0j and 1j = -.226, indicating that 8th grade classes with higher average math achievement tend to have less of a relationship between parental education and math achievement than class with lower average math achievement (i.e., achievement is more equitably distributed in the classes with higher average achievement). 0j 1j
Random or fixed effects? In the random coefficients model presented above, we allowed both the intercept, 0j, and the slope, 1j, to randomly vary across schools. Substituting the level 2 model into the level one model gives us a mixed model (which is what HLM is actually estimating):
What we would normally think of as the 'intercept' is now made up of two components a fixed effect, 00, and a random effect u0j, which varies across schools. HLM provides a t-test of the hypothesis that the fixed effect is zero (H0: 00= 0), and a 2 test that the variance in u0j is zero [H0: Var(u0j) = 0]. Similarly, the slope for HPED_CLMij is made up of a fixed effect, 11 and a random effect u1j. As noted above, one can think of the slope as an average slope 11 and an increment that is unique to each school, u1j. Again we ask whether the average slope is significant, and whether the schools vary in their slopes. The same applies for any additional level one predictor that we add to the model, Xq. If the variances of the group-level residuals are insignificant, we constrain them to be zero (e.g., if p> .10). This is done by clicking on the respective error term in the model and making it go gray (see below). By doing this you are fixing Var(u0j) = 0, i.e., that the slopes do not vary between schools.
21
To fix a parameter variance to zero, click on the level 2 error term to make it go gray. This is equivalent to setting Var(u0j) =0. Note that u1j *HPED_CLMij also disappears from the mixed model equation.
If the parameter variances are non-zero (e.g., if p<= .10), it means we have some betweengroup variance to work with for examining the effects of some group-level variables. Once these are in the model they play a similar role to the fixed portion of the within-group effects.
22
to our equation modeling between school variance in class-average math achievement (Boj). As CLHPEDij was not missing for any Russian class, no correspond dummy flag is needed.
The results from the fixed portion of the model are presented below.
Grand mean achievement (00), interpreted as class mean average achievement in classes with an average level of parental education, class size between 25 and 32, and average climate, is 507.5, similar to the estimate in the Null model and the random coefficients model run earlier. The coefficient on class average parental education (10) is 10.98, which again, is similar to the coefficient estimated on this variable in the random coefficients model and is statistically significant (p<.0001). In general, we would not expect the adding of class-level explanatory variables to level 2 models explaining variation in the intercept to affect the size of previously estimated level-1 coefficients, i.e., we would not expect class level variables to explain with class relationships. An exception to would be when a level-1 variable is correlated with a level 2 residual (e.g., cov (Xqij, uoj) 0), a violation of the assumptions underlying HLM. In this case the adding of a level 2 explanatory variable could affect the size of the estimated coefficient on Xqi, in effect, removing a form of omitted variable bias from the model (for discussion of this and other model misspecification issues see Chapter 9 of Raudenbush & Bryk, 2002, Chapter 9 of Snijders & Bosker, 2000).
23
Turning to the estimated coefficients on the level 2 variables, we see that the estimated effect of class-average parental education is large (01 = 52.95) and statistically significant (p<.0001). A class whose average parental education is one ISCED level higher than another class is expected to have a 53 point higher average math score. Note that the between school effect of parental education on student achievement (b )is nearly 5 times the within school effect (w), although the variation in parental education is considerably larger within than between classes (2 of .92 within classes compared to 2 of .47 between classes). This relationship can be seen in the graph below. A one standard deviation increase in students highest level of parental education is associated with about a 10 point increase in class average achievement (.92 * 10.98), while the expected difference in class average achievement between classes one standard deviation below the mean and a class at the mean is about 25 (.47* 52.95). For a description of how to construct graphs like this with HLM, see http://www.ssicentral.com/hlm/hlm6/hlm6.htm
The estimated effect of being in a small or large class, as opposed to a medium size class was not statistically significant in the above model. Average school academic climate (05 = 15.81), an index calculated by taking the mean of 8 questions asked of the principal regarding teachers job satisfaction; teachers understanding of the schools curricular goals; teachers degree of success in implementing the schools curriculum; teachers expectations for student achievement; parental support for student achievement; parental involvement in school activities; students regard for school property; and students desire to do well in school (with very high =1 and low=4) is marginally significant (p= .072). A standard deviation worsening in school climate (a .35 increase in the index) is associated with a -5.53 reduction in class average achievement (0.35 * -15.81). Although not statistically significant, a coefficient of -11.7 on the variable school climate missing (M_PPSCj) is an indication that sampled schools that did not fill out the school questionnaire or skipped all of the items on climate had classes that were about 11 points below average on average class mathematics achievement. Next we examine the variance components in the output:
24
Here we see that the variance in the intercept, Var(0j)= 00, is 1493.33 with a 2 of 1895.2, to be compared to a critical value of 2 with J-7 (The intercept plus the 6 coefficients estimated)=207 degrees of freedom. We can compare the variance in this model with 4 Proportional variance explained at level level 2 explanatory variables (and two dummy 00 (random coeff.)- 00 (current model) variable missing flags) to the variance 00 (random coeff.) reported in the random coefficients model reported above (where 00(random coeff.)= = 2166.24 - 1493.33 = 0.311 2166.24)). Here we see that class average parental education, class size, and school 2166.24 climate explain 31 percent of the variance in class average math achievement. Note that there remains significant variance in the intercept left to explain (H0: , Var(0j)= 0, reject with p<.0001). The variance in the HPED_CLM slope remains similar to the variance in the random coefficients model, which we would expect since we have not yet added any variables to explain this variance. Intercepts and slopes as outcomes Finally, we use class and school level variables to try and explain the between school variance in the HPED_CLM slope (Var(1j)= 11), that is, why some classes have stronger relationships between parental education and student math achievement than other classrooms. For this example, we will use the same level 2 variables to explain the variance in 1j that we used to explain variance in the class average math achievement, 0j (see below).
This model examines whether classes with higher levels of parental education, small or large classes, and classes in schools with a lower academic climate have a stronger or weaker within class relationship between students parental education and their math achievement. For example, we might hypothesize that classes with higher average parental education will have less disparity in math achievement because the teacher may have fewer at risk kids and can focus more attention on them or that there might be stronger peer effects. A similar argument could be made for why there might be a flatter parental education/math achievement slope in smaller classes or in schools with better academic climates. Although we have chosen here to keep the same class/school variables in both the level 2 equation explaining variance in the intercept and variance in the parental education/math achievement slope there is no requirement that they be the same. If either theory, prior research, or a personal hunch leads 25
you to believe that the important predictors might be different, feel free to construct different models. Be aware, however, misspecification of the model explaining variance in one random coefficient (e.g., 0j ) can lead to biased coefficients in models explaining variance in other random coefficients (1j). See Chapter 9 of Raudenbush & Bryk (2002) for a detailed discussion of specification issues at Level 2. The intercepts and slopes as outcomes model:
26
Since the coefficients on our level 2 variables explaining between class variation in the intercept (0j) are similar to the previous model, we will only focus on the coefficients for variables included to explain between school variation in the parental education/math achievement slope (1j). Although none of the coefficients are statistically significant at the p<=.05 level, we will go through their interpretation for instructional purposes. A one unit increase in class average parental education is associated with slight reduction in the parental education/math achievement slope (11 = -2.26). Thus, while a medium size class with average parental education and average school climate has a predicted parental education/math achievement slope of 8.49, a medium size class with one standard deviation higher parental education (.45) has a predicted parental education/math achievement slope of parental education/math achievement slope of 7.43, calculated as: 10+ (11 *SD(CLHPED_I)) = 8.49 + (-2.26 *.45))= 7.43 By graphing these predicted values we see a slight flattening of the association between students parental education and their math scores in classes with higher average levels of parental education.
Small classes, between 1 and 24 students, have a stronger relationship between parental education and math achievement than medium size classes (10+12 = 8.49 + 5.11 = 13.6, p=.075). Large classes also have a steeper slope, although the difference is less and not statistically significant (10+13 = 8.49 + 2.13 = 10.62, p=.755). By graphing these predicted values we can see visually how the effect of being in a small math class is stronger in classes with higher average levels of parental education.
27
Next we examine the variance components in intercepts and slopes as outcomes model:
Here we see that the variance in the slope, Var(1j)= 11, is 40.93 with a 2 of 226.85, to be compared to a critical value of 2 with J-7 (The intercept 10 plus the 6 coefficients estimated)=207 degrees of freedom. We can Proportional variance explained at level 2 compare the variance in the prior model with no explanatory variables predicting the 11 (No explan vars.)- 11 (current model) variance in 1j to this model. Here we see 11(No explan vars.) that class average parental education, class = 50.46 40.93 = .189 size, and school climate explain 19 percent of the variance in the parental 50.46 education/math achievement slope. Further, the variance in this slope is no longer statistically significant (although it was only marginally so before). For additional information on proportional variance explained statistics see the HLM FAQ at http://www.ssicentral.com/hlm/hlm6/hlm6.htm . Happy modeling!
28
References
Allison, P.D. (2002). Missing data. Thousand Oaks, CA: Sage. Beaton, A.E. (1987). The NAEP 1983-1984 Technical Report. ETS, Princeton. Cohen, J. & Cohen, P. (1983) Applied multiple regression/correlation analysis for the behavioral sciences (2nd ed.).Hillsdale, NJ: Erlbaum. Cohen, J., Cohen, P., West, SG, Aiken, L.S. (2003) Applied multiple regression/correlation analysis for the behavioral sciences (rdnd ed.).Hillsdale, NJ: Erlbaum. Goldstein, H. 2003. Multilevel Statistical Models 3rd edition. Available for download at: http://www.soziologie.uni-halle.de/langer/multilevel/books/goldstein.pdf Kreft, I, de Leeuw, J. 1998. Introduction to Multilevel Modeling. Thousand Oaks, CA: Sage Little, R.J.A. & Rubin, D.B. (1987) Statistical analysis with missing data. New York, Wiley. Mosteller, F., & Tukey, J. W. (1977). Data analysis and regression. Reading, MA: AddisonWesley. Raudenbush SW, Bryk AS. 2002. Hierarchical Linear Models: Applications and Data Analysis Methods. Thousand Oaks, CA: Sage. 2nd edition Rust, K.F. (1996). TIMSS 1995 working paper. Snijders TAB, Bosker RJ. 1999. Multilevel Analysis: An Introduction to Basic and AdvancedMultilevel Modeling. Thousand Oaks, CA: Sage. Wu, M. & Adams, R.J. (2002, April). Plausible Values why they are important. Paper presented at the International Objective Measurement Workshop. New Orleans, Louisiana.
29