Growth Model – DVAR (i.e., ΔDVAR)
|
Level 1 Model |
DVARijk = Π0jk + Π1jk (G2ijk) + Π2jk (TIMEMG1Gijk) + Π3jk (G2 × TIMEijk) + eijk
|
Level 2 Model |
Π0jk = θ0 + b00j + c00k
|
Π1jk = θ1
|
Π2jk = θ2 + b20j
|
Π3jk = θ3
|
Growth Model – Letter-Word Identification
|
Level 1 Model |
LWIDit = Π0 + Π1 (Grade 1 Month) + Π2 (Grade 2) + Π3 (Grade 2 Month) + ε |
Level 2 Model |
Π0 = β00 + β01 (DVAR) + r0
|
Π1 = β10 + β11 (DVAR) + r1
|
Π2 = β20 + β21 (DVAR) + r2
|
Π3 = β30 + β31 (DVAR) + r3
|
Growth Model – Passage Comprehension
|
Level 1 Model |
PCit = Π0 + Π1 (Grade 1 Month) + Π2 (Grade 2) + Π3 (Grade 2 Month) + ε |
Level 2 Model |
Π0 = β00 + β01 (DVAR) |
Π1 = β10 + β11 (DVAR) + r1
|
Π2 = β20 + β21 (DVAR) |
Π3 = β30 + β31 (DVAR) + r3
|
Model fitting: A series of sequential models were initially tested to examine which model best fit the data: (a) fixed intercept–fixed slope, (b) random intercept–fixed slope, (c) fixed intercept–random slope, and (d) random intercept–random slope. Following this test, the better fitting model was examined for variability in intercepts and slopes. Results from the −2 log likelihood test suggest that the random intercept–random slope model improved the description of growth for letter–word identification, χ2(7) = 40.22, p < .001, whereas the fixed intercept-random slope model was a better descriptor for growth in reading, χ2(7) = 80.45, p < .001. We then entered ΔDVAR into the model at Level 2. Nonsignificant variance components were fixed to achieve greater power in model estimation. |
Proportion Reduction in Variance |
|
where τ̑qq (UC) represents the tau estimate for a given parameter (e.g., intercept) for the unconditional model and τ̑qq (C) is a tau estimate for the conditional model (Raudenbush & Bryk, 2002). |