Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2017 Jun 29.
Published in final edited form as: Multivariate Behav Res. 2016 Nov 11;52(1):12–30. doi: 10.1080/00273171.2016.1245600

Full Information Maximum Likelihood Estimation for Latent Variable Interactions With Incomplete Indicators

Heining Cham a, Evgeniya Reshetnyak a, Barry Rosenfeld a, William Breitbart b
PMCID: PMC5489914  NIHMSID: NIHMS867956  PMID: 27834491

Abstract

Researchers have developed missing data handling techniques for estimating interaction effects in multiple regression. Extending to latent variable interactions, we investigated full information maximum likelihood (FIML) estimation to handle incompletely observed indicators for product indicator (PI) and latent moderated structural equations (LMS) methods. Drawing on the analytic work on missing data handling techniques in multiple regression with interaction effects, we compared the performance of FIML for PI and LMS analytically. We performed a simulation study to compare FIML for PI and LMS. We recommend using FIML for LMS when the indicators are missing completely at random (MCAR) or missing at random (MAR) and when they are normally distributed. FIML for LMS produces unbiased parameter estimates with small variances, correct Type I error rates, and high statistical power of interaction effects. We illustrated the use of these methods by analyzing the interaction effect between advanced cancer patients’ depression and change of inner peace well-being on future hopelessness levels.

Keywords: Latent interaction, missing data, maximum likelihood, product indicator


Missing data represent “one of the most important statistical and design problems in research” (Azar, 2002, p. 70). Inappropriate handling of missing data can result in substantially biased parameter estimates and invalid statistical inferences. Missing data can be classified into three types based on missingness mechanisms: missing completely at random (MCAR), missing at random (MAR), and missing not at random (MNAR) (Enders, 2010; Little & Rubin, 2002; Van Buuren, 2012). MCAR assumes that the probability of missingness does not depend on any observed or missing values. MAR assumes that the probability of missingness depends on the observed values but not on the missing values. MNAR assumes that the probability of missingness depends on the missing values. In the past decades, there has been an increasing interest in missing data handling approaches for all these mechanisms. For example, multiple imputation and full information maximum likelihood estimation handle MAR data, and selection modeling and pattern mixture modeling are employed when the data are MNAR (Enders, 2010; Little & Rubin, 2002). Most of these approaches assume linear relationships among variables. In recent years, researchers have developed and investigated missing data handling techniques for estimating interaction effects in multiple regression (Bartlett, Seaman, White, & Carpenter, 2015; Doove, Van Buuren, & Dusseldorp, 2014; Enders, Baraldi, & Cham, 2014; Kim, Sugar, & Belin, 2015; Seaman, Bartlett, & White, 2012; vonHippel, 2009). Specifically, these approaches handle incompletely observed predictors in multiple regression with interaction effects.

On the other hand, there is a lack of research on proper approaches to handle incompletely observed indicators in latent variable interactions when the predictors are not measured directly, but are represented by a confirmatory factor model with indicators in latent variable interactions. To address this issue, we investigated full information maximum likelihood (FIML) estimation for two latent variable interaction methods: product indicator (PI; Kenny & Judd, 1984) and latent moderated structural equations (LMS; Klein & Moosbrugger, 2000; Moosbrugger, Schermelleh-Engel, & Klein, 1997; Schermelleh-Engel, Klein, & Moosbrugger, 1998). We investigate three variants of the PI method: constrained product indicator (CPI; Jöreskog & Yang, 1996; Kenny & Judd, 1984), generalized appended product indicator (GAPI; Wall & Amemiya, 2001), and unconstrained product indicator (UPI; Marsh, Wen, & Hau, 2004). These latent variable interaction methods are available in the major SEM software packages; Mplus (Muthén & Muthén, 1998–2015), R package lavaan (Rosseel, 2012), and LISREL (Jöreskog & Sörbom, 2015) can be used for PI; Mplus and R package nlsem (Umbach et al., 2015) can be used for LMS. We conducted a simulation study to investigate whether FIML produces unbiased parameter estimates, which leads to valid statistical inferences for these methods.

The structure of the manuscript is the following. First, we reviewed three approaches to handle incompletely observed predictors in multiple regression with interaction effects that were relevant to FIML for latent variable interaction methods: listwise deletion, “just another variable” (JAV) multiple imputation, and FIML estimation. We evaluated them theoretically and in light of published simulation studies. We discussed other missing data handling approaches in the General Discussion section. Second, we introduced the latent variable interaction methods (PI and LMS) and provided the details of how FIML handles incompletely observed indicators in each method. Third, we conducted a simulation study to investigate how FIML performed under various conditions, including sample sizes, interaction effect sizes, missing data rates, and missing data mechanisms. Fourth, we applied these methods to a substantive example with incompletely observed indicators. Finally, we discussed our results and findings. Throughout the manuscript, we focused on the situation where the predictors (or indicators of predictors) had missing values. We discussed the situation where the outcome (or indicators of outcome) had missing values in the General Discussion section.

Multiple regression with interaction effects

Without loss of generality, we consider the interaction effect between two predictors ξ1 and ξ2 of outcome η (Equation [1]):

η=γ0+γ1ξ1+γ2ξ2+γ3ξ1ξ2+ζ (1)

where ξ1 ξ2 is the interaction term (product of ξ1 and ξ2); ζ is the disturbance; γ0 is the intercept; γ1, γ2, and γ3 are the regression coefficients. In this section, we described and evaluated three approaches to handle incompletely observed predictors in multiple regression with interaction effects.

Listwise deletion

In listwise deletion (or complete-case analysis), participants are eliminated from the analysis when they have missing values in any variables (ξ1, ξ2, η). Little’s (1992) theoretical work has shown that listwise deletion leads to unbiased γ0, γ1, γ2, γ3 estimates and valid statistical inferences in two situations: (a) predictors are MCAR, and/or (b) the missingness of a predictor depends on the observed values of other predictors. These two conditions are more restrictive than MAR. Another important disadvantage is that, compared to JAV and FIML (discussed later), both which use all observed data, listwise deletion often has lower statistical power due to reduced sample sizes. Simulation studies by Enders et al. (2014) and Seaman et al. (2012) have shown that when ξ1 and/or ξ2 are MCAR, listwise deletion results in unbiased γ0, γ1, γ2, γ3 estimates. However, the lower-order effect (γ0, γ1, γ2) estimates have larger variances compared to those by JAV and FIML. When ξ1 and/or ξ2 are MAR, listwise deletion can produce biased γ0, γ1, γ2, γ3 estimates with smaller variances compared to those by JAV and FIML. Therefore, listwise deletion is not recommended. We do not consider this approach for latent variable interactions.

Just another variable

Just another variable (JAV) is a Bayesian parametric multiple imputation approach (Bartlett et al., 2015; Enders et al., 2014; Kim et al., 2015; Seaman et al., 2012; von Hippel, 2009). Without loss of generality, consider that ξ1 is incompletely observed and ξ2 and η are completely observed in Equation (1). When ξ1 has missing values, the interaction term ξ1 ξ2 is also incompletely observed. JAV assumes that the conditional distribution of the incompletely observed variables (ξ1 and ξ1 ξ2) on the completely observed η and ξ2 is multivariate normal (MVN; Equation [2]):

(ξ1ξ1ξ2)~MVN[(β10+β11η+β12ξ2β20+β21η+β22ξ2),(τ11τ21τ21τ22)] (2)

where ξ1 and ξ1 ξ2 are linearly conditional on η and ξ2 with parameters βs and τs; βs are the regression coefficients; τs are the residual variances and covariances; JAV is an iterative process between two steps. In the first step, missing values of ξ1 and ξ1 ξ2 are imputed using Equation (2). We use the observed values of η and ξ2 and the estimates of βs and τs to impute ξ1 and ξ1 ξ2. In the second step, the estimates of βs and τs are updated by random draws from the β and τ matrices in Equation (2), which are based on the observed values of η and ξ2, imputed values of ξ1 and ξ1 ξ2, and the specified prior distributions of βs and τs.1 Data sets with imputed values are created after every specified number of cycles. Data augmentation algorithm (Enders 2010; Little & Rubin, 2002) or multivariate imputation via chained equations algorithm (Van Buuren, 2012) can be employed in the imputation processes.2 After creating multiple complete data sets with imputed values, we estimate the interaction model in Equation (1) for each data set. Each data set shall provide different sets of regression coefficient estimates and statistical inferences of the regression coefficients. These sets of results are combined (pooled) to one set of regression coefficients estimates and statistical inferences (Enders, 2010; Little & Rubin, 2002).

Theoretical work by Bartlett et al. (2015) and Seaman et al. (2012) has shown that JAV only produces consistent γ0, γ1, γ2, γ3 estimates when ξ1 is MCAR. The key reason is that ξ1 ξ2 is nonlinearly related to ξ2. In fact, the distribution of ξ1 ξ2 is a nonlinear function that involves ξ2 (Aroian, 1944; Bohrnstedt & Goldberger, 1969). Therefore, the imputation model for ξ1 ξ2 in Equation (2) is misspecified. Another reason is that the distribution of ξ1 ξ2 is nonnormal, which also violates the assumption of the conditional distribution of ξ1 ξ2. Given these reasons, Seaman et al. (2012) have shown that the estimates of βs and τs are only consistent when ξ1 is MCAR. Only consistent estimates of βs and τs produce imputed values of ξ1 and ξ1 ξ 2 that result in consistent γ0, γ1, γ2, γ3 estimates. In addition to the theoretical work, simulation studies by Bartlett et al. (2015), Enders et al. (2014), Kimet al. (2015), and Seaman et al. (2012) have investigated the biases and variances of γ0, γ1, γ2, γ3 estimates. When ξ1 is MCAR, JAV produces unbiased γ0, γ1, γ2, γ3 estimates. Moreover, the lower-order effect estimates have smaller variances compared to those obtained by listwise deletion. When ξ1 is MAR and normal, JAV produces unbiased γ3 estimates. When ξ1 is MAR and nonnormal, JAV can result in biased γ3 and lower-order effect estimates with larger variances compared to those produced by listwise deletion. The bias of γ3 becomes more pronounced with the increase of interaction effect size.

Full information maximum likelihood estimation

Without loss of generality, we continue to consider the same scenario that ξ1 is incompletely observed and ξ2 and η are completely observed. Full information maximum likelihood estimation (FIML) maximizes the sample log-likelihood function (Equation [3]) to estimate γ0, γ1, γ2, and γ3.

l=i=1Nln(f(wi)) (3)

where l is the sample log-likelihood function; i=1N(·) is the summation function across cases 1 to N; ln(·) is the natural logarithm function; f(wi) is the probability density function for the vector of variables w for case i. When all variables are completely observed in Equation (1), w = (η, ξ1, ξ2, ξ1ξ2). FIML allows cases with missing values on some variables. In other words, it utilizes all observed variables for each case, and thus the name “full information.” Equation (3) implies that we need to specify the distributions for w (i.e., specify f (wi)). FIML assumes that all variables (η, ξ1, ξ2, ξ1 ξ2) are linearly related with each other and are multivariate normally distributed. When ξ1 and ξ1 ξ2 are incompletely observed, FIML assumes that their conditional distribution on η and ξ2 is multivariate normal. The sample log-likelihood function in Equation (3) can be rewritten as Equation (4):

l=i=1N(-ki2ln(2π)-12lni-12(wi-μi)Ti-1(wi-μi)), (4)

where (·)T is the matrix transpose function; | · | is the determinant function; (·)−1 is the matrix inverse function; ki is the number of completely observed variables for case i; μi and Σi are the elements of the model-implied mean vector and covariance matrix that correspond to the observed data (wi) for case i. The log-likelihood function is maximized to obtain the FIML parameter estimates.

FIML is closely related to JAV. Both assume that the incompletely observed variables (ξ1 and ξ1 ξ2) are linearly related to the completely observed variables (η and ξ2). Both make the same assumption about the conditional distribution of ξ1 and ξ1 ξ2. In fact, JAV is equivalent to FIML asymptotically (Little & Rubin, 2002, p. 201). The theoretical justification by Bartlett et al. (2015) and Seaman et al. (2012) that JAV can only produce consistent γ0, γ1, γ2, γ3 estimates when ξ1 is MCAR is generalizable and applicable to FIML asymptotically. This is also supported by Yuan’s (2009) and Yuan and Savalei’s (2014) theoretical work. In the more general settings to multiple regression with interaction effects, they have shown that FIML estimation that maximizes Equation (4) does not produce consistent covariance estimates of the variables when the MAR variables (here, ξ1 and ξ1 ξ2) and the completely observed variables accounting for the missingness (here, η and ξ2) are nonlinearly related. As a result, γ0, γ1, γ2, γ3 estimates, which are derived from these covariance estimates, are not consistent. The simulation studies by Enders et al. (2014) have shown that the biases and variances of γ0, γ1, γ2, γ3 estimates are nearly the same between FIML and JAV when sample size ≥ 200. The results further support the equivalence between FIML and JAV for multiple regression with interaction effects.

Latent variable interactions

Multiple regression makes two assumptions about the predictors ξ1 and ξ2. First, it is assumed that ξ1 and ξ2 are perfectly reliable. Second, the underlying constructs of ξ1 and ξ2 are represented by a single variable. When ξ1 and ξ2 are not perfectly reliable and the underlying constructs are represented by confirmatory factor models, latent variable interaction methods are appropriate to model the interaction effects between ξ1 and ξ2 on outcome η. In this section, we introduced two methods to model latent variable interactions: product indicator (PI) and latent moderated structural equations (LMS). For each method, we described and evaluated how FIML handles the incompletely observed indicators of ξ1 and ξ2. For simplicity of the presentation, we treated outcome η as a single variable, although in both PI and LMS, η could be a factor with multiple indicators.

Product indicator

In latent variable interactions, the interaction term ξ1 ξ2 is the product of factors ξ1 and ξ2. Unlike multiple regression, we cannot multiply ξ1 and ξ2 to form the interaction term. To represent ξ1 ξ2, product indicator (PI) multiplies the indicators of ξ1 and ξ2 (product indicators) as indicators of ξ1 ξ2. There are three main issues when specifying the models in PI. The first issue is the selection of product indicators of ξ1 ξ2. Consider a widely studied situation where ξ1 and ξ2 have three indicators each (X1, X2, X3 and X4, X5, X6) (e.g., Cham, West, Ma, & Aiken, 2012; Kelava et al., 2011; Marsh et al., 2004; Wall & Amemiya, 2001), so there are nine possible product indicators (X1X4, X1X5, X1X6, X2X4, X2X5, X2X6, X3X4, X3X5, X3X6). Simulation studies suggest creating three product indicators using the three most reliable indicators of ξ1 and ξ2, respectively (Jackman, Leite, & Cochrane, 2011; Marsh et al., 2004; Wu, Wen, Marsh, & Hau, 2013). This produces unbiased γ3 estimates with unbiased estimated standard errors. It also reduces the complexity of model specification in SEM software, and it can be employed in scenarios with any number of indicators of ξ1 and ξ2.3 The second issue is centering the indicators of ξ1 and ξ2 in the presence of missing values. Centering indicators at their means and then computing the product indicators can simplify the model specification by fixing the latent intercepts of the indicators and product indicators to zero. This greatly increases the chances of successful model convergence (Algina & Moulder, 2001). When the indicators of ξ1 and ξ2 have missing values, we suggest using the FIML estimated means of the indicators. Adopting these two solutions and assuming all indicators have equal reliabilities, the measurement model of ξ1, ξ2, and ξ1 ξ2 can be expressed as in Equation (5):

(X1X2X3X4X5X6X1X4X2X5X3X6)=(λX100λX200λX3000λX400λX500λX6000λX1X400λX2X500γX3X6)(ξ1ξ2ξ1ξ2)+(δX1δX2δX3δX4δX5δX6δX1X4δX2X5δX3X6), (5)

where λs are the factor loadings; δs are the unique factors, which are set to be uncorrelated with all other unique factors. The subscripts of λs and δs correspond to the indicators. The third issue is the choice of PI method out of its three variants: constrained product indicator (CPI), generalized appended product indicator (GAPI), and unconstrained product indicator (UPI). The three variants of PI add model constraints related to ξ1 ξ2, which provide more information to describe the distribution of ξ1 ξ2 (Aroian, 1944; Bohrnstedt & Goldberger, 1969; Jöreskog & Yang, 1996). Table 1 summarizes the constraints for each PI variant (see also Cham et al., 2012; Kelava et al., 2011). CPI has the most constraints, followed by GAPI and UPI. Simulation studies find that CPI produces unbiased γ3 estimates with higher statistical power than GAPI and UPI when the indicators of ξ1 and ξ2 are normally distributed. When these indicators are nonnormally distributed, CPI produces biased γ3 estimates. GAPI and UPI produce unbiased γ3 estimates when sample size ≥ 500 (Cham et al., 2012; see also Coenders, Batista-Foguet, & Saris, 2008; Marsh et al., 2004; Wall & Amemiya, 2001). Note that these simulation studies only investigate the situation where the indicators of ξ1 and ξ2 are completely observed.

Table 1.

Model constraints of product indicator methods.

Model constraint Constrained product indicator (CPI) Generalized appended product indicator (GAPI) Unconstrained product indicator (UPI)
Mean of ξ1ξ2
Variance and covariances of ξ1ξ2 × ×
Factor loadings of product indicators ×
Unique factor variances of product indicators ×
Unique factor covariances of product indicators

When the indicators of ξ1 and/or ξ2 have missing values, FIML can be used with PI (Zhang, 2010). FIML for PI is an extension of FIML for multiple regression with interaction effects. The sample log-likelihood function is the same as Equations (3) and (4); w is a vector of indicators of ξ1 and ξ2, product indicators of ξ1ξ2, and outcome η. In our example, w = (X1, X2, X3, X4, X5, X6, X1X4, X2X5, X3X6, η). We expect that the results for JAV and FIML for multiple regression with interaction effects are generalizable to FIML for PI; FIML for PI only produces consistent γ0, γ1, γ2, γ3 estimates when the indicators of ξ1 are MCAR. The first and the key reason is that FIML for PI also maximizes the sample log-likelihood function, which incorrectly assumes that the product indicators are linearly related to the indicators of ξ1 and ξ2. When the indicators of ξ1 are MAR, FIML estimation produces inconsistent covariance estimates of w (Yuan, 2009; Yuan & Savalei, 2014). As a result, γ0, γ1, γ2, γ3 estimates are not consistent. Second, FIML for PI incorrectly assumes that the conditional distribution of the product indicators on the completely observed elements in w is multivariate normal. Similar to the distribution of ξ1ξ2, the distribution of product indicators is nonnormal and nonlinearly related to the indicators of ξ1 and ξ2 (Aroian, 1944; Bohrnstedt & Goldberger, 1969). Third, η is nonnormally distributed even when the indicators of ξ1 and ξ2 are multivariate normally distributed (Klein & Moosbrugger, 2000; more later). Our expectation is supported by the simulation studies by Zhang (2010) that have investigated FIML for CPI and UPI. When the indicators of ξ1 and ξ2 are MCAR, CPI and UPI produce unbiased γ3 estimates under normality of indicators. Only UPI produces unbiased γ3 estimates under nonnormality of the indicators. When the indicators are MAR, CPI and UPI lead to biased γ3 estimates in all indicator distribution conditions. The bias increases as the missing data rate increases. These results match the theoretical work for JAV and FIML for multiple regression with interaction effects. In the MCAR condition, the differences between CPI and UPI under different indicator distributions are consistent with those when the indicators are completely observed.

Latent moderated structural equations

Latent moderated structural equations (LMS) belong to a class of latent variable interaction methods termed distribution analytic methods (Kelava et al., 2011). In distribution analytic methods, the measurement model for ξ1ξ2 is not specified. Distribution analytic methods use estimation procedures to calculate model parameters (including γ0, γ1, γ2, γ3) and make statistical inferences directly. LMS begins by decomposing the vector of ξ1 and ξ2 as in Equation (6) (Klein &Moosbrugger, 2000, equation 8):

ξ=Az(ξ1ξ2)=(a11a0a21a22)(z1z2), (6)

where ξ is a column vector of ξ1 and ξ2; A is a lower triangular matrix produced by the Cholesky decomposition of the covariance matrix of ξ; z is a column vector of variables z1 and z2; z1 and z2 are mutually independent and are standard normally distributed. LMS also rewrites the interaction model in Equation (1) (Klein &Moosbrugger, 2000, equation 2):

η=γ0+Γξ+ξTΩξ+ζη=γ0+(γ1γ2)(ξ1ξ2)+(ξ1ξ2)(0γ300)(ξ1ξ2)+ζ (7)

where Γ is a row vector of regression coefficients of the lower-order effects except the intercept; Ω is a square matrix containing the interaction effect coefficient(s) in upper diagonal. Combining Equations (6) and (7), the interaction model can be decomposed into Equation (8) (Klein & Moosbrugger, 2000, equation 10):

η=γ0+ΓAz+zTATΩAz+ζη=γ0+ΓA(z10)+ΓA(0z2)+(z10)TATΩA(z10)+(z10)TATΩA(0z2)+ζη=(γ0+ΓA(z10)+(z10)TATΩA(z10))+(ΓA+(z10)TATΩA)(0z2)+ζ. (8)

Equations (6) to (8) decompose the interactive relationship of ξ1 and ξ2 to η into linear and nonlinear components. The last line of Equation (8) shows that η is linearly related to z2, but nonlinearly related to z1. Because of the nonlinear component in Equation (8), η is nonnormally distributed even when ξ1 and ξ2 are multivariate normally distributed. This is why FIML for PI (as well as JAV and FIML for regression with interaction effects) incorrectly assumes that η is normally distributed.

LMS correctly specifies the nonlinear relationships between ξ1, ξ2, and η, as well as the nonnormal distribution of η. In turn, LMS correctly specifies the sample log-likelihood function of FIML to obtain the parameter estimates and other results. Assuming the indicators of ξ1 and ξ2 are multivariate normal, Klein and Moosbrugger (2000, p. 463) have shown that the conditional distribution of the indicators of all factors (ξ1, ξ2, and η) on z1 are multivariate normal. Equivalently, the multivariate distribution of indicators of all factors (ξ1, ξ2, and η) is a mixture of multivariate normal distributions across the values of z1 (Equation [9]; Klein &Moosbrugger, 2000, equation 15):

f(w)=ϕ(z1)ϕ(w)dz1. (9)

Different from PI, w = (X1, X2, X3, X4, X5, X6, η) because no product indicators are involved in LMS; ϕ(z1) is the normal density function of z1, which equals 12πe-z122; ϕ(w) is the multivariate normal density function of w, which equals (12π)k(1)e-(w-μ)T-1(w-μ)2; μ and Σ are the model-implied mean vector and covariance matrix of w; ∫(·) dz1 is the integration function across the values of z1 because z1 is continuous. As discussed previously, LMS does not require any product indicators. Using Equations (3) and (9), the sample log-likelihood function of FIML for LMS is

l=i=1Nln(f(wi))=i=1Nln(ϕ(z1)ϕ(wi)dz1)=i=1Nln(12πe-z122ϕ(wi)dz1)i=1Nln(j=1Mmj2πϕ(wi)). (10)

To maximize the sample log-likelihood function in Equation (10), the computational algorithm is tailored to include numerical integration algorithm(e.g., rectangular integration, Hermite-Gaussian quadrature integration) to approximate the integration (Klein & Moosbrugger, 2000, pp. 464–467). The numerical integration algorithm approximates by splitting the integration to a total of M dimensions, each with a different weight m and a different estimate for ϕ(w). As discussed, FIML for LMS correctly specifies the nonlinear relationships between ξ1, ξ2, and η. It correctly handles the no-normal distribution of η. It assumes that all indicators of ξ1 and ξ2 are multivariate normal. When the distributional assumption holds, FIML for LMS shall produce unbiased γ0, γ1, γ2, γ3 estimates for MCAR and MAR indicators of ξ1 and ξ2. Simulation studies find that when indicators of ξ1 and ξ2 are completely observed and the distributional assumption holds, FIML for LMS produces unbiased γ0, γ1, γ2, γ3 estimates with lower variances compared to those by PI (Cham et al., 2012; Jackman et al., 2012; Marsh et al., 2004; Wu et al., 2013). When the distributional assumption is violated, FIML for LMS can produce biased γ3 estimates (Cham et al., 2012).

Current study

We conducted a simulation study to investigate how FIML for CPI, GAPI, UPI, and LMS performs when the indicators of ξ1 and/or ξ2 have missing values. We hypothesize that when the indicators are MCAR and multivariate normal, all methods lead to unbiased estimates and valid statistical inferences. LMS results in γ0, γ1, γ2, γ3 estimates with the smallest variances. When the indicators are MAR and multivariate normal, only LMS leads to unbiased γ0, γ1, γ2, γ3 estimates and valid statistical inferences. We applied these methods to a substantive example with incompletely observed indicators of ξ1 and ξ2.

Simulation study: Method

Population model

We investigated the interaction model in Equation (1). To set the values of γ0, γ1, γ2, γ3 and the variance of disturbance ζ, we assumed that ξ1, ξ2, and η were measured directly, perfectly reliable, and analyzed via multiple regression. We set γ1, γ2, and disturbance variance such that the linear effects of ξ1 and ξ2 yield a population ρ2 = 0.3. We manipulated three levels of γ3 in terms of ρ2 increment from linear model to interaction model (squared semipartial correlation, sr2) = 0 (no effect), 0.05 (small to medium effect), and 0.10 (medium effect; Cohen, 1992). We used the previous example where ξ1 had three indictors (X1, X2, X3) and ξ2 had three indicators (X4, X5, X6). To show that all methods are applicable for η as a factor, we set η having three indicators. We set the values of factor loadings and unique factor variances so that Cronbach’s α of each construct = 0.7. All unique factors were uncorrelated with each other. We set the latent intercepts of X1 to X6 to an arbitrary value (= 5) to investigate the performance of using FIML to mean center X1 to X6. Table 2 summarizes the parameters.

Table 2.

Parameters of population model.

Structural model

sr2 = 0 (no interaction effect) sr2 = 0.05 (small to medium effect)


Parameter Value Parameter Value
γ3 0.00 γ3 4.43
Variance of disturbance ζ 343.00 Variance of disturbance ζ 318.50
sr2 = 0.10 (medium effect) For all conditions


Parameter Value Parameter Value

γ3 6.26 γ0 10.00
Variance of Disturbance ζ 294.00 γ1 7.00
γ2 7.00
Measurement model of ξ1 and ξ2 Measurement model of η


Parameter Value Parameter Value

Latent intercept 5.00 Latent intercept 0.00
Factor loading 1.00 Factor loading 1.00
Unique factor Variance 1.29 Unique factor variance 630.00
Factor mean 0.00
Factor variance 1.00
Factor covariance 0.50

Sample size

To adequately represent the common situations in behavioral research, we manipulated five levels of sample size (N) = 100, 200, 500, 1,000, 5,000. N = 200 approximates the median sample size in multiple regression with interaction effects (Jaccard & Wan, 1995); N = 5,000 resembles the asymptotic situation.

Missing data rates

The indicators of ξ1 (X1, X2, X3) were incompletely observed, while those of ξ2 (X4, X5, X6) were completely observed. Our substantive example also had similar missing data patterns. Because the computational algorithms for FIML estimation handle any missing data patterns, this setting will not limit the generalizability of results; X1, X2, X3 were set to three missing data rates = 0%, 15%, 25%. The 15% and 25% conditions had been studied extensively (e.g., Collins, Schafer, & Kam, 2001; Savalei, 2010).

Missing data mechanisms

To generate incompletely observed Xs in these scenarios, we first created one set of completely observed Xs. In 15% and 25% missing data rates, X1, X2, and X3 were all MCAR or MAR. In MCAR condition, we generated two scenarios with different missing data patterns. In MAR condition, we generated six scenarios that had different missing data patterns and used different combinations of variables to explain the missingness of X1, X2, and X3. Table 3 describes these missing data scenarios.

Table 3.

Missing data scenarios for simulation study.

(A) Missing completely at random (MCAR) – two scenarios
Scenario 1 We generated one random uniform variable. We created missing values of X1, X2, and X3 for cases in the lower quartile of the random uniform variable based on the missing data rates (15% or 25%). This missing data mechanism resulted in two missing data patterns: all X1, X2, and X3 were either observed or missing.
Scenario 2 We generated three random uniform variables (U1, U2, U3). We created missing values of X1, X2, and X3 for cases in the lower quartile of U1, U2, and U3, respectively, for each variable, based on the missing data rates (15% or 25%) (i.e., X1 with U1, X2 with U2, X3 with U3). This missing data mechanism resulted in maximum number of combinations of missing data patterns for X1, X2, and X3.
(B) Missing at random (MAR) – six scenarios

Scenario 1 We computed the average scores of X4, X5, and X6, S = (X4 + X5 + X6)/3. Given the missing data rate conditions (15% or 25%), we divided the participants into quartiles according to S (i.e., 0%–24.99%, 25%–49.99%, 50%–74.99%, 75%–100%). We then used one uniform random variable to randomly delete X1, X2, and X3 in each quartile based on designated percentages. In the 15% missing data rate condition, the designated percentages from the first to the fourth quartiles were 7.5%, 12.5%, 17.5%, 22.5%. In the 25% condition, the designated percentages were 10.0%, 20.0%, 30.0%, 40.0%. These percentages result in a monotonic increasing relationship between missing data rate and S (Collins et al., 2001).
Scenario 2 The procedures were identical to those of Scenario 1 except for the designated percentages of missing data in the quartiles. In the 15% missing data rate condition, the designated percentages from the first to the fourth quartiles were 25.0%, 5.0%, 5.0%, 25.0%. In the 25% condition, the designated percentages were 10.0%, 20.0%, 30.0%, 40.0%. These percentages result in a convex relationship between missing data rate and S (Collins et al., 2001).
Scenario 3 We divided X1, X2, and X3 into quartiles according to the variables they paired with to form product indicators, respectively (i.e., X1 with X4, X2 with X5, X3 with X6). We then used three separate uniform random variable to randomly delete X1, X2, and X3 in each quartile based on designated percentages. The designated percentages were the same as those in Scenario 1.
Scenario 4 The procedures were identical to those of Scenario 3, except that we divided X1, X2, and X3 into quartiles according to the variables they did not pair with to form product indicators, respectively (i.e., X1 with X5, X2 with X6, X3 with X4).
Scenario 5 The procedures were identical to those of Scenario 3, except that we used the designated percentages that were the same as those in Scenario 2 (i.e., 15% missing data rate: designated percentages = 25.0%, 5.0%, 5.0%, 25.0%; 25% missing data rate: designated percentages = 10.0%, 20.0%, 30.0%, 40.0%).
Scenario 6 The procedures were identical to those of Scenario 4, except that we used the designated percentages that were the same as those in Scenario 2 (i.e., 15% missing data rate: designated percentages= 25.0%, 5.0%, 5.0%, 25.0%; 25% missing data rate: designated percentages = 10.0%, 20.0%, 30.0%, 40.0%).

Among these six scenarios, Scenarios 1 and 2 resulted in two missing data patterns, where all X1, X2, and X3 were either observed or missing. Scenarios 3, 4, 5, and 6 resulted in maximum number of combinations of missing data patterns for X1, X2, and X3.

Latent variable interaction methods

In FIML for CPI, GAPI, and UPI, we mean-centered X1 to X6 using the FIML estimated means and fixed their latent intercepts to zero in the models, so as to improve the chances of successful model convergence (Algina & Moulder, 2001). In FIML for LMS, we freely estimated the latent intercepts of X1 to X6, which is the default model specification in SEM software. We used Hermite-Gaussian quadrature integration with 16 integration points per integration dimension (Klein &Moosbrugger, 2000). All analyses were conducted using Mplus Version 7.11 (Muthén & Muthén, 1998–2015). In all methods, we overrode software defaults by increasing the maximum number of iterations and decreasing the convergence criterion. The Appendix presents the settings.

To sum up, we manipulated five factors (interaction effect size [three levels], sample size [five levels], missing data rate [three levels], missing data mechanism [two levels], latent variable interaction methods [four levels]) resulting in a total of 1,440 conditions. We randomly generated 1,200 replications for each condition in SAS/IML Version 9.3. In all conditions, the indicators X1 to X6 were multivariate normally distributed. We discussed the situation where X1 to X6 were nonnormal in the General Discussion section.

Dependent measures

To investigate the estimates of γ0, γ1, γ2, and γ3, we used these dependent measures: model convergence rate, relative bias, mean squared error, standard error ratio, coverage rate, actual Type I error rate (γ3 only), and statistical power (γ3 only).

Model convergence rate

A set of model results is considered to be properly converged if it satisfies three conditions: (a) no errors reported by Mplus, (b) no negative variance estimates, and (c) no out-of-bound correlation estimates. The model convergence rate in one condition is the proportion of number of properly converged replications to the total number of replications (= 1,200).

Relative bias

Relative bias of a parameter estimate is the difference between the mean parameter estimate across converged replications and the population value relative to the population value.

RelativeBias=(θ^¯-θ)/θ, (11)

where θ refers to any parameter (γ0, γ1, γ2, γ3); θ̂ is the parameter estimate. We used the suggestion by Hoogland and Boosma (1998) that an unbiased parameter estimate should have absolute value of relative bias < 0.05.

Standard error ratio

Standard error ratio is the average ratio of estimated standard error to the standard deviation of parameter estimates across converged replications.

StandardErrorRatio=(k[SE(θ^)/SD(θ^)])/k, (12)

where Σk(·) is the summation function across converged replications; k is the number of converged replications; SE(θ̂) is the estimated standard error of θ̂; SD(θ̂) is the standard deviation of θ̂ across converged replications. Standard error ratio evaluates the bias of estimated standard error. We used the suggestion by Hoogland and Boosma (1998) that an unbiased standard error estimate should have the ratio within 0.9 to 1.1.

Mean squared error

Mean squared error (MSE) is the average squared difference between the parameter estimate and its population value across converged replications.

MSE=k(θ^-θ)2/k. (13)

MSE is also the sum of squared bias and variance of parameter estimate. We prefer a parameter estimate with small MSE.

Actual Type I error rate and statistical power (γ3 only)

Actual Type I error rate is the proportion of converged replications that incorrectly reject the null hypothesis of no interaction effect (H0: γ3 = 0; two-tailed test) at α (set at .05) when the interaction effect is absent (sr2 = 0). Statistical power is the proportion of replications that correctly reject this null hypothesis when the interaction effect is present (sr2 = 0.05 or 0.10). We used the suggestion by Savalei (2010) that actual Type I error rate should be within the interval of α±z(1-α/2)α(1-α)/k. This results in the suggested interval = [0.036, 0.064].

Coverage rate

Coverage rate is defined as the proportion of converged replications that the 95% Wald confidence interval of a parameter includes its population value. Coverage rate reflects both the bias of parameter estimate and the bias of estimated standard error. We used the suggestion by Collins et al. (2001) that a satisfactory coverage rate should be > 0.9.

Results

We presented the simulation study results in the following order: model convergence rate, performance of interaction effect γ3, and performance of lower-order effects (γ0, γ1, γ2). When all indicators were completely observed, the results of all dependent measures were consistent with those in the literature (Cham et al. 2012; Coenders et al., 2008; Klein & Moosbrugger, 2000; Marsh et al., 2004; Wall & Amemiya, 2001). Given the space limit, we did not present these results in the text and summarized the results in tables and figures. We focused on the results when the indicators were MCAR or MAR. The results were consistent across the two scenarios of MCAR and the six scenarios of MAR, respectively. Therefore, we aggregated the results across the various scenarios and presented the results as MCAR and MAR conditions.4 It is noted that the results in the complete data condition were always better than those in the MCAR or MAR conditions because completely observed data have more information. To simplify the presentation, we dropped the term FIML because all four methods (CPI, GAPI, UPI, LMS) used FIML estimation. A full report of the results is available in the online supplementary materials.

Model convergence rate

Across all conditions, LMS produced the highest average convergence rate (99.7%). UPI and CPI performed similarly but were worse than LMS (average coverage rates = 92.2% and 93.7%, respectively). GAPI had the lowest average convergence rate (71.8%, range = 39.6% to 93.3%). In particular, GAPI had low convergence rates in MCAR and MAR conditions. In subsequent analyses, we investigated the first 1,000 converged replications in each condition.

The interaction effect

Parameter estimates and relative biases

Figure 1 presents the mean estimates and relative biases of γ3. The four panels in Figure 1 show the results of different combinations of missing data rate (15% or 25%) and interaction effect size (sr2 = 0.05 or 0.1). In each panel, we presented the results according to the sample sizes (horizontal axis, from N = 100 to N = 5,000), and missingness conditions (three subpanels, left: complete data; middle: MCAR; right: MAR). The vertical axis is the mean estimates. The horizontal line indicates the population value. The gray box highlights the acceptable region of γ3 estimates with relative bias < 0.05.

Figure 1.

Figure 1

Mean estimates and relative biases of interaction effect γ3.

When X1, X2, X3 were MCAR, missing data rate = 25%, and sr2 = 0.05, CPI produced unbiased estimates when N ≥ 200. GAPI and UPI produced unbiased estimates when N ≥ 500 while LMS produced unbiased estimates across all sample sizes. The ranking in terms of relative bias was LMS < CPI < GAPI < UPI. When X1, X2, X3 were MCAR and sr2 = 0.10, CPI and LMS produced unbiased estimates across all sample sizes while GAPI and UPI produced unbiased estimates when N ≥ 1,000. The ranking was LMS < CPI < GAPI ~ UPI. When X1, X2, X3 were MAR, missing data rate = 15%, and sr2 = 0.05, CPI produced unbiased estimates across all sample sizes, but GAPI failed to produce unbiased estimates across all samples sizes. UPI produced unbiased estimates only when N = 5,000, and LMS produced unbiased estimates when N ≥ 200. The ranking was CPI < LMS < UPI < GAPI. When missing data rate = 25% and sr2 = 0.05, CPI produced unbiased estimates when N ≥ 500. GAPI and UPI failed to produce unbiased estimates across all samples sizes. When N = 200, relative bias by GAPI and UPI = 33% and 31%, respectively. LMS produced unbiased estimates across all sample sizes, with slightly large bias at N = 100 (relative bias = 5.6%). The ranking was LMS<CPI <GAPI~UPI. When sr2 = 0.10, CPI produced unbiased estimates when N ≥ 200. GAPI failed to produce unbiased estimates across all sample sizes. When N = 200 and missing data rate = 25%, relative bias by GAPI = 26%. UPI only produced unbiased estimates when missing data rate = 15%. LMS resulted in unbiased estimates across all sample sizes. The ranking was LMS < CPI < UPI < GAPI.

Mean squared error

Table 4 presents the mean squared errors (MSEs) of γ3 estimates. MSE is the sum of squared bias and variance of γ3 estimates. Thus, the conditions that have large relative biases also have large MSEs. In addition, the variance of γ3 estimates decreases when sample size N increases. We prefer a latent variable interaction method with small MSEs.

Table 4.

Mean square errors (MSEs) of interaction effect γ3.

N Missing data rate Complete data Missing completely at random (MCAR) Missing at random (MAR)



CPI GAPI UPI LMS CPI GAPI UPI LMS CPI GAPI UPI LMS
(A) sr2 = 0.05
 100 15% 18.038 65.174 50.969 13.098 21.041 51.618 60.430 16.133 22.991 67.182 78.330 17.448
 100 25% 18.038 65.174 50.969 13.098 25.096 65.182 84.333 16.781 26.629 117.016 94.623 17.163
 200 15% 7.317 10.592 33.881 5.626 8.558 18.053 18.735 6.113 9.595 21.447 23.738 6.218
 200 25% 7.317 10.592 33.881 5.626 9.505 21.121 29.555 6.050 11.185 42.852 36.609 6.439
 500 15% 2.732 3.281 3.486 2.028 3.048 3.807 4.397 1.961 3.151 4.443 4.852 1.990
 500 25% 2.732 3.281 3.486 2.028 3.487 4.492 5.617 2.143 3.952 7.560 7.292 2.178
 1,000 15% 1.281 1.383 1.549 0.949 1.378 1.618 1.770 0.933 1.477 2.053 2.044 0.942
 1,000 25% 1.281 1.383 1.549 0.949 1.459 1.749 1.914 1.013 1.757 3.270 2.954 1.035
 5,000 15% 0.220 0.241 0.264 0.178 0.260 0.285 0.317 0.189 0.304 0.466 0.395 0.187
 5,000 25% 0.220 0.241 0.264 0.178 0.295 0.342 0.354 0.199 0.398 1.044 0.706 0.199
(B) sr2 = 0.10
 100 15% 19.621 72.081 70.521 16.171 21.713 56.324 67.656 18.166 24.396 60.920 72.557 18.612
 100 25% 19.621 72.081 70.521 16.171 24.100 145.940 99.795 18.382 25.988 94.977 126.243 18.380
 200 15% 8.009 13.696 15.435 5.749 8.654 14.578 21.867 6.084 9.259 22.160 26.135 6.052
 200 25% 8.009 13.696 15.435 5.749 9.971 25.702 28.345 6.789 11.956 48.520 46.797 7.029
 500 15% 2.682 3.628 3.834 2.035 3.359 4.606 5.657 2.226 3.611 5.878 6.651 2.204
 500 25% 2.682 3.628 3.834 2.035 3.788 5.391 6.218 2.541 4.431 11.290 9.233 2.552
 1,000 15% 1.304 1.615 1.701 0.908 1.435 1.817 2.063 1.048 1.640 2.540 2.505 1.063
 1,000 25% 1.304 1.615 1.701 0.908 1.724 2.281 2.530 1.101 2.138 5.149 4.077 1.164
 5,000 15% 0.253 0.295 0.320 0.184 0.309 0.366 0.422 0.199 0.377 0.626 0.517 0.205
 5,000 25% 0.253 0.295 0.320 0.184 0.343 0.421 0.455 0.214 0.511 1.583 0.944 0.215

Note. CPI = constrained product indicators; GAPI = generalized appended product indicators; UPI – unconstrained product indicators; LMS = latent moderated

structural equations.

When X1, X2, X3 were MCAR, LMS also had the lowest MSEs across all sample sizes and interaction effect sizes. CPI had larger MSEs than LMS across all conditions. When sample sizes increased, the differences of MSEs between CPI and LMS increased. GAPI and UPI had much larger MSEs than LMS when N ≤ 200 (at least two times as large). With increasing missing data rate, MSEs by GAPI and UPI further increased when N ≤ 200, suggesting large variation of γ3 estimates in these conditions. The ranking was LMS<CPI<GAPI~UPI. When X1, X2, X3 were MAR, LMS performed better than all PI methods across sample sizes, missing data rates, and interaction effect sizes. Unlike PI, MSEs of LMS were not much affected by missing data rates. CPI had larger MSEs than LMS across all conditions. At N = 5,000, the MSEs by CPI were also two times those by LMS. Given that CPI and LMS had acceptable biases of γ3 estimates, the MSE results meant that CPI was a less efficient method than LMS that produced more variable γ3 estimates. GAPI and UPI produced much larger MSEs compared to LMS across all conditions (at least two times as large). The ranking was LMS < CPI < GAPI ~ UPI.

Standard error ratio

Table 5 presents the standard error ratios of γ3 estimates. When the ratio = 1, standard error (SE) is unbiasedly estimated. When the ratio < 1, SE is underestimated. When the ratio > 1, SE is overestimated. We used the suggestion that an unbiased SE estimate should have SER within 0.9 to 1.1 (Hoogland & Boosma, 1998). The results of the complete data condition were consistent with simulation studies in literature. CPI and LMS produced unbiased SEs when N ≥ 200. CPI and LMS produced slightly underestimated SEs when N = 100. The biases of LMS were smaller than those of CPI. GAPI and UPI produced unbiased SEs when N ≥ 1,000. GAPI and UPI often underestimated SEs when N ≤ 500. The ranking of the four methods was LMS < CPI < GAPI ~ UPI.

Table 5.

Standard error ratios (SERs) of interaction effects γ3.

N Missing data rate Complete data Missing completely at random (MCAR) Missing at random (MAR)



CPI GAPI UPI LMS CPI GAPI UPI LMS CPI GAPI UPI LMS
(A) sr2 = 0.05
 100 15% 0.871 0.863 0.900 0.924 0.888 1.120 1.140 0.902 0.873 1.028 1.068 0.884
 100 25% 0.871 0.863 0.900 0.924 0.859 1.021 1.065 0.909 0.866 1.242 1.339 0.916
 200 15% 0.913 0.890 0.778 0.939 0.906 0.826 0.846 0.938 0.891 0.826 0.864 0.938
 200 25% 0.913 0.890 0.778 0.939 0.921 0.800 0.986 0.969 0.913 0.918 0.889 0.954
 500 15% 0.912 0.891 0.893 0.955 0.929 0.895 0.880 1.004 0.944 0.904 0.878 1.002
 500 25% 0.912 0.891 0.893 0.955 0.908 0.877 0.823 0.985 0.913 0.857 0.842 0.985
 1,000 15% 0.920 0.922 0.903 0.972 0.955 0.930 0.925 1.016 0.964 0.914 0.917 1.017
 1,000 25% 0.920 0.922 0.903 0.972 0.978 0.947 0.948 1.002 0.966 0.911 0.903 0.998
 5,000 15% 0.985 0.967 0.950 0.997 0.973 0.964 0.943 1.002 0.972 0.925 0.922 1.011
 5,000 25% 0.985 0.967 0.950 0.997 0.957 0.925 0.939 0.998 0.953 0.886 0.908 1.006
(B) sr2 = 0.10
 100 15% 1.000 0.897 0.863 0.832 0.873 0.987 1.103 0.873 0.849 1.050 1.072 0.878
 100 25% 1.000 0.897 0.863 0.832 0.883 0.989 1.557 0.918 0.890 1.130 1.349 0.923
 200 15% 0.857 0.982 0.910 0.842 0.912 0.869 0.821 0.978 0.918 0.858 0.872 0.985
 200 25% 0.857 0.982 0.910 0.842 0.919 0.790 0.863 0.969 0.898 0.866 0.906 0.958
 500 15% 0.897 1.003 0.938 0.873 0.909 0.872 0.845 1.005 0.915 0.867 0.835 1.011
 500 25% 0.897 1.003 0.938 0.873 0.896 0.851 0.855 0.961 0.898 0.797 0.833 0.967
 1,000 15% 0.929 1.051 0.942 0.899 1.435 1.817 2.063 1.048 1.640 2.540 2.505 1.063
 1,000 25% 0.929 1.051 0.942 0.899 0.920 0.871 0.876 1.012 0.914 0.823 0.844 0.993
 5,000 15% 0.920 1.034 0.944 0.915 0.912 0.890 0.868 1.030 0.909 0.869 0.864 1.020
 5,000 25% 0.920 1.034 0.944 0.915 0.915 0.879 0.890 1.022 0.938 0.872 0.885 1.029

Note. CPI = constrained product indicators; GAPI = generalized appended product indicators; UPI – unconstrained product indicators; LMS = latent moderated structural equations.

When X1, X2, X3 were MCAR, LMS had less biased SEs than all PI methods and produced unbiased SEs when N ≥ 200. CPI and LMS produced slightly underestimated SEs when N = 100. The biases of LMS were smaller than those of CPI. When sr2 = 0.05, GAPI and UPI produced unbiased SEs only when N ≥ 1,000. When sr2 = 0.10, GAPI and UPI often produced biased SEs. The ranking was LMS < CPI < GAPI ~ UPI.

When X1, X2, X3 were MAR, LMS also had less biased SEs than all PI methods. LMS produced unbiased SEs when N ≥ 200. CPI slightly underestimated SEs when N ≤ 500. The biases of LMS were often smaller than those of CPI. GAPI and UPI produced unbiased SEs only when N ≥ 1,000 and sr2 = 0.05. The ranking was LMS< CPI < GAPI ~ UPI.

Actual Type I error rate and statistical power

Figure 2 presents the actual Type I error rates of γ3. We averaged the results across samples sizes because the findings were consistent. We used the suggestion that actual Type I error rate should be between 0.036 and 0.064 (Savalei, 2010). All methods had acceptable Type I error rate of γ3 when X1, X2, X3 were completely observed, MCAR, or MAR. The actual Type I error rates by GAPI and UPI were on the lower bound (= 0.036) in the complete data and MCAR conditions. These results imply that GPI and UPI had lower statistical power in these conditions.

Figure 2.

Figure 2

Actual Type I error rate of interaction effect γ3. The numbers are the actual Type I error rates.

Figure 3 presents the statistical power of γ3 when sr2 = 0.05 (panel A) and 0.10 (panel B), respectively. We averaged the results across samples sizes because the findings were consistent. As expected, statistical power increased when the interaction effect size (sr2) increased. Among all methods, LMS had the highest statistical power when X1, X2, X3 were completely observed, MCAR, or MAR. The ranking of the four methods in terms of statistical power was LMS > CPI > GAPI > UPI.

Figure 3.

Figure 3

Statistical power of interaction effect γ3. The numbers are the statistical power.

Coverage rate

Table 6 shows the coverage rates of γ3. Coverage rate reflects a combination of the bias of γ3 estimate and the bias of the estimated standard error. We used the suggestion that the satisfactory coverage rate should be > 0.9 (Collins et al., 2001). When X1, X2, X3 were completely observed or MCAR, all methods produced satisfactory coverage rates across sample sizes and interaction effect sizes. When X1, X2, X3 were MAR, CPI and LMS produced satisfactory coverage rates across interaction effect sizes and missing data rates. When the sample size, interaction effect size, and missing data rate increased, CPI, GAPI, and UPI produced lower coverage rates than LMS. When N = 5,000, missing data rate = 25%, and sr2 = 0.1, the minimum coverage rate = 91.3% for LMS, 73.8% for CPI, 42.2% for GAPI, and 64.3% for UPI.

Table 6.

Coverage rates of interaction effect γ3 (%).

N Missing data rate Complete data Missing completely at random (MCAR) Missing at random (MAR)



CPI GAPI UPI LMS CPI GAPI UPI LMS CPI GAPI UPI LMS
(A) sr2 = 0.05
 100 15% 94.0 94.4 93.5 94.3 93.0 95.1 94.2 93.8 93.5 95.1 93.8 93.6
 100 25% 94.0 94.4 93.5 94.3 94.0 95.0 93.7 93.6 94.2 95.8 94.5 93.8
 200 15% 91.4 92.9 92.4 92.9 93.3 93.9 93.4 94.4 93.3 94.6 93.8 94.2
 200 25% 91.4 92.9 92.4 92.9 95.2 94.6 93.5 95.1 94.6 95.7 94.4 95.1
 500 15% 92.3 92.9 92.5 93.5 93.4 93.1 92.5 94.7 94.1 94.1 93.3 95.1
 500 25% 92.3 92.9 92.5 93.5 93.6 93.8 93.4 95.0 92.8 93.9 93.8 94.6
 1,000 15% 93.4 93.4 93.4 94.5 94.1 93.3 93.4 95.4 94.1 93.3 93.9 95.4
 1,000 25% 93.4 93.4 93.4 94.5 95.1 93.5 94.4 95.4 94.7 93.2 93.6 95.4
 5,000 15% 94.0 94.5 93.2 95.2 93.9 94.1 94.0 94.5 93.6 89.9 92.1 94.7
 5,000 25% 94.0 94.5 93.2 95.2 93.6 92.9 93.6 94.8 92.5 80.4 86.2 94.8
(B) sr2 = 0.10
 100 15% 91.7 92.1 92.2 92.1 91.7 92.6 90.8 91.4 92.0 92.9 91.0 92.2
 100 25% 91.7 92.1 92.2 92.1 92.5 93.2 91.4 93.1 92.9 93.6 91.9 93.1
 200 15% 92.9 93.4 92.1 94.3 92.7 93.6 92.3 93.7 93.3 94.5 92.4 93.9
 200 25% 92.9 93.4 92.1 94.3 93.1 93.1 92.8 94.9 93.5 94.8 93.3 94.5
 500 15% 94.0 92.4 93.1 95.1 93.3 92.9 93.0 95.9 93.2 93.7 92.9 95.4
 500 25% 94.0 92.4 93.1 95.1 92.3 92.1 91.9 94.7 93.3 93.2 93.5 94.2
 1,000 15% 93.7 92.3 93.6 95.5 93.7 94.1 92.9 94.8 93.3 93.4 93.2 95.0
 1,000 25% 93.7 92.3 93.6 95.5 93.3 92.1 92.3 95.2 92.6 89.3 91.1 95.0
 5,000 15% 94.4 92.3 93.5 95.5 92.9 92.4 91.5 95.5 91.4 86.4 90.1 95.5
 5,000 25% 94.4 92.3 93.5 95.5 92.6 91.5 92.2 95.2 89.4 74.2 84.3 95.5

Note. CPI = constrained product indicators; GAPI = generalized appended product indicators; UPI – unconstrained product indicators; LMS = latent moderated structural equations.

Performance of lower-order effects

In this section, we presented the results of the lower-order effects γ0, γ1, and γ2. A full report of the results is available in the online supplementary materials.

Parameter estimates and relative biases

When X1, X2, X3 were completely observed, CPI and LMS produced unbiased γ0 estimates across sample sizes and interaction effect sizes. GAPI and UPI sometimes underestimated γ0 when N = 100. All four methods produced unbiased γ1 estimates when N ≥ 200. CPI, GAPI, and LMS produced unbiased γ2 estimates across sample sizes and interaction effect sizes. UPI overestimated γ2 when N = 100.

When X1, X2, X3 were MCAR, the four methods also produced unbiased γ0 estimates across sample sizes and interaction effect sizes. For all methods, the relative biases of γ0 estimates decreased when sample size increased. When N = 100 or 200, the relative biases of γ0 estimates by GAPI and UPI were slightly larger than those by CPI and LMS. All methods produced unbiased γ1 estimates when N ≥ 500. When N = 200, CPI and LMS resulted in unbiased γ1 estimates, except the condition of missing data rate = 25%, sr2 = 0.05. The relative bias of γ1 estimates by GAPI and UPI were larger than those by CPI and LMS. UPI overestimated γ1 more often than the other methods. All methods produced unbiased γ2 estimates when N ≥ 200. When N = 100, CPI produced unbiased γ2 estimates, and the other methods overestimated γ2.

All four methods generally performed worse in the MAR condition than in the complete data or MCAR conditions. CPI and LMS produced unbiased γ0 estimates across sample sizes and interaction effect sizes. GAPI and UPI produced unbiased γ0 estimates when N ≥ 500. All methods could overestimate γ1 when N ≤ 200, while LMS had the least bias across conditions. When N ≥ 500, CPI and LMS produced unbiased γ1 while GAPI and UPI overestimated γ1. All methods could overestimate γ2 when N ≤ 200, while CPI had the least bias across conditions. When N ≥ 500, CPI and LMS produced unbiased γ2 while GAPI and UPI overestimated γ2.

Mean square error

Consistent with the results for interaction effect γ3, MSEs of the lower-order effects γ0, γ1, γ2 increased when sample size increased. In addition, MSEs increased slightly when the interaction effect size increased or the missing data rates increased (MCAR or MAR). In the majority of the conditions, LMS outperformed all PI methods with the lowest MSEs. The difference between the MSEs produced by LMS and those by the PI methods decreased when N ≥ 500. GAPI and UPI produced the highest MSEs of the lower-order effects, especially under conditions with N = 100.

Standard error ratio

In general, LMS produced acceptable standard error estimates of the lower-order effects with the least bias. CPI sometimes underestimated the standard errors when N = 100 and when X1, X2, X3 were MCAR or MAR. GAPI and UPI often overestimated standard errors of γ0 when N = 100 or 200.

Coverage rate

Across all conditions, all four methods produced satisfactory and similar coverage rates of the lower-order effects γ0, γ1, and γ2. LMS produced slightly better coverage rates to the reference level 95% than the PI methods. Across conditions, the variability of the coverage rates of γ1 by GAPI and UPI were noticeable. When N = 5,000, missing data rate = 25%, sr2 = 0.1, their minimum coverage rates were 15.9% and 16.3%, respectively.

Summary and discussion

The simulation results have supported our hypotheses that when the indicators are MCAR or MAR, FIML for LMS produce unbiased γ0, γ1, γ2, γ3 estimates and valid statistical inferences across sample sizes, missing data rates, and interaction effect sizes. It is because FIML for LMS makes correct distributional assumption in the sample log-likelihood function. In the simulation, the indicators of ξ1 and ξ2 were multivariate normally distributed, which met the assumptions of LMS. Our hypotheses concerning PI have been supported as well. In the simulation, CPI generally produced unbiased γ0, γ1, γ2, γ3 estimates and valid statistical inferences when the indicators were MCAR. Compared to LMS, CPI produced γ3 estimates with larger biases and MSEs. GAPI and UPI could produce unbiased γ3 estimates, especially when N ≥ 500. The results of GAPI and UPI were acceptable since they also require N ≥ 500 to produce unbiased estimates in the complete data condition (Cham et al., 2012). Therefore, the results have confirmed our hypothesis that PI produces unbiased estimates when the indicators are MCAR. When the indicators were MAR, GAPI and UPI overestimated γ3. CPI produced less overestimated γ3 and sometimes even unbiased estimates. These results have supported our hypothesis that PI produces biased estimates when the indicators are MAR. FIML for PI incorrectly assumes that ξ1ξ2 and the product indicators are normally distributed. The differences between the results by the CPI, GAPI, and UPI are due to the imposed model constraints in Table 1 that are related to the distribution of ξ1ξ2. CPI has the most model constraints, followed by GAPI and UPI. The results suggest that when more of these model constraints are imposed, the biases of parameter estimates reduce. However, compared to LMS, CPI produced γ3 estimates with larger biases and MSEs. When sample size increased, MSEs of CPI became much larger than those of LMS. The results imply that CPI is much less efficient than LMS. Thus, LMS should be preferred. The convergence rates by GAPI are surprisingly low, which may imply the dependencies between the model constraints and missing values. Given the simulation study results and the complexity to specify the model constraints in PI methods in SEM software, we recommend FIML for LMS when the indicators are MCAR or MAR and when they are multivariate normally distributed.

Substantive example

We applied FIML estimation for CPI, GAPI, UPI, and LMS to a substantive example with incompletely observed indicators. Data were drawn from a study of 266 advanced cancer patients (mean age = 59.1 years, SD = 11.6) who participated in a randomized psychotherapy study (Breitbart et al., 2015). Patients were assessed at three time points: pretreatment (T1), posttreatment (T2), and two-month follow-up (T3). We adapted the results by McClain, Rosenfeld, and Breitbart (2003) to investigate the interaction effect between patients’ depression at T1 (ξ1) and the change of their inner peace well-being between T1 and T2 (ξ2) on their hopelessness at T3 (η). Patients’ depression at T1 was self-reported by the 21-item Beck Depression Inventory (0 to 3 points; α = .88; Beck, Ward, Mendelson, Mock, & Erbaugh, 1961). Inner peace well-being (hereafter referred to as “Peace”) was self-reported by the four-item Peace subscale of the Functional Assessment of Chronic Illness Therapy Spiritual Well-being Scale (0 = not at all to 4 = very much; Brady, Peterman, Fitchett, Mo, & Cella, 1999). Change scores of the four items (T2 minus T1) had lower Cronbach’s α (.65; Edwards, 2001). Hopelessness at T3 was self-reported by the 20-item Beck Hopelessness Scale (0 = False, 1 = True; α = .92; Beck, Weissman, Lesster, & Trexler, 1974). We used the average scores of the depression items and the sum of the hopelessness items in the latent variable interaction model. To account for unreliability of the change scores of Peace, we used a factor model for the four change scores.

Table 7 shows the missing data patterns of the variables; 129 patients (48.5%)were completely observed on all variables, 84 patients (31.6%) were missing on all four change scores of Peace and hopelessness at T3, and 44 patients (16.5%) were missing on hopelessness at T3. Little’s (1988) MCAR test showed that the variables’ means did not vary significantly between missing data patterns, χ2(29) = 15.00, p = .99. Jamshidian and Jalal’s (2010) MCAR test showed that variables’ variances and covariances were not significantly different between missing data patterns, p = .21. However, there was a small to medium relationship between treatment assignment and the three major missing data patterns, χ2(2) = 5.36, p = .07, Cohen’s ω = 0.14. Given this relationship and the treatment effect on patients’ distress (including hopelessness), we set treatment assignment as an auxiliary variable to account for missingness by letting it freely correlate with all variables in the model (Enders, 2010).

Table 7.

Missing data patterns of substantive example.

Missing data pattern Depression (T1) Change of inner peace well-being (T2 – T1) Hopelessness (T3) n (%)
1 129 (48.5%)
2 × × 84 (31.6%)
3 × 44 (16.5%)
Other patterns 9 (3.4%)

Our simulation has suggested that FIML for CPI and LMS produce unbiased parameter estimates when N ≥ 200. Therefore, only these methods were used in this analysis. In CPI, we centered depression and all four Peace change scores at their EM estimated means. We chose the change score with the highest reliability to form the product indicator with depression. LMS permits that ξ1 is a single variable and ξ2 is a factor. We also centered depression (ξ1) at its FIML estimated mean and set the mean of the change in Peace factor to zero. Table 8 shows the results by CPI and LMS. The intercept and the effect of the completely observed depression at T1 had similar estimates and standard errors between CPI and LMS (differences < 5%). The effect of the change in Peace (T2 – T1) and the interaction effect had different estimates (differences ~ 22%) and standard errors (differences ~ 8%). Several reasons can explain the differences between the two methods. First, missing data rates for the four Peace change scores were higher than the conditions in simulation study. In the simulation, when the missing data rate increased, the biases of the parameter estimates by CPI were larger than those by LMS. Second, depression at T1 (ξ1) had skewness = 1.22 and kurtosis = 2.89. The four Peace change scores (ξ2) had average skewness = 0.13 and kurtosis = 0.85, which exceeded the suggested values for CPI and LMS (skewness = 0, kurtosis < 1; Cham et al., 2012). When the indicators are nonnormal, CPI and LMS can produce biased parameter estimates, though CPI produces less biased estimates than LMS (Cham et al., 2012). Third, the outcome variable (hopelessness) at T3 had missing values. In this situation, LMS shall produce unbiased parameter estimates and valid inferences while CPI may produce biased parameter estimates and standard errors. We discuss the second and third reasons in more details in the General Discussion section.

Table 8.

Illustrative example results.

Regression coefficient CPI LMS


Est. SE Est. SE
Intercept 5.74 0.41** 5.79 0.41**
Depression (T1) 5.47 1.03** 5.74 1.04**
Change of peace well-being (T2 – T1) −0.62 0.46 −0.82 0.50
Interaction effect −1.14 0.65 −1.41 0.71*

Note. CPI = constrained product indicator; LMS = latent moderated structural equations; Est. = parameter estimate; SE = estimated standard error.

*

p < .05;

**

p < .01.

In summary, Figure 4 displays the interaction effect based on the results of LMS. Patients’ depression at T1 was positively related to their hopelessness at T3. For the patients with larger increase in their inner peace well-being from T1 to T2, this positive relationship was reduced.

Figure 4.

Figure 4

Interaction between depression (T1) and change of inner peace well-being (T2 – T1) on hopelessness (T3).

General discussion

In this manuscript, we investigated FIML estimation for four methods for estimation of latent variable interactions (CPI, GAPI, UPI, LMS) to handle incompletely observed indicators of ξ1 and/or ξ2. We conducted a simulation to investigate whether FIML for these methods produces unbiased parameter estimates and valid statistical inferences, and illustrated these methods using a substantive example. As discussed previously, we recommend FIML for LMS when the indicators are MCAR or MAR and are multivariate normally distributed. Our simulation did not consider the situation that the indicators of ξ1 and ξ2 are nonnormal (Cham et al., 2012; Coenders et al., 2008; Klein & Moosbrugger, 2000; Marsh et al., 2004; Wall & Amemiya, 2001). When the indicators of ξ1 and ξ2 are completely observed, the simulation study by Cham et al. (2012) has shown that FIML for LMS produces unbiased parameter estimates when the indicators have skewness = 0 and kurtosis < 1. FIML for GAPI and UPI produces unbiased estimates when the indicators have skewness < 2 and kurtosis < 6. When indicators are MCAR or MAR, FIML for CPI, GAPI, UPI, and LMS will not perform better than with complete data; FIML for LMS will produce biased estimates when MCAR or MAR indicators have skewness ≥ 2 and kurtosis ≥ 6.

Methods are needed to properly handle incompletely observed indicators of ξ1 and/or ξ2 when they are nonnormally distributed. There are several potential candidates. Kelava, Nagengast, and Brandt (2014) proposed incorporating mixture modeling with LMS to handle nonnormally distributed indicators in the complete data condition. This method also uses FIML for model estimation. Second, promising imputation procedures for interaction effects are being developed, and these procedures need to be investigated for possible use with latent variable interaction models. Doove et al. (2014) developed multiple imputation procedures using recursive partitioning, which estimates arbitrary nonlinear relationships among variables. Bartlett and Morris (2015), Bartlett et al. (2015), and Carpenter and Kenward (2013, Chapter 7) developed the multiple imputation procedures that account for the interaction effects that are being tested in the analyzed model when imputing missing values.

On the basis of our results that FIML for LMS leads to unbiased parameter estimates and valid statistics inferences for latent variable interactions, one may consider whether FIML for LMS works for multiple regression with interaction effects. As shown in the substantive example, the model is identified when ξ1 is a single variable and ξ2 is a factor. The specification is not identified when both ξ1 and ξ2 are single variables.

We did not consider all possible situations in our simulations. One interesting situation is that η or indicator of η is MCAR or MAR, which occurs in our substantive example. This situation is also not considered in the literature for multiple regression with interaction effects. In this situation, FIML for LMS still produces unbiased parameter estimates, standard error estimates, and valid statistical inferences, because it correctly handles the distribution of η. FIML for PI incorrectly assumes that η is normally distributed. It should be noted, however, that η is linearly related to ξ1, ξ2, and ξ1ξ2 in coefficients. In such case, violating the normality assumption of η in FIML has minimal effect on parameter estimates but can lead to biased standard error estimates and invalid statistical inferences (Yuan, Bentler, & Zhang, 2005). Future research is needed to examine how FIML for PI performs when η is MCAR or MAR. Another situation is when the indicators of ξ1 and/or ξ2 are missing not at random (MNAR). In general, FIML produces biased parameter estimates and invalid statistical inferences when variables are MNAR. Therefore, FIML for PI and LMS also produce biased parameter estimates when the indicators are MNAR. Future research is needed to investigate the applicability and performance of current MNAR methods (e.g., selection model and pattern mixture model) to latent variable interaction methods.

Acknowledgments

Funding: This work was supported by Grant CA128187 from the National Cancer Institute.

The authors thank David Budescu, Se-Kang Kim, anonymous reviewers, the associate editor, and the editor for their comments on prior versions of this manuscript. The ideas and opinions expressed here in are those of the authors alone, and endorsement by the authors’ institutions or the National Cancer Institute is not intended and should not be inferred.

Appendix

Detailed settings of full information maximum likelihood (FIML) estimation for product indicator (PI) and latent moderated structural equations (LMS) methods.

Types of settings Setting Default
(A) PI (including CPI, GAPI, and UPI)
 1. Maximum number of iterations for quasi-Newton algorithm 10,000 1,000
 2. Maximum number of steepest descent iterations for quasi-Newton algorithm 100 20
 3. Convergence criterion for quasi-Newton algorithm 0.000001 0.00005
(B) LMS
 1. Numeric integration algorithm Hermite-Gaussian Quadrature Rectangular
 2. Number of integration points per dimension 16 15
 3. Maximum number of iterations for quasi-Newton algorithm 10,000 1,000
 4. Maximum number of iterations for EM algorithm 2,000 500
 5. Convergence criterion for quasi-Newton algorithm 0.000001 0.000001
 6. Absolute observed-data log-likelihood change convergence criterion for EM algorithm 0.0000001 0.001
 7. Relative observed-data log-likelihood change convergence criterion for EM algorithm 0.0000001 0.000001
 8. Observed-data log-likelihood derivative convergence criterion for EM algorithm 0.000001 0.001

Note. CPI = constrained product indicator; GAPI = generalized appended product indicator; UPI = unconstrained product indicator.

Footnotes

1

In standard multiple imputation applications, we use noninformative prior distributions of parameters (e.g., Bartlett et al., 2015). As an alternative method to update the estimates of βs and τ s, we can use the random draws from the mean vector and covariance matrix from the conditional distribution of ξ1 and ξ1ξ2 (Equation [2]). Then, we transform these estimates into βs and τs.

2

Data augmentation algorithm iteratively cycles between imputation step (I-step) and permutation step (P-step). I-step is the first step and P-step is the second step described in the main text. In the multivariate imputation via chained equations algorithm, the missing values of ξ1 and ξ1ξ2 are imputed sequentially, one variable at a time. For example, ξ1 is imputed before ξ1ξ2. To impute ξ1, first, we assume the conditional distribution of ξ1 on the observed values of η and ξ2 and the imputed values of ξ1ξ2 are normal. Second, the parameters of this conditional distribution are estimated by random draws from their Bayesian estimated posterior distributions, which are based on the observed values of η and ξ2, imputed values of ξ1ξ2, and the specified prior distributions of parameters. Third, the imputed values of ξ1 are updated using the parameter estimates, observed values of η and ξ2, and imputed values of ξ1ξ2. To impute ξ1ξ2, we assume the conditional distribution of ξ1ξ2 on the observed values of η and ξ2 and the imputed values of ξ1 is normal. We follow the same steps above to impute ξ1ξ2. The distributional assumptions of ξ1 and ξ1ξ2 in this algorithm are identical to the distributional assumption in equation 2. Therefore, data augmentation algorithm and multivariate imputation via chained equations algorithm are the same in JAV approach (Van Burren, 2012, p. 116).

3

In addition to the suggested method, Foldnes and Hagtvet (2014) suggest a post-hoc analysis to examine how γ0, γ1, γ2, γ3 estimates are affected by choosing different product indicators of ξ1 ξ2.

4

The full report of the results in each MCAR and MAR scenarios is available upon request.

Supplemental data material for this article can be accessed on the publisher’s website

Conflict of Interest Disclosures: Each author signed a form for disclosure of potential conflicts of interest. No authors reported any financial or other conflicts of interest in relation to the work described.

Ethical Principles: The authors affirm having followed professional ethical guidelines in preparing this work. These guidelines include obtaining informed consent from human participants, maintaining ethical treatment and respect for the rights of human or animal participants, and ensuring the privacy of participants and their data, such as ensuring that individual participants cannot be identified in reported results or from publicly available original or archival data.

Role of the Funders/Sponsors: None of the funders or sponsors of this research had any role in the design and conduct of the study; collection, management, analysis, and interpretation of data; preparation, review, or approval of the manuscript; or decision to submit the manuscript for publication.

References

  1. Algina J, Moulder BC. A note on estimating the Jöreskog-Yang model for latent variable interaction using LISREL 8.3. Structural Equation Modeling. 2001;8:40–52. doi: 10.1207/S15328007SEM0801_3. [DOI] [Google Scholar]
  2. Aroian LA. The probability function of the product of two normally distributed variables. Annals of Mathematical Statistics. 1944;18:265–271. doi: 10.1214/aoms/1177730442. [DOI] [Google Scholar]
  3. Azar B. Finding a solution for missing data. Monitor on Psychology. 2002;33(7):70. Retrieved from http://www.apa.org/monitor/julaug02/missingdata.aspx. [Google Scholar]
  4. Bartlett JW, Morris TP. Multiple imputation of covariates by substantive-model compatible fully conditional specification. The Stata Journal. 2015;15:437–456. Retrieved from http://www.stata-journal.com/article.html?article=st0387. [Google Scholar]
  5. Bartlett JW, Seaman SR, White IR, Carpenter JR. Multiple imputation of covariates by fully conditional specification: Accommodating the substantive model. Statistical Methods in Medical Research. 2015;24:462–487. doi: 10.1177/0962280214521348. [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Beck AT, Ward CH, Mendelson M, Mock J, Erbaugh J. An inventory for measuring depression. Archives of General Psychiatry. 1961;4:561–571. doi: 10.1001/archpsyc.1961.01710120031004. [DOI] [PubMed] [Google Scholar]
  7. Beck AT, Weissman A, Lester D, Trexler L. The measurement of pessimism: The Hopelessness Scale. Journal of Consulting and Clinical Psychology. 1974;42(6):861–865. doi: 10.1037/h0037562. [DOI] [PubMed] [Google Scholar]
  8. Bohrnstedt GW, Goldberger AS. On the exact covariance of products of random variables. Journal of the American Statistical Association. 1969;64:1439–1442. doi: 10.1080/01621459.1969.10501069. [DOI] [Google Scholar]
  9. Brady MJ, Peterman AH, Fitchett G, Mo M, Cella D. A case for including spirituality in quality of life measurement in oncology. Psycho-Oncology. 1999;8:417–428. doi: 10.1002/(SICI)1099-1611(199909/10)8:5&#x0003c;417::AIDPON398&#x0003e;3.0.CO;2-4. [DOI] [PubMed] [Google Scholar]
  10. Breitbart W, Rosenfeld B, Pessin H, Applebaum A, Kulikowski J, Lichtenthal WG. Meaning-centered group psychotherapy: An effective intervention for improving psychological well-being in patients with advanced cancer. Journal of Clinical Oncology. 2015;33:749–754. doi: 10.1200/JCO.2014.57.2198. [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Carpenter JR, Kenward MG. Multiple imputation and its application. West Sussex, United Kingdom: John Wiley & Sons; 2013. [Google Scholar]
  12. Cham H, West SG, Ma Y, Aiken LS. Estimating latent variable interactions with nonnormal observed data: A comparison of four approaches. Multivariate Behavioral Research. 2012;47:840–876. doi: 10.1080/00273171.2012.732901. [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Coenders G, Batista-Foguet JM, Saris WE. Simple, efficient and distribution-free approach to interaction effects in complex structural equation models. Quality and Quantity. 2008;42:369–396. doi: 10.1007/s11135-006-9050-6. [DOI] [Google Scholar]
  14. Cohen J. A power primer. Psychological Bulletin. 1992;112:155–159. doi: 10.1037/0033-2909.112.1.155. [DOI] [PubMed] [Google Scholar]
  15. Collins LM, Schafer JL, Kam CM. A comparison of inclusive and restrictive strategies in modern missing data procedures. Psychological Methods. 2001;6:330–351. doi: 10.1037/1082-989X.6.4.330. [DOI] [PubMed] [Google Scholar]
  16. Doove LL, Van Buuren S, Dusseldorp E. Recursive partitioning for missing data imputation in the presence of interaction effects. Computational Statistics & Data Analysis. 2014;72:92–104. doi: 10.1016/j.csda.2013.10.025. [DOI] [Google Scholar]
  17. Edwards JR. Ten difference score myths. Organizational Research Methods. 2001;4:265–287. doi: 10.1177/109442810143005. [DOI] [Google Scholar]
  18. Enders CK. Applied missing data analysis. New York, NY: Guilford; 2010. [Google Scholar]
  19. Enders CK, Baraldi AN, Cham H. Estimating interaction effects with incomplete predictor variables. Psychological Methods. 2014;19:39–55. doi: 10.1037/a0035314. [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Foldnes N, Hagtvet KA. The choice of product indicators in latent variable interaction models: Post hoc analyses. Psychological Methods. 2014;19:444–457. doi: 10.1037/a0035728. [DOI] [PubMed] [Google Scholar]
  21. Hoogland JJ, Boomsma A. Robustness studies in covariance structure modeling: An overview and a metaanalysis. Sociological Methods & Research. 1998;26:329–367. doi: 10.1177/0049124198026003003. [DOI] [Google Scholar]
  22. Jaccard J, Wan CK. Measurement error in the analysis of interaction effects between continuous predictors using multiple regression: Multiple indicator and structural equation approaches. Psychological Bulletin. 1995;117:348–357. doi: 10.1037/0033-2909.117.2.348. [DOI] [Google Scholar]
  23. Jackman MGA, Leite WL, Cochrane DJ. Estimating latent variable interactions with the unconstrained approach: A comparison of methods to form product indicators for large, unequal numbers of items. Structural Equation Modeling. 2011;18:274–288. doi: 10.1080/10705511.2011.557342. [DOI] [Google Scholar]
  24. Jamshidian M, Jalal S. Tests of homoscedasticity, normality, and missing completely at random for incomplete multivariate data. Psychometrika. 2010;75:649–674. doi: 10.1007/s11336-010-9175-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  25. Jöreskog KG, Sörbom D. LISREL 9.2 for Windows [Computer software] Skokie, IL: Scientific Software International; 2015. [Google Scholar]
  26. Jöreskog KG, Yang F. Nonlinear structural equation models: The Kenny-Judd model with interaction effects. In: Schumacker E, Marcoulides GA, editors. Advanced structural equation modeling: Issues and techniques. Mahwah, NJ: Erlbaum; 1996. pp. 57–88. [Google Scholar]
  27. Kelava A, Nagengast B, Brandt H. A nonlinear structural equation mixture modeling approach for nonnormally distributed latent predictor variables. Structural Equation Modeling. 2014;21:468–481. doi: 10.1080/10705511.2014.915379. [DOI] [Google Scholar]
  28. Kelava A, Werner C, Schermelleh-Engel K, Moosbrugger H, Zapf D, Ma Y, … West SG. Advanced nonlinear structural equation modeling: Theoretical properties and empirical application of the LMS and QML estimators. Structural Equation Modeling. 2011;18:465–491. doi: 10.1080/10705511.2011.582408. [DOI] [Google Scholar]
  29. Kenny DA, Judd CM. Estimating the nonlinear and interactive effects of latent variables. Psychological Bulletin. 1984;96:201–210. doi: 10.1037/0033-2909.96.1.201. [DOI] [Google Scholar]
  30. Kim S, Sugar CA, Belin TR. Evaluating model based imputation methods for missing covariates in regression models with interactions. Statistics in Medicine. 2015;34:1876–1888. doi: 10.1002/sim.6435. [DOI] [PMC free article] [PubMed] [Google Scholar]
  31. Klein AG, Moosbrugger H. Maximum likelihood estimation of latent interaction effects with the LMS method. Psychometrika. 2000;65:457–474. doi: 10.1007/BF02296338. [DOI] [Google Scholar]
  32. Little RJA. A test of missing completely at random for multivariate data with missing values. Journal of the American Statistical Association. 1988;83:1198–1202. doi: 10.1080/01621459.1988.10478722. [DOI] [Google Scholar]
  33. Little RJA. Regression with missing X’s: A review. Journal of the American Statistical Association. 1992;87(420):1227–1237. doi: 10.1080/01621459.1992.10476282. [DOI] [Google Scholar]
  34. Little RJA, Rubin DB. Statistical analysis with missing data. 2. Hoboken, NJ: JohnWiley & Sons; 2002. [Google Scholar]
  35. Marsh HW, Wen Z, Hau KT. Structural equation models of latent interactions: Evaluation of alternative estimation strategies and indicator construction. Psychological Methods. 2004;9:275–300. doi: 10.1037/1082-989X.9.3.275. [DOI] [PubMed] [Google Scholar]
  36. McClain CS, Rosenfeld B, Breitbart W. Effect of spiritual well-being on end-of-life despair in terminally-ill cancer patients. The Lancet. 2003;361:1603–1607. doi: 10.1016/S0140-6736(03)13310-7. [DOI] [PubMed] [Google Scholar]
  37. Moosbrugger H, Schermelleh-Engel K, Klein A. Methodological problems of estimating latent interaction effects. Methods of Psychological Research Online. 1997;2:95–111. [Google Scholar]
  38. Muthén LK, Muthén BO. Mplus User’s Guide. 7. Los Angeles, CA: Muthén & Muthén; 1998–2015. [Google Scholar]
  39. Rosseel Y. lavaan: An R package for structural equation modeling. Journal of Statistical Software. 2012;48(2):1–36. doi: 10.18637/jss.v048.i02. [DOI] [Google Scholar]
  40. Savalei V. Small sample statistics for incomplete non-normal data: Extensions of complete data formulae and a Monte Carlo comparison. Structural Equation Modeling. 2010;17:241–264. doi: 10.1080/10705511003659375. [DOI] [Google Scholar]
  41. Schermelleh-Engel K, Klein AG, Moosbrugger H. Estimating nonlinear effects using a latent moderated structural equations approach. In: Schumacker E, Marcoulides GA, editors. Interaction and nonlinear effects in structural equation modeling. Mahwah, NJ: Erlbaum; 1998. pp. 203–238. [Google Scholar]
  42. Seaman SR, Bartlett JW, White IR. Multiple imputation of missing covariates with non-linear effects and interactions: An evaluation of statistical methods. BMC Medical Research Methodology. 2012;12(46):1–13. doi: 10.1186/1471-2288-12-46. [DOI] [PMC free article] [PubMed] [Google Scholar]
  43. Umbach N, Naumann K, Hoppe D, Brandt H, Kelava A, Schmitz B. nlsem (Version 0.6) [Computer software] 2015 Retrieved from https://cran.r-project.org/web/packages/nlsem/index.html.
  44. von Hippel PT. How to impute interactions, squares, and other transformed variables. Sociological Methodology. 2009;39:265–291. doi: 10.1111/j.1467-9531.2009.01215.x. [DOI] [Google Scholar]
  45. Van Buuren S. Flexible imputation of missing data. Boca Raton, FL: CRC Press; 2012. [Google Scholar]
  46. Wall MM, Amemiya Y. Generalized appended product indicator procedure for nonlinear structural equation analysis. Journal of Educational and Behavioral Statistics. 2001;26:1–29. doi: 10.3102/10769986026001001. [DOI] [Google Scholar]
  47. Wu Y, Wen Z, Marsh HW, Hau KT. A comparison of strategies for forming product indicators for unequal numbers of items in structural equation models of latent interactions. Structural Equation Modeling. 2013;20:551–567. doi: 10.1080/10705511.2013.824772. [DOI] [Google Scholar]
  48. Yuan KH. Normal distribution based pseudo ML for missing data: With applications to mean and covariance structure analysis. Journal of Multivariate Analysis. 2009;100:1900–1918. doi: 10.1016/j.jmva.2009.05.001. [DOI] [Google Scholar]
  49. Yuan KH, Bentler PM, Zhang W. The effect of skewness and kurtosis on mean and covariance structure analysis: The univariate case and its multivariate implication. Sociological Methods & Research. 2005;34:240–258. doi: 10.1177/0049124105280200. [DOI] [Google Scholar]
  50. Yuan KH, Savalei V. Consistency, bias and efficiency of the normal-distribution-based MLE: The role of auxiliary variables. Journal of Multivariate Analysis. 2014;124:353–370. doi: 10.1016/j.jmva.2013.11.006. [DOI] [Google Scholar]
  51. Zhang W. Doctoral dissertation. University of Notre Dame; Notre Dame, IN: 2010. Estimating latent variable interactions with missing data. [Google Scholar]

RESOURCES