Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2015 Jun 1.
Published in final edited form as: Long Range Plann. 2014 Jun 1;47(3):138–145. doi: 10.1016/j.lrp.2014.02.005

On Components, Latent Variables, PLS and Simple Methods: Reactions to Rigdon's Rethinking of PLS

Peter M Bentler, Wenjing Huang
PMCID: PMC4048869  NIHMSID: NIHMS576128  PMID: 24926106

Abstract

Rigdon (2012) suggests that partial least squares (PLS) can be improved by killing it, that is, by making it into a different methodology based on components. We provide some history on problems with component-type methods and develop some implications of Rigdon's suggestion. It seems more appropriate to maintain and improve PLS as far as possible, but also to freely utilize alternative models and methods when those are more relevant in certain data analytic situations. Huang's (2013) new consistent and efficient PLSe2 methodology is suggested as a candidate for an improved PLS.


An important feature of science is that theories, methodologies, conclusions etc. are regularly reevaluated, unworkable or untestable ideas rejected, and new ideas and approaches proposed and perhaps provisionally accepted. Professor Rigdon (2012) has proposed re-evaluating the field of partial least squares (PLS) methodology. He thoroughly and coherently reviews many inadequacies of factor analysis based structural equation models (SEM), i.e., latent variable models (LVM), and proposes to eliminate LVM as a viable research strategy. In view of the historical commonalities between PLS and LVM, a rejection of LVM would require also rejecting PLS. To avoid this logical difficulty, Rigdon proposes morphing PLS into a composite variable model (CVM) based on compounds of measured variables. However, this creates a difficulty for PLS since CVM has no need for PLS, e.g., Hwang, Ho, and Lee (2010, p. 228) state that “GSCA [their variant of CVM] is recommended as an alternative to partial least squares for general SEM purposes” (p. 228). While GSCA may not be the final word in CVM (e.g., Henseler, 2012), at a minimum, the relations of PLS to CVM and LVM require some further analysis.

We certainly agree with Rigdon that observed and composite variable methods have an important role in data analysis and statistics – indeed, CVM are probably far more prevalent overall than LVM1 -- and that development of new methodological and statistical tools for their use is a valuable goal. However, in contrast to Rigdon's proposal to eliminate PLS as an LVM, we think it is more important to elevate PLS as a desirable LVM by improving its mathematical and statistical basis. Before we explain how this might be accomplished, we discuss some features of CVM and LVM not emphasized by Rigdon. A thorough review of LVM is given in Hoyle (2012). Recent reviews of the use of PLS in management and marketing research are given by Hair, Sarstedt, Pieper, and Ringle (2012), Hair, Sarstedt, Ringle, and Mena (2012), Ringle et al. (2012), Rönkkö and Evermann (2013) and Henseler et al. (2014).

Errors of Measurement

There is a lot of virtue in classical test theory and its implications (e.g., Bentler, 2009). It can serve as a foundation for looking at the real world, and in turn, provide a viewpoint toward CVM and LVM. It seems obvious to many that any observed score is liable to be fallible, i.e., to contain unwanted noise or random error (plus various potential sources of bias not discussed here). Invariably a particular score value may seem a bit arbitrary for its intended use -- a student's score on an exam, a nurse's measure of blood pressure, an economist's statement on food inflation, a rating of teacher quality from student performance, and, as we recently saw in the US election, a preference given in a political survey – no doubt only represent approximations to their intended target's concept score. Of course the basic equation X = T + E is an abstraction that may not hold precisely in any particular situation2, but it provides the basis for a nice skepticism about any particular X, implies the potential usefulness of other Xs to reflect the same T (multiple indicators), and of course leads to further questions such as whether perhaps T = a1F1+...+akFk, that is, whether an LVM based on latent factors might be appropriate.3

If one does not want to reify any X, one can also be suspicious of a CVM, since a basic building block of any CVM is a linear compound of Xs. Any such linear combination will automatically inherit the Es when they exist. Rigdon's espousal of CVM ignores these errors. We suggest that it is important to remember them, i.e., to remind oneself that under classical test theory and certain assumptions, composites will have smaller error than the individual Xs, even though in practice we may never have enough Xs in the composite to make the errors completely disappear (Bentler, 1972, 2007; Li & Bentler, 2011). Furthermore, since a given composite may well be a predictor variable in another equation of a CVM, we are forced to face the old issue that errors in predictor variables will almost always lead to parameter estimates for their coefficients that are biased proportionally to the magnitude of this irrelevant error (e.g., Cochran, 1968; Fuller, 1987). A main point of LVM is to avoid such biases (e.g., Wansbeek & Meijer, 2000). It would not be bad for PLS also to avoid such biases.

Rigdon recommends the use of various types of components, including those associated with decades-old methods that seem to have died out. Other older component methods also come to mind, e.g., Bentler (1976) proposed a CVM related to a relaxed LVM and Bartholomew (1984, 1985) proposed to use particular components that are rationalized on the basis of their closeness to factors. However, it has been noted that in such cases it would be possible just to use the factors themselves (Bentler, 1985a). While we are impressed by the surprisingly flexibility and ability to do clever things with newer component methods such as Hwang, Ho, and Lee's (2010) methodology for interactions, nonetheless, all component methods are equivalent in not being able to avoid incorporating errors. Bentler and de Leeuw (2011) showed that any components, including principal components, are linear combinations of common factors and unique factors. The latter contain random error. Why would one be interested in a linear combination that contains noise when such noise can be avoided with LVM?4

Scores

A lot of hand-wringing has gone into the issue of factor score indeterminacy. As Rigdon noted, this is an old problem that arises not only in factor analysis but also in more general models. However, many years ago Bentler (1980) noted that factor indeterminacy does not hinder model comparison and evaluation: “..although an infinite set of LVs [latent variables] can be constructed under a given LV model to be consistent with given MVs [measured variables], the goodness-of-fit of the LV model to data (as indexed, for example by a χ2 statistic) will be identical under all possible choices of such LVs. Consequently, the process of evaluating the fit of a model to data and comparing the relative fit of competing models is not affected by LV determinacy. Thus, in my view, theory testing with LV models remains a viable research strategy in spite of factor indeterminacy” (p. 442). As far as we know, this proposition has not been disproven.

There is no doubt that individual scores for specific observations are sometimes needed. In such an application it is impossible to avoid relying on observed variables and composites. While the ideal nature of such a composite will depend on the application, LVM may still be useful. For example, a composite may be an estimate of a latent factor and used, with appropriate corrections, in factor score regression (Skrondal & Laake, 2001; Hoshino & Bentler, 2013), a predictive score based methodology like PLS. Or, it may be desirable to obtain a composite with maximal reliability or minimal noise (e.g., Bentler, 1968; Raykov, 2004).

Simplicity

Rigdon recommends methods and procedures that are simple. For example, he supports the use of equal or correlation weights in regression rather than standard least-squares estimated beta weights. We thoroughly agree that researchers should have simple models and methods available, but not necessarily because they are simple. We would rationalize any virtue of parameter simplicity using standard statistical terms such as bias, efficiency, and mean squared error, as well as reduced variance and lack of statistical degradation with fewer parameters. In the SEM context, Bentler and Mooijaart (1989) provide some details. In the regression context, Bentler and Woodward (1979) developed a methodology for regression on composites that allows statistical tests of whether simple weights carry all of the information in the prediction or whether differential weighting adds any significant increment. They showed that simple a priori weights can be improved upon: rather than accept either unit or correlation weights a priori, they developed a methodology to find and use optimal binary (+1, -1) or trinary (+1, 0, -1) weights. Trinary weights simplify by automatically dropping unimportant predictors. This overlooked methodology will be available in EQS 7 (Bentler, in preparation).

Parameters are parts of models, and model simplicity in LVM and CVM is, of course, dear to our heart. Rigdon did not propose a specific (simple or complex) CVM structure for use. The modeling complexity of LISREL (Jöreskog & Sörbom, 1996) with its many Greek parameter matrices led us to the far simpler Bentler-Weeks (1980) model and the EQS program (Bentler, 1985b) for non-matrix, and then, graphical modeling setups. It is odd to see SEM continue to be presented today in far more complex ways than is necessary (see e.g., Bentler 2010). Of course, too much simplicity can be harmful. Some program practices make it hard or impossible for a researcher to determine the exact model and statistical definitions of the estimators and tests given by the program output.

Model Testing

In his approach to CVM, Rigdon does not emphasize the need for model testing, perhaps because of the typical absence of a formal model testing methodology in CVM. For example, Hwang, Ho, and Lee (2010, p. 228) provide no χ2 test of model fit for their recent approach to interaction modeling,5 while in contrast an LVM approach to interaction can provide such a test (Mooijaart & Bentler, 2010). We do recognize that model testing is but one endpoint of a modeling research program, since, in addition to prediction as emphasized in PLS, model exploration and model building also can be valuable enterprises (e.g., Jöreskog, 1993). But in our view it is a critical endpoint, in principle allowing a clean evaluation of given models and a comparison of alternative models (Bentler & Satorra, 2010; Satorra & Bentler, 2010; Yuan & Bentler, 2007).

Rather than hopelessly accept Rigdon's conclusion that “...PLS path modeling lacks an inferential test statistic like factor-based SEM's χ2 statistic” (p. 9), we recommend instead to improve PLS so that it can provide such a test. In Huang's (2013) dissertation study, we build on Dijkstra's (2011) consistent PLS estimator to derive two consistent and efficient PLS estimators (PLSe1 and PLSe2) that asymptotically have minimum sampling variability.6 Our methodology also yields standard errors for parameter estimates and allows evaluation of models via χ2 goodness of fit tests. The dissertation work is recently completed and may still need further improvements. The results show that PLSe2 is as good as standard SEM in these regards. We briefly describe PLSe2 and the results of our simulation studies in the following section.

PLSe2: An Efficient Estimator with Tests for Partial Least Squares

PLS lacks a classical parametric inferential framework (Vinzi, Trinchera & Amato, 2010). One has to resort to empirical confidence intervals and hypothesis testing procedures based on resampling methods such as jackknife and bootstrap (Chin, 1998). It also suffers from undesirable statistical properties of the estimates, e.g. coefficients are known to be biased (Cassel, Hackl, & Westlund, 1999, 2000) and only consistent at large (Dijkstra,1983). Recently, Dijkstra (2011) proposed a new estimator PLSc (a PLS-consistent estimator to distinguish it from the traditional PLS estimators) which involves corrections to the PLS estimates to account for measurement error. He showed that his new PLSc estimator is very effective in reducing the bias in parameter estimates. That is, Dijkstra's (2011) consistent PLS estimator PLSc yields a consistent estimator Σ^c of the population correlation matrix Σ. If desired, this can be rescaled into a covariance matrix.

Based on Dijkstra's PLSc estimates, we investigated two efficient PLS estimation methodologies (PLSe1 and PLSe2). Here we present PLSe2 that utilizes Browne's (1974) generalized least squares (GLS) covariance structure estimation methodology to obtain an efficient estimator and the associated parameter and model evaluation methodology.

Assuming that a sample of size N is drawn from a multivariate normal population and S is the sample covariance matrix, Browne (1974, Propositions 1-3) proved that when a consistent estimator W2 of the population inverse covariance matrix Σ-1 is used and the normal theory GLS fit function:

FGLS(θ)=12tr{[SΣ(θ)]W2}2 (1)

is minimized with respect to Θ to yield the minimizing value GLS under the assumption of multivariate normality of variables, the test statistic (N – 1)GLS is asymptotically distributed as a χmq2 variate. Here q represents the number of free parameters in the model, m* = m(m + 1)/2 where m denotes the number of manifest variables (so m* denotes the number of nonredundant elements of S) and N is the sample size. This statistic can be used to test the validity of the hypothesized model. Browne (1974) showed that the estimator resulting from the minimization of FGLS(Θ) is consistent, asymptotically normally distributed, and asymptotically efficient in the Loewner sense. He showed that estimates of the variances of the estimator can be obtained from the diagonal of (σ.WN1σ.)1 evaluated at θ^, where σ. is the Jacobian matrix, WN=.5Dm(W2W2)Dm, and Dm is the duplication matrix.

In practical implementations of Browne's theory, W2 = S-1 (for GLS) and W2=Σ^ML1 (for normal theory GLS) are used, where the latter is based on an iteratively updated estimator Σ^ under the model that leads to the normal theory maximum likelihood (ML) estimator (Lee & Jennrich, 1979). Our PLSe2 methodology makes use of Browne's results by taking W2=Σ^c1, the consistent PLS-based weight matrix. This provides a new member for the class of normal theory GLS estimators. To implement PLSe2, one could use the existing framework of any SEM package that implements normal theory GLS and simply replace the default weight with the new weight obtained from Dijkstra's PLSc procedure.

When the assumption of multivariate normality is inappropriate, the test statistics and standard errors can be corrected using the variety of robust methods available in the field such as the Satorra-Bentler (1994) scaled and adjusted test statistics, and the appropriate sandwich estimator of asymptotic variances (e.g., Bentler & Dijkstra, 1985, eq. 1.4.5).

We demonstrate the performance of PLSe2 by comparing it with ML estimation as widely used in traditional SEM through a Monte Carlo simulation study. The data generating model resembles the one from Maruyama and McGarvey (1980). We took their sample correlation matrix and estimated a model as shown below. The parameter estimates are used to generate multivariate normally distributed data at sample sizes of 200, 500 and 1000.

The measurement model is: y = Λη + , specifically

(y1y2y3y4y5y6y7y8y9y10y11y12y13)=(λ1,1λ2,1λ3,1λ4,2λ5,2λ6,3λ7,3λ8,3λ9,4λ10,4λ11,5λ12,5λ13,5)(η1η2η3η4η5)+(ε1ε2ε3ε4ε5ε6ε7ε8ε9ε10ε11ε12ε13),
whereΛ=(.609.686.611.481.364.658.844.167.631.507.522.793.400).

The covariance matrix of is a diagonal matrix Θ whose diagonal elements are .51, .37, .50, .35, .63, .57, .29, .97, .60, .74, .73, .37 and .84.

The structural model is η = + ζ, and

(η1η2η3η4η5)=(0β1,200β1,5β2,10β2,3β2,40000000000000000)(η1η2η3η4η5)+(ζ1ζ2ζ3ζ4ζ5),

This model is non-recursive. The covariance matrix of ζ is a symmetric matrix ψ with 1s on the diagonal:

whereB=(0.21200.251.2950.2661.0270000000000000000).

where ψ3,4 = .379.

The means of PLSe2 estimates across 300 Monte Carlo replications are compared with those from traditional ML estimation. We also compare Root Mean Square Error (RMSE), which can provide information of both the deviation of each parameter estimate from the true value and the variability of such distances:

RMSE=[1Mi=1M(θ^iθ)2]12 , where M is the number of replications and Θ refers to a generic parameter with θ^i the estimate from Monte Carlo replication i.

Table 1 compares the parameter estimates of factor loadings, regression coefficients, and the residual factor correlation using ML and PLSe2. We can see that at each sample size, ML and PLSe2 produce almost identical parameter estimates and RMSE. Overall for both methods, bias reduces as sample size increases. At N=200, both methods recovers parameters quite well with PLSe2 showing slightly higher RMSE than ML. Such difference goes away at N=1000 when the two estimators become virtually identical and unbiased.

Table 1.

Comparison of the mean of parameter estimates and RMSE across 300 replications using ML and PLSe2.

N=200 N=500 N=1000

true value ML PLSe2 ML PLSe2 ML PLSe2
mean RMSE mean RMSE mean RMSE mean RMSE mean RMSE mean RMSE
λ1,1 =.609 .606 .070 .598 .074 .612 .043 .609 .044 .611 .029 .610 .029
λ2,1 =.686 .675 .073 .667 .076 .690 .044 .687 .045 .686 .030 .685 .030
λ3,1 =.611 .607 .065 .600 .069 .614 .041 .611 .041 .613 .031 .612 .031
λ4,2 =.481 .494 .102 .490 .103 .464 .077 .461 .078 .474 .053 .472 .054
λ5,2 =.364 .366 .076 .360 .077 .351 .055 .348 .056 .359 .036 .357 .036
λ6,3 =.658 .667 .095 .655 .099 .664 .061 .659 .062 .658 .044 .656 .044
λ7,3 =.844 .827 .094 .815 .099 .842 .067 .836 .068 .842 .048 .839 .048
λ8,3 =.167 .180 .073 .173 .075 .170 .052 .168 .051 .166 .037 .164 .037
λ9,4 =.631 .638 .103 .624 .107 .627 .057 .621 .058 .632 .046 .629 .046
λ10,4 =.507 .520 .095 .508 .096 .505 .054 .499 .056 .502 .041 .499 .041
λ11,5 =.522 .536 .093 .522 .092 .526 .063 .522 .063 .522 .047 .520 .047
λ12,5 =.793 .789 .111 .776 .118 .790 .074 .783 .076 .799 .060 .796 .060
λ13,5 =.400 .411 .086 .398 .084 .400 .056 .395 .057 .402 .039 .399 .038

β1,2 =.212 .227 .117 .225 .123 .203 .077 .201 .078 .208 .054 .206 .055
β1,5 =.251 .286 .102 .287 .109 .255 .069 .252 .072 .250 .047 .248 .049
β2,1 =.295 .284 .215 .285 .215 .324 .170 .327 .172 .309 .107 .309 .109
β2,3 =.266 .264 .208 .245 .219 .268 .121 .261 .122 .276 .089 .274 .089
β2,4 =1.027 1.044 .342 1.032 .354 1.123 .297 1.127 .302 1.063 .182 1.063 .184

ψ4,3 =.379 .386 .110 .386 .116 .381 .067 .381 .068 .379 .047 .378 .048

It is worth noting that the estimates of the regression coefficients are generally more biased than the estimates of the loadings or the factor correlation. This happens to both ML and PLSe2 and it is probably due to the complexity of the non-recursive model we used. For example, at N=200, RMSE of the loading estimates are generally less than 0.1 whereas RMSE of the regression coefficients estimates are all larger than 0.1. We also observe a more pronounced bias in β4,5 at N=200 where both ML and PLSe2 overestimate this parameter by 14%. However, when sample size goes to 1000, such differences are diminished.

ML and PLSe2 both produce standard error estimates and minimum fit function chi-square statistics. Although standard error estimates are not tabulated, they are almost identical and are well calibrated, which means that the mean of the estimated standard errors are very close to the standard deviation of the empirical sampling distribution across 300 replications.

The advantage of using ML or normal theory GLS estimation is that they both allow for formal model fit test. Since the model we estimate is correctly specified, the expected value for the chi-square statistic is equal to the model’s degree of freedom (df = m* – q =59) and the expected variance of this test statistic is 2df = 118. At N=200, the ML test statistic has a mean of 60.54 and variance of 115.27, which are very close to the expected values. The PLSe2 test statistic has a mean of 58.44 and a slightly elevated variance of 142.02. As sample size goes to 1000, both ML and PLSe2 produce test statistics with empirical distribution closely following the theoretical expectation (mean of 58.98 and variance of 128.29 for ML; mean of 58.39 and variance of 127.05 for PLSe2). With a correctly specified model, the expected model rejection rate at α = .05 should be controlled at around 5%. Given 300 replications, the 95% confidence interval for the observed percentage of rejected model is between 2.5% and 7.5%. For both ML and PLSe2, the observed model rejection rates at all sample sizes are between 4% and 6%, well within the 95% confidence interval.

Additional Complications

New proposals and developments are often followed by unanticipated consequences, and it should be noted that Rigdon's rejection of LVM has further logical consequences that would restrict data analysis unnecessarily. If one rejects latent variables in a basic SEM context such as PLS, to be consistent one also should reject latent variables in other contexts. Illustrative simple extensions of SEM that would have to be rejected include, for example, multiple group models for study of measurement invariance (e.g., Millsap, 2011), growth curve models in which mean structures are added to covariance structures (e.g., McArdle, 2012), mixture models (e.g., Yuan & Bentler, 2010), growth mixture models (e.g., Shiyko, Ram, & Grimm, 2012), and multilevel models for within-level and between-level effects (e.g., Bentler et al., 2011). In addition, models that do not appear to be factor-based SEM contain random variables that are actually latent variables and hence would have to be rejected. These include classical statistical models such as random coefficient models, mixed models with fixed effects and random effects, and hierarchical linear models. See e.g., An, Yang, and Bentler (2013) for a modern example, de Leeuw and Meijer (2008) for a review, and Bentler and Liang (2008) on a unified approach to mixed models and multilevel models. As noted by Muthén (2003), “..the idea of latent variables captures a wide variety of statistical concepts, including random effects, missing data, sources of variation in hierarchical data, finite mixtures, latent classes, and clusters” (p. 81). In addition to specific model types such as latent class models, a rejection of LVM would include ignoring potentially useful combinations of model types such as multilevel latent class analysis (Henry & Muthén, 2010).

Furthermore, we note that Rigdon is especially interested in developing a CVM with a currently unknown “complete and consistent approach to measurement which is factor-free” (p. 33). We are actually sympathetic to such a goal, since we have ongoing research on formative measurement with binary responses that extends Guttman scaling methodology (Bentler, 2011). However, Rigdon proposes to give up LVM measurement, which also would require rejecting item response theory (IRT), today's foundational methodology for educational and psychological measurement (e.g., Embretson & Reise, 2000; van der Linden & Hambleton, 2010). IRT is, of course, applicable to scale development and measurement in a wide variety of other fields. Whether in classical unidimensional or multidimensional formats, and various response functions, these methods describe the probabilities of endorsement of binary or ordinal items or response categories in terms of parameters as well as latent variable trait scores. This is an interesting and active field, especially in the multidimensional context (e.g., An & Bentler, 2011, 2012; Cai, 2010; Cai, Yang, & Hansen, 2011; Reckase, 2009; J. Wu & Bentler, 2012, 2013). Rather than reject IRT because it is an LVM, we feel that IRT, like the various methods previously mentioned, should be utilized when it can serve a useful purpose. EQSIRT was recently completed as a companion to EQS to facilitate IRT analyses (E. Wu & Bentler, 2013).

Conclusions

A variety of fields, including the management sciences, have to evaluate theories in the context of ill-measured constructs, error-contaminated variables, extraneous influences, and other challenges to accurate scientific inferences. We do not agree with Rigdon that the best approach to dealing with these challenges is to avoid latent variable models. We urge a realistic recognition of their virtues and limitations. A major historical problem with PLS estimators has been their inconsistency. We propose to use Dijkstra's (2011) correction to PLS so that the new PLSc estimator is consistent, and then propose an efficient estimator PLSe2 that is based on PLSc so that we can obtain standard error estimates and evaluate models via χ2 goodness of fit tests. The results of our simulations show that PLSe2 is as good as standard ML estimation in SEM, thus providing one avenue for the resurrection of PLS as a fully justified statistical methodology that will be available in EQS 7 (Bentler, in preparation).

Acknowledgement

This research was supported by grants 5K05DA000017-35 and 5P01DA001070-38 from the National Institute on Drug Abuse to P. M. Bentler, who acknowledges a financial interest in EQS and its distributor, Multivariate Software. Requests for reprints should be sent to Peter M. Bentler, Departments of Psychology and Statistics, UCLA, Box 951563, Los Angeles, CA 90095-1563, USA. bentler@ucla.edu.

Biography

Peter M. Bentler is Distinguished Professor of Psychology and Statistics at the University of California, Los Angeles. An author of the EQS Structural Equations and EQSIRT computer programs, he lectures worldwide on theory and applications of structural equation models. For his research impact,see http://scholar.google.com/citations?hl=en&view_op=search_authors&mauthors=Unive rsity+of+California, bentler@ucla.edu

Wenjing Huang received her Ph.D from the University of California, Los Angeles. Her training is in applied statistics with an emphasis on psychometrics, especially Structural Equation Modeling. Her dissertation under the guidance of Dr. Peter Bentler and Dr. Theo Dijkstra explores consistent and efficient estimators for PLS-SEM. wjhuang@ucla.edu

Footnotes

Publisher's Disclaimer: This is a PDF file of an unedited manuscript that has been accepted for publication. As a service to our customers we are providing this early version of the manuscript. The manuscript will undergo copyediting, typesetting, and review of the resulting proof before it is published in its final citable form. Please note that during the production process errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain.

1

To illustrate, teachers nearly universally have used a composite score to grade student performance.

2

Especially when the observed data is binary. In such a case, we may take X to be a continuous latent variable with a probit or logit relation to its observed counterpart Y.

3

When this is done, the interpretation of E shifts to uniqueness, incorporating both specificity and random error.

4

Bentler and Molenaar (2012) discuss models that seem to contain no factors but nonetheless are LVM.

5

Hwang et al. consider “a latent interaction as a product of interacting latent variables whose individual scores are uniquely determined as weighted composites of observed variables” (p. 229). This would not define latent variables or an LVM according to Bentler’s (1982) definition (Bollen & Hoyle, 2012; Treiblmaier, Bentler, & Mair, 2011).

6

We thank Professor Dijkstra for his contributions to this work.

References

  1. An X, Bentler PM. Nesting Monte Carlo EM for high-dimensional item factor analysis. Journal of Statistical Computation and Simulation. 2011;83(1):25–36. doi: 10.1080/00949655.2011.599810. DOI: 10.1080/00949655.2011.599810. [DOI] [PMC free article] [PubMed] [Google Scholar]
  2. An X, Bentler PM. Efficient direct sampling MCEM algorithm for latent variable models with binary responses. Computational Statistics and Data Analysis. 2012;56:231–244. [Google Scholar]
  3. An X, Yang Q, Bentler PM. A latent factor linear mixed model for high-dimensional longitudinal data analysis. Statistics in Medicine. 2013;32(24):4229–4239. doi: 10.1002/sim.5825. [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Bartholomew DJ. The foundations of factor analysis. Biometrika. 1984;71:221–232. [Google Scholar]
  5. Bartholomew DJ. Foundations of factor analysis: Some practical implications. British Journal of Mathematical and Statistical Psychology. 1985;38:1–10. [Google Scholar]
  6. Bentler PM. Alpha-maximized factor analysis (Alphamax): Its relation to alpha and canonical factor analysis. Psychometrika. 1968;33:335–345. doi: 10.1007/BF02289328. [DOI] [PubMed] [Google Scholar]
  7. Bentler PM. A lower-bound method for the dimension-free measurement of internal consistency. Social Science Research. 1972;1:343–357. [Google Scholar]
  8. Bentler PM. Multistructure statistical model applied to factor analysis. Multivariate Behavioral Research. 1976;11:3–25. doi: 10.1207/s15327906mbr1101_1. [DOI] [PubMed] [Google Scholar]
  9. Bentler PM. Multivariate analysis with latent variables: Causal modeling. Annual Review of Psychology. 1980;31:419–456. [Google Scholar]
  10. Bentler PM. Linear systems with multiple levels and types of latent variables. In: Jöreskog KG, Wold H, editors. Systems under indirect observation: Causality, structure, prediction. Part I. North-Holland; Amsterdam: 1982. pp. 101–130. [Google Scholar]
  11. Bentler PM. On the implications of Bartholomew's approach to factor analysis. British Journal of Mathematical and Statistical Psychology. 1985a;38:129–131. [Google Scholar]
  12. Bentler PM. Theory and implementation of EQS, a structural equations program. BMDP Statistical Software; Los Angeles, CA: 1985b. [Google Scholar]
  13. Bentler PM. Covariance structure models for maximal reliability of unit-weighted composites. In: Lee S-Y, editor. Handbook of latent variable and related models. North-Holland; Amsterdam: 2007. pp. 1–19. [Google Scholar]
  14. Bentler PM. Alpha, dimension-free, and model-based internal consistency reliability. Psychometrika. 2009;74:137–143. doi: 10.1007/s11336-008-9100-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Bentler PM. SEM with simplicity and accuracy. Journal of Consumer Psychology. 2010;20:215–220. doi: 10.1016/j.jcps.2010.03.002. [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Bentler PM. Invited paper presented at meeting of Western Psychological Association. UCLA Statistics Preprint #621; Los Angeles CA: 2011. Can interval-level scores be obtained from binary responses? http://preprints.stat.ucla.edu/ [Google Scholar]
  17. Bentler PM. EQS 7 structural equations program manual. Multivariate Software.; Encino CA: (in preparation) ( www.mvsoft.com) [Google Scholar]
  18. Bentler PM, de Leeuw J. Factor analysis via components analysis. Psychometrika. 2011;76:461–470. [Google Scholar]
  19. Bentler PM, Dijkstra T. Efficient estimation via linearization in structural models. In: Krishnaiah PR, editor. Multivariate analysis VI. North-Holland; Amsterdam: 1985. pp. 9–42. [Google Scholar]
  20. Bentler PM, Liang J. A unified approach to two-level structural equation models and linear mixed effects models. In: Dunson D, editor. Random effects and latent variable model selection. Springer; New York: 2008. pp. 95–119. [Google Scholar]
  21. Bentler PM, Liang J, Tang M-L, Yuan K-H. Constrained maximum likelihood estimation for two-level mean and covariance structure models. Educational and Psychological Measurement. 2011;71:325–345. doi: 10.1177/0013164410381272. [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Bentler PM, Molenaar PCM. The Houdini transformation: True, but illusory. Multivariate Behavioral Research. 2012;47:442–447. doi: 10.1080/00273171.2012.673941. [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Bentler PM, Mooijaart A. Choice of structural model via parsimony: A rationale based on precision. Psychological Bulletin. 1989;106:315–317. doi: 10.1037/0033-2909.106.2.315. [DOI] [PubMed] [Google Scholar]
  24. Bentler PM, Satorra A. Testing model nesting and equivalence. Psychological Methods. 2010;15:111–123. doi: 10.1037/a0019625. [DOI] [PMC free article] [PubMed] [Google Scholar]
  25. Bentler PM, Weeks DG. Linear structural equations with latent variables. Psychometrika. 1980;45:289–308. [Google Scholar]
  26. Bentler PM, Woodward JA. Regression on linear composites: Statistical theory and applications. Multivariate Behavioral Research Monographs, No. 79-1. 1979 [Google Scholar]
  27. Bollen KA, Hoyle RH. Latent variables in structural equation modeling. In: Hoyle RH, editor. Handbook of structural equation modeling. Guilford; New York: 2012. pp. 56–67. [Google Scholar]
  28. Browne MW. Generalized least squares estimators in the analysis of covariance structures. South African Statistical Journal. 1974;8:1–24. [Google Scholar]
  29. Cai L. A two-tier full-information item factor analysis model with applications. Psychometrika. 2010;75:581–612. [Google Scholar]
  30. Cai L, Yang JS, Hansen M. Generalized full-information item bifactor analysis. Psychological Methods. 2011;16:221–248. doi: 10.1037/a0023350. [DOI] [PMC free article] [PubMed] [Google Scholar]
  31. Cassel C, Hackl P, Westlund A. Robustness of partial least-squares method for estimating latent variable quality structures. Journal of Applied Statistics. 1999;26:435–446. [Google Scholar]
  32. Cassel C, Hackl P, Westlund A. On measurement of intangible assets: a study of robustness of partial least squares. Total Quality Management. 2000;11:897–907. [Google Scholar]
  33. Chin WW. The partial least squares approach for structural equation modeling. In: Marcoulides GA, editor. Modern methods for business research. Lawrence Erlbaum Associates; London: 1998. pp. 295–336. [Google Scholar]
  34. Cochran WG. Errors of measurement in statistics. Technometrics. 1968;10:637–666. [Google Scholar]
  35. de Leeuw J, Meijer E. Introduction to multilevel analysis. In: de Leeuw J, Meijer E, editors. Handbook of multilevel analysis. Springer; New York: 2008. pp. 1–75. [Google Scholar]
  36. Dijkstra TK. Some comments on maximum likelihood and partial least squares methods. Journal of Econometrics. 1983;22:67–90. [Google Scholar]
  37. Dijkstra TK. Consistent partial least squares estimators for linear and polynomial factor models. A report of a belated, serious and not even unsuccessful attempt. 2011 www.rug.nl/staff/t.k.dijkstra/research.
  38. Embretson SE, Reise SP. Item response theory for psychologists. Erlbaum; Mahwah NJ: 2000. [Google Scholar]
  39. Fuller WA. Measurement error models. Wiley; New York: 1987. [Google Scholar]
  40. Hair JF, Sarstedt M, Pieper TM, Ringle CM. The use of partial least squares structural equation modeling in strategic management research: A review of past practices and recommendations for future applications. Long Range Planning. 2012;45:320–340. [Google Scholar]
  41. Hair JF, Sarstedt M, Ringle CM, Mena JA. An assessment of the use of partial least squares structural equation modeling in marketing research. Journal of the Academy of Marketing Science. 2012;40:414–433. [Google Scholar]
  42. Henry KL, Muthén B. Multilevel latent class analysis: An application of adolescent smoking typologies with individual and contextual predictors. Structural Equation Modeling. 2010;17:193–215. doi: 10.1080/10705511003659342. [DOI] [PMC free article] [PubMed] [Google Scholar]
  43. Henseler J. Why generalized structured component analysis is not universally preferable to structural equation modeling. Journal of the Academy of Marketing Science. 2012;40:402–413. [Google Scholar]
  44. Hoyle RH. Handbook of structural equation modeling. Guilford; New York: 2012. [Google Scholar]
  45. Hoshino T, Bentler PM. de Leon AR, Chough KC, editors. Bias in factor score regression and a simple solution. Analysis of mixed data: Methods and applications. 2013 (Ch. 4, pp. 43-61) [Google Scholar]
  46. Huang W. PLSe: Efficient estimators and tests for partial least squares. UCLA: PhD Dissertation. 2013. In progress. [Google Scholar]
  47. Hwang H, Ho RM, Lee J. Generalized structured component analysis with latent interactions. Psychometrika. 2010;75:228–242. [Google Scholar]
  48. Henseler J, Dijkstra TK, Sarstedt M, Ringle CM, Diamantopoulos A, Straub DW, Ketchen DJ, Hair JF, Hult GTM, Calantone RJ. Common Beliefs and Reality about Partial Least Squares: Comments on Rönkkö & Evermann (2013). Organizational Research Methods. 2014 (in press) [Google Scholar]
  49. Jöreskog KG. Testing structural equation models. In: Bollen KA, Long JS, editors. Testing structural equation models. Sage; Newbury Park, CA: 1993. pp. 294–316. [Google Scholar]
  50. Jöreskog KG, Sörbom D. LISREL 8: User's reference guide. Scientific Software; Chicago IL: 1996. [Google Scholar]
  51. Lee S-Y, Jennrich RI. A study of algorithms for covariance structure analysis with specific comparisons using factor analysis. Psychometrika. 1979;44:99–113. [Google Scholar]
  52. Li L, Bentler PM. The greatest lower bound to reliability: Corrected and resampling estimators. Modelling and Data Analysis. 2011;1:87–104. [Google Scholar]
  53. Maruyama G, McGarvey B. Evaluating causal models: An application of maximum likelihood analysis of structural equations. Psychological Bulletin. 1980;87:502–512. [Google Scholar]
  54. McArdle JJ. Latent curve modeling of longitudinal growth data. In: Hoyle RH, editor. Handbook of structural equation modeling. Guilford; New York: 2012. pp. 547–570. [Google Scholar]
  55. Millsap RE. Statistical approaches to measurement invariance. Routledge; New York: 2011. [Google Scholar]
  56. Mooijaart A, Bentler PM. An alternative approach for nonlinear latent variable models. Structural Equation Modeling. 2010;17(3):357–373. [Google Scholar]
  57. Muthén B. Beyond SEM: General latent variable modeling. Behaviormetrika. 2003;29:81–117. [Google Scholar]
  58. Raykov T. Estimation of maximal reliability: A note on a covariance structure modeling approach. British Journal of Mathematical and Statistical Psychology. 2004;57:21–27. doi: 10.1348/000711004849295. [DOI] [PubMed] [Google Scholar]
  59. Reckase M. Multidimensional item response theory. Springer; New York: 2009. [Google Scholar]
  60. Rigdon EE. Rethinking partial least squares path modeling: In praise of simple methods. Long Range Planning. 2012;45(5-6):341–358. [Google Scholar]
  61. Ringle CM, Sarstedt M, Straub DW. A critical look at the use of PLS-SEM in MIS Quarterly. MIS Quarterly. 2012;36:iii–xiv. S3–S8. [Google Scholar]
  62. Rönkkö M, Evermann J. A critical examination of common beliefs about partial least squares path modeling. Organizational Research Methods. 2013;16(3):425–448. [Google Scholar]
  63. Satorra A, Bentler PM. Corrections to test statistics and standard errors in covariance structure analysis. In: von Eye A, Clogg CC, editors. Latent variables analysis: Applications for developmental research. Sage; Thousand Oaks, CA: 1994. pp. 399–419. [Google Scholar]
  64. Satorra A, Bentler PM. Ensuring positiveness of the scaled difference chi-square test statistic. Psychometrika. 2010;75:243–248. doi: 10.1007/s11336-009-9135-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  65. Shiyko MP, Ram N, Grimm KJ. An overview of growth mixture modeling. In: Hoyle RH, editor. Handbook of structural equation modeling. Guilford; New York: 2012. pp. 532–546. [Google Scholar]
  66. Skrondal A, Laake P. Regression among factor scores. Psychometrika. 2001;66:563–576. [Google Scholar]
  67. Treiblmaier H, Bentler PM, Mair P. Formative constructs implemented via common factors. Structural Equation Modeling. 2011;18:1–17. [Google Scholar]
  68. van der Linden WJ, Hambleton RK. Handbook of modern item response theory. Springer-Verlag; New York: 2010. [Google Scholar]
  69. Vinzi VE, Trinchera L, Amato S. PLS path modeling: From foundations to recent developments and open issues for model assessment and improvement. In: Vinzi VE, Chin WW, Henseler J, Wang H, editors. Handbook of partial least squares. Springer-Verlag; Heidelberg: 2010. pp. 47–82. [Google Scholar]
  70. Wansbeek T, Meijer E. Measurement error and latent variables in econometrics. Elsevier; Amsterdam: 2000. [Google Scholar]
  71. Wu E, Bentler PM. EQSIRT, A comprehensive item response theory program. Multivariate Software; Encino CA: 2013. [Google Scholar]
  72. Wu J, Bentler PM. Application of H-likelihood to factor analysis models with binary response data. Journal of Multivariate Analysis. 2012;106:72–79. [Google Scholar]
  73. Wu J, Bentler PM. Limited information estimation in binary factor analysis: A review and extension. Computational Statistics & Data Analysis. 2013;57:392–403. doi: 10.1016/j.csda.2012.06.022. [DOI] [PMC free article] [PubMed] [Google Scholar]
  74. Yuan K-H, Bentler PM. Structural equation modeling. In: Rao CR, Sinharay S, editors. Handbook of statistics 26 Psychometrics. North-Holland; Amsterdam: 2007. pp. 297–358. [Google Scholar]
  75. Yuan K-H, Bentler PM. Finite normal mixture SEM analysis by fitting multiple conventional SEM models. In: Liao TF, editor. Sociological Methodology. Wiley; New York: 2010. 2010. pp. 191–245. [DOI] [PMC free article] [PubMed] [Google Scholar]

RESOURCES