To the Editors
We compliment [1] who used Item Response Theory (IRT), a statistical technique widely employed in educational and psychological assessments, to evaluate allostatic load (AL). This allows an advancement in the field compared with the traditional AL index scoring method: dichotomizing scores for various AL biomarkers based on clinically- or empirically-determined cutoffs and then adding one point for each marker that exceed these thresholds. The use of this approach in coming studies, however, should consider the following three issues.
First, although Liu et al.’s supplementary material includes a script for the Confirmatory Factor Analysis (CFA) used to investigate the dimensionality of AL, model fit indices were not described (e.g., root mean square error of approximation, comparative fit index, etc.). This is needed to indicate the model is not misspecified [2], an essential step to verify whether IRT assumptions are met [3]: local independence of the items and the unidimensionality of the latent trait.
Second, despite the authors’ claim that item parameter scores are robust even when the unidimensionality assumption is violated, model fits must still be checked if IRT is to be used to build a common scale across different datasets. This allows any arising misfits/misspecifications to be tackled to propose psychometrically sound models [4]. Understanding why Liu et al. considered AL as a unidimensional construct when it is usually hypothesized as being multidimensional would also aid in this respect.
Third, when performing CFA it is fundamental to choose the proper estimator according to variables’ levels of measurement. In the example provided in Liu et al.’s supplementary file, it seems Maximum Likelihood (ML) estimator was used, which is adequate for scalar and not dichotomous items. To avoid erroneous conclusions, Weighted Least Squares Mean and Variance (WLSMV) should be used for dichotomous data [5]. R package lavaan can automatically select the best estimator if the type of data (dichotomous, scalar) is reported correctly.
We hope our suggestions assist the development of studies designed to explore AL indicators with latent model frameworks. Researchers might also consider using biomarkers’ original scalar units of measurement rather than the traditional point-based composite score, which will likely reflect more AL variance.
Funding
This study was supported by Fundação de Amparo à Pesquisa do Estado de São Paulo (FAPESP – Process 2016/14750-0, 2017/02816-9), Conselho Nacional de Desenvolvimento Científico e Tecnológico (CNPq – #301899/2019-3 and 302511/2018-0, and Associação Fundo de Incentivo à Pesquisa (AFIP).
Declaration of competing interest
The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
References
- 1.Liu S.H., Juster R.P., Dams-O’Connor K., Spicer J. Allostatic load scoring using item response theory. Compr. Psychoneuroendocr. 2021;5:10025. doi: 10.1016/j.cpnec.2020.100025. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Schermelleh-Engel K., Moosbrugger H., Müller H. Evaluating the fit of structural equation models: tests of significance and descriptive goodness-of-fit measures. Methods Psychol. Res. Online. 2003;8:23–74. [Google Scholar]
- 3.Yang F.M., Kao S.T. Item response theory for measurement validity. Shangai Arch. Psychiatr. 2014;26(3):171–177. doi: 10.3969/j.issn.1002-0829.2014.03.010. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Zhao Y., Hambleton R.K. Practical consequences of item response theory model misfit in the context of test equating with mixed-format test data. Front. Psychol. 2017;8:484. doi: 10.3389/fpsyg.2017.00484. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Li C.H. Confirmatory factor analysis with ordinal data: comparing robust maximum likelihood and diagonally weighted least squares. Behav. Res. 2016;48:936–949. doi: 10.3758/s13428-015-0619-7. [DOI] [PubMed] [Google Scholar]
