Skip to main content
Journal of Intelligence logoLink to Journal of Intelligence
editorial
. 2020 Nov 19;8(4):39. doi: 10.3390/jintelligence8040039

Analysis of an Intelligence Dataset

Nils Myszkowski 1
PMCID: PMC7709695  PMID: 33227922

It is perhaps popular belief—at least among non-psychometricians—that there is a unique or standard way to investigate the psychometric qualities of tests. If anything, the present Special Issue demonstrates that it is not the case. On the contrary, this Special Issue on the “analysis of an intelligence dataset” is, in my opinion, a window to the present vividness of the field of psychometrics.

Much like an invitation to revisit a story with various styles or with various points of view, this Special Issue was opened to contributions that offered extensions or reanalyses of a single—and somewhat simple—dataset, which had been recently published. The dataset was from a recent paper (Myszkowski and Storme 2018), and contained responses from 499 adults to a non-verbal logical reasoning multiple-choice test, the SPM–LS, which consists of the Last Series of Raven’s Standard Progressive Matrices (Raven 1941). The SPM–LS is further discussed in the original paper (as well as through the investigations presented in this Special Issue), and most researchers in the field are likely familiar with the Standard Progressive Matrices. The SPM–LS is simply a proposition to use the last series of the test as a standalone test. A minimal description of the SPM–LS would probably characterize it as a theoretically unidimensional measure—in the sense that one ability is tentatively measured—comprised of 12 pass-fail non-verbal items of (tentatively) increasing difficulty. Here, I refer to the pass-fail responses as the binary responses, and the full responses (including which distractor was selected) as the polytomous responses. In the original paper, a number of analyses had been used, including exploratory factor analysis with parallel analysis, confirmatory factor analyses using a structural equation modeling framework, binary logistic item response theory models (1-, 2-, 3- and 4- parameter models), and polytomous (unordered) item response theory models, including the nominal response model (Bock 1972) and nested logit models (Suh and Bolt 2010). In spite of how extensive the original analysis may have seemed, the contributions of this Special Issue present several extensions to our analyses.

I will now briefly introduce the different contributions of the Special Issue, in chronoligical order of publication. In their paper, Garcia-Garzon et al. (2019) propose an extensive reanalysis of the dimensionality of the SPM–LS, using a large variety of techniques, including bifactor models and exploratory graph analysis. Storme et al. (2019) later find that the reliability boosting strategy proposed in the original paper—which consisted of using nested logit models (Suh and Bolt 2010) to recover information from distractor information—is useful in other contexts, by using the example on a logical reasoning test applied in a personnel selection context. Moreover, Bürkner (2020) later presents how to use his R Bayesian multilevel modeling package brms (Bürkner 2017) in order to estimate various binary item response theory models, and compares the results with the frequentist approach used in the original paper with the item response theory package mirt (Chalmers 2012). Furthermore, Forthmann et al. (2020) later proposed a new procedure that can be used to detect (or select) items that could present discriminating distractors (i.e., items for which distractor responses could be used to extract additional information). In addition, Partchev (2020) then discusses issues that relate to the use of distractor information to extract information on ability in multiple choice tests, in particular in the context of cognitive assessment, and presents how to use the R package dexter (Maris et al. 2020) to study the binary responses and distractors of the SPM–LS. I then present an analysis of the SPM–LS (especially of its monotonicity) using (mostly) the framework of Mokken scale analysis (Mokken 1971). Finally, Robitzsch (2020) proposes new procedures for latent class analysis applied on the polytomous responses, combined with regularization to obtain models of parsimonious complexity.

It is interesting to note that, in spite of the relative straightforwardness of the task and the relative simplicity of the dataset—which in the end, contains answers to a few pass-fail items in a (theoretically) unidimensional instrument—the contributions of this Special Issue offer a lot of original and new perspectives on analyzing intelligence test data. Admittedly, much like the story retold 99 times in Queneau’s Exercices de style, the dataset reanalysed in this Special Issue is, in of itself, of moderate interest. Nevertheless, the variety, breadth and complementarity of the procedures used, proposed and described here clearly demonstrate the creative nature of the field, giving an echo to the proposition by Thissen (2001) to see artistic value in psychometric engineering. I would like to thank Paul De Boeck for proposing the topic of this Special Issue and inviting me to act as guest editor, as well as the authors and reviewers of the articles published in this issue for their excellent contributions. I hope that the readers of Journal of Intelligence will find as much interest in them as I do.

Funding

This research received no external funding.

Conflicts of Interest

The author declares no conflict of interest.

Footnotes

Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  1. Bock R. Darrell. Estimating item parameters and latent ability when responses are scored in two or more nominal categories. Psychometrika. 1972;37:29–51. doi: 10.1007/BF02291411. [DOI] [Google Scholar]
  2. Bürkner Paul-Christian. Brms: An R Package for Bayesian Multilevel Models Using Stan. Journal of Statistical Software. 2017;80:1–28. doi: 10.18637/jss.v080.i01. [DOI] [Google Scholar]
  3. Bürkner Paul-Christian. Analysing Standard Progressive Matrices (SPM-LS) with Bayesian Item Response Models. Journal of Intelligence. 2020;8:5. doi: 10.3390/jintelligence8010005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Chalmers R. Philip. Mirt: A Multidimensional Item Response Theory Package for the R Environment. Journal of Statistical Software. 2012;48:1–29. doi: 10.18637/jss.v048.i06. [DOI] [Google Scholar]
  5. Forthmann Boris, Förster Birgit Schütze Natalie, Hebbecker Karin, Flessner Janis, Peters Martin T., Souvignier Elmar. How Much g Is in the Distractor? Re-Thinking Item-Analysis of Multiple-Choice Items. Journal of Intelligence. 2020;8:11. doi: 10.3390/jintelligence8010011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Garcia-Garzon Eduardo, Abad Francisco J., Garrido Luis E. Searching for G: A New Evaluation of SPM-LS Dimensionality. Journal of Intelligence. 2019;7:14. doi: 10.3390/jintelligence7030014. [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Maris Gunter, Bechger Timo, Koops Jesse, Partchev Ivailo. dexter: Data Management and Analysis of Tests. [(accessed on 6 November 2020)];2020 Available online: https://rdrr.io/cran/dexter/
  8. Mokken Robert J. A Theory and Procedure of Scale Analysis. Mouton/De Gruyter; The Hague and Berlin: 1971. [Google Scholar]
  9. Myszkowski Nils, Storme Martin. A snapshot of g? Binary and polytomous item-response theory investigations of the last series of the Standard Progressive Matrices (SPM-LS) Intelligence. 2018;68:109–16. doi: 10.1016/j.intell.2018.03.010. [DOI] [Google Scholar]
  10. Partchev Ivailo. Diagnosing a 12-Item Dataset of Raven Matrices: With Dexter. Journal of Intelligence. 2020;8:21. doi: 10.3390/jintelligence8020021. [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Raven John C. Standardization of Progressive Matrices, 1938. British Journal of Medical Psychology. 1941;19:137–50. doi: 10.1111/j.2044-8341.1941.tb00316.x. [DOI] [Google Scholar]
  12. Robitzsch Alexander. Regularized Latent Class Analysis for Polytomous Item Responses: An Application to SPM-LS Data. Journal of Intelligence. 2020;8:30. doi: 10.3390/jintelligence8030030. [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Storme Martin, Myszkowski Nils, Baron Simon, Bernard David. Same Test, Better Scores: Boosting the Reliability of Short Online Intelligence Recruitment Tests with Nested Logit Item Response Theory Models. Journal of Intelligence. 2019;7:17. doi: 10.3390/jintelligence7030017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Suh Youngsuk, Bolt Daniel M. Nested Logit Models for Multiple-Choice Item Response Data. Psychometrika. 2010;75:454–73. doi: 10.1007/s11336-010-9163-7. [DOI] [Google Scholar]
  15. Thissen David. Psychometric engineering as art. Psychometrika. 2001;66:473–85. doi: 10.1007/BF02296190. [DOI] [Google Scholar]

Articles from Journal of Intelligence are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES