Abstract
Existing non-verbal ability tests are typically protected by copyright, preventing them from being freely adapted or computerized. Working towards an open science framework, we provide 80 novel, open-access abstract reasoning items, an online implementation and item-level data from 659 participants aged between 11 and 33 years: the matrix reasoning item bank (MaRs-IB). Each MaRs-IB item consists of an incomplete matrix containing abstract shapes. Participants complete the matrices by identifying relationships between the shapes. Our data demonstrate age differences in non-verbal reasoning accuracy, which increased during adolescence and stabilized in early adulthood. There was a slight linear increase in response times with age, resulting in a peak in efficiency (i.e. a measure combining speed and accuracy) in late adolescence. Overall, the data suggest that the MaRs-IB is sensitive to developmental differences in reasoning accuracy. Further psychometric validation is recommended.
Keywords: non-verbal reasoning, adolescence, speed–accuracy trade-off, matrix reasoning
1. Introduction
Abstract reasoning is the ability to solve novel problems without task-specific knowledge and a core mechanism of human learning [1]. Abstract reasoning is closely related to more fundamental cognitive functions such as processing speed [2] and working memory [3]. It is also predictive of a number of mental health symptoms [4] and educational attainment [5].
Due to this predictive power and because it is typically neither trained, nor taught, abstract reasoning is often assessed as part of fluid intelligence tests, including the Wechsler Abbreviated Scale of Intelligence [6], Cattell Culture Faire Intelligence Test [7] and Raven's Progressive Matrices [8]. These IQ tests provide reliable, and well-validated abstract reasoning items [9] and are extensively used in clinical, educational, occupational and research settings [10,11]. However, these tasks are not free-to-use, and copyright usually prevents these pen-and-paper tasks from being adapted into computerized tasks. Notable exceptions here are the freely available web-based 20-item Hagen Matrices Test [12] and the 11 matrix reasoning items of the public-domain International Cognitive Ability Resource [13]. These tasks include a very limited number of items, however, and, to our knowledge, are not currently validated in developmental populations.
This limits the usefulness of existing tasks for research studies requiring large sets of computerized items, such as online, neuroimaging and longitudinal studies. In addition, the largely analogue task administration means that existing normative data for these tasks are typically limited to accuracy data and do not include response times. The latter is an important practical and theoretical limitation because of the well-known relationship between abstract reasoning and processing speed [2,14,15] and because of the potential existence of speed–accuracy trade-offs [16].
Speed–accuracy trade-offs were first demonstrated experimentally over a century ago by Woodworth [17] and Martin & Müller [18] for simple motor tasks and are ubiquitous in mental computations of all kinds and across species [16]. According to sequential sampling theories, speed–accuracy trade-offs can be viewed not as experimental noise but rather as a direct result of sequential sampling of evidence [19,20]. The more time spent on a mental computation, the more information can be accumulated. The point at which a decision is made therefore contains rich information about an individual's internal goals, as well as external task demands [16,19]. This is particularly pertinent in developmental studies.
Abstract reasoning capacity increases during childhood and adolescence and peaks in early adulthood [21–23]. These changes in reasoning performance during development have been linked to the protracted maturation of the frontal cortex [21,24,25]. During the same developmental period, impulsivity decreases markedly [26,27], raising the possibility that increases in reasoning scores could reflect either true increases in reasoning capacity, a decrease in impulsivity or both. In other words, there may be a developmental speed–accuracy trade-off.
To address these gaps in the literature, here we provide a set of 80 abstract reasoning items, each one in three shape variants, as well as accuracy and response time item-level data from a large sample of 659 participants aged 11–33 years. These items are freely available for non-commercial purposes (https://osf.io/g96f4/) and can be selected by researchers to yield age-appropriate tasks of custom difficulty and duration.
The design of the matrix reasoning item bank (MaRs-IB) was similar to Raven's matrices [8]. The items consisted of a three-by-three matrix containing abstract shapes in eight out of nine cells. Participants deduced relationships across the eight shapes, which could vary across four relations: colour, size, position and shape. Participants then identified the missing shape from an array of four options. See https://gorilla.sc/openmaterials/36164 for a demonstration of the items. The dimensionality of each item is linked to the number of relations changing in the matrix. While one-relational changes (e.g. a colour change, figure 1a) are typically easy to identify, higher-order relational changes become increasingly difficult (e.g. a three-relational change of shape, colour and position, figure 1b).
The items were originally developed as part of an online training study [28]. Here, we analyse data collected before training using adolescents and adults (N = 659, aged 11–33 years) to provide accuracy and response time data at the sample level, age group level and item level. We also assessed age differences in accuracy, response times, efficiency and productivity in the MaRs-IB to provide insights into developmental differences in response patterns as well as potential speed–accuracy trade-offs. Our aim here is to introduce a novel, open-access item bank of abstract reasoning items for studies that include adolescents and adults, and to provide a preliminary investigation of their psychometric characteristics.
2. Methods
2.1. Participants
Here, we analysed data from baseline assessments of a larger training study [28]. The research team recruited 821 participants from 16 schools in the London area and through UCL participant pools and posters around campus. Of this sample, 659 participants were included in the current analysis (396 females and 263 males, mean age at baseline = 16.21, s.e. = 0.16, age range = 11.27–33.15 years). The exclusion criteria were: missing relational reasoning data at baseline (N = 5); missing parental consent for adolescents (N = 123) and developmental conditions such as attention-deficit hyperactivity disorder and dyslexia (N = 34). For the training study, adolescents were split into three age groups of equal size and adults were included as a fourth age group (table 1). We report data for each of these age groups to provide researchers with guidelines for different age groups. We also replicated our analyses using age as a continuous variable.
Table 1.
age group | sample size |
---|---|
younger adolescents (aged 11.27–13.39) |
total: 185 females: 118 males: 67 |
mid-adolescents (aged 13.40–15.91) |
total: 184 females: 89 males: 95 |
older adolescents (aged 15.93–17.99) |
total: 184 females: 108 males: 76 |
adults (aged 18.00–33.15) |
total: 106 females: 81 males: 25 |
A follow-up study was conducted to assess the convergent validity of the MaRs-IB. A power analysis suggested that 38 participants are sufficient to detect a correlation of 0.5 at 90% power. For this study, we thus initially recruited 50 participants, with a further 50 participants tested upon reviewer request (total N = 100, 73 females, 27 males, mean age = 23.95, s.e. = 0.35, age range 19–35). The results of the initial analysis (N = 50) are available in the electronic supplementary material for transparency (see supplementary results). Studies were carried out in accordance with UCL Research Ethics Guidelines and approved by the UCL Research Ethics Committee. Informed assent or informed consent was obtained from all participants included in this study. The studies were not formally preregistered. All materials and item-level data have been made available on a permanent third-party archive (https://osf.io/g96f4/); requests for the item-level data can be sent via email to sjblakemore@psychol.cam.ac.uk. The data analysed here were derived from a training study, which was powered to detect small-to-medium effect sizes [28].
2.2. Testing procedure
The testing procedure was previously described by Knoll et al. [28]. Participants were tested on a battery of tasks including one containing MaRs-IB items. Testing was delivered via an online platform developed by the research team and a software company (www.cauldron.sc).
Adolescent participants completed the testing session in groups of 3–48 in school, while adult participants were tested in groups of 1–15 in a university computer room. Participants used computers or tablets. Responses were recorded using a mouse or touchpad. The order of the tasks was counterbalanced between testing groups.
An experimenter gave instructions before the task. Participants then completed practice items until three were completed correctly. More than three practice items were required by 121 participants. On average, these 121 participants needed 4.46 (s.e. = 0.10) practice items to proceed to the task and never more than 10. All participants completed three practice items successfully during the testing session and proceeded to the task. In all items (i.e. both practice and test items), participants were given visual feedback on their performance.
2.3. MaRs-IB design
Each MaRs-IB item consisted of a 3 × 3 matrix. Eight of the nine resulting cells contained an abstract shape, while one cell on the bottom right-hand side of the matrix was empty. The participants' task was to complete the matrix by finding the missing shape among four possible alternatives (see figure 1 for examples). To select the correct missing shape, participants had to deduce relationships between the shapes of the matrix. These shape characteristics varied along four dimensions: shape, colour, size and position in the matrix.
Items began with a 500 ms fixation cross, followed by a 100 ms white screen. Participants were then given up to 30 s to complete an item. After 25 s a clock appeared, indicating that 5 s remained before the next item began. An item ended when participants responded, or after 30 s had elapsed without response. For each participant, puzzles were taken from one of three test forms of 80 items. These test forms were created to prevent familiarity effects in future testing sessions of a training study and they were generated by drawing from one of three shape sets, in a counterbalanced fashion. The items in each test form thus tested the same abstract reasoning problems and differed only in the exact shapes used (see electronic supplementary material, figure S1 for examples). These shape set variants are also available in the online repository (https://osf.io/g96f4/). Items were presented in the same order for all participants, starting with five simple items intended to familiarize participants with the task. Participants completed the items for 8 minutes but were not informed of the total number of items and were not required, or expected, to complete 80 items in 8 min. The only time constraint stated was to provide a response to each item within 30 seconds. If a participant completed 80 items within 8 min, the items were presented again in the same order, but responses were not analysed.
Two different algorithmic strategies were employed to generate the target and distractor items in the answer options: a minimal difference strategy and a paired difference strategy. These strategies were used to control for the possibility that participants may solve the puzzle in unintended ways. Specifically, the distractor strategies were used to counterbalance the risk of pop-out effects (where the target ‘pops out’ and can be identified by reasoning in fewer dimensions than intended) with the risk that participants could reason from the answer options alone, without looking at the puzzle itself. Note that it is not possible to prevent both possibilities at the same time for three- or higher dimensional puzzles when using only four answer options. The minimal difference algorithm created distractors as variations of the target (see electronic supplementary material, figure S2). This prevented pop-out effects but theoretically allowed participants to solve the puzzle by looking at the answer options only. A paired difference algorithm created distractors that had at least one component in common with the target (electronic supplementary material, figure S2). This could theoretically induce pop-out effects but prevented participants from reasoning from the answer options alone. To counterbalance these risks, each item was pseudo-randomly assigned one of these two distractor strategies and we tested whether the distractor strategies affected MaRs-IB performance in our analyses (see below).
2.4. Convergent validity study
In the follow-up study, we asked participants to complete the MaRs-IB, as well as an existing task: the ‘International Cognitive Ability Resource’ (ICAR) [29]. The latter includes four main tasks: matrix reasoning, series completion, spatial rotations and verbal reasoning. The ICAR is relatively novel. While more traditional tasks are more thoroughly validated, the ICAR is computer-based and has been tested on a large sample of individuals (more than 90 000). The ICAR has also been indirectly related (via the Shipley-2) to more traditional tests such as the Wechsler Adult Intelligence Scale. We particularly focused on the matrix reasoning items of the ICAR, as these most closely resemble the MaRs-IB items and other tests like Raven's progressive matrices. The MaRs-IB and ICAR were administered in counterbalanced order.
2.5. Mars-IB items, data and analyses
All 80 MaRs-IB items in their test forms are freely available for non-commercial use and can be downloaded from https://osf.io/g96f4/. The repository also contains item-level dimensionality, accuracy and response time data for each age group, test form and shape set, which can be used by researchers to adapt the items to tasks of different duration and difficulty. Two further, colour vision deficiency-friendly sets of stimuli, can also be obtained from the repository.
Here, we provide sample- and age group-level accuracy and median response time data on correct items. We also analysed the effect of age on accuracy, response times, productivity and inverse efficiency. Productivity was operationally defined as the number of items completed. Inverse efficiency was calculated as median response times divided by accuracy [30].
Items with a response time under 250 ms were excluded from all analyses. We modelled each of these four dependent variables using mixed models because these allowed to accurately partition the error terms according to the hierarchical structure of the data (e.g. item-level data were clustered by participants, which were in turn nested in schools) [31]. We used generalized linear mixed models (GLMMs) for accuracy and response times, and linear mixed models (LMMs) for inverse efficiency using lme4 [31] in R [32]. GLMMs were used to model raw item-level data, with participant ID and school or university as nested random intercepts and item code as crossed random intercept. Accuracy and response times were modelled at the item level after the removal of incomplete items. Accuracy data were modelled using the binomial distribution with a logit link function while response times were modelled using the Gamma distribution with a log link function [33]. Productivity (the number of items completed) was necessarily an aggregate measure. To calculate inverse efficiency, data also had to be aggregated across items for each participant. LMMs estimating the number of items completed and inverse efficiency therefore included only school or university as a random intercept. All models included fixed effects for either age group as a categorical Helmert-coded variable or orthogonal polynomials of age with the linear, quadratic and cubic components. In exploratory models, we additionally included Helmert-coded gender and the interaction between age group and gender as fixed effects.
To assess the equivalency of the forms of the MaRs-IB, we used exploratory GLMMs specified as described above to investigate the effect of test form (test form 1, 2 or 3) on accuracy and response times. Moreover, given that test forms were created by pseudo-randomly drawing items from one of three different shape sets, and solutions were pseudo-randomly generated according to one of two distractor strategies, we additionally investigated whether such test form constituents might also independently affect accuracy and response times. This information may be useful for researchers interested in shaping novel test forms (e.g. by differently combining shape sets and distractor strategies). As a secondary analysis of equivalency, we thus assessed whether distractor type (minimal or paired difference) and shape set (shape set 1, 2 or 3) might also affect accuracy and response times.
To assess response processes, we investigated how accuracy and response times were affected by item dimensionality (a score ranging between 1 and 8, reflecting the number of dimensions changing in a puzzle), using GLMMs as specified above. For the analysis of dimensionality only, the item-related random intercept was removed because of multicollinearity between each item and the associated dimensionality. Main effects and interactions were inspected using omnibus Type III Wald χ2 tests. Planned and post hoc comparisons were performed using the emmeans package [34] and Bonferroni corrected for multiple comparisons.
The analyses above were based on completed items. This excludes (i) trials in which participants timed out and (ii) trials that were not reached. Our data stem from a previous study that aimed to address a different research question. The data are therefore not optimized for psychometric validation. In particular, participants completed varying number of items, and item-level accuracy data are based on a different number of responses for each item, with completion rates dropping for later items. This may limit their comparability. To address this issue, we used mixed models (as described above), which are able to deal with unbalanced data by modelling performance at the trial level, rather than participant level. Furthermore, to provide preliminary insight into more traditional psychometric approaches to item functioning (i.e. internal consistency, biserial correlations, p-values and differential item functioning (DIF)), we focused on a subset of data that involved no variability in completion rate.
To investigate possible DIF, we used logistic regression [35], as implemented in the ‘difR’ package in R [36]. We tested for both uniform and non-uniform DIF, correcting for multiple comparisons using the Benjamini–Hochberg method as recommended by Kim & Oshima [37]. As potential sources of DIF, we focused on age (used as both categorical and continuous) and gender.
2.6. Convergent validity analysis
To investigate convergent validity in the follow-up study, we inspected the product-moment correlation between performance in the MaRs-IB and in the matrix reasoning items of the ICAR. Although the assumptions for correcting for range restrictions could not be fully assessed, our sample was tested in a highly selective university. We therefore followed Condon & Revelle [29] and assessed the possible presence of range restriction by comparing the standard deviations of the matrix reasoning items in our sample and theirs. We obtained the latter by combining the standard deviations for the ICAR matrix reasoning items (listed in table 2 of p. 55 of the study by Condon & Revelle [29]). Range correction was performed with the ‘rangeCorrection’ function of the Psych package in R [38].
3. Results
3.1. Overall performance on the MaRs-IB and preliminary reliability measures
The mean accuracy was 69.15% (s.e. = 0.65%, range = 24.24–100.00%), and the median response time on correct items was 7914 ms (interquartile range (IQR) = 4063 ms). On average, participants completed 32.78 unique items (s.e. = 0.41, range = 13–80) of the MaRs-IB within 8 min. Timing out occurred in only 2% of items. To assess the presence of potential ceiling and floor effects, we inspected skew and kurtosis of response accuracy as well as the rate of chance performance and perfect performance across participants. A skew of −0.41 and a kurtosis of 2.39 suggested that participants' scores fell reasonably close to the normal distribution. Two participants performed below chance (i.e. ≤25% correct items), while eight participants responded correctly on all items, suggesting that 98.48% of participants were above minimum and below maximum accuracy. These results suggest that there were no obvious floor or ceiling effects.
To assess internal consistency, biserial correlations, p-values and DIF, we focused on a subset of items for which there was no variability in completion rate. Specifically, we focused on the 25 items following item 5 (since the first five items were deliberately easier and involved virtually no variance in performance; see MaRs-IB design). For this analysis, we also included trials in which participants had timed out (labelling them as incorrect), as well as those in which response latencies were shorter than 250 ms (only two of which were correct). This resulted in a subset of 25 items that were completed by N = 349 participants (at least N = 70 per age group). To assess internal consistency on this subset of items, we computed Kuder–Richardson 20 formula, which resulted in a KR-20 of 0.78 (0.95% confidence interval (CI) [0.78, 0.79]). The mean item-total biserial correlation for the same items was 0.32 (s.e. = 0.02) (biserial correlation of each item is available in electronic supplementary material, table S2). The mean item p-value of this subset of items (i.e. the proportion of participants that answered correctly to each item) was 0.59 (s.e. = 0.04) (p-values of each item are available in the electronic supplementary material, tables S1 and S2). Note that this mean p-value differs slightly from the mean accuracy value reported above because it refers to the subset of items, not to the full dataset. None of these items displayed uniform or non-uniform DIF, suggesting they are unbiased relative to the age groups studied here and relative to gender. Researchers may want to preferentially use this subset of items for which a more in-depth psychometric analysis is available.
To assess the reliability of the items, we inspected its split-half, test–retest reliability in the full dataset. The Spearman–Brown split-half reliability coefficient was computed on the raw item-level data and was 0.82 overall (younger adolescents: 0.82, mid-adolescents: 0.81, older adolescents: 0.71, adults: 0.80). To assess test–retest reliability, we focused on a subset of participants (N = 218) who were assessed on the MaRs-IB twice, 35 days apart on average (s.e. = 1 day, range = 21–52 days). During this time they trained in an unrelated face-perception task (see [28] for details). Pearson's product–moment correlation in accuracy between the two time points was r = 0.71 (t(183) = 13.58, p < 0.001), suggesting acceptable test–retest reliability. The test–retest relationship did not interact with age (, p = 0.531).
To assess the equivalency of the MaRs-IB forms, we inspected accuracy and response times across different test forms, distractor types and shape sets. The item information of one participant (but not their performance-related information) was lost due to a technical error. Test forms or distractor types did not affect accuracy or response times (test form accuracy: , p = 0.733; test form response times: , p = 0.088; distractor type accuracy: , p = 0.347; distractor type response times: , p = 0.491), suggesting good equivalency of the test forms and of the distractor strategies. Shape sets had no significant impact on accuracy (, p = 0.088), but there was a significant effect on response times (, p = 0.002). Post hoc comparisons suggested that participants showed reduced response times for shape set 1 compared to shape set 2 (z = −3.46, pBonf. = 0.002), while there were no differences between shape sets 1 and 3 (z = −2.16, pBonf. = 0.091) or shape sets 2 and 3 (z = 1.28, pBonf. = 0.601). These response time differences were unexpected. We speculate that they may be due to low-level perceptual features of the different shape sets. See electronic supplementary material, tables S3–S5 for descriptive statistics of the different test forms, shape sets and distractor types, respectively. For an item-level breakdown of the response times by test form and shape set, see https://osf.io/g96f4/.
Finally, item dimensionality had a robust impact on performance, both in terms of accuracy (, p < 0.001) and response times (, p < 0.001). Specifically, item dimensionality linearly decreased the loglikelihood of responding correctly (b = −0.38, s.e. = 0.009), and increased response times (b = 0.18, s.e. = 0.003). Taken together, these results suggest that item properties that were not intended to affect performance (i.e. test forms, shape sets and distractor strategy) did not, whereas properties that were intended to modulate performance and response processes (i.e. item dimensionality) reliably did so.
3.2. Convergent validity
To assess convergent validity, we inspected the product–moment correlation between MaRs-IB performance (at the aggregate level) and matrix reasoning scores in the ICAR [29]. The correlation was 0.61 (t(98) = 7.60, p < 0.001, 95% CI = 0.47, 0.72) and a linear regression showed that it did not interact with the order in which participants did the two tasks (p = 0.58). Performance standard deviations of the ICAR matrix reasoning items in our sample and in Condon & Revelle [29] were, respectively, 0.27 and 0.5, possibly warranting correction for range restriction [29]. This resulted in a range-corrected correlation of 0.81, to be interpreted with caution, given that the assumptions for range restriction correction could not be fully tested. To assess the extent of divergent validity, we further compared the correlation described above to the correlations between the MaRs-IB and the remaining tasks of the ICAR (table 2), namely letter–number series completion, verbal reasoning and three-dimensional (3D) rotations. The highest correlation with MaRs-IB performance was indeed with the ICAR matrix reasoning (table 2). Furthermore, Fischer's r–z transform suggested that this correlation was higher than the correlation between the MaRs-IB and the 3D rotations (p = 0.03). However, it was not significantly different from the correlation between the MaRs-IB and letter–number series completion (p = 0.05), or between the MaRs-IB and the verbal reasoning task (p = 0.11).
Table 2.
ARTOL | MR | R3D | LN | |
---|---|---|---|---|
ARTOL | ||||
MR | 0.61*** | |||
R3D | 0.45*** | 0.50*** | ||
LN | 0.44*** | 0.54*** | 0.45*** | |
VR | 0.48*** | 0.47*** | 0.29* | 0.45*** |
***p < 0.001, *p < 0.05. Bonferroni corrected.
Taken together, this evidence suggests that the MaRs-IB correlates strongly with the ICAR's matrix reasoning task, and more than with another non-verbal reasoning task of the ICAR. However, the uncorrected correlation size falls slightly short of the correlation size that could be expected from the positive manifold of fluid intelligence tests [29]. We also observed no clear evidence of divergent validity with regard to the ICAR's verbal reasoning task. We thus recommend further psychometric testing of our items, possibly using more established measures such as Raven's matrices [8] and larger samples.
We also inspected the relationship between MaRs-IB and digit span performance, as a measure of working memory, and found that the two measures were strongly correlated (r = 0.45, t(657) = 12.74, 95% CI [0.38, 0.5], p < 0.001). This is in line with previous meta-analyses, which have demonstrated an average correlation of = 0.479 between working memory and intelligence [3]. The relationship between MaRs-IB and digit span performance did not differ between age groups (, p = 0.172).
3.3. Age differences in performance
3.3.1. Accuracy
MaRs-IB accuracy differed between age groups (, p = 0.016) (table 3). Planned contrasts suggested the age effect was mainly carried by adults performing significantly better than younger adolescents (figure 2; electronic supplementary material, table S6). The age effect was also evident when age was modelled as a continuous variable (figure 2). There was a significant, positive linear effect and a significant negative quadratic trend (table 4). Performance increased with age during adolescence before tapering off in adulthood (figure 2).
Table 3.
age group | younger adolescents | mid-adolescents | older adolescents | adults |
---|---|---|---|---|
accuracy (mean) | 61 (m = 57, f = 63) | 68 (m = 67, f = 68) | 73 (m = 72, f = 73) | 81 (m = 79, f = 81) |
accuracy (s.e.) | 1 (m = 2, f = 1) | 1 (m = 2, f = 2) | 1 (m = 2, f = 1) | 1 (m = 3, f = 1) |
RT (median) | 6944 (m = 7006, f = 6872) | 7552 (m = 7964, f = 6841) | 7952 (m = 8774, f = 7582) | 9454 (m = 10 276, f = 9304) |
RT (IQR) | 3879 (m = 4301, f = 3772) | 3723 (m = 4629, f = 2669) | 3309 (m = 3338, f = 2781) | 3231 (m = 3012, f = 3297) |
items completed (mean) | 33.68 (m = 34.4, f = 33.27) | 32.29 (m = 32.17, f = 32.42) | 28.99 (m = 28.24, f = 29.52) | 33.43 (m = 31.96, f = 33.89) |
items completed (s.e.) | 0.74 (m = 1.26, f = 0.92) | 0.76 (m = 1.19, f = 0.92) | 0.75 (m = 1.22, f = 0.94) | 0.93 (m = 2.07, f = 1.04) |
inverse efficiency (median) | 11 838 (m = 12 655, f = 11 100) | 10 937 (m = 11 557, f = 10 150) | 11 057 (m = 12 285, f = 10 694) | 11 414 (m = 12 946, f = 11 226) |
inverse efficiency (IQR) | 4031 (m = 4821, f = 3263) | 3852 (m = 4456, f = 3256) | 4023 (m = 3941, f = 4470) | 3668 (m = 4452, f = 3248) |
Table 4.
trend | b | z | p-value |
---|---|---|---|
accuracy | |||
linear | 33.14 | 4.05 | <0.001 |
quadratic | −38.87 | −7.22 | <0.001 |
cubic | 10.58 | 1.78 | 0.074 |
response times | |||
linear | 5.15 | 2.43 | 0.015 |
quadratic | −1.21 | −0.8 | 0.426 |
cubic | −1.19 | −0.70 | 0.484 |
number of items completed | |||
linear | −36.90 | −1.33 | 0.188 |
quadratic | 29.17 | 2.12 | 0.035 |
cubic | −7.28 | −0.54 | 0.589 |
inverse efficiency | |||
linear | 5391.41 | 0.67 | 0.509 |
quadratic | 9755.70 | 2.29 | 0.023 |
cubic | −7861.30 | −1.86 | 0.064 |
There was no significant effect of gender on accuracy (, p = 0.052) (electronic supplementary material, figure S3) and no interaction between gender and age group (, p = 0.878).
3.3.2. Response times
There was no significant effect of age group on response times (, p = 0.673) (table 3) and none of the planned comparisons between age groups were significant (figure 3; electronic supplementary material, table S7). There was a slight positive linear trend, while the quadratic and cubic trends were non-significant (figure 3 and table 4). This indicates that there may be some increases in response times with age.
There was a significant effect of gender on response times (, p = 0.005) (electronic supplementary material, figure S3), with females responding faster than males (estimated meanFemales = 5523 ms, s.e.Females = 466 ms, estimated meanMales = 5915 ms, s.e.Males = 502 ms). There was no significant interaction between gender and age (, p = 0.311).
3.3.3. Number of items completed
The number of items completed did not differ between age groups (, p = 0.859) (table 3). None of the age group comparisons were significant (figure 4; electronic supplementary material, table S8). When modelling age as a continuous variable there was no significant linear or cubic trend. There was, however, a significant quadratic trend of age, suggesting that the number of items completed was somewhat lower in late adolescence and early adulthood than before late adolescence or after early adulthood (figure 4). This is in line with the decrease in response times observed above and indicates a slight decrease in productivity during late adolescence and early adulthood.
There was no effect of gender on the number of items completed (, p = 0.300), nor an interaction between gender and age (, p = 0.398) (electronic supplementary material, figure S3).
3.3.4. Inverse efficiency
There was no significant main effect of age group on inverse efficiency (, p = 0.288) (table 3) and none of the planned comparisons between age groups survived Bonferroni-correction (figure 5; electronic supplementary material, table S9). There was no linear or cubic effect of age on inverse efficiency (table 4). There was, however, a significant quadratic trend (figure 5 and table 4). This quadratic trend may indicate a peak in efficiency around late adolescence, consistent with the age-related nonlinear increases in accuracy and linear decreases in speed observed above.
There was a significant effect of gender on inverse efficiency (, p < 0.001), with females responding more efficiently than males (estimated meanFemales = 11222, estimated s.e.Females = 398; estimated meanMales = 12522, estimated s.e.Males = 416.58) (electronic supplementary material, figure S3). This effect matched the pattern in accuracy and response times reported above: accuracy was comparable between genders but females responded faster. There was no interaction between gender and age group (, p = 0.314).
4. Discussion
Here, we provide a large bank of novel, open-access abstract reasoning items that show acceptable internal consistency and convergent validity, and reasonable test–retest reliability. We propose that the MaRs-IB can be used for research purposes to complement existing, well-validated but copyrighted pen-and-paper tests [6,7], as well as computerized public-domain tests containing a small number of items tested in adults [12,13]. We provide a set of 80 items, as well as corresponding descriptive developmental data from adolescents and adults (aged 11–33 years). The MaRs-IB items can be implemented in computer-based experiments, allowing for flexible use ranging from online experiments to neuroimaging studies. The computerized implementation also allows for the measurement of response times, in addition to the standard accuracy measures. We further provide items in three shape variants to allow for repeated measures in longitudinal studies. We report performance and item-level statistics for different age groups to allow researchers to form item sets of custom difficulty and duration for developmental and non-developmental studies using information on accuracy, response times and efficiency.
The data analysed here suggest that accuracy in the MaRs-IB increases robustly until late adolescence, after which gains taper off. This is in line with previous research showing that reasoning capacity increases during childhood and adolescence, and into adulthood [21–23]. We extend the current literature by showing that response times increased slightly over the same developmental period and the number of items attempted decreased, resulting in a peak in efficiency in late adolescence. Overall, this suggests that accuracy increases past late-adolescence may be conveyed, in part, by response time slowing.
Two independent, but complementary lines of inquiry may provide insights into possible mechanisms of these developmental differences. First, developmental theories emphasize reductions in impulsivity over adolescence, which are linked to the protracted development of prefrontal regions over the same developmental period [26,27,39,40]. Second, computational theories of speed–accuracy trade-offs highlight that the time to arrive at a cognitive solution gives insights into the amount of evidence participants sample before deciding upon a solution [19,20]. The results obtained here suggest that participants increasingly sample evidence over the course of adolescence, resulting in more conservative decision thresholds at older ages. The peak in efficiency in late adolescence suggests that both accuracy increases and response time slowing drive developmental gains in abstract reasoning up until late adolescence, after which response time slowing may be a major driver of further improvements into adulthood. On a more general level, this finding highlights the value of collecting response time, as well as accuracy data, for providing insights into the developmental mechanisms of abstract reasoning.
We believe that the advantages of providing items open-access, including flexibility of use and equity of access, ultimately outweigh potential costs of open-access provision. Such costs include the possibility that participants may become familiar with or train on the items. This problem is not unique to open-access items—participants are already able to access many matrix reasoning tasks if they wish to. For now, some of the gold-standard abstract reasoning tests, which have been used for decades, also arguably carry a higher likelihood of familiarity than novel items like MaRs-IB. We cannot rule out the possibility that this might change in the long run, however. To address this issue, researchers may wish to implement pre- and post-screening procedures to identify (and possibly exclude) participants familiar with the MaRs-IB. The post-screening procedure may include using outlier detection methods such as examining person-fit statistics [41] or applying DIF analyses to identify participants behaving unusually on MaRs-IB items [42].
Despite strengths of the MaRs-IB, including the large number of items and developmental data, we note several limitations. First, we observed gender differences in response times and efficiency, but not accuracy, with females responding faster than males, resulting in overall higher levels of efficiency. We hypothesize that this effect may be due to low-level perceptual properties of the stimuli, particularly colour-choices. Females generally show more acute colour discrimination [43] and are less likely to present with colour-blindness than males [44]. This may have resulted in faster response times overall. To allow researchers to address colour-decimation as a potential confound in future studies, we provide a separate set of colour vision deficiency-friendly stimuli here: https://osf.io/g96f4/. Second, the data analysed here were originally collected as part of a cognitive training study [28] and were therefore not optimized for psychometric validation. In particular, we did not have enough observations to assess the psychometric properties of some of the later items of the MaRs-IB (which few participants reached). While we were able to provide evidence for acceptable internal consistency in a subset of earlier items that involved no variability in completion rate, further psychometric validation of the MaRs-IB is recommended. We also do not have data on our sample's distribution and representativeness in terms of demographic variables such as socio-economic status and ethnicity, and particular age brackets (e.g. between 18 and 20) are under-sampled. We therefore recommend a more extensive validation of these items with a representative sample in the future. While we encourage the use of the items in measuring non-verbal reasoning in development samples, our item bank should not be considered a normative measure of intelligence until further psychometric testing has been completed.
5. Conclusion
Complementing existing, mostly copyrighted, reasoning tasks, we here present a large set of abstract reasoning items which can be used and adapted, free-of-charge, for computerized research studies requiring large sets of stimuli. The MaRs-IB is sensitive to age differences in accuracy, which suggests that further psychometric validation will make the item bank a useful resource for researchers interested in abstract reasoning.
Supplementary Material
Acknowledgements
The authors would like to thank all participants and participating schools, as well as Alberto Lazari, Somya Iqbal and Fabian Stamp for help with data collection. The authors would also like to thank Yaara Erez and Rogier Kievit for helpful discussions about the items.
Ethics
All participants provided informed consent and all procedures were approved by UCL's ethical committee (Project ID: 4870/001; Project title: ‘Development of cognitive processing during adolescence’).
Data accessibility
Working towards an open science framework, the study in question provides 80 items (each in three variants) for a novel matrix reasoning item bank, a ready-to-use online implementation and descriptive data from 659 participants aged between 11 and 33 years. The manuscript is already available in preprint (https://osf.io/uvteh/). In an open repository (https://osf.io/g96f4/), we also provide the novel items and item-level data. Owing to the limitations of the ethical approvals that were granted some years in advance of this research being conducted, the authors regret that the full datasets cannot currently be made publicly available. To do so would risk revealing data that may contribute to individuals in the study (currently anonymized) becoming identifiable. Specifically, we have not made available the participant-level data. The participant-level data (and code) will be made available upon reasonable request.
Authors' contributions
S.-J.B. wrote the initial grant application and protocol. S.-J.B., D.F., L.J.K. and A.L.S. developed the original study design. Data collection for the original study were performed by D.F., L.J.K. and A.L.S. G.C. and S.-J.B. designed the validation study. Data collection for the validation study was performed by B.P.P.-S. and G.C. G.C. and D.F. performed the data analyses and interpretation, under the supervision of S.-J.B. G.C. and D.F. contributed equally to the manuscript. All authors approved the final version of the manuscript for submission.
Competing interests
The authors declare no conflict of interest.
Funding
The study was funded by the Royal Society, the Jacobs Foundation, the Nuffield Foundation, and Wellcome Trust (grant no. 104908/Z/14/Z). D.F. was supported by the Cusanuswerk, University College London and the UK Medical Research Council.
References
- 1.Krawczyk DC. 2012. The cognition and neuroscience of relational reasoning. Brain Res. 1428, 13–23. ( 10.1016/j.brainres.2010.11.080) [DOI] [PubMed] [Google Scholar]
- 2.Kail RV, Lervåg A, Hulme C. 2015. Longitudinal evidence linking processing speed to the development of reasoning. Dev. Sci. 19, 1067–1074. ( 10.1111/desc.12352) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Ackerman PL, Beier ME, Boyle MO. 2005. Working memory and intelligence: the same or different constructs? Psychol. Bull. 131, 30–60. ( 10.1037/0033-2909.131.1.30) [DOI] [PubMed] [Google Scholar]
- 4.Gale CR, Batty GD, Tynelius P, Deary IJ, Rasmussen F. 2010. Intelligence in early adulthood and subsequent hospitalisation and admission rates for the whole range of mental disorders: longitudinal study of 1,049,663 men. Epidemiology (Cambridge, Mass.) 21, 70–77. ( 10.1097/EDE.0b013e3181c17da8) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Roth B, Becker N, Romeyke S, Schäfer S, Domnick F, Spinath FM. 2015. Intelligence and school grades: a meta-analysis. Intelligence 53, 118–137. ( 10.1016/j.intell.2015.09.002) [DOI] [Google Scholar]
- 6.Wechsler D. 2011. Wechsler abbreviated scale of intelligence, 2nd edn San Antonio, TX: Pearson. [Google Scholar]
- 7.Cattell RB, Cattell HEP. 1973. Measuring intelligence with the culture fair tests. Champaign, IL: The Institute for Personality and Ability Testing. [Google Scholar]
- 8.Raven J. 2009. The Raven progressive matrices and measuring aptitude constructs. Int. J. Educ. Psychol. Assess. 2, 2–38. [Google Scholar]
- 9.Burke HR. 1972. Raven's progressive matrices: validity, reliability, and norms. J. Psychol. 82, 253–257. ( 10.1080/00223980.1972.9923815) [DOI] [Google Scholar]
- 10.Daniel MH. 1997. Intelligence testing: status and trends. Am. Psychol. 52, 1038–1045. ( 10.1037/0003-066X.52.10.1038) [DOI] [Google Scholar]
- 11.Flanagan DP, Harrison PL. 2012. Contemporary intellectual assessment: theories, tests, and issues. New York, NY: Guilford Press. [Google Scholar]
- 12.Heydasch T.2014. The Hagen matrices test (HMT) . [DOI]
- 13.The International Cognitive Ability Resource Team. 2014. The international cognitive ability resource. Retrieved from https://icar-project.com/. [Google Scholar]
- 14.Salthouse TA. 1996. The processing-speed theory of adult age differences in cognition. Psychol. Rev. 103, 403–428. ( 10.1037/0033-295X.103.3.403) [DOI] [PubMed] [Google Scholar]
- 15.Schubert A-L, Hagemann D, Frischkorn GT. 2017. Is general intelligence little more than the speed of higher-order processing? J. Exp. Psychol. Gen. 146, 1498–1512. ( 10.1037/xge0000325) [DOI] [PubMed] [Google Scholar]
- 16.Heitz RP. 2014. The speed-accuracy tradeoff: history, physiology, methodology, and behavior. Front. Neurosci. 8, 150 ( 10.3389/fnins.2014.00150) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Woodworth RS. 1899. Accuracy of voluntary movement. Psychol. Rev.: Monogr. Suppl. 3, i–114. [Google Scholar]
- 18.Martin LJ, Müller GE. 1899. Zur Analyse der Unterschiedsempfindlichkeit. Leipzig, DE: Barth, J.A. [Google Scholar]
- 19.Fitts PM. 1966. Cognitive aspects of information processing: III. Set for speed versus accuracy. J. Exp. Psychol. 71, 849 ( 10.1037/h0023232) [DOI] [PubMed] [Google Scholar]
- 20.Stone M. 1960. Models for choice-reaction time. Psychometrika 25, 251–260. ( 10.1007/BF02289729) [DOI] [Google Scholar]
- 21.Crone EA, Wendelken C, van Leijenhorst L, Honomichl RD, Christoff K, Bunge SA. 2009. Neurocognitive development of relational reasoning. Dev. Sci. 12, 55–66. ( 10.1111/j.1467-7687.2008.00743.x) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Li SC, Lindenberger U, Hommel B, Aschersleben G, Prinz W, Baltes PB. 2004. Transformations in the couplings among intellectual abilities and constituent cognitive processes across the life span. Psychol. Sci. 15, 155–163. ( 10.1111/j.0956-7976.2004.01503003.x) [DOI] [PubMed] [Google Scholar]
- 23.Richland LE, Morrison RG, Holyoak KJ. 2006. Children's development of analogical reasoning: insights from scene analogy problems. J. Exp. Child Psychol. 94, 249–273. ( 10.1016/j.jecp.2006.02.002) [DOI] [PubMed] [Google Scholar]
- 24.Bazargani N, Hillebrandt H, Christoff K, Dumontheil I. 2014. Developmental changes in effective connectivity associated with relational reasoning. Hum. Brain Mapp. 35, 3262–3276. ( 10.1002/hbm.22400) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Dumontheil I, Houlton R, Christoff K, Blakemore SJ. 2010. Development of relational reasoning during adolescence. Dev. Sci. 13, F15–F24. ( 10.1111/j.1467-7687.2010.01014.x) [DOI] [PubMed] [Google Scholar]
- 26.Bezdjian S, Tuvblad C, Wang P, Raine A, Baker LA. 2014. Motor impulsivity during childhood and adolescence: a longitudinal biometric analysis of the Go/No-Go task in 9- to 18-year-old twins. Dev. Psychol. 50, 2549–2557. ( 10.1037/a0038037) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Steinberg L, et al. 2017. Around the world, adolescence is a time of heightened sensation seeking and immature self-regulation. Dev. Sci. 21, e12532. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Knoll LJ, Fuhrmann D, Sakhardande AL, Stamp F, Speekenbrink M, Blakemore SJ. 2016. A window of opportunity for cognitive training in adolescence. Psychol. Sci. 27, 1620–1631. ( 10.1177/0956797616671327) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Condon DM, Revelle W. 2014. The international cognitive ability resource: development and initial validation of a public-domain measure. Intelligence 43, 52–64. ( 10.1016/j.intell.2014.01.004) [DOI] [Google Scholar]
- 30.Townsend JT, Ashby FG. 1978. Methods of modeling capacity in simple processing systems. Cogn. Theory 3, 200–239. [Google Scholar]
- 31.Bates DJ, Maechler M, Bolker B. 2013. lme4: linear-mixed effects models using S4 classes. Retrieved from http://CRAN.R-project.org/package=lme4.
- 32.R Core Team. 2015. R: a language and environment for statistical computing. Vienna, Austria: R Foundation for Statistical Computing; See https://www.R-project.org/. [Google Scholar]
- 33.Lo S, Andrews S. 2015. To transform or not to transform: using generalized linear mixed models to analyse reaction time data. Front. Psychol. 6, 1171 ( 10.3389/fpsyg.2015.01171) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Lenth R, Singmann H, Love J, Buerkner P, Herve M. 2018. Package ‘emmeans’. Retrieved from https://cran.r-project.org/web/packages/emmeans/index.html.
- 35.Swaminathan H, Rogers HJ. 1990. Detecting differential item functioning using logistic regression procedures. J. Educ. Meas. 27, 361–370. ( 10.1111/j.1745-3984.1990.tb00754.x) [DOI] [Google Scholar]
- 36.Magis D, Béland S, Tuerlinckx F, De Boeck P. 2010. A general framework and an R package for the detection of dichotomous differential item functioning. Behav. Res. Methods 42, 847–862. ( 10.3758/BRM.42.3.847) [DOI] [PubMed] [Google Scholar]
- 37.Kim J, Oshima TC. 2013. Effect of multiple testing adjustment in differential item functioning detection. Educ. Psychol. Meas. 73, 458–470. ( 10.1177/0013164412467033) [DOI] [Google Scholar]
- 38.Revelle W. 2018. psych: Procedures for Personality and Psychological Research, Northwestern University, Evanston, IL. See https://CRAN.Rproject.org/package=psych.
- 39.Brookman-Byrne A, Mareschal D, Tolmie AK, Dumontheil I. 2018. Inhibitory control and counterintuitive science and maths reasoning in adolescence. PLoS ONE 13, e0198973 ( 10.1371/journal.pone.0198973) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Hallquist MN, Geier CF, Luna B. 2018. Incentives facilitate developmental improvement in inhibitory control by modulating control-related networks. Neuroimage 172, 369–380. ( 10.1016/j.neuroimage.2018.01.045) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Karabatsos G. 2003. Comparing the aberrant response detection performance of thirty-six person-fit statistics. Appl. Meas. Educ. 16, 277–298. ( 10.1207/S15324818AME1604_2) [DOI] [Google Scholar]
- 42.Martinková P, Drabinová A, Liaw Y-L, Sanders EA, McFarland JL, Price RM. 2017. Checking equity: why differential item functioning analysis should be a routine part of developing conceptual assessments. CBE Life Sci. Educ. 16, rm2 ( 10.1187/cbe.16-10-0307) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Jameson KA, Highnote SM, Wasserman LM. 2001. Richer color experience in observers with multiple photopigment opsin genes. Psychon. Bull. Rev. 8, 244–261. ( 10.3758/BF03196159) [DOI] [PubMed] [Google Scholar]
- 44.Birch J. 2012. Worldwide prevalence of red-green color deficiency. J. Opt. Soc. Am. A 29, 313–320. ( 10.1364/JOSAA.29.000313) [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
Working towards an open science framework, the study in question provides 80 items (each in three variants) for a novel matrix reasoning item bank, a ready-to-use online implementation and descriptive data from 659 participants aged between 11 and 33 years. The manuscript is already available in preprint (https://osf.io/uvteh/). In an open repository (https://osf.io/g96f4/), we also provide the novel items and item-level data. Owing to the limitations of the ethical approvals that were granted some years in advance of this research being conducted, the authors regret that the full datasets cannot currently be made publicly available. To do so would risk revealing data that may contribute to individuals in the study (currently anonymized) becoming identifiable. Specifically, we have not made available the participant-level data. The participant-level data (and code) will be made available upon reasonable request.