Abstract
A growing research literature suggests that regular electronic game play and game-based training programs may confer practically significant benefits to cognitive functioning. Most evidence supporting this idea, the gaming-enhancement hypothesis, has been collected in small-scale studies of university students and older adults. This research investigated the hypothesis in a general way with a large sample of 1,847 school-aged children. Our aim was to examine the relations between young people’s gaming experiences and an objective test of reasoning performance. Using a Bayesian hypothesis testing approach, evidence for the gaming-enhancement and null hypotheses were compared. Results provided no substantive evidence supporting the idea that having preference for or regularly playing commercially available games was positively associated with reasoning ability. Evidence ranged from equivocal to very strong in support for the null hypothesis over what was predicted. The discussion focuses on the value of Bayesian hypothesis testing for investigating electronic gaming effects, the importance of open science practices, and pre-registered designs to improve the quality of future work.
Keywords: Electronic games, Cognitive ability, Human–computer interaction, Individual differences, Adolescents, Gaming-enhancement hypothesis
Introduction
Background
Electronic games are now a ubiquitous form of entertainment and it is popularly believed that the time spent playing games might have positive benefits that extend outside of gaming contexts (Lenhart et al., 2008; McGonigal, 2012). This general idea, the gaming-enhancement hypothesis, posits that electronic gaming contexts influence a range of perceptual and cognitive abilities because they present complex, dynamic, and demanding challenges. A recent representative study of British adults and young people suggests this idea is widely held; Nine in ten think brain-training games provide an effective way to improve focus, memory, and concentration, and as many as two thirds have tried using these games to improve their own cognitive abilities (Clemence et al., 2013). This view is increasingly profitable and controversial (Nuechterlein et al., 2016). Given the intense public and private interest and investment in games as a way of improving cognition and reasoning, is noteworthy that the scientific literature investigating the gaming-enhancement hypothesis, though promising, is still at an early stage.
A growing body of research suggests that common varieties of electronic gaming experience might enhance general cognitive skills and abilities. These studies show that those who opt to, or are assigned to, play a range of commercially available games show measurable differences in terms of their visual and spatial abilities (Quaiser-Pohl, Geiser & Lehmann, 2006; Green & Bavelier, 2006), executive functioning, information processing (Maillot, Perrot & Hartley, 2012), and performance at specialist skills such as laparoscopic surgery (Rosser et al., 2007). In particular, action games (Green & Bavelier, 2006), strategy games (Basak et al., 2008), and multiplayer online games (Whitlock, McLaughlin & Allaire, 2012), have been identified as having enhancing effects because they provide complex multi-tasking environments that require players to integrate a range of sensory inputs. Researchers argue these environments lead players to implement adaptive strategies to meet the varied demands of these virtual contexts (Bavelier et al., 2012).
There is good reason to think that predispositions to engage specific kinds of games may relate to cognitive performance and reasoning more broadly. For example, in studying skill acquisition among strategy game players, researchers have reported evidence that differences in brain volume are correlated with speed and performance in gaming contexts (Basak et al., 2011). Exploration and learning in online gaming contexts closely mirror their offline analogues. Those with pre-existing strengths tend to thrive at online gaming challenges initially, but those with low levels of starting ability quickly close the performance gap (Stafford & Dewar, 2013). It is possible that such inclinations guide players to games that suit them. Findings from experimental and quasi-experimental studies suggest both preference and experience matter in terms of small to moderate effects across a wide range of cognitive performance indicators (Powers et al., 2013).
Unfortunately, many of these studies have pronounced shortcomings that temper the broad promise of the gaming-enhancement hypothesis (Van Ravenzwaaij et al., 2014). For example, most of the evidence supporting this view has been derived from small-scale surveys, intervention studies in university settings, or from samples of older adults. A recent review of this literature indicates a small yet consistent link between gaming and general reasoning ability (Cohen’s d = 0.24), but the average sample size of studies examining the gaming-enhancement hypothesis is only 32 participants (Powers et al., 2013). Perhaps as a consequence of such small samples, the effects sizes reportedly linking games to cognitive abilities vary widely as a function of the methods used and the research groups investigating them. For example, studies published in top-tier journals, while typically using very small samples, report substantially larger effects compared to the rest of the literature (e.g., Cohen’s d = 0.85; Powers et al., 2013). Moreover, despite the fact that games are played by the overwhelming majority of young people (Lenhart et al., 2015), fewer than one in ten studies of the gaming enhancement study have included participants under the age of 18. Childhood and adolescence see profound development in cognitive abilities, and though negative effects of games are fiercely debated (Mills, 2014; Bell, Bishop & Przybylski, 2015), there is very little evidence concerning their possible positive effects in this cohort. Further, nearly all studies examining gaming effects do so by studying individual games or game types in isolation. So although there is reason to think that action (Green & Bavelier, 2006), strategy (Basak et al., 2011), and online game play (Whitlock, McLaughlin & Allaire, 2012) could have positive effects it is not possible to know if attitudes, preferences, or engagement with games in general, or specific subtypes in particular, are driving the effects reported. Finally, there are a number of larger-scale intervention studies that show evidence that does not support or directly contradicts the gaming-enhancement hypothesis (Chisholm et al., 2010; Kennedy et al., 2011). Given the nature of the existing evidence, research that systematically addresses these gaps in our knowledge is needed.
Present research
The aim of the present research was to evaluate the extent to which everyday electronic game engagement by young people relates to their general reasoning abilities. In particular, we were interested to test how personal preferences for specific types of games related to reasoning ability. In line with previous research, we hypothesized that those who gravitate towards action, online, and strategy games would show higher levels of cognitive performance on a test of deductive reasoning ability (Basak et al., 2011). Our second goal was to test whether regular active engagement with action, online, and strategy games was linked to cognitive ability. In line with highly cited work in the area, we hypothesized that those playing these challenging games for more than five hours each week (Green & Bavelier, 2006) would show higher levels of reasoning ability.
Method
Data source
The present study utilized data collected in the first year of the Effects of Digital Gaming on Children and Teenagers in Singapore (EDGCTS) project. This dataset has been used in numerous previous publications focusing on the effects of gaming on motivation, dysregulated behavior, and interpersonal aggression in young people (Wan & Chiou, 2006; Gentile et al., 2009; Wang, Liu & Khoo, 2009; Chen et al., 2009; Choo et al., 2010; Gentile et al., 2011; Wang et al., 2011; Li, Liau & Khoo, 2011; Chng et al., 2014; Prot et al., 2014; Gentile et al., 2014; Busching et al., 2015; Eichenbaum & Kattner, 2015; Chng et al., 2015; Liau et al., 2015b; Choo et al., 2015; Liau et al., 2015a). A subsample of data from the EDGCTS project was used for the present study because it included self-reports of game play and an objective test of participants’ reasoning abilities. Neither of these variables has been the focus of previously published studies, and a complete list of publications based on the EDGCTS dataset can be found on the Open Science Framework (osf.io/je786).
Participants, ethics, and data
Ethical clearance for data collection was granted through the Institutional Review Board of Nanyang Technological University in Singapore. Because of the combined length of the EDGCTS testing protocol, data collection was conducted over a four-day period to reduce participant burden and minimize sequence effects. The research was deemed low risk and consent was obtained from parents through liaison teachers. Participants were informed that involvement in the project was voluntary and that they could withdraw at any time. Ethical review for this secondary data analysis was conducted by the research ethics committee at the University of Oxford (SSH/OII/C1A-16-063).
In the first wave of the EDGCTS, quantitative data were collected from a total of 3,034 respondents. Gender information was present for 3,012 participants (99.3% of all cases), age data were available for 2,813 participants (92.7%), 2,135 participants provided the name of at least one electronic game they played (70.4%), and a total of 2,647 participants completed the reasoning test (87.2%). In sum, a total of 1,847 participants (60.9% of all cases), provided valid data and were included in subsequent analyses. These 1,847 participants ranged in age from 8 to 16 years (M = 10.97, SD = 1.99); 430 of these identified as female and 1,417 identified as male. The self-report materials, datasets, source code, and analysis code used in this study are available from the Open Science Framework (osf.io/je786).
Outcome variable
Reasoning ability
Participants’ deductive reasoning ability was assessed using the 60-item Raven’s Standard Progressive Matrices Plus (RPM) task (Raven, Raven & Court, 2003). The RPM, a widely used non-verbal test of reasoning ability, measures deductive intelligence by prompting takers to identify the key missing visual element that completes patterns shown in a series of increasingly complex 2 × 2, 2 × 3, 3 × 3, 4 × 4, and 6 × 6, matrices. This assessment was used because it has been well validated across a range of demographic and cross-cultural cohorts (Raven, 2000). Because our participants were school-aged children, they completed the version of this multiple-choice test designed for group administration in educational settings for students between the ages of 8 and 16 years. Participants got a median of 29 of 60 matrices (SD = 4.44) correct and age-adjusted reasoning scores were created for each participant in line with best-practices (Savage-McGlynn, 2012). To this end, participants were segregated by age and their raw performance scores were transformed into z-scores such that their performance was standardized with respect to other children their age.
Explanatory variables
Participants’ electronic gaming was assessed through a series of questions asking about the games they frequently played. Participants were requested to provide the names of up to three games they played as well as an estimate of the amount of time they spend playing each. A total of 446 (24.1%) participants named a single game, 485 (26.3%) participants named two games, and 916 (49.6%) named three games. The titles of named games were content coded to mark if they belonged to one of the three game categories of interest: action games (e.g., Call of Duty, Halo), multiplayer online games (e.g., Maple Story, World of Warcraft), or strategy games (e.g., SimCity, StarCraft).
Game preference
Preference scores were created for each participant using the game names provided through self-report. If one or more of a participant’s named games belonged to the action, online, or strategy types, they were marked as expressing a preference for this kind of game (coded 1); if not, they were counted as not having a preference for this game type (coded 0). This meant three preference scores, one for each game type, were computed for each participant.
Regular play
Data from game preference scores and participants’ self-reported play behavior were used to determine if participants were regular players of specific game types. Scores were created for participants by combing information about the kinds of games they expressed preferences for and their self-reported amounts of weekly play. Amounts of weekly play were computed summing participant estimates of weekday engagement, multiplied by five, and estimates of weekend day engagement multiplied by two. Codes for game types were then used to create one game engagement score for each type of game. In line with the approach prescribed in previous research (Green & Bavelier, 2006), participants were considered regular active players of a game type if they invested five or more hours in a given game type in a week (coded 1), and were coded 0 if they did not spend any time playing this game type in a typical week. Table 1 presents summary statistics for participant game preferences and proportions of active players of each game type.
Table 1. Participant game preferences and engagement.
Game preference | Regular play | |||||
---|---|---|---|---|---|---|
Action | Strategy | Online | Action | Strategy | Online | |
Males | 21.7% | 25.1% | 45.1% | 13.0% | 17.2% | 33.7% |
Females | 2.3% | 11.4% | 52.8% | 1.9% | 7.4% | 41.8% |
Total | 17.2% | 21.9% | 47.2% | 10.4% | 14.9% | 34.1% |
Results
Preliminary analyses
Zero-order bivariate correlations between observed variables are presented in Table 2. Because 36 correlations were conducted, we adjusted our p value threshold for rejecting the null hypothesis from 0.05 to 0.0014 (Holm, 1979). Analyses indicated that older participants were more likely to report higher levels of engagement with games as compared to younger ones (rs = .130–.278). Similarly, female participants were less likely to engage action and strategy games (rs = .128–.217). Gender was not significantly related to online game play nor was it related to age-adjusted reasoning ability (all ps > 0.0014). Because gender was not associated with our target outcome, which was centered on age, neither age nor gender were considered as covariates in hypothesis testing.
Table 2. Observed zero-order correlations.
2. | 3. | 4. | 5. | 6. | 7. | 8. | 9. | ||
---|---|---|---|---|---|---|---|---|---|
1. Age | Pearson’s r | −0.070 | 0.132* | 0.175* | 0.130* | 0.125* | 0.215* | 0.198* | −0.000 |
2. Female | Pearson’s r | — | −0.217* | −0.140* | 0.062 | −0.167* | −0.126* | 0.040 | −0.029 |
3. Action game preference | Pearson’s r | — | 0.078* | −0.123* | 0.943* | 0.067 | −0.111* | −0.053 | |
4. Strategy game preference | Pearson’s r | — | −0.061 | 0.052 | 0.959* | −0.060 | 0.041 | ||
5. Online game preference | Pearson’s r | — | −0.093* | −0.046 | 0.946* | 0.051 | |||
6. Regular action game play | Pearson’s r | — | 0.090* | −0.055 | −0.028 | ||||
7. Regular strategy game play | Pearson’s r | — | −0.019 | 0.038 | |||||
8. Regular online game play | Pearson’s r | — | 0.020 | ||||||
9. Deductive reasoning ability | Pearson’s r | — |
Notes.
Denotes p value below Bonferroni-Holm adjusted value of 0.0014.
Game preference and reasoning ability
In line with meta-analytic evidence a series of one-sided Bayesian independent samples t-tests were used to quantify evidence for the game-enhancement hypothesis (Rouder et al., 2012; Powers et al., 2013; Morey, Romeijn & Rouder, 2016), that specified preference for action, online, and strategy games would be related to better deductive reasoning ability. Table 3 presents a summary of these results and observed means using a Cauchy prior of 0.24, effect size for quasi-experiments on measures of reasoning and intelligence, and a second prior effect size for the enhancement hypothesis as reported in top-tier journals (Cohen’s d of 0.85; Powers et al., 2013). Results provided very strong support for the null hypothesis over the alternative for action games (BF10 = 0.06; M0 = 0.02, SD0 = 1.00, M1 = − 0.12, SD1 = 0.97) and equivocal support for alternative hypothesis for those who preferred strategy games (BF10 = 1.45; M0 = − 0.02, SD0 = 1.02, M1 = 0.08, SD1 = 0.93), or online multiplayer games (BF10 = 2.815; M0 = − 0.5, SD0 = 1.04, M1 = 0.05, SD1 = 0.95). In examining the robustness of these Bayes factors it is clear that evidence for effects larger than the average in the literature are also not supported. Results derived using the larger effect sizes reported in top-tier journals for the enhancement effect, d = 0.85, appeared less likely as evidence was 1.25 and 57.01 times more likely to have been observed under the null hypothesis than under the gaming-enhancement hypothesis (see Table 3). Figure 1A–1C, present relative evidence for the gaming enhancement hypothesis for effect sizes ranging from a 0.0 to 1.5. Taken together with the focused hypothesis tests, these results indicated participants’ preferences for these specific types of games were not reliably linked to their deductive reasoning ability in these data.
Table 3. Evidence from Bayesian hypothesis testing.
Participants who did not express preference or play game type | Participants who did express preference or play game type | Average Enhancement Effecta | Enhancement Effect in Top Tier Journalsb | |||||||
---|---|---|---|---|---|---|---|---|---|---|
Count | Mean | SD | Count | Mean | SD | BF01 | BF10 | BF01 | BF10 | |
Action game preference | 1,530 | 0.02 | 1.00 | 317 | −0.12 | 0.97 | 16.38 | 0.06 | 57.01 | 0.02 |
Strategy game preference | 1,442 | −0.02 | 1.02 | 405 | 0.08 | 0.93 | 0.69 | 1.45 | 2.06 | 0.48 |
Online game preference | 975 | −0.05 | 1.04 | 872 | 0.05 | 0.95 | 0.36 | 2.82 | 1.07 | 0.94 |
Regular action game play | 1,551 | 0.02 | 1.00 | 192 | −0.07 | 0.99 | 8.61 | 0.12 | 29.15 | 0.03 |
Regular strategy game play | 1,462 | −0.02 | 1.02 | 276 | 0.08 | 0.90 | 0.84 | 1.20 | 2.45 | 0.41 |
Regular online game play | 1,019 | −0.04 | 1.02 | 630 | 0.01 | 0.97 | 2.97 | 0.03 | 9.84 | 0.10 |
Notes.
BF01 denotes evidence favoring the Null hypothesis. BF10 denotes evidence favoring the alternative hypothesis.
Average effect size for quasi-experiments on measures of reasoning and intelligence (d = 0.24; Powers et al., 2013).
Average effect size for quasi-experiments on measures of general cognitive abilities published in top-tier journals (d = 0.85; Powers et al., 2013).
Regular game play and reasoning ability
To examine the relations between regular action, strategy, or online game play and reasoning ability, three additional Bayesian hypothesis tests were conducted following the procedure used for game preferences. These models evaluated the relative evidence for the null hypothesis as well as the gaming-enhancement hypothesis that postulates that playing these games for more than five hours each week would be associated with players’ reasoning abilities. Results provided moderate support for the null over the alternative hypothesis for action games (BF10 = 0.12; M0 = 0.02, SD0 = 1.00, M1 = − 0.07, SD1 = 0.99), equivocal evidence for strategy games (BF10 = 1.120; M0 = − 0.023, SD0 = 1.02, M1 = 0.08, SD1 = 0.90), and equivocal to moderate evidence for the null over the alternative hypothesis for multiplayer online games (BF10 = 0.34; M0 = − 0.4, SD0 = 1.04, M1 = 0.01, SD1 = 0.97). Robustness checks using effect size for the gaming enhancement hypothesis as reported in top-tier journals (d = 0.85; Powers et al., 2013) indicated evidence against the this hypothesis. Data were between 2.45 and 29.15 times more likely to be observed under the null hypothesis. Figure 1D–1F present relative evidence for the gaming enhancement hypothesis for effect sizes ranging from of 0.0 to 1.5. Taken together with the previous results, these findings suggest it is unlikely that regular active play of these games is systematically related to higher levels of general reasoning abilities.
Discussion
Overview
The promise that electronic games might positively influence human cognition is one that generates intense public, corporate, and scientific interest. The present research drew on a large sample of school-aged children and considered both their electronic gaming and cognitive performance to test of the gaming-enhancement hypothesis. Of central interest was the nature of the potential relations between children’s self-reported preferences and gaming habits, and performance on a widely used test of deductive reasoning ability. Contrary to our expectations, the results did not provide substantial evidence in support of the idea that the complex and interactive experiences provided by commercially available games generalize to functioning outside of gaming contexts. In most cases, the evidence was in favor of the null hypothesis over this account.
We hypothesized that those who express preferences for action, strategy, and multiplayer online games, modes of play would show modestly higher levels of cognitive performance as seen in previous smaller-scale studies (Basak et al., 2011). In contrast to what was expected, we found equivocal to very strong evidence favoring the null hypothesis over this prediction. Second, we hypothesized that regular engagement with games—playing five or more hours a week of action, strategy, and multiplayer online games—would be linked with better reasoning ability. Results from our analyses did not support this prediction. Evidence for regular strategy game players were equivocal, but ranged from moderately to strongly in favor of the null for these types of games. Taken together, our findings disconfirmed the gaming-enhancement hypothesis, especially in terms of the larger effects reported in top-tier journals.
Our approach carries a number of strengths that should inform future studies of gaming effects. First, evidence from this study relied on data provided by more than 1,800 young people, a sample over 50 times larger than the average for studies examining gaming and cognition (Powers et al., 2013). If research is to sort out the cognitive dynamics of play, larger and more robust sampling is needed. Second, the Bayesian hypothesis testing approach we adopted used open source software (JASP; JASP Team, 2016) and allowed our study to quantify evidence for the both the null hypothesis as well as a plausible alternative based on the existing literature. Although the analysis plan for this study was not registered before the data were known, the framework provides valuable empirical data that researchers can use as the basis, or prior, to inform their own pre-registered designs. Finally, if indeed scholars are increasingly skeptical of corporate attempts to sell games based on their purported upsides (e.g., Allaire et al., 2014), this vigor should be extended to all scientific inquiry in this area, for example by making the materials, data, source and analysis code openly available. Future research making both positive and negative claims regarding the effects of gaming on young people should do likewise.
Limitations and future directions
The present study presents a number of limitations that suggest promising avenues for future work. First, because the data under study were cross-sectional they capture an easy to interpret pattern of results that represent a snapshot in time. The data structure did so at the expense of being able to draw causal inferences, and a complementary approach would examine long-term salutary effects on cognition, ideally as a function of experimental manipulations of game exposure. Second, data regarding participants’ game preferences and regular play were collected through self-report. Research indicates that some participants, and young people in particular, provide exaggerated data when it comes to taboo activities such as sexual habits and drug use (Robinson-Cimpian, 2014). It is possible that the average levels of engagement reported by our participants disguised interesting patterns of engagement which merit inquiry. For example, infrequent periods of high engagement (e.g., binge-playing) might have its own special relations with reasoning abilities. If so, an experience-sampling based approach would be needed to assess both between- and within-person variability with respect to the relation between gaming and reasoning ability. Finally, the present study only used a single assessment of general cognitive abilities, the Ravens Progressive Matrices task (Raven, 2000). There are many other facets to intelligence and executive control that might be more sensitive to influence by regular electronic gaming. Measures of naïve reasoning (Masson, Bub & Lalonde, 2011), short and long-term memory (Melby-Lervåg & Hulme, 2013), audio processing (Liu & Holt, 2011) might be more liable to be influenced by gaming. If the gaming-enhancement hypothesis is not broadly accurate, it may find empirical support under conditions where these alternative aspects of intelligence and reasoning abilities are under study.
Closing remarks
The promise that popular games can enhance cognitive skills is an alluring one. Our findings suggest there is no relation between interest in, or regular play of, electronic games and general reasoning ability. As such, we advise that future research examining the potential influences of gaming contexts on players should pre-register their analysis plans or follow the registered reports process (e.g., Chambers, 2013; Elson, Przybylski & Krämer, 2015). Such steps would go a long way to reduce researcher degrees of freedom which might, along with publication bias, affect conclusions drawn about the effects of gaming and cognitive enhancement (Feynman, 1974; Gelman & Loken, 2013; Nissen et al., 2016). While the research presented here might be further informed by additional work conducted to these standards, our findings above offer an early exploration of the gaming-enhancement hypothesis which is well-powered and guided by open-science.
Funding Statement
This research was partially funded by a John Fell Fund Grant (CZD08320) through the University of Oxford to Dr. Przbylski, and a joint grant awarded by the Ministry of Education, Singapore and the Media Development Authority f (EPI/06AK) to Professor Wang. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Additional Information and Declarations
Competing Interests
The authors declare there are no competing interests.
Author Contributions
Andrew K. Przybylski conceived and designed the experiments, performed the experiments, analyzed the data, contributed reagents/materials/analysis tools, wrote the paper, prepared figures and/or tables, reviewed drafts of the paper.
John C. Wang conceived and designed the experiments, performed the experiments, contributed reagents/materials/analysis tools, wrote the paper, reviewed drafts of the paper.
Ethics
The following information was supplied relating to ethical approvals (i.e., approving body and any reference numbers):
Ethical clearance for the Effects of Digital Gaming on Children and Teenagers in Singapore project (EDGCTS) was sought and granted through the Institutional Review Board of Nanyang Technological University in Singapore. The Research and Ethics Committee at the Oxford Internet Institute conducted ethical review for secondary data analysis on the EDGCTS dataset (SSH/OII/C1A-16-063).
Data Availability
The following information was supplied regarding data availability:
All study materials and data are available for download using the Open Science Framework: https://osf.io/je786/.
References
- Allaire et al. (2014).Allaire JC, Bäckman L, Balota DA, Bavelier D, Bjork RA, Bower GH. Max planck institute for human development and stanford center on longevity. 2014. [15 January 2015]. A consensus on the brain training industry from the scientific community. [Google Scholar]
- Basak et al. (2008).Basak C, Boot WR, Voss MW, Kramer AF. Can training in a real-time strategy video game attenuate cognitive decline in older adults? Psychology and Aging. 2008;23:765–777. doi: 10.1037/a0013494. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Basak et al. (2011).Basak C, Voss MW, Erickson KI, Boot WR, Kramer AF. Regional differences in brain volume predict the acquisition of skill in a complex real-time strategy videogame. Brain and Cognition. 2011;76:407–414. doi: 10.1016/j.bandc.2011.03.017. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bavelier et al. (2012).Bavelier D, Green CS, Pouget A, Schrater P. Brain plasticity through the life span: learning to learn and action video games. Annual Review of Neuroscience. 2012;35:391–416. doi: 10.1146/annurev-neuro-060909-152832. [DOI] [PubMed] [Google Scholar]
- Bell, Bishop & Przybylski (2015).Bell V, Bishop DVM, Przybylski AK. The debate over digital technology and young people. BMJ. 2015;351:h3064. doi: 10.1136/bmj.h3064. [DOI] [PubMed] [Google Scholar]
- Busching et al. (2015).Busching R, Gentile DA, Krahé B, Möller I, Khoo A, Walsh DA, Anderson CA. Testing the reliability and validity of different measures of violent video game use in the United States, Singapore, and Germany. Psychology of Popular Media Culture. 2015;4:97–111. doi: 10.1037/ppm0000004. [DOI] [Google Scholar]
- Chambers (2013).Chambers CD. Registered reports: a new publishing initiative at Cortex. Cortex. 2013;49:609–610. doi: 10.1016/j.cortex.2012.12.016. [DOI] [PubMed] [Google Scholar]
- Chen et al. (2009).Chen VH-H, Lin W, Ng CW, Chai SL, Khoo ACE, Duh HB-L. Children’s choice of games: the influence of prosocial tendency and education-level. In: Natkin S, Dupire J, editors. Entertainment Computing—ICEC 2009. Springer; Berlin Heidelberg: 2009. pp. 110–119. (Lecture Notes in Computer Science). [Google Scholar]
- Chisholm et al. (2010).Chisholm JD, Hickey C, Theeuwes J, Kingstone A. Reduced attentional capture in action video game players. Attention, Perception & Psychophysics. 2010;72:667–671. doi: 10.3758/APP.72.3.667. [DOI] [PubMed] [Google Scholar]
- Chng et al. (2015).Chng GS, Li D, Liau AK, Khoo A. Moderating effects of the family environment for parental mediation and pathological internet use in youths. Cyberpsychology, Behavior, and Social Networking. 2015;18:30–36. doi: 10.1089/cyber.2014.0368. [DOI] [PubMed] [Google Scholar]
- Chng et al. (2014).Chng GS, Liau A, Khoo A, Li D. Parental mediation and cyberbullying—a longitudinal study. Studies in Health Technology and Informatics. 2014;199:98–102. [PubMed] [Google Scholar]
- Choo et al. (2010).Choo H, Gentile DA, Sim T, Li D, Khoo A, Liau AK. Pathological video-gaming among Singaporean youth. Annals of the Academy of Medicine, Singapore. 2010;39:822–829. [PubMed] [Google Scholar]
- Choo et al. (2015).Choo H, Sim T, Liau AKF, Gentile DA, Khoo A. Parental influences on pathological symptoms of video-gaming among children and adolescents: a prospective study. Journal of Child and Family Studies. 2015;24:1429–1441. doi: 10.1007/s10826-014-9949-9. [DOI] [Google Scholar]
- Clemence et al. (2013).Clemence M, Gilby N, Shah J, Swiecicka J, Warren D. Wellcome trust monitor wave 2. Wellcome Trust, Londonhttps://wellcome.ac.uk/sites/default/files/monitor-wave2-full-wellcome-may13.pdf Tracking public views on science, research and science education. 2013
- Eichenbaum & Kattner (2015).Eichenbaum A, Kattner F. The role of game genres and the development of internet gaming disorder in school—aged children. Journal of Addictive Behaviors, Therapy & Rehabilitation. 2015;4:3. doi: 10.4172/2324-9005.1000141. [DOI] [Google Scholar]
- Elson, Przybylski & Krämer (2015).Elson M, Przybylski AK, Krämer N. Technology and human behavior. Journal of Media Psychology. 2015;27:203–204. doi: 10.1027/1864-1105/a000170. [DOI] [Google Scholar]
- Feynman (1974).Feynman RP. Cargo cult science. Engineering and Science. 1974;37:10–13. [Google Scholar]
- Gelman & Loken (2013).Gelman A, Loken E. The garden of forking paths: why multiple comparisons can be a problem, even when there is no fishing expedition or “p-hacking” and the research hypothesis was posited ahead of time. 2013. Department of Statistics, Columbia University, New York.
- Gentile et al. (2009).Gentile DA, Anderson CA, Yukawa S, Ihori N, Saleem M, Lim KM, Shibuya A, Liau AK, Khoo A, Bushman BJ, Rowell Huesmann L, Sakamoto A. The effects of prosocial video games on prosocial behaviors: international evidence from correlational, longitudinal, and experimental studies. Personality and Social Psychology Bulletin. 2009;35:752–763. doi: 10.1177/0146167209333045. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Gentile et al. (2011).Gentile DA, Choo H, Liau A, Sim T, Li D, Fung D, Khoo A. Pathological video game use among youths: a two-year longitudinal study. Pediatrics. 2011;127:450–457. doi: 10.1542/peds.2010-1353. [DOI] [PubMed] [Google Scholar]
- Gentile et al. (2014).Gentile DA, Li D, Khoo A, Prot S, Anderson CA. Mediators and moderators of long-term effects of violent video games on aggressive behavior: practice, thinking, and action. JAMA Pediatrics. 2014;168:450–457. doi: 10.1001/jamapediatrics.2014.63. [DOI] [PubMed] [Google Scholar]
- Green & Bavelier (2006).Green CS, Bavelier D. Effect of action video games on the spatial distribution of visuospatial attention. Journal of experimental psychology. Human perception and performance. 2006;32:1465–1478. doi: 10.1037/0096-1523.32.6.1465. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Holm (1979).Holm S. A simple sequentially rejective multiple test procedure. Scandinavian. Journal of Statistics. 1979;6:65–70. [Google Scholar]
- JASP Team (2016).JASP Team JASP. https://jasp-stats.org 2016
- Kennedy et al. (2011).Kennedy AM, Boyle EM, Traynor O, Walsh T, Hill ADK. Video gaming enhances psychomotor skills but not visuospatial and perceptual abilities in surgical trainees. Journal of Surgical Education. 2011;68:414–420. doi: 10.1016/j.jsurg.2011.03.009. [DOI] [PubMed] [Google Scholar]
- Lenhart et al. (2008).Lenhart A, Kahne J, Middaugh E, Macgill A, Vitak J. Pew Research Center, Washington, D.C.http://www.pewinternet.org/2008/09/16/teens-video-games-and-civics/ Teens, video games and civics. 2008
- Lenhart et al. (2015).Lenhart A, Smith A, Anderson M, Duggan M, Perrin A. Pew Research Center, Washington, D.C.http://www.pewinternet.org/2015/08/06/teens-technology-and-friendships/ Teens, technology and friendships. 2015
- Li, Liau & Khoo (2011).Li D, Liau A, Khoo A. Examining the influence of actual-ideal self-discrepancies, depression, and escapism, on pathological gaming among massively multiplayer online adolescent gamers. Cyberpsychology, Behavior, and Social Networking. 2011;14:535–539. doi: 10.1089/cyber.2010.0463. [DOI] [PubMed] [Google Scholar]
- Liau et al. (2015a).Liau AK, Choo H, Li D, Gentile DA, Sim T, Khoo A. Pathological video-gaming among youth: a prospective study examining dynamic protective factors. Addiction Research & Theory. 2015a;23:301–308. doi: 10.3109/16066359.2014.987759. [DOI] [Google Scholar]
- Liau et al. (2015b).Liau AK, Neo EC, Gentile DA, Choo H, Sim T, Li D, Khoo A. Impulsivity, self-regulation, and pathological video gaming among youth: testing a mediation model. Asia-Pacific Journal of Public Health. 2015b;27:NP2188–NP2196. doi: 10.1177/1010539511429369. [DOI] [PubMed] [Google Scholar]
- Liu & Holt (2011).Liu R, Holt LL. Neural changes associated with nonspeech auditory category learning parallel those of speech category acquisition. Journal of Cognitive Neuroscience. 2011;23:683–698. doi: 10.1162/jocn.2009.21392. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Maillot, Perrot & Hartley (2012).Maillot P, Perrot A, Hartley A. Effects of interactive physical-activity video-game training on physical and cognitive function in older adults. Psychology and Aging. 2012;27:589–600. doi: 10.1037/a0026268. [DOI] [PubMed] [Google Scholar]
- Masson, Bub & Lalonde (2011).Masson MEJ, Bub DN, Lalonde CE. Video-game training and naïve reasoning about object motion. Applied Cognitive Psychology. 2011;25:166–173. doi: 10.1002/acp.1658. [DOI] [Google Scholar]
- McGonigal (2012).McGonigal J. Reality is broken: why games make us better and how they can change the world; [includes practical advice for gamers] Vintage Books; London: 2012. [Google Scholar]
- Melby-Lervåg & Hulme (2013).Melby-Lervåg M, Hulme GC. Is working memory training effective? A meta-analytic review. Developmental Psychology. 2013;49:270–291. doi: 10.1037/a0028228. [DOI] [PubMed] [Google Scholar]
- Mills (2014).Mills KL. Effects of Internet use on the adolescent brain: despite popular claims, experimental evidence remains scarce. Trends in Cognitive Sciences. 2014;18:385–387. doi: 10.1016/j.tics.2014.04.011. [DOI] [PubMed] [Google Scholar]
- Morey, Romeijn & Rouder (2016).Morey RD, Romeijn J-W, Rouder JN. The philosophy of Bayes factors and the quantification of statistical evidence. Journal of Mathematical Psychology. 2016;72:6–18. doi: 10.1016/j.jmp.2015.11.001. [DOI] [Google Scholar]
- Nissen et al. (2016).Nissen SB, Magidson T, Gross K, Bergstrom CT. Publication bias and the canonization of false facts. 2016. 1609.00494 [DOI] [PMC free article] [PubMed]
- Nuechterlein et al. (2016).Nuechterlein J, Rusk M, Soberats A, Johnson M. 2016. Federal trade commission, plaintiff, V. Lumos Labs, Inc. https://www.ftc.gov/system/files/documents/cases/160105lumoslabscmpt.pdf .
- Powers et al. (2013).Powers KL, Brooks PJ, Aldrich NJ, Palladino MA, Alfieri L. Effects of video-game play on information processing: a meta-analytic investigation. Psychonomic Bulletin & Review. 2013;20:1055–1079. doi: 10.3758/s13423-013-0418-z. [DOI] [PubMed] [Google Scholar]
- Prot et al. (2014).Prot S, Gentile DA, Anderson CA, Suzuki K, Swing E, Lim KM, Horiuchi Y, Jelic M, Krahe B, Liuqing W, Liau AK, Khoo A, Petrescu PD, Sakamoto A, Tajima S, Toma RA, Warburton W, Zhang X, Lam BCP. Long-term relations among prosocial-media use, empathy, and prosocial behavior. Psychological Science. 2014;25:358–368. doi: 10.1177/0956797613503854. [DOI] [PubMed] [Google Scholar]
- Quaiser-Pohl, Geiser & Lehmann (2006).Quaiser-Pohl C, Geiser C, Lehmann W. The relationship between computer-game preference, gender, and mental-rotation ability. Personality and Individual Differences. 2006;40:609–619. doi: 10.1016/j.paid.2005.07.015. [DOI] [Google Scholar]
- Raven (2000).Raven J. The raven’s progressive matrices: change and stability over culture and time. Cognitive Psychology. 2000;41:1–48. doi: 10.1006/cogp.1999.0735. [DOI] [PubMed] [Google Scholar]
- Raven, Raven & Court (2003).Raven J, Raven JC, Court JH. Manual for Raven’s progressive matrices and vocabulary scales. Section 1: general overview. Harcourt Assessment; San Antonio: 2003. [Google Scholar]
- Robinson-Cimpian (2014).Robinson-Cimpian JP. Inaccurate estimation of disparities due to mischievous responders several suggestions to assess conclusions. Educational Researcher. 2014;43:171–185. doi: 10.3102/0013189X14534297. [DOI] [Google Scholar]
- Rosser et al. (2007).Rosser JC, Lynch PJ, Cuddihy L, Gentile DA, Klonsky J, Merrell R. The impact of video games on training surgeons in the 21st century. Archives of Surgery. 2007;142:181–186. doi: 10.1001/archsurg.142.2.181. [DOI] [PubMed] [Google Scholar]
- Rouder et al. (2012).Rouder JN, Morey RD, Speckman PL, Province JM. Default Bayes factors for ANOVA designs. Journal of Mathematical Psychology. 2012;56:356–374. doi: 10.1016/j.jmp.2012.08.001. [DOI] [Google Scholar]
- Savage-McGlynn (2012).Savage-McGlynn E. Sex differences in intelligence in younger and older participants of the Raven’s Standard Progressive Matrices Plus. Personality and Individual Differences. 2012;53:137–141. doi: 10.1016/j.paid.2011.06.013. [DOI] [Google Scholar]
- Stafford & Dewar (2013).Stafford T, Dewar M. Tracing the trajectory of Skill learning with a very large sample of online game players. Psychological Science. 2013;25(2):511–518. doi: 10.1177/0956797613511466. [DOI] [PubMed] [Google Scholar]
- Van Ravenzwaaij et al. (2014).Van Ravenzwaaij D, Boekel W, Forstmann BU, Ratcliff R, Wagenmakers E-J. Action video games do not improve the speed of information processing in simple perceptual tasks. Journal of Experimental Psychology. General. 2014;143:1794–1805. doi: 10.1037/a0036923. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wan & Chiou (2006).Wan C-S, Chiou W-B. Psychological motives and online games addiction: a test of flow theory and humanistic needs theory for taiwanese adolescents. CyberPsychology & Behavior. 2006;9:317–324. doi: 10.1089/cpb.2006.9.317. [DOI] [PubMed] [Google Scholar]
- Wang et al. (2011).Wang CKJ, Liu WC, Chye S, Chatzisarantis NLD. Understanding motivation in internet gaming among Singaporean youth: the role of passion. Computers in Human Behavior. 2011;27:1179–1184. doi: 10.1016/j.chb.2010.12.012. [DOI] [Google Scholar]
- Wang, Liu & Khoo (2009).Wang CKJ, Liu WC, Khoo A. The psychometric properties of dispositional flow scale-2 in internet gaming. Current Psychology. 2009;28:194–201. doi: 10.1007/s12144-009-9058-x. [DOI] [Google Scholar]
- Whitlock, McLaughlin & Allaire (2012).Whitlock LA, McLaughlin AC, Allaire JC. Individual differences in response to cognitive training: using a multi-modal, attentionally demanding game-based intervention for older adults. Computers in Human Behavior. 2012;28:1091–1096. doi: 10.1016/j.chb.2012.01.012. [DOI] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
The following information was supplied regarding data availability:
All study materials and data are available for download using the Open Science Framework: https://osf.io/je786/.