Skip to main content
PLOS ONE logoLink to PLOS ONE
. 2020 Oct 26;15(10):e0240032. doi: 10.1371/journal.pone.0240032

Evaluating the quality of evidence for gaming disorder: A summary of systematic reviews of associations between gaming disorder and depression or anxiety

Michelle Colder Carras 1,¤,*, Jing Shi 2,3, Gregory Hard 4, Ian J Saldanha 5
Editor: Florian Naudet6
PMCID: PMC7588081  PMID: 33104730

Abstract

Gaming disorder has been described as an urgent public health problem and has garnered many systematic reviews of its associations with other health conditions. However, review methodology can contribute to bias in the conclusions, leading to research, policy, and patient care that are not truly evidence-based. This study followed a pre-registered protocol (PROSPERO 2018 CRD42018090651) with the objective of identifying reliable and methodologically-rigorous systematic reviews that examine the associations between gaming disorder and depression or anxiety in any population. We searched PubMed and PsycInfo for published systematic reviews and the gray literature for unpublished systematic reviews as of June 24, 2020. Reviews were classified as reliable according to several quality criteria, such as whether they conducted a risk of bias assessment of studies and whether they clearly described how outcomes from each study were selected. We assessed possible selective outcome reporting among the reviews. Seven reviews that included a total of 196 studies met inclusion criteria. The overall number of participants was not calculable because not all reviews reported these data. All reviews specified eligibility criteria for studies, but not for outcomes within studies. Only one review assessed risk of bias. Evidence of selective outcome reporting was found in all reviews—only one review incorporated any of the null findings from studies it included. Thus, none were classified as reliable according to prespecified quality criteria. Systematic reviews related to gaming disorder do not meet methodological standards. As clinical and policy decisions are heavily reliant on reliable, accurate, and unbiased evidence synthesis; researchers, clinicians, and policymakers should consider the implications of selective outcome reporting. Limitations of the current summary include using counts of associations and restricting to systematic reviews published in English. Systematic reviewers should follow established guidelines for review conduct and transparent reporting to ensure evidence about technology use disorders is reliable.

Introduction

Gaming disorder or Internet gaming disorder (IGD) is a disorder related to excessive video, computer, or online game play that results in psychological distress and/or functional impairment [1,2]. Internet gaming disorder was included as a condition for further research in the 5th edition of the Diagnostic and Statistical Manual (DSM-5) and the diagnosis of gaming disorder has been added to the 11th edition of the World Health Organization (WHO) International Classification of Diseases (ICD-11) [1,2]. Gaming disorder includes symptoms related to substance use disorder, such as loss of control (that continues despite negative consequences), functional impairment, distress, and/or interference with daily activities. The disorder is distinguished from other related disorders, such as technology overuse, Internet addiction, and social networking addiction [3]. Recent commentaries have described gaming disorder (which we will define here broadly to include the diagnoses of IGD or gaming disorder, problematic/pathological video gaming, and other concepts related to excessive video game play) as a clinical and public health problem in urgent need of advancements in treatment development [4,5].

Delineation and measurement of a clear construct with no overlap with other related conditions, such as gambling, Internet use, and technology use, are crucial to this field. Many recent commentaries on the need for a diagnosis of gaming disorder use terms like “Internet addiction or Gaming disorder” [6], “Internet-related disorders including gaming disorder” [4], and “Internet addiction including gaming addiction” [5], pointing to the persistent overlap in measurement of these problematic behaviors. From a public health perspective, many forms of Internet use—not just gaming—continue to be recognized as potentially problematic, as evidenced by a recently-funded international research collaborative on problematic Internet use [7].

Systematic reviews are research activities that follow established, rigorous methods to summarize all relevant evidence on specific research questions that are vital for decision-making by clinicians, patients, policy-makers, and other stakeholders. The methods include framing the research question, searching for the evidence, screening studies for eligibility, assessing risk of bias and extracting data from included studies, conducting qualitative and, where merited, quantitative syntheses, and reporting the findings. Recent decades have witnessed a surge in the number of systematic reviews conducted [8]. Multiple standards have been developed for the conduct and reporting of systematic reviews [9]. However, research has shown that reviews in some fields provide low-quality evidence, are unreliable, and can be sources of bias themselves [8,10,11]. Bias can sometimes be introduced due to methods used in the systematic review (‘meta-bias’) [12].

One source of meta-bias can potentially occur when a given study included in a review reports results for a given relevant outcome in multiple ways, and the reviewer must make a choice among these to determine which result(s) to extract for the review [13,14]. In this situation, choice of the result based on the largest (or smallest) magnitude of treatment effect, on statistical significance, and/or on the result that supports the reviewer’s conscious or subconscious preconceptions can be problematic and lead to bias. Such bias can be preempted by completely prespecifying the five elements of an outcome (Fig 1) [10,15]. However, complete prespecification is not always possible and/or may be considered too restrictive. Moreover, choosing specific results from multiple reported analyses from multiple data sources for a given study is a multi-dimensional problem. In one study of meta-analytic methods, an examination of outcomes reported in 14 clinical trials revealed that, depending on which outcomes from the trials were chosen by the reviewers, over 34 trillion meta-analyses were possible [13].

Fig 1. Defining outcomes for a systematic review or meta-analysis.

Fig 1

Elements of outcome domains required for complete outcome specification in health research. Figure adapted from [15]; see also the PRISMA-P [16] statement or description of PICOS [15].

Now that gaming disorder has been recognized as a disorder by the WHO, ensuring systematic and accurate measurement of gaming disorder in studies and accurate reporting of exposures, outcomes, and conclusions in reviews are vital for ongoing decision-making regarding diagnosis, treatment, and public health interventions. Given the established association between gaming disorder and two common mental health outcomes—depression and anxiety—we limited the scope of our study to systematic reviews that included data about these outcomes. This allowed us to explore the issue of selective outcome reporting in reviews.

In this summary of systematic reviews, we assess the reliability of current reviews that have examined the association between gaming disorder and depression or between gaming disorder and anxiety in any population. We aimed to answer the following research questions to inform directions for future research and policymaking:

  1. Do systematic reviews of the associations between gaming disorder and depression and between gaming disorder and anxiety meet reliability standards for systematic reviews?

  2. Do systematic reviews of the associations between gaming disorder and depression and between gaming disorder and anxiety distinguish between gaming disorder and other constructs, such as Internet addiction?

  3. Do systematic reviews of the associations between gaming disorder and depression and between gaming disorder and anxiety report outcomes selectively?

  4. What are the associations between gaming disorder and depression and between gaming disorder and anxiety reported in reliable systematic reviews?

Methods

This study is a summary of systematic reviews of the associations between gaming disorder and depression and between gaming disorder and anxiety in any population. The review methods, including the research question, search strategy, inclusion/exclusion criteria, and risk of bias assessment, were developed a priori and described in the registered protocol (PROSPERO 2018 CRD42018090651); these are also available in S1 Protocol. All data, the protocol, a list of articles excluded at the full-text screening stage with reasons for exclusion, and other supporting documentation are available on our Open Science Framework website (see Project on OSF website) and in Supporting Information files. In this paper, we discuss two groups of research studies: the systematic reviews (henceforth called ‘reviews’) and the primary studies included in those reviews (henceforth called ‘studies’).

We examined reviews that included studies of the associations between the exposure of gaming disorder (as defined by the review authors) and the outcomes of depression or anxiety. We restricted to reviews published in English by June 24, 2020. We excluded reviews that:

  • Were narrative reviews, overviews of reviews, commentaries, and other non-systematic reviews of studies;

  • Only examined Internet addiction or other technological addiction; or

  • Did not report results for the associations between gaming disorder and depression or anxiety separately (e.g., we excluded reviews that only reported pooled outcomes for "mental health").

Fig 2 illustrates how we defined the domains of depression and anxiety in our study. For the outcome of depression, we restricted to scales, subscales, diagnosis, or clinical interviews for depression or more severe single symptoms related to depression, such as suicidal ideation, but excluded measurements of nonspecific symptoms, such as low energy, sleep problems, sadness, or withdrawal from social activities. For the outcome of anxiety, we included scales, subscales, diagnosis, or clinical interviews for anxiety, social anxiety, and social phobia, but excluded measurements that combined anxiety with other constructs (e.g., anxiety/depression).

Fig 2. Domains used to define depression and anxiety as constructs for analysis.

Fig 2

YSR = Youth Self Report scale.

Search strategy and screening process

We conducted electronic searches of PubMed and PsycInfo for published reviews and meta-analyses (searches were current as of June 24, 2020). Searches combined terms related to gaming disorder and terms related to depression or anxiety (S1 Search Strategies). In addition, we reviewed all years of the Journal of Behavioral Addictions, including its supplements, and all proceedings of the International Conference on Behavioral Addictions.

Assessment of reliability of reviews

We adapted the definition of “reliability” of systematic reviews developed by Cochrane Eyes and Vision [1721]. This definition, in turn, was informed by items identified from the Critical Appraisal Skills Programme (CASP), A Measurement Tool to Assess systematic Reviews (AMSTAR), and the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) tools [9,22,23]. According to this definition, a review is reliable when its authors did each of the following:

  • (1)

    Defined eligibility criteria for including studies;

  • (2)

    Conducted a comprehensive literature search for studies (i.e., searched at least one relevant electronic database, such as PubMed and PsycInfo; used at least one other method of searching, such as searching the grey literature, searching for unpublished studies, and searching the reference lists of included articles; and were not limited to English language citations);

  • (3)

    Assessed risk of bias in individual included studies;

  • (4)

    Used appropriate methods for meta-analysis, when conducted (e.g., adequately accounting for any heterogeneity); and

  • (5)

    Presented conclusions that were supported by the evidence reported in the review.

Because we also examined each study included in the reviews, we added an additional criterion that review authors should have:

  • (6)

    Specified in the methods or protocol which outcomes from their eligible studies were included in the synthesis or synthesized all reported outcomes from each included study.

We classified a review as reliable only if all six of the criteria were met. Finally, we conducted a full assessment of the quality of the included reviews using A Measurement Tool to Assess systematic Reviews—version 2 (AMSTAR 2) [24]; the full results of this assessment are found in S1 Data Extraction.

Assessment of other outcomes

Other outcomes included the proportion of all studies within a review that measured gaming disorder with a gaming disorder-specific instrument; the proportion of reviews that specified all elements of an outcome; and the specific review- and study-level associations between gaming disorder and depression and anxiety. All reported associations within the studies were extracted from the original study reports and characterized as present and positive, present and negative, present and null, unclear, or absent. The count and type (positive, null, negative, unclear, or absent) of results for each study were compared with the results reported for each study in the reviews. We also made several comparisons regarding overall conclusions about the associations between gaming disorder and depression and between gaming disorder and anxiety by comparing bivariable versus multivariable analyses, cross-sectional versus longitudinal analyses, and results from reviews classified as reliable versus results from all reviews.

Data extraction

We developed and pilot tested a data extraction form using Microsoft Excel®, based on the form developed by Mayo-Wilson et al. [17]. We added questions relevant to reviews of epidemiological studies [25]. During the initial data extraction, we noticed discrepancies in how specific studies were reported in the reviews, resulting in potential selective outcome reporting at the review level. To ensure that we evaluated this potential source of bias, we expanded the scope of our preregistered protocol to include examining study-level outcomes and how they were reported in reviews.

Two investigators from among MCC, JS, and GH extracted data from each review, consulting the third investigator for resolution of discrepancies where needed. If a review did not have a summary of findings table that included the total number of studies mentioned in the results or in supplementary material, we extracted data for all studies mentioned in text or tables of the Results section. Data on depression and anxiety outcomes within each study of each review were extracted by one investigator. Extracted data for a 10% random sample of studies were validated by the second and third investigator.

Data extracted from the reviews included information on methods for specifying eligibility criteria and outcomes, specific measurements (e.g., scales) of depression and anxiety in included studies, analyses conducted, whether and how review authors assessed risk of bias in included studies, specific measurement (e.g., scales) of gaming disorder in included studies, and all items from the AMSTAR 2 tool.

We summarize below the three conditions that had to be met for a specific measurement or scale to be classified as asssesing gaming disorder (Fig 3):

  • The specific measurement or scale asked questions about computer, video, online, or digital game use in general, rather than just a single game (e.g., World of Warcraft®).

  • The specific measurement or scale asked questions about gaming or online gaming rather than Internet or computer use in general (e.g., did not use only an Internet addiction measure, such as the Young Internet Addiction Test or the Compulsive Internet Use Test). If a study mentioned adapting a scale for video games and gave an example of an adapted question, we classified that scale as measuring gaming disorder. Otherwise, we classified the measurement according to the original scale from which it was adapted. We also conducted a sensitivity analysis to examine how our findings differed when other measurements (e.g., the Young Internet Addiction Test) were used with a clinical population diagnosed with gaming disorder. When the clinical population was unclear or was not diagnosed with gaming disorder and Internet addiction scales or other specific measurements/scales/interviews were used, we did not characterize this as gaming disorder (e.g., Young Internet Addiction Test in a clinical population of patients with gambling disorder).

  • The specific measurement or scale asked questions about specific symptoms of gaming disorder rather than only experiences related to video game use in general, such as time spent playing games or the experience of time loss.

Fig 3. Domains used to define gaming disorder as a construct for analysis.

Fig 3

IGD = Internet gaming disorder; PG = problematic gaming; PIU = problematic Internet use. (a) Sensitivity analysis: Clinical population of those seeking help for gaming-related problems but an Internet addiction scale was used. (b) Including those adapted from Internet addiction scales where an example question is given. (c) Where scales referenced appendices or other papers, these were also searched for example questions.

Data on depression and anxiety consisted of study scale, type of analysis, direction of association (positive, negative, or null), and how each review reported the outcome of the study (positive, negative, null, unclear, or absent).

Quality assessment

See section above entitled ‘Assessment of reliability of reviews’.

Strategy for data synthesis and reporting

We narratively describe the characteristics of included reviews and their reliability. Because measurements of exposures and outcomes were heterogeneous, we present counts of positive or null/negative outcomes from studies and how they were reported in reviews [26]. Because consistency is one factor that supports strength of evidence, we compared tallies of qualitative associations from the multiple outcomes reported in studies. We described associations to be ‘positive and consistent’ at the study level if the count of statistically-significant positive associations was greater than the total number of negative or null associations. We described an association as ‘null’ if there were more null findings or negative associations than positive. We conducted a sensitivity analysis to examine the impact of measuring gaming disorder with a scale for Internet addiction in a clinical population of individuals with gaming disorder. All extracted data and derived variables are available in S1 Dataset.

The PRISMA checklist [9] for the current study is available in S1 PRISMA Checklist. This study was conducted using publicly-available information and therefore did not require Institutional Board (IRB) approval.

Results

The searches yielded 842 records, of which, seven reviews were eligible for inclusion in this overview (Fig 4). The most frequent reasons for excluding articles (at the full-text screening stage) were that they were not a systematic review or did not specify methods (n = 35), did not report associations between gaming disorder and anxiety or depression (n = 23), and were not specific to gaming disorder (e.g., being about behavioral addictions in general) (n = 9).

Fig 4. PRISMA flow diagram.

Fig 4

Review characteristics

The characteristics of the seven included reviews are reported in Table 1. They included a total of 196 unique studies. The number of included studies per review ranged from 24 to 63, with a mean of 46. Most studies (61.7%) were included in only one review each.

Table 1. Review characteristics and reliability criteria.

Reliability criteria
First author, year Number of included studiesa Total number of participants across all included studiesb Participant populations Years of publication of included studiesc Number (%) studies measuring problematic gamingd (1) Defined eligibility criteria? (2) Conducted a comprehensive search? (3) Assessed of risk of bias? (4) Used appropriate methods to combine results? e (5) Conclusions about depression and anxiety supported by evidence? (6) Specified which outcomes would be included in the synthesis?
Sugaya 2019 [27] 51 Unclear Age 0–28 Until 2018 25(56.8) Yes Yes No Yes No No
González-Bueso 2018 [28] 24 53,889 Any 2011–2017 17 (70.8) Yes Yes No Unclear No No
Mihara 2017 [29] 47 127,749 Any Until 2016 44 (93.6) Yes No No Yes No No
Männikkö 2017 [30] 50 129,430 >12.5 years 2005–2016 41 (82.0) Yes Yes Yes No No No
King 2013 [31] 63 58,415 Any 2000–2012 39 (61.9) Yes Yes No Yes No No
Kuss 2012 [32] 58 Unclear Any 2000–2010 33 (56.9) Yes No No Yes No No
Kuss 2012 [33] 30 72,825 Children 2000–2011 20 (66.7) Yes Yes No Yes No No

Notes:

(a) The number of included studies for a review is taken from the PRISMA flow diagram (where possible) or from reports in the text or tables of each review.

(b)The number of participants was taken directly from the text where possible or calculated from other information that was reported in the review.

(c) If no years were given, the end year was listed as one year prior to the year of publication.

(d) Proportion of studies measuring problematic gaming was assessed out of all studies mentioned in the review. This did not always match the number of studies that were said to be included in the review in the abstract, methods, or results.

(e) most reviews did not combine results quantitatively.

Research question 1: Assessment of review reliability

We found that none of the seven included reviews fulfilled all six criteria for reliability. All reviews defined eligibility criteria and most reviews (six of seven) conducted comprehensive database searches (Table 1). No review defined outcomes using all five elements of completely-specified outcomes (i.e., domain, specific measurement, specific metric, method of aggregation, and time points). No reviews specified which outcomes of a study would be used in synthesis. One review specified that it would consider only study effect sizes from multivariable analyses, classifying full associations as “…a correlation was found for both genders after multivariable analyses” or partial associations as “…correlation was identified for only one gender” [28]. Other reviews did not specify how outcomes would be included, although some mentioned that "factors", "disorders", "comorbidity", "health-related outcomes", or "psycho-social features" "associated with" problematic gaming were "identified" [28], "ascertained" [29], or "extracted" [30,31].

Although all reviews acknowledged heterogeneity in measurement of problematic gaming, only one review assessed risk of bias systematically [30]. In this context, because five studies chose to conduct qualitative syntheses instead of quantitative syntheses (i.e., meta-analyses), we considered their results to have been combined appropriately. In one review, results were combined quantitatively despite a very high amount of statistical heterogeneity among studies (suggested by an I2 value of 98%) [30]. Another review classified effect sizes as small, medium, or large and presented a table of counts of effect sizes for four mental health outcomes as a way to address heterogeneity in measurement [28]. Most reviews discussed limitations at the study, outcome, and review level, but two reviews did not discuss limitations systematically [27,29].

Assessment of AMSTAR 2 criteria showed that no study met all criteria, and some criteria were lacking in all studies. Full results can be found in S1 Data Extraction.

Because of the lack of clarity around how study outcomes were selected, the reporting of outcomes that was inconsistent with study findings (see Figs 5 and 6), the inclusion of studies that did not measure gaming disorder, and the lack of systematic assessment of bias (except for one review [30]), we determined that review conclusions were not supported by the evidence from included studies. This is further explored in the following sections.

Fig 5. Associations between problematic gaming and depression.

Fig 5

a = Composite reporting of outcomes in review made comparisons difficult.

Fig 6. Study reporting of associations between problematic gaming and anxiety.

Fig 6

a = Composite reporting of outcomes in review made comparisons difficult.

Research question 2: Distinguishing between gaming disorder and other concepts

Based on our definition for measurement of gaming disorder (Fig 3), no review focused only on studies that measured gaming disorder. The percentage of studies within a review that measured gaming disorder ranged from 56.8% to 93.6%. On sensitivity analysis, where measurement of gaming disorder also included using an Internet addiction scale in a gaming disorder clinical population, the percent remained similar, ranging from 58.6% to 93.6%.

Research question 3: Reporting of associations between gaming disorder and depression or anxiety

Figs 5 and 6 report the positive and null associations for the depression and anxiety outcomes according to analysis type (bivariable/multivariable, cross-sectional, and longitudinal), their frequency of being incorporated into reviews, and how they are represented/reported in reviews (e.g., not reported, not eligible, report conflicts with outcomes). The shades of blue highlighting pertain to different percentages of reviews that incorporated a given relevant result from a given study (darker highlighting indicates higher percentages). Note that only two negative (inverse) associations were found (between gaming disorder and anxiety) and because these represent findings that were not positive and significant, they were included in the count of null findings. Overall, only the review by González-Bueso and colleagues [28] reported any null results about depression or anxiety from any study.

Associations between gaming disorder and depression

For the depression outcome (Fig 5, including citations [3464]), of the 31 studies reporting associations between gaming disorder and depression, results from 25 were included in at least one review. We found frequent under-incorporation of null results for the depression outcome by the reviews, as suggested by the paucity of blue cell highlights in the null columns. For example, the 2010 study by Rehbein and colleagues [34] reported two findings related to depression—a positive association between gaming disorder and suicidal thoughts in one subsample, but a null association between gaming disorder and self-reported depression in the full sample. However, the three reviews that included this study and reported results for depression all reported them as positive [30,31,33].

Ten of the 31 studies reporting associations between gaming disorder and depression reported both bivariable and multivariable analyses. In five of these 10 studies, results from both analyses were robust and positive, while five studies reported inconsistent results. Whether consistent or inconsistent in the studies, positive results were incorporated into five of the six reviews that included depression findings from the study.

Only one study reporting an association between gaming disorder and depression examined both cross-sectional and longitudinal associations, and the results were inconsistent [38]. However, the results were incorporated into two of three reviews as showing a positive association. The final review used a composite definition when reporting associations, which made comparisons difficult [28].

Six studies reported additional cross-sectional depression results that were not incorporated into any review. Three of these studies reported null findings and in one of those cases, results were null in both bivariable and multivariable analyses. An additional 15 studies were mentioned by reviews as reporting associations between gaming disorder and depression, but using domain definitions in Figs 2 and 3, these were not found (S1 Output contains full results). All but one of the six reviews that included these studies reported these as positive associations. Some reasons for this were: studies used a measure of Internet addiction or other exposure (e.g., "excessive" gaming), studies reported a composite measure (depression/anxiety/stress) as depression, and possible mistake in citation or data extraction (e.g., reporting data for a problematic Internet use subgroup rather than problematic gaming subgroup).

In a sensitivity analysis that included studies where a broad Internet addiction scale (rather than a gaming disorder scale) was used to measure gaming disorder in a clinical population identified as having gaming disorder, one additional study [65] was found to have positive associations and was reported in the single review that included it as positive, while another three studies [6668] had null findings which were not reported by the three reviews that included them.

Associations between gaming disorder and anxiety

Of the 28 studies that reported associations between anxiety and gaming disorder, results from only 22 of these studies were incorporated into reviews (Fig 6, including citations [3475]).

Six studies reported both bivariable and multivariable associations; half of these showed inconsistent results. Whether consistent or inconsistent, reviews incorporated only positive findings. Six studies reported results that were not incorporated into any review; four of these had inconsistent or null findings. An additional nine studies were mentioned by reviews as reporting associations between an gaming disorder and anxiety, but using domain definitions in Fig 3, these were not found. All but one of the three reviews that incorporated these studies reported these associations as positive.

In the sensitivity analysis, one additional study [65] reported inconsistent associations in bivariable and multivariable analysis and was reported as positive in the one review that contained it.

Research question 4: Association between gaming disorder and depression or anxiety in reliable reviews

Overall, no review satisfied all the criteria we used to identify reliable reviews, so we could not address this research question.

Discussion

This summary of systematic reviews found methodological problems in all seven systematic reviews that reported on associations between gaming disorder and depression or anxiety; no reviews could be classified as reliable based on established criteria. Although most systematic reviews studied herein defined their criteria for selecting studies and conducted a comprehensive search, each review was rated as unreliable because of one or more of the other criteria. Because of the poor pre-specification of how outcomes would be included, it is difficult to draw conclusions from these reviews regarding associations between gaming disorder and depression or anxiety that are supported by evidence. These findings suggest that the way systematic reviews of gaming disorder have been reporting results and drawing conclusions may have introduced bias into the gaming disorder literature, possibly misleading future research, policy-making, and patient care.

Various concerns identified during this summary of systematic reviews are worthy of further discussion. We present these in the hope that the current work drives important progress in research on gaming disorder and other types of behavioral addictions in the coming years.

First, the existing reviews seldom incorporated null findings (i.e., lack of associations) or negative findings (i.e., inverse associations) from included studies even when the studies reported such findings. This is a major concern because it seems to represent selective outcome reporting at the review level. It is vital to conduct systematic reviews and meta-analyses in ways that are replicable and consistent with best practices to ensure that all evidence is reported and that relevant studies and findings are not overlooked. Selecting which outcomes of studies to include in a review without specifying the process, which has been labelled “cherry-picking” in the clinical epidemiology literature, can lead to biased conclusions at the review level [10,13]. Completely specifying all elements of outcomes (i.e., domain, specific measurements, specific metrics, methods of aggregation, and time-points of interest) or explicitly noting whether all variations of a given outcome element will be extracted is the current standard for evidence synthesis [15,26,76]. As incomplete outcome specification may lead to trillions of potential combinations of meta-analytic results [13], it is inappropriate to draw meaningful and reliable conclusions about associations between gaming disorder and the common mental health problems of depression and anxiety from the reviews summarized in this paper. Selective reporting of outcomes can be hard to detect, and further research into the impact of selective inclusion of results in reviews is needed to advance the understanding of this form of bias on evidence synthesis [77,78].

A second major concern is that reviews did not limit evidence synthesis and conclusions to studies that measured the construct of gaming disorder and at times used overly-broad definitions of depression and anxiety (e.g., combined depression, anxiety, and stress), which might have led to reports of associations between gaming disorder and depression or anxiety when none might exist.

Although more recent reviews had higher proportions of gaming-only measures, even recent reviews included studies that used Internet addiction questions to measure gaming disorder. Distinguishing between problematic behaviors is vital in ongoing research of problematic technology use and will continue to be relevant to shaping the future of health policy and government regulation of the Internet, video games, and other forms of media and technology. Ensuring that systematic and accurate measurement of gaming disorder in studies and accurate measurement and reporting of exposures, outcomes, and conclusions in reviews are vital to inform ongoing decision making regarding diagnosis, treatment, and public health interventions.

A third major concern is that only one review [30] reported a systematic assessment of risk of bias using multiple domains, which has long been a best practice in conducting systematic reviews [7983]. When the risk of bias is not systematically assessed and reported, conclusions from studies included in reviews may be seen as valid and reliable when they may actually reflect biases, such as selection bias, information bias, and/or confounding [84]. When evidence of questionable methodologic quality is used to inform public health or policy decisions, such decisions may be misguided.

To our knowledge, the current analysis is the first comprehensive examination of selective outcome reporting in systematic reviews of gaming disorder, a relatively new clinical entity. Due to this selective outcome reporting, incomplete outcome specification, and lack of systematic assessment of risk of bias, we found no reviews that could be considered reliable. These findings suggest that the evidence base of systematic reviews of associations between gaming disorder and the most common mental health problems must be improved.

Limitations

The current overview is subject to certain limitations. First, at the level of the studies we found significant inconsistencies in measurement and analysis, which were dealt with by describing counts of associations by type. While this is a somewhat reductionist approach to summarizing results, it helps paint a picture. Relatedly, no reviews defined outcomes completely. Second, we limited our analysis to systematic reviews published in English. It is possible that our findings may have been different had we included reviews in other languages. Third, we focused on the outcomes of depression or anxiety. This narrow scope made a detailed analysis possible, but findings regarding associations between gaming disorder and other outcomes (e.g., attention-deficit hyperactivity disorder) may have been different. However, due to the ubiquitous nature of selective outcome reporting, in particular, in the reviews herein, we consider this to be unlikely. Fourth, we defined the constructs of gaming disorder, depression, and anxiety very specifically; had we used broader definitions, our findings would likely be different. However, using a narrow definition was our aim. We do not attempt to draw conclusions at the study level (the 196 studies) due to the inconsistency within studies and the uncertain nature of the examined evidence. Finally, in our search of PubMed we used the PubMed publication type filters of “systematic review”, “review,” or “meta-analysis”, while we broadened our search of PsycInfo to include these terms as text-words in all fields. For this reason, it is possible that we missed some systematic reviews that were only available in PubMed and were not indexed using these terms or did not contain these terms in the title, abstract, publication type, or keywords.

Conclusions

To advance the field of addictive behaviors and ensure that research measures and reports constructs rigorously and with clarity, existing standards for systematic review conduct and reporting should be followed. Increasing transparency of reviews and minimizing the risk of bias requires the effort of multiple agents. Authors must prospectively register protocols (including adequately specifying outcomes); use reporting guidelines, such as those from the EQUATOR Network; and share data, analysis code, and other study materials. Journals and editors must verify authors’ adherence to reporting guidelines [77]. Although public health decision-making should always proceed on the best available evidence [85], the data provided in this paper suggest that limiting technology-related diagnoses to video game play is not likely to accurately reflect the findings of years of research surrounding problematic technology use. A highly rigorous systematic review that fully specifies outcome domains is needed to clarify the potential mental health problems associated with problematic technology behaviors, including video gaming and Internet use.

Supporting information

S1 PRISMA Checklist. PRISMA checklist for reporting of our systematic review.

(PDF)

S1 Protocol. PROSPERO registration for our systematic review protocol.

(PDF)

S1 Search Strategy. Search strategies.

(PDF)

S1 Data Extraction. Data extraction at the review level, including AMSTAR 2.

(XLSX)

S1 Dataset. Complete analysis dataset containing extracted and derived variables.

(DTA)

S1 Output. Output of analysis.

(DOCX)

Acknowledgments

The authors are grateful to Michael M. Hughes for assistance with the graphic design and formatting of figures for publication.

Data Availability

All relevant data are within the manuscript and its Supporting Information files.

Funding Statement

The author(s) received no specific funding for this work.

References

  • 1.World Health Organization. International classification of diseases for mortality and morbidity statistics (11th Revision). World Health Organization; 2018. https://icd.who.int/browse11/l-m/en.
  • 2.American Psychiatric Association. Diagnostic and Statistical Manual of Mental Disorders, Fifth Edition. Arlington, VA: American Psychiatric Association; 2013. dsm.psychiatryonline.org.
  • 3.Starcevic V, Billieux J. Does the construct of Internet addiction reflect a single entity or a spectrum of disorders? Clin Neuropsychiatry J Treat Eval. 2017;14: 5–10. Available: http://search.ebscohost.com/login.aspx?direct=true&db=psyh&AN=2017-10293-002&site=ehost-live&scope=site. [Google Scholar]
  • 4.Rumpf H-J, Achab S, Billieux J, Bowden-Jones H, Carragher N, Demetrovics Z, et al. Including gaming disorder in the ICD-11: The need to do so from a clinical and public health perspective. J Behav Addict. 2018; 1–6. 10.1556/2006.7.2018.59 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Saunders JB, Hao W, Long J, King DL, Mann K, Fauth-Bühler M, et al. Gaming disorder: Its delineation as an important condition for diagnosis, management, and prevention. J Behav Addict. 2017;6: 271–279. 10.1556/2006.6.2017.039 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Higuchi S, Nakayama H, Mihara S, Maezono M, Kitayuguchi T, Hashimoto T. Inclusion of gaming disorder criteria in ICD-11: A clinical perspective in favor. J Behav Addict. 2017;6: 293–295. 10.1556/2006.6.2017.049 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Fineberg NA, Demetrovics Z, Stein DJ, Ioannidis K, Potenza MN, Grünblatt E, et al. Manifesto for a European research network into Problematic Usage of the Internet. Eur Neuropsychopharmacol J Eur Coll Neuropsychopharmacol. 2018;28: 1232–1246. 10.1016/j.euroneuro.2018.08.004 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Page MJ, Shamseer L, Altman DG, Tetzlaff J, Sampson M, Tricco AC, et al. Epidemiology and Reporting Characteristics of Systematic Reviews of Biomedical Research: A Cross-Sectional Study. PLoS Med. 2016;13: e1002028 10.1371/journal.pmed.1002028 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Moher D, Liberati A, Tetzlaff J, Altman DG, PRISMA Group. The PRISMA Group. Preferred reporting items for systematic reviews and meta-analyses: the PRISMA statement. PLoS Med. 2009;6: e1000097 10.1371/journal.pmed.1000097 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Mayo-Wilson E, Fusco N, Li T, Hong H, Canner JK, Dickersin K, et al. Multiple outcomes and analyses in clinical trials create challenges for interpretation and research synthesis. J Clin Epidemiol. 2017;86: 39–50. 10.1016/j.jclinepi.2017.05.007 [DOI] [PubMed] [Google Scholar]
  • 11.Pussegoda K, Turner L, Garritty C, Mayhew A, Skidmore B, Stevens A, et al. Systematic review adherence to methodological or reporting quality. Syst Rev. 2017;6: 131 10.1186/s13643-017-0527-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Goodman S, Dickersin K. Metabias: a challenge for comparative effectiveness research. Ann Intern Med. 2011;155: 61–62. 10.7326/0003-4819-155-1-201107050-00010 [DOI] [PubMed] [Google Scholar]
  • 13.Mayo-Wilson E, Li T, Fusco N, Bertizzolo L, Canner JK, Cowley T, et al. Cherry-picking by trialists and meta-analysts can drive conclusions about intervention efficacy. J Clin Epidemiol. 2017. 10.1016/j.jclinepi.2017.07.014 [DOI] [PubMed] [Google Scholar]
  • 14.Saldanha IJ, Lindsley KB, Money S, Kimmel HJ, Smith BT, Dickersin K. Outcome choice and definition in systematic reviews leads to few eligible studies included in meta-analyses: a case study. BMC Med Res Methodol. 2020;20: 30 10.1186/s12874-020-0898-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Saldanha IJ, Dickersin K, Wang X, Li T. Outcomes in Cochrane systematic reviews addressing four common eye conditions: an evaluation of completeness and comparability. PloS One. 2014;9: e109400 10.1371/journal.pone.0109400 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Moher D, Shamseer L, Clarke M, Ghersi D, Liberati A, Petticrew M, et al. Preferred reporting items for systematic review and meta-analysis protocols (PRISMA-P) 2015 statement. Syst Rev. 2015;4: 1 10.1186/2046-4053-4-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Mayo-Wilson E, Ng SM, Chuck RS, Li T. The quality of systematic reviews about interventions for refractive error can be improved: a review of systematic reviews. BMC Ophthalmol. 2017;17: 164 10.1186/s12886-017-0561-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Saldanha IJ, Lindsley KB, Lum F, Dickersin K, Li T. Reliability of the Evidence Addressing Treatment of Corneal Diseases: A Summary of Systematic Reviews. JAMA Ophthalmol. 2019;137: 775–785. 10.1001/jamaophthalmol.2019.1063 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Le JT, Qureshi R, Twose C, Rosman L, Han G, Fapohunda K, et al. Evaluation of Systematic Reviews of Interventions for Retina and Vitreous Conditions. JAMA Ophthalmol. 2019. 10.1001/jamaophthalmol.2019.4016 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Golozar A, Chen Y, Lindsley K, Rouse B, Musch DC, Lum F, et al. Identification and Description of Reliable Evidence for 2016 American Academy of Ophthalmology Preferred Practice Pattern Guidelines for Cataract in the Adult Eye. JAMA Ophthalmol. 2018;136: 514–523. 10.1001/jamaophthalmol.2018.0786 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Lindsley K, Li T, Ssemanda E, Virgili G, Dickersin K. Interventions for Age-Related Macular Degeneration: Are Practice Guidelines Based on Systematic Reviews? Ophthalmology. 2016;123: 884–897. 10.1016/j.ophtha.2015.12.004 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Shea BJ, Bouter LM, Peterson J, Boers M, Andersson N, Ortiz Z, et al. External Validation of a Measurement Tool to Assess Systematic Reviews (AMSTAR). PLOS ONE. 2007;2: e1350 10.1371/journal.pone.0001350 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Critical Appraisal Skills Programme. CASP systematic review checklist. 2018. https://casp-uk.net/wp-content/uploads/2018/01/CASP-Systematic-Review-Checklist_2018.pdf.
  • 24.Shea BJ, Reeves BC, Wells G, Thuku M, Hamel C, Moran J, et al. AMSTAR 2: a critical appraisal tool for systematic reviews that include randomised or non-randomised studies of healthcare interventions, or both. BMJ. 2017;358 10.1136/bmj.j4008 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Stroup DF, Berlin JA, Morton SC, Olkin I, Williamson GD, Rennie D, et al. Meta-analysis of observational studies in epidemiology: a proposal for reporting. Meta-analysis Of Observational Studies in Epidemiology (MOOSE) group. JAMA J Am Med Assoc. 2000;283: 2008–2012. [DOI] [PubMed] [Google Scholar]
  • 26.Chandler J, Cumpston M, Li T, Page MJ, Welch VA. Cochrane Handbook for Systematic Reviews of Interventions. 2nd edition Higgins JPT, Thomas J, editors. Chichester (UK): John Wiley & Sons; 2019. [Google Scholar]
  • 27.Sugaya N, Shirasaka T, Takahashi K, Kanda H. Bio-psychosocial factors of children and adolescents with internet gaming disorder: a systematic review. Biopsychosoc Med. 2019;13: 3 10.1186/s13030-019-0144-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.González-Bueso V, Santamaría JJ, Fernández D, Merino L, Montero E, Ribas J. Association between Internet Gaming Disorder or Pathological Video-Game Use and Comorbid Psychopathology: A Comprehensive Review. Int J Environ Res Public Health. 2018;15 10.3390/ijerph15040668 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Mihara S, Higuchi S. Cross‐sectional and longitudinal epidemiological studies of internet gaming disorder: A systematic review of the literature. Psychiatry Clin Neurosci. 2017;71: 425–444. 10.1111/pcn.12532 [DOI] [PubMed] [Google Scholar]
  • 30.Männikkö N, Ruotsalainen H, Miettunen J, Pontes HM, Kääriäinen M. Problematic gaming behaviour and health-related outcomes: A systematic review and meta-analysis. J Health Psychol. 2017; 1359105317740414. 10.1177/1359105317740414 [DOI] [PubMed] [Google Scholar]
  • 31.King DL, Haagsma MC, Delfabbro PH, Gradisar M, Griffiths MD. Toward a consensus definition of pathological video-gaming: A systematic review of psychometric assessment tools. Clin Psychol Rev. 2013;33: 331–342. 10.1016/j.cpr.2013.01.002 [DOI] [PubMed] [Google Scholar]
  • 32.Kuss DJ, Griffiths MD. Internet gaming addiction: A systematic review of empirical research. Int J Ment Health Addict. 2012;10: 278–296. 10.1007/s11469-011-9318-5 [DOI] [Google Scholar]
  • 33.Kuss DJ, Griffiths MD. Online gaming addiction in children and adolescents: A review of empirical research. J Behav Addict. 2012;1: 3–22. 10.1556/JBA.1.2012.1.1 [DOI] [PubMed] [Google Scholar]
  • 34.Rehbein F, Kleimann M, Mößle T. Prevalence and risk factors of video game dependency in adolescence: results of a German nationwide survey. Cyberpsychology Behav Soc Netw. 2010;13: 269–277. [DOI] [PubMed] [Google Scholar]
  • 35.Allison SE, von Wahlde L, Shockley T, Gabbard GO. The development of the self in the era of the internet and role-playing fantasy games. Am J Psychiatry. 2006;163: 381–385. 10.1176/appi.ajp.163.3.381 [DOI] [PubMed] [Google Scholar]
  • 36.Andreassen CS, Griffiths MD, Gjertsen SR, Krossbakken E, Kvam S, Pallesen S. The relationships between behavioral addictions and the five-factor model of personality. J Behav Addict. 2013;2: 90–99. Available: http://search.ebscohost.com/login.aspx?direct=true&db=psyh&AN=2013-30916-004&site=ehost-live&scope=site. [DOI] [PubMed] [Google Scholar]
  • 37.Bouna-Pyrrou P, Mühle C, Kornhuber J, Lenz B. Internet gaming disorder, social network disorder and laterality: Handedness relates to pathological use of social networks. J Neural Transm. 2015;122: 1187–1196. 10.1007/s00702-014-1361-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Brunborg GS, Mentzoni RA, Froyland LR. Is video gaming, or video game addiction, associated with depression, academic achievement, heavy episodic drinking, or conduct problems? J Behav Addict. 2014;3 10.1556/JBA.3.2014.002 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Desai RA, Krishnan-Sarin S, Cavallo D, Potenza MN. Video-gaming among high school students: health correlates, gender differences, and problematic gaming. Pediatrics. 2010;126: e1414–1424. 10.1542/peds.2009-2706 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Gentile DA, Choo H, Liau A, Sim T, Li D, Fung D, et al. Pathological video game use among youths: a two-year longitudinal study. Pediatrics. 2011;127: e319–329. 10.1542/peds.2010-1353 [DOI] [PubMed] [Google Scholar]
  • 41.Jiménez-Murcia S, Fernández-Aranda F, Granero R, Chóliz M, La Verde M, Aguglia E, et al. Video Game Addiction in Gambling Disorder: Clinical, Psychopathological, and Personality Correlates. BioMed Res Int. 2014;2014: 11 10.1155/2014/315062. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Kim NR, Hwang SS-H, Choi J-S, Kim D-J, Demetrovics Z, Király O, et al. Characteristics and Psychiatric Symptoms of Internet Gaming Disorder among Adults Using Self-Reported DSM-5 Criteria. Psychiatry Investig. 2015;13: 58–66. 10.4306/pi.2016.13.1.58 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.King DL, Delfabbro PH, Zwaans T, Kaptsis D. Clinical features and axis I comorbidity of Australian adolescent pathological Internet and video game users. Aust N Z J Psychiatry. 2013;47: 1058–1067. 10.1177/0004867413491159 [DOI] [PubMed] [Google Scholar]
  • 44.King DL, Delfabbro PH. The Cognitive Psychopathology of Internet Gaming Disorder in Adolescence. J Abnorm Child Psychol. 2016;44: 1635–1645. 10.1007/s10802-016-0135-y [DOI] [PubMed] [Google Scholar]
  • 45.Király O, Griffiths MD, Urban R, Farkas J, Kokonyei G, Elekes Z, et al. Problematic internet use and problematic online gaming are not the same: findings from a large nationally representative adolescent sample. Cyberpsychology Behav Soc Netw. 2014;17: 749–754. 10.1089/cyber.2014.0475 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46.Laconi S, Pirès S, Chabrol H. Internet gaming disorder, motives, game genres and psychopathology. Comput Hum Behav. 2017;75: 652–659. 10.1016/j.chb.2017.06.012 [DOI] [Google Scholar]
  • 47.Lemos IL, Cardoso A, Sougey EB. Validity and reliability assessment of the Brazilian version of the game addiction scale (GAS). Compr Psychiatry. 2016;67: 19–25. 10.1016/j.comppsych.2016.01.014 [DOI] [PubMed] [Google Scholar]
  • 48.Li D, Liau A, Khoo A. Examining the influence of actual-ideal self-discrepancies, depression, and escapism, on pathological gaming among massively multiplayer online adolescent gamers. Cyberpsychology Behav Soc Netw. 2011;14: 535–539. 10.1089/cyber.2010.0463 [DOI] [PubMed] [Google Scholar]
  • 49.Männikkö N, Billieux J, Kaariainen M. Problematic digital gaming behavior and its relation to the psychological, social and physical health of Finnish adolescents and young adults. J Behav Addict. 2015;4: 281–288. 10.1556/2006.4.2015.040 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.Mentzoni RA, Brunborg GS, Molde H, Myrseth H, Mår Skouverøe KJ, Hetland J, et al. Problematic video game use: Estimated prevalence and associations with mental and physical health. Cyberpsychology Behav Soc Netw. 2011;14: 591–596. Available: http://search.ebscohost.com/login.aspx?direct=true&db=psyh&AN=2011-24261-006&site=ehost-live&scope=site. [DOI] [PubMed] [Google Scholar]
  • 51.Müller KW, Janikian M, Dreier M, Wölfling K, Beutel ME, Tzavara C, et al. Regular gaming behavior and internet gaming disorder in European adolescents: results from a cross-national representative survey of prevalence, predictors, and psychopathological correlates. Eur Child Adolesc Psychiatry. 2015;24: 565–574. 10.1007/s00787-014-0611-2 [DOI] [PubMed] [Google Scholar]
  • 52.Na E, Lee H, Choi I, Kim D-J. Comorbidity of Internet gaming disorder and alcohol use disorder: A focus on clinical characteristics and gaming patterns. Am J Addict. 2017;26: 326–334. 10.1111/ajad.12528 [DOI] [PubMed] [Google Scholar]
  • 53.Pápay O, Urbán R, Griffiths MD, Nagygyörgy K, Farkas J, Kökönyei G, et al. Psychometric properties of the Problematic Online Gaming Questionnaire Short-Form and prevalence of problematic online gaming in a national sample of adolescents. Cyberpsychology Behav Soc Netw. 2013;16: 340–348. 10.1089/cyber.2012.0484 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54.Peng W, Liu M. Online gaming dependency: a preliminary study in China. Cyberpsychology Behav Soc Netw. 2010;13: 329–333. [DOI] [PubMed] [Google Scholar]
  • 55.Starcevic V, Berle D, Porter G, Fenech P. Problem video game use and dimensions of psychopathology. Int J Ment Health Addict. 2011;9: 248–256. Available: http://search.ebscohost.com/login.aspx?direct=true&db=psyh&AN=2011-11090-003&site=ehost-live&scope=site. [Google Scholar]
  • 56.Van Rooij AJ, Schoenmakers TM, Van den Eijnden RJJM, Vermulst AA, Van de Mheen D. Video game addiction test: validity and psychometric characteristics. Cyberpsychology Behav Soc Netw. 2012;15: 507–511. 10.1089/cyber.2012.0007 [DOI] [PubMed] [Google Scholar]
  • 57.Vadlin S, Aslund C, Hellstrom C, Nilsson KW. Associations between problematic gaming and psychiatric symptoms among adolescents in two samples. Addict Behav. 2016;61: 8–15. 10.1016/j.addbeh.2016.05.001 [DOI] [PubMed] [Google Scholar]
  • 58.Vukosavljevic-Gvozden T, Filipovic S, Opacic G. The mediating role of symptoms of psychopathology between irrational beliefs and Internet gaming addiction. J Ration-Emotive Cogn-Behav Ther. 2015;33: 387–405. 10.1007/s10942-015-0218-7 [DOI] [Google Scholar]
  • 59.Colder Carras M, Kalbarczyk A, Wells K, Banks J, Kowert R, Gillespie C, et al. Connection, meaning, and distraction: A qualitative study of video game play and mental health recovery in veterans treated for mental and/or behavioral health problems. Soc Sci Med 1982. 2018;216: 124–132. 10.1016/j.socscimed.2018.08.044 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 60.King DL, Delfabbro PH, Zajac IT. Preliminary validation of a new clinical tool for identifying problem video game playing. Int J Ment Health Addict. 2011;9: 72–87. Available: http://search.ebscohost.com/login.aspx?direct=true&db=psyh&AN=2011-02644-006&site=ehost-live&scope=site. [Google Scholar]
  • 61.Liu M, Peng W. Cognitive and Psychological Predictors of the Negative Outcomes Associated with Playing MMOGs (Massively Multiplayer Online Games). Comput Hum Behav. 2009;25: 1306–1311. 10.1016/j.chb.2009.06.002 [DOI] [Google Scholar]
  • 62.Topor DR, Swenson LP, Liguori GM, Spirito A, Lowenhaupt EA, Hunt JI. Problematic video game use scale: initial psychometric properties with psychiatrically hospitalized adolescents. J Clin Psychiatry. 2011;72: 1611–1615. 10.4088/JCP.11m07040 [DOI] [PubMed] [Google Scholar]
  • 63.Walther B, Morgenstern M, Hanewinkel R. Co-occurrence of addictive behaviours: Personality factors related to substance use, gambling and computer gaming. Eur Addict Res. 2012;18: 167–174. 10.1159/000335662 [DOI] [PubMed] [Google Scholar]
  • 64.Yu H, Cho J. Prevalence of Internet Gaming Disorder among Korean Adolescents and Associations with Non-psychotic Psychological Symptoms, and Physical Aggression. Am J Health Behav. 2016;40: 705–716. 10.5993/AJHB.40.6.3 [DOI] [PubMed] [Google Scholar]
  • 65.Hyun GJ, Han DH, Lee YS, Kang KD, Yoo SK, Chung U-S, et al. Risk factors associated with online game addiction: A hierarchical model. Comput Hum Behav. 2015;48: 706–713. 10.1016/j.chb.2015.02.008 [DOI] [Google Scholar]
  • 66.Han DH, Lyoo IK, Renshaw PF. Differential regional gray matter volumes in patients with on-line game addiction and professional gamers. J Psychiatr Res. 2012;46: 507–515. 10.1016/j.jpsychires.2012.01.004 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 67.Kim SM, Han DH, Lee YS, Kim JE, Renshaw PF. Changes in brain activity in response to problem solving during the abstinence from online game play. J Behav Addict. 2012;1: 41–49. 10.1556/JBA.1.2012.2.1 [DOI] [PubMed] [Google Scholar]
  • 68.Han DH, Hwang JW, Renshaw PF. Bupropion sustained release treatment decreases craving for video games and cue-induced brain activity in patients with Internet video game addiction. Exp Clin Psychopharmacol. 2010;18: 297–304. 10.1037/a0020023 [DOI] [PubMed] [Google Scholar]
  • 69.Batthyány D, Muller KW, Benker F, Wolfling K. [Computer game playing: clinical characteristics of dependence and abuse among adolescents]. Wien Klin Wochenschr. 2009;121: 502–509. 10.1007/s00508-009-1198-3 [DOI] [PubMed] [Google Scholar]
  • 70.Kim MG, Kim J. Cross-validation of reliability, convergent and discriminant validity for the problematic online game use scale. Comput Hum Behav. 2010;26: 389–398. Available: http://search.ebscohost.com/login.aspx?direct=true&db=psyh&AN=2010-04134-016&site=ehost-live&scope=site. [Google Scholar]
  • 71.Kim N, Hughes TL, Park CG, Quinn L, Kong ID. Resting-State Peripheral Catecholamine and Anxiety Levels in Korean Male Adolescents with Internet Game Addiction. Cyberpsychology Behav Soc Netw. 2016;19: 202–208. 10.1089/cyber.2015.0411 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 72.Mehroof M, Griffiths MD. Online gaming addiction: The role of sensation seeking, self-control, neuroticism, aggression, state anxiety, and trait anxiety. Cyberpsychology Behav Soc Netw. 2010;13: 313–316. Available: http://search.ebscohost.com/login.aspx?direct=true&db=psyh&AN=2010-13062-011&site=ehost-live&scope=site. [DOI] [PubMed] [Google Scholar]
  • 73.Achab S, Nicolier M, Mauny F, Monnin J, Trojak B, Vandel P, et al. Massively multiplayer online role-playing games: comparing characteristics of addict vs non-addict online recruited gamers in a French adult population. BMC Psychiatry. 2011;11 10.1186/1471-244X-11-144 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 74.Coyne SM, Dyer WJ, Densley R, Money NM, Day RD, Harper JM. Physiological indicators of pathologic video game use in adolescence. J Adolesc Health. 2015. 10.1016/j.jadohealth.2014.10.271 [DOI] [PubMed] [Google Scholar]
  • 75.Tejeiro RA, Gómez-Vallecillo JL, Pelegrina M, Wallace A, Emberley E. Risk factors associated with the abuse of video games in adolescents. Psychology. 2012;3: 310–314. 10.4236/psych.2012.34044 [DOI] [Google Scholar]
  • 76.Zarin D, Tse T, Williams R, Califf R, Ide N. The ClinicalTrials.gov Results Database—Update and Key Issues. NEJM. 2011;364: 852–860. 10.1056/NEJMsa1012065 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 77.McLeroy KR, Garney W, Mayo-Wilson E, Grant S. Scientific Reporting: Raising the Standards. Health Educ Behav Off Publ Soc Public Health Educ. 2016;43: 501–508. 10.1177/1090198116668522 [DOI] [PubMed] [Google Scholar]
  • 78.Page MJ, McKenzie JE, Kirkham J, Dwan K, Kramer S, Green S, et al. Bias due to selective inclusion and reporting of outcomes and analyses in systematic reviews of randomised trials of healthcare interventions. Cochrane Database Syst Rev. 2014; MR000035. 10.1002/14651858.MR000035.pub2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 79.Higgins J, Green S, editors. Cochrane Handbook for Systematic Reviews of Interventions Version 5.1.0 [updated March 2011]. The Cochrane Collaboration; 2011. www.cochrane-handbook.org.
  • 80.Moher D, Jadad AR, Nichol G, Penman M, Tugwell P, Walsh S. Assessing the quality of randomized controlled trials: an annotated bibliography of scales and checklists. Control Clin Trials. 1995;16: 62–73. 10.1016/0197-2456(94)00031-w [DOI] [PubMed] [Google Scholar]
  • 81.Huwiler-Müntener K, Jüni P, Junker C, Egger M. Quality of reporting of randomized trials as a measure of methodologic quality. JAMA. 2002;287: 2801–2804. 10.1001/jama.287.21.2801 [DOI] [PubMed] [Google Scholar]
  • 82.Higgins JPT, Altman DG, Gøtzsche PC, Jüni P, Moher D, Oxman AD, et al. The Cochrane Collaboration’s tool for assessing risk of bias in randomised trials. BMJ. 2011;343: d5928 10.1136/bmj.d5928 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 83.Sterne JAC, Savović J, Page MJ, Elbers RG, Blencowe NS, Boutron I, et al. RoB 2: a revised tool for assessing risk of bias in randomised trials. BMJ. 2019;366: l4898 10.1136/bmj.l4898 [DOI] [PubMed] [Google Scholar]
  • 84.Institute of Medicine (US) Committee on Standards for Systematic Reviews of Comparative Effectiveness Research. Finding What Works in Health Care: Standards for Systematic Reviews. Eden J, Levit L, Berg A, Morton S, editors. Washington (DC): National Academies Press (US); 2011. http://www.ncbi.nlm.nih.gov/books/NBK209518/. [PubMed] [Google Scholar]
  • 85.Brownson RC, Chriqui JF, Stamatakis KA. Understanding Evidence-Based Public Health Policy. Am J Public Health. 2009;99: 1576–1583. 10.2105/AJPH.2008.156224 [DOI] [PMC free article] [PubMed] [Google Scholar]

Decision Letter 0

Florian Naudet

3 Jan 2020

PONE-D-19-32964

Evaluating the quality of evidence for gaming disorder: A systematic review of systematic reviews of associations between gaming disorder and depression or anxiety

PLOS ONE

Dear Dr Colder Carras,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

First, I would like to thank here the two reviewers for providing very insightful comments. The comments were all the more important for me to reach a decision, and I want you to address all of these comments in details. Importantly, I think that some additional work is needed

- Please update the searches. The topic is hot and the literature certainly expanding very fast. This is therefore important to provide up to date results. Being up to date is all the more important as your paper will attract a lot of attention ;

- Please use AMSTAR 2, as far as possible, as a secondary outcome (as suggested by reviewer 2) ; 

- Please extract information by two reviewers in an independent manner. A second reviewer must extract the information at the study level and in case of disagreement a third reviewer must arbitrate. This must be done for all the studies and not only for a sample. Vote counting is not an optimal method and therefore subject to bias during data extraction. As an editor for PLOS one I must ensure that the method is sound enough and this is an important shortcoming. Please also provide inter-rater agreement. 

- You might be interested to map the evidence base using a figure similar to Figure 2 in this reference : doi: 10.1093/ije/dyx138. Importantly, you don't have to cite this study (I'm against encouraging for self citation) but the visual display in figure 2 would be of interest in the context of your study (this is to give you an idea and it deserves surely a lot of adaptation to your specific study). To detail wether an empirical study is included in a given systematic review, a heatmap would be more readable than the current tables with numbers. In addition a plot mapping all existing empirical evidence used in the SRs across time / versus all empirical evidence available (with a line) / versus the number of empirical study in each SR would be very helpful. Indeed as it is presented now it is very difficult to understand for the reader.

- You must detail in depth the criteria used to define "an empirical" study in your methods section.

- You must provide details for selection of empirical studies in your flow charts. 

- You must search for empirical studies that were not included in the review. Indeed, all systematic reviews are incomplete in comparisons with the others as you point. But it is also important to assess wether all these SR taken together were able to find all the available empirical studies (see my comment above about the suggested figure). 

- Please make a specific section "changes to the initial protocol" to make it clear that these changes were performed during the peer review process. 

- Last please review the abstract using PRISMA statement for abstracts. Please, make sure that sufficient details are provided in the abstract and please write also a few words about the limitations of your study to avoid any over-interpretation of the findings. 

I appreciate that the changes needed are extensive but these changes will surely help to improve the paper. 

We would appreciate receiving your revised manuscript by Feb 17 2020 11:59PM. When you are ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter.

To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). This letter should be uploaded as separate file and labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. This file should be uploaded as separate file and labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. This file should be uploaded as separate file and labeled 'Manuscript'.

Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out.

We look forward to receiving your revised manuscript.

Kind regards,

Florian Naudet, M.D., M.P.H., Ph.D.

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

http://www.journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and http://www.journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: N/A

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: Carras et al. conducted an interesting study analyzing the quality of evidence for gaming disorder and found that no reviews were classified as reliable. This topic is of great importance and I commend the authors for taking on this task. I additionally commend the authors for following a pre-registered protocol, for reporting results according to the PRISMA reporting standards, and for making all materials available online through OSF. Overall, this is a robust study and, with minor corrections, should be considered for publication.

- The introduction and discussion is well written and appropriately introduces and discusses the topic at hand and the importance of having high-quality reviews to provide reliable evidence for healthcare and policy decision-making. I am pleased the authors have made all materials and protocols available online. This improves the transparency and reproducibility of their work.

- Line 135 says Open Science Foundation, but the correct name is Open Science Framework.

- Please provide a statement regarding IRB approval.

- While I agree the with the classification criteria chosen in this study, it would benefit from further explanation why these 6 aspects of the criteria were chosen. The authors mention Mayo-Wilson and colleagues as part of their reasoning, yet Mayo-Wilson and colleagues reference including components of the Critical Appraisal Skills Programme (CASP), the Assessment of Multiple Systematic Reviews (AMSTAR), and the Preferred Reporting Items for Systematic Reviews (PRISMA).

- The second line in the conclusion has an extra space. While the authors mention systematic reviews should follow existing standards and should be improved, this review could benefit from providing a few more concrete suggestions on how to do so.

Reviewer #2: Evaluating the quality of evidence for gaming disorder: A systematic review of systematic reviews of associations between gaming disorder and depression or anxiety

This study attempts to present a review of the quality evidence of systematic review for gaming disorder. The topic is interesting, but the rationale, the aim and the methodology of the study proposed need to be clarified.

Comments

Title

In my opinion, the proposed title is inconsistent with the main aim of the study. After reading all the manuscript, I am wondering if your main research question is relative to selective reporting of outcomes, or if it rather of focused on the quality of evidence in systematic reviews.

Abstract

Currently, there are opportunities to improve the understanding of the abstract. The terminology used is not clear and the aim of the study is not clearly defined either.

Introduction

In my opinion, the rationale for this study could be improved. This section should be edited explaining the background to the study, an accurate summary of the existing literature and why this study is necessary.

Some specific comments:

1. I think the second part of the introduction “Synthesizing evidence from systematic reviews” needs to be improved and reworked. Moreover, I do not totally agree with your statement (line 85-90), they seem to suggest that the reliability of a systematic review is only due to the definition of an outcome. Please, revise your definition.

2. In my opinion, the figure 1 should not be presented in the introduction part but in the methodological part as a rational to characterize an outcome.

3. Regarding your sentence in line 96-98, I think that to claim “Reviews that actively minimize bias and are reported transparently can be considered high-quality” could be (too) over-simplifying the methodological quality of a well conduct systematic review.

4. In my opinion, the rationale for this study is not clear (and accurate). This section should explain the background to the study, its aims, an accurate summary of the existing literature and why this study is necessary or its contribution to the field. Specifically, authors should describe the rationale for the analysis in the context of what is already known.

Methods

1. Page 9-10: Why did the authors choose the scale proposed by Mayo-Wilson and colleagues? Some others tools could be more suited (for example the tool AMSTAR2). I am very surprised by the way you classified the review as reliable. Please provide a rationale and more explanation about the use of the tool.

2. Page 11, search strategy: Usually, it is recommended not to use the publication type as limitation to be the most exhaustive. Moreover, to be “systematic”, the methodology of a systematic review requires consulting also grey literature (for example by checking the reference list of the retrieved article or by searching the congress abstract). Did the authors consult other sources than scientific database? It could be a limitation of your paper.

3. Page 11, line 215-217: What was the reliability of the coding? Did the authors calculated a Kappa coefficient to justify the validation of a random sample of 10%?

4. Page 12, line 242: See my comment above (Page9-10, #1).

5. Page 13, line 252-258: Regarding your “vote counting approach”, I am not sure to really understand the aim of this approach. Could you provide more information about this method and the rationale to use this approach?

Results

1. I am very surprised that the authors found only 6 studies. Have you been the most exhaustive? See my comment above (Introduction #4 and method #2).

2. The legend for table 1 is missing.

3. Because the primary outcome of this review is to assess the review quality, the result for this part should be more detailed. In my opinion, the result presented do not really represent the review quality of a systematic review. Please, revise.

4. In table 2, It could be relevant to add “empirical study”. The difference between review and “empirical study” is not clear in the result part.

Discussion and conclusions

1. I am wondering about your rationale and your methodology to assess the selective outcome reporting in the review. In my opinion, the presentation of an outcome or not, depends of the aim of the review. All the outcome presented in the empirical study are not always necessary presented in the review. Did the authors check the presence of a research protocol of the 6 reviews? It could be helpful to assess the selective outcome reporting (for example: to identify the difference between the outcome specify in the research protocol and the outcome presented in the review paper).

Thank you for the opportunity to review this paper. I look forward to receiving feedback about the

peer-review process of this manuscript.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: Yes: Austin L. Johnson

Reviewer #2: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files to be viewed.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org. Please note that Supporting Information files do not need this step.

Attachment

Submitted filename: PLOS One Review .docx

PLoS One. 2020 Oct 26;15(10):e0240032. doi: 10.1371/journal.pone.0240032.r002

Author response to Decision Letter 0


27 Jul 2020

Please also see the submitted file Response to reviewers. Text is pasted below.

- Please update the searches. The topic is hot and the literature certainly expanding very fast. This is therefore important to provide up to date results. Being up to date is all the more important as your paper will attract a lot of attention ;

Response: We have updated our searches to June 24, 2020.

- Please use AMSTAR 2, as far as possible, as a secondary outcome (as suggested by reviewer 2) ;

Response: AMSTAR 2 has now been applied and is available in S3 Data extraction.

- Please extract information by two reviewers in an independent manner. A second reviewer must extract the information at the study level and in case of disagreement a third reviewer must arbitrate. This must be done for all the studies and not only for a sample. Vote counting is not an optimal method and therefore subject to bias during data extraction. As an editor for PLOS one I must ensure that the method is sound enough and this is an important shortcoming. Please also provide inter-rater agreement.

Response: Unfortunately, we do not have the resources do independent verification of extracted data for all primary studies. We have addressed this by revising the description of our study to that of a summary of systematic reviews. We acknowledge this as a limitation in the Discussion section. That being said, we extracted data at the review level in duplicate, but do did not calculate inter-rater agreement.

- You might be interested to map the evidence base using a figure similar to Figure 2 in this reference : doi: 10.1093/ije/dyx138. Importantly, you don't have to cite this study (I'm against encouraging for self citation) but the visual display in figure 2 would be of interest in the context of your study (this is to give you an idea and it deserves surely a lot of adaptation to your specific study). To detail wether an empirical study is included in a given systematic review, a heatmap would be more readable than the current tables with numbers. In addition a plot mapping all existing empirical evidence used in the SRs across time / versus all empirical evidence available (with a line) / versus the number of empirical study in each SR would be very helpful. Indeed as it is presented now it is very difficult to understand for the reader.

Response: This is a very helpful graphic. We have incorporated a heatmap-type presentation of these findings (see Tables 2 and 3) and descriptions of them in the Results section. We believe that this and other improvements to the manuscript greatly improve its quality and reporting.

- You must detail in depth the criteria used to define "an empirical" study in your methods section.

Response: We did not actually choose which empirical studies to include; we only used those included in the reviews. This is clarified in the Methods section.

- You must provide details for selection of empirical studies in your flow charts.

Response: See response to previous comment.

- You must search for empirical studies that were not included in the review. Indeed, all systematic reviews are incomplete in comparisons with the others as you point. But it is also important to assess wether all these SR taken together were able to find all the available empirical studies (see my comment above about the suggested figure).

Response: Our goal was to assess the reliability of reviews based on defined criteria. While we feel it is important to ensure reviews capture relevant studies, to do a separate systematic review for all empirical studies would be outside the scope of the present study.

- Please make a specific section "changes to the initial protocol" to make it clear that these changes were performed during the peer review process.

As we did not make the changes suggested above, we did not include a section "changes to the initial protocol".

- Last please review the abstract using PRISMA statement for abstracts. Please, make sure that sufficient details are provided in the abstract and please write also a few words about the limitations of your study to avoid any over-interpretation of the findings.

Response: We have made the recommended changes to the abstract.

I appreciate that the changes needed are extensive but these changes will surely help to improve the paper.

Response: We are grateful for your careful review and suggestions.

5. Review Comments to the Author

Reviewer #1: Carras et al. conducted an interesting study analyzing the quality of evidence for gaming disorder and found that no reviews were classified as reliable. This topic is of great importance and I commend the authors for taking on this task. I additionally commend the authors for following a pre-registered protocol, for reporting results according to the PRISMA reporting standards, and for making all materials available online through OSF. Overall, this is a robust study and, with minor corrections, should be considered for publication.

Response: Thank you.

- The introduction and discussion is well written and appropriately introduces and discusses the topic at hand and the importance of having high-quality reviews to provide reliable evidence for healthcare and policy decision-making. I am pleased the authors have made all materials and protocols available online. This improves the transparency and reproducibility of their work.

Response: Thank you.

- Line 135 says Open Science Foundation, but the correct name is Open Science Framework.

Response: We have made that correction.

- Please provide a statement regarding IRB approval.

Response: We have added that on page 14 at the end of the Methods section.

- While I agree the with the classification criteria chosen in this study, it would benefit from further explanation why these 6 aspects of the criteria were chosen. The authors mention Mayo-Wilson and colleagues as part of their reasoning, yet Mayo-Wilson and colleagues reference including components of the Critical Appraisal Skills Programme (CASP), the Assessment of Multiple Systematic Reviews (AMSTAR), and the Preferred Reporting Items for Systematic Reviews (PRISMA).

Response: Our goal was to apply an existing approach to determining reliability of reviews to an entirely new topic area. We described this in more detail on pages 9-10 and included the full AMSTAR 2 assessment in S3 Data extraction.

- The second line in the conclusion has an extra space.

Response: We made sure there were no extra spaces.

While the authors mention systematic reviews should follow existing standards and should be improved, this review could benefit from providing a few more concrete suggestions on how to do so.

Response: We have made additional suggestions as to how systematic review authors could improve their reviews on pages 29-30.

Reviewer #2: Evaluating the quality of evidence for gaming disorder: A systematic review of systematic reviews of associations between gaming disorder and depression or anxiety

This study attempts to present a review of the quality evidence of systematic review for gaming disorder. The topic is interesting, but the rationale, the aim and the methodology of the study proposed need to be clarified.

Response: We have revised the manuscript extensively to conform with the reviewer's suggestions.

Comments

Title

In my opinion, the proposed title is inconsistent with the main aim of the study. After reading all the manuscript, I am wondering if your main research question is relative to selective reporting of outcomes, or if it rather of focused on the quality of evidence in systematic reviews.

Response: The title now reflects that we are summarizing systematic reviews. However, our research questions remain unchanged. Our first question addresses quality, while subsequent questions address aspects of quality that may be lacking, such as conflation of domains of the exposure (gaming disorder/IGD/etc) and selective outcome reporting.

Abstract

Currently, there are opportunities to improve the understanding of the abstract. The terminology used is not clear and the aim of the study is not clearly defined either.

Response: The abstract has been extensively revised and reported in accordance with the PRISMA for Abstracts reporting guideline.

Introduction

In my opinion, the rationale for this study could be improved. This section should be edited explaining the background to the study, an accurate summary of the existing literature and why this study is necessary.

Response: We have revised the Introduction extensively to conform with the reviewer's suggestions.

Some specific comments:

1. I think the second part of the introduction “Synthesizing evidence from systematic reviews” needs to be improved and reworked. Moreover, I do not totally agree with your statement (line 85-90), they seem to suggest that the reliability of a systematic review is only due to the definition of an outcome. Please, revise your definition.

Response: The revised Introduction section now provides a clear and comprehensive description of various relevant potential sources of bias in systematic reviews and the revised Methods section now clarifies how our study assesses reliability of systematic reviews.

2. In my opinion, the figure 1 should not be presented in the introduction part but in the methodological part as a rational to characterize an outcome.

Response: Figure 1 is an important piece of background information that many who study gaming disorder but are not systematic review specialists may not be familiar with. It describes how outcomes are specified in general. We have retained this in the Introduction but have added a second figure to the Methods section that more clearly describes how we defined the broad domains of gaming disorder, depression and anxiety in our data extraction.

3. Regarding your sentence in line 96-98, I think that to claim “Reviews that actively minimize bias and are reported transparently can be considered high-quality” could be (too) over-simplifying the methodological quality of a well conduct systematic review.

Response: We have removed that statement.

4. In my opinion, the rationale for this study is not clear (and accurate). This section should explain the background to the study, its aims, an accurate summary of the existing literature and why this study is necessary or its contribution to the field. Specifically, authors should describe the rationale for the analysis in the context of what is already known.

Response: We have extensively added text to the Introduction to address these issues. As we explain in the Introduction on page 7 and the Discussion on page 28, our study is the first comprehensive examination of selective outcome reporting in systematic reviews of gaming disorder, a new clinical entity. There have been no prior studies focusing on any type of bias assessment in systematic reviews of gaming disorder.

Methods

1. Page 9-10: Why did the authors choose the scale proposed by Mayo-Wilson and colleagues? Some others tools could be more suited (for example the tool AMSTAR2). I am very surprised by the way you classified the review as reliable. Please provide a rationale and more explanation about the use of the tool.

Response: As we mention in our response to Reviewer 1, our goal with this was to apply an existing approach to determining reliability of reviews to an entirely new topic area. We adopted the definition of reliable systematic reviews that has been used quite widely. We described this in more detail on pages 9-10 and included the full AMSTAR 2 assessment in S3 Data extraction.

2. Page 11, search strategy: Usually, it is recommended not to use the publication type as limitation to be the most exhaustive.

Response: We recognize that the use of review publication type filters is a limitation; we mention it as a limitation in the Discussion section on page 29.

Moreover, to be “systematic”, the methodology of a systematic review requires consulting also grey literature (for example by checking the reference list of the retrieved article or by searching the congress abstract). Did the authors consult other sources than scientific database? It could be a limitation of your paper.

Response: We have improved our methods and now include a grey literature search, as described on page 9.

3. Page 11, line 215-217: What was the reliability of the coding? Did the authors calculated a Kappa coefficient to justify the validation of a random sample of 10%?

Response: We did not estimate Kappa coefficients, but did a few rounds of pilot-testing. Estimation of Kappa coefficients is not widely recommended for screening and data extraction.

4. Page 12, line 242: See my comment above (Page9-10, #1).

Response: Our revised manuscript clarifies that our focus on quality assessment (pages 9-10) addresses risk of bias.

5. Page 13, line 252-258: Regarding your “vote counting approach”, I am not sure to really understand the aim of this approach. Could you provide more information about this method and the rationale to use this approach?

Response: Instead of vote counting, we present tallies of positive, negative, and null results for transparency. These can be compared directly to results found in the individual studies of the reviews we summarized.

Results

1. I am very surprised that the authors found only 6 studies. Have you been the most exhaustive? See my comment above (Introduction #4 and method #2).

Response: Yes, our search was comprehensive, including a search of multiple electronic databases and a grey literature search. We have now updated our search and found one more relevant systematic review. We also present a list of excluded studies with reasons for exclusion on the OSF webpage for the project.

2. The legend for table 1 is missing.

Response: All tables now have legends.

3. Because the primary outcome of this review is to assess the review quality, the result for this part should be more detailed. In my opinion, the result presented do not really represent the review quality of a systematic review. Please, revise.

Response: This section has been extensively revised. We have clarified the assessment of review quality (reliability).

4. In table 2, It could be relevant to add “empirical study”. The difference between review and “empirical study” is not clear in the result part.

Response: We now refrain from using the ‘empirical’ when denoting a study or a review because it is an ambiguous term. Early in the Methods section, we define the eligible systematic reviews and the primary studies they included as ‘reviews’ and ‘studies’, respectively.

Discussion and conclusions

1. I am wondering about your rationale and your methodology to assess the selective outcome reporting in the review. In my opinion, the presentation of an outcome or not, depends of the aim of the review. All the outcome presented in the empirical study are not always necessary presented in the review. Did the authors check the presence of a research protocol of the 6 reviews? It could be helpful to assess the selective outcome reporting (for example: to identify the difference between the outcome specify in the research protocol and the outcome presented in the review paper).

Response: In the Introduction, we have bolstered the argument for the need for assessing selective outcome reporting. One problem that we were trying to address is the lack of attention to standards for systematic reviews in this area. Of the now seven systematic reviews included in the manuscript, only one included a protocol, and indeed this was the only review to specify which outcomes would be reported and to report any non-positive associations. We have noted the lack of pre-specifications in the AMSTAR 2 reporting now found in S3 Review data extraction.

Thank you for the opportunity to review this paper. I look forward to receiving feedback about the peer-review process of this manuscript.

Response: Thank you for your helpful comments.

Attachment

Submitted filename: Response to reviewers.docx

Decision Letter 1

Florian Naudet

2 Sep 2020

PONE-D-19-32964R1

Evaluating the quality of evidence for gaming disorder: A summary of systematic reviews of associations between gaming disorder and depression or anxiety

PLOS ONE

Dear Dr. Colder Carras,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

I would like to thank the 2 reviewers. Thank you for the changes that were made. One of the two suggest that there are still important points to address and I agree. In addition, Table 2 needs to be attached as a Figure. Idem for table 3. Else, I suspect that the editorial staff will not be able to used colors in a Table (and with such colors, it is now rather a figure than a table).

Please submit your revised manuscript by Oct 17 2020 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

We look forward to receiving your revised manuscript.

Kind regards,

Florian Naudet, M.D., M.P.H., Ph.D.

Academic Editor

PLOS ONE

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

Reviewer #2: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: (No Response)

Reviewer #2: This study attempts to present a review of the quality evidence of systematic review for gaming disorder. I also reviewed a previous version of this manuscript. The authors did a good job in revising the manuscript, but I still have some minus remaining comments.

Comments

1. There is a syntax error in the sentence on line 167 page 9.

2. The authors have chosen to use your own definition using various sources to determine a reliable systematic review. Therefore, I think it would be useful for a better understanding of all, to better define your 6 criteria. For example: how did you judge the criterion 'comprehensive literature search'? Did you give a yes when the authors had consulted at least one database and had not limited their searches in English or other criteria were taken into account? The remark is valid for all items. How did you assign yes and no to the different items for each systematic review?

3. Thank you for appending the search strategy used in Medline. Could you add the one used in Psycinfo to be completely transparent?

4. Finally, the authors used the AMSTAR2 tool to assess the methodological quality of systematic reviews as outlined in the method. However, the results are not presented even briefly in the results section, let alone discussed in the discussion. A commentary on these results would be very interesting and could allow a comparison of the results with the existing literature on this topic.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2020 Oct 26;15(10):e0240032. doi: 10.1371/journal.pone.0240032.r004

Author response to Decision Letter 1


14 Sep 2020

We respond here to revision requests and suggestions, but this information can also be found in the file "Response to reviewers". Thank you again for your assistance in improving our manuscript for publication.

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: (No Response)

Reviewer #2: This study attempts to present a review of the quality evidence of systematic review for gaming disorder. I also reviewed a previous version of this manuscript. The authors did a good job in revising the manuscript, but I still have some minus remaining comments.

Comments

1. There is a syntax error in the sentence on line 167 page 9.

Response: Thank you for bringing this to our attention. We have fixed this error.

2. The authors have chosen to use your own definition using various sources to determine a reliable systematic review. Therefore, I think it would be useful for a better understanding of all, to better define your 6 criteria. For example: how did you judge the criterion 'comprehensive literature search'? Did you give a yes when the authors had consulted at least one database and had not limited their searches in English or other criteria were taken into account? The remark is valid for all items. How did you assign yes and no to the different items for each systematic review?

Response: We have clarified the definition of reliability. As we stated in the opening sentence of the section “Assessment of reliability of reviews”, this definition was not created de novo, but was adapted from one developed and published by Cochrane Eyes and Vision. Further, we have made changes throughout the manuscript to avoid confusion between “reliability” and “quality”.

3. Thank you for appending the search strategy used in Medline. Could you add the one used in Psycinfo to be completely transparent?

Response: We have now also appended the search strategy that we used for searching PsycInfo (in S2 (Search Strategies)).

4. Finally, the authors used the AMSTAR2 tool to assess the methodological quality of systematic reviews as outlined in the method. However, the results are not presented even briefly in the results section, let alone discussed in the discussion. A commentary on these results would be very interesting and could allow a comparison of the results with the existing literature on this topic.

Response: As mentioned in our response to your comment #2 above, we have clarified the distinction between ‘reliability’ (assessed using a few items from AMSTAR and other tools) and ‘quality’ (assessed using all AMSTAR items). The focus of this paper is on reliability of reviews, so we describe the AMSTAR items relevant to reliability in the main paper, and we include the results of the full AMSTAR assessment in the appendix table (S3).

Attachment

Submitted filename: Response to reviewers.docx

Decision Letter 2

Florian Naudet

18 Sep 2020

Evaluating the quality of evidence for gaming disorder: A summary of systematic reviews of associations between gaming disorder and depression or anxiety

PONE-D-19-32964R2

Dear Dr. Colder Carras,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Florian Naudet, M.D., M.P.H., Ph.D.

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #2: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #2: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #2: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #2: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #2: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #2: I have no further comments. Good work.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #2: No

Acceptance letter

Florian Naudet

24 Sep 2020

PONE-D-19-32964R2

Evaluating the quality of evidence for gaming disorder: A summary of systematic reviews of associations between gaming disorder and depression or anxiety

Dear Dr. Colder Carras:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Pr. Florian Naudet

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 PRISMA Checklist. PRISMA checklist for reporting of our systematic review.

    (PDF)

    S1 Protocol. PROSPERO registration for our systematic review protocol.

    (PDF)

    S1 Search Strategy. Search strategies.

    (PDF)

    S1 Data Extraction. Data extraction at the review level, including AMSTAR 2.

    (XLSX)

    S1 Dataset. Complete analysis dataset containing extracted and derived variables.

    (DTA)

    S1 Output. Output of analysis.

    (DOCX)

    Attachment

    Submitted filename: PLOS One Review .docx

    Attachment

    Submitted filename: Response to reviewers.docx

    Attachment

    Submitted filename: Response to reviewers.docx

    Data Availability Statement

    All relevant data are within the manuscript and its Supporting Information files.


    Articles from PLoS ONE are provided here courtesy of PLOS

    RESOURCES