Skip to main content
Royal Society Open Science logoLink to Royal Society Open Science
. 2025 Mar 5;12(3):241248. doi: 10.1098/rsos.241248

The academic impact of Open Science: a scoping review

Thomas Klebel 1,, Vincent Traag 2, Ioanna Grypari 3,4, Lennart Stoy 5, Tony Ross-Hellauer 1
PMCID: PMC11879623  PMID: 40046663

Abstract

Open Science seeks to make research processes and outputs more accessible, transparent and inclusive, ensuring that scientific findings can be freely shared, scrutinized and built upon by researchers and others. To date, there has been no systematic synthesis of the extent to which Open Science (OS) reaches these aims. We use the PRISMA scoping review methodology to partially address this gap, scoping evidence on the academic (but not societal or economic) impacts of OS. We identify 485 studies related to all aspects of OS, including Open Access (OA), Open/FAIR Data (OFD), Open Code/Software, Open Evaluation and Citizen Science (CS). Analysing and synthesizing findings, we show that the majority of studies investigated effects of OA, CS and OFD. Key areas of impact studied are citations, quality, efficiency, equity, reuse, ethics and reproducibility, with most studies reporting positive or at least mixed impacts. However, we also identified significant unintended negative impacts, especially those regarding equity, diversity and inclusion. Overall, the main barrier to academic impact of OS is lack of skills, resources and infrastructure to effectively re-use and build on existing research. Building on this synthesis, we identify gaps within this literature and draw implications for future research and policy.

Keywords: academic impact, Open Science, Open Access, FAIR Data, Citizen Science, scoping review

1. Introduction

In this paper, we aim to scope and summarize existing evidence on the academic impact of Open Science. Open Science seeks to make research processes and outputs more accessible, transparent and inclusive, ensuring that scientific findings can be freely shared, scrutinized and built-upon by researchers and others [1]. Key practices include Open Access publishing (OA), where articles are made freely available to readers [24], open (or FAIR, meaning findable, accessible, interoperable and re-usable) sharing of data and code [58], transparency of processes and methods [912], transparent models of peer review [13,14] and engaging publics and others beyond academia (e.g. civil society, policymakers, industry) in participatory research processes [1517].

In promoting these distinct practices, the Open Science (OS) movement has developed as a synthesis of what Field [18] rightfully terms ‘a constellation of communities, rather than a single, homogeneous one’ (p. 149). Open Science’s roots can be found in the foundational Open Access statements of Budapest [4], Berlin [19] and Bethesda [20], all focusing on access to knowledge and highlighting OA’s potential to democratize and increase use of knowledge. Rising awareness of poor levels of robustness of results [21], labelled a ‘reproducibility crisis’ in some disciplines in the mid-2010s [22], also led to emphasis on transparency of methods and reporting, as well as availability of research materials. The latter has been further boosted by a focus on enabling reuse of research data, code and other materials, most notably via the FAIR agenda among researchers and infrastructure providers [8,23]. Meanwhile, quasi-independent movements towards Citizen Science (CS) and participatory research have tended to be incorporated under the umbrella of OS, given alignment of aims such as public engagement and participation [24,25].

This thumbnail sketch already indicates that the aims of Open Science are far from uniform. Reviewing the OS literature in 2014, Fecher & Friesike [26] identified five ‘schools of thought’ in advocating for Open Science: to make knowledge freely available, increase efficiency of knowledge production, create open infrastructures and tools, make science accessible for the public, and develop alternative systems for evaluation and assessment. Since then, the picture has only become more complex (for example, a notable omission in the Fecher & Friesike paper is transparency, quality and robustness of methods and results). Recognizing such difficulties, UNESCO [24] attempted to establish global consensus on the OS programme, detailing principles including transparency, equality, responsibility, collaboration, flexibility and sustainability, and citing quality and integrity, collective benefit, equity and fairness, and diversity and inclusiveness as core values.

As a constellation of communities with often distinct motivations, and requiring support and action from various actors with each their own priorities (research communities, research performing organizations, research funding organizations, publishers, governments, as well as individual researchers), OS has what we recognize as an ‘interpretative flexibility’ [27]. Its complex meanings and interpretations among different groups need not always align, and along the various routes to implementation, unintended consequences may be expected [28,29]. This complexity, along with the heavy investments required to realize OS, necessitate careful monitoring. Moreover, OS as a transformative change may require not only monitoring of new indicators, but a different monitoring approach altogether [30]. Yet thus far, large-scale monitoring of OS (e.g. the EC’s Open Science Monitor [31], and various national monitors [3234]) seem mainly focused on understanding the spread of open practices (uptake), rather than whether that uptake is having the downstream real-world effects envisioned by the various manifestos and declarations made by OS communities. Understanding the latter is essential for policymakers, stakeholders and researchers to evaluate the effectiveness of activities and to make informed decisions.

Understanding the real-world change resulting from projects or programmes is the task of impact evaluation. We define impact as ‘long-lasting, elementary and wide-spread change’ upon academia, society or the economy which may be either ‘direct or indirect, intended or unintended, [and] relate to behavioural and/or systemic changes’ [35]. In the realm of policy evaluation, impact usually denotes population level effects, which result from outputs and intermediary outcomes of a given intervention [36]. Mapping the chain of effects from intervention to impacts is done through frameworks such as intervention logics, which aim to explicate the causal mechanisms at play.

In this scoping review, we consolidate evidence to date for the academic impacts of Open Science. We define academic impact with Ravenscroft et al. as ‘the impact that scientific research has within the academic sphere’ [37], hence distinguishing it from broader impacts upon society or the economy.1 Related to the above discussion, we consider the effects of Open Science on research quality (standards of research processes and products), trust and integrity (i.e. honest and verifiable methods), equity (fairness, equal participation, minimizing biases), collaboration (work among research groups), efficiency and productivity (increased speed or economic effectiveness), reproducibility, reuse of research outputs, as well as citations (included since citations are often taken as direct indicators of academic impact).

We follow the PRISMA Extension for Scoping Reviews methodology (PRISMA-ScR) [39] to scope, appraise, summarize and synthesize knowledge from existing literature demonstrating the academic impact of OS in general, as well as that of its constituent aspects: OA, Open and FAIR Data (OFD), Open Methods, Open Code/Software, Open Evaluation and CS. Given this is a scoping (not systematic) review, we do not aim to provide a systematic assessment of certain effects and weigh the evidence and critically assess questions around causal inference. Nonetheless, we appraise the literature and provide indications of strengths and weaknesses, also in terms of causality.

Our main research question (RQ1) is: what evidence exists in the literature regarding the effect of OS on the academic impact of research? Secondary research questions are:

  • SRQ1: what types of positive or negative, direct or indirect academic impact are observed?

  • SRQ2: what kinds of mechanisms produce them?

  • SRQ3: what specific enabling and/or inhibiting factors (drivers and barriers) are associated with these impacts?

  • SRQ4: what knowledge gaps emerge from this analysis?

2. Methods

To address these research questions, the study used the PRISMA framework to align study selection with the research question and followed the relevant aspects of the PRISMA Extension for Scoping Reviews (PRISMA-ScR) to ensure thorough mapping, reporting and analysis of the literature [39]. The study proceeded in four phases: identifying potentially pertinent literature, screening and selecting for eligibility, data-charting included studies, and analysing and reporting our results. Our methodological protocol was pre-registered on 31 October 2022 [40], with an addendum further specifying processes for ‘snowball’ searching and identification of relevant ‘grey literature’ sources published on 29 June 2023 [41]. Changes to pre-registration are detailed in electronic supplementary material, S1.

2.1. Identification of relevant studies

Search was conducted for relevant peer-reviewed literature published in English from 1 January 2000 using Scopus and Web of Science.2 Relevant keywords were assembled and prioritized through search piloting prior to pre-registration (see electronic supplementary material, S1 for more details). The search strategy was developed iteratively. First, relevant high-level terms were defined for ‘impact’ generally (i.e. impact, effect, outcome) and academic impact specifically (i.e. efficiency, productivity, quality, education, reproducibility, reuse, citations, collaboration, equity, diversity and inclusion). From this, the following primary search strings for Web of Science (WoS) and Scopus database searches were specified (see table 1), with database queries run and search result retrieved on 2 November 2022. Across both databases a total 11 420 initial results were retrieved.

Table 1.

Primary search strings for Web of Science (WoS) and Scopus. Explanation of abbreviations: TI = title, AB = abstract, TS = topic, TITLE-ABS = title and abstract, TITLE-ABS-KEY = title, abstracts and keywords.

Web of Science

[to be run in all databases, for time 01-01-2000 to 31-12-2022]

(TI= (‘open scien*’ OR ‘science 2.0’ OR ‘open data’ OR ‘FAIR data’ OR ‘open access’ OR (‘open code’ OR ‘open software’ OR ‘open tool*’) OR ‘open method*’ OR ‘citizen science’ OR ‘open peer review’ OR ‘open metric*’) OR AB= (‘open scien*’ OR ‘science 2.0’ OR ‘open data’ OR ‘FAIR data’ OR (‘open code’ OR ‘open software’ OR ‘open tool*’) OR ‘open method*’ OR ‘citizen science’ OR ‘open peer review’ OR ‘open metric*’ OR ‘open access publ*’ OR ‘open access paper*’ OR ‘open access journal*’ OR ‘open access book*’)) AND TS = ((impact* OR effect* OR outcome*) AND (quality OR citation* OR integrity OR equi* OR collaborat* OR trust OR efficien* OR re-us* OR reus* OR productiv*))

Scopus

TITLE-ABS (‘open scien*’ OR ‘science 2.0’ OR ‘open data’ OR ‘FAIR data’ OR (‘open access’ W/1 publ* OR paper* OR journal* OR book*) OR (‘open code’ OR ‘open software’ OR ‘open tool*’) OR ‘open method*’ OR ‘citizen science’ OR ‘open peer review’ OR ‘open metric*’) OR TITLE (‘open access’) AND TITLE-ABS-KEY ((impact* OR effect* OR outcome*) AND (quality OR citation* OR integrity OR equi* OR collaborat* OR trust OR efficien* OR re-us* OR reus* OR productiv*)) AND (PUBYEAR> 1999) AND (LIMIT-TO (LANGUAGE , ‘English’))

The next phase involved identifying further literature via (i) ‘snowball’ search, i.e. analysing relevant citations to and from already included studies (via OpenAlex), and (ii) ‘grey literature’ search for non-peer-reviewed material from websites (including the EC, UNESCO, OECD, for example), preprint servers (OSF preprint search) and web search (Google). Full documentation, code and data for this phase are shared with the full dataset [42].

2.2. Selection of eligible studies

We used the PRISMA-ScR checklist (electronic supplementary material, S2) for the selection process. The following inclusion criteria were used for both title/abstract and full-text screening:

  • Studies including evidence of the academic impact of OS3 (i.e. OA, Open/FAIR Data4, Open Methods, Open Code/Software, CS5, or Open Evaluation)

  • Conducted internationally or nationally

  • Published from 1 January 2000 until the date of search

  • Text in English

  • Full-text available

  • Study is either a research article, review article, conference paper, or other peer-reviewed output, or a grey literature study from a recognized stakeholder

  • All methodologies (quantitative, qualitative, mixed, etc.) eligible

The primary focus was on original studies. However, we included review articles where the synthesized findings enabled broader conclusions in the literature. For example, the inclusion of reviews on the Open Access Citation Advantage, where the literature is particularly dense.

For title/abstract screening, following an initial screening of the WoS/Scopus results to exclude clear false positives, results were merged and de-duplicated, leaving 4556 initial records. Titles and abstracts were then screened by two researchers, the first coding items either ‘yes’, ‘no’ or ‘unsure’ for inclusion, and the second reviewing all results judged ‘unsure’ to either include or exclude. The aspect of OS addressed by the study (whether Open Access, Open/FAIR data, etc.) was also recorded at this stage.

The full-texts of the 541 remaining studies were then sought using all available means (i.e. search via library log-in credentials, inter-library loans and emailing authors). A total of 541 full-texts were retrieved and stored in a shared folder using the open source Zotero reference management software. Full-texts were then screened and assessed for inclusion by one researcher, following which 301 total studies remained.

Title/abstract screening of snowball and grey literature identified 2664 studies, reduced to 184 following full-text screening. Across all searches, 485 relevant studies were identified for inclusion in this Scoping Review. The full process of selection is summarized using the PRISMA-P chart below (figure 1).

Figure 1.

PRISMA-P chart of selection process.

PRISMA-P chart of selection process.

2.3. Extraction of data

Studies were assigned to individual co-authors based on their relevance to specific aspects of OS. We used a collaborative spreadsheet, hosted via MS Teams, to extract data for specific items including data sources, study aims, types of impact studied, key findings, coverage and our confidence assessment. The data-charting categories used are below in table 2.

Table 2.

Categories extracted from included studies in the data charting process.

data chart heading

description

author

name of author/s

date

date article sourced

title of study

title of the article or study

publication year

year that the article was published

publication type

journal, website, conference, etc.

DOI/URL

unique identifier

exclusion

out of scope, non-English, duplicate

justification

if a study was deemed to be out of scope, a justification had to be provided.

study details and design (if applicable)

type of study, empirical or review, etc.; notes on methods used in study (whether qualitative or quantitative, which population demographics studied, etc.)

types of data sources included

detail the data sources

study aims

overview of the main objectives of the study

relevance to which aspect of Open Science

Open Access, Open/FAIR Data, Open Methods, Citizen Science, Open Evaluation, Open Science General

relevance to which aspect of impact

quality, citations, integrity, equity, collaboration, trust, efficiency, productivity, reuse

key findings

noteworthy results of the study that contribute to the scoping review question(s)

coverage

optional field to note any relevant information about the level of coverage of the study, e.g. only specific countries, disciplines, demographics covered

confidence assessment

optional field to note any concerns about reliability/generalizability of findings (e.g. conflict of interest, potential biases, small sample sizes or other methodological issues) within the study

Periodic checks on quality and consistency of extraction were carried out by the lead author, with inconsistencies discussed in team meetings and revisions conducted based on this feedback. Part of this process involved the elaboration and re-categorization of our initial list of impacts (cf. table 2, ‘relevance to which aspect of impact’). We merged the terms ethics and integrity, as well as efficiency and productivity, since they refer to the same underlying concepts. Based on the extracted content, we also added the new category novelty, as well as expanding the scope of equity to equity, diversity, and inclusion.. See table 3 for the final list of categories and their definitions.

Table 3.

Categories/definitions of academic impact used for data charting and synthesis.

category

definition

citations

changes (positive or negative) to the citation measures of research outputs (raw or standardized citations, citedness)

collaboration

changes (positive or negative) in cooperation or teamwork regarding the disciplinary, institutional, demographic or geographic composition of researchers in research activities

efficiency and productivity

changes (positive or negative) in terms of resources, time or effort required for research activities

equity, diversity and inclusion (EDI)

changes (positive or negative) to inclusion, fairness, diversity and justice in research, including: distribution of resources, opportunities or credit; addressing systemic inequalities/discrimination, and inclusion of diverse perspectives or populations

ethics and integrity

changes (positive or negative) in accountability, objectivity, fairness, responsibility, or adherence to ethical and professional standards/principles in research processes

novelty

changes (positive or negative) in the development of new ideas, approaches, discoveries or insights

quality

changes (positive or negative) to the rigour, relevance, reliability or soundness of scholarly processes and outputs

reproducibility

changes (positive or negative) in the reproducibility, replicability, robustness or generalizability of research findings

reuse

changes (positive or negative) in the further uptake and exploitation of research outputs (e.g. findings, data, code, methods, materials) for new research activities

trust

changes (positive or negative) within the academic community to confidence in the integrity, reliability or credibility of researchers, research processes or research findings

2.4. Collating, summarizing and reporting the results

The data on all included studies from the data-charting process is available at [42]. Following the data-charting process, the study team discussed main emergent themes including impact types and mechanisms. We then delegated responsibilities for drafting individual sections into our results below, summarizing the extracted information. These results are described in relation to the research question and in the context of the overall study purpose. Following the presentations of results, we present a final synthesis across all studies including gap identification of areas where further research is required. Finally, we conclude with a more general discussion of our findings and implications for future research and policy.

3. Results

In this study, we systematically scoped the evidence to date regarding the academic impact of OS. Our analysis is based on 485 studies that provide evidence towards impacts of OS on academia. The largest share of the reviewed literature concerns the impact of Open Access (233 studies), followed by Citizen Science (129), Open/FAIR data (67). Fewer studies reported on the impacts of general forms of OS (21), Open Evaluation (16), Open Methods (10) and Open Code (9) (see figure 2A). The main type of impact reported concerns citations, but other types of impact, such as towards quality, efficiency and equity, were also found (figure 2B). Note that the sum of impact types is larger than the number of included studies, since some studies demonstrated multiple types of impact.

Figure 2.

(A) Number of studies reporting academic impact of Open Science by relevance to aspects of Open Science.

(A) Number of studies reporting academic impact of Open Science by relevance to aspects of Open Science. (B) Number of studies reporting academic impact of Open Science by type of impact. Note that the numbers reported in (B) do not sum to 485 because some studies reported multiple types of impact.

To provide further context, we also analysed the included studies according to their publication year, article type and the topic they covered (figure 3). The majority of studies investigated the impact of Open Science among research in the natural sciences, followed by the medical and health sciences, as well as cases covering multiple or unspecified fields. The median publication year was 2019, indicating a substantial growth of the literature in recent years, despite our main search only partially covering 2022. In terms of article types, the included studies were mostly research articles (84% of included articles), followed by review articles (8%), conference papers (4%) and other types of articles (preprints, book chapters and grey literature).

Figure 3.

Disciplinary coverage and publication year of included studies.

Disciplinary coverage and publication year of included studies. (A) Top-level Frascati classification. (B) Second-level Frascati classification. We collapsed categories with fewer than 10 included publications. (C) Publication year.

3.1. Open Access

Our review identified 233 relevant articles related to OA [44276], making it the most studied element of OS for academic impact. Of these, 160 studies investigate effects on citations and the so-called Open Access Citation Advantage (OACA). Other areas of inquiry include the effect of OA on quality (40 relevant studies), equity, diversity and inclusion (37), efficiency and productivity (31) and other areas (20). In the following, we review the literature along the following dimensions: Open Access Citation Advantage, equity in OA publishing, preprints and other changes to the landscape of scholarly communication brought about by OA publishing.

3.1.1. Open Access Citation Advantage

The effect of OA publishing on citations, especially regarding differences with closed access, has been studied extensively. A major discussion has been whether there exists an ‘Open Access Citation Advantage’ (henceforth OACA), the assumption being that easier access to literature leads to higher readership and subsequently higher citation counts. Given the large number of studies (160 studies identified as relevant to citations), we here mainly report on previously conducted reviews [103,150,209], complemented by more recent studies.

There is substantial heterogeneity in methodological approaches among the studies identified as relevant to the OACA in our review. Asai [209] identifies three types of approaches: (i) comparing measures of citations ‘between randomly sampled subscription and open access articles or journals’, (ii) employing econometric models with control variables, and (iii) examining ‘the changes in citation scores for individual journals that shifted from subscription to open access using time trend data’ (p. 108). In addition, most studies tend to focus on specific research fields or on certain modes of OA, typically using classifications like the following (cf. Piwowar et al. [266])6 :

  • Gold OA: immediate free access with publication costs typically covered by authors or institutions

  • Green OA: self-archiving of a free version in repositories, often with an embargo

  • Hybrid OA: subscription-based journals offering optional paid open access for individual articles

  • Bronze OA: free access provided by publishers without an explicit open license

  • Diamond OA: immediate free access with no charges to authors or readers, funded by institutions or consortia

Beyond differences in general methodological approaches, there are also substantial differences in how the two key concepts, citations and Open Access status, are measured. Citation impact is measured in a variety of ways, comparing raw or standardized citations, or comparing journal impact factors between OA and non-OA journals, or comparing a number of further available indicators that rely on citations [209]. Similarly, approaches to determine whether a given article is OA differ substantially among the studies included in our review. Even in established databases such as Web of Science and Scopus, definitions and operationalizations of OA were unclear as of 2015 [176]. While the introduction of the Unpaywall service2 in 2017 (an open database of OA research articles), the highly cited article by Piwowar et al. [266] and the subsequent availability of Unpaywall’s data have led to some standardization, earlier studies relied on very different ways of defining and selecting individual publications or journals as being OA.

A key OACA study is the systematic review by Langham-Putrow et al. [150], which synthesizes evidence for and against the OACA. The authors report finding substantial heterogeneity in how studies aim to measure the OACA, for which reason they did not conduct a quantitative meta-analysis. Moreover, they found remarkably high levels of risk of bias. In their assessment, only three of 134 studies [88,224,277] were found to have a low risk of bias across the domains ‘population’, ‘data collection’, ‘study design’ and ‘results’. One of the three studies found to have a low risk of bias detected an OACA in general, one reported an OACA in subsets of data and one found no OACA. Disregarding risk of bias to report across all 134 relevant studies, the authors did identify an OACA in general, reporting that 47.8% of reviewed studies found an OACA, with the remaining studies finding no OACA (27.6%) or only in subsets (23.9%) with respect to journal, discipline or time. The authors did not find differences in how often studies reported an OACA by type of OA, but did find that studies with broader disciplinary coverage tended to find an OACA more often.

The degree of risk of bias across the literature surveyed by Langham-Putrow et al. [150] shows that causal effects are difficult to substantiate. The main sources for risk of bias reported by the authors were poorly described samples or samples insufficient to support conclusions, and missing justification for the choice of study period (e.g. length of citation window). In addition, a large number of confounding factors have been suggested to affect estimates of the OACA, which were also apparent in our review (e.g. ‘journal impact factor, number of authors, length of article, type of study’). Besides the confounding factors reported by Langham-Putrow et al., the studies identified in our review indicate further factors which might bias causal estimates of the OACA [103]:

  • Selection bias: in the case of hybrid7 OA publishing, some claim that researchers might only choose to make their best works OA, which in turn would lead to increased citations based on publication quality, rather than OA status [130,236]. However, Gargouri et al. [216] find the OACA to be independent of self-selection, thus questioning the existence of this bias.

  • Early-view bias: another hypothesized effect is the early-view effect on citations. Publications made available as preprints (i.e. prior to publication in a peer-reviewed journal) might receive more citations, since citations usually take some time to accrue [236].

  • Research funding: research emanating from publicly funded projects might be of higher quality than the average research conducted in a field, due to selection effects at proposal stage. Since research funders often mandate OA, not controlling for funding could introduce spurious effects [113].

  • Journal prestige: a common approach in studies assessing the OACA is to compare journal impact factors between OA and non-OA journals. However, these studies often do not control for other factors influencing impact factors, such as the extent to which journals are established and central to subdisciplines. Comparing new (OA) with established (non-OA) journals regarding their citation performance ignores these aspects.

  • Other confounding factors discussed in the literature, such as the ‘absence of an appropriate control group of non-OA articles’ are summarized by Rovira-Esteva et al. [168].

Given this complexity, a clear evidence gap seems to be the lack of any systematic account of the causal pathways related to the OACA. A systematized model of known and potential causal pathways might bring clarity to the study of the OACA and inform design choices for future studies.

A shortcoming of the review by Langham-Putrow et al. [150] is the question of differences in OACA between modes of OA (hybrid, green, gold, bronze). Comparing outcomes on whether the OACA exists or not across OA modalities, Langham-Putrow et al. report finding no statistically significant effect. In their analysis, Langham-Putrow et al. merged hybrid and gold OA, because they report the two modes to be ‘sometimes conflated within the included studies’. Our review of the literature suggests that merging gold and hybrid OA masks an important difference, where hybrid articles tend to receive more citations than their closed counterparts, whereas gold articles tend to receive fewer.

To substantiate these claims, we conducted a post hoc analysis of all studies we identified which relate to the OACA. Given the complexities in assigning studies to different modes of OA and the lack of a thorough protocol, we provide a broad summary of our findings but no exact counts of outcomes.

The majority of studies find that gold OA journals have lower impact factors than closed-access journals (see for example [198,199,204,266]). This is especially salient in studies which compare multiple OA modes, where many find an OACA for hybrid and/or green OA, but not gold OA. However, the low citation rates of gold OA might be partly driven by other factors than OA, most prominently the fact that many gold OA journals are relatively new and thus have had less time to build reputation than closed-access journals [176].

By contrast, our analysis indicates that the majority of studies comparing citations received between hybrid OA and closed articles from the same journals find an OACA (see for example [85,137,262]; but see [140]). However, this effect might be partially driven by selection bias and effects of research funding.

Finally, studies assessing citations towards green OA articles tend to report finding an OACA (see for example [133,170,199,257,266,271]).

Taken together, we find clear evidence for differences in the OACA between modes of OA. Our results suggest that it might be futile to investigate and claim an overall OACA, but that efforts should probably focus on context-specific accounts of an OACA.

Lastly, the studies included in our review also reveal vastly different estimates of the size of the OACA. Studies with more elaborate methods that aim to make estimates of causal effects, such as the one by Staudt [156], tend to report low estimates of the OACA (in this case: 3%), compared with much larger estimates in other studies [85,170,225].

3.1.2. Equity in Open Access publishing

Prominent among financial models for OA publishing are author-facing charges, so-called article processing charges (APCs). This business model has been identified as being a threat to equity in publishing very early [259], but is still widespread today [70,160]. APCs have been found to be positively correlated with journal metrics [54,59,60,62,75,131,218], although the effect is not present in all fields [178]. This inequity is further aggravated by APCs due to global differences in purchasing power parity [46], restricting publishing in high-impact journals to those with sufficient funding. Waivers have been found to be ineffective at countering this issue, in particular for researchers from lower- and upper-middle income countries [68,158].

Investigating the difference in geographic diversity of authors between closed and OA journals, Smith et al. [65] report lower geographic diversity among authors in OA journals, suggesting a prohibitive effect of publication costs for scientists from the Global South. By contrast, there is evidence that OA is improving equity in readership. Analysing a large bibliographic dataset from 2010 to 2019, Huang et al. [182] report finding higher diversity of citations in terms of (among others) institutions, countries and fields of research for OA publications.

Overall, the dynamics around APC-based publishing create a barrier for researchers aiming to publish their work, stratifying global publishing even further and undermining initial goals of the OA movement in terms of democratization [223]. Knöchelmann [234] discusses these trends against the objective of Open Access to democratize access to knowledge, arguing that dominant solutions in the form of Global North-driven OA policies create an accessibility problem and therefore publication injustices for authors from other regions. Thus, even though OA increases equity in terms of physical access, it has been found to decrease equity on the side of publishing research.

3.1.3. Preprints

Another innovation aligned with Open Science is the use of preprints (public posting of non-peer-reviewed manuscripts, usually in advance of journal submission). Although preprinting has been common in disciplines like Physics (via the arXiv preprint server, established in 1991), rates have expanded rapidly across disciplines in recent years, especially during the COVID-19 pandemic [93,269,270]. We identified 14 studies reporting impacts on efficiency, quality and citations.

The use of preprints aims to enhance the efficiency (i.e. speed) of scholarly communication. Rates of improvement in dissemination time differ across disciplines, however. Abdill & Blekhman [270] reported that around two-thirds of bioRxiv preprints were eventually published in peer-reviewed journals, with a median time of around five and a half months. Studying arXiv preprints, Larivière et al. [63] reported a similar lag in Physics, but significantly longer times in Mathematics of over a year. Wang et al. [197] confirmed the latter, finding an average lag of 16.3 months in three subdisciplines of Mathematics between preprinting on arXiv and eventual article publication. The emergency response to the COVID-19 pandemic significantly reduced time lag, however, as Oikonomidi et al. [74] report a median lag of just two months for bioRxiv preprints on COVID-19 interventions.

Seven studies assess the quality of preprints compared with their published versions. Most focus on biomedical research, especially that related to the COVID-19 pandemic. Overall, these studies demonstrate that preprints which are eventually published within journals change relatively little, and are of similar or only somewhat lower quality across a range of outcome measures. These include: basic differences in text [92]; epidemiological estimates and expert assessment of quality in COVID-19 research [210]; study characteristics, outcomes and interpretative ‘spin’ [99]; validity of COVID-19 parameter estimations [228]; quality of reporting [93]; and numbers of figures/tables and conclusions reported [269]. Oikonomidi et al. [74] found, however, that one in six COVID-19 preprints did have at least one significant change in evidence reporting in their published version. In interpreting this evidence, it should be kept in mind that the included studies do not include comparison with preprints which did not later go on to be published (for obvious reason). This potentially leads to an overestimation of preprint quality (selection bias), where non-published preprints would presumably tend to be of lower quality on average.

Lastly, across various fields, published articles that were preprinted are associated with increased levels of citations compared with articles where no preprint was made available [205,230,250,264]. This literature makes no claims for direct causality, however, with potential for confounding factors like that those who preprint may be more likely to publish OA, or be generally more active in self-promotion, or that works of perceived higher potential impact are more likely to be preprinted.

3.1.4. Other changes in the scholarly publishing landscape

The advent of OA publishing has been accompanied by multiple changes in the landscape of scholarly communication and publishing (see e.g. Eysenbach [80]). While the internet and modes of online publication have enabled OA, the move to OA publishing has in turn led to new forms of publishing, such as OA mega-journals, predatory publishing and an uptake in preprinting, which in turn has been sparking new modes of organizing peer review (see e.g. F1000, or eLife from 2023).

A total of 33 studies assessed impacts on quality in scientific publishing. Due to lack of basic editorial practices, predatory publishing is by definition expected to lead to publications of lower quality. While empirical investigations of articles published in potentially predatory journals indeed find them to be of lower quality [175,194], these articles have been found to accrue much fewer citations than comparable articles, and therefore have a low impact on science [138]. The use of blacklists (such as Beall’s list) to identify predatory journals has been subject of substantial debate and can be considered a secondary impact of OA publishing on equity, given its purported effect on ‘divisiveness, discrimination and stigmatization’ [278].

Beyond predatory publishing, multiple studies have analysed the impact of OA publishing on article quality. While acceptance rates have been found to be higher among OA journals [152], there is conflicting evidence as to differences in actual article quality. Some authors have found no difference in quality between OA and non-OA articles [58,135,202], while others report higher quality among non-OA articles [96]. In addition, evidence on the effect of OA mega-journals on article quality is sparse [190]. Retraction rates have been found to be higher among OA mega-journals [151], but this is arguably related to better editorial practices rather than lower quality research, as indicated by the finding that OA journals provide more detailed information on the reason for retractions [78].

In addition, the move to OA publishing has changed the business models of publishers. Quantifying costs and benefits of OA business models and self-archiving, Houghton [127] reports ‘substantial net benefits in the longer term’ in terms of academic, societal and economic impact for ‘more open access’, due to increases in speed and breadth of access, as well as lower costs of publishing, driven by the move from print to electronic distribution [127]. A publisher’s revenue under gold OA depends on the number of accepted publications and the level of APCs, which might incentivize publishers to lower their acceptance criteria [75,240].8 This has given rise to the concerns about article quality discussed above, but also led to business models where OA mega-journals cross-subsidize more selective and thus less profitable journals from the same publisher [190].

3.2. Open/FAIR Data

Our review identified 67 studies which assessed the academic impact of Open Data [279345]. Of these, 25 studies reported impacts on data reuse, 22 studies reported impacts on citations, 10 each on ethics/integrity and reproducibility, seven on efficiency and productivity, six on quality and equity, diversity and inclusion each, with further impacts on collaboration, and trust (two each).

Overall, few publications that are included in the review provide concrete evidence of impact of Open Data. Various publications study effects of data sharing policies, instead of effects of data sharing itself. The most obvious direct effect of data sharing is whether the data are actually re-used, and most literature focuses on this. By contrast to OA, if data are not available at all, they cannot be re-used, whereas research that is not OA can still be read and cited. Therefore, most studies do not compare usage statistics with a possible counterfactual. Another sizable part of the literature analyses whether making data openly available has some effect related to the publication that makes the data openly available itself, most notably, studying an ‘Open Data’ citation advantage, similar to OA. Other studies focus instead on whether studies that make data openly available increase the robustness of the results, related also to questions of reproducibility.

3.2.1. Data reuse

In a wide variety of fields there are ongoing discussions of data sharing. A number of publications examine the use of datasets from existing repositories, such as in agriculture [339], biodiversity [105], ocean science [319], clinical trials [292,342], satellite imaging [302], environmental sciences [334], network sensing [306] and genomics [341]. Other publications introduce novel repositories, such as for topography [344], COVID-19 [333] or neurology [335,338] or at a specific university [308]. This does not represent an exhaustive list of repositories that are introduced, only publications that also report on actual usage.

Many papers discuss problems of data sharing, including challenges of common formats, standardization of vocabularies and ontologies [338]. Common standards also bring larger benefits for disciplinary specific repositories, such as the OpenNeuro repository [338] or the Genome Expression Omnibus repository [341], over general repositories, such as Zenodo, Dryad or Figshare. At the same time, Quarati & Raffaghelli [298], report that metadata quality is not associated with usage. Some argue for standards for citation of data [281], while others argue for also considering electronic supplementary material in considering data use [295].

Other papers discuss challenges in making data re-usable, as part of the findable, accessible, interoperable and re-usable (FAIR) definition [8]. Smale et al. [280] report a scarcity of published systematic evidence that data management plans (DMPs) have positive effects on data FAIRness. Bishop & Collier [303] found that researchers are influenced by ‘trusted brands’, such as well-known repositories, in trying to determine whether data will probably be re-usable or not. Bezuidenhout & Chakauya [305] report that reuse of data is constrained by limited resources in low/middle-income countries.

Qualitative data seems to be used most often for learning, not research, report Bishop & Kuula-Luumi [329], based on repositories in the UK and in Finland. Bishop [343] argues that sharing qualitative data is especially useful for teaching.

Several papers use theoretical models to explore the data sharing behaviour of researchers. Mueller-Langer & Andreoli-Versbach [320] found that authors may strategically delay publications to fully enjoy the benefits of studying the data for additional papers. Spiegelman [300] studied a model of data sharing in low- and high-quality science. They find that high-quality science is more likely to share data than low-quality science, and that data sharing can then be used as a marker of high-quality science.

3.2.2. Open Data citation advantage

Piwowar & Vision [290] is the key study concerning the Open Data citation advantage, and is cited by most other articles for evidence of an Open Data citation advantage. They study gene expression microarray data and find that publications that share data openly receive 9% more citations than publications that did not share data openly, while controlling for a number of relevant confounders. This contrasts an earlier finding of Piwowar et al. [345] which found a much larger increase of 69%. Piwowar & Vision [290] find that 6% of the citations were made in the context of data reuse, suggesting that much of the citation benefit comes from additional citations due to data reuse. The exact mechanism is not clear, but the presence of some effect seems quite convincing.

Leitner et al. [289] also observe that publications which share data openly are cited more frequently, but they only control for publication year, and not for a range of other confounders, thereby offering little evidence of a causal effect. Drachen et al. [331] find evidence for a citation advantage of about 15% in astrophysics, which is possibly mostly caused by citations to datasets being shared. McGillivray et al. [297] report that special data papers have a positive impact on citations of the original research, suggesting that additional citations might indeed be associated with data reuse. Colavizza et al. [332] report about 25% higher citations for PLoS and BMC papers that link to data in a repository compared with papers that report that no data are available.

Zhang & Ma [313] studied the effect of the introduction of a data sharing policy at a journal using a difference-in-differences approach. They found that the Open Data policy increased citations by 1−4 times. The estimate of receiving four times more citations seems extremely large, and quite different from the more conservative estimate of 9% of Piwowar & Vision [290]. Huang et al. [321] performed a very similar study on the same journal, but reported an increase in citations of 1.5−2 times. Christensen et al. [282] also used a journal policy change in data sharing requirements, but reported no statistical differences between citations to articles shortly before and after the policy change. However, they did report almost double citation rates for articles that actually share data, although this might be due to some confounding factors.

Kwon & Motohashi [310] suggest that there are two competing effects of sharing data on the accrual of citations: a positive credit effect and a negative competition effect. While the credit effect leads to an increase in citations to publications that share data, it might lead to greater competition by other researchers, thereby crowding out citations to the original research. Both effects are reported to be present, where in the short term publications that make data available are cited more frequently, due to the credit effect, but in the long term, they are cited less frequently, due to the competition effect. Zhang & Ma [313] perform a similar analysis, and also find that that citation benefit is positive initially, but becomes negative later on.

A number of other articles also study citations to publications that either share or use Open Data. AlRyalat et al. [336] found that publications that used Open Data from BioLINCC receive a higher number of citations. However, they do not compare their results with other publications, making it unclear whether this is a causal effect. Coady et al. [342] report no significant differences between citations of publications using BioLINCC data and citations of other publications supported by the National Heart, Lung and Blood Institute, suggesting that there might not be an effect. Heller et al. [340] study conference proceedings in medical image computing (MICCAI) and find that papers that use open, public data are cited (or mentioned—formal citations were sometimes missing) 60% more frequently than papers using private data.

Raffaghelli & Manca [301] found that reading or citing of the data does not seem to be associated with the FAIRness of the data. Robinson-García et al. [284] study the Data Citation Index provided by Clarivate. They find that ‘data citation practices are far from common within the scientific community, with a high rate of uncitedness (88%)’ (p. 2970). They also find different data citation practices, with the natural sciences tending to cite datasets, while the social sciences and humanities tend to cite publications that introduce the data. Aleixandre-Benavent et al. [325] study open data policies and data availability across journals, and find no association with a journal’s impact factor. Piwowar & Chapman [326] on the other hand do report an association of data sharing with a journal’s impact factor, in addition to author experience. Neither study makes any causal claim.

3.2.3. Reproducibility

One important reason for sharing data is that other researchers can then replicate the original research. However, research by Coady et al. [342] suggests that data are rarely re-used for this purpose.

Hardwicke et al. [288] studied whether the introduction of a data sharing policy at a journal affected the reproducibility of articles published in the journal. They reported that not all data appeared re-usable. Of the 35 articles for which data was re-usable, only 11 articles could be reproduced without author assistance, and an additional 11 could be reproduced with help. For the remaining 13 articles, at least one value could not be reproduced. Overall, 95% of the numbers produced in the articles could be reproduced (within a 10% error margin) using the open data, but it did take considerable time and effort. Hardwicke et al. [283] extended their previous study to articles with open data badges, and found similar results. Laurinavichyute et al. [330] also found that data sharing policies led to limited improvements in reproducibility. When code was shared, in addition to data, reproducibility greatly increased. Naudet et al. [291] analysed randomized controlled trials in BMJ and PLoS Medicine. Although only about half of the studies (17/37) shared data, most of those studies (14/17) could be reproduced but again required substantial effort and contact with the authors.

Nuijten et al. [314] compared papers and journals with data sharing (policies) with those without such data sharing (policies), and studied how they differ in terms of statistical inconsistencies. They found that data sharing (policies) do not affect statistical inconsistencies (as found through statcheck). Claesen et al. [293] corroborate this finding and also report no fewer inconsistencies when data are shared. Berberi & Roche [318] also report no association between data sharing mandates and article retractions or corrections, but this finding might suffer from selection bias [346].

One interesting issue raised by Thompson et al. [296] is that open datasets can be analysed multiple times, which, coupled with publication bias, may result in increases in false positives in the literature. Besides some potential technical fixes within the realm of null hypothesis testing, the authors suggest three alternatives: Bayesian analysis, predictive cross-validation with held-out data, or most radically, considering all research on open data as exploratory. Drewry et al. [347] discuss the benefits of Open Science practices in pharmaceutical research, such as SGC laboratories’ open sharing of chemogenomic data. They highlight enhanced collaboration, increased reproducibility and faster scientific progress as significant benefits.

3.2.4. Efficiency/productivity

Cannon et al. [309] found no effects of a new data sharing policy on submission rates, acceptance rates and peer-review times. Holt et al. [312] did find that the introduction of a data sharing policy increased the time editorial staff spent on processing a manuscript, which was mitigated after some modifications to the workflow. Grant & Hrynaszkiewicz [337] found that simpler data availability statements required less additional editorial time than data availability statements that pointed to public data.

3.2.5. Ethics and equity of data sharing

Cummings et al. [307] studied whether Open Data sharing makes a difference to participants consenting to take part in research, but found no such influence. They do raise one issue, namely that participants’ consent is usually only valid for the original study, while data sharing provisions allow the reuse of data for other purposes, without acquiring consent from the original participants, a point also made by Bishop [343] and Mozersky et al. [279] in the context of qualitative data. Although researchers often raise concerns of privacy and ethics in the context of qualitative data, research participants may be comfortable sharing for instance interview transcripts [316]. Nonetheless, there might also be challenges and ethical concerns in secondary analyses of qualitative data as these data are situated in their context, although with care, it might be possible to integrate and analyse qualitative data across projects [327], and in some fields, such as health science, the situatedness may be less salient in the research [279].

Xia & Liu [341] report a divide in the use of genomic data between the Global North and the Global South, with essentially none of the data being used by Latin America and Africa, indicating that benefits of shared data mainly accrue for the Global North.

Abebe et al. [317] argue that those whose data is being shared might not enjoy the benefits, and that a deeper understanding of, and engagement with, the local context is necessary to make data practices more equitable. A similar argument is made by Carroll et al. [311] who explored the intersection between data and tribal rights of indigenous people, arguing that indigenous people should regain sovereignty and governance over their data. Carroll et al. [324] argue for integrating the CARE Principles for Indigenous Data Governance (Collective Benefit, Authority to Control, Responsibility, and Ethics), in addition to the FAIR principles, when considering open data. Khalil et al. [304] plead for considering adopting ethical views on open science development to avoid potential issues when implementing open science.

Bezuidenhout et al. [285] report that access to data alone is not enough to make effective use of it. They highlight that ‘an emphasis on access fails to capture the social and material conditions under which data can be made usable, and the multiplicity of conversion factors required for researchers to engage with data’ (p. 473). Moreover, similar barriers may affect data sharing: ‘The absence of pre-publication data and the slow rate of academic publications diminish the impact that LMIC [low- and middle-income countries] research has in the global research community. Nonetheless, it is precisely because of the slow rate of publication that LMIC scientists are hesitant to share pre-publication data with anyone with whom they do not have a personal connection’ [305, p. 47]. Using an innovative approach, Shanahan & Bezuidenhout [328] report that online material is also not equally accessible around the world, not only due to connectivity issues such as low bandwidth, but also due to explicit geoblocking and restricted access.

3.3. Open Methods

We identified 10 studies related to Open Methods [348357]. Three studies investigated effects of Open Methods on methodological quality. Bakker et al. [348] investigated whether requiring explicit pre-data collection power analyses increased statistical power in reported results. The intervention seemed ineffective overall, however, with no difference in the planned sample sizes stated. In a related study, Bakker et al. [351] examined the impact of pre-registrations that were either structured (‘with detailed instructions’) or unstructured (‘minimal direct guidance’) to find that structured pre-registrations were more effective in reducing deviations in published studies. Studying the effects of registered reports on adherence to OS practices in psychological research, Susanin et al. [355] found an increase in ‘the use of data sharing, reporting of statistical form, and inclusion of start-stop rules’.

Next, related to quality and efficiency, Hocker et al. [357] investigated the use of an open tool for reuse of coding schemas for qualitative research, finding the tool improved both time taken to complete the task as well as coding accuracy.

Regarding quality and citations, Markowitz et al. [354] observed no effect of OS practices in communication research on citation rates or the prevalence of p-values just below p < 0.05, but they did note an increase in non-significant p-values, suggesting a potential impact on transparency and honesty.

Relating to citations, Schafmeister [352] studied whether publication of independent replications affected levels of citedness of original studies. They found that successful replication was associated with around a 5% increase, and unsuccessful replication with a 4% decrease in citations.

Finally, regarding reproducibility, Ebersole et al. [349] found that review of protocols for replication studies in advance of replication attempts had ‘little impact’ on levels of replication.

3.4. Open Code

Nine studies were identified addressing this theme [358366], five relating to efficiency gains in the production of research-enabling software via open source community efforts [367], three relating to a citation advantage for publications with Open Research Software [368] and two examining effects of Open Code on computational reproducibility.

Regarding efficiency, Ratib et al. [365] found that open source enabled faster development of tools that are better adapted to user needs (via users contributing to development, and larger numbers involved in development and testing). Blasco et al. [358] found that use of open data and algorithm development competitions in computational biology and bioinformatics leads to better performing algorithms, especially through quick ‘exploration of the solution space’ and creation of ensemble techniques through collaboration. Coetzee et al. [359] reviewed open source software and open data initiatives in the geospatial domain, concluding that open source geospatial software and open geospatial data have changed data collection (including via crowdsourcing), processing, analysis and visualization. Similarly, Ratib & Rosset [362], in their case study on the OsiriX Biomedicine tool, reported that its open source nature and embedding in a very responsive community equated to ‘software updates and new features at a rate that exceeded by far the rate of software updates in the industry’ (note that the authors do not quantify this claim, however). McCormick et al. [361] describe the open source ITK (Insight Toolkit) software, a ‘reproducibility verification infrastructure’ with community code review. They report efficiency gains (operationalized as fewer ‘fix-up commits’). Finally, Wallace et al. [363] report on development of RT-Cloud, an open source cloud-based Python software package for real-time functional magnetic resonance imaging (fMRI) experiments, finding efficiency gains in terms of ease of set-up/maintenance, reduced costs and ease of scaling.

Regarding impact on citations, Heumüller et al. [360] studied papers from Software Engineering conference and found a slight citation advantage for papers which had made software artefacts available over those which had not. However, Winter et al. [364] also studied levels of citations of articles with associated Open Code and found no general effect of Open Code on citations.

Finally, Schröder et al. [366] assessed the computational reproducibility of Jupyter notebooks associated with journal publications. They found that just 3 of 22 notebooks could be reproduced (i.e. ‘not only all cells run successfully but also the output is equal to the originally published version’), suggesting the need for improvements in code quality.

3.5. Citizen Science

We identified 129 studies as relevant to Citizen Science (CS) [43,369496]. More than two-thirds (89 studies) of these report on the impact of CS projects in ecology or environmental research, and especially in activities to monitor flora and fauna. The main types of academic impacts of employing CS in these monitoring activities relate to efficiency in data collection, as well as concerns over data quality. Academic impacts demonstrated by a range of studies are the increased temporal and spatial scales of monitoring activities enabled by CS methods, coupled with moderate to substantial cost savings. Many state that CS can be a useful complement to existing approaches [393,410,420].

A key question is which specific mechanisms drive the various forms of academic impact attributed to CS. While there are certain challenges related to organizing large groups of volunteers, it is not surprising that more people will be able to conduct monitoring activities on larger temporal and spatial scales. It is thus clear that there exist trade-offs ‘between cost-effectiveness and involvement (i.e. societal impact), as well as between data richness and data quality’ [463]. Besides the impacts on efficiency and quality in data collection, CS has also been suggested to lead to novel insights and challenge existing approaches, due to the diversity of backgrounds, perspectives and skills of citizens [399,404,426,429,465]. At the same time, CS activities seem to be concentrated in the Global North, limiting the global diversity [375,411,412,414,423,455]. Some researchers have raised ethical questions around how to properly credit CS volunteers [375,403,414].

3.5.1. Data quality

Earlier literature, such as the review by Dickinson et al. [402], reported quality issues with citizen-generated data. More recent reviews generally agree that CS projects can generate data of sufficient quality for monitoring activities [383,412,477]. As shown below in table 4, most studies we examined (46 studies) report that the quality of data gathered through CS is of the same, if not higher, as data gathered by professionals. However, as Aceves-Bueno et al. [477] make clear, there is not a single agreed upon definition of when data are deemed ‘reliable’ or of ‘sufficient quality’. Indeed, eight studies also report worse data quality and 24 report neutral or mixed impacts on quality, and data might be good enough for some purposes, but insufficient for others.

Table 4.

Overview of CS studies relevant to ‘quality’, and whether outcomes were positive, neutral/mixed or negative.

type of change

relevant studies

positive

Aceves-Bueno et al. [401]; Agostini et al. [416]; Aplin et al. [370]; Barrows et al. [492]; Biraghi et al. [410]; Branchini et al. [490]; Brown et al. [389]; Cruickshank et al. [468]; Crumley et al. [390]; Dallaqua et al. [446]; Danielsen et al. [460]; Delaney et al. [456]; Dodson et al. [436]; Facchinelli et al. [443]; Farhadinia et al. [407]; Fehri et al. [431]; Fehri et al. [475]; Flesch and Belt [419]; Fowler et al. [371]; Fuccillo et al. [386]; Hochmair et al. [440]; Jones et al. [459]; Kasten et al. [464]; Keeping et al. [395]; Lawson et al. [437]; Leocadio et al. [432]; Lin et al. [489]; Lovell et al. [380]; Manda et al. [384]; McKinley et al. [405]; Meentemeyer et al. [409]; Meschini et al. [470]; Njue et al. [412]; Petrovan et al. [491]; Porter et al. [466]; Pusceddu et al. [494]; Quinlivan et al. [383]; Shumba et al. [379]; Soroye et al. [462]; Starkey et al. [434]; Swanson et al. [373]; Van Der Velde et al. [422]; Van Emmerik et al. [425]; Vianna et al. [377]; Watson and Floridi [427]; Weigelhofer, Pölz, and Hein [415]

neutral/mixed

Beck et al. [429]; Bison et al. [421]; Buytaert et al. [413]; Cox et al. [442]; Crall et al. [387]; DiBattista et al. [418]; Dickinson et al. [402]; Engel and Voshell [495]; English et al.s [447]; Foster-Smith and Evans [485]; Gardiner et al. [452]; Giovos et al. [417]; Gunko et al. [449]; Harvey et al. [420]; Kallimanis et al. [467]; Kremen et al. [441]; Meixner et al. [469]; Périquet et al. [476]; Ratnieks et al. [433]; Shin et al. [398]; Soroye et al. [496]; Swan [426]; Thel et al. [392]; Tran et al. [439]

negative

Alabri and Hunter [438]; Galloway et al. [484]; Johansson et al. [394]; Robinson et al. [391]; Roy et al. [445]; Smale et al. [473]; Theobald et al. [448]; Yang et al. [388]

Many CS studies focus on biology/ecology, and problems of data quality often relate to less accurate identification of species, or over- and under-reporting of the abundance of certain species. In some instances, more common, or more ‘likeable’ species are more likely to be reported (e.g. [468,476]). In other cases, CS seemed to be biased towards over-reporting rare species (e.g. [398,417,484]), possibly due to the greater ‘surprise’ of observing such a species. Other rare events might also be more likely to be observed by CS [434,443]. The results might vary according to which people are involved in data collection [430].

Aceves-Bueno et al. [477] suggest that studies might be overstating the accuracy of CS, because many studies have been explicitly designed to test the accuracy. Moreover, it is possible that positive outcomes are more likely to be reported in the literature; Theobald et al. [448] report that only 12% of the surveyed projects have contributed data to peer-reviewed articles (see also Conrad & Hilchey [374]). In addition, these projects are more likely to have invested a lot of effort in training and gathering good data. Indeed, training [464] and protocols [495] in data collection seem to be important determinants of data quality [384,388,411,433,474]. On the other hand, the literature may underestimate the accuracy of CS, because many studies ignore uncertainty, or potential errors, in professionally gathered data [477]. As several others also argue [378,482,483], the study design might be more relevant to data quality than the use of CS per se.

3.5.2. Efficiency

Sauermann & Franzoni [424] estimate the average contributions of volunteers to represent a value of about $200 000 during the first half year of a project. Generally, CS was found to have great cost savings compared with traditional surveying for monitoring [378,407,452,454,460] or experiments [415,493]. Most studies report higher start-up costs, but lower running costs, so that longer studies [401] or larger studies [453] could be more efficient. Indeed, CS was found to be most effective at large spatial and temporal scales [411,443,465,478]. Combining traditional and CS monitoring might be more effective [434,444], and combining CS with AI possibly even more so [385].

3.6. Open Evaluation

We identified 16 articles relating to Open Evaluation [14,497511], i.e. open peer review, including disclosure of reviewer identities (Open Identities) and publication of review reports (Open Reports). We here report those findings under the themes of effects on quality (eight studies), ethics and integrity (five studies), efficiency and productivity (five studies), citations (three studies) and equity, diversity and inclusion (two studies).

3.6.1. Quality

Eight articles investigated effects on review quality of models of Open Identities. In Bianchi & Squazzoni’s [497] theoretical model simulating competition and status dynamics in peer review, Open Identities led to reductions in quality of processes under conditions of high levels of status awareness and competition among reviewers. Such results are conceded by the authors, however, to be ‘only abstract and highly hypothetical’. In fact, empirical investigations of effects of Open Peer Review on review quality seem to generally show neutral to positive effects.

Two studies found neutral impact of Open Identities on quality. Van Rooyen et al. [502] conducted a randomized trial at a medical journal, where manuscripts were evaluated under either Open Identities or standard (anonymized) review. No significant differences in review quality were found. A similar randomized control trial (RCT), also at a medical journal [509] found no difference in review quality.

Three studies found a positive correlation between Open Identities and indicators of quality, however. Walsh et al. [507] performed an RCT at one psychiatry journal with reviews performed under Open Identities found to be of significantly higher quality than those conducted under blinded review. Thelwall [506] analysed open review reports from MDPI journals, where Open Identities is optional, finding that signed reviews were on average 15% longer (although with no difference in publication recommendations). Thelwall speculated that increased review length is associated with review quality. Fox’s [511] study of optional Open Identities at one ecology journal similarly found that signed reviews tended to be longer. However, interpreting these latter two findings requires caution. When optional, it is possible (perhaps likely) that reviewers are simply less likely to sign reviews they perceive themselves to be of lower quality (measured here by the questionable proxy of review length).

Finally, two studies looked at effects of Open Identities and Open Reports in combination upon review quality. Kowalczuk et al. [500] performed a comparative assessment of peer review reports employing open or closed review models at three otherwise comparable medical journals (BMC Infectious Diseases—Open Identities and Open Reports; BMC Microbiology—single-blind; Journal of Inflammation—both models). Reviews at the journal employing open peer review were found to be of 5% higher quality, although at Journal of Inflammation there was no difference in quality in reviews under different models. Finally, Bornmann & Daniel [508] studied levels of inter-reviewer agreement in the journal Atmospheric Chemistry and Physics, which uses Open Identities and Open Reports. They found similarly low levels of agreement to that reported in other studies of reviews conducted under closed review models.

3.6.2. Integrity

Open Identities’ effects upon the integrity of review processes and reviewer behaviour seem complex. Wolfram et al. [14] studied how journal policies for Open Identities (optional or mandatory) impacted reviewers’ use of ‘hedging’ (terms indicating lack of certainty) and ‘research-related terms’ when evaluating manuscripts, finding that opening identities made ‘no appreciable difference’ and concluding that ‘reviewers are not influenced by the identification requirements’.

Other studies did find differences, however. Regarding tone, courtesy and constructiveness of comments, Walsh et al.’s [507] RCT study of optional Open Identities found that where reviews were signed, they tended to be more ‘courteous’. Felizardo et al. [499] conducted a small survey (12 reviewers, six authors from Software Engineering) to understand how respondents believed their review behaviour changed when not anonymous. The majority reported being more likely to write ‘bland and cautious’ reviews that avoid issues of novelty/general interest and focus on more ‘objective’ issues like technical concerns. Matsui et al. [501] used sentiment analysis of review reports to assess (among other aims) how reviewer disclosure of identities (published optionally online after single-blind process) is associated with review sentiment, finding that reviews written by reviewers who choose to ‘sign’ were ‘more subjective and more positive than the anonymous reviewers’ reviews’. The authors contend this implies ‘possible social pressure from name association’, although since authors chose to sign, we might more readily posit that authors were more inclined to sign when reviewing positively. Finally, regarding effects from both revealing reviewer identities and publishing reports, Bravo et al. [498] found that male reviewers tended to be more ‘constructive’ under these conditions.

3.6.3. Efficiency

The aforementioned simulation work of Bianchi & Squazzoni [497] suggested that efficiency could be negatively affected in terms of reviewer time. Similarly, the simulation study of Radzvilas et al. [505] suggested that Open Identities could ‘increase the effort of authors in a range of circumstances’, but this effect is influenced by multiple factors including reviewers’ own levels of effort. Supporting these model-driven findings, Walsh et al. [507] and Van Rooyen et al. [502] found that under Open Identities, signed reviews took longer to complete than those not signed. Finally, Bravo et al. [498] found that open reports and optional open identities had no significant change on referees’ willingness to review or time taken to review.

3.6.4. Citations

Regarding effects on citations, Zong et al. [503] conducted a scientometric study using data from the PeerJ journal, where revealing identities and publishing reports are optional, to find that ‘articles with open peer review history could be expected to have significantly greater citation counts than articles with closed peer review history’. Wei et al. [510] also studied effects of open peer review (OPR) on article citations at six PLoS journals, comparing reports under open to those under closed review. They found that articles reviewed under open conditions were associated with higher average page views, saving and sharing, but found mixed effects upon citations, with between 6.2% and 7.5% more citations in Google Scholar, Dimensions and Semantic Scholar, but no significant association between OPR and citations in the more selective Scopus and Web of Science databases (hence the boost in citations may be attributable to higher citations of openly reviewed articles in ‘grey literature’ found in Google Scholar, Dimensions and Semantic Scholar but not Scopus or WoS).

Regarding reviewer requests for self-citations, Levis et al. [504] compared prevalence of citation requests for reviewers’ own work in open and closed conditions, controlling for overall number of citation requests. They found that ‘in reasonably similar journals that use single-blind and open review, there were no substantive differences in the pattern of peer reviewer self-citations’.

3.6.5. Equity, diversity and inclusion

Finally, two studies reported findings applicable to equity, diversity and inclusion. Firstly, Bianchi & Squazzoni’s [497] simulated model suggested that perceived status of open reviewers impacts reciprocity strategies. If so, they advise that ‘transparency can be especially detrimental in cases of young scholars who could be sensitive to the risk of retaliation when asked to review work by more advanced senior scholars’. Fox [511] found that male reviewers were 1.8 times more likely to sign reviews than female reviewers, which may have negative consequences for reviewer recruitment and the ‘diversity of reviewers recruited by journals’.

3.7. Open Science in general

In this section, we cover studies that demonstrated academic impacts but did not fit pre-specified categories, addressing multiple types of Open Science (OS) or a broader perspective on OS. Our review identified 21 studies from general Open Science [28,347,512530], showing impacts (or lack thereof) on diverse areas such as reuse (eight), efficiency and productivity (six), quality, (six), citations (five), collaboration (five), equity, diversity and inclusion (four), ethics and integrity (three), reproducibility (three), trust (three) and others (two).

McKiernan et al. [522] reviewed the prior literature showing ‘open research is associated with increases in citations, media attention, potential collaborators, job opportunities, and funding opportunities’.

An emerging approach to signify the application of OS practices is to award ‘badges’ to journals or articles. These badges display different types of adherence to OS practices, such as Open Data and Open Code. Schneider et al. [526] observed a positive effect of badges on trust in results among scientists and students, but not the general public, possibly due to the public already assuming openness. However, Schneider et al. [526] also noted that this effect might not be generalizable without additional research. Zong et al. [529] reported that while Open Science badges increased social media attention, they did not affect citation counts.

Research on OS impacts often appears as case studies. Anagnostou et al. [512] highlighted a platform in Ghana based on open repositories and digital research tools for ageing research, enabling local insights for research and policy. Investigating the case of ‘blue economy’ (sustainable use of ocean resources for economic growth), Coro [517] described how OS practices improved computational process speed, collaboration, interdisciplinarity, virtual laboratories, model reuse, data longevity and findings dissemination.

Other case studies address efficiency and equity in research. A review covering four case studies from Argentina [513] suggests that OS practices support ‘efficiency, democratization and social responsiveness’. The authors also note ‘several directions of openness could lead to different benefits’, implying that partial openness can also be beneficial.

On the other hand, Besançon et al. [515] highlight how the urgency of the COVID-19 pandemic led to suboptimal adoption of OS practices (especially Open/FAIR Data, Open Code and Open Peer Review) and (in some cases) misuse of preprinting slowed the response of the academic community. On the same topic, Tse et al. [530] argue that Open Data, by sharing protein structures, and Open Source practices, such as sharing information on potential inhibitors to COVID-19, have greatly contributed to accelerating research into COVID-19.

OS may also create negative effects. Hofmann [519] and Ross-Hellauer et al. [28] warned of risks like reduced quality, inclusiveness and reinforced gender divides, along with reliance on specific skills and infrastructure. These effects are difficult to quantify, but awareness is crucial for understanding OS impact pathways.

Lastly, LeBel et al. [521] scrutinized concerns about OS practice costs, such as increased false negatives and opportunity costs associated with larger samples and direct replications. They argue these concerns are unwarranted, as OS aims to reduce Type I error rates and increase statistical power. Despite short-term challenges, the benefits of open and high-powered research outweigh the costs in the long run.

In conclusion, while Open Science practices show benefits across various fields, they also present potential challenges and risks that need careful management. Further studies and refined models of impact pathways are necessary to fully understand and optimize these practices in diverse research contexts.

4. Discussion

Our scoping review shows that there is a substantial body of literature (485 studies) demonstrating academic impacts of OS practices (RQ1). These studies covered all OS practices (Open Access—OA, Citizen Science—CS, Open/FAIR Data—OFD the most addressed practices), and effects on various aspects of impact (most common were effects on citations, quality, and efficiency and productivity) (see figure 2).

Regarding SRQ1 (types of impact, including whether positive/negative), table 5 below presents an overview of our findings. We found a range of positive impacts, most notably effects of OA upon levels of citations, although importantly, effects vary by OA type and differ across disciplines. We further found moderate to neutral effects of open peer review upon review quality, efficiency gains due to cost savings in Citizen Science (ecological monitoring), OA (publishing costs, speed of dissemination due to preprints) and general gains in effectiveness of research in response to the COVID-19 pandemic. We also found positive impacts on reuse of Open/FAIR data, reproducibility (due to pre-registration), as well as some limited evidence of positive effects on novelty, collaboration and trust. No or mixed impacts included effects upon quality of articles due to OA and data from CS, efficiency of editorial processes due to OFD and Open Peer Review, reproducibility (OFD not associated with levels of statistical consistency, but still associated with high reproduction costs; Open Methods of pre-registration or protocol-sharing not effective in increasing sample-sizes or reducing p-values just below 0.05, but increasing levels of reporting of non-significant p-values), and integrity in open peer review processes. Negative impacts included those on equity/diversity/inclusion (APC barrier and potential for reuse of OFD barriers for less well-resourced researchers, as well as negative effects on inclusivity and gender equity), efficiency in OA (time lost due to engagement with predatory publishing), as well as ethics due to potential for re-identifying participants and data colonialism (lack of benefit-sharing) in OFD. These effects are most commonly directly contributable to OS practices.

Table 5.

Summary of key findings. Findings based on individual studies are highlighted with an asterisk (*). Abbreviations: Open Access (OA), Open/FAIR Data (OFD), Open Methods (OM), Open Code (OC), Citizen Science (CS), Open Evaluation (OE), Open Science in general (OSG).

positive impact

no or mixed impact

negative impact

citations

OA: small to moderate increase for OA papers, although with differences across OA types (findings of a citation advantage less common for gold OA compared with hybrid and/or green OA); significant increase for publications which were earlier available as preprints.

OFD: small to moderate increase.

OE: neutral to moderate effect of OPR.

OC: unclear effects.

OSG: positive impact of OS badges on social media attention (tweets) but not citations.

quality

OE: open peer review has moderately positive to neutral effects on review quality.

OA: conflicting evidence as to the effect of OA on article quality. Preprints found to be of same or similar quality as published versions.

CS: given adequate training, data quality often comparable with other sources, although some studies report negative effects.

efficiency and productivity

CS: cost savings in ecological monitoring; expanding spatial and temporal scales of data collection.

OA: lower publishing costs; significant increase in speed of dissemination due to preprints.

OSG: OS contributed to speed and scope of research activities during COVID-19 pandemic response.

OFD: no substantial effect of current data sharing policies on editorial effort.*

OE: unclear impact on overall efficiency, but tendency towards slightly longer review times under open identities.

OA: mixed effect on number of published articles per journal.

OA: wasted time (esp. among junior researchers) engaging with predatory publishers.*

equity, diversity, inclusion

OA: research published via OA involves more (international) collaboration*; OA publications receive more diverse citations, suggesting more universal use of research.*

OA: APC model has effects of stratification upon patterns of authorship, tending to marginalize those with fewer resources.

OE: gender and status dynamics regarding open identities (male, more senior reviewers more likely to adopt).*

OFD: reuse of data more challenging for those with fewer resources.

CS: CS activities tend to be focused in the Global North.

reuse

OFD: positive effect on reuse, but largely constrained to well-resourced researchers.

reproducibility

OM: structured preregistrations decrease deviations from protocol*; positive effect of registered reports on Open Science practices related to sound statistical practice.*

OFD: data sharing policies have no clear effect on statistical inconsistencies*; high costs for reproduction even though data shared.

OM: no impact of requiring sample size justification on sample size*; no impact of reviews of protocols for replication studies on replicability of findings*, no effect of Open Science practices on citations or p-values just below p < 0.05, but increase in reporting of non-significant p-values.*

novelty

OSG: positive effect of OS practices on rate and frequency of true discoveries, despite short-term costs.*

ethics and integrity

OE: unclear impact of open reports and open identities on integrity of review process: some find no impact, some do.

OE: no effect of OPR on reviewer requests for self-citation.*

OFD: making data public might involve risk of re-identifying participants.*

OFD: whose data is being shared might not enjoy the benefits.

trust

OSG: positive impact of OS badges on trust in results by scientists.*

In terms of specific mechanisms responsible for the impacts of OS (SRQ2), we found similarities, but also important differences, among OS practices. Impacts of OA and OFD on citations and reuse work in similar ways: open publications/data → higher readership/reuse → increased citations. This mechanism is only part of the picture, as there are multiple moderating factors (see below). For CS, the main mechanism of impact is reduced cost of data collection (efficiency), which leads to the ability to conduct data collection across increased temporal and spatial scales (effectiveness). For Open Methods (OM), we did not find many impacts, but where we did, they are driven by increased guidance and structure in terms of research methodology. In the case of Open Evaluation (OE), increased accountability due to open reports or identities leads to neutral to moderate effects on quality of reviews, but may lower efficiency with some studies finding that reviews take longer to complete (although others find no effect).

As stated, the implied mechanisms involve certain enabling and inhibiting factors that moderate the impact of OS (SRQ3). An important inhibiting factor is the requirement for potential beneficiaries of open resources to have sufficient capacity in terms of skills, resources and infrastructure. Examples are of the OA APC-barrier, which presents a financial obstacle that influences who is able to publish in which journals (disadvantaging researchers from less resourced institutions and the Global South more broadly) and reuse of OFD which often requires sufficient skills, equipment and infrastructure, lack of which at least partly explains the low reuse of data by researchers in the Global South. While skills and resources can act as barriers to the impact of OS, the reverse is also true: our review found sufficient training of volunteers in CS projects to be a key factor in ensuring sufficient data quality. Similarly, the existence of funding and institutional support enables researchers to overcome potential efficiency costs of OM more broadly (including sharing of data and code, preregistration and similar practices). Lastly, epistemic diversity (different approaches to generating knowledge) also acts as a factor moderating the impact of OS. For example, quantitative approaches might benefit more directly from shared data than qualitative methods, while missing context of anonymized qualitative data inhibits further reuse.

Although our review identified a variety of academic impacts stemming from a range of OS practices, substantial knowledge gaps remain (SRQ4). We found, for instance, relatively few studies addressing efficiency gains or losses due to OS workflows (e.g. data or code sharing), the efficacy of DMPs to promote higher quality data sharing, or to which extent Open Methods directly improve quality and reproducibility of research.

A salient issue that posed a challenge in conducting our review is that many studies we encountered report on the uptake of OS, rather than OS’s impact. This is not problematic per se, but becomes an issue when policy is built on assumptions about impacts deriving from uptake, where the impact itself has not been rigorously studied. In addition, from the overview gained via this process, we note that many studies which were initially thought potentially relevant were eventually excluded for their lack of empirical observations, presenting arguments rather than evidence.

Much of the research summarized in our study comes via case studies or is based on small samples (e.g. individual data repositories or Citizen Science projects). Findings from case studies can be difficult to generalize. Although such case reports provide evidence of the potential impact of OS, it may also bias the evidence. projects that failed to set up a successful data repository, or failed to finish a Citizen Science project, may be less likely to report on it. In addition, we note that many of the authors of such reports are involved in the projects themselves, potentially bringing other forms of bias or conflicts-of-interest. For these reasons, more independent research on OS is necessary.

One set of challenges in studying the effects of OS on academic impact is methodological. At a broader level of impact assessment, studies often rely on post hoc investigations of existing datasets, where establishing a credible counterfactual is often difficult. For example, while it is quite straightforward to compare citations of OA and closed access, it is much harder to compare reuse of Open Data versus closed data, since closed data that has been re-used is hard to identify. Moreover, the central problem in this challenge is distinguishing impacts of Open Science from the effect of Open Science on impact. That is, Open Science may have some impact, but it might not immediately be clear whether this impact depends on the fact that it was ‘open’. For example, consider software that is developed by researchers. Suppose it is released as Open Source. The software might be used by thousands of researchers, and hence, it could be considered to have academic impact. However, it is possible that the exact same software might have also been used by thousands of researchers had it been licensed by a commercial company. If that were the case, whether the software was open or closed would have had no effect on the overall use by researchers, and there would be no effect of Open Science on academic impact.

At the same time, many studies are not able to disentangle effects of OS from other potentially confounding factors (such as selection bias or journal prestige in the case of the OA citation advantage), thus challenging claims of causal effects. Many studies lacked consideration of causality, simply reporting overall figures. Some studies do compare ‘Open’ outputs with ‘Closed’ outputs, but are not explicit about causality, even if some of the conclusions are being drawn and stated in causal terms. This is a particular issue in much of the citation impact literature considered here, where despite the large number of studies investigating the OACA, we note the lack of a consolidated understanding of causal factors affecting citations. In addition, the literature on OA and OFD employs a large variety of bibliometric methods, some of which are inadequate to answer the posed questions. We therefore echo the call for reporting guidelines for bibliometric studies by Langham-Putrow et al. [150].

A small minority of publications explicitly considers causality, and only a few of them (including some randomized controlled trials) are able to make a strong case for causality. Drawing causal conclusions is indeed something that is very challenging, and sometimes simply impossible. Although we call for greater attention on causality in studies on Open Science, it is even more important to consider the limitations when interpreting studies without a clear causal understanding, and draw conclusions and provide policy advice that is commensurate with the (lack of causal) evidence [346].

Beyond methodological issues around evidence of academic impact of OS, we must be careful in appraising the available evidence, keeping in mind that what has been researched most might not be the most relevant aspects of OS impact. For instance, much of the literature focuses on citations, and the citation advantage, be it of Open Access, Open Data or Open Code. This is a clear instance of a streetlight effect: it is easy to measure and study, even if drawing causal conclusions remains challenging. This is not to say that citations are irrelevant; they do provide some insight as to the increased (and potentially broader, cf. [182]) use of OS. But there are other academic impacts, such as quality, reproducibility, efficiency and equity, that merit much more attention, even if they are more difficult to study. As stated by UNESCO, OS should be about ‘making science more accessible, inclusive and equitable for the benefit of all’ [531]. The research community should ensure that we study the impacts that are relevant in achieving that.

5. Conclusion

Our scoping review demonstrates a range of academic impacts of Open Science, both positive and negative. The majority of reviewed studies investigated effects of Open Access, Citizen Science and Open/FAIR Data, with fewer studies reporting effects of Open Evaluation, Open Code and Open Methods. Key areas of impact studied are citations, quality, efficiency, equity, reuse, ethics and reproducibility, with most studies reporting positive or at least mixed impacts. However, we also identified significant unintended negative impacts, especially those regarding equity, diversity and inclusion. Overall, the main barrier to academic impact of OS is lack of skills, resources and infrastructure to effectively re-use and build on existing research.

This scoping review has several limitations. The inclusion only of English-language studies and use of a limited number of cross-disciplinary databases for initial search (Web of Science and Scopus) may have led to the exclusion of relevant studies. Web of Science and Scopus are known to have issues of representation regarding disciplines (i.e. Social Sciences and Humanities) and regions (i.e. Global South) [532535]. These issues are probably somewhat mitigated by our use of OpenAlex, which has much broader coverage [536,537], for the second search phase, however. Relatedly, pilot searching indicated the large numbers of results returned by key terms. Hence, our search strings were somewhat restricted to balance results retrieved and pragmatic concerns for resource constraints. This means some key terms for OS practices, like preprints, preregistration or open hardware, were not included. This may have excluded relevant studies that used narrower or less common terms in their titles or abstracts. Some omissions (e.g. preprints) were specifically targeted during the snowballing phase to somewhat make up for any such limitations. These limitations could help explain our inclusion of only nine references each for Open Methods and Open Code, which may have been mitigated by inclusion of keywords like protocols, research software, open source software and academic software (our thanks to one reviewer for pointing this out). Further, because of our limitations on terms to include in search strings, and our updating of categorizations of academic impacts as we analysed the literature, our search terms excluded specific mention of some aspects reported here (ethics, diversity, inclusion, reproducibility, novelty, reproducibility) meaning some relevant studies related to those aspects may have been missed in our initial search. Finally, our screening approach involved single screening at certain stages. Despite regular check-ups and feedback within the team to strengthen inter-coder reliability, this nonetheless may impact coding reliability.

While it is not our goal to provide specific recommendations for stakeholders, we can nonetheless draw some conclusions regarding implications for future research and policy. First, despite the wealth of studies reviewed, important knowledge gaps around the efficacy of Open Methods and DMPs to improve reproducibility, and the effect of OM on efficiency remain and should be addressed. Second, independent investigations across initiatives, instead of individual case studies of successful initiatives, would be better equipped to build a systematic evidence base of the academic impacts of Open Science. In methodological terms, monitoring OS should expand from a focus on uptake towards impact. In extension, monitoring of impact must go beyond what is easy to measure or where data sources are readily available, which could be achieved with increased funding of research infrastructure for meta-research and monitoring efforts. Finally, more systematic agreement on standards/guidelines for reporting of bibliometric studies, as well as a greater focus on causality and limitations of correlational approaches would aid in establishing rigorous evidence for the academic impact of OS.

Acknowledgements

We thank Nicki Lisa Cole, Ron Dekker, Silvia Vignetti, Jessica Catalano and two anonymous reviewers for feedback on earlier drafts of this manuscript.

Footnotes

1

Complementary reviews by our collaborators separately tackle societal [38] and economic (forthcoming) impacts.

3

We hence do not include literature which reports only on the uptake of or attitudes towards Open Science and its various aspects.

4

We excluded Open Government Data from this study because we focused on OS practices within academia.

5

We excluded literature on crowdsourcing. Although some papers refer to a crowdsourcing approach as Citizen Science [43], approaches to crowdsourcing might also involve paying (online) workers, such as using Amazon Mechanical Turk. For this reason, we considered crowdsourcing out of scope for our review.

6

In addition, ‘diamond’ OA more recently refers to OA journals with no author facing charges, which can be understood as a subset of gold OA. Only one study identified in our review analysed diamond OA journals.

7

Closed-access journals that offer the option to publish manuscripts OA under an open licence by paying an article processing charge (APC) are commonly referred to as ‘hybrid’.

8

Note, however, that many OA journals listed in the Directory of Open Access Journals do not charge an APC at all (see footnote above on the terminology of ‘diamond OA’). Thus, alternative funding models for OA publishing which do not necessarily incentivize low acceptance thresholds exist [190].

Contributor Information

Thomas Klebel, Email: tklebel@know-center.at.

Vincent Traag, Email: V.A.Traag@cwts.leidenuniv.nl.

Ioanna Grypari, Email: igrypari@athenarc.gr.

Lennart Stoy, Email: lennart.stoy@technopolis-group.com.

Tony Ross-Hellauer, Email: tross@know-center.at.

Ethics

This work did not require ethical approval from a human subject or animal welfare committee.

Data accessibility

Datasets and additional materials supporting this article are published on Zenodo [42].

Supplementary material is available online [538].

Declaration of AI use

ChatGPT was used at revision stage to improve conciseness and flow of the section ‘Open Science in general’.

Authors’ contributions

T.K.: conceptualization, data curation, formal analysis, investigation, methodology, project administration, software, supervision, validation, visualization, writing—original draft, writing—review and editing; V.T.: funding acquisition, investigation, methodology, validation, writing—original draft, writing—review and editing; I.G.: funding acquisition, investigation, writing—original draft, writing—review and editing; L.S.: investigation, writing—original draft, writing—review and editing; T.R.-H.: conceptualization, funding acquisition, investigation, methodology, project administration, supervision, validation, writing—original draft, writing—review and editing.

All authors gave final approval for publication and agreed to be held accountable for the work performed therein.

Conflict of interest declaration

We declare we have no competing interests.

Funding

This work was supported by the project PathOS, funded by the European Commission under the Horizon Europe programme (grant no. 101058728). The Know-Center is funded within COMET—Competence Centers for Excellent Technologies—under the auspices of the Austrian Federal Ministry of Transport, Innovation and Technology, the Austrian Federal Ministry for Digital and of Economic Affairs, the Austrian Research Promotion Agency (FFG), and by the province of Styria. The COMET programme is managed by the FFG.

References

  • 1. Vicente-Saez R, Gustafsson R, Martinez-Fuentes C. 2021. Opening up science for a sustainable world: An expansive normative structure of open science in the digital era. Sci. Publ. Policy 48, 799–813. ( 10.1093/scipol/scab049) [DOI] [Google Scholar]
  • 2. Suber P. 2012. Open access. Cambridge, MA: MIT Press. See https://library.oapen.org/handle/20.500.12657/26065. [Google Scholar]
  • 3. Willinsky J. 2006. The access principle: the case for open access to research and scholarship. Cambridge, MA: MIT Press. ( 10.1096/fasebj.20.4.A439) [DOI] [Google Scholar]
  • 4. Chan L, et al. 2002. Budapest Open Access Initiative. See https://web.archive.org/web/20171213093708/http://www.budapestopenaccessinitiative.org/read (accessed 29 March 2022).
  • 5. Easterbrook SM. 2014. Open code for open science? Nat. Geosci. 7, 779–781. ( 10.1038/ngeo2283) [DOI] [Google Scholar]
  • 6. Mendez D, Graziotin D, Wagner S, Seibold H. 2020. Open science in software engineering. In Contemporary empirical methods in software engineering (eds Felderer M, Travassos GH), pp. 477–501. Cham, Switzerland: Springer International Publishing. ( 10.1007/978-3-030-32489-6_17) [DOI] [Google Scholar]
  • 7. Murray-Rust P. 2008. Open Data in Science. Nat. Prec 1. ( 10.1038/npre.2008.1526.1) [DOI] [Google Scholar]
  • 8. Wilkinson MD, et al. 2016. The FAIR Guiding Principles for scientific data management and stewardship. Sci. Data 3, 160018. ( 10.1038/sdata.2016.18) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9. Goble C, et al. 2021. Implementing FAIR digital objects in the EOSC-Life workflow collaboratory. Zenodo. ( 10.5281/zenodo.4605654) [DOI]
  • 10. Teytelman L, Stoliartchouk A. 2015. Protocols.io: reducing the knowledge that perishes because we do not publish it. Inf. Serv. Use 35, 109–115. ( 10.3233/ISU-150769) [DOI] [Google Scholar]
  • 11. National Research Council . 2003. Sharing publication-related data and materials: responsibilities of authorship in the life sciences. Washington, DC: National Academies Press. ( 10.17226/10613) [DOI] [PubMed] [Google Scholar]
  • 12. Nosek BA, Ebersole CR, DeHaven AC, Mellor DT. 2018. The preregistration revolution. Proc. Natl Acad. Sci. USA 115, 2600–2606. ( 10.1073/pnas.1708274114) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13. Ross-Hellauer T. 2017. What is open peer review? A systematic review. F1000Res. 6, 588. ( 10.12688/f1000research.11369.2) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14. Wolfram D, Wang P, Hembree A, Park H. 2020. Open peer review: promoting transparency in open science. Scientometrics 125, 1033–1051. ( 10.1007/s11192-020-03488-4) [DOI] [Google Scholar]
  • 15. Strasser BJ, Baudry J, Mahr D, Sanchez G, Tancoigne E. 2019. ‘Citizen science’? Rethinking science and public participation. Sci. Technol. Stud. 32, 52–76. ( 10.23987/sts.60425) [DOI] [Google Scholar]
  • 16. Peters MA. 2020. Citizen science and ecological democracy in the global science regime: the need for openness and participation. Educ. Philos. Theory 52, 221–226. ( 10.1080/00131857.2019.1584148) [DOI] [Google Scholar]
  • 17. Hecker S, Haklay M, Bowser A, Makuch Z, Vogel J, Bonn A. 2018. Citizen science: innovation in open science, society and policy. London, UK: University College London. See https://muse.jhu.edu/pub/354/oa_edited_volume/book/81913. [Google Scholar]
  • 18. Field S. 2022. Charting the constellation of science reform. Groningen, The Netherlands: University of Groningen. ( 10.33612/diss.229114775) [DOI] [Google Scholar]
  • 19. 2003. Berlin Declaration on Open Access to Knowledge in the Sciences and Humanities. See https://openaccess.mpg.de/Berlin-Declaration (accessed 8 July 2024).
  • 20. Brown PO, et al. 2003. Bethesda Statement on Open Access Publishing. See https://dash.harvard.edu/handle/1/4725199 (accessed 8 July 2024).
  • 21. Ioannidis JPA. 2005. Why most published research findings are false. PLoS Med. 2, e124. ( 10.1371/journal.pmed.0020124) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22. Baker M. 2016. 1,500 scientists lift the lid on reproducibility. Nature 533, 452–454. ( 10.1038/533452a) [DOI] [PubMed] [Google Scholar]
  • 23. Barker M, et al. 2022. Introducing the FAIR principles for research software. Sci. Data 9, 622. ( 10.1038/s41597-022-01710-x) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24. UNESCO . 2021. UNESCO Recommendation on Open Science. ( 10.54677/MNMH8546) [DOI]
  • 25. Center for Open Science . What is open science? In press. See https://www.cos.io/open-science (accessed 8 July 2024).
  • 26. Fecher B, Friesike S. 2014. Open science: one term, five schools of thought. In Opening science (eds Bartling S, Friesike S), pp. 17–47. Cham, Switzerland: Springer International Publishing. ( 10.1007/978-3-319-00026-8_2) [DOI] [Google Scholar]
  • 27. Bijker WE, Hughes TP, Pinch T (eds). 2005. The social construction of technological systems: new directions in the sociology and history of technology. In Papers of a workshop held at the University of Twente, The Netherlands, in July 1984. Cambridge, MA: MIT Press. [Google Scholar]
  • 28. Ross-Hellauer T, Reichmann S, Cole NL, Fessl A, Klebel T, Pontika N. 2022. Dynamics of cumulative advantage and threats to equity in open science: a scoping review. R. Soc. Open Sci. 9, 211032. ( 10.1098/rsos.211032) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29. Leonelli S. 2023. Philosophy of open science. Cambridge, UK: Cambridge University Press. ( 10.1017/9781009416368) [DOI] [Google Scholar]
  • 30. Rafols I, Meijer I, Molas-Gallart J. 2024. Monitoring Open Science as transformative change: towards a systemic framework. F1000Res. 13, 320. ( 10.12688/f1000research.148290.1) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31. European Commission . 2019. About the Open Science Monitor. See https://research-and-innovation.ec.europa.eu/strategy/strategy-2020-2024/our-digital-future/open-science/open-science-monitor/about-open-science-monitor_en (accessed 8 July 2024).
  • 32. Bracco L. 2022. Promoting Open Science through bibliometrics: a practical guide to build an open access monitor. Liber Q. 32, 1–18. ( 10.53377/lq.11545) [DOI] [Google Scholar]
  • 33. Barbers I, Stanzel F, Mittermaier B. 2022. Open access monitor Germany: best practice in providing metrics for analysis and decision-making. Serials Rev. 48, 49–62. ( 10.1080/00987913.2022.2066968) [DOI] [Google Scholar]
  • 34. Bracco L, L’Hôte A, Jeangirard E, Torny D. 2022. Extending the open monitoring of open science. See https://hal.science/hal-03651518. [Google Scholar]
  • 35. Karasz I, Stoy L, Seminaroti E, Grapengiesser IM. 2024. PathOS - D1.3 key impact pathways for the Open Science Framework. Zenodo. ( 10.5281/ZENODO.11108567) [DOI]
  • 36. White H. 2009. Theory-based impact evaluation: principles and practice. J. Dev. Effect. 1, 271–284. ( 10.1080/19439340903114628) [DOI] [Google Scholar]
  • 37. Ravenscroft J, Liakata M, Clare A, Duma D. 2017. Measuring scientific impact beyond academia: an assessment of existing impact metrics and proposed improvements. PLoS One 12, e0173152. ( 10.1371/journal.pone.0173152) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38. Cole NL, Kormann E, Klebel T, Apartis S, Ross-Hellauer T. 2024. The societal impact of Open Science: a scoping review. R. Soc. Open Sci. 11, 240286. ( 10.1098/rsos.240286) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39. Tricco AC, et al. 2018. PRISMA Extension for scoping reviews (PRISMA-ScR): checklist and explanation. Ann. Intern. Med. 169, 467–473. ( 10.7326/M18-0850) [DOI] [PubMed] [Google Scholar]
  • 40. Ross-Hellauer T, Klebel T, Pitelis A. 2022. Protocol for scoping reviews of (1) academic, (2) societal and (3) economic impacts of open science. OSF. ( 10.17605/OSF.IO/GM57C) [DOI] [Google Scholar]
  • 41. Cole NL, Klebel T, Ross-Hellauer T. 2023. In Addendum to protocol for scoping reviews of (1) academic, (2) societal and (3) economic impacts of open science: protocol for grey literature search and indexing of sources, and for screening and data charting in the syrf platform. OSF. See https://osf.io/3b6xj (accessed 22 January 2024). [Google Scholar]
  • 42. Klebel T, Traag V, Grypari I, Stoy L, Ross-Hellauer T. 2024. Data from ‘The academic impact of Open Science: a scoping review' [dataset]. Zenodo. ( 10.5281/zenodo.14674544) [DOI] [PMC free article] [PubMed]
  • 43. Noel-Storr A. 2019. Working with a new kind of team: harnessing the wisdom of the crowd in trial identification. EFSA J. 17, e170715. ( 10.2903/j.efsa.2019.e170715) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44. Ay IE, Tazegul G, Duranoğlu Y. 2022. A comparison of scientometric data and publication policies of ophthalmology journals. Indian J. Ophthalmol. 70, 1801–1807. ( 10.4103/ijo.IJO_2720_21) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45. Bjork BC. 2012. A comparison of subscription and open access journals in construction management and related fields. CEB 12, 42. ( 10.5130/AJCEB.v12i2.2560) [DOI] [Google Scholar]
  • 46. Dove C, Chan TM, Thoma B, Roland D, Bruijns SR. 2019. A cross-sectional description of open access publication costs, policies and impact in emergency medicine and critical care journals. Afr. J. Emerg. Med. 9, 150–155. ( 10.1016/j.afjem.2019.01.015) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47. Wilkinson TA, Russell CJ, Bennett WE, Cheng ER, Carroll AE. 2019. A cross-sectional study of predatory publishing emails received by career development grant awardees. BMJ Open 9, e027928. ( 10.1136/bmjopen-2018-027928) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48. Solomon DJ, Laakso M, Björk BC. 2013. A longitudinal comparison of citation rates and growth among open access journals. J. Informetr. 7, 642–650. ( 10.1016/j.joi.2013.03.008) [DOI] [Google Scholar]
  • 49. Vaughan L, Shaw D. 2008. A new look at evidence of scholarly citation in citation indexes and from web sources. Scientometrics 74, 317–330. ( 10.1007/s11192-008-0220-2) [DOI] [Google Scholar]
  • 50. Kirkham JJ, Penfold N, Murphy F, Boutron I, Ioannidis JP, Polka JK, Moher D. 2020. A systematic examination of preprint platforms for use in the medical and biomedical sciences setting. Sci. Commun. Educ 10, e041849. ( 10.1101/2020.04.27.063578) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 51. Bjórk BC. 2019. Acceptance rates of scholarly peer-reviewed journals: A literature survey. EPI 28, 1–9. ( 10.3145/epi.2019.jul.07) [DOI] [Google Scholar]
  • 52. Gaulé P. 2009. Access to scientific literature in India. J. Am. Soc. Inf. Sci. 60, 2548–2553. ( 10.1002/asi.21195) [DOI] [Google Scholar]
  • 53. Peidu C. 2020. An Empirical Examination of Citation in Life Science. JSCIRES 9, 70–76. ( 10.5530/jscires.9.1.8) [DOI] [Google Scholar]
  • 54. Tennant JP, Lomax DR. 2019. An overview of Open Access publishing in palaeontology. Palaeontol. Electron 22, 1–9. ( 10.26879/968) [DOI] [Google Scholar]
  • 55. Xu L, Liu J, Fang Q. 2011. Analysis on open access citation advantage: an empirical study based on Oxford open journals. In Proc. of the 2011 iConf., pp. 426–432. Seattle, WA: ACM. ( 10.1145/1940761.1940819) [DOI] [Google Scholar]
  • 56. Boczar J, Schmidt L. 2022. Analyzing faculty open access publishing: a citation analysis of select colleges at the University of South Florida. J. Acad. Librariansh. 48, 102493. ( 10.1016/j.acalib.2022.102493) [DOI] [Google Scholar]
  • 57. Borrego Á. 2021. Are mega-journals a publication outlet for lower quality research? A bibliometric analysis of Spanish authors in PLoS ONE. OIR 45, 261–269. ( 10.1108/OIR-04-2018-0136) [DOI] [Google Scholar]
  • 58. Meerpohl JJ, Wolff RF, Antes G, von Elm E. 2011. Are pediatric Open Access journals promoting good publication practice? An analysis of author instructions. BMC Pediatr. 11, 27. ( 10.1186/1471-2431-11-27) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 59. Maddi A, Sapinho D. 2021. Article processing charges based publications: to which extent the price explains scientific impact? In 18th Int. Conf. on Scientometrics & Informetrics. Leuven, Belgium. [Google Scholar]
  • 60. Björk BC, Solomon D. 2015. Article processing charges in OA journals: relationship between price and quality. Scientometrics 103, 373–385. ( 10.1007/s11192-015-1556-z) [DOI] [Google Scholar]
  • 61. Maddi A, Sapinho D. 2022. Article processing charges, altmetrics and citation impact: Is there an economic rationale? Scientometrics 127, 7351–7368. ( 10.1007/s11192-022-04284-y) [DOI] [Google Scholar]
  • 62. Schönfelder N. 2020. Article processing charges: mirroring the citation impact or legacy of the subscription-based model? Quant. Sci. Stud. 1, 6–27. ( 10.1162/qss_a_00015) [DOI] [Google Scholar]
  • 63. Larivière V, Sugimoto CR, Macaluso B, Milojević S, Cronin B, Thelwall M. 2014. arXiv E‐prints and the journal of record: an analysis of roles and relationships. J. Assoc. Inf. Sci. Technol. 65, 1157–1169. ( 10.1002/asi.23044) [DOI] [Google Scholar]
  • 64. AlRyalat SA, Alessa Z, Mansour M, Hamidi M, Obeidat M, Odtallah O. 2021. Assessing number and quality of urology open access journals: 2011 to 2018. Curr. Urol. 15, 59–62. ( 10.1097/CU9.0000000000000007) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 65. Smith AC, Merz L, Borden JB, Gulick CK, Kshirsagar AR, Bruna EM. 2021. Assessing the effect of article processing charges on the geographic diversity of authors using Elsevier’s ‘Mirror Journal’ system. Quant. Sci. Stud. 2, 1123–1143. ( 10.1162/qss_a_00157) [DOI] [Google Scholar]
  • 66. Togia A, Korobili S. 2014. Attitudes towards open access: a meta-synthesis of the empirical literature. Inf. Serv. Use 34, 221–231. ( 10.3233/ISU-140742) [DOI] [Google Scholar]
  • 67. Faber Frandsen T. 2009. Attracted to open access journals: a bibliometric author analysis in the field of biology. J. Doc. 65, 58–82. ( 10.1108/00220410910926121) [DOI] [Google Scholar]
  • 68. Asai S. 2021. Author choice of journal type based on income level of country. J. Sch. Publ. 53, 24–34. ( 10.3138/jsp.53.1.03) [DOI] [Google Scholar]
  • 69. Davis PM. 2009. Author‐choice open‐access publishing in the biological and medical literature: a citation analysis. J. Am. Soc. Inf. Sci. Technol. 60, 3–8. ( 10.1002/asi.20965) [DOI] [Google Scholar]
  • 70. Siler K, Haustein S, Smith E, Larivière V, Alperin JP. 2018. Authorial and institutional stratification in open access publishing: the case of global health research. PeerJ 6, e4269. ( 10.7717/peerj.4269) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 71. Frisch NK, Nathan R, Ahmed YK, Shidham VB. 2014. Authors attain comparable or slightly higher rates of citation publishing in an open access journal (CytoJournal) compared to traditional cytopathology journals - a five year (2007-2011) experience. Cytojournal 11, 10. ( 10.4103/1742-6413.131739) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 72. Gurnani B, Kaur K. 2021. Avoiding predatory publishing for early-career ophthalmologists. Indian J. Ophthalmol. 69, 3719–3725. ( 10.4103/ijo.IJO_1639_21) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 73. Komaritsa VN. 2022. Benefits of using open access: citation analysis. Autom. Doc. Math. Linguist. 56, 197–202. ( 10.3103/S0005105522040045) [DOI] [Google Scholar]
  • 74. Oikonomidi T, Boutron I, Pierre O, Cabanac G, Ravaud P, COVID-19 NMA Consortium . 2020. Changes in evidence for studies assessing interventions for COVID-19 reported in preprints: meta-research study. BMC Med. 18, 402. ( 10.1186/s12916-020-01880-8) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 75. Asai S. 2019. Changes in revenue structure of a leading open access journal publisher: the case of BMC. Scientometrics 121, 53–63. ( 10.1007/s11192-019-03200-1) [DOI] [Google Scholar]
  • 76. Wilson K, Huang CK (Karl), Montgomery L, Neylon C, Handcock RN, Ozaygen A, Roelofs A. 2022. Changing the academic gender narrative through open access. Publications 10, 22. ( 10.3390/publications10030022) [DOI] [Google Scholar]
  • 77. Fukuzawa N. 2017. Characteristics of papers published in journals: an analysis of open access journals, country of publication, and languages used. Scientometrics 112, 1007–1023. ( 10.1007/s11192-017-2414-y) [DOI] [Google Scholar]
  • 78. Peterson GM. 2013. Characteristics of retracted open access biomedical literature: a bibliographic analysis. J. Am. Soc. Inf. Sci. Technol. 64, 2428–2436. ( 10.1002/asi.22944) [DOI] [Google Scholar]
  • 79. Alkhawtani RHM, Kwee TC, Kwee RM. 2020. Citation advantage for open access articles in European Radiology. Eur. Radiol. 30, 482–486. ( 10.1007/s00330-019-06389-0) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 80. Eysenbach G. 2006. Citation advantage of open access articles. PLoS Biol. 4, e157. ( 10.1371/journal.pbio.0040157) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 81. Donovan JM, Watson CA. 2011. Citation advantage of open access legal scholarship. SSRN Journal 11, 07. ( 10.2139/ssrn.1777090) [DOI] [Google Scholar]
  • 82. Dorta-González P, Dorta-González MI. 2023. Citation differences across research funding and access modalities. J. Acad. Librariansh. 49, 102734. ( 10.1016/j.acalib.2023.102734) [DOI] [Google Scholar]
  • 83. Ezema IJ, Onyancha OB. 2017. Citation impact of health and medical journals in Africa: does open accessibility matter? EL 35, 934–952. ( 10.1108/EL-11-2016-0245) [DOI] [Google Scholar]
  • 84. Caulley L, Cheng W, Catalá-López F, Whelan J, Khoury M, Ferraro J, Husereau D, Altman DG, Moher D. 2020. Citation impact was highly variable for reporting guidelines of health research: a citation analysis. J. Clin. Epidemiol. 127, 96–104. ( 10.1016/j.jclinepi.2020.07.013) [DOI] [PubMed] [Google Scholar]
  • 85. Abbasi Z, Shekofteh M, Shahbodaghi A, Kazemi E. 2019. Citation indicators’ comparison of LIS open access and subscription publications based on Scopus. Glob. Knowl. Mem. Commun. 68, 288–299. ( 10.1108/GKMC-02-2018-0016) [DOI] [Google Scholar]
  • 86. Wani ZA, Shah TS. 2023. Citation pattern of open access and toll-based research articles in the field of biological and physical sciences: a comparative study. OIR 47, 1302–1319. ( 10.1108/OIR-01-2021-0029) [DOI] [Google Scholar]
  • 87. Crossley JR, Almasri M, Samaha N, Deklotz TR, Harley EH, Davidson BJ, Malekzadeh S, Kim HJ. 2023. Citations and author characteristics in open-access and subscription-based otolaryngology journals. Laryngoscope 133, 79–82. ( 10.1002/lary.30167) [DOI] [PubMed] [Google Scholar]
  • 88. Nelson GM, Eggett DL. 2017. Citations, mandates, and money: author motivations to publish in chemistry hybrid open access journals. J. Assoc. Inf. Sci. Technol. 68, 2501–2510. ( 10.1002/asi.23897) [DOI] [Google Scholar]
  • 89. McCabe MJ, Snyder CM. 2021. Cite unseen: theory and evidence on the effect of open access on cites to academic articles across the quality spectrum. Manage. Decis. Econ. 42, 1960–1979. ( 10.1002/mde.3353) [DOI] [Google Scholar]
  • 90. Gentil-Beccot A, Mele S, Brooks TC. 2010. Citing and reading behaviours in high-energy physics. Scientometrics 84, 345–355. ( 10.1007/s11192-009-0111-1) [DOI] [Google Scholar]
  • 91. Asai S. 2021. Collaboration between research institutes and large and small publishers for publishing open access journals. Scientometrics 126, 5245–5262. ( 10.1007/s11192-021-03949-4) [DOI] [Google Scholar]
  • 92. Klein M, Broadwell P, Farb SE, Grappone T. 2016. Comparing published scientific journal articles to their pre-print versions. In Proc. of the 16th ACM/IEEE-CS on Joint Conf. on Digital Libraries, pp. 153–162. Newark, New Jersey, USA: ACM. ( 10.1145/2910896.2910909) [DOI] [Google Scholar]
  • 93. Carneiro CFD, et al. 2020. Comparing quality of reporting between preprints and peer-reviewed articles in the biomedical literature. Res. Integr. Peer Rev. 5, 16. ( 10.1186/s41073-020-00101-3) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 94. Özay AC, Emekci Ozay O, Gün İ. 2022. Comparison of subscription access and open access obstetrics and gynecology journals in the SCImago database. Ginekol. Pol. 93, 381–388. ( 10.5603/GP.a2021.0124) [DOI] [PubMed] [Google Scholar]
  • 95. Polat B, Özmanevra R, Özmanevra PT, Kazıkdaş KÇ. 2019. Comparison of the impact factors of subscription access and open access orthopedics and sports medicine journals in the SCImago database. Eklem Hastalik. Cerrahisi 30, 163–167. ( 10.5606/ehc.2019.64729) [DOI] [PubMed] [Google Scholar]
  • 96. Jerčić Martinić-Cezar I, Marušić A. 2019. Completeness of reporting in abstracts of randomized controlled trials in subscription and open access journals: cross-sectional study. Trials 20, 669. ( 10.1186/s13063-019-3781-x) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 97. Harnad S. 2008. Confirmation bias and the open access advantage: some methodological suggestions for the davis citation study. arXiv ( 10.48550/arXiv.0808.3296) [DOI] [Google Scholar]
  • 98. Dorta-González P, Dorta-González MI. 2022. Contribution of the open access modality to the impact of hybrid journals controlling by field and time effects. J. Data Inf. Sci. 7, 57–83. ( 10.2478/jdis-2022-0007) [DOI] [Google Scholar]
  • 99. Bero L, et al. 2021. Cross-sectional study of preprints and final journal publications from COVID-19 studies: discrepancies in results reporting and spin in interpretation. BMJ Open 11, e051821. ( 10.1136/bmjopen-2021-051821) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 100. Pagliaro M. 2021. Did you ask for citations? An insight into preprint citations en route to open science. Publications 9, 26. ( 10.3390/publications9030026) [DOI] [Google Scholar]
  • 101. Holmberg K, Hedman J, Bowman TD, Didegah F, Laakso M. 2020. Do articles in open access journals have more frequent altmetric activity than articles in subscription-based journals? An investigation of the research output of Finnish universities. Scientometrics 122, 645–659. ( 10.1007/s11192-019-03301-x) [DOI] [Google Scholar]
  • 102. Bautista-Puig N, Lopez-Illescas C, de Moya-Anegon F, Guerrero-Bote V, Moed HF. 2020. Do journals flipping to gold open access show an OA citation or publication advantage? Scientometrics 124, 2551–2575. ( 10.1007/s11192-020-03546-x) [DOI] [Google Scholar]
  • 103. Craig I, Plume A, Mcveigh M, Pringle J, Amin M. 2007. Do open access articles have greater citation impact? A critical review of the literature. J. Informetr. 1, 239–248. ( 10.1016/j.joi.2007.04.001) [DOI] [Google Scholar]
  • 104. Basson I, Blanckenberg JP, Prozesky H. 2021. Do open access journal articles experience a citation advantage? Results and methodological reflections of an application of multiple measures to an analysis by WoS subject areas. Scientometrics 126, 459–484. ( 10.1007/s11192-020-03734-9) [DOI] [Google Scholar]
  • 105. Khan D, Ashar M, Yuvaraj M. 2023. Do open access journals have a greater citation impact? A study of journals in library and information science. CC 42, 13–24. ( 10.1108/CC-03-2022-0010) [DOI] [Google Scholar]
  • 106. Ingwersen P, Elleby A. 2011. Do Open Access Working Papers Attract more Citations Compared to Printed Journal Articles from the same Research Unit? In Proc. of ISSI 2011 – the 13th Int. Conf. of the International Society for Scientometrics and Informetrics, Durban, South Africa. [Google Scholar]
  • 107. Antelman K. 2004. Do open-access articles have a greater research impact? CRL 65, 372–382. ( 10.5860/crl.65.5.372) [DOI] [Google Scholar]
  • 108. Xie F, Ghozy S, Kallmes DF, Lehman JS. 2022. Do open-access dermatology articles have higher citation counts than those with subscription-based access? PLoS One 17, e0279265. ( 10.1371/journal.pone.0279265) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 109. Yang S, Qi F, Diao H, Ajiferuke I. 2024. Do retraction practices work effectively? Evidence from citations of psychological retracted articles. J. Inf. Sci. 50, 531–545. ( 10.1177/01655515221097623) [DOI] [Google Scholar]
  • 110. Clark AD, et al. 2022. Does it pay to pay? A comparison of the benefits of open-access publishing across various sub-fields in Biology. Sci. Commun. Educ. 12, e16824. ( 10.1101/2022.12.10.519925) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 111. McCabe MJ, Snyder CM. 2015. Does online availability increase citations? Theory and evidence from a panel of economics and business journals. Rev. Econ. Stat. 97, 144–165. ( 10.1162/REST_a_00437) [DOI] [Google Scholar]
  • 112. Sotudeh H. 2020. Does open access citation advantage depend on paper topics? J. Inf. Sci. 46, 696–709. ( 10.1177/0165551519865489) [DOI] [Google Scholar]
  • 113. Lansingh VC, Carter MJ. 2009. Does open access in ophthalmology affect how articles are subsequently cited in research? Ophthalmology 116, 1425–1431. ( 10.1016/j.ophtha.2008.12.052) [DOI] [PubMed] [Google Scholar]
  • 114. Davis PM. 2010. Does open access lead to increased readership and citations? A randomized controlled trial of articles published in APS journals. Physiologist 53, 197–200. ( 10.1096/fj.11-183988) [DOI] [PubMed] [Google Scholar]
  • 115. Davis PM, Fromerth MJ. 2007. Does the arXiv lead to higher citations and reduced publisher downloads for mathematics articles? Scientometrics 71, 203–215. ( 10.1007/s11192-007-1661-8) [DOI] [Google Scholar]
  • 116. Boselli B, Galindo-Rueda F. 2016. Drivers and implications of scientific open access publishing: findings from a pilot oecd international survey of scientific authors. Paris, France: OECD Publishing. ( 10.1787/5jlr2z70k0bx-en) [DOI] [Google Scholar]
  • 117. Xia J. 2019. Economic modelling of predatory journal publishing. Pub. Res. Q. 35, 377–390. ( 10.1007/s12109-019-09661-9) [DOI] [Google Scholar]
  • 118. Perianes-Rodríguez A, Olmeda-Gómez C. 2019. Effects of journal choice on the visibility of scientific publications: a comparison between subscription-based and full Open Access models. Scientometrics 121, 1737–1752. ( 10.1007/s11192-019-03265-y) [DOI] [Google Scholar]
  • 119. Lin WYC. 2021. Effects of open access and articles-in-press mechanisms on publishing lag and first-citation speed: a case on energy and fuels journals. Scientometrics 126, 4841–4869. ( 10.1007/s11192-021-03956-5) [DOI] [Google Scholar]
  • 120. Macgregor G. 2020. Enhancing content discovery of open repositories: an analytics-based evaluation of repository optimizations. Publications 8, 8. ( 10.3390/publications8010008) [DOI] [Google Scholar]
  • 121. Houghton J, Sheehan P. 2009. Estimating the potential impacts of open access to research findings. Econ. Anal. Policy 39, 127–142. ( 10.1016/S0313-5926(09)50048-3) [DOI] [Google Scholar]
  • 122. Wei M, Noroozi Chakoli A. 2020. Evaluating the relationship between the academic and social impact of open access books based on citation behaviors and social media attention. Scientometrics 125, 2401–2420. ( 10.1007/s11192-020-03678-0) [DOI] [Google Scholar]
  • 123. Cheng W, Ren S. 2008. Evolution of open access publishing in Chinese scientific journals. Learn. Publ. 21, 140–152. ( 10.1087/095315108X288884) [DOI] [Google Scholar]
  • 124. Fraser N, Momeni F, Mayr P, Peters I. 2019. Examining the citation and altmetric advantage of bioRxiv preprints. In 17th International Conference on Scientometrics & Informetrics (ISSI 2019), September 2-5, 2019, Sapienza University, Rome, Italy, Edizioni Efesto, Rom, pp. 667–672. http://issi-society.org/publications/issi-conference-proceedings/proceedings-of-issi-2019/. [Google Scholar]
  • 125. Wang M, Zhang J, Chen G, Chai KH. 2019. Examining the influence of open access on journals’ citation obsolescence by modeling the actual citation process. Scientometrics 119, 1621–1641. ( 10.1007/s11192-019-03085-0) [DOI] [Google Scholar]
  • 126. van Leeuwen TN, Tatum C, Wouters PF. 2018. Exploring possibilities to use bibliometric data to monitor gold open access publishing at the national level. J. Assoc. Inf. Sci. Technol. 69, 1161–1173. ( 10.1002/asi.24029) [DOI] [Google Scholar]
  • 127. Houghton J. 2009. Exploring the costs and benefits of alternative publishing models. In Rethinking electronic publishing: innovation in communication paradigms and technologie, pp. 207–238. Milano, Italy: ELPUB. [Google Scholar]
  • 128. Salisbury L, Chowdhury AR, Smith JJ. 2017. Faculty publications from a research university: the scholarly impact of open access versus non-open access. Sci. Technol. Libr. 36, 187–199. ( 10.1080/0194262X.2016.1273815) [DOI] [Google Scholar]
  • 129. Lawrence S. 2001. Free online availability substantially increases a paper’s impact. Nature 411, 521–521. ( 10.1038/35079151) [DOI] [PubMed] [Google Scholar]
  • 130. Gaulé P, Maystre N. 2011. Getting cited: does open access help? Res. Policy 40, 1332–1338. ( 10.1016/j.respol.2011.05.025) [DOI] [Google Scholar]
  • 131. Ezema IJ. 2021. Gold route open access journals in engineering and technology: analysis of research impact and implications for scholarly communication. ISTL 96, 1–16. ( 10.29173/istl43) [DOI] [Google Scholar]
  • 132. Mikki S, Gjesdal ØL, Strømme TE. 2018. Grades of openness: open and closed articles in Norway. Publications 6, 46. ( 10.3390/publications6040046) [DOI] [Google Scholar]
  • 133. Young JS, Brandes PM. 2020. Green and gold open access citation and interdisciplinary advantage: a bibliometric study of two science journals. J. Acad. Librariansh. 46, 102105. ( 10.1016/j.acalib.2019.102105) [DOI] [Google Scholar]
  • 134. Gargouri Y, Larivière V, Gingras Y, Carr L, Harnad S. 2012. Green and gold open access percentages and growth, by discipline. arXiv ( 10.48550/arXiv.1206.3664) [DOI] [Google Scholar]
  • 135. Hall LM, Hendricks AE. 2020. High-throughput analysis suggests differences in journal false discovery rate by subject area and impact factor but not open access status. BMC Bioinform. 21, 564. ( 10.1186/s12859-020-03817-7) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 136. Jamorabo DS, Koulouris V, Briggs WM, Buscaglia JM, Renelus BD. 2022. Higher author fees in gastroenterology journals are not associated with faster processing times or higher impact. Dig. Dis. Sci. 67, 3562–3567. ( 10.1007/s10620-021-07247-0) [DOI] [PubMed] [Google Scholar]
  • 137. Sotudeh H, Arabzadeh H, Mirzabeigi M. 2019. How do self-archiving and author-pays models associate and contribute to OA citation advantage within hybrid journals. J. Acad. Librariansh. 45, 377–385. ( 10.1016/j.acalib.2019.05.004) [DOI] [Google Scholar]
  • 138. Björk BC, Kanto-Karvonen S, Harviainen JT. 2020. How frequently are articles in predatory open access journals cited. Publications 8, 17. ( 10.3390/publications8020017) [DOI] [Google Scholar]
  • 139. Krawczyk F, Kulczycki E. 2021. How is open access accused of being predatory? The impact of Beall’s lists of predatory journals on academic publishing. J. Acad. Librariansh. 47, 102271. ( 10.1016/j.acalib.2020.102271) [DOI] [Google Scholar]
  • 140. Mueller‐Langer F, Watt R. 2018. How many more cites is a $3,000 Open Access fee buying you? Empirical evidence from a natural experiment. Econ. Inq. 56, 931–954. ( 10.1111/ecin.12545) [DOI] [Google Scholar]
  • 141. Saravudecha C, Na Thungfai D, Phasom C, Gunta-in S, Metha A, Punyaphet P, Sookruay T, Sakuludomkan W, Koonrungsesomboon N. 2023. Hybrid gold open access citation advantage in clinical medicine: analysis of hybrid journals in the Web of Science. Publications 11, 21. ( 10.3390/publications11020021) [DOI] [Google Scholar]
  • 142. McCABE MJ, Snyder CM. 2014. Identifying the effect of Open Access on citations using a panel of science journals. Econ. Inq. 52, 1284–1300. ( 10.1111/ecin.12064) [DOI] [Google Scholar]
  • 143. Kolpekwar JA, Shidham VB. 2021. Impact of cytopathology authors work: comparative analysis based on open-access cytopathology publications versus non-open-access conventional publications. Cytojournal 18, 20. ( 10.25259/Cytojournal_32_2021) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 144. Razumova IK, Kuznetsov A. 2019. Impact of open access models on citation metrics. J. Inf. Sci. Theory Pract. 7, 23–31. ( 10.1633/JISTAP.2019.7.2.2) [DOI] [Google Scholar]
  • 145. Koler-Povh T, Južnič P, Turk G. 2014. Impact of open access on citation of scholarly publications in the field of civil engineering. Scientometrics 98, 1033–1045. ( 10.1007/s11192-013-1101-x) [DOI] [Google Scholar]
  • 146. Asaad M, Elmorsi R, Winocour S, Mericli AF, Reece E, Selber JC, Butler CE, Chu CK. 2022. Impact of predatory journals in plastic surgery literature: researchers beware. Plast. Reconst. Surg. 149, 1234e–1243e. ( 10.1097/PRS.0000000000009054) [DOI] [PubMed] [Google Scholar]
  • 147. Campos JLA, Sobral A, Silva JS, Araújo TAS, Ferreira-Júnior WS, Santoro FR, dos Santos GC, Albuquerque UP. 2016. Insularity and citation behavior of scientific articles in young fields: the case of ethnobiology. Scientometrics 109, 1037–1055. ( 10.1007/s11192-016-2067-2) [DOI] [Google Scholar]
  • 148. Amjad T, Sabir M, Shamim A, Amjad M, Daud A. 2022. Investigating the citation advantage of author-pays charges model in computer science research: a case study of Elsevier and Springer. LHT 40, 685–703. ( 10.1108/LHT-05-2021-0154) [DOI] [Google Scholar]
  • 149. Crowe M, Carlyle D. 2015. Is open access sufficient? A review of the quality of open‐access nursing journals. Int. J. Ment. Health Nurs. 24, 59–64. ( 10.1111/inm.12098) [DOI] [PubMed] [Google Scholar]
  • 150. Langham-Putrow A, Bakker C, Riegelman A. 2021. Is the open access citation advantage real? A systematic review of the citation of open access and subscription-based articles. PLoS One 16, e0253129. ( 10.1371/journal.pone.0253129) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 151. Erfanmanesh M, Teixeira da Silva JA. 2019. Is the soundness-only quality control policy of open access mega journals linked to a higher rate of published errors? Scientometrics 120, 917–923. ( 10.1007/s11192-019-03153-5) [DOI] [Google Scholar]
  • 152. Sugimoto CR, Larivière V, Ni C, Cronin B. 2013. Journal acceptance rates: a cross-disciplinary analysis of variability and relationships with journal measures. J. Informetr. 7, 897–906. ( 10.1016/j.joi.2013.08.007) [DOI] [Google Scholar]
  • 153. Oviedo-García MÁ. 2021. Journal citation reports and the definition of a predatory journal: the case of the Multidisciplinary Digital Publishing Institute (MDPI). Res. Eval. 30, 405–419a. ( 10.1093/reseval/rvab020) [DOI] [Google Scholar]
  • 154. Smith E, Haustein S, Mongeon P, Shu F, Ridde V, Larivière V. 2017. Knowledge sharing in global health research - the impact, uptake and cost of open access to scholarly literature. Health Res. Policy Syst. 15, 73. ( 10.1186/s12961-017-0235-3) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 155. Iyandemye J, Thomas MP. 2019. Low income countries have the highest percentages of open access publication: a systematic computational analysis of the biomedical literature. PLoS One 14, e0220229. ( 10.1371/journal.pone.0220229) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 156. Staudt J. 2020. Mandating access: assessing the NIH’s public access policy. Econ. Policy 35, 269–304. ( 10.1093/epolic/eiaa015) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 157. Zhang L, Watson EM. 2017. Measuring the impact of gold and green open access. J. Acad. Librariansh. 43, 337–345. ( 10.1016/j.acalib.2017.06.004) [DOI] [Google Scholar]
  • 158. Druelinger D, Ma L. 2023. Missing a golden opportunity? An analysis of publication trends by income level in the Directory of Open Access Journals 1987–2020. Learn. Publ. 36, 348–358. ( 10.1002/leap.1543) [DOI] [Google Scholar]
  • 159. Xia J, Lynette Myers R, Kay Wilhoite S. 2011. Multiple open access availability and citation impact. J. Inf. Sci. 37, 19–28. ( 10.1177/0165551510389358) [DOI] [Google Scholar]
  • 160. Alordiah CO, Owamah HI, Ogbinaka EJA, Alordiah MO. 2021. Nigeria’s low contribution to recognized world research literature: causes and remedies. Account. Res. 28, 471–491. ( 10.1080/08989621.2020.1855984) [DOI] [PubMed] [Google Scholar]
  • 161. Valderrama-Zurián JC, Aguilar-Moya R, Gorraiz J. 2019. On the bibliometric nature of a foreseeable relationship: open access and education. Scientometrics 120, 1031–1057. ( 10.1007/s11192-019-03175-z) [DOI] [Google Scholar]
  • 162. Jamali HR, Nabavi M. 2015. Open access and sources of full-text articles in Google Scholar in different subject fields. Scientometrics 105, 1635–1651. ( 10.1007/s11192-015-1642-2) [DOI] [Google Scholar]
  • 163. Vadhera AS, Lee JS, Veloso IL, Khan ZA, Trasolini NA, Gursoy S, Kunze KN, Chahla J, Verma NN. 2022. Open access articles garner increased social media attention and citation rates compared with subscription access research articles: an altmetrics-based analysis. Am. J. Sports Med. 50, 3690–3697. ( 10.1177/03635465221124885) [DOI] [PubMed] [Google Scholar]
  • 164. Clements JC. 2017. Open access articles receive more citations in hybrid marine ecology journals. Facets 2, 1–14. ( 10.1139/facets-2016-0032) [DOI] [Google Scholar]
  • 165. Miller GW, Stoner R, Boissaud-Cooke MA, Lim J, Furness H, Putt O, Lewis TL. 2021. Open access availability of anatomy papers presented at meetings of the American and British Associations of Clinical Anatomists. Clin. Anat. 34, 660–667. ( 10.1002/ca.23580) [DOI] [PubMed] [Google Scholar]
  • 166. Dorta-González P, Suárez-Vega R, Dorta-González MI. 2020. Open access effect on uncitedness: a large-scale study controlling by discipline, source type and visibility. Scientometrics 124, 2619–2644. ( 10.1007/s11192-020-03557-8) [DOI] [Google Scholar]
  • 167. Koundouri P, Chatzistamoulou N, Dávila OG, Giannouli A, Kourogenis N, Xepapadeas A, Xepapadeas P. 2021. Open access in scientific information: sustainability model and business plan for the infrastructure and organization of OpenAIRE. J. Benefit Cost Anal. 12, 170–198. ( 10.1017/bca.2020.26) [DOI] [Google Scholar]
  • 168. Rovira-Esteva S, Olalla-Soler C, Aixelá JF. 2021. Open access in translation and interpreting studies: a bibliometric overview of its impact (1996-2015). Meta 66, 532. ( 10.7202/1088349ar) [DOI] [Google Scholar]
  • 169. Tang M, Bever JD, Yu F. 2017. Open access increases citations of papers in ecology. Ecosphere 8, e01887. ( 10.1002/ecs2.1887) [DOI] [Google Scholar]
  • 170. De Filippo D, Mañana-Rodríguez J. 2020. Open access initiatives in European universities: analysis of their implementation and the visibility of publications in the YERUN network. Scientometrics 125, 2667–2694. ( 10.1007/s11192-020-03705-0) [DOI] [Google Scholar]
  • 171. Cooke R, Jain N. 2021. Open access journals are as likely to be referenced by the orthopaedic literature, despite having a lower impact factor than subscription-based journals. SICOT J 7, 64. ( 10.1051/sicotj/2021062) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 172. Bernius S, Hanauske M, König W, Dugall B. 2009. Open access models and their implications for the players on the scientific publishing market. Econ. Anal. Policy 39, 103–116. ( 10.1016/S0313-5926(09)50046-X) [DOI] [Google Scholar]
  • 173. Pollock D, Michael A. 2019. Open access mythbusting: testing two prevailing assumptions about the effects of open access adoption. Learn. Publ. 32, 7–12. ( 10.1002/leap.1209) [DOI] [Google Scholar]
  • 174. Lewison G. 2017. Open access papers: their growth over time and from different countries, and their citations. In Proc. of ISSI 2017, pp. 1247–1255. Wuhan, China: University of Wuhan. [Google Scholar]
  • 175. Bianchini C, Cosentino C, Paci M, Baccini M. 2020. Open access physical therapy journals: do predatory journals publish lower-quality randomized controlled trials? Arch. Phys. Med. Rehabil. 101, 969–977. ( 10.1016/j.apmr.2019.12.012) [DOI] [PubMed] [Google Scholar]
  • 176. van Leeuwen T, Tatum C, Wouters P. 2015. Open access publishing and citation impact - an international study. In Proc. of ISSI 2015 Istanbul: 15th Int. Society of Scientometrics and Informetrics Conf. Istanbul, Turkey: Bogaziçi University Printhouse. [Google Scholar]
  • 177. Kazikdas KC, et al. 2019. Open access publishing in otorhinolaryngology: an emerging trend among Turkish scientists? Cyprus J. Med. Sci. 3, 164–167. ( 10.5152/cjms.2018.634) [DOI] [Google Scholar]
  • 178. Pourret O, Hursthouse A, Irawan DE, Johannesson K, Liu H, Poujol M, Tartèse R, van Hullebusch ED, Wiche O. 2020. Open Access publishing practice in geochemistry: overview of current state and look to the future. Heliyon 6, e03551. ( 10.1016/j.heliyon.2020.e03551) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 179. Nguyen HTT, Nguyen MH, Le TT, Ho MT, Vuong QH. 2021. Open access publishing probabilities based on gender and authorship structures in Vietnam. Publications 9, 45. ( 10.3390/publications9040045) [DOI] [Google Scholar]
  • 180. Davis PM, Lewenstein BV, Simon DH, Booth JG, Connolly MJL. 2008. Open access publishing, article downloads, and citations: randomised controlled trial. BMJ 337, a568. ( 10.1136/bmj.a568) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 181. Tahim A, Bansal H, Goodson AMC, Payne KFB, Sabharwal S. 2016. Open access publishing: a study of current practice in oral and maxillofacial surgery research. J. Maxillofac. Oral Surg. 15, 517–520. ( 10.1007/s12663-016-0898-2) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 182. Huang CK, Neylon C, Montgomery L, Hosking R, Diprose JP, Handcock RN, Wilson K. 2024. Open access research outputs receive more diverse citations. Scientometrics 129, 825–845. ( 10.1007/s11192-023-04894-0) [DOI] [Google Scholar]
  • 183. Hua F, Sun H, Walsh T, Worthington H, Glenny AM. 2016. Open access to journal articles in dentistry: prevalence and citation impact. J. Dent. 47, 41–48. ( 10.1016/j.jdent.2016.02.005) [DOI] [PubMed] [Google Scholar]
  • 184. Hua F, Sun H, Walsh T, Glenny AM, Worthington H. 2017. Open access to journal articles in oncology: current situation and citation impact. Ann. Oncol. 28, 2612–2617. ( 10.1093/annonc/mdx398) [DOI] [PubMed] [Google Scholar]
  • 185. Björk BC, Solomon D. 2012. Open access versus subscription journals: a comparison of scientific impact. BMC Med. 10, 73. ( 10.1186/1741-7015-10-73) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 186. Davis PM. 2011. Open access, readership, citations: a randomized controlled trial of scientific journal publishing. FASEB J. 25, 2129–2134. ( 10.1096/fj.11-183988) [DOI] [PubMed] [Google Scholar]
  • 187. Matthias L, Jahn N, Laakso M. 2019. Open and closed – what do reverse flips tell us about the scholarly publishing landscape? In Impact of social sciences. LSE Impact Blog. See https://blogs.lse.ac.uk/impactofsocialsciences/2019/05/13/open-and-closed-what-do-reverse-flips-tell-us-about-the-scholarly-publishing-landscape/ (accessed 3 February 2025). [Google Scholar]
  • 188. Marshall Z, Brunger F, Welch V, Asghari S, Kaposy C. 2018. Open availability of patient medical photographs in Google Images search results: cross-sectional study of transgender research. J. Med. Internet Res. 20, e70. ( 10.2196/jmir.8787) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 189. Louderback ER, Gainsbury SM, Heirene RM, Amichia K, Grossman A, Bernhard BJ, LaPlante DA. 2023. Open science practices in gambling research publications (2016-2019): a scoping review. J. Gambl. Stud. 39, 987–1011. ( 10.1007/s10899-022-10120-y) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 190. Spezi V, Wakeling S, Pinfield S, Creaser C, Fry J, Willett P. 2017. Open-access mega-journals: the future of scholarly communication or academic dumping ground? A review. JD 73, 263–283. ( 10.1108/JD-06-2016-0082) [DOI] [Google Scholar]
  • 191. Clauson KA, Veronin MA, Khanfar NM, Lou JQ. 2008. Open-access publishing for pharmacy-focused journals. Am. J. Health Syst. Pharm. 65, 1539–1544. ( 10.2146/ajhp070625) [DOI] [PubMed] [Google Scholar]
  • 192. Campbell JD. 2015. Ownership and pricing of information: a model and application to open access. Inf. Econ. Policy 33, 29–42. ( 10.1016/j.infoecopol.2015.10.001) [DOI] [Google Scholar]
  • 193. Calver MC, Bradley JS. 2010. Patterns of citations of open access and non-open access conservation biology journal papers and book chapters. Conserv. Biol. 24, 872–880. ( 10.1111/j.1523-1739.2010.01509.x) [DOI] [PubMed] [Google Scholar]
  • 194. Clements JC, Daigle RM, Froehlich HE. 2018. Predator in the pool? A quantitative evaluation of non-indexed open access journals in aquaculture research. Front. Mar. Sci. 5, 106. ( 10.3389/fmars.2018.00106) [DOI] [Google Scholar]
  • 195. O’Kelly F, Fernandez N, Koyle MA. 2019. Predatory publishing or a lack of peer review transparency?—a contemporary analysis of indexed open and non-open access articles in paediatric urology. J. Pediatr. Urol. 15, 159. ( 10.1016/j.jpurol.2018.08.019) [DOI] [PubMed] [Google Scholar]
  • 196. Rangaswamy BR. 2021. Preferences and citation impact of open access publishing by Indian research community. In Library philosophy and practice (e-journal). Lincoln, NE: University of Nebraska-Lincoln. [Google Scholar]
  • 197. Wang Z, Chen Y, Glänzel W. 2020. Preprints as accelerator of scholarly communication: an empirical analysis in Mathematics. J. Informetr. 14, 101097. ( 10.1016/j.joi.2020.101097) [DOI] [Google Scholar]
  • 198. Dorta-González P, Santana-Jiménez Y. 2018. Prevalence and citation advantage of gold open access in the subject areas of the Scopus database. Res. Eval. 27, 1–15. ( 10.1093/reseval/rvx035) [DOI] [Google Scholar]
  • 199. Eger T, Mertens A, Scheufen M. 2021. Publication cultures and the citation impact of open access. Manage. Decis. Econ. 42, 1980–1998. ( 10.1002/mde.3429) [DOI] [Google Scholar]
  • 200. Alrawadieh Z. 2020. Publishing in predatory tourism and hospitality journals: mapping the academic market and identifying response strategies. Tourism Hosp. Res. 20, 72–81. ( 10.1177/1467358418800121) [DOI] [Google Scholar]
  • 201. Maleki A. 2015. PubMed and arXiv vs. gold open access: citation, Mendeley, and Twitter uptake of academic articles of Iran. In Proc. of ISSI 2015 Istanbul: 15th Int. Society of Scientometrics and Informetrics Conf. Istanbul, Turkey: Bogaziçi University Printhouse. [Google Scholar]
  • 202. Pastorino R, Milovanovic S, Stojanovic J, Efremov L, Amore R, Boccia S. 2016. Quality assessment of studies published in open access and subscription journals: results of a systematic evaluation. PLoS One 11, e0154217. ( 10.1371/journal.pone.0154217) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 203. Erfanmanesh M. 2019. Quantitative portrait of open access mega-journals. MJLIS 24, 115–131. ( 10.22452/mjlis.vol24no2.7) [DOI] [Google Scholar]
  • 204. Dorta-González P, González-Betancor SM, Dorta-González MI. 2017. Reconsidering the gold open access citation advantage postulate in a multidisciplinary context: an analysis of the subject categories in the Web of Science database 2009–2014. Scientometrics 112, 877–901. ( 10.1007/s11192-017-2422-y) [DOI] [Google Scholar]
  • 205. Fu DY, Hughey JJ. 2019. Releasing a preprint is associated with more attention and citations for the peer-reviewed article. Elife 8, e52646. ( 10.7554/eLife.52646) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 206. Ahmad S, Qureshi IU, Ramzan M, Rehman SU. 2022. Research elite of Pakistan: profile and determinants of productivity and impact. Pub. Res. Q. 38, 263–280. ( 10.1007/s12109-022-09874-5) [DOI] [Google Scholar]
  • 207. Ming W, Zhao Z. 2022. Rethinking the open access citation advantage: evidence from the ‘reverse‐flipping’ journals. J. Assoc. Inf. Sci. Technol. 73, 1608–1620. ( 10.1002/asi.24699) [DOI] [Google Scholar]
  • 208. Snijder R. 2016. Revisiting an open access monograph experiment: measuring citations and tweets 5 years later. Scientometrics 109, 1855–1875. ( 10.1007/s11192-016-2160-6) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 209. Asai S. 2023. Revisiting methodology for identifying open access advantages. Pub. Res. Q. 39, 107–119. ( 10.1007/s12109-023-09946-0) [DOI] [Google Scholar]
  • 210. Nelson L, Ye H, Schwenn A, Lee S, Arabi S, Hutchins BI. 2022. Robustness of evidence reported in preprints during peer review. Lancet Glob. Health 10, e1684–e1687. ( 10.1016/S2214-109X(22)00368-0) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 211. Arendt J, Peacemaker B, Miller H. 2019. Same question, different world: replicating an open access research impact study. C&RL 80, 303–318. ( 10.5860/crl.80.3.303) [DOI] [Google Scholar]
  • 212. Antell K, Foote JS, Foote JB. 2016. Scholarly publishing’s evolving landscape: impact metrics, electronic-only journals, and open access in journalism and communication research. J. Mass Commun. Educ. 71, 309–328. ( 10.1177/1077695816668864) [DOI] [Google Scholar]
  • 213. Breugelmans JG, Roberge G, Tippett C, Durning M, Struck DB, Makanga MM. 2018. Scientific impact increases when researchers publish in open access and international collaboration: a bibliometric analysis on poverty-related disease papers. PLoS One 13, e0203156. ( 10.1371/journal.pone.0203156) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 214. Jokić M, Mervar A, Mateljan S. 2018. Scientific potential of European fully open access journals. Scientometrics 114, 1373–1394. ( 10.1007/s11192-017-2629-y) [DOI] [Google Scholar]
  • 215. Wang HY, Wei SX, Cao C, Wang X, Ye FY. 2021. Scientific value weights more than being open or toll access: an analysis of the OA advantage in Nature and Science. J. Data Inf. Sci. 6, 62–75. ( 10.2478/jdis-2021-0033) [DOI] [Google Scholar]
  • 216. Gargouri Y, Hajjem C, Larivière V, Gingras Y, Carr L, Brody T, Harnad S. 2010. Self-selected or mandated, open access increases citation impact for higher quality research. PLoS One 5, e13636. ( 10.1371/journal.pone.0013636) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 217. Xia J, Nakanishi K. 2012. Self‐selection and the citation advantage of open access articles. Online Inf. Rev. 36, 40–51. ( 10.1108/14684521211206953) [DOI] [Google Scholar]
  • 218. Gray RJ. 2020. Sorry, we’re open: Golden open-access and inequality in non-human biological sciences. Scientometrics 124, 1663–1675. ( 10.1007/s11192-020-03540-3) [DOI] [Google Scholar]
  • 219. Erfanmanesh M. 2017. Status and quality of open access journals in Scopus. CB 36, 155–162. ( 10.1108/CB-02-2017-0007) [DOI] [Google Scholar]
  • 220. Oermann MH, Conklin JL, Nicoll LH, Chinn PL, Ashton KS, Edie AH, Amarasekara S, Budinger SC. 2016. Study of predatory open access nursing journals. J. Nurs. Scholarsh. 48, 624–632. ( 10.1111/jnu.12248) [DOI] [PubMed] [Google Scholar]
  • 221. Sotudeh H, Estakhr Z. 2018. Sustainability of open access citation advantage: the case of Elsevier’s author-pays hybrid open access journals. Scientometrics 115, 563–576. ( 10.1007/s11192-018-2663-4) [DOI] [Google Scholar]
  • 222. Tennant JP, Waldner F, Jacques DC, Masuzzo P, Collister LB, Hartgerink CHJ. 2016. The academic, economic and societal impacts of Open Access: an evidence-based review. F1000Res. 5, 632. ( 10.12688/f1000research.8460.1) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 223. Klebel T, Ross-Hellauer T. 2023. The APC-barrier and its effect on stratification in open access publishing. Quant. Sci. Stud. 4, 22–43. ( 10.1162/qss_a_00245) [DOI] [Google Scholar]
  • 224. Sotudeh H, Ghasempour Z, Yaghtin M. 2015. The citation advantage of author-pays model: the case of Springer and Elsevier OA journals. Scientometrics 104, 581–608. ( 10.1007/s11192-015-1607-5) [DOI] [Google Scholar]
  • 225. Norris M, Oppenheim C, Rowland F. 2008. The citation advantage of open‐access articles. J. Am. Soc. Inf. Sci. Technol. 59, 1963–1972. ( 10.1002/asi.20898) [DOI] [Google Scholar]
  • 226. Metcalfe TS. 2006. The citation impact of digital preprint archives for solar physics papers. Sol. Phys. 239, 549–553. ( 10.1007/s11207-006-0262-7) [DOI] [Google Scholar]
  • 227. Kousha K, Abdoli M. 2010. The citation impact of Open Access agricultural research: a comparison between OA and non‐OA publications. Online Inf. Rev. 34, 772–785. ( 10.1108/14684521011084618) [DOI] [Google Scholar]
  • 228. Wang Y, Cao Z, Zeng DD, Zhang Q, Luo T. 2021. The collective wisdom in the COVID-19 research: comparison and synthesis of epidemiological parameter estimates in preprints and peer-reviewed articles. Int. J. Infect. Dis. 104, 1–6. ( 10.1016/j.ijid.2020.12.040) [DOI] [PubMed] [Google Scholar]
  • 229. Aguado-López E, Becerril-Garcia A. 2020. The commercial model of academic publishing underscoring plan S weakens the existing open access ecosystem in Latin America. In Impact of social sciences. LSE Impact Blog. See https://blogs.lse.ac.uk/impactofsocialsciences/2020/05/20/the-commercial-model-of-academic-publishing-underscoring-plan-s-weakens-the-existing-open-access-ecosystem-in-latin-america/. [Google Scholar]
  • 230. Xu F, Ou G, Ma T, Wang X. 2021. The consistency of impact of preprints and their journal publications. J. Informetr. 15, 101153. ( 10.1016/j.joi.2021.101153) [DOI] [Google Scholar]
  • 231. Walters WH, Wilder EI. 2007. The cost implications of open-access publishing in the life sciences. Bioscience 57, 619–625. ( 10.1641/B570709) [DOI] [Google Scholar]
  • 232. Estakhr Z, Sotudeh H, Abbaspour J. 2021. The cost-effectiveness of the article-processing-charge-funded model across countries in different scientific blocks: the case of Elsevier’s hybrid open access journals. Information Research 26, 897. ( 10.47989/irpaper897) [DOI] [Google Scholar]
  • 233. Ilva J, Laitinen MA, Saarti J. 2016. The costs of open and closed access: using the Finnish research output as an example. Liber Q. 26, 13–27. ( 10.18352/lq.10137) [DOI] [Google Scholar]
  • 234. Knöchelmann M. 2021. The democratisation myth. S&TS 34, 65–89. ( 10.23987/sts.94964) [DOI] [Google Scholar]
  • 235. Liljekvist MS, Andresen K, Pommergaard HC, Rosenberg J. 2014. The Directory of Open Access Journals covers more biomedical open access journals than other databases. PeerJ. ( 10.7287/peerj.preprints.717v1) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 236. Moed HF. 2007. The effect of ‘open access’ on citation impact: an analysis of arXiv’s condensed matter section. J. Am. Soc. Inf. Sci. Technol. 58, 2047–2054. ( 10.1002/asi.20663) [DOI] [Google Scholar]
  • 237. Huang C, Yue X, Chen J, Xu W, Li J. 2019. The effect of ‘open access’ on journal impact factors: a causal analysis of medical journals. Physica A 533, 122043. ( 10.1016/j.physa.2019.122043) [DOI] [Google Scholar]
  • 238. Asai S. 2020. The effect of collaboration with large publishers on the internationality and influence of open access journals for research institutions. Scientometrics 124, 663–677. ( 10.1007/s11192-020-03426-4) [DOI] [Google Scholar]
  • 239. Zhang Y. 2006. The effect of open access on citation impact: a comparison study based on Web citation analysis. Libri 56, 145–156. ( 10.1515/LIBR.2006.145) [DOI] [Google Scholar]
  • 240. van Vlokhoven H. 2019. The effect of open access on research quality. J. Informetr. 13, 751–756. ( 10.1016/j.joi.2019.04.001) [DOI] [Google Scholar]
  • 241. Frandsen TF. 2009. The effects of open access on un-published documents: a case study of economics working papers. J. Informetr. 3, 124–133. ( 10.1016/j.joi.2008.12.002) [DOI] [Google Scholar]
  • 242. Kortlever JTP, Tran TTH, Ring D, Menendez ME. 2019. The growth of poorly cited articles in peer-reviewed orthopaedic journals. Clin. Orthop. Relat. Res. 477, 1727–1735. ( 10.1097/CORR.0000000000000727) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 243. Chua SK, et al. 2017. The impact factor of an open access journal does not contribute to an article’s citations. F1000Res. 6, 208. ( 10.12688/f1000research.10892.1) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 244. Gunasekaran S, Arunachalam S. 2014. The impact factors of open access and subscription journals across fields. Curr. Sci. 107, 380–388. [Google Scholar]
  • 245. Cintra PR, Furnival AC, Milanez DH. 2018. The impact of open access citation and social media on leading top Information Science journals. IB 32, 117. ( 10.22201/iibi.24488321xe.2018.77.57874) [DOI] [Google Scholar]
  • 246. Mwambari D, Ali FA, Barak C. 2022. The impact of open access on knowledge production, consumption and dissemination in Kenya’s higher education system. Third World Q. 43, 1408–1424. ( 10.1080/01436597.2022.2056010) [DOI] [Google Scholar]
  • 247. Bernius S. 2010. The impact of open access on the management of scientific knowledge. Online Inf. Rev. 34, 583–603. ( 10.1108/14684521011072990) [DOI] [Google Scholar]
  • 248. Malkawi L, Al-Ryalat N, Al Hadidi E, Serrieh D, AlRyalat SA. 2019. The impact of open access status on journal indexes of radiology journals. Am. J. Roentgenol. 213, 736–739. ( 10.2214/AJR.19.21215) [DOI] [PubMed] [Google Scholar]
  • 249. AlRyalat SA, Abu Nassar A, Tamimi F, Al-Fraihat E, Assaf L, Ghareeb R, Masoudi M, Al-Essa M. 2019. The impact of open-access status on journal indices: respiratory and pulmonology journals. Curr. Respir. Med. Rev. 15, 26–31. ( 10.2174/1573398X15666190214154531) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 250. Wang Z, Glänzel W, Chen Y. 2020. The impact of preprints in Library and Information Science: an analysis of citations, usage and social attention indicators. Scientometrics 125, 1403–1423. ( 10.1007/s11192-020-03612-4) [DOI] [Google Scholar]
  • 251. AlRyalat SA, et al. 2020. The impact of the open-access ophthalmology journals status on journal indices. Bahrain Med. Bull 42, 44–46. [Google Scholar]
  • 252. AlRyalat SA, Saleh M, Alaqraa M, Alfukaha A, Alkayed Y, Abaza M, Abu Saa H, Alshamiry M. 2019. The impact of the open-access status on journal indices: a review of medical journals. F1000Res. 8, 266. ( 10.12688/f1000research.17979.1) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 253. AlRyalat SA, Nassar AA, Tamimi F, Al-Fraihat E, Assaf L, Ghareeb R, Masoudi M, Al-Essa M. 2019. The impact of the open-access status on journal indices: oncology journals. J. Gastrointest. Oncol. 10, 777–782. ( 10.21037/jgo.2019.02.13) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 254. Zheng E, Fu H. 2022. The influence of open access on the academic purification effect of retracted publications: a causal inference analysis. Proc. Assoc. Inf. Sci. Technol. 59, 853–855. ( 10.1002/pra2.749) [DOI] [Google Scholar]
  • 255. Yu X, Meng Z, Qin D, Shen C, Hua F. 2022. The long-term influence of Open Access on the scientific and social impact of dental journal articles: an updated analysis. J. Dent. 119, 104067. ( 10.1016/j.jdent.2022.104067) [DOI] [PubMed] [Google Scholar]
  • 256. Wang X, Liu C, Mao W, Fang Z. 2015. The open access advantage considering citation, article usage and social media attention. Scientometrics 103, 555–564. ( 10.1007/s11192-015-1547-0) [DOI] [Google Scholar]
  • 257. Clayson PE, Baldwin SA, Larson MJ. 2021. The open access advantage for studies of human electrophysiology: impact on citations and altmetrics. Int. J. Psychophysiol. 164, 103–111. ( 10.1016/j.ijpsycho.2021.03.006) [DOI] [PubMed] [Google Scholar]
  • 258. Lewis CL. 2018. The open access citation advantage: does it exist and what does it mean for libraries? ITAL 37, 50–65. ( 10.6017/ital.v37i3.10604) [DOI] [Google Scholar]
  • 259. Papin-Ramcharan J, Dawe RA. 2006. The other side of the coin for open access publishing – a developing country view. Libri 56, 16–27. ( 10.1515/LIBR.2006.16) [DOI] [Google Scholar]
  • 260. Ottaviani J. 2016. The post-embargo open access citation advantage: it exists (probably), its modest (usually), and the rich get richer (of course). PLoS One 11, e0159614. ( 10.1371/journal.pone.0159614) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 261. Barbaro A, Greenblatt RL, Gentili D, Zedda M. 2015. The presence of high-impact factor open access journals in science, technology, engineering and medicine (STEM) disciplines. JLIS 6, 57–76. ( 10.4403/jlis.it-11257) [DOI] [Google Scholar]
  • 262. Zhang L, Wei Y, Huang Y, Sivertsen G. 2021. The prevalence and impact of different types of open access articles from China and USA. In 18th Int. Conf. of the International Society for Scientometrics and Informetrics. Leuven, Belgium. [Google Scholar]
  • 263. Snijder R. 2010. The profits of free books: an experiment to measure the impact of open access publishing. Learn. Publ. 23, 293–301. ( 10.1087/20100403) [DOI] [Google Scholar]
  • 264. Fraser N, Momeni F, Mayr P, Peters I. 2020. The relationship between bioRxiv preprints, citations and altmetrics. Quant. Sci. Stud 1, 1–21. ( 10.1162/qss_a_00043) [DOI] [Google Scholar]
  • 265. Nguyen MH, Nguyen HTT, Ho MT, Le TT, Vuong QH. 2022. The roles of female involvement and risk aversion in open access publishing patterns in Vietnamese social sciences and humanities. J. Data Inf. Sci. 7, 76–96. ( 10.2478/jdis-2022-0001) [DOI] [Google Scholar]
  • 266. Piwowar H, Priem J, Larivière V, Alperin JP, Matthias L, Norlander B, Farley A, West J, Haustein S. 2018. The state of OA: a large-scale analysis of the prevalence and impact of Open Access articles. PeerJ 6, e4375. ( 10.7717/peerj.4375) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 267. Chen M, Du Y. 2016. The status of open access library and information science journals in SSCI. EL 34, 722–739. ( 10.1108/EL-05-2015-0070) [DOI] [Google Scholar]
  • 268. Heneberg P. 2019. The troubles of high-profile open access megajournals. Scientometrics 120, 733–746. ( 10.1007/s11192-019-03144-6) [DOI] [Google Scholar]
  • 269. Brierley L, Nanni F, Polka JK, Dey G, Pálfy M, Fraser N, Coates JA. 2022. Tracking changes between preprint posting and journal publication during a pandemic. PLoS Biol. 20, e3001285. ( 10.1371/journal.pbio.3001285) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 270. Abdill RJ, Blekhman R. 2019. Tracking the popularity and outcomes of all bioRxiv preprints. Elife 8, e45133. ( 10.7554/eLife.45133) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 271. Chen X, Bharti N, Marsteller MR. 2021. Use of bibliometrics data to understand the citation advantages of different open access categories in Covid-19 related studies. Proc. Assoc. Inf. Sci. Technol. 58, 410–414. ( 10.1002/pra2.469) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 272. Momeni F, Mayr P, Fraser N, Peters I. 2021. What happens when a journal converts to open access? A bibliometric analysis. Scientometrics 126, 9811–9827. ( 10.1007/s11192-021-03972-5) [DOI] [Google Scholar]
  • 273. Yan E, Li K. 2018. Which domains do open‐access journals do best in? A 5‐year longitudinal study. J. Assoc. Inf. Sci. Technol. 69, 844–856. ( 10.1002/asi.24002) [DOI] [Google Scholar]
  • 274. Xia J, Harmon JL, Connolly KG, Donnelly RM, Anderson MR, Howard HA. 2015. Who publishes in ‘predatory’ journals? J. Assoc. Inf. Sci. Technol. 66, 1406–1417. ( 10.1002/asi.23265) [DOI] [Google Scholar]
  • 275. Atchison A, Bull J. 2015. Will open access get me cited? An analysis of the efficacy of open access publishing in political science. APSC 48, 129–137. ( 10.1017/S1049096514001668) [DOI] [Google Scholar]
  • 276. Li Y, Wu C, Yan E, Li K. 2018. Will open access increase journal CiteScores? An empirical investigation over multiple disciplines. PLoS One 13, e0201885. ( 10.1371/journal.pone.0201885) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 277. Tonia T, Van Oyen H, Berger A, Schindler C, Künzli N. 2016. If I tweet will you cite? The effect of social media exposure of articles on downloads and citations. Int. J. Public Health 61, 513–520. ( 10.1007/s00038-016-0831-y) [DOI] [PubMed] [Google Scholar]
  • 278. Teixeira da Silva JA, Kimotho SG. 2022. Signs of divisiveness, discrimination and stigmatization caused by Jeffrey Beall’s ‘predatory’ open access publishing blacklists and philosophy. J. Acad. Librariansh. 48, 102418. ( 10.1016/j.acalib.2021.102418) [DOI] [Google Scholar]
  • 279. Mozersky J, Friedrich AB, DuBois JM. 2022. A content analysis of 100 qualitative health research articles to examine researcher-participant relationships and implications for data sharing. Int. J. Qual. Methods 21, 160940692211050. ( 10.1177/16094069221105074) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 280. Smale NA, Unsworth K, Denyer G, Magatova E, Barr D. 1970. A review of the history, advocacy and efficacy of data management plans. IJDC 15, 30. ( 10.2218/ijdc.v15i1.525) [DOI] [Google Scholar]
  • 281. Marwick B, Birch SEP. 2018. A standard for the scholarly citation of archaeological data as an incentive to data sharing. Adv. Archaeol. Pract. 6, 125–143. ( 10.1017/aap.2018.3) [DOI] [Google Scholar]
  • 282. Christensen G, Dafoe A, Miguel E, Moore DA, Rose AK. 2019. A study of the impact of data sharing on article citations using journal policies as a natural experiment. PLoS One 14, e0225883. ( 10.1371/journal.pone.0225883) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 283. Hardwicke TE, et al. 2021. Analytic reproducibility in articles receiving open data badges at the journal Psychological Science: an observational study. R. Soc. Open Sci. 8, 201494. ( 10.1098/rsos.201494) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 284. Robinson‐García N, Jiménez‐Contreras E, Torres‐Salinas D. 2016. Analyzing data citation practices using the data citation index. J. Assoc. Inf. Sci. Technol. 67, 2964–2975. ( 10.1002/asi.23529) [DOI] [Google Scholar]
  • 285. Bezuidenhout LM, Leonelli S, Kelly AH, Rappert B. 2017. Beyond the digital divide: towards a situated approach to open data. Sci. Publ. Policy 44, 464–475. ( 10.1093/scipol/scw036) [DOI] [Google Scholar]
  • 286. Nguyen PY, et al. 2022. Changing patterns in reporting and sharing of review data in systematic reviews with meta-analysis of the effects of interventions: a meta-research study. Epidemiology. ( 10.1101/2022.04.11.22273688) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 287. Bressan P. 2019. Confounds in ‘failed’ replications. Front. Psychol. 10, 1884. ( 10.3389/fpsyg.2019.01884) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 288. Hardwicke TE, et al. 2018. Data availability, reusability, and analytic reproducibility: evaluating the impact of a mandatory open data policy at the journal Cognition. R. Soc. Open Sci. 5, 180448. ( 10.1098/rsos.180448) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 289. Leitner F, Bielza C, Hill SL, Larrañaga P. 2016. Data publications correlate with citation impact. Front. Neurosci. 10, 419. ( 10.3389/fnins.2016.00419) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 290. Piwowar HA, Vision TJ. 2013. Data reuse and the open data citation advantage. PeerJ 1, e175. ( 10.7717/peerj.175) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 291. Naudet F, Sakarovitch C, Janiaud P, Cristea I, Fanelli D, Moher D, Ioannidis JPA. 2018. Data sharing and reanalysis of randomized controlled trials in leading biomedical journals with a full data sharing policy: survey of studies published in The BMJ and PLoS Medicine. BMJ 360, k400. ( 10.1136/bmj.k400) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 292. Huser V, Shmueli-Blumberg D. 2018. Data sharing platforms for de-identified data from human clinical trials. Clin. Trials 15, 413–423. ( 10.1177/1740774518769655) [DOI] [PubMed] [Google Scholar]
  • 293. Claesen A, Vanpaemel W, Maerten AS, Verliefde T, Tuerlinckx F, Heyman T. 2023. Data sharing upon request and statistical consistency errors in psychology: A replication of Wicherts, Bakker and Molenaar (2011). PLoS One 18, e0284243. ( 10.1371/journal.pone.0284243) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 294. Wiley C. 2021. Data sharing: an analysis of medical faculty journals and articles. Sci. Technol. Libr. 40, 104–115. ( 10.1080/0194262X.2020.1781740) [DOI] [Google Scholar]
  • 295. Kafkas Ş, Kim JH, Pi X, McEntyre JR. 2015. Database citation in supplementary data linked to Europe PubMed Central full text biomedical articles. J. Biomed. Semantics 6, 1–7. ( 10.1186/2041-1480-6-1) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 296. Thompson WH, Wright J, Bissett PG, Poldrack RA. 2020. Dataset decay and the problem of sequential analyses on open datasets. eLife 9, e53498. ( 10.7554/eLife.53498) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 297. McGillivray B, Marongiu P, Pedrazzini N, Ribary M, Wigdorowitz M, Zordan E. 2022. Deep impact: a study on the impact of data papers and datasets in the humanities and social sciences. Publications 10, 39. ( 10.3390/publications10040039) [DOI] [Google Scholar]
  • 298. Quarati A, Raffaghelli JE. 2022. Do researchers use open research data? Exploring the relationships between usage trends and metadata quality across scientific disciplines from the Figshare case. J. Inf. Sci. 48, 423–448. ( 10.1177/0165551520961048) [DOI] [Google Scholar]
  • 299. Zhang L, Ma L. 2021. Does open data boost journal impact: evidence from Chinese economics. Scientometrics 126, 3393–3419. ( 10.1007/s11192-021-03897-z) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 300. Spiegelman E. 2021. Esteemed colleagues: a model of the effect of open data on selective reporting of scientific results. Front. Psychol. 12, 761168. ( 10.3389/fpsyg.2021.761168) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 301. Raffaghelli JE, Manca S. 2023. Exploring the social activity of open research data on ResearchGate: implications for the data literacy of researchers. OIR 47, 197–217. ( 10.1108/OIR-05-2021-0255) [DOI] [Google Scholar]
  • 302. Wulder MA, et al. 2022. Fifty years of Landsat science and impacts. Remote Sens. Environ. 280, 113195. ( 10.1016/j.rse.2022.113195) [DOI] [Google Scholar]
  • 303. Bishop BW, Collier H. 2022. Fitness for use of data: scientists’ heuristics of discovery and reuse behaviour framed by the FAIR Data Principles. Inf. Res. 27, 942. ( 10.47989/irpaper942) [DOI] [Google Scholar]
  • 304. Khalil AT, Shinwari ZK, Islam A. 2022. Fostering openness in open science: an ethical discussion of risks and benefits. Front. Polit. Sci. 4, 930574. ( 10.3389/fpos.2022.930574) [DOI] [Google Scholar]
  • 305. Bezuidenhout L, Chakauya E. 2018. Hidden concerns of sharing research data by low/middle-income country scientists. Glob. Bioet. 29, 39–54. ( 10.1080/11287462.2018.1441780) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 306. Wallis JC, Rolando E, Borgman CL. 2013. If we share data, will anyone use them? Data sharing and reuse in the long tail of science and technology. PLoS One 8, e67332. ( 10.1371/journal.pone.0067332) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 307. Cummings JA, Zagrodney JM, Day TE. 2015. Impact of open data policies on consent to participate in human subjects research: discrepancies between participant action and reported concerns. PLoS One 10, e0125208. ( 10.1371/journal.pone.0125208) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 308. Wang H, Gainey M, Campbell P, Young S, Behrman K. 2022. Implementation and assessment of an end-to-end Open Science & Data Collaborations program. F1000Res. 11, 501. ( 10.12688/f1000research.110355.2) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 309. Cannon M, Kelly A, Freeman C. 2022. Implementing an Open & FAIR data sharing policy—a case study in the earth and environmental sciences. Learn. Publ. 35, 56–66. ( 10.1002/leap.1442) [DOI] [Google Scholar]
  • 310. Kwon S, Motohashi K. 2021. Incentive or disincentive for research data disclosure? A large-scale empirical analysis and implications for open science policy. Int. J. Inf. Manage. 60, 102371. ( 10.1016/j.ijinfomgt.2021.102371) [DOI] [Google Scholar]
  • 311. Carroll SR, Rodriguez-Lonebear D, Martinez A. 2019. Indigenous data governance: strategies from United States native nations. Data Sci. J. 18, 31. ( 10.5334/dsj-2019-031) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 312. Holt J, Walker A, Jones P. 2021. Introducing a data availability policy for journals at IOP Publishing: measuring the impact on authors and editorial teams. Learn. Publ. 34, 478–486. ( 10.1002/leap.1386) [DOI] [Google Scholar]
  • 313. Zhang L, Ma L. 2023. Is open science a double-edged sword?: data sharing and the changing citation pattern of Chinese economics articles. Scientometrics 128, 2803–2818. ( 10.1007/s11192-023-04684-8) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 314. Nuijten MB, Borghuis J, Veldkamp CLS, Dominguez-Alvarez L, van Assen MALM, Wicherts JM. 2017. Journal data sharing policies and statistical reporting inconsistencies in psychology. Collabra Psychol. 3, 31. ( 10.1525/collabra.102) [DOI] [Google Scholar]
  • 315. Khan N, Thelwall M, Kousha K. 2021. Measuring the impact of biodiversity datasets: data reuse, citations and altmetrics. Scientometrics 126, 3621–3639. ( 10.1007/s11192-021-03890-6) [DOI] [Google Scholar]
  • 316. Kuula A. 2011. Methodological and ethical dilemmas of archiving qualitative data. IASSIST Q. 34, 12. ( 10.29173/iq455) [DOI] [Google Scholar]
  • 317. Abebe R, Aruleba K, Birhane A, Kingsley S, Obaido G, Remy SL, Sadagopan S. 2021. Narratives and counternarratives on data sharing in Africa. In Proc. of the 2021 ACM Conf. on Fairness, Accountability, and Transparency, virtual event Canada, pp. 329–341. New York, NY: ACM. ( 10.1145/3442188.3445897). https://dl.acm.org/doi/proceedings/10.1145/3442188. [DOI] [Google Scholar]
  • 318. Berberi I, Roche DG. 2022. No evidence that mandatory open data policies increase error correction. Nat. Ecol. Evol. 6, 1630–1633. ( 10.1038/s41559-022-01879-9) [DOI] [PubMed] [Google Scholar]
  • 319. Tanhua T, et al. 2019. Ocean FAIR data services. Front. Mar. Sci. 6, 440. ( 10.3389/fmars.2019.00440) [DOI] [Google Scholar]
  • 320. Mueller-Langer F, Andreoli-Versbach P. 2018. Open access to research data: strategic delay and the ambiguous welfare effects of mandatory data disclosure. Inf. Econ. Policy 42, 20–34. ( 10.1016/j.infoecopol.2017.05.004) [DOI] [Google Scholar]
  • 321. Huang HC, Ma Y, Wang Y. 2024. Open data policy and journal impacts: a synthetic difference-in-differences approach. Appl. Econ. Lett. 31, 1368–1373. ( 10.1080/13504851.2023.2187017) [DOI] [Google Scholar]
  • 322. Carris NW, Cheon B, Wolfson J. 2021. Open data, trials and new ethics of using others’ work. J. Med. Ethics 47, e34–e34. ( 10.1136/medethics-2019-105898) [DOI] [PubMed] [Google Scholar]
  • 323. Waltman L, European Commission and Directorate-General for Research and Innovation . 2019. Open metadata of scholarly publications: open science monitor case study. Brussels: Publications Office of the European Union. ( 10.2777/132318) [DOI] [Google Scholar]
  • 324. Carroll SR, Herczog E, Hudson M, Russell K, Stall S. 2021. Operationalizing the CARE and FAIR Principles for Indigenous data futures. Sci. Data 8, 108. ( 10.1038/s41597-021-00892-0) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 325. Aleixandre-Benavent R, Sapena AF, Ferrer SC, Peset F, García AG. 2019. Policies regarding public availability of published research data in pediatrics journals. Scientometrics 118, 439–451. ( 10.1007/s11192-018-2978-1) [DOI] [Google Scholar]
  • 326. Piwowar HA, Chapman WW. 2010. Public sharing of research datasets: a pilot study of associations. J. Informetr. 4, 148–156. ( 10.1016/j.joi.2009.11.010) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 327. Irwin S. 2013. Qualitative secondary data analysis: ethics, epistemology and context. Prog. Dev. Stud. 13, 295–306. ( 10.1177/1464993413490479) [DOI] [Google Scholar]
  • 328. Shanahan H, Bezuidenhout L. 2022. Rethinking the A in FAIR data: issues of data access and accessibility in research. Front. Res. Metr. Anal. 7, 912456. ( 10.3389/frma.2022.912456) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 329. Bishop L, Kuula-Luumi A. 2017. Revisiting qualitative data reuse. Sage Open 7, 215824401668513. ( 10.1177/2158244016685136) [DOI] [Google Scholar]
  • 330. Laurinavichyute A, Yadav H, Vasishth S. 2022. Share the code, not just the data: a case study of the reproducibility of articles published in the Journal of Memory and Language under the open data policy. J. Mem. Lang. 125, 104332. ( 10.1016/j.jml.2022.104332) [DOI] [Google Scholar]
  • 331. Drachen TM, Ellegaard O, Larsen AV, Fabricius Dorch SB. 2016. Sharing data increases citations. Liber Q. 26, 67–82. ( 10.18352/lq.10149) [DOI] [Google Scholar]
  • 332. Colavizza G, Hrynaszkiewicz I, Staden I, Whitaker K, McGillivray B. 2020. The citation advantage of linking publications to research data. PLoS One 15, e0230416. ( 10.1371/journal.pone.0230416) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 333. Harrison PW, et al. 2021. The COVID-19 data portal: accelerating SARS-CoV-2 and COVID-19 research through rapid open access data sharing. Nucleic Acids Res. 49, W619–W623. ( 10.1093/nar/gkab417) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 334. Gries C, Hanson PC, O’Brien M, Servilla M, Vanderbilt K, Waide R. 2023. The environmental data initiative: connecting the past to the future through data reuse. Ecol. Evol. 13, e9592. ( 10.1002/ece3.9592) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 335. Reinertsen I, Collins DL, Drouin S. 2020. The Essential role of open data and software for the future of ultrasound-based neuronavigation. Front. Oncol. 10, 619274. ( 10.3389/fonc.2020.619274) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 336. AlRyalat SA, El Khatib O, Al-qawasmi O, Alkasrawi H, al Zu’bi R, Abu-Halaweh M, alkanash Y, Habash I. 2020. The impact of the National Heart, Lung, and Blood Institute data: analyzing published articles that used BioLINCC open access data. F1000Res. 9, 30. ( 10.12688/f1000research.21884.1) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 337. Grant R, Hrynaszkiewicz I. 2018. The impact on authors and editors of introducing Data Availability Statements at Nature journals. Sci. Commun. Educ. ( 10.1101/264929) [DOI] [Google Scholar]
  • 338. Markiewicz CJ, et al. 2021. The OpenNeuro resource for sharing of neuroscience data. Elife 10, e71774. ( 10.7554/eLife.71774) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 339. Ali B, Dahlhaus P. 2022. The role of FAIR data towards sustainable agricultural performance: a systematic literature review. Agriculture 12, 309. ( 10.3390/agriculture12020309) [DOI] [Google Scholar]
  • 340. Heller N, et al. 2019. The role of publicly available data in MICCAI papers from 2014 to 2018. In Large-scale annotation of biomedical data and expert label synthesis and hardware aware learning for medical imaging and computer assisted intervention (ed. Zhou L), pp. 70–77. Cham, Switzerland: Springer International Publishing. ( 10.1007/978-3-030-33642-4_8) [DOI] [Google Scholar]
  • 341. Xia J, Liu Y. 2013. Usage patterns of open genomic data. CRL 74, 195–207. ( 10.5860/crl-324) [DOI] [Google Scholar]
  • 342. Coady SA, Mensah GA, Wagner EL, Goldfarb ME, Hitchcock DM, Giffen CA. 2017. Use of the National Heart, Lung, and Blood Institute data repository. N. Engl. J. Med. 376, 1849–1858. ( 10.1056/NEJMsa1603542) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 343. Bishop L. 2012. Using archived qualitative data for teaching: practical and ethical considerations. Int. J. Soc. Res. Methodol. 15, 341–350. ( 10.1080/13645579.2012.688335) [DOI] [Google Scholar]
  • 344. Crosby CJ, Arrowsmith JR, Nandigam V. 2020. Zero to a trillion: Advancing Earth surface process studies with open access to high-resolution topography. In Developments in earth surface processes, pp. 317–338. Oxford, UK: Elsevier. ( 10.1016/B978-0-444-64177-9.00011-4) [DOI] [Google Scholar]
  • 345. Piwowar HA, Day RS, Fridsma DB. 2007. Sharing detailed research data is associated with increased citation rate. PLoS One 2, e308. ( 10.1371/journal.pone.0000308) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 346. Klebel T, Traag V. 2024. Introduction to causality in science studies. SocArXiv. ( 10.31235/osf.io/4bw9e) [DOI] [Google Scholar]
  • 347. Drewry DH, Wells CI, Zuercher WJ, Willson TM. 2019. A perspective on extreme open science: companies sharing compounds without restriction. SLAS Discov. 24, 505–514. ( 10.1177/2472555219838210) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 348. Bakker M, Veldkamp CLS, van den Akker OR, van Assen MALM, Crompvoets E, Ong HH, Wicherts JM. 2020. Recommendations in pre-registrations and internal review board proposals promote formal power analyses but do not increase sample size. PLoS One 15, e0236079. ( 10.1371/journal.pone.0236079) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 349. Ebersole CR, et al. 2020. Many Labs 5: testing pre-data-collection peer review as an intervention to increase replicability. Adv. Methods Pract. Psychol. Sci. 3, 309–331. ( 10.1177/2515245920958687) [DOI] [Google Scholar]
  • 350. Vilaró M, et al. 2019. Adherence to reporting guidelines increases the number of citations: the argument for including a methodologist in the editorial process and peer-review. BMC Med. Res. Methodol. 19, 112. ( 10.1186/s12874-019-0746-4) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 351. Bakker M, Veldkamp CLS, van Assen MALM, Crompvoets EAV, Ong HH, Nosek BA, Soderberg CK, Mellor D, Wicherts JM. 2020. Ensuring the quality and specificity of preregistrations. PLoS Biol. 18, e3000937. ( 10.1371/journal.pbio.3000937) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 352. Schafmeister F. 2021. The effect of replications on citation patterns: evidence from a large-scale reproducibility project. Psychol. Sci. 32, 1537–1548. ( 10.1177/09567976211005767) [DOI] [PubMed] [Google Scholar]
  • 353. Shmueli-Blumberg D, Hu L, Allen C, Frasketi M, Wu LT, Vanveldhuisen P. 2013. The national drug abuse treatment clinical trials network data share project: website design, usage, challenges, and future directions. Clin. Trials 10, 977–986. ( 10.1177/1740774513503522) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 354. Markowitz DM, Song H, Taylor SH. 2021. Tracing the adoption and effects of Open Science in communication research. J. Commun. 71, 739–763. ( 10.1093/joc/jqab030) [DOI] [Google Scholar]
  • 355. Susanin A, Boyar A, Costello K, Fraiman A, Misrok A, Sears M, Hildebrandt T. 2022. Rigor and reproducibility for data analysis and design in the study of eating disorders. Int. J. Eat. Disord. 55, 1267–1278. ( 10.1002/eat.23774) [DOI] [PubMed] [Google Scholar]
  • 356. Switters J, Osimo D, CWTS, Directorate-General for Research, Elsevier, Lisbon Council . 2019. Electronic Laboratory Notebooks (ELNs) as key enablers of open science: open science monitor case study. Luxembourg: Publications Office of the European Union. See https://data.europa.eu/doi/10.2777/07890. [Google Scholar]
  • 357. Hocker J, Bipat T, McDonald DW, Zachry M. 2021. Evaluating QualiCO: an ontology to facilitate qualitative methods sharing to support open science. J. Internet Serv. Appl. 12, 5. ( 10.1186/s13174-021-00135-w) [DOI] [Google Scholar]
  • 358. Blasco A, et al. 2019. Advancing computational biology and bioinformatics research through open innovation competitions. PLoS One 14, e0222165. ( 10.1371/journal.pone.0222165) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 359. Coetzee S, Ivánová I, Mitasova H, Brovelli M. 2020. Open geospatial software and data: a review of the current state and a perspective into the future. ISPRS Int. J. Geoinf. 9, 90. ( 10.3390/ijgi9020090) [DOI] [Google Scholar]
  • 360. Heumüller R, Nielebock S, Krüger J, Ortmeier F. 2020. Publish or perish, but do not forget your software artifacts. Empir. Software Eng. 25, 4585–4616. ( 10.1007/s10664-020-09851-6) [DOI] [Google Scholar]
  • 361. McCormick M, Liu X, Jomier J, Marion C, Ibanez L. 2014. ITK: enabling reproducible research and open science. Front. Neuroinform. 8, 13. ( 10.3389/fninf.2014.00013) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 362. Ratib O, Rosset A. 2006. Open-source software in medical imaging: development of OsiriX. Int. J. Comput. Assist. Radiol. Surg. 1, 187–196. ( 10.1007/s11548-006-0056-2) [DOI] [Google Scholar]
  • 363. Wallace G, et al. 2022. RT-Cloud: a cloud-based software framework to simplify and standardize real-time fMRI. Neuroimage 257, 119295. ( 10.1016/j.neuroimage.2022.119295) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 364. Winter S, Timperley CS, Hermann B, Cito J, Bell J, Hilton M, Beyer D. 2022. A retrospective study of one decade of artifact evaluations. In Proc. of the 30th ACM Joint European Software Engineering Conf. and Symp. on the Foundations of Software Engineering, Singapore Singapore, pp. 145–156. New York, NY: ACM. ( 10.1145/3540250.3549172) [DOI] [Google Scholar]
  • 365. Ratib O, Rosset A, Heuberger J. 2011. Open Source software and social networks: disruptive alternatives for medical imaging. Eur. J. Radiol. 78, 259–265. ( 10.1016/j.ejrad.2010.05.004) [DOI] [PubMed] [Google Scholar]
  • 366. Schröder M, Krüger F, Spors S. 2019. Reproducible research is more than publishing research artefacts: a systematic analysis of Jupyter notebooks from research articles. Arxiv. ( 10.48550/arXiv.1905.00092) [DOI] [Google Scholar]
  • 367. Open Source Initiative . 2007. The open source definition. See https://opensource.org/osd (accessed 16 January 2025).
  • 368. Gruenpeter M, et al. 2021. Defining research software: a controversial discussion. Zenodo. ( 10.5281/zenodo.5504016) [DOI]
  • 369. Odenwald SF. 2020. A citation study of earth science projects in citizen science. PLoS One 15, e0235265. ( 10.1371/journal.pone.0235265) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 370. Aplin LM, Major RE, Davis A, Martin JM. 2021. A citizen science approach reveals long-term social network structure in an urban parrot, Cacatua galerita. J. Anim. Ecol. 90, 222–232. ( 10.1111/1365-2656.13295) [DOI] [PubMed] [Google Scholar]
  • 371. Fowler PW, et al. 2022. A crowd of BashTheBug volunteers reproducibly and accurately measure the minimum inhibitory concentrations of 13 antitubercular drugs from photographs of 96-well broth microdilution plates. eLife 11, e75046. ( 10.7554/eLife.75046) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 372. Resnik DB, Elliott KC, Miller AK. 2015. A framework for addressing ethical issues in citizen science. Environ. Sci. Policy 54, 475–481. ( 10.1016/j.envsci.2015.05.008) [DOI] [Google Scholar]
  • 373. Swanson A, Kosmala M, Lintott C, Packer C. 2016. A generalized approach for producing, quantifying, and validating citizen science data from wildlife images. Conserv. Biol. 30, 520–531. ( 10.1111/cobi.12695) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 374. Conrad CC, Hilchey KG. 2011. A review of citizen science and community-based environmental monitoring: issues and opportunities. Environ. Monit. Assess. 176, 273–291. ( 10.1007/s10661-010-1582-5) [DOI] [PubMed] [Google Scholar]
  • 375. Lee KA, Lee JR, Bell P. 2020. A review of Citizen Science within the Earth Sciences: potential benefits and obstacles. Proc. Geol. Assoc. 131, 605–617. ( 10.1016/j.pgeola.2020.07.010) [DOI] [Google Scholar]
  • 376. Marks L, Laird Y, Trevena H, Smith BJ, Rowbotham S. 2022. A scoping review of Citizen Science approaches in chronic disease prevention. Front. Public Health 10, 743348. ( 10.3389/fpubh.2022.743348) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 377. Vianna GMS, Meekan MG, Bornovski TH, Meeuwig JJ. 2014. Acoustic telemetry validates a citizen science approach for monitoring sharks on coral reefs. PLoS One 9, e95565. ( 10.1371/journal.pone.0095565) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 378. Schmeller DS, et al. 2009. Advantages of volunteer-based biodiversity monitoring in Europe. Conserv. Biol. 23, 307–316. ( 10.1111/j.1523-1739.2008.01125.x) [DOI] [PubMed] [Google Scholar]
  • 379. Shumba T, Montgomery RA, Rasmussen GSA, Macdonald DW. 2018. African wild dog habitat use modelling using telemetry data and citizen scientist sightings: are the results comparable? Afr. J. Wildl. Res. 48, 013002. ( 10.3957/056.048.013002) [DOI] [Google Scholar]
  • 380. Lovell S, Hamer M, Slotow R, Herbert D. 2009. An assessment of the use of volunteers for terrestrial invertebrate biodiversity surveys. Biodivers. Conserv. 18, 3295–3307. ( 10.1007/s10531-009-9642-2) [DOI] [Google Scholar]
  • 381. Alfonso L, Gharesifard M, Wehn U. 2022. Analysing the value of environmental citizen-generated data: complementarity and cost per observation. J. Environ. Manage. 303, 114157. ( 10.1016/j.jenvman.2021.114157) [DOI] [PubMed] [Google Scholar]
  • 382. Tsueng G, Nanis M, Fouquier JT, Mayers M, Good BM, Su AI. 2020. Applying citizen science to gene, drug and disease relationship extraction from biomedical abstracts. Bioinformatics 36, 1226–1233. ( 10.1093/bioinformatics/btz678) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 383. Quinlivan L, Chapman DV, Sullivan T. 2020. Applying citizen science to monitor for the Sustainable Development Goal Indicator 6.3.2: a review. Environ. Monit. Assess. 192, 218. ( 10.1007/s10661-020-8193-6) [DOI] [PubMed] [Google Scholar]
  • 384. Manda AK, Etheridge JR, Grace‐McCaskey C, Allen TR, Howard R. 2021. Are groundwater level data collected by citizen scientists trustworthy? A cautionary tale. Ground Water 59, 123–130. ( 10.1111/gwat.13019) [DOI] [PubMed] [Google Scholar]
  • 385. McClure EC, et al. 2020. Artificial intelligence meets citizen science to supercharge ecological monitoring. Patterns 1, 100109. ( 10.1016/j.patter.2020.100109) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 386. Fuccillo KK, Crimmins TM, de Rivera CE, Elder TS. 2015. Assessing accuracy in citizen science-based plant phenology monitoring. Int. J. Biometeorol. 59, 917–926. ( 10.1007/s00484-014-0892-7) [DOI] [PubMed] [Google Scholar]
  • 387. Crall AW, Newman GJ, Stohlgren TJ, Holfelder KA, Graham J, Waller DM. 2011. Assessing citizen science data quality: an invasive species case study: assessing citizen science data quality. Conserv. Lett. 4, 433–442. ( 10.1111/j.1755-263X.2011.00196.x) [DOI] [Google Scholar]
  • 388. Yang J, Xing D, Luo X. 2021. Assessing the performance of a citizen science project for monitoring urban woody plant species diversity in China. Urban For. Urban Green. 59, 127001. ( 10.1016/j.ufug.2021.127001) [DOI] [Google Scholar]
  • 389. Brown G, et al. 2018. Assessing the validity of crowdsourced wildlife observations for conservation using public participatory mapping methods. Biol. Conserv. 227, 141–151. ( 10.1016/j.biocon.2018.09.016) [DOI] [Google Scholar]
  • 390. Crumley RL, Hill DF, Wikstrom Jones K, Wolken GJ, Arendt AA, Aragon CM, Cosgrove C, Community Snow Observations Participants . 2021. Assimilation of citizen science data in snowpack modeling using a new snow data set: Community Snow Observations. Hydrol. Earth Syst. Sci. 25, 4651–4680. ( 10.5194/hess-25-4651-2021) [DOI] [Google Scholar]
  • 391. Robinson WD, Hallman TA, Hutchinson RA. 2021. Benchmark bird surveys help quantify counting accuracy in a citizen-science database. Front. Ecol. Evol. 9, 568278. ( 10.3389/fevo.2021.568278) [DOI] [Google Scholar]
  • 392. Thel L, Chamaillé-Jammes S, Keurinck L, Catala M, Packer C, Huebner SE, Bonenfant C. 2021. Can citizen science analysis of camera trap data be used to study reproduction? Wildlife Biol 2021, 1–9. ( 10.2981/wlb.00833) [DOI] [Google Scholar]
  • 393. Freihardt J. 2020. Can Citizen Science using social media inform sanitation planning? J. Environ. Manage. 259, 110053. ( 10.1016/j.jenvman.2019.110053) [DOI] [PubMed] [Google Scholar]
  • 394. Johansson F, Heino J, Coiffard P, Svanbäck R, Wester J, Bini LM. 2020. Can information from citizen science data be used to predict biodiversity in stormwater ponds? Sci. Rep. 10, 9380. ( 10.1038/s41598-020-66306-0) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 395. Keeping D, Burger JH, Keitsile AO, Gielen MC, Mudongo E, Wallgren M, Skarpe C, Foote AL. 2018. Can trackers count free-ranging wildlife as effectively and efficiently as conventional aerial survey and distance sampling? Implications for citizen science in the Kalahari, Botswana. Biol. Conserv. 223, 156–169. ( 10.1016/j.biocon.2018.04.027) [DOI] [Google Scholar]
  • 396. Lasky M, et al. 2021. Candid critters: challenges and solutions in a large-scale citizen science camera trap project. Citiz. Sci. Theory Pract. 6, 4. ( 10.5334/cstp.343) [DOI] [Google Scholar]
  • 397. Siano R, Chapelle A, Antoine V, Michel-Guillou E, Rigaut-Jalabert F, Guillou L, Hégaret H, Leynaert A, Curd A. 2020. Citizen participation in monitoring phytoplankton seawater discolorations. Mar. Policy 117, 103039. ( 10.1016/j.marpol.2018.01.022) [DOI] [Google Scholar]
  • 398. Shin Y, Kim K, Groffen J, Woo D, Song E, Borzée A. 2022. Citizen science and roadkill trends in the Korean herpetofauna: the importance of spatially biased and unstandardized data. Front. Ecol. Evol. 10, 944318. ( 10.3389/fevo.2022.944318) [DOI] [Google Scholar]
  • 399. Sauermann H, et al. 2020. Citizen science and sustainability transitions. Res. Policy 49, 103978. ( 10.1016/j.respol.2020.103978) [DOI] [Google Scholar]
  • 400. Lawson B, Petrovan SO, Cunningham AA. 2015. Citizen science and wildlife disease surveillance. Ecohealth 12, 693–702. ( 10.1007/s10393-015-1054-z) [DOI] [PubMed] [Google Scholar]
  • 401. Aceves-Bueno E, et al. 2015. Citizen science as an approach for overcoming insufficient monitoring and inadequate stakeholder buy-in in adaptive management: criteria and evidence. Ecosystems 18, 493–506. ( 10.1007/s10021-015-9842-4) [DOI] [Google Scholar]
  • 402. Dickinson JL, Zuckerberg B, Bonter DN. 2010. Citizen science as an ecological research tool: challenges and benefits. Annu. Rev. Ecol. Evol. Syst. 41, 149–172. ( 10.1146/annurev-ecolsys-102209-144636) [DOI] [Google Scholar]
  • 403. Riesch H, Potter C. 2014. Citizen science as seen by scientists: methodological, epistemological and ethical dimensions. Public Underst. Sci. 23, 107–120. ( 10.1177/0963662513497324) [DOI] [PubMed] [Google Scholar]
  • 404. van de Gevel J, van Etten J, Deterding S. 2020. Citizen science breathes new life into participatory agricultural research: a review. Agron. Sustain. Dev. 40, 35. ( 10.1007/s13593-020-00636-1) [DOI] [Google Scholar]
  • 405. McKinley DC, et al. 2017. Citizen science can improve conservation science, natural resource management, and environmental protection. Biol. Conserv. 208, 15–28. ( 10.1016/j.biocon.2016.05.015) [DOI] [Google Scholar]
  • 406. Crall AW, Jarnevich CS, Young NE, Panke BJ, Renz M, Stohlgren TJ. 2015. Citizen science contributes to our knowledge of invasive plant species distributions. Biol. Invasions 17, 2415–2427. ( 10.1007/s10530-015-0885-4) [DOI] [Google Scholar]
  • 407. Farhadinia MS, Moll RJ, Montgomery RA, Ashrafi S, Johnson PJ, Hunter LTB, Macdonald DW. 2018. Citizen science data facilitate monitoring of rare large carnivores in remote montane landscapes. Ecol. Indic. 94, 283–291. ( 10.1016/j.ecolind.2018.06.064) [DOI] [Google Scholar]
  • 408. Trouille L, Lintott CJ, Fortson LF. 2019. Citizen science frontiers: efficiency, engagement, and serendipitous discovery with human-machine systems. Proc. Natl Acad. Sci. USA 116, 1902–1909. ( 10.1073/pnas.1807190116) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 409. Meentemeyer RK, Dorning MA, Vogler JB, Schmidt D, Garbelotto M. 2015. Citizen science helps predict risk of emerging infectious disease. Front. Ecol. Environ. 13, 189–194. ( 10.1890/140299) [DOI] [Google Scholar]
  • 410. Biraghi CA, Carrion D, Brovelli MA. 2022. Citizen science impact on environmental monitoring towards SDGs indicators: the CASE of SIMILE project. Sustainability 14, 8107. ( 10.3390/su14138107) [DOI] [Google Scholar]
  • 411. Adler FR, Green AM, Şekercioğlu ÇH. 2020. Citizen science in ecology: a place for humans in nature. Ann. N. Y. Acad. Sci. 1469, 52–64. ( 10.1111/nyas.14340) [DOI] [PubMed] [Google Scholar]
  • 412. Njue N, Stenfert Kroese J, Gräf J, Jacobs SR, Weeser B, Breuer L, Rufino MC. 2019. Citizen science in hydrological monitoring and ecosystem services management: state of the art and future prospects. Sci. Total Environ. 693, 133531. ( 10.1016/j.scitotenv.2019.07.337) [DOI] [PubMed] [Google Scholar]
  • 413. Buytaert W, et al. 2014. Citizen science in hydrology and water resources: opportunities for knowledge generation, ecosystem service management, and sustainable development. Front. Earth Sci. 2, 1–21. ( 10.3389/feart.2014.00026) [DOI] [Google Scholar]
  • 414. Kawabe LA, Ghilardi-Lopes NP, Turra A, Wyles KJ. 2022. Citizen science in marine litter research: a review. Mar. Pollut. Bull. 182, 114011. ( 10.1016/j.marpolbul.2022.114011) [DOI] [PubMed] [Google Scholar]
  • 415. Weigelhofer G, Pölz EM, Hein T. 2019. Citizen science: how high school students can provide scientifically sound data in biogeochemical experiments. Freshw. Sci. 38, 236–243. ( 10.1086/698765) [DOI] [Google Scholar]
  • 416. Agostini G, SturtzSreetharan C, Wutich A, Williams D, Brewis A. 2019. Citizen sociolinguistics: a new method to understand fat talk. PLoS One 14, e0217618. ( 10.1371/journal.pone.0217618) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 417. Giovos I, et al. 2019. Citizen-science for monitoring marine invasions and stimulating public engagement: a case project from the eastern Mediterranean. Biol. Invasions 21, 3707–3721. ( 10.1007/s10530-019-02083-w) [DOI] [Google Scholar]
  • 418. DiBattista JD, West KM, Hay AC, Hughes JM, Fowler AM, McGrouther MA. 2021. Community‐based citizen science projects can support the distributional monitoring of fishes. Aquat. Conserv. 31, 3580–3593. ( 10.1002/aqc.3726) [DOI] [Google Scholar]
  • 419. Flesch EP, Belt JJ. 2017. Comparing citizen science and professional data to evaluate extrapolated mountain goat distribution models. Ecosphere 8, e01638. ( 10.1002/ecs2.1638) [DOI] [Google Scholar]
  • 420. Harvey GKA, Nelson TA, Paquet PC, Ferster CJ, Fox CH. 2018. Comparing citizen science reports and systematic surveys of marine mammal distributions and densities. Biol. Conserv. 226, 92–100. ( 10.1016/j.biocon.2018.07.024) [DOI] [Google Scholar]
  • 421. Bison M, Yoccoz NG, Carlson BZ, Delestrade A. 2019. Comparison of budburst phenology trends and precision among participants in a citizen science program. Int. J. Biometeorol. 63, 61–72. ( 10.1007/s00484-018-1636-x) [DOI] [PubMed] [Google Scholar]
  • 422. van der Velde T, Milton DA, Lawson TJ, Wilcox C, Lansdell M, Davis G, Perkins G, Hardesty BD. 2017. Comparison of marine debris data collected by researchers and citizen scientists: is citizen science data worth the effort? Biol. Conserv. 208, 127–138. ( 10.1016/j.biocon.2016.05.025) [DOI] [Google Scholar]
  • 423. Chandler M, et al. 2017. Contribution of citizen science towards international biodiversity monitoring. Biol. Conserv. 213, 280–294. ( 10.1016/j.biocon.2016.09.004) [DOI] [Google Scholar]
  • 424. Sauermann H, Franzoni C. 2015. Crowd science user contribution patterns and their implications. Proc. Natl Acad. Sci. USA 112, 679–684. ( 10.1073/pnas.1408907112) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 425. van Emmerik T, Seibert J, Strobl B, Etter S, den Oudendammer T, Rutten M, bin Ab Razak MS, van Meerveld I. 2020. Crowd-based observations of riverine macroplastic pollution. Front. Earth Sci. 8, 298. ( 10.3389/feart.2020.00298) [DOI] [Google Scholar]
  • 426. Swan M. 2012. Crowdsourced health research studies: an important emerging complement to clinical trials in the public health research ecosystem. J. Med. Internet Res. 14, e46. ( 10.2196/jmir.1988) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 427. Watson D, Floridi L. 2018. Crowdsourced science: sociotechnical epistemology in the e-research paradigm. Synthese 195, 741–764. ( 10.1007/s11229-016-1238-2) [DOI] [Google Scholar]
  • 428. Tucker JD, Day S, Tang W, Bayus B. 2019. Crowdsourcing in medical research: concepts and applications. PeerJ 7, e6762. ( 10.7717/peerj.6762) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 429. Beck S, Brasseur TM, Poetz M, Sauermann H. 2022. Crowdsourcing research questions in science. Res. Policy 51, 104491. ( 10.1016/j.respol.2022.104491) [DOI] [Google Scholar]
  • 430. Comber A, Mooney P, Purves RS, Rocchini D, Walz A. 2016. Crowdsourcing: it matters who the crowd are. The impacts of between group variations in recording land cover. PLoS One 11, e0158329. ( 10.1371/journal.pone.0158329) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 431. Fehri R, Bogaert P, Khlifi S, Vanclooster M. 2020. Data fusion of citizen-generated smartphone discharge measurements in Tunisia. J. Hydrol. 590, 125518. ( 10.1016/j.jhydrol.2020.125518) [DOI] [Google Scholar]
  • 432. Leocadio JN, Ghilardi-Lopes NP, Koffler S, Barbiéri C, Francoy TM, Albertini B, Saraiva AM. 2021. Data reliability in a citizen science protocol for monitoring stingless bees flight activity. Insects 12, 766. ( 10.3390/insects12090766) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 433. Ratnieks FLW, Schrell F, Sheppard RC, Brown E, Bristow OE, Garbuzov M. 2016. Data reliability in citizen science: learning curve and the effects of training method, volunteer background and experience on identification accuracy of insects visiting ivy flowers. Methods Ecol. Evol. 7, 1226–1235. ( 10.1111/2041-210X.12581) [DOI] [Google Scholar]
  • 434. Starkey E, Parkin G, Birkinshaw S, Large A, Quinn P, Gibson C. 2017. Demonstrating the value of community-based (‘citizen science’) observations for catchment modelling and characterisation. J. Hydrol. 548, 801–817. ( 10.1016/j.jhydrol.2017.03.019) [DOI] [Google Scholar]
  • 435. Ashley EA, Pattengill-Semmens CV, Orr JW, Nichols JD, Gaydos JK. 2022. Documenting fishes in an inland sea with citizen scientist diver surveys: using taxonomic expertise to inform the observation potential of fish species. Environ. Monit. Assess. 194, 227. ( 10.1007/s10661-022-09857-1) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 436. Dodson JB, Robles MC, Taylor JE, DeFontes CC, Weaver KL. 2019. Eclipse across America: citizen science observations of the 21 August 2017 total solar eclipse. J. Appl. Meteorol. Climatol. 58, 2363–2385. ( 10.1175/JAMC-D-18-0297.1) [DOI] [Google Scholar]
  • 437. Lawson KN, Tracy BM, Sharova M, Muirhead JR, Cawood A. 2022. Engaging online citizen scientists and the consensus method to monitor the marine biofouling community. Front. Mar. Sci. 9, 862430. ( 10.3389/fmars.2022.862430) [DOI] [Google Scholar]
  • 438. Alabri A, Hunter J. 2010. Enhancing the quality and trust of citizen science data. In 2010 IEEE 6th Int. Conf. on E Science (e-Science), Brisbane, Australia, pp. 81–88. Brisbane, Australia: IEEE. ( 10.1109/eScience.2010.33) [DOI] [Google Scholar]
  • 439. Tran T, Porter WT, Salkeld DJ, Prusinski MA, Jensen ST, Brisson D. 2021. Estimating disease vector population size from citizen science data. J. R. Soc. Interface 18, 20210610. ( 10.1098/rsif.2021.0610) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 440. Hochmair HH, Scheffrahn RH, Basille M, Boone M. 2020. Evaluating the data quality of iNaturalist termite records. PLoS One 15, e0226534. ( 10.1371/journal.pone.0226534) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 441. Kremen C, Ullman KS, Thorp RW. 2011. Evaluating the quality of citizen-scientist data on pollinator communities: citizen-scientist pollinator monitoring. Conserv. Biol. 25, 607–617. ( 10.1111/j.1523-1739.2011.01657.x) [DOI] [PubMed] [Google Scholar]
  • 442. Cox TE, Philippoff J, Baumgartner E, Smith CM. 2012. Expert variability provides perspective on the strengths and weaknesses of citizen-driven intertidal monitoring program. Ecol. Appl. 22, 1201–1212. ( 10.1890/11-1614.1) [DOI] [PubMed] [Google Scholar]
  • 443. Facchinelli F, et al. 2022. Extreme citizens science for climate justice: linking pixel to people for mapping gas flaring in Amazon rainforest. Environ. Res. Lett. 17, 024003. ( 10.1088/1748-9326/ac40af) [DOI] [Google Scholar]
  • 444. Eritja R, et al. 2019. First detection of Aedes japonicus in Spain: an unexpected finding triggered by citizen science. Parasit. Vectors 12, 53. ( 10.1186/s13071-019-3317-y) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 445. Roy HE, Baxter E, Saunders A, Pocock MJO. 2016. Focal plant observations as a standardised method for pollinator monitoring: opportunities and limitations for mass participation citizen science. PLoS One 11, e0150794. ( 10.1371/journal.pone.0150794) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 446. Dallaqua FBJR, Fazenda ÁL, Faria FA. 2021. ForestEyes project: conception, enhancements, and challenges. Fut. Gener. Comput. Syst. 124, 422–435. ( 10.1016/j.future.2021.06.002) [DOI] [Google Scholar]
  • 447. English PB, Richardson MJ, Garzón-Galvis C. 2018. From crowdsourcing to extreme citizen science: participatory research for environmental health. Annu. Rev. Public Health 39, 335–350. ( 10.1146/annurev-publhealth-040617-013702) [DOI] [PubMed] [Google Scholar]
  • 448. Theobald EJ, et al. 2015. Global change and local solutions: tapping the unrealized potential of citizen science for biodiversity research. Biol. Conserv. 181, 236–244. ( 10.1016/j.biocon.2014.10.021) [DOI] [Google Scholar]
  • 449. Gunko R, Rapeli L, Scheinin M, Vuorisalo T, Karell P. 2022. How accurate is citizen science? Evaluating public assessments of coastal water quality. Env. Pol. Gov. 32, 149–157. ( 10.1002/eet.1975) [DOI] [Google Scholar]
  • 450. Chen SL. 2019. How empowering is citizen science? Access, credits, and governance for the crowd. East Asian Sci. Technol. Soc. 13, 215–234. ( 10.1215/18752160-7497711) [DOI] [Google Scholar]
  • 451. Camila D, Bilenca D, Agostini G. 2017. In search of the horned frog (Ceratophrys ornata) in Argentina: complementing field surveys with citizen science. Herpetol. Conserv. Biol. 12, 664–672. ( 10.2305/IUCN) [DOI] [Google Scholar]
  • 452. Gardiner MM, Allee LL, Brown PM, Losey JE, Roy HE, Smyth RR. 2012. Lessons from lady beetles: accuracy of monitoring data from US and UK citizen‐science programs. Front. Ecol. Environ. 10, 471–476. ( 10.1890/110185) [DOI] [Google Scholar]
  • 453. Goldstein EA, Lawton C, Sheehy E, Butler F. 2014. Locating species range frontiers: a cost and efficiency comparison of citizen science and hair-tube survey methods for use in tracking an invasive squirrel. Wildl. Res. 41, 64. ( 10.1071/WR13197) [DOI] [Google Scholar]
  • 454. Encarnação J, Baptista V, Teodósio MA, Morais P. 2021. Low-cost citizen science effectively monitors the rapid expansion of a marine invasive species. Front. Environ. Sci. 9, 752705. ( 10.3389/fenvs.2021.752705) [DOI] [Google Scholar]
  • 455. Rathnayake C, Joshi S, Cerratto-Pargman T. 2020. Mapping the current landscape of citizen-driven environmental monitoring: a systematic literature review. Sustain. Sci. Pract. Policy. 16, 326–334. ( 10.1080/15487733.2020.1829845) [DOI] [Google Scholar]
  • 456. Delaney DG, Sperling CD, Adams CS, Leung B. 2008. Marine invasive species: validation of citizen science and implications for national monitoring networks. Biol. Invasions 10, 117–128. ( 10.1007/s10530-007-9114-0) [DOI] [Google Scholar]
  • 457. Roger E, Tegart P, Dowsett R, Kinsela MA, Harley MD, Ortac G. 2020. Maximising the potential for citizen science in New South Wales. Austral. Zool. 40, 449–461. ( 10.7882/AZ.2019.023) [DOI] [Google Scholar]
  • 458. Bedessem B, Julliard R, Montuschi E. 2021. Measuring epistemic success of a biodiversity citizen science program: a citation study. PLoS One 16, e0258350. ( 10.1371/journal.pone.0258350) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 459. Jones JS, et al. 2022. Microplastic distribution and composition on two Galápagos island beaches, Ecuador: verifying the use of citizen science derived data in long-term monitoring. Environ. Pollut. 311, 120011. ( 10.1016/j.envpol.2022.120011) [DOI] [PubMed] [Google Scholar]
  • 460. Danielsen F, Burgess ND, Balmford A. 2005. Monitoring matters: examining the potential of locally-based approaches. Biodivers. Conserv. 14, 2507–2542. ( 10.1007/s10531-005-8375-0) [DOI] [Google Scholar]
  • 461. van der Zee T, Reich J. 2018. Open educational science. SocArXiv. ( 10.31235/osf.io/d9bme) [DOI] [Google Scholar]
  • 462. Soroye P, Ahmed N, Kerr JT. 2018. Opportunistic citizen science data transform understanding of species distributions, phenology, and diversity gradients for global change research. Glob. Chang. Biol. 24, 5281–5291. ( 10.1111/gcb.14358) [DOI] [PubMed] [Google Scholar]
  • 463. Onwezen MC, Bouwman EP, van Trijp HCM. 2021. Participatory methods in food behaviour research: a framework showing advantages and disadvantages of various methods. Foods 10, 470. ( 10.3390/foods10020470) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 464. Kasten P, Jenkins SR, Christofoletti RA. 2021. Participatory monitoring—a citizen science approach for coastal environments. Front. Mar. Sci. 8, 681969. ( 10.3389/fmars.2021.681969) [DOI] [Google Scholar]
  • 465. MacPhail VJ, Colla SR. 2020. Power of the people: a review of citizen science programs for conservation. Biol. Conserv. 249, 108739. ( 10.1016/j.biocon.2020.108739) [DOI] [Google Scholar]
  • 466. Porter WT, Barrand ZA, Wachara J, DaVall K, Mihaljevic JR, Pearson T, Salkeld DJ, Nieto NC. 2021. Predicting the current and future distribution of the western black-legged tick, Ixodes pacificus, across the Western US using citizen science collections. PLoS One 16, e0244754. ( 10.1371/journal.pone.0244754) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 467. Kallimanis AS, Panitsa M, Dimopoulos P. 2017. Quality of non-expert citizen science data collected for habitat type conservation status assessment in Natura 2000 protected areas. Sci. Rep. 7, 8873. ( 10.1038/s41598-017-09316-9) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 468. Cruickshank SS, Bühler C, Schmidt BR. 2019. Quantifying data quality in a citizen science monitoring program: false negatives, false positives and occupancy trends. Conservat. Sci. Prac. 1, e54. ( 10.1111/csp2.54) [DOI] [Google Scholar]
  • 469. Meixner T, et al. 2021. Rapid assessment and long-term monitoring of green stormwater infrastructure with citizen scientists. Sustainability 13, 12520. ( 10.3390/su132212520) [DOI] [Google Scholar]
  • 470. Meschini M, et al. 2021. Reliability of data collected by volunteers: a nine-year citizen science study in the Red Sea. Front. Ecol. Evol. 9, 694258. ( 10.3389/fevo.2021.694258) [DOI] [Google Scholar]
  • 471. Cooper S, Sterling ALR, Kleffner R, Silversmith WM, Siegel JB. 2018. Repurposing citizen science games as software tools for professional scientists. In Proc. of the 13th Int. Conf. on the Foundations of Digital Games, pp. 1–6. Malmö, Sweden: ACM. ( 10.1145/3235765.3235770) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 472. Earp HS, Liconti A. 2020. Science for the future: the use of citizen science in marine research and conservation. In Youmares 9 - the oceans: our research (eds Future O, Jungblut S, Liebich V, Bode-Dalby M), pp. 1–19. Cham, Switzerland: Springer International Publishing. ( 10.1007/978-3-030-20389-4_1) [DOI] [Google Scholar]
  • 473. Smale DA, Epstein G, Parry M, Attrill MJ. 2019. Spatiotemporal variability in the structure of seagrass meadows and associated macrofaunal assemblages in southwest England (UK): using citizen science to benchmark ecological pattern. Ecol. Evol. 9, 3958–3972. ( 10.1002/ece3.5025) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 474. Oturai NG, Bille Nielsen M, Clausen LPW, Hansen SF, Syberg K. 2021. Strength in numbers: how citizen science can upscale assessment of human exposure to plastic pollution. Curr. Opin. Toxicol. 27, 54–59. ( 10.1016/j.cotox.2021.08.003) [DOI] [Google Scholar]
  • 475. Fehri R, Khlifi S, Vanclooster M. 2020. Testing a citizen science water monitoring approach in Tunisia. Environ. Sci. Policy 104, 67–72. ( 10.1016/j.envsci.2019.11.009) [DOI] [Google Scholar]
  • 476. Périquet S, Roxburgh L, le Roux A, Collinson WJ. 2018. Testing the value of citizen science for roadkill studies: a case study from South Africa. Front. Ecol. Evol. 6, 15. ( 10.3389/fevo.2018.00015) [DOI] [Google Scholar]
  • 477. Aceves‐Bueno E, Adeleye AS, Feraud M, Huang Y, Tao M, Yang Y, Anderson SE. 2017. The accuracy of citizen science data: a quantitative review. Bulletin Ecologic. Soc. America 98, 278–290. ( 10.1002/bes2.1336) [DOI] [Google Scholar]
  • 478. Catlin-Groves CL. 2012. The citizen science landscape: from volunteers to citizen sensors and beyond. Int. J. Zool. 2012, 1–14. ( 10.1155/2012/349630) [DOI] [Google Scholar]
  • 479. de Camargo Barbosa KV, Develey PF, Ribeiro MC, Jahn AE. 2021. The contribution of citizen science to research on migratory and urban birds in Brazil. Ornithol. Res. 29, 1–11. ( 10.1007/s43388-020-00031-0) [DOI] [Google Scholar]
  • 480. Hidalgo-Ruz V, Thiel M. 2015. The contribution of citizen scientists to the monitoring of marine litter. In Marine anthropogenic litter (eds Litter MA, Bergmann M, Gutow L, Klages M), pp. 429–447. Cham, Switzerland: Springer International Publishing. ( 10.1007/978-3-319-16510-3_16) [DOI] [Google Scholar]
  • 481. Cooper CB, Shirk J, Zuckerberg B. 2014. The invisible prevalence of citizen science in global research: migratory birds and climate change. PLoS One 9, e106508. ( 10.1371/journal.pone.0106508) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 482. Brown ED, Williams BK. 2019. The potential for citizen science to produce reliable and useful information in ecology. Conserv. Biol. 33, 561–569. ( 10.1111/cobi.13223) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 483. Abourashed A, Doornekamp L, Escartin S, Koenraadt CJM, Schrama M, Wagener M, Bartumeus F, van Gorp ECM. 2021. The potential role of school citizen science programs in infectious disease surveillance: a critical review. Int. J. Environ. Res. Public Health 18, 7019. ( 10.3390/ijerph18137019) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 484. Galloway AWE, Tudor MT, Haegen WMV. 2006. The reliability of citizen science: a case study of Oregon white oak stand surveys. Wildl. Soc. Bull. 34, 1425–1429. ( 10.2193/0091-7648(2006)34[1425:TROCSA]2.0.CO;2) [DOI] [Google Scholar]
  • 485. Foster-Smith J, Evans SM. 2003. The value of marine ecological data collected by volunteers. Biol. Conserv. 113, 199–213. ( 10.1016/S0006-3207(02)00373-7) [DOI] [Google Scholar]
  • 486. Mannino AM, Borfecchia F, Micheli C. 2021. Tracking marine alien macroalgae in the mediterranean sea: the contribution of citizen science and remote sensing. JMSE 9, 288. ( 10.3390/jmse9030288) [DOI] [Google Scholar]
  • 487. Anhalt-Depies C, Stenglein JL, Zuckerberg B, Townsend PA, Rissman AR. 2019. Tradeoffs and tools for data quality, privacy, transparency, and trust in citizen science. Biol. Conserv. 238, 108195. ( 10.1016/j.biocon.2019.108195) [DOI] [Google Scholar]
  • 488. Klemann-Junior L, Villegas Vallejos MA, Scherer-Neto P, Vitule JRS. 2017. Traditional scientific data vs. uncoordinated citizen science effort: a review of the current status and comparison of data on avifauna in southern Brazil. PLoS One 12, e0188819. ( 10.1371/journal.pone.0188819) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 489. Lin YP, Deng D, Lin WC, Lemmens R, Crossman ND, Henle K, Schmeller DS. 2015. Uncertainty analysis of crowd-sourced and professionally collected field data used in species distribution models of Taiwanese moths. Biol. Conserv. 181, 102–110. ( 10.1016/j.biocon.2014.11.012) [DOI] [Google Scholar]
  • 490. Branchini S, et al. 2015. Using a citizen science program to monitor coral reef biodiversity through space and time. Biodivers. Conserv. 24, 319–336. ( 10.1007/s10531-014-0810-7) [DOI] [Google Scholar]
  • 491. Petrovan SO, Vale CG, Sillero N. 2020. Using citizen science in road surveys for large-scale amphibian monitoring: are biased data representative for species distribution? Biodivers. Conserv. 29, 1767–1781. ( 10.1007/s10531-020-01956-0) [DOI] [Google Scholar]
  • 492. Barrows CW, Hoines J, Vamstad MS, Murphy-Mariscal M, Lalumiere K, Heintz J. 2016. Using citizen scientists to assess climate change shifts in desert reptile communities. Biol. Conserv. 195, 82–88. ( 10.1016/j.biocon.2015.12.027) [DOI] [Google Scholar]
  • 493. Kaartinen R, Hardwick B, Roslin T. 2013. Using citizen scientists to measure an ecosystem service nationwide. Ecology 94, 2645–2652. ( 10.1890/12-1165.1) [DOI] [PubMed] [Google Scholar]
  • 494. Pusceddu M, Floris I, Mannu R, Cocco A, Satta A. 2019. Using verified citizen science as a tool for monitoring the European hornet (Vespa crabro) in the island of Sardinia (Italy). NB 50, 97–108. ( 10.3897/neobiota.50.37587) [DOI] [Google Scholar]
  • 495. Engel SR, Voshell JR Jr. 2002. Volunteer biological monitoring: can it accurately assess the ecological condition of streams? Am. Entomol. 48, 164–177. ( 10.1093/ae/48.3.164) [DOI] [Google Scholar]
  • 496. Soroye P, et al. 2022. The risks and rewards of community science for threatened species monitoring. Conservat. Sci. Prac. 4, e12788. ( 10.1111/csp2.12788) [DOI] [Google Scholar]
  • 497. Bianchi F, Squazzoni F. 2022. Can transparency undermine peer review? A simulation model of scientist behavior under open peer review. Sci. Publ. Policy 49, 791–800. ( 10.1093/scipol/scac027) [DOI] [Google Scholar]
  • 498. Bravo G, Grimaldo F, López-Iñesta E, Mehmani B, Squazzoni F. 2019. The effect of publishing peer review reports on referee behavior in five scholarly journals. Nat. Commun. 10, 322. ( 10.1038/s41467-018-08250-2) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 499. Felizardo K, Flach C, Pereira R, Oliveirajr E. 2022. Surveying the audience effect in open peer review of a software engineering workshop. In Proc. of the XXXVI Brazilian Symp. on Software Engineering, Virtual Event Brazil, pp. 446–451. New York, NY, USA: ACM. ( 10.1145/3555228.3555271) [DOI] [Google Scholar]
  • 500. Kowalczuk MK, Dudbridge F, Nanda S, Harriman SL, Patel J, Moylan EC. 2015. Retrospective analysis of the quality of reports by author-suggested and non-author-suggested reviewers in journals operating on open or single-blind peer review models. BMJ Open 5, e008707. ( 10.1136/bmjopen-2015-008707) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 501. Matsui A, Chen E, Wang Y, Ferrara E. 2021. The impact of peer review on the contribution potential of scientific papers. PeerJ 9, e11999. ( 10.7717/peerj.11999) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 502. van Rooyen S, Delamothe T, Evans SJW. 2010. Effect on peer review of telling reviewers that their signed reviews might be posted on the web: randomised controlled trial. BMJ 341, c5729. ( 10.1136/bmj.c5729) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 503. Zong Q, Xie Y, Liang J. 2020. Does open peer review improve citation count? Evidence from a propensity score matching analysis of PeerJ. Scientometrics 125, 607–623. ( 10.1007/s11192-020-03545-y) [DOI] [Google Scholar]
  • 504. Levis AW, Leentjens AFG, Levenson JL, Lumley MA, Thombs BD. 2015. Comparison of self-citation by peer reviewers in a journal with single-blind peer review versus a journal with open peer review. J. Psychosom. Res. 79, 561–565. ( 10.1016/j.jpsychores.2015.08.004) [DOI] [PubMed] [Google Scholar]
  • 505. Radzvilas M, De Pretis F, Peden W, Tortoli D, Osimani B. 2023. Incentives for research effort: an evolutionary model of publication markets with double-blind and open review. Comput. Econ. 61, 1433–1476. ( 10.1007/s10614-022-10250-w) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 506. Thelwall M. 2023. Journal and disciplinary variations in academic open peer review anonymity, outcomes, and length. J. Librariansh. Inf. Sci. 55, 299–312. ( 10.1177/09610006221079345) [DOI] [Google Scholar]
  • 507. Walsh E, Rooney M, Appleby L, Wilkinson G. 2000. Open peer review: a randomised controlled trial. Br. J. Psychiatry 176, 47–51. ( 10.1192/bjp.176.1.47) [DOI] [PubMed] [Google Scholar]
  • 508. Bornmann L, Daniel H-D. 2010. Reliability of reviewers’ ratings when using public peer review: a case study. Learn. Publ. 23, 124–131. ( 10.1087/20100207) [DOI] [Google Scholar]
  • 509. Vinther S, Nielsen OH, Rosenberg J, Keiding N, Schroeder TV. 2012. Same review quality in open versus blinded peer review in ‘Ugeskrift for Læger’. Dan. Med. J. 59, A4479. [PubMed] [Google Scholar]
  • 510. Wei C, Zhao J, Ni J, Li J. 2023. What does open peer review bring to scientific articles? Evidence from PLoS journals. Scientometrics 128, 2763–2776. ( 10.1007/s11192-023-04683-9) [DOI] [Google Scholar]
  • 511. Fox CW. 2021. Which peer reviewers voluntarily reveal their identity to authors? Insights into the consequences of open-identities peer review. Proc. R. Soc. B 288, 20211399. ( 10.1098/rspb.2021.1399) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 512. Anagnostou A, Taylor SJE, Groen D, Suleimenova D, Anokye N, Bruno R, Barbera R. 2019. Building global research capacity in public health: the case of a science gateway for physical activity lifelong modelling and simulation. In 2019 Winter Simulation Conf. (WSC), National Harbor, MD, pp. 1067–1078. National Harbor, MD: IEEE. ( 10.1109/WSC40007.2019.9004845) [DOI] [Google Scholar]
  • 513. Arza V, Fressoli M. 2017. Benefits of open science: an analytical framework illustrated with case study evidence from Argentina. In Expanding perspectives on open science: communities, cultures and diversity in concepts and practices, pp. 153–173. IOS Press. ( 10.3233/978-1-61499-769-6-153) [DOI] [Google Scholar]
  • 514. Baldassarre MT, Ernst N, Hermann B, Menzies T, Yedida R. 2023. (Re)use of research results (is rampant). Commun. ACM 66, 75–81. ( 10.1145/3554976) [DOI] [Google Scholar]
  • 515. Besançon L, Peiffer-Smadja N, Segalas C, Jiang H, Masuzzo P, Smout C, Billy E, Deforet M, Leyrat C. 2021. Open science saves lives: lessons from the COVID-19 pandemic. BMC Med. Res. Methodol. 21, 117. ( 10.1186/s12874-021-01304-y) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 516. Bush KA, Calvert ML, Kilts CD. 2022. Lessons learned: a neuroimaging research center’s transition to open and reproducible science. Front. Big Data 5, 988084. ( 10.3389/fdata.2022.988084) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 517. Coro G. 2020. Open Science and artificial intelligence supporting blue growth. Environ. Eng. Manag. J. 19, 1719–1729. ( 10.30638/eemj.2020.162) [DOI] [Google Scholar]
  • 518. Ale Ebrahim N, Bong YB. 2017. Publishing research support documents in open access platform to improve research impact. Social Sci. ( 10.20944/preprints201706.0111.v1) [DOI] [Google Scholar]
  • 519. Hofmann B. 2022. Open science knowledge production: addressing epistemological challenges and ethical implications. Publications 10, 24. ( 10.3390/publications10030024) [DOI] [Google Scholar]
  • 520. Lakomý M, Hlavová R, Machackova H. 2019. Open science and the science-society relationship. Society 56, 246–255. ( 10.1007/s12115-019-00361-w) [DOI] [Google Scholar]
  • 521. LeBel EP, Campbell L, Loving TJ. 2017. Benefits of open and high-powered research outweigh costs. J. Pers. Soc. Psychol. 113, 230–243. ( 10.1037/pspi0000049) [DOI] [PubMed] [Google Scholar]
  • 522. McKiernan EC, et al. 2016. How open science helps researchers succeed. eLife 5, e16800. ( 10.7554/eLife.16800) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 523. Nieminen P, Uribe SE. 2021. The quality of statistical reporting and data presentation in predatory dental journals was lower than in non-predatory journals. Entropy 23, 468. ( 10.3390/e23040468) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 524. Roman D, et al. 2021. An analysis of pollution Citizen Science projects from the perspective of Data Science and Open Science. DTA 55, 622–642. ( 10.1108/DTA-10-2020-0253) [DOI] [Google Scholar]
  • 525. Rzayeva N, Henriques SO, Pinfield S, Waltman L. 2022. The experiences of COVID-19 preprint authors: a survey of researchers about publishing and receiving feedback on their work during the pandemic. SocArXiv. ( 10.31235/osf.io/d96yj) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 526. Schneider J, Rosman T, Kelava A, Merk S. 2022. Do open-science badges increase trust in scientists among undergraduates, scientists, and the public? Psychol. Sci. 33, 1588–1604. ( 10.1177/09567976221097499) [DOI] [PubMed] [Google Scholar]
  • 527. Strømme CB, et al. 2021. Applying and promoting open science in ecology - surveyed drivers and challenges. Sci. Commun. Educ. ( 10.1101/2021.10.12.464125) [DOI] [Google Scholar]
  • 528. Tamminen KA, Poucher ZA. 2018. Open science in sport and exercise psychology: review of current approaches and considerations for qualitative inquiry. Psychol. Sport Exerc. 36, 17–28. ( 10.1016/j.psychsport.2017.12.010) [DOI] [Google Scholar]
  • 529. Zong Q, Huang Z, Huang J. 2023. Do open science badges work? Estimating the effects of open science badges on an article’s social media attention and research impacts. Scientometrics 128, 3627–3648. ( 10.1007/s11192-023-04720-7) [DOI] [Google Scholar]
  • 530. Tse EG, Klug DM, Todd MH. 2020. Open science approaches to COVID-19. F1000Res. 9, 1043. ( 10.12688/f1000research.26084.1) [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 531. UNESCO . 2024. Open Science. See https://www.unesco.org/en/open-science (accessed 17 July 2024).
  • 532. Visser M, van Eck NJ, Waltman L. 2021. Large-scale comparison of bibliographic data sources: Scopus, Web of Science, Dimensions, Crossref, and Microsoft Academic. Quant. Sci. Stud. 2, 20–41. ( 10.1162/qss_a_00112) [DOI] [Google Scholar]
  • 533. Harzing AW, Alakangas S. 2016. Google Scholar, Scopus and the Web of Science: a longitudinal and cross-disciplinary comparison. Scientometrics 106, 787–804. ( 10.1007/s11192-015-1798-9) [DOI] [Google Scholar]
  • 534. Asubiaro T, Onaolapo S, Mills D. 2024. Regional disparities in Web of Science and Scopus journal coverage. Scientometrics 129, 1469–1491. ( 10.1007/s11192-024-04948-x) [DOI] [Google Scholar]
  • 535. van Bellen S, Alperin JP, Larivière V. 2024. The oligopoly of academic publishers persists in exclusive database. Arxiv ( 10.48550/arXiv.2406.17893) [DOI] [Google Scholar]
  • 536. Alperin JP, Portenoy J, Demes K, Larivière V, Haustein S. 2024. An analysis of the suitability of OpenAlex for bibliometric analyses. Arxiv ( 10.48550/arXiv.2404.17663) [DOI] [Google Scholar]
  • 537. Simard MA, Basson I, Hare M, Lariviere V, Mongeon P. 2024. The open access coverage of OpenAlex, Scopus and Web of Science. Arxiv ( 10.48550/arXiv.2404.01985) [DOI] [Google Scholar]
  • 538. Klebel T, Traag V, Grypari I, Stoy L, Ross-Hellauer T. 2025. Supplementary material from: The academic impact of Open Science: a scoping review. FigShare ( 10.6084/m9.figshare.c.7700096) [DOI] [PMC free article] [PubMed]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

Datasets and additional materials supporting this article are published on Zenodo [42].

Supplementary material is available online [538].


Articles from Royal Society Open Science are provided here courtesy of The Royal Society

RESOURCES