Abstract
Sharing research code in an open access version-controlled repository offers significant benefits for both science as a whole and for individual researchers. In this article, we focus on this practice, which is fully aligned with the NIH’s Gold Standard Science (GSS) program as well as FAIR (findable, accessible, interoperable, reusable) and TRUST (transparency, responsibility, user focus, sustainability, technology) principles. Gold Standard Science supports open science by emphasizing transparency, reproducibility, and the use of best practices that enable others to verify and extend research. Pairing a research article’s cited data snapshot with a versioned, environment-specific code release, deposited in a companion code repository, ensures that, upon submission to a medical journal, readers and reviewers can directly verify results. An executable and updatable companion code repository complements, rather than replaces, established research data repositories. When code underlying medical research results is made openly available, then other scientists can inspect, run, and validate analyses. These activities enhance reproducibility, which is a core aim of GSS. Shared code also facilitates collaborative innovation by allowing researchers to extend the utility of the code to new datasets and applications. For researchers, code sharing can increase visibility, credibility, and citation impact. Demonstrating transparency through shared executable and updatable code builds trust with journal readers, peer reviewers, funders, and peers. Shared code in an open access repository signals adherence to high standards of scientific integrity and attracts opportunities for collaboration. A researcher who shares code receives recognition as a leader in reproducible, trustworthy research consistent with NIH’s GSS principles.
Keywords: code, data, diabetes, Gold Standard Science, reproducibility, repository
Introduction
In this article, we focus on a practical and important complement to data sharing that is aligned with Gold Standard Science (GSS), FAIR (findable, accessible, interoperable, and reusable), and TRUST (transparency, responsibility, user focus, sustainability, technology): pairing each journal article’s cited data snapshot with a versioned and environment-specific code release, deposited in TRUST-aligned repositories. Through this approach, readers can verify results and build reliably on prior work.
The integrity of medical science relies on a foundation of credibility, reproducibility, and transparency to drive innovation and progress in clinical care. These values, which are broadly understood as TRUST, are long recognized by the scientific and medical communities and are now further articulated in the National Institutes of Health’s (NIH) recently published Plan to Drive Gold-Standard Science. 1 In that plan, NIH outlines nine tenets of scientific rigor, beginning with reproducibility and transparency.
These recommendations build on over three decades of the open science movement, which strives to make every stage of research (from planning and data collection to analysis, publication, and reuse) transparent, accessible, and reusable by default. 2 Early anchors include arXiv (1991) for preprints, 3 the Bermuda Principles (1996-1997) for rapid genomic data release, 4 and the Budapest Open Access Initiative (2002) that defined open access and catalyzed policy adoption. 5 In the last decade, communities have standardized how to share and preserve research outputs: the FAIR principles (2016) for making data and code findable, accessible, interoperable, and reusable 6 ; the TRUST principles (2020) for trustworthy repositories 7 ; and FAIR for Research Software (FAIR4RS, 2022) 8 and software-citation guidance for code. 9 Collectively, they have laid the groundwork for credible, reusable science, making it practical to verify findings and build reliably on prior work.
Principles of Open Access Science
Open science emphasizes transparency, reproducibility, and collaboration to make scientific research results—including publications, data, code, physical samples, and methods—openly accessible to everyone. Three quality frameworks—GSS, FAIR, and TRUST—are examples of open science principles. They are complementary, and each provides guidance for different dimensions of the scientific process (Figure 1). NIH’s GSS sets expectations for how science should be done (eg, rigor, transparency, falsifiability, and bias-mitigating peer review), shaping the behaviors of investigators and institutions. Those behaviors yield the scientific record in three parts: (1) manuscripts (the narrative, claims, and insights), (2) datasets (the evidence), and (3) code and computing environment specifications (the method to regenerate results). The FAIR principles specify the qualities of these objects so others can find, access, interoperate with, and reuse them. The TRUST principles specify that these objects should live in repositories with transparent governance, versioning and durable landing pages, user focus, sustainability, and sound technology to ensure long-term citability and stewardship.
Figure 1.
Three open science frameworks (GSS, FAIR, and Trust) that are complementary to each other and guide different aspects of the scientific process.
Abbreviations: FAIR, findable, accessible, interoperable, and reusable; TRUST, transparency, responsibility, user focus, sustainability, technology.
The 2023 NIH Data Management and Sharing (DMS) Policy codifies parts of this agenda into funder requirements for data: planning, budgeting, and appropriate sharing aligned with privacy, consent, and legal/ethical limits. 2 The policy requires an approved DMS Plan at the time of application, requires updates in progress reports, and allows budgeting for data management and sharing activities. 2 It directs investigators to maximize appropriate sharing of scientific data, typically via trusted research data repositories that support persistent identifiers, versioning, preservation, and clear access models (including controlled access for human-participant data when needed). 2 While the DMS Policy’s primary focus is data, NIH notes that sharing analysis scripts, software, and code is a best practice that strengthens reproducibility; it encourages such sharing but does not mandate public code release. 3
Companion Repositories for Code
The National Institutes of Health does not specify a uniform format for deposited code or that it be made easily sharable. Building on the DMS Policy’s focus on data, code sharing can be improved with standardized practices focused on sharing and reproducibility. For code, there are two practical and complementary approaches: (1) A permanent archive of code releases for citation (eg, archiving versioned releases with unique DOIs) that creates non-mutable, citable snapshots that match a manuscript’s analyses; and (2) An updatable, executable companion code repository for collaboration and reuse where authors can maintain issues, pull requests, and documentation. In the latter setting, code should be open access, meaning that the source is publicly available under a license that permits use, modification, and redistribution. Without an explicit open-source or permissive license, others generally cannot legally modify or redistribute code. 10 Providing a clear license and archiving versioned releases ensures that readers can both reuse the software and cite the exact version used in the study. 11 An updatable and executable companion repository for code complements, rather than replaces, established research data repositories. The contents of a companion repository are listed in Table 1.
Table 1.
Contents of a Companion Repository for Executable and Updatable Code.
| • README File: Plain-English run instructions and repository map; link to the manuscript and to the data repository accession/DOI for the exact data snapshot used. • Analysis code and workflows: Notebooks and scripts (eg, R, Python, Stata/MP syntax exports) and/or workflow files that regenerate all tables, figures, and results without manual edits. • Computational environment: A machine-readable spec so others can run the code (eg, Dockerfile, environment.yml, requirements.txt), and any version/seed settings needed for deterministic output. • Provenance and transformation notes: Stepwise description of data preparation (filters, recodes, exclusions, imputation rules), with a data dictionary, naming conventions, schema and references to any taxonomies/ontologies used. • Public exemplars (optional): Small, non-sensitive sample, aggregate or clearly labeled synthetic files only to demonstrate scripts; the real study data live in a data repository. • Citation and licensing: CITATION.cff (or equivalent) with how to cite the software, a clear LICENSE (OSI-approved where possible), and instructions for submitting issues, pull requests, etc. • Versioning and releases: Tagged releases with notes; archive each release to a DOI-issuing service (eg, Zenodo) for a citable, non-mutable snapshot. • Do not include participant-level human data, secrets, tokens, or other sensitive information in the code repository. Use controlled-access data repositories as appropriate and reference their accession in the README. |
Abbreviation: OSI, Open Source Initiative.
The National Institutes of Health encourages the use of collaborative code platforms with long-term sustainability that support web-based development, sharing, integration, issue tracking, and version control. These platforms are often called code repositories or code forges; examples include GitHub, GitLab, and Bitbucket. 12
Executable Code in a Companion Repository
In this context, “executable code” means analysis code that a third party can run end-to-end in a documented computational environment. Authors should provide scripts, notebooks, and an environment specification (eg, container, environment.yml, requirements.txt, or renv.lock) so others can rerun the pipeline. If code is provided in publicly available executable, structured formats, then peer reviewers and secondary researchers can validate analyses, rerun workflows, and detect discrepancies. 13 Even with access to a complete and/or raw dataset with the code to generate the precise analytic dataset, it is possible for code to be non-executable in an environment other than that of the author because of missing supporting code, errors in code, or intentional sharing of illustrative/teaching code rather than fully runnable scripts. As much as 76% of public notebook code is technically non-executable under strict standards. 14 A 2022 study by Trisovic et al 15 of 9000 code files in an open-source data plus code repository platform for sharing, archiving, and citing research data, showed that 74% failed to complete without error in the initial execution, and 56% still failed following code cleaning. For analyses that involve proprietary tools or sensitive code, authors may provide a temporary private link or an access-controlled repository for reviewers to examine the work in an environment that mimics the proprietary context. 16 Public release of code is encouraged for reproducibility but is not required under the NIH DMS Policy. 17
Updatable Code in a Companion Repository
At scale, there are two practical pathways to support updatable, executable code alongside frozen datasets. First, some generalist/archival repositories (and certain research data repositories) let authors co-locate a versioned code release with the cited data snapshot, preserving both with persistent identifiers (eg, DOIs) to ensure their long-term availability and integrity as an official record of research artifacts. 18 Second, data repositories can partner with modern code platforms (eg, via integrations) so that when authors create a formal code release, that exact version is archived and assigned its own DOI for citation and reproducibility. 19 In both scenarios, each new code release receives a distinct DOI (with links between versions), ensuring that readers can always retrieve the precise code for a specific analysis. Figure 2 illustrates how journals, data repositories, code repositories, and collaborative code platforms interact in the scientific process.
Figure 2.
How journals, data repositories, code repositories, and collaborative code platforms interact to produce research objects and research records leading to dissemination and discovery.
Abbreviations: DOI, Digital Object Identifier; PID, Persistent Identifier.
An updatable companion repository lets code evolve through bug fixes, documentation, and new methods without changing what was cited in the publication. Authors can issue new releases with clear changelogs, while the manuscript continues to reference the original archived release/DOI used for its results. Framed this way, the repository becomes an open, version-controlled, community-maintained record. If code is submitted to a companion updatable repository, then future viewers may extend, update, or recompute the analyses, ensuring that the original work remains transparent and verifiable long after its initial appearance. 20
Benefits of Sharing Research Code in Companion Repositories
Submitting research code in a transparent companion repository with a persistent identifier such as a DOI, and citing or linking it within a journal article, enhances the study’s visibility and overall impact within the research community. 21 Code sharing also demonstrates transparency and a commitment to reproducibility, which can bolster the author’s reputation and that of their affiliated institution. 22 Practically, code sharing supports methodological rigor and error detection through verification by peers as well as serving as a foundation for new collaborations. 23
A citable companion code repository allows others to validate, reproduce, and build upon the work. These actions increase the credibility and trustworthiness of the published results and could lead to higher citation rates and greater recognition for the research, 24 although the topic has not been extensively studied. Furthermore, non-citation quantitative measures of impact, such as forks (new repositories created from existing ones) and downloads, reveal engagement and reuse of code. They can provide meaningful evidence of impact, although they are not represented by citations. 25 COARA (Coalition for Advancing Research Assessment) endorses several qualitative methods, such as peer review and expert opinion, for recognizing the impact of code that goes beyond traditional citation metrics. 26 In the same way that article- and dataset-level metrics have matured (eg, Altmetric attention scores for papers/software, 27 Make Data Count/DataCite usage, 28 and citation tracking), 29 code repositories are on a similar trajectory. As community norms around software citation, versioned DOIs for releases, and standardized usage signals continue to evolve and gain traction, companion code repositories will increasingly serve as recognized, citable scholarly outputs with impact that is measurable alongside traditional citations.
Potential Barriers
Direct costs are typically minimal for public repositories on platforms such as GitHub or GitLab, but indirect costs related to the time and effort burden of documentation, organizing code, environment capture, and maintenance can be substantial. Dedicated funding for these efforts may be needed to secure contributions and ensure sustainability. Researchers who share code may worry that competing groups may use their code without giving them adequate credit, especially if they have not yet published all intended results or completed larger projects that rely on that code. Clear license terms can help define how the code may be used.
The primary goal of most open research code repositories is to make code freely available under recognized open-source licenses for reuse, as well as to collaborate in research and software development under licenses like MIT (Massachusetts Institute of Technology), 30 GNU (GNU’s Not Unix), 31 or CC (Creative Commons). 32 Without an explicit open-source license, code is proprietary and not legally reusable. However, practical constraints can override or limit openness. Sharing code in a code repository is subject to privacy and confidentiality protection as well as (1) policies of the funding bodies and their institutions, (2) restrictions and policies set by applicable federal, Tribal, state, and local laws, and (3) existing or anticipated agreements. 17 In industry-sponsored studies, contracts often constrain code release (eg, intellectual property ownership, embargoes, or review clauses). When sharing is limited, journals can still encourage verification bundles so reviewers can evaluate methods without disclosing proprietary data. Finally, code from industry-sponsored studies is typically not shared, and research contracts often prohibit such sharing. 33
Conclusions
We expect two reinforcing trends in code sharing over the next few years. First, we anticipate increased encouragement and infrastructure for research code sharing across funders and journals. Many medical researchers—NIH-funded or not—are already archiving code snapshots alongside data deposits. Second, we anticipate that voluntary archiving of code in companion repositories by medical researchers will become more prevalent, which will facilitate greater transparency, reproducibility, and reuse of research software.
We commend NIH for its Gold Standard Science initiative and encouragement of data and code sharing. Evidence from multiple assessments shows persistent reproducibility challenges across parts of biomedicine, and surveys report that many researchers perceive a problem, although estimates vary by field and method.19-21 Sharing executable, environment-pinned code with a cited data snapshot makes work more transparent and testable, and thus more trustworthy and potentially more impactful. Over time, such contributions will shift the culture of medical science toward one that embodies the rigor envisioned in NIH’s Gold Standard plan.
Acknowledgments
The authors thank Annamarie Sucher-Jones for her expert editorial assistance.
Footnotes
Abbreviations: AI, artificial intelligence; CC, Creative Commons; CCR, Companion Code Repositories; COARA, Coalition for Advancing Research Assessment; DMS, data management and sharing; DOI, digital object identifier; FAIR, findable, accessible, interoperable, reusable; FAIR4RS, findable, accessible, interoperable, reusable for research software; GNU, GNU’s Not Unix; GSS, Gold Standard Science; MIT, Massachusetts Institute of Technology; NIH, National Institutes of Health; PID, Persistent Identifier; TRUST, transparency, responsibility, user focus, sustainability, technology.
The authors declared the following potential conflicts of interest with respect to the research, authorship, and/or publication of this article: DCK is a consultant for Afon, Atropos Health, Embecta, Glooko, Glucotrack, Lifecare, Novo Nordisk, SynchNeuro, and Thirdwayv. JE receives funding the Helmsley Charitable Trust, and is a consultant for Sanofi, Glooko, and Dexcom. JKM is a member of advisory boards of Abbott Diabetes Care, Becton-Dickinson, Biomea Fusion, DexCom, Eli Lilly, Embecta, Medtronic, myLife, Novo Nordisk A/S, Pharmasens, Roche Diabetes Care, Sanofi-Aventis, Tandem, and Viatris and received speaker honoraria from A. Menarini Diagnostics, Abbott Diabetes Care, DexCom, Eli Lilly, Medtrust, MSD, Novo Nordisk A/S, Roche Diabetes Care, Sanofi, Viatris, and Ypsomed. She is a shareholder of decide Clinical Software GmbH and elyte Diagnostics and serves as CMO of elyte Diagnostics. LH has received honoraria from Roche Diagnostics, Lifecare, Medtronic, Liom, Dexcom, OneTwenty, Perfood, Boydsense, PharmaSens, Unomedical, and Sinocare for lectures and participation in advisory boards. LH is a shareholder in Profil Institut für Stoffwechselforschung GmbH, diateam GmbH, and Science Consulting in Diabetes GmbH. DK’s institution has received research support from Abbott Diabetes Care. BKo reports receiving research support from DexCom, Inc and Tandem Diabetes Care handled by the University of Virginia and patent royalties from DexCom, Inc. handled by the University of Virginia’s Licensing and Ventures Group. BN has served as a consultant for BioSensics and Mölnlycke on studies unrelated to the scope of this manuscript. PP has consulted for Sanofi. DMM has consulted for Abbott, the Helmsley Charitable Trust, Lifescan, Sanofi, Medtronic, Provention Bio, Kriya, Biospex, and Bayer. GEU has received research support for Emory University from Bayer, Corcept, Abbott, Glucotrack, and Dexcom, and has participated in advisory boards for Dexcom, Corcept, Glucotrack, and Glycare. TYW is a consultant for Bayer, Boehringer-Ingelheim, Carl Zeiss, Genentech, Quaerite Biopharm Research Ltd, Roche, and Shanghai Henlius. He is an inventor, holds patents, and is a cofounder of start-up companies EyRiS and Visre, which have interests in, and develop digital solutions for eye diseases. AMA is an independent Board Advisor. No funding from any agency or Industry. MSDA is presently engaged in investigator-initiated research sponsored by Dexcom with discounted CGM devices. DTA has received speaker’s and/or consulting fees from Abbott, Ascensia, Lilly, Mannkind, Insulet, Novo, Sequel, and Xeris. He has received consulting/advisory fees from Lilly, Ascensia, and Mannkind. MEA has served on an advisory panel for Medtronic, Insulet, Abbott, VitalAire, Sanofi, and Dexcom, has received honoraria for speaking from Abbott, Eli Lilly, Medtronic, Novo Nordisk, Sanofi, and VitalAire; and has received research support from Medtronic and Sanofi. MAA Consultant for NNOXX, Inc. RB is a shareholder of Biomeris s.r.l. and Engenome s.r.l. EC is a Scientific Advisory Board Member/Consultant for Novo Nordisk, Eli Lilly, Arecor, Proventionbio, Portal Insulin, MannKind, Tandem, and Abbott. SLC has received research funding from i-SENS Inc., holds shares in Novo Nordisk A/S, and has received consultancy fees from Roche Diagnostics and Medicus Engineering. MAC is employed by Glooko and receives research support from Dexcom and Abbott Diabetes Care. KLC reports subscription fees and nonprofit contributions from multiple organizations. A full listing is available online at closeconcerns.com and diaTribe.org. JC is Medical Director of Teleophthalmology at Optain Health (formerly EyePACS). KD has received honoraria, travel, fees for speaking or advisory boards from Abbott Diabetes, AstraZeneca, Boehringer-Ingelheim, Eli Lilly, Menarini, Novo Nordisk, Roche, and Sanofi Diabetes. AFa is the principal investigator of research projects funded by Dexcom Inc. and co-investigator of research projects funded by Roche Diagnostics International Ltd. The funding was provided to his institution, and not directly to him personally. MALG has speaker fees: Medtronic, Sanofi, Roche, Abbott, and Novo Nordisk. TDH is a consultant for Acella and Amgen. PGJ reports receiving grants from the Leona M. and Harry B. Helmsley Charitable Trust, Breakthrough T1D, Dexcom, Eli Lilly, Oregon State University, and the Oregon Health & Science University Foundation; consultancy fees from Dexcom, Eli Lilly, Roche, 4YouAndMe, CDISC; US patents 62/352 939, 63/269 094, 62/944 287, 8 810 388, 9 480 418, 8 317 700, 61/570 382, 8 810 388, 7 976 466, and 6 558 321; and reports stock options from Pacific Diabetes Technologies, outside submitted work. BKu is head of the research institute of the diabetes academy Bad Mergentheim (FIDAM). BK has received speakers’ honoraria or consulting fees from Abbott, Bayer, Berlin Chemie, Dexcom, Embecta, Emperra, Lilly, Novo Nordisk, Roche, Sanofi, and Ypsomed. JLB is presently funded by: Helmsley Charitable Trust, Arizona Commerce Authority in Collaboration with Medtronic. AYL reports grants from Santen, personal fees from Genentech, personal fees from Johnson and Johnson, personal fees from Apellis, personal fees from Boehringer-Ingelheim, non-financial support from iCareWorld, grants from Topcon, grants from Carl Zeiss Meditec, personal fees from Gyroscope, non-financial support from Optomed, non-financial support from Heidelberg, non-financial support from Microsoft, grants from Regeneron, grants from Amazon, grants from Meta, outside the submitted work. DM has had research support from Breakthrough T1D, and the Helmsley Charitable Trust and his institution has had research support from Dexcom. DM has consulted for Abbott, Sanofi, Eli Lilly, Medtronic, Biospex, Kriya, and Enable Biosciences. NM has research support from Breakthrough T1D, and the Helmsley Charitable Trust. AAM is an employee of Alphabet and may own stock as part of the standard compensation package. SM has received speaker honoraria from Lilly UK, Menarini and Sanofi for educational presentations. SM is funded by a Wellcome Trust Career Development Award [223024/Z/21/Z] and is supported by the NIHR Imperial Biomedical Research Center. CR has received honoraria from AstraZeneca, Dexcom, Fresenius, and Vifor. DS is an advisor to Carta Healthcare. VNS’s institute has received research funding from Lilly, Enable Bioscience, Zucara Therapeutics, Cystic Fibrosis foundation, and Breakthrough T1D. VNS has received honoraria from Sanofi, Novo Nordisk, Lilly, Dexcom, Insulet, Tandem Diabetes Care, Medtronic, Sequel Med Tech, Abbott Diabetes, Roche, Biomea Fusion, and T1D Scout for advising, consulting, or speaking outside of this submitted work. MPS a cofounder and scientific advisor of Xthera, Exposomics, Filtricine, Fodsel, iollo, InVu Health, January AI, Marble Therapeutics, Mirvie, Next Thought AI, Orange Street Ventures, Personalis, Protos Biologics, Qbio, RTHM, and SensOmics. MPS is a scientific advisor of Abbratech, Applied Cognition, Enovone, Jupiter Therapeutics, M3 Helium, Mitrix, Neuvivo, Onza, Sigil Biosciences, Captify Inc, WndrHLTH, Yuvan Research, and Ovul. EKS is receiving research support (to Baltimore VA and University of Maryland) from Dexcom, Medtronic, Tandem Diabetes and Deka/Twist for the conduction of clinical trials. JS has served as an educator for Medtronic. She has received funding from Medtronic and Novo Nordisk; owns shares in Novo Nordisk and has been invited as part of the advisory board for Sanofi-Aventis; has received fees for speaking on behalf of Medtronic, Sanofi-Aventis, Rubin Medical, and Novo Nordisk. JV receives research support from Dexcom. AW conducts research with UnitedHealth Care. KW is a consultant for Astellas, a shareholder in WaShiLa Health, and has received research support from Astellas and Raxi. EW is an affiliate of the Diabetes Research Hub. RMW reports receiving research support as the site PI of sponsored clinical trials by Novo Nordisk, Lilly Diabetes and Sanofi, unrelated to this work. RMW reports serving as a consultant for Sanofi. MZ reports consulting for Dexcom, Inc. AC AD, AFle, AFS, AYD, BS, BWB, CC, CSL, DA, DL, EH, ER, FJD, GC, GD, HC, HR, IC, JCW, JF, JYC, KS, MAK, MMS, MV, NV, RA, RG, SAM, SJM, SNS, SY, TW, WAL, YMB, and YZ have nothing to disclose.
Funding: The authors received no financial support for the research, authorship, and/or publication of this article.
ORCID iDs: David C. Klonoff
https://orcid.org/0000-0001-6394-6862
Juan Espinoza
https://orcid.org/0000-0003-0513-588X
Julia K. Mader
https://orcid.org/0000-0001-7854-4233
Lutz Heinemann
https://orcid.org/0000-0003-2493-1304
Claudio Cobelli
https://orcid.org/0000-0002-0169-6682
David Kerr
https://orcid.org/0000-0003-1335-1857
Boris Kovatchev
https://orcid.org/0000-0003-0495-3901
Bijan Najafi
https://orcid.org/0000-0002-0320-8101
Priya Prahalad
https://orcid.org/0000-0002-3894-4344
Yaguang Zheng
https://orcid.org/0000-0002-8400-1398
Mandy M. Shao
https://orcid.org/0009-0004-9550-9965
Agatha F. Scheideman
https://orcid.org/0009-0008-4211-4934
Ashley Y. DuNova
https://orcid.org/0000-0002-1478-7065
Michael Kohn
https://orcid.org/0000-0001-5459-5044
Guillermo E. Umpierrez
https://orcid.org/0000-0002-3252-5026
Tien Y. Wong
https://orcid.org/0000-0002-8448-1264
Aiman Abdel Malek
https://orcid.org/0009-0000-5549-1589
Michael S. D. Agus
https://orcid.org/0000-0001-6454-6828
David T. Ahn
https://orcid.org/0000-0002-8941-8459
Rawan AlSaad
https://orcid.org/0000-0002-3235-0860
Mohammed E. Al-Sofiani
https://orcid.org/0000-0003-4420-9378
David Armstrong
https://orcid.org/0000-0003-1887-9175
Mark A. Arnold
https://orcid.org/0000-0001-9375-1863
Yong Mong Bee
https://orcid.org/0000-0002-5482-2646
B. Wayne Bequette
https://orcid.org/0000-0002-6472-1902
Riccardo Bellazzi
https://orcid.org/0000-0002-6974-9808
Eda Cengiz
https://orcid.org/0000-0001-7992-9506
J. Geoffrey Chase
https://orcid.org/0000-0001-9989-4849
Haipeng Chen
https://orcid.org/0000-0003-0572-8888
Jake Y. Chen
https://orcid.org/0000-0001-8829-7504
Simon L. Cichosz
https://orcid.org/0000-0002-3484-7571
Ali Cinar
https://orcid.org/0000-0002-1607-9943
Mark A. Clements
https://orcid.org/0009-0002-4341-1697
Kelly L. Close
https://orcid.org/0000-0001-7332-1380
Jorge Cuadros
https://orcid.org/0000-0002-7804-5386
Ivan Contreras
https://orcid.org/0000-0001-6896-818X
Gora Datta
https://orcid.org/0000-0002-7783-296X
Ketan Dhatariya
https://orcid.org/0000-0003-3619-9579
Francis J. Doyle III
https://orcid.org/0000-0002-3293-9114
Andjela Drincic
https://orcid.org/0000-0001-8365-7662
Andrea Facchinetti
https://orcid.org/0000-0001-8041-2280
G. Alexander Fleming
https://orcid.org/0000-0002-6549-0288
Joshua Foreman
https://orcid.org/0000-0002-3685-4054
Monica A. L. Gabbay
https://orcid.org/0000-0002-1300-9675
Ricardo Gutierrez-Osuna
https://orcid.org/0000-0003-2817-2085
Elizabeth Healey
https://orcid.org/0000-0002-7307-8429
Thanh D. Hoang
https://orcid.org/0000-0001-7437-5604
Peter G. Jacobs
https://orcid.org/0000-0001-9897-4783
Bernhard Kulzer
https://orcid.org/0000-0001-9120-4479
Jeff La Belle
https://orcid.org/0000-0003-4691-2702
Aaron Y. Lee
https://orcid.org/0000-0002-7452-1648
Cecilia S. Lee
https://orcid.org/0000-0003-1994-7213
Wei-An Lee
https://orcid.org/0000-0002-2928-7338
Dorian Liepmann
https://orcid.org/0000-0002-2591-4031
David Maahs
https://orcid.org/0000-0002-4602-7909
Nestoras Mathioudakis
https://orcid.org/0000-0002-0210-655X
Sultan A. Meo
https://orcid.org/0000-0001-9820-1852
Ahmed A. Metwally
https://orcid.org/0000-0002-0155-7412
Shivani Misra
https://orcid.org/0000-0003-2886-0726
Helge Raeder
https://orcid.org/0000-0001-9465-8580
Viswanathan Mohan
https://orcid.org/0000-0001-5038-6210
Sun-Joon Moon
https://orcid.org/0000-0002-6286-7254
Connie Rhee
https://orcid.org/0000-0002-9703-6469
Eun-Jung Rhee
https://orcid.org/0000-0002-6108-7758
David Scheinker
https://orcid.org/0000-0001-5885-8024
Viral N. Shah
https://orcid.org/0000-0002-3827-7107
Bin Sheng
https://orcid.org/0000-0001-8678-2784
Michael P. Snyder
https://orcid.org/0000-0003-0784-7987
Koji Sode
https://orcid.org/0000-0002-9833-2091
Elias K. Spanakis
https://orcid.org/0000-0002-9352-7172
Jannet Svensson
https://orcid.org/0000-0002-9365-0728
Nitin Vaswani
https://orcid.org/0009-0009-1586-4091
Maryam Vareth
https://orcid.org/0000-0002-7827-2570
Josep Vehi
https://orcid.org/0000-0001-6884-9789
Amisha Wallia
https://orcid.org/0000-0002-3183-4062
Kayo Waki
https://orcid.org/0000-0003-0046-2523
Tao Wang
https://orcid.org/0000-0002-0865-0062
Eric Williams Jr
https://orcid.org/0009-0002-8700-6722
Risa M. Wolf
https://orcid.org/0000-0001-7674-520X
Jenise C. Wong
https://orcid.org/0000-0003-0573-6650
Sewagegn Yeshiwas
https://orcid.org/0000-0002-8153-7672
Mihail Zilbermint
https://orcid.org/0000-0003-4047-7260
Shahid N. Shah
https://orcid.org/0000-0001-8481-6493
References
- 1. NIH publishes plan to drive gold standard science. National Institutes of Health (NIH). Published August 22, 2025. Accessed September 15, 2025. https://www.nih.gov/about-nih/nih-director/statements/nih-publishes-plan-drive-gold-standard-science.
- 2. UNESCO recommendation on open science. UNESCO Digital Library. Accessed October 8, 2025. https://unesdoc.unesco.org/ark:/48223/pf0000379949.
- 3. About arXiv. arXiv info. Accessed October 8, 2025. https://info.arxiv.org/about/index.html.
- 4. Bermuda meeting affirms principle of data release, 1997. Accessed October 8, 2025. https://www.genome.gov/25520385/online-education-kit-1997-bermuda-meeting-affirms-principle-of-data-release.
- 5. Read the declaration—Budapest open access initiative. Accessed October 8, 2025. https://www.budapestopenaccessinitiative.org/read/.
- 6. Wilkinson MD, Dumontier M, Aalbersberg IJJ, et al. The FAIR Guiding Principles for scientific data management and stewardship. Sci Data. 2016;3(1):160018. doi: 10.1038/sdata.2016.18. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7. Lin D, Crabtree J, Dillo I, et al. The TRUST principles for digital repositories. Sci Data. 2020;7(1):144. doi: 10.1038/s41597-020-0486-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8. Barker M, Chue Hong NP, Katz DS, et al. Introducing the FAIR principles for research software. Sci Data. 2022;9(1):622. doi: 10.1038/s41597-022-01710-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9. Smith AM, Katz DS, Niemeyer KE; FORCE11 Software Citation Working Group. Software citation principles. PeerJ Comput Sci. 2016;2:e86. doi: 10.7717/peerj-cs.86. [DOI] [Google Scholar]
- 10. About. Open source initiative. Accessed October 8, 2025. https://opensource.org/about/.
- 11. Sonabend R, Gruson H, Wolansky L, Kiragga A, Katz DS. FAIR-USE4OS: guidelines for creating impactful open-source software. Plos Comput Biol. 2024;20(5):e1012045. doi: 10.1371/journal.pcbi.1012045. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12. Best practices for sharing research software. Data Science at NIH. Accessed October 8, 2025. https://datascience.nih.gov/tools-and-analytics/best-practices-for-sharing-research-software-faq.
- 13. Lotterhos KE, Moore JH, Stapleton AE. Analysis validation has been neglected in the Age of Reproducibility. PLoS Biol. 2018;16(12):e3000070. doi: 10.1371/journal.pbio.3000070. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14. Are the majority of public computational notebooks pathologically non-executable? Accessed September 29, 2025. https://arxiv.org/html/2502.04184v1.
- 15. Trisovic A, Lau MK, Pasquier T, Crosas M. A large-scale study on research code quality and execution. Sci Data. 2022;9(1):60. doi: 10.1038/s41597-022-01143-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16. Gomes DGE, Pottier P, Crystal-Ornelas R, et al. Why don’t we share data and code? Perceived barriers and benefits to public archiving practices. Proc Biol Sci. 2022;289(1987):20221113. doi: 10.1098/rspb.2022.1113. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17. Writing a data management & sharing plan. Accessed September 29, 2025. https://grants.nih.gov/policy-and-compliance/policy-topics/sharing-policies/dms/writing-dms-plan.
- 18. Neumann J, Brase J. DataCite and DOI names for research data. J Comput Aided Mol Des. 2014;28(10):1035-1041. doi: 10.1007/s10822-014-9776-5. [DOI] [PubMed] [Google Scholar]
- 19. Cheifet B. Promoting reproducibility with Code Ocean. Genome Biol. 2021;22(1):65. doi: 10.1186/s13059-021-02299-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20. A living document: reincarnating the research article. Trials. 2015;16:151. Accessed September 15, 2025. https://link.springer.com/article/10.1186/s13063-015-0666-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21. Baker JW, Crowley H, Wald D, et al. Sharing data and code facilitates reproducible and impactful research. Earthq Spectra. 2024;40:2210-2218. doi: 10.1177/87552930241259397. [DOI] [Google Scholar]
- 22. Zuiderwijk A, Shinde R, Jeng W. What drives and inhibits researchers to share and use open research data? A systematic literature review to analyze factors influencing open research data adoption. PLoS ONE. 2020;15(9):e0239283. doi: 10.1371/journal.pone.0239283. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23. Ott AW, Sol-Church K, Deshpande GM, et al. Rigor, reproducibility, and transparency in shared research resources: follow-up survey and recommendations for improvements. J Biomol Tech JBT. 2022;33(3):3Fc1f5fe.fa789303. doi: 10.7171/3fc1f5fe.fa789303. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24. McKiernan EC, Bourne PE, Brown CT, et al. How open science helps researchers succeed. eLife. 2016;5:e16800. doi: 10.7554/eLife.16800. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25. Afiaz A, Ivanov AA, Chamberlin J, et al. Best practices to evaluate the impact of biomedical research software-metric collection beyond citations. Bioinforma Oxf Engl. 2024;40(8):btae469. doi: 10.1093/bioinformatics/btae469. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26. The agreement. CoARA. Accessed September 29, 2025. https://coara.org/agreement/the-agreement-full-text/.
- 27. Reveal online attention to research. Altmetric. Accessed October 8, 2025. https://www.altmetric.com/.
- 28. Make Data Count. Make data count. Published May 17, 2024. Accessed October 8, 2025. https://makedatacount.org/.
- 29. Vierkant P. DataCite—connecting research, advancing knowledge. Datacite. Accessed October 8, 2025. https://datacite.org/. [Google Scholar]
- 30. The MIT License. Open source initiative. Accessed October 8, 2025. https://opensource.org/license/mit/.
- 31. The GNU general public license v3.0-GNU project. Free Software Foundation. Accessed October 8, 2025. https://www.gnu.org/licenses/gpl-3.0.en.html.
- 32. About CC Licenses. Creative commons. Accessed October 8, 2025. https://creativecommons.org/share-your-work/cclicenses/.
- 33. Siena LM, Papamanolis L, Siebert MJ, Bellomo RK, Ioannidis JPA. Industry involvement and transparency in the most cited clinical trials, 2019-2022. JAMA Netw Open. 2023;6(11):e2343425. doi: 10.1001/jamanetworkopen.2023.43425. [DOI] [PMC free article] [PubMed] [Google Scholar]


