Skip to main content
PLOS One logoLink to PLOS One
. 2021 Jul 9;16(7):e0253538. doi: 10.1371/journal.pone.0253538

A tool for assessing alignment of biomedical data repositories with open, FAIR, citation and trustworthy principles

Fiona Murphy 1,#, Michael Bar-Sinai 2,#, Maryann E Martone 3,*
Editor: Florian Naudet4
PMCID: PMC8270168  PMID: 34242248

Abstract

Increasing attention is being paid to the operation of biomedical data repositories in light of efforts to improve how scientific data is handled and made available for the long term. Multiple groups have produced recommendations for functions that biomedical repositories should support, with many using requirements of the FAIR data principles as guidelines. However, FAIR is but one set of principles that has arisen out of the open science community. They are joined by principles governing open science, data citation and trustworthiness, all of which are important aspects for biomedical data repositories to support. Together, these define a framework for data repositories that we call OFCT: Open, FAIR, Citable and Trustworthy. Here we developed an instrument using the open source PolicyModels toolkit that attempts to operationalize key aspects of OFCT principles and piloted the instrument by evaluating eight biomedical community repositories listed by the NIDDK Information Network (dkNET.org). Repositories included both specialist repositories that focused on a particular data type or domain, in this case diabetes and metabolomics, and generalist repositories that accept all data types and domains. The goal of this work was both to obtain a sense of how much the design of current biomedical data repositories align with these principles and to augment the dkNET listing with additional information that may be important to investigators trying to choose a repository, e.g., does the repository fully support data citation? The evaluation was performed from March to November 2020 through inspection of documentation and interaction with the sites by the authors. Overall, although there was little explicit acknowledgement of any of the OFCT principles in our sample, the majority of repositories provided at least some support for their tenets.

Introduction

Best practices emerging from the open science movement emphasize that for data to be effectively shared, they are to be treated as works of scholarship that can be reliably found, accessed, reused and credited. To achieve these functions, the open science movement has recommended that researchers formally publish their data by submitting them to a data repository [1], which assumes stewardship of the data and ensures that data are made FAIR: Findable, Accessible, Interoperable and Reusable [2]. Publishing data can therefore be seen as equivalent to publishing narrative works in that the locus of responsibility for stewardship transfers from the researcher to other entities, who ensure consistent metadata, future-friendly formats, stable and reliable access, long term availability, indexing and tools for crediting the contributors. As these types of responsibilities are traditionally supported by journals and libraries, it is not surprising that many publishers and libraries are now developing platforms for hosting research data. At the same time, data are not exactly the same as narrative works. They require additional functionality to increase their utility, which explains why the most well known scientific data repositories are led by individual researchers or research communities. Scientific data repositories such as the Protein Data Bank [3] predated the internet and are viewed as important infrastructures for data harmonization, integration and computation.

Although there is general agreement that repositories should support FAIR data, there have been several other community-led initiatives to develop principles in support of open science and data sharing. The “Defining the Scholarly Commons” project at FORCE 11.org identified over 100 sets of principles issued by organizations and groups around the world that cover a range of activities involved in scholarship and how it should be conducted in the 21st century [4]. Common threads included: 1) the need to include not only narrative works, but data, code and workflows; 2) the desire to make these products “as open as possible; as closed as necessary”; 3) FAIRness, i.e., designing the products of scholarship so that they operate efficiently in a digital medium; 4) Citability, i.e., expanding our current citation systems to cover other research outputs like data, and 5) Trustworthiness, i.e., ensuring that those who assume responsibility for stewardship of scholarly output operate in the best interests of scholarship. FORCE11 conducted workshops and exercises to define what a system of scholarly communication should look like in the 21st century. One theme that emerged based on the workshops was that in the imagined scholarly commons, data repositories were central players that provided the human and technical infrastructure for publishing research data. Therefore, data repositories themselves should align with principles governing Open, FAIR, Citable and Trustworthy (OFCT) science.

The FORCE11 exercise was hypothetical but over the years, scholarly communications has been moving towards this vision. As sharing of data and code are increasingly expected and in some cases required, more attention is being paid to the infrastructures that host them and the functions they support. As documented by the Neuroscience Information Framework (NIF; neuinfo.org), on-line data repositories are diverse, each with their own custom user interfaces and few standards as to how they should be designed and the functions they should support [5]. With data repositories increasing in importance, groups have been developing recommendations on a basic set of functions that these repositories should support (e.g., [611]. Many of these focus on FAIR, e.g., FAIRshake [6] but they are by no means the only criteria. Although there is considerable agreement across all of these lists, e.g., the use of persistent identifiers, each has a slightly different focus and therefore they are not identical. Rather, they reflect priorities arising out of different contexts. In October 2020, the Coalition of Open Access Repositories (COAR) issued a set of recommendations for data repositories that, like FORCE11, built upon openness, FAIR, Data Citation and TRUST principles, reinforcing the view of the FORCE11 working group that together, OFCT provide a framework for scientific data repositories.

In the work presented here, we developed a set of evaluation criteria based on OFCT by selecting community principles that cover each of these dimensions and operationalizing them (Table 1). Our aim was to conduct a review of data repositories of relevance to a specific domain, diabetes, digestive and kidney diseases. This work was conducted in the context of the NIDDK Information Network (dkNET.org; [12]). dkNET was established in 2012 to help basic and clinical biomedical researchers find scientific resources relevant to diabetes, digestive and kidney diseases (collectively referred to here as “dk”). dkNET is taking an active role in interpreting and facilitating compliance with FAIR on behalf of this community. Part of this effort involves creating tools to help researchers select an appropriate repository for their data. dkNET maintains a curated list of recommended data repositories that cover domains relevant to dk science, extracted from the resource catalog originally developed by the Neuroscience Information Framework (Cachat et al. 2012 [5]), and cross referenced to repositories recommended by major journals and the National Library of Medicine.

Table 1. Guiding principles for OFCT used in this study to develop the assessment instrument.

Principle Description Guiding principles/charters
Open Research outputs should be as open as possible and as closed as necessary Open Definition 2.1 [13]
FAIR Research outputs should be designed to be Findable, Accessible, Interoperable and Reusable for humans and computers FAIR Data Principles [2]
Citable Research outputs should be supported by formal systems of citation for the purposes of provenance and credit. Joint Declaration of Data Citation Principles (JDDCP) [14]; Software Citation Principles
Trustworthy Data repositories should demonstrate that they are responsible for long term sustainability and access of data entrusted to them Principles of Open Infrastructures [15]; Core Trust Seal [16]

The goal of this repository listing is to make it easier for dk researchers to find an appropriate data repository in support of FAIR, open science and current and upcoming NIH mandates (NOT-OD-21-013: Final NIH Policy for D…). We therefore developed an instrument to evaluate these repositories against the OFCT principles to add information to the repository listings that might be important to a researcher in satisfying a mandate, e.g., does the repository issue a persistent identifier as per FAIR, or personal preference, e.g., does the repository support data citation. The instrument was developed using the open source PolicyModels toolkit, a software tool for developing decision trees based on specific policies. In this report, we describe the development and design of an OFCT decision tree, our criteria and strategy for evaluating data repositories for compliance and the results of their application to eight biomedical data repositories from the dkNET listing.

Materials and methods

We developed a set of 31 questions (Table 2) operationalizing the major elements of each of the principles listed in Table 1. We did not attempt to cover all aspects of the principles, but selected those that were relevant for repositories and for which clear criteria could be developed. At the time we conducted this study, the TRUST principles had not yet been issued and so are not included explicitly in our instrument, although much of what is covered in the CoreTrustSeal is relevant to the TRUST principles. The methods used in this study were not pre-registered prior to conducting the study. For a list of abbreviations used in this text, see S2 Table.

Table 2. Questions and properties used for the final interview, the table shows the question order (Q#), the text of the question posed in the interview (Question text), possible answers (Answers), whether or not the question is conditional (“C”), the dependencies of conditional questions (D) and the principle(s) the question is meant to cover (P).

Question text Answers C D P
1. Does the repository provide access to the data with minimal or no restrictions? (acc) no restrictions
minimal restrictions
significant restrictions
significant but not justified
N O
2. Are you free to reuse the data with no or minimal restrictions? (reuse) yes
somewhat
no
N O
3. Does the repository provide a clear license for reuse of the data? (lic-clr) dataset level
repository level
no license
N F
4. Are the data covered by a commons-compliant license? (lic-cc) best
good
somewhat open
closed
Y #3 O
5. Does the repository platform make it easy to work with (e.g. download/re-use) the data? (plat) yes
no
N F
6. Does the repository require or support documentation that aids in proper (re)-use of the data? (ru-doc) best
good
adequate
lacking
N F
7. Does the repository provide a search facility for the data and metadata? (sch-ui) yes
no
N F
8. Does the repository assign globally unique and persistent identifiers (PIDs)? (pid-g) yes
no
N F
9. Does the repository allow you to associate your ORCID ID with a dataset? (orcid) required
supported
not available
N C
10. Does the repository support the addition of rich metadata to promote search and reuse of data? (md-level) rich
limited
minimal
N F
11. Are the (meta)data associated with detailed provenance? (md-prov) best
good
worst
N F
12. Does the repository provide the required metadata for supporting data citation? (md-daci) full
partial
none
N C
13. Do the metadata include qualified references to other (meta)data? (md-ref) best
good
worst
N F
14. Does the repository support bidirectional linkages between related objects such that a user accessing one object would know that there is a relationship to another object? (md-lnk) best
good
unclear
worst
N F
15. Does the repository enforce or allow the use of community standards for data format or metadata? (fmt-com) yes
no
N F
16.Does the repository accept metadata that is applicable to the dkNET community disciplines? (md-dkn) best
good
worst
N F
17. Does the repository have a policy that ensures the metadata (landing page) will persist even if the data are no longer available? (md-psst) no
by evidence
by policy
N F
18. Do the metadata use vocabularies that follow FAIR principles? (md-FAIR) enforced
allowed
minimal
N F
19. Does the machine-readable landing page support data
Citation? (land-ctsp)
yes
no
N C
20. Does the repository use a recognized community standard for representing basic metadata? (md-cs) yes
no
N F
21. Can the (meta)data be accessed via a standards compliant API? (acc-api) yes
no
N F
22. Do the metadata use a formal accessible shared and broadly applicable language for knowledge representation? (md-vcb) yes
no
N F
23. Does the repository provide API-based search of the data and metadata? (sch-api) yes
no
N F
24. Is the governance of the repository transparent? (gov-tsp) best
good
worst
N T
25. Is the code that runs the data infrastructure covered under an open source license? (oss) best
good
no
N T
26. Has the repository been certified by Data Seal of Approval or the Core Trust Seal or equivalent? (tr-seal) yes
no
N T
27. Is the repository stakeholder governed? (gov-stk) full
good
weak
none
N T
28. Does the repository provide a machine-readable landing page? (land-api) yes
no
Y #29 F
29. Does the PID or other dataset identifier resolve to a landing page that describes the data? (land-pg) yes
no
Y #8 C
30. Does the metadata clearly and explicitly include identifiers of the data it describes? (md-pid) all
some
none
Y #8, #29 F
31. Does the repository assign, or the contributor provides, a locally unique identifier to the dataset or the data contribution? (pid-l) yes
no
Y F

A “Y” in the conditional column indicates that whether or not the question is shown to the interviewer depends upon a prior answer. The questions that elicit the conditional questions are shown in the Dependencies column. Each question is assigned a unique ID which is shown in parentheses after each question. Y = Yes, N = No, O = Open, F = FAIR, C = Citable, T = Trustworthy. The full instrument, which also includes explanatory text and appropriate links, is available at [17].

The instrument was used to evaluate eight repositories listed by dkNET (RRID:SCR_001606) provided in Table 3. We selected these repositories to represent different data types or different research foci. Excluded from consideration were repositories that required an approved account to access the data, e.g., the NIDDK Central Repositories. We also did not consider knowledge bases, defined here as a database that extracts observations from the literature or as a result of analyses of primary data, but not the primary data themselves. We did, however, include AMP-T2D which presents statistical summaries of clinical data although it does not host the primary data. We also excluded some of the most well known of the biomedical databases, e.g., the Protein Data Bank and GEO, in order to focus on dk-relevant but perhaps lesser known repositories. We included two generalist repositories, Zenodo and NIH-Figshare, as the generalist repositories are likely to play a significant role for diverse domains like dk, where specialist repositories for all data types and research foci may not be available. NIH-Figshare at the time of evaluation was made available as a pilot by the National Library of Medicine for data deposition by NIH-supported researchers. Many of these repositories are complex websites with multiple tools, services and databases, and so for each of the repositories, we indicate in Table 3 which specific component(s) were reviewed.

Table 3. List of repositories evaluated in this study.

Repository Description Section URL
Accelerating Medicine Partnership Type 2 Diabetes (RRID:SCR_003743) Portal and database of DNA sequence, functional and epigenomic information, and clinical data from studies on type 2 diabetes and analytic tools to analyze these data. T2DKP Datasets under “Data” tab http://www.kp4cd.org/datasets/t2d
Cell Image Library (RRID:SCR_003510) Freely accessible, public repository of vetted and annotated microscopic images, videos, and animations of cells from a variety of organisms, showcasing cell architecture, intracellular functionalities, and both normal and abnormal processes. Main site representing single image and datasets http://www.cellimagelibrary.org
Flow Repository (RRID:SCR_013779) A database of flow cytometry experiments where users can query and download data collected and annotated according to the MIFlowCyt data standard. Public site http://flowrepository.org
Image Data Resource (IDR) (RRID:SCR_017421) Public repository of reference image datasets from published scientific studies. IDR enables access, search and analysis of these highly annotated datasets. Cell-IDR http://idr.openmicroscopy.org/cell/
Mass Spectrometry Interactive Virtual Environment (MassIVE) (RRID:SCR_013665) MassIVE is a community resource developed by the NIH-funded Center for Computational Mass Spectrometry to promote the global, free exchange of mass spectrometry data. Access public datasets https://massive.ucsd.edu/ProteoSAFe/datasets.jsp#%7B%22query%22%3A%7B%7D%2C%22table_sort_history%22%3A%22createdMillis_dsc%22%7D
Metabolomics Workbench (RRID:SCR_013794) Repository for metabolomics data and metadata which provides analysis tools and access to various resources. NIH grantees may upload data and general users can search metabolomics database. Data Repository https://www.metabolomicsworkbench.org/data
NIH Figshare (RRID:SCR_017580) Repository to make datasets resulting from NIH funded research more accessible, citable, shareable, and discoverable. Public portal and password protected space https://nih.figshare.com/
Zenodo (RRID:SCR_004129) Repository for all research outputs from across all fields of science in any file format. Public site and data submission forms https://zenodo.org/

The specific section of the repository evaluated is indicated in the Section column, along with the corresponding URL.

Developing and testing the instrument

To design the instrument, we adapted the decision tree originally designed by the FORCE11 Scholarly Commons project for evaluating repositories on OFCT principles [4]. We benchmarked the instrument against a range of surveys and other tools then available for similar uses. The process involved reading the background rationale and information about these other instruments, and assessing them for their objectives, projected user profiles, scope and outputs. Factors such as: is this publisher/publication facing; is it community-led or commercial; if referring to, say, repositories, is it intended to inform potential users about the repositories, or to support repository managers themselves? The materials in question included the repository finder tool developed by DataCite for the Enabling FAIR Data project; the Scientific Data journal repository questionnaire; the FAIRsFAIR data assessment tool; and the Core Trustworthy Data Requirements. From this exercise, we determined that the answers to the questions were sometimes difficult to ascertain as clear criteria for evaluation had not been specified. Some areas were clearly missing while some of the questions were duplicative. We thus modified the questionnaire by removing duplicates, adding additional questions, developing specific evaluation criteria and adding tips as to where to look for certain types of information. Definitions and links to supporting materials were also provided for each question where appropriate. The complete version of the questionnaire used here, which includes the criteria used for each question, was deposited in Zenodo [17].

The final questionnaire comprised 31 questions, listed in order in Table 2. Some of the questions are conditional, that is, their presentation is dependent upon a prior answer. For example, if an interviewer answered “No” to question #3 “Does the repository provide a clear license for reuse of the data?” then question #4 “Are the data covered by a commons-compliant [i.e., open] license?” is not presented. Thus, the total number of questions asked may differ across repositories. Each of the questions was given a unique ID which is supplied after the question in Table 2 can be found organized alphabetically in S1 Table.

Table 2 also lists the principle set it covers (OFCT). Although the questions were originally grouped by principle, when testing the questionnaire we noted that many questions were logically related to one another, e.g., under the FAIR section we asked about licenses, while under the open section we asked about open licenses. Therefore, we reordered the questions to reflect better the actual workflow a reviewer might implement by grouping together related questions.

Encoding the instrument in policy models

The questionnaire was encoded using the PolicyModels software (RRID:SCR_019084). PolicyModels uses formal modeling to help humans interactively assess artifacts or situations against a set of rules. A PolicyModels model consists of an n-dimensional space (called "policy space"), and a decision graph that guides users through that space using questions. Each of the policy space’s dimensions describes a single assessed aspect using ordinal values. Thus, every location in a policy space describes a single, discrete situation with regards to the modeled guidelines [18].

The dimensions of the policy space defined for this work formally capture the assessment aspects implied by OFCT. It contains 45 dimensions that are assessed by the 31 questions shown in Table 2, such as Documentation Level (lacking/adequate/good/full), Metadata Provenance (unclear/adequate/full), and overall ratings of each criteria, e.g., FAIR Accessibility level (none/partial/full) and so forth. The full policy space for this instrument is shown in Fig 1, and is also available via the questionnaire landing page and in [17]. Some dimensions are assigned based on the answer to a single question, while some are calculated based on values on other dimensions. Using an interactive interview guided by our model’s decision graph, we were able to find the location of each of the evaluated repositories in the space we defined. To visualize this space, we developed an interactive viewer available at http://mbarsinai.com/viz/dknet. This allowed us to formally compare repositories across multiple dimensions, and to collect overall statistics.

Fig 1. Policy space defined by the PolicyModels software illustrating the relationship of the dimensions assessed to the properties (rectangles) and flags (blue ovals).

Fig 1

A full resolution view is available in Martone et al., (2020) [17].

The main features of the tool are shown in Fig 2. The online version allows interviewers to annotate the response to each question with notes (Fig 2B) and export the outcomes of the evaluation (Fig 2G). Currently, the results can only be exported as.json or.xml. However, to save a human readable version.pdf version of the questionnaire results, users can use the browser’s print function to save the interview summary page as a PDF.

Fig 2. Main features of policy models questionnaire.

Fig 2

The panel on the right provides an example of the question panel and the left panel shows the results of a survey after it is completed. A) each question is presented in sequence and can be accompanied by explanatory material and links to additional material; B) The interviewer may add notes to each question; C) Interviewer records an answer by selecting the appropriate response; D) The answer feed may be displayed and used to track progress and also to allow an interviewer to revisit a question to change an answer; E) Policy models tallies the answers and assigns tags assessing compliance with OFCT; F) Final tags assigned for each category; G) The results may be downloaded as json or xml.

Scoring

Five of the sites were reviewed independently by FM and MM between March and May 2020 and three in December 2020. Results were compared and a final score assigned for each question. The reviewers made a good faith effort to find information on the site to provide an accurate answer for each question. The evaluation included checking of information on the repository site, examination of the metadata provided by the site, investigations into the PID system, including what information was exported to DataCite if DOIs were used, inspection of the underlying platform code, documentation and tutorials. For some of the repositories, we created accounts in order to evaluate practices and further documentation for uploading data, e.g, can one associate an ORCID with a dataset, although in no case did we actually upload any data. To check machine-readability for data citation, we attempted to import the citation metadata into an on-line reference manager to see if they were recognized. We did not attempt to read papers that described the site. If we could not find explicit evidence for a criterion, we assumed that it was not present. Therefore, a “No” answer to a question such as “Does the repository provide an API” could mean either that the repository has a statement saying that it will not provide an API, or that we could find no evidence that it did. After the study was completed, we sent a copy of the assessments to the owner or contact for each of the 8 repositories asking them to review our results for accuracy. We received acknowledgements from 7 and responses from 3 of them. We provide both the original and the corrected versions in [19].

After a model-based interview regarding a given repository is completed, PolicyModels displays a coded evaluation of the repository. Formally, PolicyModels locates the coordinate that best describes that repository in our model’s policy space. While mathematically all dimensions are equally important, PolicyModels allows its users to organize them hierarchically, to make working with them more comfortable.

Our proposed model’s policy space is organized as follows. High-level property descriptions, such as openness and citability levels, are each represented in a dimension of their own. These dimensions have three levels, corresponding to “not at all”, “somewhat”, and “fully”. For example, the Reusable dimension contains the levels “not reusable”, “partially reusable”, and “fully reusable”.

The high-level properties are a summary of lower-level assertions, each describing a narrow aspect of these high-level properties. These assertions can be binary or detailed. For example, “open format”, one of the openness sub-aspects, is “yes” for repositories that use an open format and "no" for the others. On the other hand, “Study Linkage”, an interoperability sub-aspect, can be “none”, “free text”, “textual metadata”, or “machine readable metadata”.

Each interview starts by pessimistically setting all high-level dimensions to their lowest possible value: "not at all". During the interview, while lower-level aspect results are collected, high-level repository coordinates may be advanced to their corresponding "somewhat" levels. After the last question, if the evaluated repository achieved an acceptable for all sub-aspects of a certain higher property, that property is advanced to its "fully" level.

As a concrete example, consider the "Findable" dimension. At the interview’s start, we set it to "not findable". During the interview, our model collects results about persistent identifiers used by the repository (none/internal/external), the grade of the metadata it uses (minimal/limited/rich), whether ids are stored in the metadata (none/partial/all), and whether the repository offers an internal search feature (yes/no). If a repository achieves the lowest values in all these dimensions, it maintains its "not findable" score. If it achieves at least one non-lowest value, it is advanced to "partially findable". After the interview is completed, if it achieved the highest value in each of these dimensions, it is advanced to "fully findable".

Results

Overall impressions

Fig 3 provides the average score, scaled to a 10 point scale for each question, with 1 = lowest score and 10 = best score. Scaling was performed because each question can have a different number of answers, making it difficult to compare repositories across questions. A full list of question IDs is available in Supplemental Material S1. On over half of the questions (18/31), repositories scored on average higher than the midpoint, indicating at least some alignment. On under half they were at or below the midline (13/31), indicating poor alignment or no information available, with all repositories receiving the lowest score on 3 of the questions. We note, however, that of the 3 lowest scores, only tr-seal (“Has the repository been certified by Data Seal of Approval or the Core Trust Seal or equivalent”) was assessed across all repositories. The other two were conditional questions with only 1 assessment for each.

Fig 3. Average scaled score for each question across all repositories.

Fig 3

Questions are ordered on the Y axis according to highest average score (top) to lowest score (bottom). The data underlying the figure is available in Bar-Sinai et al., 2020 in the summary-transcript.tsv file. The average scaled score was calculated per question and then the results were sorted from highest to lowest. Question ID key is found in S1 Table.

The answers to these questions are used to assign OFCT properties and flags in the Policy Space. Flags represent a binary rating; if the flag is assigned, then the repository meets that criterion, e.g., openFormat means that the repository makes data available in an open format. Properties are rated on an ordinal scale generally that indicates full, partial or minimal compliance. The properties and the flags assigned by the PolicyModels software and their meaning are provided in Table 4.

Table 4. Ratings for each OFCT property and flag.

Properties and flags Counts Description
Open
Restrictions none:6 minimal:2 significant:0 Level of restrictions imposed by the repository in order to access datasets.
CCLicenseCompliance nonCompliant:0 none:3 adequate:1 good:3 full:1 Commons-compliance level of the repository license
openFormat no:4 yes:4 Is the data available in an open (non-proprietary) format?
platformSupportsDataWork no:1 yes:7 Does the repository platform make it easy to work with (e.g. download/re-use) the data?
ccLicenseOK no:3 yes:5 Are the data covered by a commons-compliant license? (any answer except "closed" is considered a "yes")
restrictionsNotJustified no:8 yes:0 Does the repository impose "significant but not justified restrictions" on accessing the data?
FAIR:Findable
PersistentIdentifier none:1 internalPID:0 externalPID:7 Scope of persistent identifier assigned to the data, if any
IdInMetadata none:1 partial:2 all:4 Does the metadata clearly and explicitly include the identifier of the data it describes?}
MetadataGrade minimal:0 limited:5 rich:3 Level of additional metadata that can be added to promote search and reuse of data
FindableFlags/internalSearchOK no:0 yes:8 Does the repository provide a search facility for the data and metadata?
FAIR:Accessible
humanAccessible no:1 yes:7 Does the repository provide access to the data with minimal or no restrictions?
machineAccessible no:2 yes:6 Can the data be accessed by a computer? Note that this includes access both via UI and API, as web-based UI is by definition machine-accessible.
persistentMetadata no:7 yes:1 Does the repository have a policy that ensures the metadata (landing page) will persist even if the data are no longer available, either by policy or example?
licenseOK no:3 yes:5 Does the repository provide a clear license for reuse of the data? (any answer except "no license")
stdApi no:1 yes:7 Can the (meta)data be accessed via a standards compliant API?
MetadataPersistence no:7 byEvidence:0 byStatedPolicy:1 Does the repository have a policy that ensures the metadata (landing page) will persist even if the data are no longer available?
FAIR:Inteoperable
MetadataFAIRness minimal:3 allowed:3 enforced:2 Do the metadata use vocabularies that follow FAIR principles?
StudyLinkage none:0 freeText:6 textualMetadata:1 machineReadableMetadata:1 Type of linkage between the published dataset and the paper that accompanied it
formalMetadataVocabularyOK no:5 yes:3 Do the metadata use a formal, accessible, shared and broadly applicable language for knowledge representation?
fairMetadataOK no:3 yes:5 Do the metadata use vocabularies that follow FAIR principles? (any answer except "minimal")
qualifiedMetadataReferencesOK no:3 yes:5 Do the metadata include qualified references to other (meta)data? (any answer except "worst")
studyLinkageOK no:6 yes:2 Linkage between the published dataset and the paper that accompanied it is "good" or "best".
MetadataReferenceQuality freeText:3 informal:3 formal:2 Type of qualified references to other (meta)data, included in the (meta)data stored in the repository
FAIR:Reusable
DocumentationLevel lacking:4 adequate:3 good:1 full:0 Level of support offered by the repository for documentation that aids in proper (re)-use of the data
MetadataProvenance unclear:0 adequate:5 full:3 Are the (meta)data associated with detailed provenance?
documentationOK no:4 yes:4 Does the repository require or support documentation that aids in proper (re)-use of the data? (any answer except "worst")
dkNetMetadataOK no:5 yes:3 Does the repository accept metadata that is applicable to the dkNET community disciplines? (any answer except "worst")
communityStandard no:3 yes:5 Does the repository enforce or allow the use of community standards for data format or metadata?
generalMetadata no:4 yes:4 Does the repository use a recognized community standard for representing basic metadata?
metadataProvenanceOK no:0 yes:8 Are the (meta)data associated with detailed provenance? (any answer except "worst")
DkNetMetadataLevel none:5 dataset:1 datasetAndSubject:2 Does the repository accept metadata that is applicable to the dkNET community disciplines?
ReuseLicense none:3 repositoryLevel:0 datasetLevel:5 Level at which the repository provides a clear license for reuse of the data
Citable
MachineReadableLandingPage none:1 exists:5 supportsDataCitation:2 Level of machine-readability of the dataset landing page (if any) provided by the repository
CitationMetadataLevel none:2 partial:3 full:3 Does the repository provide the required metadata for supporting data citation?
OrcidAssociation none:6 supported:2 required:0 Does the repository allow the authors to associate their ORCID ID with a dataset?
Trustworthy
GovernanceTransparency opaque:2 partial:5 full:1 Transparency level of the repository governance
SourceOpen no:5 partially:0 yes:3 Is the code that runs the data infrastructure covered under an open source license?
StakeholderGovernance none:0 weak:2 good:2 full:2 Level of control stakeholders have in the repository’s governance

Overall ratings on each dimension measuring OFCT. Properties (Props) and Flags are assigned by the PolicyModels software based on the answers given. Properties are assigned at multiple levels depending on level of compliance, whereas all flags are binary and are only assigned if the repository meets the criteria. Repository Count = number of repositories with each rating; QID: ID of question that assigns the property/flag; Short explanation: Meaning of the property or flag.

Our instrument calculates an overall rating per OFCT dimension, as shown in Fig 4. For a repository to be rated fully compliant, it would have to receive an acceptable score for all dimensions that evaluate that principle; conversely to be rated non-compliant would require an unacceptable score on all dimensions. This calculation is performed using PolicyModels, and is based on the range of acceptable and unacceptable values in various dimensions of the instrument’s policy space. Note that we do not provide scores for individual repositories in this paper, as our intent is not to grade them. However, the completed questionnaires for the individual repositories are available in [20].

Fig 4. Overall ratings of repositories on OFCT criteria.

Fig 4

The Y axis shows the individual dimensions and the X axis shows the number of repositories assigned each rating out of the 8 assessed. Red = Not compliant; Gold = Partially compliant; Blue = Fully compliant.

We sent the completed versions to the repositories for their inspection. We received acknowledgements from 7 of the repositories and responses from 3 of them. Two of the repositories each noted two errors in our evaluation. For one repository, one error was a misinterpretation on our part and one was made because the documentation of that function was not clearly visible on the site. For the other, the errors were on our part. For the third repository, AMP-T2D, we went over our responses with the repository representative in person. This site was difficult to review as it is not a repository for primary data but rather presents harmonized summary results from clinical studies. We agreed that there were 4 errors on our part, and two where there was some disagreement as to whether an aggregator such as AMP-T2D was responsible for such a function. One question was also scored negatively because of lack of relevant documentation on the site. These results suggest that overall, we were reasonably accurate in our evaluation of the repositories. In the following figures, we use the corrected results if we made the error but not if it was a difference of opinion or if the necessary documentation was not present.

As seen in Fig 4, at least one repository scored as fully compliant in each of the Open, Findability, Accessibility, Reusability and Citability dimensions. Conversely, three repositories received the lowest rating for Findability and one for Citability. No single repository was equally good—or bad—on all dimensions, that is, the same repositories did not receive either all of the highest or lowest scores. The most flags assigned to a single repository was 15 while the fewest was 5.

Open dimension

Seven repositories were scored as “Partially Open” and one as fully open (Fig 4) with details of the policy space for open criteria shown in Table 4. As biomedical repositories can deal with sensitive information that cannot be openly shared, they should adhere to the “As open as possible; as closed as necessary” principle. However, none of the repositories we evaluated had sensitive data and all were judged to make their data available with minimal to no restrictions, i.e., no approval process for accessing the data. We also evaluated repositories’ policies against the open definition: “Knowledge is open if anyone is free to access, use, modify, and share it—subject, at most, to measures that preserve provenance and openness.” Thus, data have to be available to anyone, including commercial entities, and users must be free to share them with others. We thus examined the licenses against those rated by the Open Knowledge Foundation as adhering to their definition (https://opendefinition.org/licenses/). One repository was considered fully compliant, 4 were rated as “good” with respect to open licenses, 3 had no licenses (Table 4; CCLicenseCompliance). The four rated as “good” did not receive the best score due to allowing the user to select from a range of licenses, some of which restricted commercial use.

FAIR dimension

Our questions on FAIR evaluated both compliance with specific FAIR criteria, e.g., the presence of a persistent identifier or with practices that support FAIR, e.g., providing landing pages and providing adequate documentation to promote reuse. Evaluating a repository against some principles also required that we define concepts such as “rich metadata” (FAIR principle F2) and a “plurality of relevant attributes” (FAIR principles R1).

Rich metadata were considered to comprise basic descriptive metadata, i.e., dataset title, description, authors but also metadata specific to biomedical data, e.g., organism, disease conditions studied and techniques employed (Q:md-level). “A plurality of relevant attributes” was defined in question md-dkn as providing sufficient metadata to understand the necessary context required to interpret a dkNET relevant biomedical dataset. Such metadata includes subject level attributes, e.g., ages, sex and weight along with detailed experimental protocols. Fig 5 positions each repository in the metadata policy space and shows that only one repository fully satisfied both metadata requirements.

Fig 5. Assessment of the degree of descriptive metadata (X) vs relevant biomedical metadata (dkNET metadata level) (Y).

Fig 5

The Metadata Grade assesses whether the repository complies with the Findable principle for Rich Metadata, while the dkNET metadata measures the degree to which the repository supports the Reusable principle requiring “a plurality of relevant attributes”. Relevance here was assessed with respect to dkNET. Only one repository received the highest score for both categories.

Fig 4 shows that the majority of repositories were either partially or fully compliant with all the Findability and Accessibility dimensions. Two repositories achieved the highest rating in Findability. Seven out of the 8 repositories supported external PIDs, either DOIs or accession numbers registered to identifiers.org. One repository issued no identifiers. Only 1 repository was considered fully accessible because only 1 repository had a clear persistence policy (Q:md-psst). Both the Data Citation and FAIR principles state that metadata should persist even if the accompanying data are removed. We considered either an explicit policy or clear evidence of such a practice as acceptable, e.g., a dataset that had been withdrawn but whose metadata remained.

Overall scores were lowest for the interoperability dimensions, with 3 repositories being judged non-interoperable. Only one of the repositories achieved the StudyLinkage flag which indicated that they had fully qualified references to other data, in other words, that the relationship between a metadata attribute and a value was both machine readable and informative. We measured this property by looking at how repositories handled supporting publications in their metadata, e.g., did they specify the exact relationship between the publication and the dataset? To measure this, we looked at the web page markup (“view source”) and also checked records in DataCite.

Two repositories achieved the highest score for reusability, while the remainder were considered partially reusable. Five repositories were judged as having inadequate metadata for providing experimental context, 4 as having inadequate user documentation, while 3 did not provide a clear license.

Citable dimension

Data citation criteria included the availability of full citation metadata and machine-readable citation metadata according to the JDDCP ( [10, 21, 22] ). We also evaluated the use of ORCIDs, as linking ORCIDs to datasets facilitates assigning credit to authors. As shown in Fig 6, only two repositories supported ORCID and provided full citation metadata. Consequently, 2 repositories were judged to fully support data citation, while the remainder were judged as partially (N = 5) or not supporting (N = 1) data citation. Many of the repositories had a citation policy, but most of these policies requested citation of a paper describing the repository and contributor of the data acknowledged rather than creating full citations of a particular dataset. Two were judged not to have sufficient metadata to support full citation, e.g., listing only the submitter and not other authors [see question med-daci].

Fig 6. Repositories plotted against two dimensions of data citation.

Fig 6

The Y axis shows support for citation metadata and the X axis for ORCID support. Two repositories support ORCID and provide full citation metadata. Two repositories have no support for data citation and the others have partial support.

Trustworthy dimension

Trustworthiness was largely assessed against the Principles of Open Infrastructures (Bilder et al., 2015) and the CoreTrustSeal criteria. As noted in the introduction, the TRUST principles had not been issued when we developed our instrument and so are not explicitly included. The questionnaire originally probed the different certification criteria recommended by the CoreTrustSeal but we dropped this approach in favor of a single binary question on whether or not the repository was certified by CoreTrustSeal or equivalent. If a repository was certified, it would automatically be rated fully trustworthy. However, none of the eight repositories provided evidence of such a certification.

In accordance with the Principles of Open Infrastructures, we measured the degree to which the governance of the repository was transparent and documented and whether the repository was stakeholder governed. Only one repository received the highest rating for each of these, while 1 had virtually no information on how the repository is governed, e.g., who is the owner of the repository, or how decisions are made. Although 6 of the repositories were researcher-led, it wasn’t always clear how the stakeholder community was involved in oversight, e.g., a scientific advisory board. Finally, the Principles of Open Infrastructures recommends that the software underlying the repository be open source, so that if the repository ceases to be responsive to the community, it could be forked. Two of the repositories provided links to a GitHub repository with a clear open source license.

Discussion

As part of dkNET.org’s efforts to promote data sharing and open science, we undertook an evaluation of current biomedical repositories. Two of the repositories evaluated in this initial round had a dk focus, the Metabolomics Workbench and the AMP-T2D, while the others were either focused on a particular data type, e.g., proteomic, images or flow cytometry, or were generalist repositories that accepted all data types. This study focused primarily on the development and testing of the instrument and the criteria and approach we used for evaluating data repositories. Our intention is to apply it to the entire list of repositories provided by dkNET.

The goal of developing this evaluation instrument was two fold. First, dkNET is developing an on-line catalog of specialist and generalist data repositories suitable for dk researchers to deposit data. The repository catalog is part of a larger effort to help the dk community with FAIR data practices and open science. As one of the most important steps researchers can take towards both open and FAIR is to deposit data into a trustworthy repository that supports both, we wanted to evaluate the extent to which our current ecosystem of biomedical data repositories support practices that are consistent with these principles. Similarly, as researchers may want to receive credit for sharing their data, we wanted to determine the degree to which biomedical repositories supported the requirements for data citation laid out in the data citation principles.

Second, as more attention is now being paid in biomedicine to the services that biomedical repositories should support, including the recently released set of criteria by the National Institutes of Health [16], community organizations like dkNET can serve as a resource both for those maintaining data repositories or developing new ones by helping to define a consistent set of criteria for how dk repositories should operate.

The instrument itself is fairly generic in that many questions would apply to any type of scientific repository. However, we did interpret questions that dealt with community specifics with respect to the needs of biomedicine. For example, we interpreted “rich metadata” as including critical and basic biomedical information such as the type of subject, type of technique and disease condition, that are often left out of general metadata recommendations, e.g., the Data Cite schema, Dublin Core or schema.org. For reusability, we probed for deeper biomedical data, e.g., individual subject attributes such as age and sex, that will promote reuse. As we were specifically interested in open sharing of data, the questionnaire did not contain any questions relating to sensitive data that could not be openly shared, e.g., human subjects information. However, the instrument itself is applicable to all biomedical repositories that have data freely available to the public. We have published the instrument as a FAIR object under a CC-By license. It is built on open-source software and can be adapted and extended by others.

Overall, as shown in Fig 4, the biomedical repositories we evaluated were considered partially or fully compliant with the OFCT dimensions even though only two of the repositories gave any indication that their functions or design were informed by any of the OFCT principles, in this case specifically mentioning FAIR. The lack of explicit engagement with these principles is not surprising given that most of the repositories were established before these principles came into existence. For this reason, we gave credit for what we called “OFCT potential” rather than strict adherence to a given practice. We used a sliding scale for many questions that would assign partial credit. For example, if the repository did have landing pages at stable URLs we gave them some credit, even if the identifier was not strictly a PID. Such IDs could easily be turned into PIDs by registering them with a resolving service such as Identifiers.org or N2T.org [17]. A good-faith effort was made to try to answer the questions accurately, although reviewing biomedical repositories is challenging. To evaluate specific dimensions required significant engagement with the site, even in some cases requiring us to establish accounts to see what metadata was gathered at time of upload. Discovery of these types of routes, e.g., that ORCIDs are only referenced when you establish an account, required us often to go back and re-evaluate the other repositories using this same method. A follow up with repository managers presenting the result of our review indicated fairly good agreement for most of our evaluations. We acknowledged errors where we missed information that was present on the site. In other cases, we were informed that the repository did support a particular function although no documentation was available through the site. These findings suggest that reviews by independent parties like dkNET can serve a useful function by identifying areas where documentation is missing or unclear or pointing out services that a repository might want to implement, but also point to the importance of verifying any information with repository owners before rendering an evaluation.

In addition to finding relevant information, consistent scoring of the repository was also a challenge. Principles are designed to be aspirational and to provide enough flexibility that they will be applicable across multiple domains. There is therefore a certain amount of subjectivity in their evaluation particularly in the absence of validated, established standards. For example, one of the repositories issued persistent identifiers at the project level but not to the data coming from the individual studies. In another website not included in the final evaluation sample, DOIs were available upon request. Are these considered compliant? One could argue both ways.

As described in the methods, we did not attempt to cover all aspects of the underlying principles, we selected those for which we could develop reasonable evaluation criteria. For example, one very important issue covered by CoreTrustSeal, the newly published TRUST principles [23] and Principles of Open Infrastructure [15] is long term sustainability. Although critical, we do not think that an external party such as ourselves is in a position to comment on the long term sustainability plan for a given repository. Long term sustainability for biomedical infrastructure is a known problem and one for which there are currently few concrete answers as support of most researcher-led infrastructures is in the form of time-limited grants. Our instrument is relevant to this issue, however, as OFCT practices such as FAIR, open formats, open software and good governance practices make repositories more likely to be sustainable as they facilitate transfer of data across organizations.

Since the issuance of the FAIR data principles, several initiatives have invested in the development of tools that are designed to assess the level of data FAIRness, including those that are meant to evaluate on-line data repositories. Some funders such as the EU and NIH are developing policies around FAIR data which may include a more formal assessment of FAIRness. Such tools include FAIRmetrics (https://github.com/FAIRMetrics/Metrics), FAIR Maturity Indicators [7]), FAIRshake [6] and the FORCE11/Research Data Alliance evaluation criteria [9]. Other efforts, however, extend beyond FAIR and have published lists of functions that should be supported by scientific data repositories in general, e.g., COAR [8], criteria that matter [24] or specifically by biomedical repositories, e.g., NIH [11], Elixir [25]. Our instrument aligns most closely in spirit with COAR, released in October 2020 after this study was completed, as COAR also references the Data Citation Principles, Core Trust Seal and TRUST principles in addition to FAIR.

The FAIRshake toolkits have developed some fully automated or semi-automated approaches for determining FAIRness, e.g., for checking resolvable identifiers. Such tools will make certain of our evaluations easier, e.g., to determine whether machine-readable metadata were available. However, as we show here, some aspects of FAIR require interpretation, e.g., “a plurality of relevant attributes”, making it difficult to employ fully automated approaches. In the case of “rich metadata” and “plurality of relevant attributes”, dkNET is evaluating these based on our criteria, that is, the type of metadata we think are critical for biomedical studies in our domain. These may not be universal.

While evaluation tools can be powerful, there are downsides to rushing into too rigid an interpretation of any of the principles underlying OFCT, particularly FAIR as that is currently the target of many of these efforts. First, communities are still coming together around good data practices and standards for their constituents based on what can be reasonably implemented at this time. Second, recommendations are being issued by specific stakeholders in the scientific data ecosystem and may not be relevant to other contexts. As noted in the introduction, data repositories have to straddle two worlds: providing traditional publishing/library functions to ensure findability and stability, while at the same fulfilling more traditional roles of scientific infrastructures for harmonizing and reusing data. Thus, evaluating a repository from a journal’s or funder’s perspective may not be the same as from a researcher’s perspective.

Biomedicine also has a very diverse set of repositories, including knowledge bases and sites that are hard to classify as one type or another. One of our sites, AMP-T2D, fell into that category. It is billed as a knowledge portal but provides access to summary statistics from GWAS studies. So although it hosts data, it also has the characteristics of an aggregator site and also a knowledge base. Many of our criteria were hard to interpret in this context and it was a judgement call as to whether all were relevant. However, sites such as AMP = T2D point to the complexity of on-line biomedical data resources. Such results indicate that it is still perhaps early days for understanding what constitutes best practices for a data repository across all disciplines. Our understanding of such practices may evolve over time as data sharing becomes more mainstream. As already noted, for example, early efforts in data sharing necessarily focused on deposition of data. Less attention, perhaps, was paid to what it takes for the effective reuse of the data. While the FAIR principles emphasize machine-readable attributes for achieving reusability without human intervention, some studies suggest that the human factor may be more critical for some types of data [26] For these types, having a contact person and an accompanying publication makes it much easier to understand key contextual details [26,27]. As we start to see more reuse of data, it may be possible to employ more analytical methods for determining best practices based on actual use cases.

For these reasons, we deliberately refrained from assigning grades or calling out individual repositories in the work presented here. [7] noted that many repositories which were evaluated early on using FAIRmetrics expressed resentment. We recognize the struggles that those who develop and host scientific data repositories undergo to keep the resource up and running, particularly in the face of uncertain funding. Generally, these repositories were founded to serve a particular community, and the community itself may not be demanding or engaging with OFCT principles. We therefore favor flexible approaches that allow individual communities to interpret OFCT within the norms of their community and not entirely according to the dictates of external evaluators. Nevertheless, research data repositories, after operating largely on their own to determine the best way to serve research data, are going to have to adapt to meet the challenges and opportunities of making research data a primary product of scientific research.

Supporting information

S1 Table. Question identifiers and their abbreviated meaning. PolicyModels allows specifying an id for each question. This id is later used to identify that question, and to localize its text.

The ids we use here pertain to the subject of the question. The table below explains each abbreviation.

(DOCX)

S2 Table. Abbreviations used in text.

(DOCX)

Acknowledgments

The authors wish to thank Drs. Ko Wei Lin and Jeffrey Grethe for helpful comments.

Data Availability

The data outputs and completed questionnaires from the interview analysis are in Zenodo (RRID:SCR_004129): https://zenodo.org/record/4069364. The latest version of the dkNET evaluation instrument is available at http://trees.scicrunch.io/models/dkNET-DRP/start and is made available under a CC-BY 4.0 license. The version used for this study, V1.0 is available at: http://trees.scicrunch.io/models/dkNET-DRP/7/?localizationName=en-US. A copy of the codebook for the instrument along with the visualizations produced by the PolicyModels software is available through Zenodo at (Martone, Murphy, and Bar-Sinai 2020) Additional explication of the Policy Models dimension usage: https://github.com/codeworth-gh/dkNET-DecisionTrees/blob/master/data-repo-compliance/dimension-usage.adoc A snapshot of the code underlying this study is available at: https://doi.org/10.5281/zenodo.4275004 PolicyModels is managed in GitHub (https://github.com/IQSS/DataTaggingLibrary) under an Apache v2 Open-source license. The summary tools (https://github.com/michbarsinai/PolicyModelsSummary) are released under an MIT license.

Funding Statement

This work was supported by NIH grant# 3U24DK097771-08S1 from the National Institutes of Diabetes and Digestive and Kidney Diseases. The original decision tree was developed through the FORCE11 Scholarly Commons working group supported by an award from The Leona M. and Harry B. Helmsley Charitable Trust Biomedical Research Infrastructure Program to FORCE11. The authors wish to thank Drs. Ko Wei Lin and Jeffrey Grethe for helpful comments. Fiona Murphy is employed by MoreBrains Cooperative Ltd, Chichester, UK, a commercial company that did not provide any direct funding for this project and does not have any direct interests in the repositories, dkNET or NIH. The funder provided support to UCSD for this work which was paid in the form of salaries for MM and subcontracts to FM and MBS, but did not have any additional role in the study design, data collection and analysis, decision to publish, or preparation of the manuscript. The specific roles of these authors are articulated in the ‘author contributions ’section.

References

  • 1.OpenAire. How to select a data repository? [cited 13 Oct 2020]. https://www.openaire.eu/opendatapilot-repository-guide.
  • 2.Wilkinson MD, Dumontier M, Aalbersberg IJ, Appleton G, Axton M, Baak A, et al. The FAIR Guiding Principles for scientific data management and stewardship. Scientific Data. 2016;3: 160018. doi: 10.1038/sdata.2016.18 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Berman HM, Kleywegt GJ, Nakamura H, Markley JL. The Protein Data Bank at 40: reflecting on the past to prepare for the future. Structure. 2012;20: 391–396. doi: 10.1016/j.str.2012.01.010 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Bosman J.; Bruno I; Chapman C; Greshake Tzovaras B; Jacobs N; Kramer B; et al. (2017) The Scholarly Commons—principles and practices to guide research communication OSF Preprints, 10.31219/osf.io/6c2xt. [DOI] [Google Scholar]
  • 5.Cachat J, Bandrowski A, Grethe JS, Gupta A, Astakhov V, Imam F, et al. A survey of the neuroscience resource landscape: perspectives from the neuroscience information framework. Int Rev Neurobiol. 2012;103: 39–68. doi: 10.1016/B978-0-12-388408-4.00003-4 [DOI] [PubMed] [Google Scholar]
  • 6.Clarke DJB, Wang L, Jones A, Wojciechowicz ML, Torre D, Jagodnik KM, et al. FAIRshake: Toolkit to Evaluate the FAIRness of Research Digital Resources. Cell Syst. 2019;9: 417–421. doi: 10.1016/j.cels.2019.09.011 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Wilkinson MD, Dumontier M, Sansone S-A, Bonino da Silva Santos LO, Prieto M, Batista D, et al. Evaluating FAIR maturity through a scalable, automated, community-governed framework. Sci Data. 2019;6: 174. doi: 10.1038/s41597-019-0184-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Shearer K. COAR Community Framework for Best Practices in repositories. [cited 3 Apr 2021]. https://www.coar-repositories.org/news-updates/coar-community-framework-for-best-practices-in-repositories/.
  • 9.McQuilton P, Sansone S-A, Cousijn H, Cannon M, Chan W, Carnevale I, et al. FAIRsharing Collaboration with DataCite and Publishers: Data Repository Selection, Criteria That Matter. Open Science Framework; 2020. doi: 10.17605/OSF.IO/N9QJ7 [DOI] [Google Scholar]
  • 10.Fenner M, Crosas M, Grethe JS, Kennedy D, Hermjakob H, Rocca-Serra P, et al. A data citation roadmap for scholarly data repositories. Sci Data. 2019;6: 28. doi: 10.1038/s41597-019-0031-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.NOT-OD-21-016: Supplemental Information to the NIH Policy for Data Management and Sharing: Selecting a Repository for Data Resulting from NIH-Supported Research. [cited 3 Apr 2021]. https://grants.nih.gov/grants/guide/notice-files/NOT-OD-21-016.html.
  • 12.Whetzel PL, Grethe JS, Banks DE, Martone ME. The NIDDK Information Network: A Community Portal for Finding Data, Materials, and Tools for Researchers Studying Diabetes, Digestive, and Kidney Diseases. PLoS One. 2015;10: e0136206. doi: 10.1371/journal.pone.0136206 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Open Knowledge Open Definition Group. The Open Definition—Open Definition—Defining Open in Open Data, Open Content and Open Knowledge. [cited 25 Feb 2020]. https://opendefinition.org/.
  • 14.Data Citation Synthesis Group. Joint Declaration of Data Citation Principles. 2013. https://www.force11.org/datacitationprinciples.
  • 15.Bilder G, Lin J, Neylon C. Principles for open scholarly infrastructures. Science in the Open. 2015. https://cameronneylon.net/blog/principles-for-open-scholarly-infrastructures/.
  • 16.CoreTrustSeal Standards and Certification Board. CoreTrustSeal Trustworthy Data Repositories Requirements: Glossary 2020–2022. 2019. doi: 10.5281/zenodo.3632563 [DOI] [Google Scholar]
  • 17.Martone ME, Murphy F, Bar-Sinai M. dkNET Repository Compliance. 2020.
  • 18.Bar-Sinai M, Sweeney L, Crosas M. DataTags, Data Handling Policy Spaces and the Tags Language. 2016 IEEE Security and Privacy Workshops (SPW). 2016. pp. 1–8. [Google Scholar]
  • 19.Bar-Sinai M, Murphy F, Martone M. Data Appendix for “Alignment of data sharing practices in biomedical data repositories with open, FAIR, citable and trustworthy principles.” 2020. doi: 10.5281/zenodo.4662165 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Bar-Sinai M, Murphy F, Martone M. codeworth-gh/dkNET-DecisionTrees: Paper submittion. 2020.
  • 21.Starr J, Castro E, Crosas M, Dumontier M, Downs RR, Duerr R, et al. Achieving human and machine accessibility of cited data in scholarly publications. PeerJ Computer Science. 2015;1: e1. doi: 10.7717/peerj-cs.1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Cousijn H, Kenall A, Ganley E, Harrison M, Kernohan D, Lemberger T, et al. A Data Citation Roadmap for Scientific Publishers. bioRxiv. 2018. p. 100784. doi: 10.1038/sdata.2018.259 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Lin D, Crabtree J, Dillo I, Downs RR, Edmunds R, Giaretta D, et al. The TRUST Principles for digital repositories. Sci Data. 2020;7: 144. doi: 10.1038/s41597-020-0486-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Sansone S-A, McQuilton P, Cousijn H, Cannon M, Chan WM, Callaghan S, et al. Data Repository Selection: Criteria That Matter. 2020. doi: 10.5281/zenodo.4084763 [DOI] [Google Scholar]
  • 25.Durinx C, McEntyre J, Appel R, Apweiler R, Barlow M, Blomberg N, et al. Identifying ELIXIR Core Data Resources. F1000Res. 2016;5. doi: 10.12688/f1000research.9656.2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Faniel IM, Yakel E. Practices Do Not Make Perfect: Sharing, Disciplinary Data; Practices, Reuse. In: Johnson LR, editor. Curating Research Data Volume One: Practical Strategies for Your Digital Repository. Association of College and Research Libraries; 2017. pp. 103–125. [Google Scholar]
  • 27.Turner CF, Pan H, Silk GW, Ardini M-A, Bakalov V, Bryant S, et al. The NIDDK Central Repository at 8 years—ambition, revision, use and impact. Database. 2011;2011: bar043. doi: 10.1093/database/bar043 [DOI] [PMC free article] [PubMed] [Google Scholar]

Decision Letter 0

Florian Naudet

27 Jan 2021

PONE-D-20-39290

Alignment of biomedical data repositories with open, FAIR, citable and trustworthy principles

PLOS ONE

Dear Dr. Martone,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

I have a few specific comments

- Please state if the research was pre-registered or not explicitely in the method section ; 

- Figure 2 would be better in a supplementary material (in my opinion) ;

- On Figures 5 and 6, please provide the name of each repository as a legend (if possible) ; 

Please submit your revised manuscript by Mar 13 2021 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

We look forward to receiving your revised manuscript.

Kind regards,

Florian Naudet, M.D., M.P.H., Ph.D.

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Thank you for stating the following in the Financial Disclosure section:

"This work was supported by NIH grant# 3U24DK097771-08S1 from the National Institutes of Diabetes and Digestive and Kidney Diseases.  The original decision tree was developed through the  FORCE11 Scholarly Commons working group supported by an award from The Leona M. and Harry B. Helmsley Charitable Trust Biomedical Research Infrastructure Program to FORCE11.  The authors wish to thank Drs. Ko Wei Lin and Jeffrey Grethe for helpful comments.  The funders played no role in the design, data collection, analysis, decision to publish or preparation of the manuscript."

We note that one or more of the authors are employed by a commercial company: MoreBrains Cooperative Ltd, Chichester, UK

2. 1. Please provide an amended Funding Statement declaring this commercial affiliation, as well as a statement regarding the Role of Funders in your study. If the funding organization did not play a role in the study design, data collection and analysis, decision to publish, or preparation of the manuscript and only provided financial support in the form of authors' salaries and/or research materials, please review your statements relating to the author contributions, and ensure you have specifically and accurately indicated the role(s) that these authors had in your study. You can update author roles in the Author Contributions section of the online submission form.

Please also include the following statement within your amended Funding Statement.

“The funder provided support in the form of salaries for authors [insert relevant initials], but did not have any additional role in the study design, data collection and analysis, decision to publish, or preparation of the manuscript. The specific roles of these authors are articulated in the ‘author contributions’ section.”

If your commercial affiliation did play a role in your study, please state and explain this role within your updated Funding Statement.

2. 2. Please also provide an updated Competing Interests Statement declaring this commercial affiliation along with any other relevant declarations relating to employment, consultancy, patents, products in development, or marketed products, etc.  

Within your Competing Interests Statement, please confirm that this commercial affiliation does not alter your adherence to all PLOS ONE policies on sharing data and materials by including the following statement: "This does not alter our adherence to  PLOS ONE policies on sharing data and materials.” (as detailed online in our guide for authors http://journals.plos.org/plosone/s/competing-interests) . If this adherence statement is not accurate and  there are restrictions on sharing of data and/or materials, please state these. Please note that we cannot proceed with consideration of your article until this information has been declared.

Please include both an updated Funding Statement and Competing Interests Statement in your cover letter. We will change the online submission form on your behalf.

3. Thank you for stating the following in the Competing Interests section:

"I have read the journal's policy and the authors of this manuscript have the following competing interests: Dr. Martone is on the board and has equity interest in SciCrunch Inc., a tech startup that develops tools and services in support of Research Resource Identifiers.

Dr Murphy is on the board of Dryad Data Repository."

Please confirm that this does not alter your adherence to all PLOS ONE policies on sharing data and materials, by including the following statement: "This does not alter our adherence to  PLOS ONE policies on sharing data and materials.” (as detailed online in our guide for authors http://journals.plos.org/plosone/s/competing-interests). If there are restrictions on sharing of data and/or materials, please state these. Please note that we cannot proceed with consideration of your article until this information has been declared.

Please respond by return email with your amended Competing Interests Statement and we will change the online submission form on your behalf.

Please know it is PLOS ONE policy for corresponding authors to declare, on behalf of all authors, all potential competing interests for the purposes of transparency. PLOS defines a competing interest as anything that interferes with, or could reasonably be perceived as interfering with, the full and objective presentation, peer review, editorial decision-making, or publication of research or non-research articles submitted to one of the journals. Competing interests can be financial or non-financial, professional, or personal. Competing interests can arise in relationship to an organization or another person. Please follow this link to our website for more details on competing interests: http://journals.plos.org/plosone/s/competing-interests

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Partly

Reviewer #3: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: N/A

Reviewer #3: I Don't Know

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: First of all, I want to congratulate the authors on this interesting piece of research.

In times of open data, data repositories are evolving rapidly and thus, it is even more important to have scientific articles analyzing them.

A further plus of this paper is the openness and transparency which is shown by sharing the code and the data. Finally, the paper shows a well thought through study design.

However, there are some points that need to be improved.

I recommend an acceptance with minor improvements.

As the lines are not numbered, I will go through the paper section by section.

Title:

Please include the design of the research in the title and its outcome

Abstract:

Please include that you are looking for repositories focusing on diabetes and metabolic diseases.

Again, the design of the study is missing

Please include the time frame of when the research was conducted.

Introduction:

Overall, I am missing references to already established research, but particularly in the introduction section. Was this the only research initiative on NIH repositories.

Methods:

Please explain how many repositories are on the dkNet website and describe the outsourcing process. A flowchart could be useful here for readers.

Did you preregister a protocol on OSF, etc? If so, please insert the link. If not, please explain why.

Did you pilot test the questionnaire on other repositories?

Please include references for the TRUST principles and the CoreTrust Seal.

Table 2 is kind of confusing with the Q#ID for the reader. In my opinion not necessary as it is in the supplementary material S1 again listed. However, personal preference here.

Same problem in Table 3.

Is listing RRID necessary?

What is the point of the section “section”. For the AMP-T2D you put data. My understanding is that “data” would be applicable to all repositories.

Please take care that you explain abbreviations. These might not always be clear for readers who are not familiar with the subject (RRID, API, …). It would be good to have a list of abbreviations in the beginning of the manuscript.

Please state if you followed a reporting guideline, STROBE would be a possibility here. If you don’t use one, please explain why.

Results:

When you finished the research and didn’t find certain items, did you then contact the responsible or repository managers?

Keep in mind that it is highly appreciated that you created accounts in some cases.

I am sorry if I didn’t understand this point but how did you get to a 10 point scale if in some questions you had only 3 or 4 options to choose.

On page 16 you are referring to table 3 when you mean table 2.

Discussion:

I am again missing the focus on the specific disease area. You are writing in a general way on repositories but in which way do your results have external validity for other biomedical areas?

The study design is very well done, however only applied on 8 repositories; should be further discussed/ mentioned.

Reviewer #2: The authors present a framework for data repositories to support Open, FAIR, Citable and Trustworthy (OFCT) data. They describe the application of the open source PolicyModels toolkit to operationalize key aspects of OFCT principles. They use the framework to assess the current state of eight biomedical community repositories listed by the NIDDK Information Network (dkNET). The dkNET is taking an active role in interpreting and facilitating compliance with FAIR on behalf of this community. By defining a set of important functions repositories should support, the work will support two goals: (i) serve as a best practice for those developing new dk data repositories, and (ii) inform the creation of a dkNET tool to help researchers select an appropriate repository for their data.

In general, assessment of repositories and is an important area, and more attention is being paid in biomedicine to certification instruments. I like very much the online PolicyModels toolkit, its functionality and usability, and the machine-readable exports.

While I applaud the goals, in general, I have the following concerns.

1. The process, and the rationale for the development of the OFCT principles are not explained. As the authors say “To our knowledge, this is the first evaluation instrument that was designed specifically around OFCT. However, since the issuance of the FAIR data principles, several initiatives have invested in the development of tools that are designed to assess the level of data FAIRness, including those that are meant to evaluate on-line data repositories”. To my knowledge, beside FAIR and the CTS certifications, there are the several community-driven ‘principles’ to assess repositories and/or provide best pratices for repositories developers and maintainers, by the following groups: COAR (https://www.coar-repositories.org/coar-community-framework-for-good-practices-in-repositories), NIH (https://grants.nih.gov/grants/guide/notice-files/NOT-OD-21-016.html), ELIXIR (https://doi.org/10.12688/f1000research.9656.2) and a group of publishers (https://doi.org/10.5281/zenodo.4084763). To address this, I would like to see the authors explaining: (i) how OFCT was developed, in terms of community participation, and how the range of acceptable and unacceptable values in various dimensions (properties and flags) were defined; (ii) the rationale about creating these new set of principles, and how OFCT is different, complementary or more usable than the others; and (i) how the OFCT maps to these existing efforts, and to the CTS.

2. Although FAIR is only one of the elements of the OFCT principles, much of the discussion is only centered on comparing them to FAIR assessment tools, such as the FAIR Maturity Indicators, FAIR Evaluator and FAIRshake, which are solely based on FAIRness. To address this, I would like to see the authors expanding the discussion to on the rest of the elements in the OFCT principles, which can be done by addressing my point 1.

3. It seems that the application of the OFCT principles is based on manual assessment and on the interpretation of the results. The authors say: “…we deliberately refrained from assigning grades or calling out individual repositories in the work presented here. (Wilkinson et al. 2019) noted that many repositories which were evaluated early on using FAIRmetrics expressed resentment. We recognize the struggles that those who develop and host scientific data repositories undergo to keep the resource up and running, particularly in the face of uncertain funding……We therefore favor flexible approaches that allow individual communities to interpret OFCT within the norms of their community and not entirely according to the dictates of external evaluators.” And also: “The FAIR Maturity Indicators and FAIRshake toolkits differ from ours in that they are intended to employ either fully automated or semi-automated approaches for determining FAIRness. As we show here, some aspects of FAIR require interpretation, e.g., “a plurality of relevant attributes”, making it difficult to employ fully automated approaches”. I understand that fully automated approaches may not be easy, but I do not see how manual assessment would be better and work in practice, in particular be objective and scalable. To address this, I would like to see the authors explaining how they see the manual assessment (with OFCT principles), and what they call the “OFCT potential”, to help them to develop a tool to select an appropriate repository (one of their goals).

Other comments

4. The assessment was performed through inspection of documentation and interaction with the sites, without interacting with the repository’s owners. However, this is not clear. Fig 2 talks about “survey and questionnaire and interactive interview”, but then in the Scoring section authors say : “Five of the sites were reviewed independently by FM and MM…The reviewers made a good faith effort to find information on the site to provide an accurate answer for each question”. Therefore, the authors should make clear that this not a survey sent to repository’s owners or a self-assessment done by the repository’s owners. Nevertheless, the authors could contact the repository’s owners to vet the results.

5. The authors say that they have benchmarked their instrument against others for ‘similar use’, such as: the Datacite’s Repository Finders, and the Scientific Data’s Repository Questionnaire, and the FAIRsFAIR assessment tool. However, there is not information on what this benchmarked has produced since these efforts are not equivalent and the OFCT are principles and indeed not a tool.

6. The data is well stored and cited, as expected, but there are a lot of key details scattered across figures and in the GitHub repositories, and in places it is quite hard to follow.

Reviewer #3: The paper and study have were clearly set out. In particular, the discussion provided insight into the difficulty of other toolkits, limitations on a universal approach and how this study addressed these challenges.

The paper could have included more detail about how the toolkit will impact the workflows of researchers (for the better) and how the community can apply the toolkit in different contexts. It is not immediately clear how the toolkit operationalises key aspects of OFCT principles and this could be further highlighted. It should be noted, however, the toolkit and related materials have been made available for reuse.

It was an interesting paper that has built on work that has been done by other groups and provided a good framework for the community to adopt.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

Reviewer #3: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2021 Jul 9;16(7):e0253538. doi: 10.1371/journal.pone.0253538.r002

Author response to Decision Letter 0


7 Apr 2021

We responded to all comments raised by the editor and reviewers. I uploaded a response to the review but I'm including them here as well.

Response to reviewer comments

We thank the editor and the reviewers for their thoughtful and helpful comments. We have revised the manuscript substantially in response to these critiques.

Editor’s comments

PO-1: Attached - marked-up copy of the manuscript labelled 'Revised Manuscript with Track Changes', and clean version labelled ‘Manuscript’

Response: We have attached both a version with track changes on and a clean version labeled as per above.

PO-2: Please state if the research was pre-registered or not explicitly in the method section ;

Response: Done

PO-3: Figure 2 would be better in a supplementary material (in my opinion) ;

Reponse: Would prefer to leave in main text as one of our audience targets is repository owners or evaluators who may want to use the instrument. So we wanted to show them the user-friendliness of the user interface. However, if the editor prefers, we can move it.

PO-4: On Figures 5 and 6, please provide the name of each repository as a legend (if possible) ;

Response: We prefer not to identify individual repositories in the manuscript, as our intent is to examine overall trends and practices, not to critique individual repositories. The scores for individual repositories are, however, available in the data provided.

PO-5: If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future.

Response: We don’t feel it necessary as the protocol is documented in the instrument and the particular instrument used was given a DOI with a clear version number.

Journal Requirements:

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Thank you for stating the following in the Financial Disclosure section:

"This work was supported by NIH grant# 3U24DK097771-08S1 from the National Institutes of Diabetes and Digestive and Kidney Diseases. The original decision tree was developed through the FORCE11 Scholarly Commons working group supported by an award from The Leona M. and Harry B. Helmsley Charitable Trust Biomedical Research Infrastructure Program to FORCE11. The authors wish to thank Drs. Ko Wei Lin and Jeffrey Grethe for helpful comments. The funders played no role in the design, data collection, analysis, decision to publish or preparation of the manuscript.

We note that one or more of the authors are employed by a commercial company: MoreBrains Cooperative Ltd, Chichester, UK

2. 1. Please provide an amended Funding Statement declaring this commercial affiliation, as well as a statement regarding the Role of Funders in your study. If the funding organization did not play a role in the study design, data collection and analysis, decision to publish, or preparation of the manuscript and only provided financial support in the form of authors' salaries and/or research materials, please review your statements relating to the author contributions, and ensure you have specifically and accurately indicated the role(s) that these authors had in your study. You can update author roles in the Author Contributions section of the online submission form.

Response: We appended the following statement to our Funding statement: “ Fiona Murphy is employed by MoreBrains Cooperative Ltd, Chichester, UK, a commercial company that did not provide any direct funding for this project and does not have any direct interests in the repositories, dkNET or NIH."

Please also include the following statement within your amended Funding Statement.

“The funder provided support in the form of salaries for authors [insert relevant initials], but did not have any additional role in the study design, data collection and analysis, decision to publish, or preparation of the manuscript. The specific roles of these authors are articulated in the ‘author contributions’ section.”

If your commercial affiliation did play a role in your study, please state and explain this role within your updated Funding Statement.

Response: Done

2. 2. Please also provide an updated Competing Interests Statement declaring this commercial affiliation along with any other relevant declarations relating to employment, consultancy, patents, products in development, or marketed products, etc.

Within your Competing Interests Statement, please confirm that this commercial affiliation does not alter your adherence to all PLOS ONE policies on sharing data and materials by including the following statement: "This does not alter our adherence to PLOS ONE policies on sharing data and materials.” (as detailed online in our guide for authors http://journals.plos.org/plosone/s/competing-interests) . If this adherence statement is not accurate and there are restrictions on sharing of data and/or materials, please state these. Please note that we cannot proceed with consideration of your article until this information has been declared.

Response: Done

Please include both an updated Funding Statement and Competing Interests Statement in your cover letter. We will change the online submission form on your behalf.

3. Thank you for stating the following in the Competing Interests section:

"I have read the journal's policy and the authors of this manuscript have the following competing interests: Dr. Martone is on the board and has equity interest in SciCrunch Inc., a tech startup that develops tools and services in support of Research Resource Identifiers.

Dr Murphy is on the board of Dryad Data Repository."

Please confirm that this does not alter your adherence to all PLOS ONE policies on sharing data and materials, by including the following statement: "This does not alter our adherence to PLOS ONE policies on sharing data and materials.” (as detailed online in our guide for authors http://journals.plos.org/plosone/s/competing-interests). If there are restrictions on sharing of data and/or materials, please state these. Please note that we cannot proceed with consideration of your article until this information has been declared.

Response: Done

Please respond by return email with your amended Competing Interests Statement and we will change the online submission form on your behalf.

Please know it is PLOS ONE policy for corresponding authors to declare, on behalf of all authors, all potential competing interests for the purposes of transparency. PLOS defines a competing interest as anything that interferes with, or could reasonably be perceived as interfering with, the full and objective presentation, peer review, editorial decision-making, or publication of research or non-research articles submitted to one of the journals. Competing interests can be financial or non-financial, professional, or personal. Competing interests can arise in relationship to an organization or another person. Please follow this link to our website for more details on competing interests: http://journals.plos.org/plosone/s/competing-interests

Reviewer #1

Reviewer #1: First of all, I want to congratulate the authors on this interesting piece of research.

In times of open data, data repositories are evolving rapidly and thus, it is even more important to have scientific articles analyzing them.A further plus of this paper is the openness and transparency which is shown by sharing the code and the data. Finally, the paper shows a well thought through study design.

Response: We thank the reviewer for these positive comments

1-1: Title: Please include the design of the research in the title and its outcome

Response: We modified the title to: “A new tool for assessing aAlignment of biomedical data repositories with open, FAIR, citable and trustworthy principles”

1-2: Abstract: Please include that you are looking for repositories focusing on diabetes and metabolic diseases. Again, the design of the study is missing. Please include the time frame of when the research was conducted.

Response: As only two of the repositories were specifically focused on diabetes and metabolomics, we added the following statement: Repositories included both specialist repositories that focused on a particular data type or domain, in this case diabetes and metabolomics, and generalist repositories that accept all data types and domains.

We also added some additional methodological details and the timeline over which the repositories were evaluated.

1-3: Introduction: Overall, I am missing references to already established research, but particularly in the introduction section. Was this the only research initiative on NIH repositories.

Response: We included references to other instruments that have been designed to evaluate scientific and biomedical data repositories and added the following text:

Because of the central importance of data repositories to achieving goals in open science and FAIR, many groups have been developing criteria and instruments to assess the state of scientific data repositories, including biomedical repositories, e.g., FAIRmetrics, FAIRshake, Criteria that matter, to name a few. In the work presented here, we wanted to expand the evaluation criteria based on OFCT and to specifically assess repositories that are relevant to a specific domain, diabetes, digestive and kidney diseases, on developed an instrument to assess the current state of data repositories on behalf of the NIDDK Information Network (dkNET.org)

1-4: Methods:

Please explain how many repositories are on the dkNet website and describe the outsourcing process. A flowchart could be useful here for readers.

Response: There are over 80 repositories listed on the dkNET website. We are not sure what is meant by “outsourced” here. We added more information to the manuscript about how dkNET chooses repositories for listing. We explain in the manuscript how we chose a sample for evaluation.

Did you preregister a protocol on OSF, etc? If so, please insert the link. If not, please explain why.

Response: No, we did not and we have added a statement to the manuscript to that effect. However, a preprint was posted in BiorXiv. We view that as a form of community review prior to publication. However, it is a good suggestion to register all such studies in the future and we appreciate you bringing it to our attention.

Did you pilot test the questionnaire on other repositories?

Response: Yes, we tested the questionnaire on 13 repositories initially to pilot the questionnaire. Some of them are the same repositories that we evaluated for the final sample, but not all repositories evaluated were in the initial set.

Please include references for the TRUST principles and the CoreTrust Seal.

Response: Both references to the Core Trust Seal and the TRUST principles are included.

Table 2 is kind of confusing with the Q#ID for the reader. In my opinion not necessary as it is in the supplementary material S1 again listed. However, personal preference here.

Response: Agreed. We removed the IDs but left in the question number for ease of reference in the text.

Same problem in Table 3.

Is listing RRID necessary?

Response: Yes, the RRID is used to make it easy to track use of repositories and is supported by PLoS. However, we removed the column entitled “RRID” and just listed it under each repository in smaller font so it isn’t so prominent.

What is the point of the section “section”. For the AMP-T2D you put data. My understanding is that “data” would be applicable to all repositories.

Response: Many of the websites are complicated scientific applications with multiple sections. We wanted to be clear about which section we reviewed for compliance. In the AMP-T2D, the tab heading is entitled “Data” and that is what we referred to. However, we modified the text to indicate the title of the page and the section header to make it more clear.

Please take care that you explain abbreviations. These might not always be clear for readers who are not familiar with the subject (RRID, API, …). It would be good to have a list of abbreviations in the beginning of the manuscript.

Response: Done. We included it in the supplemental materials.

Please state if you followed a reporting guideline, STROBE would be a possibility here. If you don’t use one, please explain why.

Response: We followed the guidelines recommended for FAIR metrics (https://www.nature.com/articles/sdata2018118/tables/1

1-5: Results:

When you finished the research and didn’t find certain items, did you then contact the responsible or repository managers?

Keep in mind that it is highly appreciated that you created accounts in some cases.

Response: We did not but in response to this comment, we have sent the completed questionnaires to all of the repositories and have described the results in the text. At the time of deadline for resubmission, we received 3 responses. If by the time of rereview we receive more, we will update accordingly.

I am sorry if I didn’t understand this point but how did you get to a 10 point scale if in some questions you had only 3 or 4 options to choose.

Response: Questions could have 2, 3 or 4 possible answers so to make it easier to compare across questions to gauge the overall alignment with each principle , we normalized all questions to a 10 point scale. We added a sentence to the text to explain.

On page 16 you are referring to table 3 when you mean table 2.

Response: As we removed the IDs from that table, we removed the erroneous reference to Table 3 as it was no longer relevant.

1-6: Discussion:

I am again missing the focus on the specific disease area. You are writing in a general way on repositories but in which way do your results have external validity for other biomedical areas?

Response: We added a paragraph to the discussion on this point. While the majority of questions would apply to all scientific data repositories, for questions that probed the level of metadata, we specifically looked for metadata relevant for biomedicine.

The study design is very well done, however only applied on 8 repositories; should be further discussed/ mentioned.

Response: We mention that this study focused on the design of the instrument and an initial evaluation of eight representative repositories to see how it performed. We intend to apply it to all repositories listed in dkNET.

Reviewer #2:

Reviewer #2: The authors present a framework for data repositories to support Open, FAIR, Citable and Trustworthy (OFCT) data. They describe the application of the open source PolicyModels toolkit to operationalize key aspects of OFCT principles. They use the framework to assess the current state of eight biomedical community repositories listed by the NIDDK Information Network (dkNET). The dkNET is taking an active role in interpreting and facilitating compliance with FAIR on behalf of this community. By defining a set of important functions repositories should support, the work will support two goals: (i) serve as a best practice for those developing new dk data repositories, and (ii) inform the creation of a dkNET tool to help researchers select an appropriate repository for their data.

In general, assessment of repositories and is an important area, and more attention is being paid in biomedicine to certification instruments. I like very much the online PolicyModels toolkit, its functionality and usability, and the machine-readable exports.

Response: We thank the reviewer for the positive comments

While I applaud the goals, in general, I have the following concerns.

2-1. The process, and the rationale for the development of the OFCT principles are not explained. As the authors say “To our knowledge, this is the first evaluation instrument that was designed specifically around OFCT. However, since the issuance of the FAIR data principles, several initiatives have invested in the development of tools that are designed to assess the level of data FAIRness, including those that are meant to evaluate on-line data repositories”. To my knowledge, beside FAIR and the CTS certifications, there are the several community-driven ‘principles’ to assess repositories and/or provide best practices for repositories developers and maintainers, by the following groups: COAR (https://www.coar-repositories.org/coar-community-framework-for-good-practices-in-repositories), NIH (https://grants.nih.gov/grants/guide/notice-files/NOT-OD-21-016.html), ELIXIR (https://doi.org/10.12688/f1000research.9656.2) and a group of publishers (https://doi.org/10.5281/zenodo.4084763). To address this, I would like to see the authors explaining: (i) how OFCT was developed, in terms of community participation, and how the range of acceptable and unacceptable values in various dimensions (properties and flags) were defined; (ii) the rationale about creating these new set of principles, and how OFCT is different, complementary or more usable than the others; and (i) how the OFCT maps to these existing efforts, and to the CTS.

Response: We added additional background material in the introduction that explains the origins of the OFCT framework through the work of FORCE11. We note that the principles referenced in OFCT were not developed by us but rather through large community efforts, e.g., the Data Citation Principles, FAIR and Trust, or third parties (Principles of open infrastructure and the Open Definition. We arrived at OFCT through a process of community workshops in FORCE11 and through analysis of dozens of principle sets issued around the world that emphasized openness, reusability, credit and trustworthiness. We operationalized these concepts first by selecting community principles that embody these concepts and then selecting criteria through which they could be evaluated. In addition, we add references to COAR, Elixir and NIH which were missing from our article. We note that COAR, which was released in Oct 2020 after our instrument was complete actually is close in spirit to OFCT and therefore complementary to our efforts. We include a discussion of that point.

2-2. Although FAIR is only one of the elements of the OFCT principles, much of the discussion is only centered on comparing them to FAIR assessment tools, such as the FAIR Maturity Indicators, FAIR Evaluator and FAIRshake, which are solely based on FAIRness. To address this, I would like to see the authors expanding the discussion to on the rest of the elements in the OFCT principles, which can be done by addressing my point 1.

Response: We had referenced some of these other efforts in the discussion, but agree that they should be brought up earlier, so we have added more background to the introduction to reference additional efforts.

2-3. It seems that the application of the OFCT principles is based on manual assessment and on the interpretation of the results. The authors say: “…we deliberately refrained from assigning grades or calling out individual repositories in the work presented here. (Wilkinson et al. 2019) noted that many repositories which were evaluated early on using FAIRmetrics expressed resentment. We recognize the struggles that those who develop and host scientific data repositories undergo to keep the resource up and running, particularly in the face of uncertain funding……We therefore favor flexible approaches that allow individual communities to interpret OFCT within the norms of their community and not entirely according to the dictates of external evaluators.” And also: “The FAIR Maturity Indicators and FAIRshake toolkits differ from ours in that they are intended to employ either fully automated or semi-automated approaches for determining FAIRness. As we show here, some aspects of FAIR require interpretation, e.g., “a plurality of relevant attributes”, making it difficult to employ fully automated approaches”. I understand that fully automated approaches may not be easy, but I do not see how manual assessment would be better and work in practice, in particular be objective and scalable. To address this, I would like to see the authors explaining how they see the manual assessment (with OFCT principles), and what they call the “OFCT potential”, to help them to develop a tool to select an appropriate repository (one of their goals).

Response: We agree that we did not tie the two themes together effectively and have not added more material as to how the manual evaluation helps an organization like dkNET to improve the quality of repositories serving their domains and provide useful information to our constituents. We have strengthened the basic premise that while repositories should have a set of core functions they support, e.g., PIDs, different groups may evaluate their fitness of purpose differently.

Other comments

2-4. The assessment was performed through inspection of documentation and interaction with the sites, without interacting with the repository’s owners. However, this is not clear. Fig 2 talks about “survey and questionnaire and interactive interview”, but then in the Scoring section authors say : “Five of the sites were reviewed independently by FM and MM…The reviewers made a good faith effort to find information on the site to provide an accurate answer for each question”. Therefore, the authors should make clear that this not a survey sent to repository’s owners or a self-assessment done by the repository’s owners. Nevertheless, the authors could contact the repository’s owners to vet the results.

Response: We have made that point clearer in the methods. In addition, following the reviewers’ recommendation, we sent the results of our assessment to the 8 repositories. We had acknowledgements from 7 of them and received feedback from 3 of them to date. We have included this in our methods and results.

2-5. The authors say that they have benchmarked their instrument against others for ‘similar use’, such as: the Datacite’s Repository Finders, and the Scientific Data’s Repository Questionnaire, and the FAIRsFAIR assessment tool. However, there is not information on what this benchmarked has produced since these efforts are not equivalent and the OFCT are principles and indeed not a tool.

Response: We have expanded on the explanation for this section in the methods section to explain that we were seeking both context for our own work and information as to what is required for a thorough questionnaire that will be of use to its potential community.

2-6. The data is well stored and cited, as expected, but there are a lot of key details scattered across figures and in the GitHub repositories, and in places it is quite hard to follow.

Response: We tried to simplify the figures and the amount of cross referencing that is necessary. We have also deposited all of the code and data underlying this study in Zenodo and we make that clear in the first paragraph of the data availability statement.

Reviewer #3

Reviewer #3: The paper and study were clearly set out. In particular, the discussion provided insight into the difficulty of other toolkits, limitations on a universal approach and how this study addressed these challenges.

Response: We thank the reviewer for his positive comments

3-1. The paper could have included more detail about how the toolkit will impact the workflows of researchers (for the better) and how the community can apply the toolkit in different contexts. It is not immediately clear how the toolkit operationalises key aspects of OFCT principles and this could be further highlighted. It should be noted, however, the toolkit and related materials have been made available for reuse.

Response: We added additional detail about how this instrument will be used in dkNET and can be used in different contexts.

3-2. It was an interesting paper that has built on work that has been done by other groups and provided a good framework for the community to adopt.

Thank you for the positive comment.

Attachment

Submitted filename: Response to reviews.docx

Decision Letter 1

Florian Naudet

8 Jun 2021

A tool for assessing alignment of biomedical data repositories with open, FAIR, citation and trustworthy principles

PONE-D-20-39290R1

Dear Dr. Martone,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Florian Naudet, M.D., M.P.H., Ph.D.

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

I would like to thank the 3 reviewers for their comments and for pointing that the manuscript is in a better shape now. I'm happy to accept it for publication. 

As I have informed the authors, it was not possible for one of the reviewer to provide a fast answer. He informed me appropriately and I have decided to give him more time because I wanted to have his opinion on the revised manuscript. I apologize for the time taken but it was very important for me to have his opinion on this new version. I hope that the authors will understand. 

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

Reviewer #2: All comments have been addressed

Reviewer #3: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: N/A

Reviewer #3: I Don't Know

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: I thank the authors for addressing all my comments. This is a nice piece of work and I further wish them a smooth publication process.

Reviewer #2: I thank the authors for their thorough response and revision of the manuscript, which addresses my comments in a satisfactory manner.

Reviewer #3: The revised submission has adequately addressed my comments raised previously. There is growing interest in the assessment of repositories globally and the authors have developed a framework and tool that can be adapted across other domains. The authors have provided a comprehensive overview of the current biomedical repository assessment landscape and developed an instrument that goes beyond the FAIR principles. This is an important step in the Open Science community as repositories need to adapt to best practice. The toolkit can be used for repositories to benchmark against best practice as they continue to evolve. It would have been good to see more about the exploitation or proposed uptake of the toolkit, beyond making the code openly available, although this does not detract from the merit of the paper.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: Yes: Susanna-Assunta Sansone

Reviewer #3: No

Acceptance letter

Florian Naudet

28 Jun 2021

PONE-D-20-39290R1

A tool for assessing alignment of biomedical data repositories with open, FAIR, citation and trustworthy principles

Dear Dr. Martone:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Pr. Florian Naudet

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Table. Question identifiers and their abbreviated meaning. PolicyModels allows specifying an id for each question. This id is later used to identify that question, and to localize its text.

    The ids we use here pertain to the subject of the question. The table below explains each abbreviation.

    (DOCX)

    S2 Table. Abbreviations used in text.

    (DOCX)

    Attachment

    Submitted filename: Response to reviews.docx

    Data Availability Statement

    The data outputs and completed questionnaires from the interview analysis are in Zenodo (RRID:SCR_004129): https://zenodo.org/record/4069364. The latest version of the dkNET evaluation instrument is available at http://trees.scicrunch.io/models/dkNET-DRP/start and is made available under a CC-BY 4.0 license. The version used for this study, V1.0 is available at: http://trees.scicrunch.io/models/dkNET-DRP/7/?localizationName=en-US. A copy of the codebook for the instrument along with the visualizations produced by the PolicyModels software is available through Zenodo at (Martone, Murphy, and Bar-Sinai 2020) Additional explication of the Policy Models dimension usage: https://github.com/codeworth-gh/dkNET-DecisionTrees/blob/master/data-repo-compliance/dimension-usage.adoc A snapshot of the code underlying this study is available at: https://doi.org/10.5281/zenodo.4275004 PolicyModels is managed in GitHub (https://github.com/IQSS/DataTaggingLibrary) under an Apache v2 Open-source license. The summary tools (https://github.com/michbarsinai/PolicyModelsSummary) are released under an MIT license.


    Articles from PLoS ONE are provided here courtesy of PLOS

    RESOURCES