Skip to main content
PLOS ONE logoLink to PLOS ONE
. 2020 Feb 27;15(2):e0229182. doi: 10.1371/journal.pone.0229182

The views, perspectives, and experiences of academic researchers with data sharing and reuse: A meta-synthesis

Laure Perrier 1,*,#, Erik Blondal 2,#, Heather MacDonald 3,#
Editor: Pablo Dorta-González4
PMCID: PMC7046208  PMID: 32106224

Abstract

Background

Funding agencies and research journals are increasingly demanding that researchers share their data in public repositories. Despite these requirements, researchers still withhold data, refuse to share, and deposit data that lacks annotation. We conducted a meta-synthesis to examine the views, perspectives, and experiences of academic researchers on data sharing and reuse of research data.

Methods

We searched the published and unpublished literature for studies on data sharing by researchers in academic institutions. Two independent reviewers screened citations and abstracts, then full-text articles. Data abstraction was performed independently by two investigators. The abstracted data was read and reread in order to generate codes. Key concepts were identified and thematic analysis was used for data synthesis.

Results

We reviewed 2005 records and included 45 studies along with 3 companion reports. The studies were published between 2003 and 2018 and most were conducted in North America (60%) or Europe (17%). The four major themes that emerged were data integrity, responsible conduct of research, feasibility of sharing data, and value of sharing data. Researchers lack time, resources, and skills to effectively share their data in public repositories. Data quality is affected by this, along with subjective decisions around what is considered to be worth sharing. Deficits in infrastructure also impede the availability of research data. Incentives for sharing data are lacking.

Conclusion

Researchers lack skills to share data in a manner that is efficient and effective. Improved infrastructure support would allow them to make data available quickly and seamlessly. The lack of incentives for sharing research data with regards to academic appointment, promotion, recognition, and rewards need to be addressed.

Introduction

Research communities, including funding agencies and scholarly journals, have moved towards greater access to data through the development of policies that promote data sharing [14]. Examples include the development of data sharing requirements for clinical trials by the International Committee of Medical Journal Editors [5], the creation of a data repository for all researchers working towards a solution to the Zika virus so that all data is published as soon as it becomes available [6], and large funding bodies such as the Bill & Melinda Gates Foundation implementing strong open data policies [7].

These global developments require more researchers to share their data and make it available for reuse. Proponents for open data maintain that it offers the opportunity for others to freely reuse data, makes research more reproducible, uses public funds more effectively, and expands the potential to combine data sets for increased statistical power or creating new knowledge [8]. Sharing data is routine and embedded into the research process for some disciplines such as genomics and astronomy [910]. However, in many fields data produced by researchers has traditionally only been shared at the discretion of the principal investigator upon request, and otherwise kept in filing cabinets or on hard drives. These global shifts around research data have left some feeling uneasy and argue that those who generate the data own the data, certain studies (e.g., those with human subjects) require protection that may be difficult to assure with open data, and data sharing puts an increased administrative burden upon researchers [11]. There are also concerns of the inequity of a career built on data reuse versus the hard work of writing grants, being ‘scooped’, or being falsely discredited [1112].

Although funding agencies, institutions, and journals have implemented policies on data sharing and archiving, these practices have not produced the anticipated results. Researchers still withhold data [13], refuse to share data upon request [1415], publish without data availability statements [16], and fail to put their data into repositories [16] even after agreeing to share their data when publishing a journal article. Problems encountered when data is retrieved from repositories include inadequate annotation [17], limited structured data (Marwick), and incomplete specifications for data processing and analysis [17]. To gain insights on these behaviors, it is important to understand researchers’ perspectives. In this study, we report on researchers’ views and experiences on data sharing and reuse.

Aim

Our metasynthesis focuses on the individuals conducting research, and synthesizes the available qualitative literature that examines academic researchers and data sharing. This study addresses the question: what are the views, perspectives, and experiences of academic researchers on data sharing and reuse of research data?

Materials and methods

A protocol was developed and is available upon request to the authors. Although the PRISMA statement has not been modified for meta-syntheses, it was used to guide the reporting of this review and can be viewed in S1 Appendix.

Types of studies

This is a metasynthesis of qualitative primary studies. Qualitative research seeks to discover how people perceive and experience the world around them [18]. Direct communication (e.g., interviews, focus groups) or observation are used to explore people’s perceptions. Data is explored using qualitiatve analytical methods and findings are then presented narratively using thick description rather than through numbers [19]. Thick description presents the findings as they were interpreted or explained by the authors as opposed to simply providing descriptive summaries of each study [20]. This provides the opportunity to translate the findings into a richer, more complete understanding of a phenomenon [21]. We included studies that reported qualitative methodologies and utilized qualitative methods for data analysis. Studies that collected data using qualitative methods but did not use qualitative analysis (e.g., surveys with open-ended questions that used descriptive statistics) were excluded. Mixed methods studies were included if it was possible to retrieve findings exclusively from the qualitative research.

Identification of studies

The studies used for our meta-synthesis were derived from two sources. The first source was a dataset [22] generated from a scoping review on research data management in academic institutions [23] which provided records from inception to April 2016. The purpose of the scoping review was to describe the volume, topics, and methodological nature of the existing research literature on research data management as it specifically related to academic institutions. The search strategy included the terms data sharing, sharing research, data reuse, and research reuse, along with spelling variations and wildcards to ensure all relevant records were captured. The second source for data came from re-running the literature searches from the scoping review with the addition of a validated qualitative search filter [24] in order to retrieve current records from April 2016 to October 2018. When the searches were conducted for the update, four of the original literature databases were unavailable and were replaced with comparable platforms upon consultation with subject matter specialists. Both the original and the updated search included a total of 40 literature databases representing a wide range of disciplines. The search strategy for MEDLINE can be found in S2 Appendix and the full search strategies for the other databases can be obtained by contacting the author. S3 Appendix lists all the literature databases searched. As well, the grey literature, conference proceedings, and a search of the reference list of all included studies were completed. No restrictions were placed on publication date or language. Thus, a comprehensive search of the literature was conducted.

Eligibility criteria

We aimed to identify all studies that investigated the views, perspectives, and experiences of academic researchers with data sharing and reuse. Studies were included if they were original research and reported qualitative methodologies, specifically focus groups or interviews. Studies had to include researchers (full- or part-time) conducting studies in academic institutions. We defined an academic institution as a higher education degree-granting organization dedicated to education and research. If studies included a mixed population, 50% or more of the total sample had to be researchers from academic institutions in order to be eligible for inclusion. Data sharing is defined as the practice of making data available for reuse [25]; reuse is defined as the use of content outside of its original intention [25]. Examples of this include depositing data into a digital repository or publishing raw data. Mixed methods studies that used both qualitative and quantitative methods within the same study were eligible if the qualitative portion met our inclusion criteria.

Study selection

Two investigators independently screened all records from the scoping review dataset in order to identify qualitative studies that met the eligibility criteria. The records from the updated search were assessed for eligibility by two investigators independently at level 1 (title and abstract) and level 2 (full-text) screening. Discrepancies were resolved by discussion or by a third investigator at every phase of study selection.

Quality appraisal

The CASP (Critical Appraisal Skills Programme) Qualitative Checklist [26] was used for quality appraisal of all included studies. The CASP Qualitative Checklist is a 10-item checklist that examines three domains: validity of results, reporting of the results, and value of the research (S4 Appendix). Each study was assessed independently by two investigators and discrepancies were resolved by discussion.

Data abstraction and analysis

The authors read and reread all articles to become familiar with each study [2728]. A data abstraction sheet was developed that included the study characteristics (e.g., year study conducted), participant characteristics (e.g., sample size), and key concepts. Key concepts or interpretations included all findings along with associated quotes from study participants [29]. Thematic analysis using constant comparison was used for data synthesis [30]. An initial set of 10 studies were coded independently by two investigators. These codes were then compared and refined in order to create an initial set of codes that were used going forward. Two investigators continued to independently code the remainder of the studies in duplicate and the team met regularly to discuss and iteratively refine the codes. All discrepancies were resolved through discussion or by a third member of the team. Finally, analytical themes were generated that offered a higher level interpretation beyond a descriptive synthesis [27,31]. The codes were grouped and categorized by making comparisons across articles in order to ensure that we appropriately captured similar themes from multiple studies. Meetings were used to review all constructs and resolve discrepancies, resulting in a refinement of the analytical themes.

Results

Forty-five studies and three companion reports were included in the review (Fig 1). Included studies are listed in S5 Appendix.

Fig 1. Study flow diagram.

Fig 1

Study characteristics

The studies included in the review were published during a 15 year period between 2003 and 2018. The most studies were published in 2014 (11 out of 45) and the method for data collection were interviews (37), a combination of interviews/focus groups (5), and focus groups (4). Over half of the studies (27 out of 45) were conducted in the United States. Table 1 provides a summary of study characteristics.

Table 1. Characteristics of included studies.

Study Country Type (no.) of participants Data collection Methodology Principal experiences explored
Allard 2012 [32] Turkey • Environmental scientist (10)
• Information science academic (2)
Interviews Grounded theory; Analytic induction Understanding knowledge and attitudes of information science and environmental towards scientific data and information
Bamkin 2014 [33] United Kingdom Participant (9) Focus Groups NR Identify the opinions of potential users of a policy databank service
Broom 2009 [34] Australia • Education (12)
• Sociology (5) Anthropology (5)
• Social Work (5)
• Public Health (5)
• Psychology (2) Journalism (1)
• Politics (2)
Focus Groups Interpretive qualitative approach Explore the perceived challenges posed by contemporary innovations in data management, access, and analysis through electronic archiving
Carlson 2013 [35] USA • PhD student (5)
• Masters student (1)
Interviews NR Understand grad students practices with data, the challenges they face, and their attitudes toward managing and sharing data
Cheah 2015 [36] Thailand • Interview participant (15)
• Focus group participant (10)
• Interviews
• Focus Groups
NR Understand attitudes and experiences of relevant stakeholders about what constitutes good data sharing practice
Colledge 2014 [37] Switzerland Stakeholders (includes clinicians, pathologists, lawyers, ethicists, and biobank managers, researchers) (36) Interviews Classical qualitative method Opinions regarding getting consent for sharing samples with biobanks, the role of ethics committees
Cragin 2010 [38] USA • Agronomy and soil science (5)
• Anthropology (3)
• Earth and atmospheric science (2)
• Geology (3)
• Horticulture
• and plant science (2)
Interviews NR Investigate how data-related scholarly activities vary among disciplines and research communities
Delasalle 2013 [39] United Kingdom Participant (8) Interviews NR Charts the steps taken and possible ways forward to research data management, providing a typical example of a UK research university’s approach in two strands: requirements and support
Denny 2015 [40] South Africa • Community research support team (2)
• Junior research staff (10)
• Research manager (4)
• Senior researcher (10)
• Policy and department manager (3)
• Executive member (3)
• Interviews
• Focus Groups
Grounded theory; Thematic framework approach Examine the perceptions, experiences and concerns of research
stakeholders about data-sharing practices
Diekmann 2012 [41] USA Participant (14) Interviews NR Examine data practices of agricultural scientists
Faniel 2010 [42] USA • Assistant professor (4)
• Associate professor (2)
• Full professor (6)
• PhD student (1)
• Postdoctoral student (1)
Interviews NR Examine how earthquake engineer researchers assess the reusability of colleagues’ experimental data for model validation
Faniel 2013 [43] NR Participant (22) Interviews NR Examine the needs of archaeological data re-users, particularly the context they need to understand, verify, and trust data others collect during field studies
Finn 2014 [44] Multiple countries in Europe, USA Participant (5) Interviews NR Identify legal and ethical issues relevant to open access to research data, identify examples that illuminate these issues, and identify potential solutions currently being used to address these issues
Frank 2015 [45] USA Archaeologists (22)
Zoologists (27)
Interviews Not explicitly stated but declares, “combined deductive and inductive approaches” • Practices and norms affect how archaeologists and zoologists view/understand preservation as it relates to their own research data
• External factors influencing attitudes of archaeologists and zoologists toward the feasibility of long-term preservation of research data
Hall 2013 [46] USA Environmental science (14) Interviews Phenomenological approach Determining where metadata re-use is most common or lacking
Henty 2008 [47] Australia NR • Focus Groups
• Interviews
NR Needs related to and provisions of data management infrastructure
Higman 2015 [48] United Kingdom Participants (librarians, research
office staff and IT professionals) (11)
Interviews Interpretivist perspective • Relationship between research data management (RDM) and data sharing in formulations of RDM policies
• Clarify what is influencing decisions, how different actors are behaving and how
networks are being formed
Hunt 2018 [49] USA Professor (12)
Associate professor (8)
Assistant professor (4)
Interviews Grounded theory To assess the comprehensive information science needs and behaviors of public health research faculty
Johnston 2014 [50] USA Faculty (1)
Graduate student (4)
Interviews NR Needs and data management skills required by graduate student in engineering field
Johri 2016 [51] USA Associate professor (2)
Assistant professor (2)
Graduate students (2)
Interviews NR To get better insights into the current state of data sharing in engineering education and what needs to be done if data sharing is to be supported
Kervin 2012 [52] PhD student (3)
Post-doc (1)
Professor (1)
Interviews NR How researchers handled data in a research project from start to finish
Kim 2012 [53] USA Tenured (full and associate) professors (11)
Assistant professors (8)
Emeritus professor (1)
Professor of practice (1)
Post-doctoral research
Associates (2)
Doctoral candidates (2)
Interviews Inductive approach (mentioned with regards to coding scheme) Sharing practices in diverse fields and factors motivating or preventing data sharing
Lage 2011 [54] USA NR Interviews Ethnographic Represent the range of attitudes and needs regarding the type of datasets created, existing data storage and maintenance support, disciplinary culture or personal feelings on data sharing, and receptivity to the library’s role in data curation
Manion 2009 [55] USA Participant (24) • Interviews
• Focus Groups
NR Collect policy statements, expectations, and requirements from regulatory decision makers at academic
cancer centers in the United States; use these statements to examine fundamental assumptions regarding data sharing using data federations and grid computing
Marcus 2007 [56] USA Interview participant (7)
Focus group participant (NR)
Interviews NR Capture the practical and conceptual challenges of research in the sciences
McGuire 2012 [57] USA Investigators (63)
Recruits (50)
Interviews Thematic content analysis Explore core ethical, legal, and social implication issues that arose during the first phase of the Human Microbiome Project from the perspective of individuals involved in the research
McLure 2014 [58] USA Researcher (31) Focus Groups Thematic analysis Understanding nature of researcher data sets, their management, need for assistance/support, library support
Murillo 2014 [59] USA Participant (14) Focus Groups Inductive content analysis Scientists’ perceptions on the topic of data at risk; re-use/sharing;
Data At Risk Inventory
Noorman 2014 [60] United Kingdom Data centre manager, project coordinator,
librarian (15)
Interviews NR • Focus on challenges faced by institutions, such as
archives, libraries, universities, data centers and funding bodies, in making open access to research data possible
• Explore current strategies, the remaining barriers and possible solutions for overcoming these barriers
Ochs 2017 [61] USA NR Interviews NR To examine various aspects of the research life and process of faculty and research staff in the agriculture discipline
Oleksik 2012 [62] United Kingdom Professor (1)
Post-docs (3)
PhD students (5)
Interviews Thematic analysis Understand the interdependencies of technologies, practices, and artifacts that emerge as part of the scientific activities
Pepe 2014 [63] USA Astronomers (12) Interviews NR Gather a first-hand account of the needs and challenges of data referencing and archiving in astronomy
Read 2015 [64] USA Basic scientists (11)
Clinical researchers (19)
Interviews Grounded theory Obtain information to plan data-related products and services
Stamatolos 2016 [65] NR Faculty (14) Interviews Inductive approach Seek an understanding of the thinking and practices of a small, but diverse population of faculty researchers regarding data management
Stapleton 2017 [66] USA Professors (6)
Associate professors (4)
Assistant professors (4)
Non-tenure track research assistant (1)
Interviews Grounded theory The research practices of academics in agriculture in order to understand the resources and services these faculty members need to be successful in their teaching and research
Sturges 2014 [67] NR Participant (23) Focus Groups Grounded theory Views and practices of stakeholders to data sharing
Valentino 2015 [68] USA Graduate student (5) Interviews NR Allow students to explain their research covering the areas of data analysis, storage, organization, and format, and data back-up practices
Van den Eynden 2014 [69] Europe (Denmark, UK, Germany, Netherlands, Finland) Participant (22) Interviews Comparative analysis Data sharing practices and motivation for data sharing
Van Tuyl 2015 [70] USA NR Interviews NR Formalize assessment of research data management practices of researchers at the institution by launching a faculty survey and conducting a number of interviews with researchers
Wallis 2013 [71] USA Participant (43) Interviews NR Motivation for sharing data; conditions placed on data that is shared;
sharing data with others outside own research group; data used that were not generated by a researcher’s own group; how are data from external sources is used
Williams 2013 [72] USA Assistant professor (3)
Associate professor (1)
Full professor (3)
Interviews NR Summarize the participants' reasons for making data publicly available but also describes the challenges that they faced when sharing data
Yatcilla 2017 [73] USA Agricultural & Biological Engineering (3)
Agricultural Economics (1)
Agronomy (4)
Botany & Plant Pathology (1)
Entomology (3)
Forestry & Natural Resources (1)
Youth Development & Agricultural Education (3)
Agricultural Administration (1)
Interviews NR To understand the resources and services these faculty members (agriculture) need to be successful in their research and teaching
Yoon 2014 [74] USA Faculty (17)
Research associate (2)
Interviews Inductive approach Enhance our understanding of trust in repositories from the users’ point of view
Yoon 2017 [75] USA PhD students
Postdocs Professors
Research scientists
No numbers provided
Interviews Interpretive qualitative approach To investigate reusers’ trust beyond trust formation and tracks those changes to trust that happen during the experiences of using data
Zimmerman 2003 [76] NR Ecologists (13)
Data managers (4)
Interviews Inductive approach Experiences of ecologists who use shared data

Findings

We identified four major themes and several sub-themes. The four major themes were data integrity, responsible conduct of research, feasibility of sharing data, and value of sharing data. Themes and sub-themes along with illustrative quotes are summarized in Table 2. Each theme is described below and details are provided for each sub-theme.

Table 2. Themes derived and illustrative quotes.

Theme and sub-theme Illustrative quotes Reference
Theme: Data Integrity
Data quality There are definitely different comfort levels for people. Some people will forever be confined to studying their own system because they are unable to accept any degree of, you know, sort of taking other people’s word—sort of dealing with data that they didn’t actually see collected themselves. Zimmerman 2003 [76]
What had been reported, what had been presented and discussed were, kinda, the best view of the data. [I]n reality, the data did have some problems that weren’t apparent until you got deeply inside and started looking. Yoon 2017 [74]
Data documentation …a lot of the contextual data that you need is not provided. Faniel 2013 [43]
You can tell from the documentation whether or not a research[er] was thorough and careful. Yoon 2017 [75]
It’s so easy to generate this digital data, but if you’re not careful how you name things and how you document stuff and making sense of it later, particularly for someone else, is going to be a real challenge. Yatcilla 2017 [73]
What is worth sharing Am I worried it won’t be there in 20 years? No. Am I worried it won’t be there in 100? It doesn’t matter. By that point, data become irrelevant except as historical curiosity. Marcus 2007 [56]
Biospecimens are very valuable because they were collected before the disease, so they’re good for looking at developing disease…I think it could be used for many years. Read 2015 [64]
Theme: Responsible Conduct of Research
Misuse of data …my main concern is I don’t want people to misuse it … and if I don’t have some relationship of trust then I don’t know whether they’re going to, you know, just go off and do something and never check with me to see, well, was this a good interpretation. Cragin 2010 [38]
…a whole cadre of people whose only job is pilfering other people’s stuff, or parasitically using it. Hunt 2018 [49]
Work culture I completed an NSF grant in December and… you have to have now a section that describes what you are going to do with your data…Data availability and where you’re going to archive it… So you’re being forced to deal with it now whereas in the past you’re like, ‘Well it’s in my file cabinet. Frank 2015 [45]
I think perhaps it’s just tradition or it’s a thing of the past where people have held their data somewhat closely… Ochs 2017 [61]
Protecting one’s own work / Intellectual property We all collect samples together in the field, but when you come back to process the samples, people want the data without any understanding or agreement about ownership. Marcus 2007 [56]
But it’s also the notion of intellectual property, isn’t it? … How are we going to know if other people are picking it up and using it elsewhere, unless they’re being absolutely… Broom 2009 [34]
Control of data If someone were to use the data would be good to know, what did they do with it, some form of communication… Johri 2016 [51]
You would have to describe your intended use of the data. And then the people who originally were the researchers who gathered that data, would all have to agree to consent to each application. And so they still retain the control of the data. Finn 2014 [44]
Privacy/Confidentiality/Ethics If the systems are such that they can get into our data, we might need to think for the first time about being a little bit more circumspect and think about what qualifications we would want to impose … I think there would probably be a lot of regulatory compliance pieces we might want to spell out more than we do now. Manion 2009 [55]
…we can never actually, never guarantee confidentiality of all data, because it could be hacked into and we can’t anymore say that your data will be anonymous because that is nonsense too, because we are able to bring in so many different kinds of data, … that the potential for people to be re-identified or distinguished in the data are quite high… Finn 2014 [44]
Theme: Feasibility of Sharing Data
Infrastructure I do think that from an institutional level there should be a governing body to provide guidance and to enforce policy, and to make policy for all the systems that will interact and handle activity with other institutions. As far as what functions they would dictate, [they] would be all around the authorization, authentication, and accounting of access to that data. Manion 2009 [55]
It’s very easy to see how having a central, university wide, storage and dissemination system for data would be much more cost effective, and probably better executed, than anything we could do ourselves. McLure 2014 [58]
Time/work required If there's someone in the institute who can [deposit data], instead of individual researchers, that would save lots of our time and [we could] be more productive… Williams 2013 [72]
To be quite honest, the biggest hurdle when you’re dealing with genetic data in like depositing … the information and the sequence data onto GenBank is associating that with museum specimens or locality data ….It’s really kind of clunky and it really takes a lot of time to do that. Frank 2015 [45]
Skills We are not thinking too much about data management. We are thinking more about the approach and methodology… Diekmann 2012 [41]
They are resistant to having to learn how to
use new tools that make open data and reproducibility easier. They generally kind of
just have their process and they feel like they're tested already in terms of their time
and their commitment and they don’t really want to add this to the list of things that
they have to worry about.
Noorman 2014 [60]
Theme: Value of Sharing Data
Promote future discovery …there is no sense in collecting data if it can't be used [by other researchers]. Lage 2011 [54]
We truly believe that sharing data is the right thing to do, simply because the original data we used for this study was not ours. Our study was only possible because other astronomers made their data publicly available in the first place! Pepe 2014 [63]
Researcher perspective To incentivize data sharing there should be follow-on grants on data analysis and dissemination grant to bring other researchers on board. If NSF changed their model for a year, there is a lot of data out there. I think there has to be some stipulation about who gets authorship when the data is used but I think funding to bring new people on board is essential. There can also be a solicitation focused on secondary analysis. Johri 2016 [51]
I think one barrier to data sharing is the merit review process within institutions for tenure and promotion; things such as ‘how many people accessed your dataset’ are not valued. Johri 2016 [51]

Data integrity

The theme data integrity addresses researcher’s perspectives on data that are available from repositories and their expectations around the prospects of reusing this data based on the quality, documentation available, and what individual researchers deemed as worthy of sharing.

  • Data quality: Researchers acknowledged that although there may be interest or willingness to consider using open data, there would always be people that would not trust the quality of a dataset unless they collected it themselves. They recognized the need to manage expectations with some suggesting that lowering their standards related to data quality may help with increasing the likelihood of being able to use another researcher’s dataset. Nuances around the conditions, context, or materials that are not normally recorded were identified as one of the challenges to data quality. One example was offered in the field of engineering where equipment may not be anchored solidly and produce a variation that would impact the quality of data outputs [42]. The range of skill levels of researchers (e.g., junior versus senior researchers) was flagged as potentially affecting the quality of data collected and it was noted that there was no way of knowing this when reusing a dataset. Similarly, the quality of datasets may also vary depending on the person’s intentions or purpose when collecting data. It was felt that data collected with the intention of reporting only to people within their own discipline may look different from data collected for external groups, (i.e., these datasets may include more details).

  • Data documentation: For a dataset to be truly reusable, researchers indicated adequate documentation was necessary including a significant amount of detail and metadata. The importance of contextual information was noted as providing layers of information that offered necessary insight. Providing this was seen as a time- and energy-intensive endeavor for the researcher collecting the data and making it available for reuse. Comprehensive documentation signaled a reliable dataset to researchers that were looking for datasets to reuse.

  • What is worth sharing: When considering their data, researchers varied on what was worth sharing. It spanned from believing the preservation of datasets to be a top priority, to complete lack of interest. For those that felt their data was not worth sharing, they believed their data would be irrelevant after a period of time and nothing more than a “historical curiosity” if it was offered for reuse [56]. Others thought their data had the potential to be useful but had clear views on what was worth sharing and felt it had to have ‘scholarly value’. As an example, researchers described biospecimens they had collected as valuable since it provided the opportunity to look at the development of a disease [64].

Responsible conduct of research

The responsible conduct of research emerged as a theme that encompassed the professional standards, ethical principles, and tacit norms that researchers described when considering data sharing. The five sub-themes under this theme are the misuse of data, protecting one’s own work/intellectual property, privacy/confidentiality/ethics, control of data, and work culture.

  • Misuse of data: Researchers expressed concern about the potential for the inappropriate use of their data. This included what was termed as ‘fishing expeditions’ which involved dredging data with no particular research question in the hopes of stumbling upon possible relationships that could be presented as convincing results. The potential for data to be misunderstood and thus produce inappropriate or misguided conclusions was also considered a possibility, even if researchers reused datasets with a focused research question. The people reusing data were given names such as ‘free riders’ [42] and it was believed that misunderstood data could lead to false conclusions and ultimately threaten the original work related to the data.

  • Protecting one’s own work/intellectual property: Clarity around who owns data, along with intellectual property rights, were raised as issues when sharing data. Collaborations were mentioned as making it difficult to determine ownership since multiple people and institutions were involved. Licensing data was seen as both a potential solution, as well as a potential barrier (e.g., the cost could mean it would not be accessible to all) for providing access to research data. Since data had publication value, this was considered a major deterrent to sharing data. This obvious connection between publications and data made researchers feel that data needed to be protected. Publications were seen as a key research output with a relationship between this and future funding.

  • Privacy/confidentiality/ethics: Privacy and confidentiality were taken seriously when researchers considered their data, particularly when it came to human subjects. When data was shared between institutions, the precautions undertaken were complex. This included considerations such as de-identification, re-identification risks (along with the potential for this to happen unintentionally), consent (e.g., whether re-consent was necessary), and the challenge of future use if the purpose for future use was not pre-defined. Precautions implemented included each institution independently obtaining ethics approval before data was exchanged. Views were divergent around whether data should be freely accessible with some favoring restrictions on the re-use of data and others indicating it should be freely accessible.

  • Control of data: It was acknowledged that the relationship between research being publicly funded and making data available for public benefit had merit. However, some felt they would like to know who was using their data and for what purpose. It was considered important to have a relationship with the person who wished to reuse data. As well, access should be controlled and only given to those that could be identified as a research professional who were qualified to do research. The level of control ranged from wanting systems in place that would allow them to monitor data sharing with people who were not known to them personally (although this was acknowledged as labor-intensive), to simply wanting to have a list of who was using their data and for what purpose as a minimum level of communication. The need to protect data until publication was considered a deterrent to sharing data and it was necessary to have control over data until this was completed.

  • Work culture: Work culture highlights the beliefs of how research should be conducted that are influenced by shared attitudes, views, and written/unwritten rules developed over time. These normative values were described as being taught to junior researchers by senior academics. In the past, the cultural norm was to rely on informal processes, such as personal relationships, for sharing data. Usually, these were people who were known and trusted either through direct contact or by reputation. As new requirements were introduced by funding agencies and journals, researchers observed changes in their practice of sharing data over time. There was an acknowledgement that a shift in culture that favored a more open view of data was needed. It was also noted that even if researchers were to understand the benefits of sharing data, this transition would not be immediate and that incentives must be identified for researchers in this process.

Feasibility of sharing data

The feasibility of sharing data examines the ease with which researchers can make their data available to others, along with the related barriers and facilitators. Infrastructure, time/work required, and skills are the three sub-themes that emerged and are described below.

  • Infrastructure: Researchers described the structures and supports that would help ensure data sharing. Infrastructure support included data storage, file migration, and funding for making data available. These challenges involved both local (e.g., individual research labs) and institution-wide settings. Data handling was a fragmented activity managed by researchers who devised their own independent strategies that generally lacked sustainability. One example was the number of people (i.e., students, staff) that rotated through a research lab where everyone was responsible for their own data [62]. Although the lab manager encouraged best practices (e.g., including sufficient documentation), this did not necessarily translate into being adopted and applied. As a result, it was not guaranteed that datasets could be easily shared or be accessible in the future. Appropriate data storage infrastructure and support were associated with good data management, which in turn laid the groundwork for data sharing. An institution-level policy to support data sharing, along with resources, were identified as important to ensure good quality data being deposited and made available.

  • Time/work required: The effort to prepare data for sharing was seen as time-consuming, expensive, and labor-intensive. Barriers included the lack of time to organize the necessary documentation, challenges with repository interfaces, and the lack of resources. For those that chose to offer their data upon request, the administrative aspect of filling requests for data was considered an added burden.

  • Skills: For many disciplines, data sharing was a new activity that was typically imposed by funding agencies or journals. As a result, researchers were looking for services or resources that would help with this task. The lack of technical skills and knowledge included how to anonymize data, how to create metadata, and unfamiliarity with depositing data into repositories. It was felt that providing open access to data was complex. Providing adequate support may not be feasible given that each discipline had a variety of data types, different amount of data being generated, disparities in what is considered data, and varying norms in research culture.

Value of sharing data

The value of sharing data theme describes researchers’ views on the importance placed on making data available to others. While the sub-theme promote future discovery identifies a benefit to society with sharing data, researchers’ perspectives focused on the benefits to researchers themselves.

  • Promote future discovery: The importance of making data accessible for possible use in the future was understood as a benefit. Those that described proactively sharing data (before they were required to) also noted the importance of sharing computer code as well. There was recognition that research funded by public money should be open and available. It was felt that taxpayers provided an investment and the public deserved a return on their investment. In some instances, researchers were able to identify examples of data sharing that helped promote scientific progress, such as the development of a new drug or containment of a disease. It was felt that data sharing had the potential to move a field forward by closing knowledge gaps and further opening new avenues of investigation.

  • Researchers’ perspectives: Data was identified as a research product that helped achieve a goal such as completing a publication and there was the recognition that amongst researchers that data sharing would provide greater accountability and transparency. For those that were already reusing data, its value was recognized as helpful for writing proposals and training students. The importance of providing incentives for sharing data was emphasized with researchers unable to identify significant benefits. Suggestions included creating grants that focused specifically on the reuse of data generated from earlier grants.

Quality assessment

The CASP tool, used to assess the quality of studies, identified 27 (out of 45) studies that had seven out of the ten items present. Most studies adequately addressed the methods (41 out of 45 studies) and aims (40 out of 45 studies) (Fig 2). Author reflexivity, which asked if the relationship between the researcher and participants was adequately considered, was not apparent in any of the studies. No studies were excluded due to a low score as this may have eliminated those with relevant and insightful results [7778].

Fig 2. Quality appraisal of included studies.

Fig 2

Discussion

We conducted a comprehensive review that included 45 studies along with 3 companion reports on the views, perspectives, and experiences of academic researchers on sharing their research data. The National Institutes of Health (NIH) in the United States were one of the first funding agencies to introduce a policy on sharing research data in 2001 [4]. This aligns with beginning to see research published on this topic starting in 2003, along with over half of the studies being conducted in the United States. Our results show that some of the themes and sub-themes offer positive support for sharing data however, most highlight areas of discomfort for researchers. In particular, researchers identified concern with issues related to data quality, misuse of data, protecting data, lack of time and skills, and deficiencies in infrastructure and support.

By default, researchers believed the quality of datasets available for reuse were poor and there is support for this in the literature. Studies assessing data available in public repositories have found incomplete datasets, saved in a way that compromised reuse [7980]. Researchers who felt their data had value describe using a tacit set of criteria to determine if it had ‘scholarly value’ [38]. These criteria are based on discretion and would vary from person to person thus adding further to factors that affect the quality of datasets. For researchers who felt their data was not worth sharing, this may be reflected in how they prepare their data for depositing into repositories (e.g., providing poor documentation) and ultimately its final quality.

The lack of supportive infrastructure, lack of time, and skills deficit had an influence on data quality as well as data availability. Researchers indicated that the lack of time and skills impacted the production of sufficient data documentation, creation of suitable metadata, and appropriately anonymized data. They also lacked skills in navigating repository interfaces in order to deposit data. While training and education may address these issues [81], a more effective pathway is to focus energy and resources on creating user-friendly interfaces that allow users to accomplish their goal of depositing datasets as quickly and easily as possible [82]. At an institutional level, the lack of procedures, policies, and guidance contributed to challenges in sharing data. This was particularly true for sensitive data that requires more vetting and scrutiny before sharing. Solutions for this include using a trusted party regulated by an ethics board that manages requests and maintains the de-identified records and original identifiers [55].

Our results show that a major concern of researchers is the possibility of misuse or misinterpretation of their data, and this is reported as well in surveys [69,8384]. Traditionally, research data has been shared through professional networks and by personal request [32,36,38]. These ‘traditions’ were incorporated into research processes as early-career researchers were indoctrinated by mentors and senior researchers [52]. This approach allowed those who owned datasets to scrutinize requests and all aspects of the requestor, including the reputation of their institution, their publications, and any other factors they felt important. Data producers had a hand in assuring their work and intellectual property were protected, privacy and confidentiality were safe, and it allowed them to exercise caution if there were any concerns around the misuse of their data, including the option to decline the request to share. Currently, funding agencies and journals are moving researchers in the direction of sharing data which is not embraced by all members of the research community [12,40,69,8384]. In a recent paper, Campbell and colleagues [85] identified senior researchers as less likely to support data sharing while their early-career colleagues were more willing to make their data available for reuse. Researchers describe shifting to a culture of open data as a gradual transition in our findings, and stage of career may contribute to this need for a gradual shift.

Incentives were also identified as necessary for researchers within the research process that promoted open data [47,51,86]. In 2016, more than 500 researchers that received grants from the Wellcome Trust (welcome.ac.uk) in the United Kingdom were surveyed and although over half indicated that they made their data available for reuse, few reported direct benefits [69]. The lack of benefits appeared in our results and were identified as necessary yet lacking in the realm of data sharing. Suggestions for incentives included offering research grants that focused specifically on the reuse of data generated from earlier grants [51], and creating systems that ensure credit is awarded to data generators [8790]. In one example, Pierce and colleagues [88] proposed creating enduring links between those who generate data and any time it was used in the future. This would involve linking persistent identifier (PIDs) to all datasets and provide infrastructure to link the identifiers to publications. In this strategy, data authorship would be listed on curriculum vitae, considered in academic institutions promotions criteria, and be considered by granting agencies as an element for review for funding.

There is a global movement towards openness in research that includes open data. Data sharing and reuse is a key part of this movement and anticipated benefits include promoting research transparency, verification of findings, and gaining new insights from re-analysis [8]. Despite this, it has not become a common practice [1316]. Investing in strategies that improve skills amongst researchers that focus on improving data integrity in repositories and identifying incentives that provide motivation for data sharing are essential.

Limitations

Quality assessment indicated that some items in the CASP tool were addressed poorly in the studies. This included author reflexivity, analysis, and ethical issues. Limitations set by journals (i.e., word counts) may restrict authors from providing rich data and thick descriptions which are characteristic of qualitative studies. Studies based on low reporting quality were not excluded as this may have eliminated those with highly relevant and insightful results [77] and were used to judge the relative contribution in developing explanations in the study findings.

The qualitative data collected for this review originates from multiple disciplines and each may use a variety of data collection methods and research processes. However, when examining the studies by discipline, over a third of the studies in our review are listed as ‘combined’ (37% or 17 out of 45) [91] (i.e., participants came from multiple disciplines) yet none of the authors reported this as an issue in their analysis or impacting their results (S6 Appendix). Similarly, one of the study authors (LP) conducted focus groups with academic researchers in the area of research data management (including data sharing) and found that data saturation was reached after conducting four focus groups despite collecting data from discrete disciplines (i.e., health science, humanities, natural science) [81]. While diverse tools and methods may be employed by researchers in distinct disciplines to conduct their studies, issues related to research data management were identified as a commonality [32,38,39,4445, 47,5254,56,58,5960,65,7071,75,81]

Most of the studies accepted into our review are interviews (82% or 37 out of 45 studies). While the group setting of a focus group may prompt ideas and memories from group members by listening to other participants [92], interviews provide the opportunity to go deeper into a topic and gather in-depth information [93]. When Guest and colleagues [93] performed a randomized controlled trial comparing focus groups and interviews, they found that individual interviews were more effective at generating a broad range of items at an individual level [93].

Conclusions

Misuse and misinterpretation of data is a significant concern amongst researchers when sharing their data. Preparation of data so that it is truly reusable requires an investment in time and resources as well as skills that researchers indicate they lacked. Deficiencies in infrastructure may hamper sharing data effectively, particularly sensitive data. The availability of data is marked by researchers’ decision making around what they determine is worth sharing. Currently, there is a lack of incentives for researchers to share their data with regards to academic appointment, promotion, recognition, and rewards. As such, enhancements need to be considered that focus on providing direct benefits to researchers who share their data. Identifying appropriate incentives may help improve motivation to share data and enhance the integrity of data put into repositories.

Supporting information

S1 Appendix. PRISMA checklist.

(DOC)

S2 Appendix. MEDLINE search strategy.

(DOCX)

S3 Appendix. Literature databases searched.

(DOCX)

S4 Appendix. CASP (critical appraisal skills programme): Qualitative checklist.

(DOCX)

S5 Appendix. Included studies.

(DOCX)

S6 Appendix. Included studies by discipline.

(DOCX)

Acknowledgments

We thank Jordan Raghunandan for collating information after the data was coded.

Data Availability

The data are available from the Zenodo Repository, DOI: doi.org/10.5281/zenodo.3258850.

Funding Statement

The authors received no specific funding for this work.

References

  • 1.Holdren JP. Increasing access to the results of federally funded scientific research. February 22, 2013. Office of Science and Technology Policy. Executive Office of the President. United States of America. Available at: https://obamawhitehouse.archives.gov/blog/2016/02/22/increasing-access-results-federally-funded-science Accessed October 11, 2019.
  • 2.OECD (Organization for Economic Co-Operation and Development). Declaration on access to research data from public funding. 2004. Available at: https://legalinstruments.oecd.org/en/instruments/157 Accessed October 11, 2019.
  • 3.DCC (Digital Curation Centre). Overview of funders’ data policies. Available at: http://www.dcc.ac.uk/resources/policy-and-legal/overview-funders-data-policies Accessed October 11, 2019.
  • 4.Shearer K. Comprehensive Brief on Research Data Management Policies. April 2015. Available at: http://web.archive.org/web/20151001135755/http://science.gc.ca/default.asp?lang=En&n=1E116DB8-1 Accessed October 11, 2019.
  • 5.Taichman DB, Sahni P, Pinborg A, Peiperl L, Laine C, James A, et al. Data sharing statements for clinical trials: a requirement of the International Committee of Medical Journal Editors. Lancet. 2017. June 10;389(10086):e12–e14. 10.1016/S0140-6736(17)31282-5 [DOI] [PubMed] [Google Scholar]
  • 6.Butler D. Zika researchers release real-time data on viral infection study in monkeys. Nature. 2016; 530:5 10.1038/530005a [DOI] [PubMed] [Google Scholar]
  • 7.Bill & Melinda Gates Foundation. Open Access Policy. Available at: https://www.gatesfoundation.org/How-We-Work/General-Information/Open-Access-Policy Accessed October 11, 2019.
  • 8.PolicyWise for Children and Families. The current state of data sharing. July 2016. Available at: https://policywise.com/wp-content/uploads/SAGE/Current-State-of-Data-Sharing-Funders-2016-06JUN-30.pdf Accessed October 11, 2019.
  • 9.Hayes J. The data-sharing policy of the World Meteorological Organization: The case for international sharing of scientific data In: Mathae KB, Uhlir PF, editors. Committee on the Case of International Sharing of Scientific Data: A Focus on Developing Countries. National Academies Press; 2012. p. 29–31. [Google Scholar]
  • 10.Ivezic Z. Data sharing in astronomy In: Mathae KB, Uhlir PF, editors. Committee on the Case of International Sharing of Scientific Data: A Focus on Developing Countries. National Academies Press; 2012. p. 41–45. [Google Scholar]
  • 11.Dijkers MP. A beginner's guide to data stewardship and data sharing. Spinal Cord. 2019. March;57(3):169–182. 10.1038/s41393-018-0232-6 [DOI] [PubMed] [Google Scholar]
  • 12.Longo DL, Drazen JM. Data sharing. New England Journal of Medicine. 2016; 374:276 10.1056/NEJMe1516564 [DOI] [PubMed] [Google Scholar]
  • 13.Naudet F, Sakarovitch C, Janiaud P, Cristea I, Fanelli D, Moher D, et al. Data sharing and reanalysis of randomized controlled trials in leading biomedical journals with a full data sharing policy: survey of studies published in The BMJ and PLOS Medicine. BMJ. 2018. February 13;360:k400 10.1136/bmj.k400 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Savage CJ, Vickers AJ. Empirical study of data sharing by authors publishing in PLoS journals. PLoS One. 2009. September 18;4(9):e7078 10.1371/journal.pone.0007078 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Vanpaemel W, Vermorgen M, Deriemaecker L, Storms G. Are we wasting a good crisis? The availability of psychological research data after the storm. Collabra. 2015;1(1):1–5. [Google Scholar]
  • 16.Federer LM, Belter CW, Joubert DJ, Livinski A, Lu YL, Snyders LN, et al. Data sharing in PLOS ONE: An analysis of Data Availability Statements. PLoS One. 2018. May 2;13(5):e0194768 10.1371/journal.pone.0194768 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Ioannidis JP, Allison DB, Ball CA, Coulibaly I, Cui X, Culhane AC, et al. Repeatability of published microarray gene expression analyses. Nat Genet. 2009. February;41(2):149–55. 10.1038/ng.295 [DOI] [PubMed] [Google Scholar]
  • 18.Sandelowski M, Barroso J. Classifying the findings in qualitative studies. Qual Health Res. 2003. September;13(7):905–23. 10.1177/1049732303253488 [DOI] [PubMed] [Google Scholar]
  • 19.Glenton C, Lewin S. Using evidence from qualitative research to develop WHO guidelines WHO Handbook for Guideline Development. 2nd Edition. Geneva: WHO, 2014. [Google Scholar]
  • 20.Sandelowski M, Barroso J. Reading qualitative studies. International Journal of Qualitative Methods. 2002;1(1):74–108. [Google Scholar]
  • 21.Sherwood G. Meta-synthesis: merging qualitative studies to develop nursing knowledge. International Journal for Human Caring. 1999;3:37–42. [Google Scholar]
  • 22.Perrier L, Blondal E, Ayala AP, Dearborn D, Kenny T, Lightfoot D, et al. Dataset for: Research data management in academic institutions: a scoping review. 2017a. 10.5281/zenodo.557043 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Perrier L, Blondal E, Ayala AP, Dearborn D, Kenny T, Lightfoot D, et al. Research data management in academic institutions: A scoping review. PLoS One. 2017b;12(5): e0178261. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.McMaster University. McMaster PLUS Projects: Hedges. Available at: https://hiru.mcmaster.ca/hiru/hiru_hedges_home.aspx Accessed October 11, 2019.
  • 25.CASRAI. Dictionary. Available at: http://dictionary.casrai.org Accessed October 11, 2019.
  • 26.CASP. CASP Qualitative Checklist. Available at: https://casp-uk.net/wp-content/uploads/2018/01/CASP-Qualitative-Checklist-2018.pdf Accessed June 13, 2019.
  • 27.Thomas J, Harden A. Methods for the thematic synthesis of qualitative research in systematic reviews. BMC Med Res Methodol (2008) 8:45 10.1186/1471-2288-8-45 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Paterson BL, Thorne SE, Canam C, Jillings C. Meta-Study of Qualitative Health Research: A Practical Guide to Meta-Analysis and Metasynthesis. 2001, London: Sage. [Google Scholar]
  • 29.Schütz A. Collected papers 1. The Hague, Netherlands: Martinus Nijhoff; 1962. [Google Scholar]
  • 30.Miles MB, Huberman AM, Saldaña J. Qualitative Data Analysis: A Methods Sourcebook. Thousand Oaks, CA: SAGE, 2014. [Google Scholar]
  • 31.Lachal J, Revah-Levy A, Orri M, Moro MR. Metasynthesis: an original method to synthesize qualitative literature in psychiatry. Frontiers in Psychiatry. 2017; 8:269 10.3389/fpsyt.2017.00269 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Allard S, Aydinoglu AU. Environmental researchers' data practices: An exploratory study in Turkey. International Symposium on Information Management in a Changing World. Springer Berlin Heidelberg. 2012; 317:13–24. [Google Scholar]
  • 33.Bamkin M. Report of Findings from Focus Group and Online Questionnaire: The opinions of potential users of a policy databank service. JORD Project. 2014. Available at: https://jordproject.files.wordpress.com/2014/06/report-of-findings-from-focus-group-and-online-questionnaire.pdf Accessed June 19, 2019.
  • 34.Broom A, Cheshire L, Emmison M. Qualitative researchers' understandings of their practice and the implications for data archiving and sharing. Sociology. 2009; 43(6):1163–1180. [Google Scholar]
  • 35.Carlson J, Stowell-Bracke M. Data management and sharing from the perspective of graduate students: an examination of the culture and practice at the water quality field station. portal: Libraries and the Academy. 2013; 13(4):343–361. [Google Scholar]
  • 36.Cheah PY, Tangseefa D, Somsaman A, Chunsuttiwat T, Nosten F, Day NPJ, et al. Perceived benefits, harms, and views about how to share data responsibly: a qualitative study of experiences with and attitudes toward data sharing among research staff and community representatives in Thailand. Journal of Empirical Research on Human Research Ethics. 2015; 10(3):278–289. 10.1177/1556264615592388 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Colledge F, Persson K, Elger B, Shaw D. Sample and data sharing barriers in biobanking: consent, committees, and compromises. Annals of Diagnostic Pathology. 2014; 18(2):78–81. 10.1016/j.anndiagpath.2013.12.002 [DOI] [PubMed] [Google Scholar]
  • 38.Cragin MH, Palmer CL, Carlson JR, Witt M. Data sharing, small science and institutional repositories. Philosophical transactions. Series A, Mathematical, physical, and engineering sciences. 2010; 368(1926):4023–4038. 10.1098/rsta.2010.0165 [DOI] [PubMed] [Google Scholar]
  • 39.Delasalle J. Research data management at the University of Warwick: recent steps towards a joined-up approach at a UK university. LIBREAS. Library Ideas. 2013; 23. Available at: http://libreas.eu/ausgabe23/10delasalle Accessed June 19, 2019.
  • 40.Denny SG, Silaigwana B, Wassenaar D, Bull S, Parker M. Developing ethical practices for public health research data sharing in South Africa: the views and experiences from a diverse sample of research stakeholders. Journal of Empirical Research on Human Research Ethics. 2015; 10(3):290–301. 10.1177/1556264615592386 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Diekemann AR, Wesolek A, Walters CD. The NSF/NIH effect: surveying the effect of data management requirements on faculty, sponsored programs, and institutional repositories The Journal of Academic Librarianship. 2014; 40:322–331. [Google Scholar]
  • 42.Faniel IM, Jacobsen TE. Reusing scientific data: how earthquake engineering researchers assess the reusability of colleagues data. Computer Supported Cooperative Work. 2010; 19:355–375. [Google Scholar]
  • 43.Faniel I, Kansa E, Kansa SW, Barrera-Gomez J, Yakel E. The challenges of digging data: A study of context in archaeological data reuse. Proceedings of the 13th ACM/IEEE-CS Joint Conference on Digital Libraries. 2013; 295–304.
  • 44.Finn R, Wadhwa K, Taylor M, Sveinsdottir T, Noorman M, Sondervan J. Legal and ethical issues in open access and data dissemination and preservation. Recode Project. 2014. Available at: https://zenodo.org/record/1297492#.W3MvaM5KhhE Accessed June 19, 2019. [Google Scholar]
  • 45.Frank RD, Yakel E, Faniel IM. Destruction/reconstruction: preservation of archaeological and zoological research data. Archival Science. 2015; 15(2):141–167. [Google Scholar]
  • 46.Hall N. Environmental studies faculty attitudes towards sharing of research data. Proceedings of the 13th ACM/IEEE-CS Joint Conference on Digital Libraries. 2013; 383–384.
  • 47.Henty M, Weaver B, Bradbury SJ, Porter S. Investigating Data Management Practices in Australian Universities. 2008. Available at: http://eprints.qut.edu.au/14549 Accessed June 19, 2019.
  • 48.Higman R, Pinfield S. Research data management and openness: the role of data sharing in developing institutional policies and practices. Program: Electronic Library and Information Systems. 2015; 49(4):364–381. [Google Scholar]
  • 49.Hunt SL, Bakker CJ. A qualitative analysis of the information science needs of public health researchers in an academic setting. Journal of the Medical Library Association: JMLA. 2018. April;106(2):184 10.5195/jmla.2018.316 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.Johnston L, Jeffryes J. Data management skills needed by structural engineering students: case study at the University of Minnesota. Journal of Professional Issues in Engineering Education and Practice. 2014; 140(2):05013002. [Google Scholar]
  • 51.Johri A, Yang S, Vorvoreanu M, Madhavan K. Perceptions and practices of data sharing in engineering education. Advances in Engineering Education. 2016;5(2):n2. [Google Scholar]
  • 52.Kervin K, Finholt T, Hedstrom M. Macro and micro pressures in data sharing. IEEE 13th International Conference. 2012; 525–32.
  • 53.Kim Y, Stanton JM. Institutional and Individual Influences on Scientists’ Data Sharing Practices. Journal of Computational Science Education. 2012; 3(1):47–56. [Google Scholar]
  • 54.Lage K, Losoff B, Maness J. Receptivity to library involvement in scientific data curation: a case study at the University of Colorado Boulder. portal: Libraries and the Academy. 2011; 11(4):915–937. [Google Scholar]
  • 55.Manion FJ, Robbins RJ, Weems WA, Crowley RS. Security and privacy requirements for a multi-institutional cancer research data grid: an interview-based study. BMC Medical Informatics and Decision Making. 2009; 9:31 10.1186/1472-6947-9-31 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 56.Marcus C, Ball S, Delserone L, Hribar A, Loftus W. Understanding Research Behaviors, Information Resources, and Service Needs of Scientists and Graduate Students: A Study by the University of Minnesota Libraries. 2007. Available at: https://conservancy.umn.edu/handle/11299/5546 Accessed June 19, 2019.
  • 57.McGuire AL, Achenbaum LS, Whitney SN, Slashinski MJ, Versalovic J, Keitel WA, et al. Perspectives on human microbiome research ethics. Journal of Empirical Research on Human Research Ethics. 2012; 7(3):1–14. 10.1525/jer.2012.7.3.1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58.McLure M, Level AV, Cranston CL, Oehlerts B, Culbertson M. Data curation: a study of researcher practices and needs. portal: Libraries and the Academy. 2014; 14(2):139–164. [Google Scholar]
  • 59.Murillo AP. Data at risk initiative: examining and facilitating the scientific process in relation to endangered data. Data Science Journal. 2014; 12:207–219. [Google Scholar]
  • 60.Noorman M, Kalaitzi V, Angelaki M, Tsoukala V, Linde P, Sveinsdottir T, et al. Institutional barriers and good practice solutions. 2014. Available at: https://zenodo.org/record/1297494#.W3MwWc5KhhE Accessed June 19, 2019.
  • 61.Ochs M, Andrews C, Downs A, Morris-Knower J, Young S. Research practices and support needs of scholars in the field of agriculture at Cornell University. Journal of Agricultural & Food Information. 2017. July 3;18(3–4):200–19. [Google Scholar]
  • 62.Oleksik G, Milic-Frayling N, Jones R. Beyond data sharing: Artifact ecology of a collaborative nanophotonics research centre. Proceedings of the ACM 2012 conference on Computer Supported Cooperative Work. 2012; 1165–1174.
  • 63.Pepe A, Goodman A, Muench A, Crosas M, Erdmann C. How do astronomers share data? Reliability and persistence of datasets linked in AAS publications and a qualitative study of data practices among US astronomers. PloS One. 2014; 9(8):e104798 10.1371/journal.pone.0104798 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 64.Read KB, Surkis A, Larson C, McCrillis A, Graff A, Nicholson J, et al. Starting the data conversation: informing data services at an academic health sciences library. Journal of the Medical Library Association. 2015; 103(3):131–5. 10.3163/1536-5050.103.3.005 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 65.Stamatolos A, Neville T, Henry D. Analyzing the Data Management Environment in a Master's-level Institution. Journal of Academic Librarianship. 2016; 42(2):154–160. [Google Scholar]
  • 66.Stapleton S, Minson V, Spears L. Investigating the research practices of agriculture scholars: findings from the University of Florida. Journal of Agricultural & Food Information. 2017. July 3;18(3–4):327–46. [Google Scholar]
  • 67.Sturges P, Bamkin M, Anders JHS, Hubbard B, Hussain A, Heeley M. Research Data Sharing: Developing a Stakeholder-Driven Model for Journal Policies. Journal of the Association for Information Science and Technology. 2014; 66(12):2445–2455. [Google Scholar]
  • 68.Valentino M, Boock M. Data Management for Graduate Students: A Case Study at Oregon State University. Practical Academic Librarianship. 2015; 5(2):77–91. [Google Scholar]
  • 69.Van den Eynden V, Knight G, Vlad A, Radler B, Tenopir C, Leon D, et al. Towards Open Research: practices, experiences, barriers and opportunities. Wellcome Trust. 2016. Available at: 10.6084/m9.figshare.4055448 Accessed October 11, 2019. [DOI] [Google Scholar]
  • 70.Van Tuyl S, Michalek G. Assessing Research Data Management Practices of Faculty at Carnegie Mellon University. Journal of Librarianship and Scholarly Communication. 2015; 3(3):eP1258. [Google Scholar]
  • 71.Wallis JC, Rolando E, Borgman CL. If we share data, will anyone use them? Data sharing and reuse in the long tail of science and technology. PLoS One. 2013; 8(7):e67332 10.1371/journal.pone.0067332 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 72.Williams SC. Data sharing interviews with crop sciences faculty: why they share data and how the library can help. Issues in Science and Technology Librarianship. 2013. Available at: http://www.istl.org/13-spring/refereed2.html Accessed June 19, 2019.
  • 73.Yatcilla JK, Bracke MS. Investigating the needs of agriculture scholars: the Purdue Case. Journal of Agricultural & Food Information. 2017. July 3;18(3–4):293–305. [Google Scholar]
  • 74.Yoon A. End users’ trust in data repositories: definition and influences on trust development. Archival Science. 2014; 14(1):17–35. [Google Scholar]
  • 75.Yoon A. Data reusers' trust development. Journal of the Association for Information Science and Technology. 2017. April;68(4):946–56. [Google Scholar]
  • 76.Zimmerman AS. New knowledge from old data: the role of standards in the sharing and reuse of ecological data. 2008. September; 33(5):631–652. [Google Scholar]
  • 77.Campbell R, Pound P, Morgan M, Daker-White G, Britten N, Pill R, et al. Evaluating meta ethnography: systematic analysis and synthesis of qualitative research. Health Technology Assessment. 2011;15(43):1–164 10.3310/hta15430 [DOI] [PubMed] [Google Scholar]
  • 78.Atkins S, Lewin S, Smith H, Engel M, Fretheim A, Volmink J. Conducting a Meta-Ethnography of Qualitative Literature: Lessons Learnt BMC Med Res Methodol 2008;8(1):8–21. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 79.Roche DG, Kruuk LE, Lanfear R, Binning SA. Public Data Archiving in Ecology and Evolution: How Well Are We Doing? PLoS Biol. 2015. November 10;13(11):e1002295 10.1371/journal.pbio.1002295 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 80.Pope LC, Liggins L, Keyse J, Carvalho SB, Riginos C. Not the time or the place: the missing spatio-temporal link in publicly available genetic data. Mol Ecol. 2015. August;24(15):3802–9. 10.1111/mec.13254 [DOI] [PubMed] [Google Scholar]
  • 81.Perrier L, Barnes L. Developing research data management services and support for researchers: a mixed methods study. Partnership: The Canadian Journal of Library and Information Practice and Research. 2018;13(1). 10.21083/partnership.v13i1.4115 [DOI] [Google Scholar]
  • 82.Krug S. Don't make me think. Berkeley CA: New Riders, 2014. [Google Scholar]
  • 83.Weng C, Friedman C, Rommel CA, Hurdle JF. A two-site survey of medical center personnel's willingness to share clinical data for research: implications for reproducible health NLP research. BMC Med Inform Decis Mak. 2019. April 4;19(Suppl 3):70 10.1186/s12911-019-0778-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 84.Stuart S, Baynes G, Hrynaszkiewicz I, Allin K, Penny D, Lucraft M, et al. Practical challenges for researching in sharing data. White paper. March 2018. Available at: 10.6084/m9.figshare.5975011 Accessed June 11, 2019. [DOI] [Google Scholar]
  • 85.Campbell HA, Micheli-Campbell MA, Udyawer V. Early career researchers embrace data sharing. Trends in Ecology and Evolution. 2019. February;34(2):98. [DOI] [PubMed] [Google Scholar]
  • 86.Murillo AP. Data at risk initiative: examining and facilitating the scientific process in relation to endangered data. Data Science Journal. 2014; 12:207–219. [Google Scholar]
  • 87.Bierer BE, Crosas M, Pierce HH. Data authorship as an incentive to data sharing. N Engl J Med. 2017. April 27;376(17):1684–1687 10.1056/NEJMsb1616595 [DOI] [PubMed] [Google Scholar]
  • 88.Pierce HH, Dev A, Statham E, Bierer BE. Credit data generators for data reuse. Nature. 2019. June;570(7759):30–32. 10.1038/d41586-019-01715-4 [DOI] [PubMed] [Google Scholar]
  • 89.Ioannidis JP. Anticipating consequences of sharing raw data and code and of awarding badges for sharing. J Clin Epidemiol. 2016. February;70:258–60. 10.1016/j.jclinepi.2015.04.015 [DOI] [PubMed] [Google Scholar]
  • 90.Moher D, Naudet F, Cristea IA, Miedema F, Ioannidis JPA, Goodman SN. Assessing scientists for hiring, promotion, and tenure. PLoS Biol. 2018. March 29;16(3):e2004089 10.1371/journal.pbio.2004089 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 91.HESA (Higher Education Statistics Agency). The Higher Education Classification of Subjects. Available at: https://www.hesa.ac.uk/innovation/hecos Accessed October 11, 2019.
  • 92.Lindlof TR, Taylor BC. Qualitative Communication Research Methods, 2nd Edition. Thousand Oaks, CA: Sage, 2002. [Google Scholar]
  • 93.Guest G, Namey E, Taylor J, Eley N, McKenna K. Comparing focus groups and individual interviews: findings from a randomized study. International Journal of Social Research Methodology. 2017;20(6):693–708. [Google Scholar]

Decision Letter 0

Cassidy Rose Sugimoto

4 Oct 2019

PONE-D-19-18302

The views, perspectives, and experiences of academic researchers with data sharing and reuse: a meta-synthesis

PLOS ONE

Dear Dr. Perrier,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

The reviewers found the manuscript valuable; however, several things need to be addressed before it meets PLOS publication criteria. The most immediate is the need for additional analysis, to ensure the robustness of the results. There are also some minor concerns about the legibility of figures. I would encourage the authors to address all concerns raised in the reviews. 

We would appreciate receiving your revised manuscript by Nov 18 2019 11:59PM. When you are ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter.

To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). This letter should be uploaded as separate file and labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. This file should be uploaded as separate file and labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. This file should be uploaded as separate file and labeled 'Manuscript'.

Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out.

We look forward to receiving your revised manuscript.

Kind regards,

Cassidy Rose Sugimoto, Ph.D.

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at http://www.journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and http://www.journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. We note that you have stated that you will provide repository information for your data at acceptance. Should your manuscript be accepted for publication, we will hold it until you provide the relevant accession numbers or DOIs necessary to access your data. If you wish to make changes to your Data Availability statement, please describe these changes in your cover letter and we will update your Data Availability statement to reflect the information you provide.

Additional Editor Comments (if provided):

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: I Don't Know

Reviewer #2: I Don't Know

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: 1. Is the manuscript technically sound, and do the data support the conclusions?

The use of CASP and the supporting Appendices provided the necessary information to ensure that the research was conducted in a technically sound way. The findings are described thoroughly and are supported by the data. The appendices and the quotations from the articles make it easier to determine if the data supports the conclusions.

2. Has the statistical analysis been performed appropriately and rigorously?

No statistical analysis was conducted in this study; however, the independent coders used the appropriate metric CASP to ensure that the documents were of high quality. However, surprisingly, the authors did not describe or include any intercoder reliability statistics for either the CASP coding nor the analytical themes coding. A discussion of this should be included in the study.

3. Have the authors made all data underlying the findings in their manuscript fully available?

Yes, the data is provided in a repository. Additionally, the authors provided a list of all manuscripts analyzed and the databases that were used in the appendices.

4. Is the manuscript presented in an intelligible fashion and written in standard English?

Yes, the language used in the manuscript is intelligible, with very few copyediting mistakes. There are occasionally lack of commas and other minor copyediting errors. However, the language throughout is vague, I would suggest editing throughout the make the paper more specific.

5. Review Comments to the Author

Other comments:

Figure 2 was nearly impossible to see, a higher quality image needs to be provided.

While this paper is well written and easy to follow, the findings leave the reading wanting to know why this study is important. It provides a summary of the other studies in data sharing, however, does not provide an understanding outside what is already known from the current data sharing research. I would suggest adding more to the discussion and conclusion to indicate why this study is important and what the academic community and data sharers, as well as repositories managers should learn from it.

Reviewer #2: This is an interesting “meta-synthesis” of 45 studies previously published that addressed data sharing. The conclusions are not surprising, and the results (the ‘synthesis’) has all been published previously, as well as in review articles about the same topic. This is not novel, nor are new suggestions made for further work to be done or solutions to be tested. Nevertheless, it represents a comprehensive review of 45 studies.

While the authors have aggregated concerns and opinions about sharing data, they provide no statistical analysis of the frequency of those concerns. This is a major limitation of the work. How often did a concern or opinion need to be observed in the meta-analysis in order for it to be listed as a major concern? Did any concern(s) rise above others or for different types of research data? What weight should be assigned to the importance of each of these many concerns?

While a majority of the studies utilized interviews as the principal methodology, 8 studies did not. Were there any differences between focus groups (or other) and interview results? Were any surveys found?

The authors appear to comingle studies that interviewed (or reported on focus groups of) researchers that are required to (or encouraged to) deposit data as well as attitudes of others (e.g. data scientists, students, data users, lawyers, etc.). One would anticipate that the opinions of each of these would be quite different, and different again from participants in clinical trials or research. How these different subgroups weighted their concerns, and how analyzed should be discussed. This may require incremental analysis by the authors.

It is surprising to this reviewer that there is no differentiation between the attitudes of scientists in the different fields: Investigators in some research disciplines are accustomed to sharing data (e.g. astronomy, genetics) and have standards in place for doing so. Researchers in other disciplines, most notably in human participant research where issues of privacy and confidentiality are central concerns, appear to drive reticence to share. The authors state that 17 of 45 studies were from “mixed” disciplines; nevertheless, this reviewer requests an analysis by type of research, and by ‘concern’.

The authors mention that one of the authors “conducted focus groups with academic researchers…and found that data saturation was reached after conducting four focus groups despite collecting data from discrete disciplines (i.e., health science, humanities, natural science).” How did that research factor into this current analysis and what significance does it hold?

Did the authors perceive any differences in opinions and attitudes over time?

Figure 2 is illegible. Why not present in a tabular form?

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy

Reviewer #1: No

Reviewer #2: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files to be viewed.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/ PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2020 Feb 27;15(2):e0229182. doi: 10.1371/journal.pone.0229182.r002

Author response to Decision Letter 0


11 Oct 2019

Response to Peer Reviewers

Reviewer #1

No statistical analysis was conducted in this study; however, the independent coders used the appropriate metric CASP to ensure that the documents were of high quality. However, surprisingly, the authors did not describe or include any intercoder reliability statistics for either the CASP coding nor the analytical themes coding. A discussion of this should be included in the study.

Completed: Page 4-5

We added a new section under Materials and Methods entitled ‘Types of Studies’ and provided the references and support to outline more information on metasynthesis as a study type. This includes the clarification that metasyntheses is a qualitative review and therefore does not use quantitative approaches/analysis, e.g., statistics (Walsh & Downe 2005).

Other comments:

Figure 2 was nearly impossible to see, a higher quality image needs to be provided.

Completed: Figure 2

The font size and dimensions have been increased on the image.

While this paper is well written and easy to follow, the findings leave the reading wanting to know why this study is important. It provides a summary of the other studies in data sharing, however, does not provide an understanding outside what is already known from the current data sharing research. I would suggest adding more to the discussion and conclusion to indicate why this study is important and what the academic community and data sharers, as well as repositories managers should learn from it.

Completed: Page 3-4; Page 27; Page 29

Page 3-4: A paragraph was added to the Introduction to clarify why this study is important.

Page 27, Last paragraph of Discussion: A paragraph was added to confirm why this study is important.

Page 29, Last sentence of Conclusions: The Conclusions were enhanced to clarify why this study is important.

Reviewer #2 Comments

This is an interesting “meta-synthesis” of 45 studies previously published that addressed data sharing. The conclusions are not surprising, and the results (the ‘synthesis’) has all been published previously, as well as in review articles about the same topic. This is not novel, nor are new suggestions made for further work to be done or solutions to be tested. Nevertheless, it represents a comprehensive review of 45 studies.

No Changes Made

We searched for a review on this topic before undertaking our study but were not able to locate any. Reviewer #2 states “…the results (the ‘synthesis’) has all been published previously, as well as review articles about the same topic”.

• We would be very appreciative if the citations of the qualitative reviews on the perspectives/experiences of academic researchers and data sharing could be shared to determine if there is overlap with our study

While the authors have aggregated concerns and opinions about sharing data, they provide no statistical analysis of the frequency of those concerns. This is a major limitation of the work. How often did a concern or opinion need to be observed in the meta-analysis in order for it to be listed as a major concern? Did any concern(s) rise above others or for different types of research data? What weight should be assigned to the importance of each of these many concerns?

Completed: Page 4-5

We added a new section under Materials and Methods entitled ‘Types of Studies’ and provided the references and support to outline more information on metasynthesis as a study type. This includes the clarification that metasyntheses is a qualitative review and therefore does not use quantitative approaches/analysis, e.g., statistical analysis, meta-analysis, or weighting, i.e., quantitative approaches (Zimmer 2006; Walsh & Downe 2005; Sandelowski & Barroso 2007).

While a majority of the studies utilized interviews as the principal methodology, 8 studies did not. Were there any differences between focus groups (or other) and interview results? Were any surveys found?

Completed: Page 4-5

We added a new section under Materials and Methods entitled ‘Types of Studies’ and provided the references and support to outline more information on metasynthesis as a study type. This includes the clarification that metasyntheses is a qualitative review and therefore does not use quantitative approaches/analysis, e.g., subgroup analyses (Zimmer 2006; Walsh & Downe 2005; Sandelowski & Barroso 2007).

Surveys are eligible if they included mixed methods and is described on Page 5-6 under Eligibility Criteria: “Mixed methods studies that used both qualitative and quantitative methods within the same study were eligible if the qualitative portion met our inclusion criteria.”

The authors appear to comingle studies that interviewed (or reported on focus groups of) researchers that are required to (or encouraged to) deposit data as well as attitudes of others (e.g. data scientists, students, data users, lawyers, etc.). One would anticipate that the opinions of each of these would be quite different, and different again from participants in clinical trials or research. How these different subgroups weighted their concerns, and how analyzed should be discussed. This may require incremental analysis by the authors.

Completed: Page 4-5

We added a new section under Materials and Methods entitled ‘Types of Studies’ and provided the references and support to outline more information on metasynthesis as a study type. This includes the clarification that metasyntheses is a qualitative review and therefore does not use quantitative approaches/analysis, e.g., subgroup analyses (Zimmer 2006; Walsh & Downe 2005; Sandelowski & Barroso 2007).

We stated on Page 5 that “50% or more of the total sample had to be researchers from academic institutions in order to be eligible for inclusion”. As a result, the majority of participants being analysed would be researchers (i.e., the other participants would be present in lesser numbers).

It is surprising to this reviewer that there is no differentiation between the attitudes of scientists in the different fields: Investigators in some research disciplines are accustomed to sharing data (e.g. astronomy, genetics) and have standards in place for doing so. Researchers in other disciplines, most notably in human participant research where issues of privacy and confidentiality are central concerns, appear to drive reticence to share. The authors state that 17 of 45 studies were from “mixed” disciplines; nevertheless, this reviewer requests an analysis by type of research, and by ‘concern’.

Completed: Page 4-5

We added a new section under Materials and Methods entitled ‘Types of Studies’ and provided the references and support to outline more information on metasynthesis as a study type. This includes the clarification that metasyntheses is a qualitative review and therefore does not use quantitative approaches/analysis, e.g., sub-group analysis such as by type of research and by ‘concern’ (Walsh & Downe 2005).

The authors mention that one of the authors “conducted focus groups with academic researchers…and found that data saturation was reached after conducting four focus groups despite collecting data from discrete disciplines (i.e., health science, humanities, natural science).” How did that research factor into this current analysis and what significance does it hold?

Completed: Page 27-28

Further explanation has been provided to this statement in order to provide clarification that despite different research methods and processes being used in various disciplines, researchers aligned with issues in research data management.

Did the authors perceive any differences in opinions and attitudes over time?

Completed: Page 4-5

We added a new section under Materials and Methods entitled ‘Types of Studies’ and provided the references and support to outline more information on metasynthesis as a study type. This includes the clarification that metasyntheses is a qualitative review and therefore does not use quantitative approaches/analysis (Glenton 2014).

Figure 2 is illegible. Why not present in a tabular form? Completed: Figure 2

The font size and dimensions have been increased on the image.

References

Glenton C, Lewin S. Using evidence from qualitative research to develop WHO guidelines. WHO Handbook for Guideline Development. 2nd Edition. Geneva: WHO, 2014.

Sherwood, G. (1999). Meta-synthesis: Merging qualitative studies to develop nursing knowledge. International

Journal for Human Caring, 3, 37-42

Sandelowski M, Barroso J. Handbook for synthesizing qualitative research. New York, NY: Springer, 2007.

Walsh D, Downe S. Meta-synthesis method for qualitative research: A literature review. Journal of Advanced Nursing. 2005; 50:204-211.

Zimmer L. Qualitative meta-synthesis: A question of dialoguing with texts. Journal of Advanced Nursing. 2006; 53:311-318.

Attachment

Submitted filename: Response to Reviewers Comments.docx

Decision Letter 1

Pablo Dorta-González

3 Feb 2020

The views, perspectives, and experiences of academic researchers with data sharing and reuse: a meta-synthesis

PONE-D-19-18302R1

Dear Dr. Perrier,

We are pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it complies with all outstanding technical requirements.

Within one week, you will receive an e-mail containing information on the amendments required prior to publication. When all required modifications have been addressed, you will receive a formal acceptance letter and your manuscript will proceed to our production department and be scheduled for publication.

Shortly after the formal acceptance letter is sent, an invoice for payment will follow. To ensure an efficient production and billing process, please log into Editorial Manager at https://www.editorialmanager.com/pone/, click the "Update My Information" link at the top of the page, and update your user information. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to enable them to help maximize its impact. If they will be preparing press materials for this manuscript, you must inform our press team as soon as possible and no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

With kind regards,

Pablo Dorta-González, Ph.D.

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

Reviewer #2: (No Response)

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Partly

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: N/A

Reviewer #2: No

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The authors have sufficiently addressed all of the comments from the previous review and have added additional discussion of the methods, as well as a more thorough discussion of the importance of the study.

There are some very minor types throughout and the paper should be proofread one last time before publication.

Reviewer #2: The authors have not adequately addressed the comments of either of the two reviewers, both of whom had substantive concerns with the data, the presentation of the data, and the conclusion drawn. Further, both reviewers expressed concern that there were no to limited new observations or insights from this study. Observations are thematic, and not novel. Perhaps it is more appropriate for a more specialized journal.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy

Reviewer #1: No

Reviewer #2: No

Acceptance letter

Pablo Dorta-González

18 Feb 2020

PONE-D-19-18302R1

The views, perspectives, and experiences of academic researchers with data sharing and reuse: a meta-synthesis

Dear Dr. Perrier:

I am pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please notify them about your upcoming paper at this point, to enable them to help maximize its impact. If they will be preparing press materials for this manuscript, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

For any other questions or concerns, please email plosone@plos.org.

Thank you for submitting your work to PLOS ONE.

With kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Mr. Pablo Dorta-González

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Appendix. PRISMA checklist.

    (DOC)

    S2 Appendix. MEDLINE search strategy.

    (DOCX)

    S3 Appendix. Literature databases searched.

    (DOCX)

    S4 Appendix. CASP (critical appraisal skills programme): Qualitative checklist.

    (DOCX)

    S5 Appendix. Included studies.

    (DOCX)

    S6 Appendix. Included studies by discipline.

    (DOCX)

    Attachment

    Submitted filename: Response to Reviewers Comments.docx

    Data Availability Statement

    The data are available from the Zenodo Repository, DOI: doi.org/10.5281/zenodo.3258850.


    Articles from PLoS ONE are provided here courtesy of PLOS

    RESOURCES