Skip to main content
PLOS Biology logoLink to PLOS Biology
. 2024 Apr 2;22(4):e3002562. doi: 10.1371/journal.pbio.3002562

Shortcut citations in the methods section: Frequency, problems, and strategies for responsible reuse

Kai Standvoss 1,2,3,, Vartan Kazezian 4,, Britta R Lewke 5, Kathleen Bastian 6, Shambhavi Chidambaram 7, Subhi Arafat 3,8, Ubai Alsharif 9, Ana Herrera-Melendez 10, Anna-Delia Knipper 11, Bruna M S Seco 12,¤, Nina Nitzan Soto 13, Orestis Rakitzis 14, Isa Steinecker 2,15, Philipp van Kronenberg Till 16, Fereshteh Zarebidaki 17, Parya Abbasi 4, Tracey L Weissgerber 4,*
Editor: Cilene Lino de Oliveira18
PMCID: PMC10986953  PMID: 38564513

Abstract

Methods sections are often missing essential details. Methodological shortcut citations, in which authors cite previous papers instead of describing the method in detail, may contribute to this problem. This meta-research study used 3 approaches to examine shortcut citation use in neuroscience, biology, and psychiatry. First, we assessed current practices in more than 750 papers. More than 90% of papers used shortcut citations. Other common reasons for using citations in the methods included giving credit or specifying what was used (who or what citation) and providing context or a justification (why citation). Next, we reviewed 15 papers to determine what can happen when readers follow shortcut citations to find methodological details. While shortcut citations can be used effectively, they can also deprive readers of essential methodological details. Problems encountered included difficulty identifying or accessing the cited materials, missing or insufficient descriptions of the cited method, and shortcut citation chains. Third, we examined journal policies. Fewer than one quarter of journals had policies describing how authors should report previously described methods. We propose that methodological shortcut citations should meet 3 criteria; cited resources should provide (1) a detailed description of (2) the method used by the citing authors’, and (3) be open access. Resources that do not meet these criteria should be cited to give credit, but not as shortcut citations. We outline actions that authors and journals can take to use shortcut citations responsibly, while fostering a culture of open and reproducible methods reporting.


Methods sections often lack critical details needed to reproduce an experiment; the practice of citing previous papers instead of describing the methods in detail may contribute to this problem. An analysis of >800 papers across three fields shows that >90% of papers use at least one shortcut citation, that these significantly impair reconstruction of the original method, and that <25% of journals have policies relating to previously described methods.

Introduction

Methods sections should serve several purposes, each of which requires a different level of detail (Fig 1). Well-written methods sections provide an overview of the study design and techniques used to answer the research question, help readers to evaluate the risk of bias and provide details needed to replicate the experiment. Unfortunately, methods sections are often missing critical details. The Reproducibility Project: Cancer Biology aimed to replicate high profile findings from 193 experiments in cancer research [1]. Unfortunately, none of the papers contained sufficient details to allow researchers to design and conduct a replication study [2]. When the original authors were contacted to obtain methodological details, 41% were extremely or very helpful, 9% were minimally helpful, and 32% were not helpful or did not respond [2]. An assessment of 300 fMRI studies revealed that key information, such as how the task was optimized for efficiency or the distribution of inter-trial intervals, was frequently missing [3]. Methodological details in randomized controlled trials are also frequently missing. In approximately half of randomized controlled trials included in Cochrane reviews, methods were not reported in enough detail to allow researchers to complete risk of bias assessments for items such as randomization sequence generation and allocation concealment [4].

Fig 1. The level of methodological detail required depends on the reader.

Fig 1

All readers need an overview of the study design, methods used to answer the research question, and information needed to assess scientific rigor and the risk of bias. These details should always be presented in the methods section of the paper. While fewer readers need the details required to reproduce or reuse the method, these individuals are particularly important because they are most likely to perform follow-up experiments. Very simple methods that can be explained and reproduced easily may be described in the methods section. Protocol repositories or method or protocol journals are better for many methods, as it is difficult to provide the details needed to implement or reuse the method within the methods section.

A methodological shortcut citation is a citation that is used to replace a full description of a method or a part of a method, under the assumption that the resource being cited fully describes the method. Shortcut citations are sometimes accompanied by phrases like “as described previously” or “see (citation) for details.” Methodological shortcut citations are one factor that could affect readers’ ability to reproduce an experiment.

While some scientists and editors view the use of shortcut citations as a good practice [5,6], others worry that these shortcuts may adversely affect reproducibility [7]. Ideally, shortcut citations should reference resources that describe the method, as it is currently performed, in detail. This may include protocol papers, diagnostic guidelines, or original research articles with detailed methods. In practice, anecdotal reports show that shortcut citations can cause problems [8,9]. Readers may be unable to access the cited paper. The cited paper may not contain methodological details or may itself use a shortcut citation instead of describing the methods. Those in favor of shortcut citations note that authors do not waste time repeating details that have been written elsewhere and avoid potential copyright issues that might emerge if one copied methods from another publication. Those who are skeptical of shortcuts argue that the cited paper may no longer reflect current practice or may not accurately describe the procedures of the authors who cite it. Furthermore, consulting cited resources to obtain details needed to interpret the study is time consuming. There are also disagreements about how shortcut citations should be used. While some authors cite the paper that introduced the method to give its creators credit, others cite the paper whose methods most closely resemble their own. Some journals require authors to use shortcut citations to avoid repeating published methods, or incentivize this practice through strict word limits, whereas other journals have excluded the methods section from word limits to encourage detailed reporting [10].

Our meta-research study used 3 approaches to systematically examine the use of methodological shortcut citations in neuroscience, biology, and psychiatry. First, we examined papers to determine why authors use citations in the methods section and to assess how often shortcut citations were used. Next, we reviewed shortcut citations for 15 papers to determine what can happen when readers follow these citations to find methodological details. Third, we reviewed journal policies related to shortcut citations and methodological reporting.

Methods

This study was performed as part of a participant guided, learn-by-doing course [11], in which graduate students in different fields at 4 Berlin universities learned meta-research skills by working together to design, conduct, and publish a meta-research study.

We conducted 3 distinct but related studies.

  1. Methodological citations study: This examined the reasons why authors use citations in the methods section of papers and assessed how often shortcut citations were used. Additional data collected included the number of resources cited in a shortcut citation and the years in which papers cited as shortcuts were published.

  2. Shortcut citation chains study: This study examined problems that may occur when readers consult shortcut citations to find further details of the study methods.

  3. Journal policy study: This study examined journal policies related to shortcut citations and methodological reporting.

Each study examined 3 fields: neuroscience, biology, and psychiatry, to improve generalizability. These fields were selected based on the study teams’ expertise. The abstraction protocols, data, and code for the methodological citations study and journal policy studies were deposited on the Open Science Framework (RRID:SCR_003238) at https://osf.io/d2sa3/ [12].

Methodological citations study

This was a cross-sectional study of research articles.

Systematic review

We followed all relevant items in the PRISMA guidelines [13]. Items that only applied to meta-analyses or were not relevant to literature surveys were not followed. Ethical approval was not required.

Journal screening

Our sampling frame included journals with the highest impact factors that publish original research in each field. Journals for each category were ranked according to 2019 impact factors listed for the specified categories in Journal Citation Reports. We excluded journals that did not publish original research or did not publish a March 2020 issue. Fewer journals were used for biology (n = 15), compared to neuroscience and psychiatry (n = 20), due to the large number of publications in some biology journals. The neuroscience journals examined generally focus on biomedical neuroscience, which may include basic science, translational and clinical research (S1 Table). The biology journals examined publish a wide range of general biology research and are not exclusively biomedical (S2 Table).

Search strategy

Articles were identified through a PubMed search. We performed a supplemental Web of Science search to identify articles published in journals that were not indexed in PubMed. The full search strategy is available on the OSF repository [12].

Inclusion and exclusion criteria

The study included all full-length, original research articles that included a methods section published in each included journal between March 1 and March 31, 2020. Among journals that publish print issues, we examined all articles included in print issues of the journal that were published in March 2020 (S1S3 Tables and S1 Fig). Articles for online journals that did not publish print issues were included if the publication date was between March 1 and March 31, 2020. Articles were excluded if they were not full-length original research articles, did not contain methods sections, or were methods articles.

Screening

Screening for each article was performed by 2 independent reviewers (Biology: PvKT, SC, ADK, VK; Neuroscience: KS, FZB, SA; Psychiatry: IS, OR, UA) using Rayyan software (RRID:SCR_017584), and disagreements were resolved by consensus discussions between the 2 reviewers. A list of articles was uploaded into Rayyan. Reviewers independently examined each article and marked whether the article was included or excluded.

Abstraction

All abstractors completed a training set of 35 articles before abstracting data. Data abstraction for each article was performed by 2 independent reviewers (Biology: SC, ADK, VK, PvKT; Neuroscience: KS, FZB, SA; Psychiatry: BL, IS, AHM). When disagreements could not be resolved by consensus, ratings were assigned after a group review of the paper. Eligible manuscripts were reviewed in detail to evaluate the following questions according to a predefined protocol (available at: https://osf.io/d2sa3/ [12]).

The following data were abstracted:

  1. Is the paper related to the SARS-CoV-2 pandemic? This information was abstracted as, at the beginning of the pandemic, members of the scientific community was concerned about the quality of COVID-19 papers due to the speed at which these studies were conducted and published.

  2. Does the paper include additional methodological details in the supplement?

  3. Does the paper reference a repository or repositories as a shortcut for any method used? If so, what is the name of the repository? (Note that code used to analyze data was considered a method.)

Abstractors then reviewed each citation in the methods section, including citations listed in STAR (Structure Transparent Accessible Reporting) methods tables [14]. Some journals use these tables to provide an overview of all of the key reagents used in the study. Notations or hyperlinks that appeared only in the text, and not as entries in the reference list (e.g., company names), were not counted as citations. Links to supplemental files were not counted as citations. Multiple papers cited as part of the same reference were treated as a single citation (e.g., 2 citations appearing within the same set of brackets). Papers cited in different locations in the same sentence were treated as different citations. Citations in the supplemental methods were not evaluated.

Each methodological citation was classified into one of the categories outlined in Table 1, according to the inferred purpose of the citation. The “How (Explain a method)” and “Other” categories could be shortcut citations, whereas other categories could not. Each citation in these 2 categories was classified as a probable shortcut, possible shortcut, or no shortcut. The conceptual goal was to distinguish between situations where readers would likely need to consult the cited paper to implement the method (probable shortcut), situations where the reader may need to consult the shortcut citation to implement the method (possible shortcut), and cases where the reader would not need to consult the shortcut citation to implement the method. However, these conceptual definitions are highly subjective and depend on the reader’s knowledge of the reported methods. We therefore used the syntactic definitions described below to classify shortcut citations.

Table 1. Reasons for citations in the methods section.
Category Example Potential shortcut
How (Explain a method): The citation was intended to explain how something was done. This category included 4 subcategories for specific types of citations (1. General, 2. Protocol, 3. Prior publication describing the study design and/or results, and 4. Guideline or manual). Liver samples were collected, sectioned, and frozen as described previously (citation). Yes
Who or what (Give credit or specify what was used): The citation is used to give credit to the group that created the method, tool, resource or substance, or to specify exactly what method, tool, resource, or substance was used. The citation is not used to explain how the method was performed, or how the tool, resource or substance was used. This category includes 3 subcategories (1. Software, 2. Atlas, 3. Other). Analyses were conducted using R (citation of R). No
From where (Source of materials): The citation is used to show where a substance or organism was obtained from, not how the substance or organism was created. Mice were obtained from the Smith lab (citation). No
Why (Provide context or a justification): The citation refers to previous studies to provide context. This includes citations that compare the authors’ results with results from previous studies, demonstrate that others have used a similar approach, or explain why the authors chose a particular method, model, formula, etc. The citation is not intended to provide insight into how the existing study was conducted. In accordance with previous studies (citation), we observed that the optimal treatment dose was 20 ml. No
Formula or value: The citation referred to a formula or specified a value for a parameter that was used. The formula or value had to be stated in the article, so that the reader would not need to look up this information in the cited resource. The citation was not used to explain how that parameter was calculated, how the formula was derived, or why the parameter or formula was chosen. Drug cost $X for drug Y was obtained from publicly available reports (citation). No
Other: Citations that do not fit into the other categories. Yes

A complete protocol with examples is available in the online repository.

  • Probable shortcut: The sentence that includes the shortcut citation is the only description of the method. Additional details are not provided in the following sentences or elsewhere in the methods section.

  • Possible shortcut: Additional details of the cited method are provided in the sentences following the sentence containing the shortcut citation, or elsewhere in the methods section.

  • Not a shortcut: A reader would not need to consult the cited paper to implement the method. This rare category was generally used when the citation referred to concentrations, parameters, or other details that were fully specified in the sentence where the shortcut citation appeared.

Two independent reviewers (SA and UA) abstracted the following data for each paper:

  • The minimum and maximum number of resources cited within each probable shortcut.

  • The minimum and maximum number of resources cited within each possible shortcut.

  • The publication year of the youngest and oldest probable shortcut citations.

  • The publication year of the youngest and oldest possible shortcut citations.

Protocol modification

During peer review, a reviewer requested that we add publication years for all resources cited as shortcuts, rather than using the minimum and maximum to quantify the range of values observed. This data was abstracted by a single reviewer (PA).

Shortcut citation chains study

Selection of articles

Detailed assessments were performed on a cohort of studies. In each field, papers from the methodological citations study were divided into quintiles based on the total number of shortcuts (possible + probable). Ten articles from each quintile were randomly selected using a computer algorithm and placed on an ordered list. Potential reviewers identified articles on the list that fell within their area of expertise. We then chose the first article in each list that had a self-identified expert evaluator. This approach was used to select 15 parent articles (1 article per quintile per field).

Data abstraction

Each reviewer carefully examined all papers or materials cited as shortcuts to determine whether they could locate the cited content. While articles were almost always accessible, reviewers were sometimes unable to access other types of cited resources. Books were only checked if they could be accessed online, as students could not visit libraries due to COVID-19 pandemic restrictions. Newer versions of books and manuals were reviewed if older, cited versions were unavailable or inaccessible.

For each cited article or resource that was found, the reviewer documented the cited material type (paper, protocol, book, etc.), publication year, whether the article or resource was open access or behind a paywall, and any other problems encountered while searching for information on the cited method. Additionally, reviewers noted whether book citations included chapters or pages. Finally, reviewers also noted whether the article or resource contained an adequate description of the method cited in the parent article. Descriptions might be judged as inadequate if they were clearly missing essential details that would be needed to implement the methods. If the methodological description was not adequate and the shortcut citation also used a shortcut citation to explain the method cited in the parent paper, then the reviewer repeated the abstraction process for each new shortcut citation, adding these new shortcuts as additional steps in the shortcut citation chain. Abstraction was complete when the reviewer either found a complete and comprehensive description of the method or reached a dead-end in the chain of shortcut citations. Dead ends included an inability to locate the cited article or resource, or an article or resource that did not describe the cited method.

A second reviewer assessed the accuracy of all abstracted information. A graphic illustrating the number of shortcuts and chains of shortcut citations was created for each parent paper.

Journal policy study

In this cross-sectional analysis of journal policies, we examined policies of all eligible journals listed in the Journal Citation Reports 2019 ranking for neuroscience, psychiatry, and biology.

Inclusion and exclusion criteria

Journals that publish original research were included, whereas journals that only publish review articles, book series, correspondence, perspectives, or editorials were excluded. Methods journals were excluded, as these journals often require authors to report extensive methodological details. Journals were also excluded if they did not have a website, had suspended publishing activities, or planned to cease publishing in 2021.

Search strategy

Journal webpages were examined to confirm that journal policies were accessible. Electronic searches were performed using the terms “[journal name],” “journal citation reports ranking,” “author guidelines,” “journal policy,” and “impact factor.” When journals with similar names were identified, impact factors were used to confirm that the correct journal was selected.

Screening

Each journal was screened by 2 of the 3 independent reviewers (BMSS, KB, NNS) to determine whether the journal was eligible according to the prespecified criteria listed above. Discrepancies were resolved by consensus. Information on whether the journal publishes original research articles was assessed through the journal descriptions (e.g., “About the Journal” or “Aims and Scope”), or by determining whether the submission guidelines listed original research as an article type. If this information was inconclusive, the 2 most recent issues were examined to determine whether the journal published original research.

Abstraction

Training was performed on the 20 eligible journals with the highest impact factor for each category prior to data abstraction. Data were collected by 2 of the 3 independent abstractors (BMSS, KB, NNS). When disagreements could not be resolved by consensus, discrepancies were resolved through discussion with the third reviewer. An additional study team member (TLW) was consulted to resolve complex discrepancies. A trained abstractor abstracted data from non-English journal webpages with the help of a native speaker.

Webpages containing author and or submission guidelines were identified by looking for the following terms in the website menu, or by using the website search function: “policy,” “policies,” “author,” “author/s guidelines,” “author/s instructions,” “submission guidelines,” “submit your article,” “recommendation,” “about,” “publish.” Each webpage section was visually examined using the search terms “method,” “methods,” “experiment,” “reproducibility,” “replicate,” “replication,” “repository,” “repositories,” “self-plagiarism,” “supplementary,” “supplementaries,” “citation,” “protocol,” “journal protocol,” “scientific society,” and “society.”

Reviewers determined whether the instructions for authors or journal policies addressed the following points:

  1. Explicitly asked authors to provide sufficient methodological details to allow others to reproduce the experiment.

  2. Specified how authors should describe methods that have been reported elsewhere (e.g., provide a citation instead of describing the method, briefly summarize the method and provide a citation).

  3. Encouraged authors to use supplemental files, protocol repositories, or protocol journals to explain their methods.

Reviewers assessed whether this information was found in the material and methods section of the author guidelines or journal policies, in other sections of the author guidelines, or elsewhere on the journal’s website.

Reviewers also consulted the journal website to determine whether the journal was affiliated with a scientific society, and whether the journal endorsed the TOP (Transparency and Openness Promotion) guidelines. The Center for Open Science list of journals that have implemented the TOP guidelines (https://www.cos.io/initiatives/top-guidelines, https://osf.io/mwxb3/) was consulted to identify journals that endorsed TOP, but did not provide this information on their website.

Data analysis

Code for data figures and color schemes was adapted from a previous paper [15]. Summary statistics were calculated using Python (RRID:SCR_008394, version 3.7.7, libraries NumPy 1.18.5 [16], Pandas 1.2.4 [17] and Matplotlib 3.4.1 [18,19]). Charts were prepared with a Python-based Jupyter Notebook (Jupyter-client, RRID:SCR_018413 [20], Python version 3.7.7, libraries NumPy 1.18.5 [16], Pandas 1.2.4 [17] and Matplotlib 3.4.1 [18,19]) and assembled into figures with vector graphic software.

Results

Methodological citations study

Study sample

The study sample consisted of 224 articles with 2,756 methodological citations in neuroscience, 431 papers with 6,226 methodological citations in biology, and 160 papers with 1,870 methodological citations in psychiatry (Figs 2A and S1). The sample contained few publications related to COVID-19 (neuroscience: n = 0, 0%, biology: n = 2, 0.5%, psychiatry: n = 0, 0%). While it was not feasible to collect data on the types of methods used, the study team’s subjective impression is that the papers examined covered an extensive range of methods used in each field. The biology journals published both biomedical and non-biomedical research, including field studies, computational research, and laboratory studies. Neuroscience and psychiatry journals had a strong biomedical focus. Psychiatry research is predominantly clinical, and resources cited as shortcuts often included diagnostic guidelines and surveys. Neuroscience included basic, translational, and clinical studies.

Fig 2. Understanding the use of citations in the methods sections, methods supplements, and methods repositories.

Fig 2

(A) The most common reasons why authors cite papers in the methods section are to explain how a method was performed, give credit or specify what was used (who or what), or provide context or a justification (why). These numbers should be regarded as approximations. Small changes in the wording or position of the reference could alter the categorization, as could variations in reader expertise (see limitations). (B) Citations and shortcut citations often appear in the methods section of published papers. In the violin plot on the right, values for probable shortcut citations are shown in darker hues on the left half of each violin, whereas possible shortcut citations are shown in lighter hues on the right side of each violin. Most papers contain both probable and possible shortcut citations. Dashed lines on each half of the violin indicate the 25th, 50th, and 75th percentiles. Median (25th percentile, 75th percentile) for the number of probable shortcut citations per paper were as follows: neuroscience 3 (1, 6); biology 2 (1, 4); psychiatry 3 (1, 5). Median and 25th and 75th percentiles for the number of possible shortcut citations per paper were as follows: neuroscience 2 (1, 4); biology 1 (0, 2); psychiatry 2 (1, 4). (C) Methods are often shared in supplements but are less likely to be deposited on methods repositories. Percentages may not total 100% due to rounding errors. Data are available at https://osf.io/d2sa3/, in the methodological citations study folder [12].

Use of citations in the methods section

Citations were common in the methods sections of published papers (Fig 2B, left panel). The median number of citations in the methods section was 10 [interquartile range: 6, 16.25] in neuroscience, 12 [6, 19] in biology, and 10 [7, 16] in psychiatry.

In neuroscience and psychiatry, 53% to 54% of citations in the methods section were used to explain study methods (How citations, Fig 2A). Citations were also commonly used to give credit or specify what was used (Who or what, 11% to 21% of papers), and provide context or a justification (Why, 19% to 26%). In biology, the most common reason for citing a paper in the methods section was to give credit or specify what was used (Who or what, 41%), followed by explaining a method (How, 31%), and providing context or a justification (Why, 19%). Citations specifying the source of data or materials (From where), or referring to a formula or value, were uncommon in all 3 fields. Citations that did not fit into any of these categories were rare.

Depending on the field, 55% and 60% of papers provided some methodological information in supplemental files, whereas only 8% to 16% provided methodological information in a repository (Fig 2C). While reviewers did not systematically collect data on the content or quality of information in the supplemental methods, reviewers’ subjective impression was that supplemental methods were rarely detailed. Common examples of supplemental methods included tables that list primers or sequences or provided basic information on study participants. Methodological information in repositories included a mixture of study protocols and code for data analysis. The most common repository was GitHub, followed by ClinicalTrials.gov, OSF and FigShare (S4 Table). Other repositories were rarely used in this dataset.

Shortcut citations

Methodological shortcut citations were common in all 3 fields, with 96% of neuroscience papers, 90% of biology papers and 92% of psychiatry papers containing at least 1 possible or probable shortcut. Fig 2B (right panel) shows the median number of possible and probable shortcuts for each field.

Reviewers assessed the age of all resources cited as shortcuts. Fig 3A shows the number of probable (left side of violin plot) and possible (right side of violin plot) shortcut citations in each field. The median age of the youngest shortcuts citation ranged between 3 and 5 years, whereas the median age for the median shortcut citation ranged between 6 and 11 years. Median age for the oldest shortcut citations ranged between 9 and 24 years. Summary statistics for Fig 3A are reported in S5 Table.

Fig 3. Age and number of resources cited within a shortcut citations.

Fig 3

(A) The first violin plot shows the age of all resources cited as shortcuts, whereas the remaining box plots show the age of the youngest, median, and oldest resource cited as a shortcut for each article. The left side of the violin plot shows data for possible shortcut citations, whereas the right side shows data for probable citations. The youngest and median shortcut citation cited in a paper is typically less than 10 years old. The age of the oldest shortcut citation is more variable, and many of these citations were published more than a decade ago. The line with long dashes represents the median, whereas the lines with short dashes represent the 25th and 75th percentiles. The number of papers appears below the field name (n for probable shortcut citations/n for possible shortcut citations). (B) The first bar graph shows the number of resources cited within all shortcut citations in the dataset, whereas the next 3 bar graphs show the minimum, median, and maximum number of resources cited as shortcuts within an article. Percentages may not total 100% due to rounding errors. Data are available at https://osf.io/d2sa3/, in the methodological citations study folder [12].

When using a shortcut citation, authors typically cite 1 resource (Fig 3) Shortcut citations citing 2 resources are also common, whereas citations of 3 or more resources are less common.

Shortcut citation chains study

Figs 4 and 5 show 2 examples of the process of finding cited methodological information for each article. S3 Fig contains illustrations for all 15 articles in the case series. In many cases, reviewers were able to locate additional information. However, this study revealed 5 types of issues that arise when readers follow shortcut citations in search of detailed methods (summarized in Fig 6). The first problem was an inability to identify the cited material due to incomplete or inaccurate citations (e.g., incorrect author names, years, or DOIs) or dead website links. In some cases, reviewers could not find evidence that the cited source existed. The second problem was accessing the cited source. PDFs for some older articles were difficult or impossible to obtain. Many articles, supplemental files, and books were not open access. Subscriptions vary among institutions, and even scientists from well-funded institutions may have difficulty accessing paywalled articles. The third problem was that the cited method was difficult or impossible to find. Some textbook citations failed to provide specific chapters or pages, making it difficult to locate the cited content. In one case, the cited source was in a different language. The fourth problem was that the cited method was not adequately described. Examples included cited sources that did not describe the method or provided the same details as the citing paper. Authors sometimes cited older resources as shortcuts, raising concerns that the cited resource may not have accurately reflected the modern methods used in the citing study. The fifth problem was that in many cases, reviewers had to follow a chain of shortcut citations to locate a complete methodological description. Following a citation chain is time-consuming and readers may have difficulty identifying and accessing the cited material at each step in the chain.

Fig 4. Shortcut citations in example paper 1.

Fig 4

This diagram maps the process of finding methodological details for a biology paper in the fifth quintile of probable + possible shortcut citations in the shortcut citation chains study. The diagram shows the publication year and type of each cited resource and whether the resource was behind a paywall. Text on the diagram provides information describes problems encountered when searching for details about the cited method.

Fig 5. Shortcut citations in example paper 2.

Fig 5

This diagram maps the process of finding methodological details for a neuroscience paper in the fourth quintile of probable + possible shortcut citations in the shortcut citation chains study. The diagram shows the publication year and type of each cited resource and whether the resource was behind a paywall. Text on the diagram provides information, describes problems encountered when searching for details about the cited method.

Fig 6. Problems that arose when searching shortcut citations for detailed methods.

Fig 6

While methodological shortcut citations can be used effectively, reviewers encountered some problems when consulting shortcut citations to find details of cited methods. These included problems identifying the citation, problems accessing the citation, problems finding the cited method within the shortcut citation, and an insufficient description of the cited method. Chains of shortcut citations, in which the cited shortcut citation also used a shortcut citation to describe the cited method, were common.

Journal policy

Study sample

Among the 519 journals screened, 465 were eligible for the study (S4 Fig; 244 neuroscience journals; 76 biology journals; 145 psychiatry journals). Twenty-one journals appeared on the neuroscience and psychiatry lists and were included in both groups.

Policies on details needed to reproduce experiments

Policies that explicitly instruct authors to provide sufficient methodological details to allow others to reproduce the experiment were found in 40% of neuroscience journals, 18% of psychiatry journals, and 44% of biology journals (Fig 7A).

Fig 7. Journal policies for methodological reporting.

Fig 7

Policies for all journals in 3 fields were assessed to determine (A) whether journals ask authors to provide sufficient information about the methods to allow others to reproduce the experiments; (B) how journals ask authors to describe methods that have been reported previously; (C) whether journals ask authors to describe modifications of previously reported methods; and (D) where journals ask authors to share detailed methods. Journals were represented in more than one category of panel D if they encouraged authors to share detailed methods and protocols in more than one place. Percentages in panels A–C may not total 100% due to rounding errors. Data are available at https://osf.io/d2sa3/, in the journal policy study folder [12].

Reporting of previously described methods

Most journals had no policies concerning the reporting of methods that have been described previously (72% to 87%, Fig 7B). Some journals asked authors to summarize the method (9% to 19%) or to provide a citation instead of describing previously published methods in detail (4% to 8%). Policies asking authors to fully describe previously published methods were rare (0% to 2%). Policies asking authors to report modifications of previously described methods were also uncommon (10% to 21%, Fig 7C).

Where to share detailed methods

The percentage of journals that encouraged authors to share detailed methods in protocol repositories ranged between 12% and 23% (Fig 7D). A similar proportion of journals (9% to 21%) encouraged authors to share detailed methods elsewhere, without specifying where methods should be shared. Policies encouraging authors to share detailed methods in a protocol journal (0% to 8%) or as supplemental files (2% to 5%) were rare.

Discussion

This exploratory study of papers in biology, neuroscience, and psychiatry revealed several important findings. First, citations are often used in methods sections. More than 90% of papers used a shortcut citation, explaining how a method was performed. Other common reasons for using citations in the methods included “who or what” citations, that give credit to the authors of another paper or specify what was used, and “why” citations, which provide context or a justification. Different methods evolve at different rates; however, citation age assessments suggested that some methods described in shortcut citations may no longer reflect current practice. The shortcut citation chains study showed that while shortcut citations can be used effectively, they can also create problems for readers seeking detailed methods. These problems included difficulty identifying the correct citation, accessing the cited materials, finding the cited method within the cited materials, and insufficient descriptions of the cited method. Following chains of shortcut citations to find methodological details was time-consuming, and each additional step in the chain can amplify the problems described in the previous sentence. Journals typically lack policies addressing methodological reporting. Fewer than one quarter of journals had policies addressing how authors should report methods that have been described previously, or address modifications of previously described methods. While some journals (18% to 43%, depending on the field) asked authors to provide sufficient methodological details to allow others to reproduce the method, most journals (57% to 81%) had no such policy.

Using shortcut citations to foster a culture of open and reproducible methods

When using shortcut citations, authors replace a section of their methods with a citation referring to another resource. The details contained within that resource are essential to implement the method. We therefore propose that methodological shortcut citations should meet higher standards than other types of citations. Box 1 outlines 3 proposed criteria that authors can use to determine whether a resource should be cited as a shortcut. The open access criterion may be controversial for some, as it suggests that scientists who have a paywalled resource that meets the other 2 criteria should cite this resource to give credit, and create a second, open access resource (e.g., a deposited protocol) to cite as a shortcut. Nevertheless, we believe that the open access criteria is particularly important. Unlike other types of citations, readers who want to implement the study methods will need to read shortcut citations. Paywalled shortcut citations systematically deprive some scientists of information needed to reproduce experiments.

Box 1. How to use methodological shortcut citations responsibly

Methodological shortcut citations replace a section of the methods. Detailed methods are essential for those seeking to implement the method. We therefore propose that resources cited as methodological shortcuts should meet higher standards than materials cited for other purposes.

We propose that authors should use 3 criteria to determine whether a paper, or another resource, should be used as a shortcut citation. Resources that do not meet these criteria can be cited to give credit to those who developed the method, but should not be used as shortcuts.

  1. Detailed description: The resource should provide enough detail about the method that was used to allow others, including those who have little prior experience with the method, to implement the method. Resources with sufficient detail to be shortcut citations might include protocols, methods papers that are recent enough to reflect the current practices, or original research articles with unusually detailed methods.

  2. Similar or identical method: The method described in the resource should be similar or identical to the method used by the authors. The authors should be able to describe any modifications to the methods in the methods section of their paper.

  3. Open access: Paywalled or inaccessible shortcut citations deprive some readers of the information needed to implement the method. This creates disproportionate obstacles for researchers with limited access to publications, including researchers in countries with limited research funding and scientists who are not affiliated with a major institution.

Authors using shortcut citations should:

  1. Describe modifications in detail: Deviations from the published method should be clearly described, in enough detail to allow others to implement the method. When a protocol posted on a repository is cited as a shortcut citation, the authors can version or fork the protocol to share their exact methods. Versioning is sharing an updated version of one’s own protocol, whereas forking is sharing an adapted version of a protocol shared by others.

  2. Specify the exact location of the cited methods: This might include providing page numbers where the cited method was described for books and manuals or describing the method name and location in the cited resource.

When a resource does not meet the criteria proposed in Box 1, we recommend that authors cite the resource to give credit to its authors and use other strategies to share detailed methods. Options for sharing detailed methods include supplemental files, protocol repositories, and protocol journals (S6 Table). Authors who deposit or publish protocols can cite these resources as shortcut citations.

Sharing methods in supplemental files is suboptimal for several reasons. First, readers who have access to the paper may not have access to the supplement. While completing this research, we noticed that some publishers and journals make supplemental files freely available, whereas others do not. Papers obtained through online repositories or interlibrary loan programs may not include supplements. Second, methods in supplemental files are not findable. While scientists can quickly search protocol repositories to identify relevant protocols, there is no way to identify papers that contain detailed methods in the supplemental files. Third, supplemental methods are often written as general descriptions, which are typically less useful than the detailed, step-by-step protocols shared in protocol repositories and methods journals. Fourth, supplemental methods cannot be updated after publication; hence, this format is not useful for tracking the evolution of protocols within and across research groups. This is not a problem when investigators’ goal is to describe their methods for a completed study. Methods, however, are constantly evolving and are one of the most useful outputs that researchers create. Given this, it is more efficient for research teams to use platforms that allow versioning and forking when sharing research procedures, and cite the version of the method that was used for a particular research study.

A key advantage of protocol journals is that they allow authors to obtain credit for their methods development work through a peer-reviewed publication. One disadvantage is that the publication process takes time and effort. Furthermore, the published method cannot be updated—it only shows what one research group is doing at a single point in time. In contrast, protocol repositories allow authors to create living protocols by quickly sharing updated versions [21]. Some repositories also allow forking, in which scientists can post a modified version of a protocol posted by someone else [21]. Versions and forks are linked to the original protocol. This credits the original authors for their work, while allowing researchers to see how the protocol evolved. PLOS ONE offers a new “Lab protocol” publication [22] that combines the advantages of both formats. Lab protocol publications consist of a protocol, deposited on protocols.io [21], paired with a brief peer-reviewed publication that provides context for the protocol. Authors can demonstrate that the protocol works by citing previous publications that used the protocol, or providing new data obtained using the protocol.

The decision tree shown in Fig 8 may aid scientists in using shortcut citations responsibly, while reporting methods more transparently. Achieving these goals requires a shift in incentives—the scientific community needs to value protocols as a product of scientific work, on par with publications. While depositing methods takes time, it benefits authors in the long term. Depositing protocols in a repository that allows versioning and forking allows researchers to track changes as the protocol evolves and determine what version of a protocol was used for a particular publication. Furthermore, repositories provide long-term access to protocols even when researchers have not used the protocol in years, have moved to another lab or institution, or the person responsible for the protocol has left the research group. Detailed deposited protocols can also make it easier for new team members to learn protocols. Finally, other scientists can use and cite deposited protocols. This may facilitate collaborations with others who are building on one’s methods or help to identify previously unknown factors that affect protocol outcomes.

Fig 8. Decision tree for the responsible use of shortcut citations.

Fig 8

This decision tree helps authors to prepare reproducible methods sections by determining when to use shortcut citations and when to share detailed methods through protocol repositories or methods articles.

While some scientists believe that one should always cite the first paper to use a method, others prefer to cite recent papers that fully describe methods that are similar to their own. These 2 beliefs reflect different reasons for citing a paper. Scientists who cite the original paper are using a “who or what” citation to give the authors of the cited paper credit for developing the method. In contrast, those who cite recent papers that describe current protocols recognize that these citations are more useful as shortcut citations. Fortunately, authors can do both. Authors should structure the citing sentence to help readers to distinguish between the “who or what” citation and the “shortcut” citation. For example, authors might write: “Experiments were performed using an updated version [citation 1] of the protocol originally developed by Smith and colleagues [citation 2]. This phrasing clearly demonstrates that the first citation is the shortcut citation, while the second citation is intended to give credit.

The role of journals in fostering open and reproducible methods

The scientific community has rapidly expanding groups focused on open data and open code [2326], yet comparatively little attention has been dedicated to open methods. Open data are most useful when they are generated through rigorous experimental designs, using reproducible methods. Data are of little value when we do not understand or cannot evaluate the quality of the methods used to generate them. By implementing the actions outlined below, journals can foster a culture of protocol sharing, and open and reproducible methods, within the scientific communities that they serve.

  • Make all methods and protocol publications open access. When publishing paywalled papers, move methods sections in front of the paywall [27]. Methods and protocol papers are extremely likely to be cited as shortcut citations, and readers need to access these resources to determine how a method was performed. Eliminating paywalls for methods and protocol papers would allow everyone to access the information needed to implement the methods described. Moving methods sections for all closed access papers, including historical content, in front of the paywall, would give everyone access methodological details from papers cited as shortcuts. Publishers already place other parts of the paper, such as abstracts and references, in front of the paywall.

  • Replace or augment static methods and protocol papers with dynamic formats where authors publish a living protocol on an open access protocol repository: The static formats used to publish traditional research papers, which are primarily focused on results, do not work well for methods. Protocols evolve over time. The question is not whether protocols will change, but when and how they will change. Furthermore, step-by-step protocols are often more useful that the text descriptions found in methods section. Journals should embrace dynamic formats for publishing methods, such as the PLOS ONE Lab Protocols article [22]. Authors publish a detailed living protocol on an open access protocol repository [21], which can be versioned and forked to track the evolution of methods within and across research teams.

  • Ensure that the protocol repositories used to publish dynamic methods allow versioning and forking: Versioning and forking allow readers to share and track the evolution of methods over time, both within and across labs. Forking benefits protocol creators by giving them credit for their work, while allowing them to see how others are adapting their methods. Without forking, researchers may be reluctant to share adapted versions of protocols created by others, as they do not want to inadvertently claim credit for work that was not their own.

  • Exempt methods sections from word count limits. Ask authors to provide sufficient details to allow others to implement the method. Word count exemption policies (e.g., [28]) eliminate a barrier to sharing detailed methods.

  • Encourage authors to share methods in protocol repositories, not as supplemental files. Protocols deposited in repositories are findable and dynamic, whereas methods deposited in supplemental files are not.

  • Require authors use methodological citation shortcuts responsibly, according to the criteria described in Box 1. Eliminate policies requiring authors to cite previous publications instead of fully describing methods. The research community should rethink the way that shortcut citations are used to ensure that these citations contribute to a culture of open and reproducible methods reporting. Policies that ask authors to use shortcut citations irresponsibly deprive others of details needed to implement the method. The driving force behind some of these policies may be concerns about plagiarism and copyright violations [5]. Laws in some countries specifically exempt detailed descriptions of protocols and methods from copyright restrictions (e.g., United States [29,30], European Union [31]), whereas other countries take a more indirect approach by not including detailed descriptions of methods under items that can be copyrighted (e.g., United Kingdom [32]). Other countries may not offer similar protection. Resolving these legal issues is crucial to allow scientists to fully describe their methods whenever they publish their work.

Limitations

This study has several important limitations. Data examining the reasons for citing a paper in the methods (Fig 2A) should be viewed as a rough approximation. Abstractors were making distinctions about the reasons for a citation that the authors themselves likely did not make when inserting citations. Small changes in the wording or position of the reference could alter the categorization, as could variations in reader expertise. We may have systematically underestimated the number of shortcut citations, as we did not have the resources to assess all citations in supplemental methods files. When abstractors encountered a citation that could potentially fall under multiple categories, they were instructed to choose the most likely category. Abstractors encountered many citations that could have been classified under more than one category. Researchers who use this protocol in the future may wish to allow abstractors to select more than one category for each citation. Probable and possible shortcuts were defined using syntactic definitions. Conceptual definitions were not feasible, as reviewers’ familiarity with the methods described was highly variable. Our results may not apply to journals with lower impact factors or non-English language journals. The shortcut citation chains study was designed to determine what problems one might encounter when following shortcut citations to find detailed methods. Data from this small, nonrepresentative sample (1 paper per quintile of possible plus probable shortcut citations per field) should not be used to determine how often each of these problems occurred. Larger samples would be needed to answer this question.

This study focused on shortcut citations. We did not examine the completeness of methodological reporting when authors described a method without using a shortcut citation. The 3 fields examined are within the life sciences and results may not be generalizable to other domains.

Conclusions

Authors routinely used methodological shortcut citations to explain their research methods. While shortcut citations can be used effectively, they can also make it difficult for readers to find methodological details that would be needed to implement the method. Journals often lack clear policies to encourage open and reproducible methods reporting. We propose that methodological shortcut citations should meet 3 criteria. Cited resources should describe a method similar to the one used by the authors, provide enough detail to allow others to implement the method, and be open access. We outline additional steps that authors can take to use methodological shortcut citations responsibly, and also propose actions that journals can take to foster a culture of open and reproducible methods reporting. Future studies should explore strategies for creating and sharing detailed protocols that others can use. The scientific community should also identify opportunities to incentivize and reward open methods.

Supporting information

S1 Fig. Methodological citations study flow chart.

This flow chart illustrates the journal and article screening process and shows the number of observations excluded and reasons for exclusion at each phase of screening. Data are available at https://osf.io/d2sa3/, in the methodological citations study folder [12].

(TIF)

pbio.3002562.s001.tif (2.9MB, tif)
S2 Fig. Reasons for citing a resource in the methods section of a paper.

The box plots illustrate the number of times that each type of citation was used, per article, for neuroscience (yellow), biology (blue), and psychiatry (red). The horizontal line within each box shows the median, whereas the top and bottom of the box show the 25th and 75th percentiles. Whiskers represent the furthest datapoint that is within 1.5* the interquartile range from the box. Dots above the whiskers show outliers. Data are available at https://osf.io/d2sa3/, in the methodological citations study folder [12].

(TIF)

pbio.3002562.s002.tif (4.2MB, tif)
S3 Fig. Diagrams for individual articles in the shortcut citation chains study.

These diagrams map the process of finding methodological details for each of the 15 papers in the shortcut citation chains study. Reviewers consulted resources cited in shortcut citations to find methodological details. The diagrams show the publication year and type of each cited resource and whether the resource was behind a paywall. Chains of shortcut citations occur when the cited source also uses a methodological shortcut citation to describe the method. Text on the diagram provides information describes problems encountered when searching for details about the cited method.

(PDF)

pbio.3002562.s003.pdf (213.1KB, pdf)
S4 Fig. Flow diagram for journal policy study.

This flow chart illustrates the journal screening process and shows the number of observations excluded and reasons for exclusion at each phase of screening. Data are available at https://osf.io/d2sa3/, in the journal policy study folder [12].

(TIF)

pbio.3002562.s004.tif (2.5MB, tif)
S1 Table. Number of articles examined for each neuroscience journal.

Values are n, or n (% of all articles). Screening was performed to exclude articles that were not full-length original research articles (e.g., reviews, editorials, perspectives, commentaries, letters to the editor, short communications), were not published in March 2020, or did not have a methods section. No issue indicates that the journal did not publish an issue or any articles in March 2020. Data are available at https://osf.io/d2sa3/, in the methodological citations study folder [12]. * Journals were included on both the neuroscience and psychiatry (S3 Table) lists.

(DOCX)

pbio.3002562.s005.docx (25.9KB, docx)
S2 Table. Number of articles examined for each biology journal.

Values are n, or n (% of all articles). Screening was performed to exclude articles that were not full-length original research articles (e.g., reviews, editorials, perspectives, commentaries, letters to the editor, short communications), were not published in March 2020, or did not have a methods section. No issue indicates that the journal did not publish an issue or any articles in March 2020. Data are available at https://osf.io/d2sa3/, in the methodological citations study folder [12].

(DOCX)

pbio.3002562.s006.docx (14.1KB, docx)
S3 Table. Number of articles examined for each psychiatry journal.

Values are n, or n (% of all articles). Screening was performed to exclude articles that were not full-length original research articles (e.g., reviews, editorials, perspectives, commentaries, letters to the editor, short communications), were not published in March 2020, or did not have a methods section. No issue indicates that the journal did not publish an issue or any articles in March 2020. Data are available at https://osf.io/d2sa3/, in the methodological citations study folder [12]. * Journals were included on both the neuroscience and psychiatry (S1 Table) lists.

(DOCX)

pbio.3002562.s007.docx (14.4KB, docx)
S4 Table. Methods repositories used in neuroscience, biology, and psychiatry.

Values are n (% of articles). Data are available at https://osf.io/d2sa3/, in the methodological citations study folder [12].

(DOCX)

pbio.3002562.s008.docx (13.5KB, docx)
S5 Table. Age of the youngest, median, and oldest shortcut citations within a paper for each field.

This table shows summary statistics for Fig 3A. IQR, interquartile range. Data are available at https://osf.io/d2sa3/, in the methodological citations study folder [12].

(DOCX)

pbio.3002562.s009.docx (25.4KB, docx)
S6 Table. Comparing methods of sharing detailed protocols.

JOVE, Journal of Visualized Experiments; N/A, not applicable; OA, open access. * protocols.io offers a partnership with PLOS ONE where authors can publish a methods article linked to their protocol.

(DOCX)

pbio.3002562.s010.docx (26.4KB, docx)

Acknowledgments

The authors thank Małgorzata Anna Gazda and Alberto Antonietti for their assistance in translating policies of Polish and Italian journals.

Data Availability

The abstraction protocols, data and code for the methodological citations study and journal policy study were deposited on the Open Science Framework (RRID:SCR_003238) at https://osf.io/d2sa3/.

Funding Statement

This study was completed through a participant guided, learn-by doing meta-research course funded by the Berlin University Alliance within the Excellence Strategy of the federal and state governments (301_TrainIndik to TLW). VK received salary support from the Berlin University Alliance grant. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Morrison SJ. Time to do something about reproducibility. Elife. 2014;3. Epub 2014/12/11. doi: 10.7554/eLife.03981 ; PubMed Central PMCID: PMC4260475. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Errington TM, Denis A, Perfito N, Iorns E, Nosek BA. Challenges for assessing replicability in preclinical cancer biology. Elife. 2021;10. Epub 2021/12/08. doi: 10.7554/eLife.67995 ; PubMed Central PMCID: PMC8651289. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Carp J. The secret lives of experiments: methods reporting in the fMRI literature. Neuroimage. 2012;63(1):289–300. Epub 2012/07/17. doi: 10.1016/j.neuroimage.2012.07.004 . [DOI] [PubMed] [Google Scholar]
  • 4.Dechartres A, Trinquart L, Atal I, Moher D, Dickersin K, Boutron I, et al. Evolution of poor reporting and inadequate methods over time in 20 920 randomised controlled trials included in Cochrane reviews: research on research study. BMJ. 2017;357:j2490. Epub 2017/06/10. doi: 10.1136/bmj.j2490 . [DOI] [PubMed] [Google Scholar]
  • 5.Zhang Y-H, Jia X-Y, Lin H-F, Tan X-F. Be careful! Avoid duplication: A case study. J Zhejiang Univ Sci B. 2013;14(4):355–358. doi: 10.1631/jzus.B1300078 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Moro-Martin L. How to write the methods section of your research paper. Kolabtree Blog: Kolabtree Blog. 2020. [updated 2020 Sep 5; cited 2022 June 13, 2022]. Available from: https://www.kolabtree.com/blog/how-to-write-the-methods-section-of-your-research-paper/. [Google Scholar]
  • 7.Copy-pasting your methods section is good, actually. Stanford Center for Reproducible Neuroscience: Stanford Center for Reproducible Neuroscience; 2020. [updated 2020 Oct 28; cited 2022 June 13]. Available from: https://reproducibility.stanford.edu/copy-pasting-your-methods-section-is-good-actually/. [Google Scholar]
  • 8.Quintana D. Twitter2022 [cited 2022 Jun 13]. Available from: https://twitter.com/dsquintana/status/1514931134341296131.
  • 9.Halane M. Twitter2017 [cited 2022 Jun 13]. Available from: https://twitter.com/themorgantrail/status/925940615199600641.
  • 10.Brief guide for submission to Nature. Nature. 2022. [cited 2022 Jun 13]. Available from: https://www.nature.com/nature/for-authors/initial-submission. [Google Scholar]
  • 11.Weissgerber TL. Training early career researchers to use meta-research to improve science: A participant-guided “learn by doing” approach. PLoS Biol. 2021;19(2):e3001073. Epub 2021/02/25. doi: 10.1371/journal.pbio.3001073 ; PubMed Central PMCID: PMC7904195. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Standvoss K, Soto N, Kazezian V, Lewke BR, Bastian K, Chidambaram S, et al. Shortcut citations in the methods section: frequency, problems and strategies for responsible reuse. Open Science Framework. 2024. [cited 2024]. Available from: https://osf.io/d2sa3/. doi: 10.17605/OSF.IO/D2SA3 [DOI] [PubMed] [Google Scholar]
  • 13.Moher D, Liberati A, Tetzlaff J, Altman DG. Preferred reporting items for systematic reviews and meta-analyses: the PRISMA statement. J Clin Epidemiol. 2009;62(10):1006–12. Epub 2009/07/28. doi: 10.1016/j.jclinepi.2009.06.005 . [DOI] [PubMed] [Google Scholar]
  • 14.Marcus E, whole Cell team. A STAR Is Born. Cell. 2016;166(5):1059–60. Epub 2016/08/28. doi: 10.1016/j.cell.2016.08.021 . [DOI] [PubMed] [Google Scholar]
  • 15.Jambor H, Antonietti A, Alicea B, Audisio TL, Auer S, Bhardwaj V, et al. Creating clear and informative image-based figures for scientific publications. PLoS Biol. 2021;19(3):e3001161. Epub 2021/04/01. doi: 10.1371/journal.pbio.3001161 ; PubMed Central PMCID: PMC8041175. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Harris CR, Millman KJ, van der Walt SJ, Gommers R, Virtanen P, Cournapeau D, et al. Array programming with NumPy. Nature. 2020;585(7825):357–62. Epub 2020/09/18. doi: 10.1038/s41586-020-2649-2 ; PubMed Central PMCID: PMC7759461. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.The pandas development team. pandas-dev/pandas: Pandas. Zenodo; 2020.
  • 18.Hunter JD. Matplotlib: A 2D graphics environment. Comput Sci Eng. 2007;9(3):90–95. doi: 10.1109/MCSE.2007.55 [DOI] [Google Scholar]
  • 19.Caswell TA, Droettboom M, Lee A, Hunter J, Sales de Andrade E, Firing E, et al. matplotlib/matplotlib: REL: v3.3.3. v3.3.3 ed: Zenodo; 2020. [Google Scholar]
  • 20.Kluyver T, Ragan-Kelley B, Pérez F, Granger B. Jupyter Notebooks—a publishing format for reproducible computational workflows. In: Scmidt FLaB, editor. Positioning and Power in Academic Publishing: Players, Agents and Agendas. the Netherlands: IOS Press; 2016. [Google Scholar]
  • 21.Teytelman L, Stoliartchouk A, Kindler L, Hurwitz BL. Protocols.io: Virtual Communities for Protocol Development and Discussion. PLoS Biol. 2016;14(8):e1002538. Epub 2016/08/23. doi: 10.1371/journal.pbio.1002538 ; PubMed Central PMCID: PMC4993360. Teytelman and Alexei Stoliartchouk are employees of protocols.io and both own equity in the company. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.PLOS announces two new peer-reviewed article types for PLOS ONE [Internet]. STMPublishing News; 2020 [cited 2022 Mar 27]. Available from: https://www.stm-publishing.com/plos-announces-two-new-peer-reviewed-article-types-for-plos-one/.
  • 23.The Carpentries. 2022 [cited 2022 Jul 14]. Available from: https://carpentries.org.
  • 24.FAIRsharing.org 2022 [cited 2022 Jul 14]. Available from: https://fairsharing.org.
  • 25.FAIRsFAIR 2022 [cited 2022 Jul 14]. Available from: https://www.fairsfair.eu.
  • 26.ReproNim 2022 [cited 2022 Jul 14]. Available from: https://www.repronim.org.
  • 27.Leite SB, Brooke M, Carusi A, Collings A, Deceunick P, Dechamp JF, et al. Promoting Reusable and Open Methods and Protocols (PRO-MaP): Draft recommendations to improve methodological clarity in life sciences publications. OSFPreprints. 2023. doi: 10.31219/osf.io/x85gh [DOI] [Google Scholar]
  • 28.Announcement: Reducing our irreproducibility. Nature. 2013;496(398). doi: 10.1038/496398a [DOI] [Google Scholar]
  • 29.Office USC. Circular 31: Ideas, Methods or Systems 2012. [cited 2022 Jun 13]. Available from: https://www.copyright.gov/circs/circ31.pdf. [Google Scholar]
  • 30.U.S. Copyright Office. Circular 1: Copyright basics: Library of Congress; 2021 [updated 2021 Sep; cited 2022 Jul 17]. Available from: https://www.copyright.gov/circs/circ01.pdf, p. 2.
  • 31.European Parlimentary Research Service. Copyright Law in the EU: Salient features of copyright law across the EU Member states: Comparative Law Library Unit, Directorate-General for arlimentary Research Services, European Parliment; 2018 [cited 2022 Jul 17]. Available from: https://www.europarl.europa.eu/RegData/etudes/STUD/2018/625126/EPRS_STU(2018)625126_EN.pdf, p. 52 (III.2.3. Protected works and exceptions to copyright protection), paragraph 6.
  • 32.How copyright protects your work GOV.UK [cited 2022 Jul 17]. Available from: https://www.gov.uk/copyright.

Decision Letter 0

Roland G Roberts

1 Feb 2023

Dear Tracey,

Thank you for submitting your manuscript entitled "Taking shortcuts: Great for travel, but not for reproducible methods sections" for consideration as a Meta-Research Article by PLOS Biology.

Your manuscript has now been evaluated by the PLOS Biology editorial staff, as well as by an academic editor with relevant expertise, and I'm writing to let you know that we would like to send your submission out for external peer review.

However, before we can send your manuscript to reviewers, we need you to complete your submission by providing the metadata that is required for full assessment. To this end, please login to Editorial Manager where you will find the paper in the 'Submissions Needing Revisions' folder on your homepage. Please click 'Revise Submission' from the Action Links and complete all additional questions in the submission questionnaire.

Once your full submission is complete, your paper will undergo a series of checks in preparation for peer review. After your manuscript has passed the checks it will be sent out for review. To provide the metadata for your submission, please Login to Editorial Manager (https://www.editorialmanager.com/pbiology) within two working days, i.e. by Feb 03 2023 11:59PM.

If your manuscript has been previously peer-reviewed at another journal, PLOS Biology is willing to work with those reviews in order to avoid re-starting the process. Submission of the previous reviews is entirely optional and our ability to use them effectively will depend on the willingness of the previous journal to confirm the content of the reports and share the reviewer identities. Please note that we reserve the right to invite additional reviewers if we consider that additional/independent reviewers are needed, although we aim to avoid this as far as possible. In our experience, working with previous reviews does save time.

If you would like us to consider previous reviewer reports, please edit your cover letter to let us know and include the name of the journal where the work was previously considered and the manuscript ID it was given. In addition, please upload a response to the reviews as a 'Prior Peer Review' file type, which should include the reports in full and a point-by-point reply detailing how you have or plan to address the reviewers' concerns.

During the process of completing your manuscript submission, you will be invited to opt-in to posting your pre-review manuscript as a bioRxiv preprint. Visit http://journals.plos.org/plosbiology/s/preprints for full details. If you consent to posting your current manuscript as a preprint, please upload a single Preprint PDF.

Feel free to email us at plosbiology@plos.org if you have any queries relating to your submission.

Kind regards,

Roli

Roland Roberts, PhD

Senior Editor

PLOS Biology

rroberts@plos.org

Decision Letter 1

Roland G Roberts

15 Mar 2023

Dear Tracey,

Thank you for your patience while your manuscript "Taking shortcuts: Great for travel, but not for reproducible methods sections" was peer-reviewed at PLOS Biology. It has now been evaluated by the PLOS Biology editors, an Academic Editor with relevant expertise, and by four independent reviewers.

In light of the reviews, which you will find at the end of this email, we would like to invite you to revise the work to thoroughly address the reviewers' reports.

You 'll see that reviewer #1 is positive, but thinks that the assessment of the scale of the problem is somewhat weak, and probably requires a more quantitative treatment (and larger sample size); he also questions the field categorisation, asks about types of protocol (and field-specific citation culture), and makes a number of helpful textual and presentational requests. Reviewer #2 is very positive, and only has textual requests (plus a suggestion to include an example from Fig S2 in the main text). Reviewers #3 and #4 are also very positive and only have textual requests.

IMPORTANT: Please note that, contrary to a recommendation from reviewer #3, we do not accept structured abstracts. Also, note reviewer #2's recommendation to incorporate some aspects of Fig S2 into the main manuscript. During cross-commenting, one of the other reviewers also strongly supported this recommendation. I looked at this Figure to see what the fuss was about, and I agree that it's fascinating and interestingly presented - I encourage you to include a few "case studies" as part of a new main Figure.

Given the extent of revision needed, we cannot make a decision about publication until we have seen the revised manuscript and your response to the reviewers' comments. Your revised manuscript is likely to be sent for further evaluation by all or a subset of the reviewers.

We expect to receive your revised manuscript within 3 months. Please email us (plosbiology@plos.org) if you have any questions or concerns, or would like to request an extension.

At this stage, your manuscript remains formally under active consideration at our journal; please notify us by email if you do not intend to submit a revision so that we may withdraw it.

**IMPORTANT - SUBMITTING YOUR REVISION**

Your revisions should address the specific points made by each reviewer. Please submit the following files along with your revised manuscript:

1. A 'Response to Reviewers' file - this should detail your responses to the editorial requests, present a point-by-point response to all of the reviewers' comments, and indicate the changes made to the manuscript.

*NOTE: In your point-by-point response to the reviewers, please provide the full context of each review. Do not selectively quote paragraphs or sentences to reply to. The entire set of reviewer comments should be present in full and each specific point should be responded to individually, point by point.

You should also cite any additional relevant literature that has been published since the original submission and mention any additional citations in your response.

2. In addition to a clean copy of the manuscript, please also upload a 'track-changes' version of your manuscript that specifies the edits made. This should be uploaded as a "Revised Article with Changes Highlighted" file type.

*Re-submission Checklist*

When you are ready to resubmit your revised manuscript, please refer to this re-submission checklist: https://plos.io/Biology_Checklist

To submit a revised version of your manuscript, please go to https://www.editorialmanager.com/pbiology/ and log in as an Author. Click the link labelled 'Submissions Needing Revision' where you will find your submission record.

Please make sure to read the following important policies and guidelines while preparing your revision:

*Published Peer Review*

Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out. Please see here for more details:

https://blogs.plos.org/plos/2019/05/plos-journals-now-open-for-published-peer-review/

*PLOS Data Policy*

Please note that as a condition of publication PLOS' data policy (http://journals.plos.org/plosbiology/s/data-availability) requires that you make available all data used to draw the conclusions arrived at in your manuscript. If you have not already done so, you must include any data used in your manuscript either in appropriate repositories, within the body of the manuscript, or as supporting information (N.B. this includes any numerical values that were used to generate graphs, histograms etc.). For an example see here: http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5

*Blot and Gel Data Policy*

We require the original, uncropped and minimally adjusted images supporting all blot and gel results reported in an article's figures or Supporting Information files. We will require these files before a manuscript can be accepted so please prepare them now, if you have not already uploaded them. Please carefully read our guidelines for how to prepare and upload this data: https://journals.plos.org/plosbiology/s/figures#loc-blot-and-gel-reporting-requirements

*Protocols deposition*

To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols

Thank you again for your submission to our journal. We hope that our editorial process has been constructive thus far, and we welcome your feedback at any time. Please don't hesitate to contact us if you have any questions or comments.

Sincerely,

Roli

Roland Roberts, PhD

Senior Editor

PLOS Biology

rroberts@plos.org

------------------------------------

REVIEWERS' COMMENTS:

Reviewer #1:

[identifies himself as Olavo B. Amaral]

Summary:

The manuscript addresses the question of "shortcut citations" in methods description. Although this problem is frequently mentioned in debates about methodological reproducibility, it is understudied and it's nice to see actual research about it.

The results contain three main sections, which study (a) the prevalence of various types of citations in the methods sections of articles in highly cited journals, including shortcut ones, (b) examples of what happens when shortcut citations are followed and (c) a review of journal policies. This is followed by a reasonably extensive discussion focused on (d) guidelines on how to use shortcut citations.

I generally agree that this is an interesting structure, as it (a) documents the phenomenon, (b) evaluates to what degree it represents a problem, (c) inquires what is being made to address it and (d) suggests additional measures. The weakest link in the chain, however, seems to be point (b) (i.e. measuring the impact of the problem), as I am not sure the case studies provided are enough to quantify this. I will try to make this clear in my major points below.

Main concern:

- While the numbers of articles and citations in the first section of the study are probably sufficient to provide an overview of the use of citations, the 15 articles included as case studies in the second section are not. The authors seem to acknowledge this limitation, as they refrain from making a quantitative synthesis of these articles. That said, this leads this section of the manuscript to fall short in accurately presenting the importance of the problem.

Although I found the visualization for each case study provided in Fig. S2 interesting, I would doubt that most readers will really make the effort to go through each one of them, much less be able to synthesize the data in their own heads to reach meaningful conclusions. Thus, I would strongly recommend that the authors provide some kind of quantitative synthesis of the problem in this section (i.e. What percentage of shortcut citations can ultimately be traced to the original reference? What's the average number of steps? What percentage is behind a paywall? What percentage reaches a dead end or an insufficient description?).

I note that 15 articles are probably too few for this purpose, and that the sample of articles in which citations are followed would have to be expanded. Thus, I would recommend that the authors perform a sample size calculation to reach the number of citations/articles that can provide reliable estimates within a given confidence interval. For this purpose, it's worth noting that it would be desirable to perform synthesis both at the level of citations (i.e. what percentage of citations in the sample can be traced?) and at the level of articles (i.e. what percentage of articles in the sample have at least one untraceable citation?), as citations within a single article should not be considered as fully independent units when it comes to representing the whole population of citations. Thus, using articles as units for the purpose of sample size calculation might be the better option.

Other general concerns:

- The categorization of scientific fields is somewhat strange: most people would probably consider neuroscience is a subfield of biology, so presenting both as separate categories may puzzle some readers. I understand that this is a consequence of the JCR categories used, but making this clearer from the start (e.g. "examine the use of shortcut citations in neuroscience, biology and psychiatry journals in the abstract) and perhaps referring to the biology journals as "general biology" would help to avoid confusion.

Still on this point, the selection of fields is narrow and ad hoc. I understand that this is a limitation posed by the authors' own expertise, but it is nevertheless one of the main weaknesses of the manuscript. Thus, the narrow range of scientific fields examined should probably be mentioned in the limitations section.

- Even within this relatively narrow sample of fields, the kinds of methods that deserve a protocol probably varies a lot: I'd guess that psychiatry journals include a lot of surveys and instruments, while biology and neuroscience might have predominantly wet lab protocols. It would be interesting if somewhere in the paper (possibly in the example cases provided) we could get a feeling of what kind of "protocols" we are talking about, even if only in a general sense. If quantifying/classifying them is not feasible, at least some illustrative examples could be provided. Are we talking about methods to quantify proteins? Scales to measure depression? Electrophysiology setups for rodents)? The citation culture probably depends a lot on the particular method, so the whole discussion sounds a bit disembodied without touching on this point somewhere.

- Why are only minimum/maximum numbers of citations within shortcuts and the youngest/oldest citation coded? This looks like an approach to simplify data extraction, but it ends up providing very limited information (i.e. especially if there are many citations per paper, the oldest and youngest ones give very little information on the actual range).

Moreover, this ends up making data visualization in Fig. 3 much less intuitive than it could be (i.e. it would clearer and more informative to provide the full range of citation ages). If the authors could provide the full ranges (although I'm not sure that this is feasible), this would likely strengthen the paper. If not, I'd reconsider whether Fig. 3 should be included in the main results, as I don't think the results as displayed say much about the sample of citations as a whole.

- Some points in the case series description and discussion mention that some references "provided a description that was no longer state-of-the-art" and that this may be a problem. I don't really get the idea here: methods citation are supposed to provide an accurate description of what was done in a study, not of what's the current state of the art of the method. In this sense, descriptions shouldn't age badly or become "not-state-of-the art".

I understand the concern that a very old shortcut citation raises suspicions that it might not really describe what was done in the paper (as it may be likely that no one uses certain methods in exactly the same way after 50 years). But if this is what the authors meant, this should be stated more clearly, as it is not really the impression that comes out of reading these passages.

In the same vein, mentioning in the discussion that "supplemental methods cannot be updated" is technically correct, but is not a limitation in terms of making methods sections reproducible (which seems to be the point of the paper). For the purpose of methods description, whatever was used in a paper should remain static, even though the method may evolve in subsequent study.

- In terms of data sharing, one thing I could not find in the manuscript or in the OSF was the DOI and title of the articles used as case studies in Fig. S2. I may have missed it, but as there was no folder for the case series section I didn't know where to look for it. As this seems important for reproducing the findings, this list should be provided somewhere (possibly as a document within the OSF) and cited within the text and legend to figure S2.

Minor concerns:

Introduction:

- The correct name of the project mentioned in the first paragraph is Reproducibility Project: Cancer Biology (not "for Cancer Biology").

- "This risk of bias for randomization sequence generation and allocation concealment was unclear…" - this sentence seems odd (in particular the "This" at the start), please revise the wording.

Figure 1:

- Isn't the methods section a viable alternative for sharing details needed to reproduce experiments as well? While I agree that in many cases a separate protocol may be a better option, that depends on the length of detail that is needed, which will vary greatly depending on the method. Therefore, I would argue that the methods section should be included as an option in the figure - saying that the information "should" be shared in a separate document sounds overprescriptive.

- The second "readers" can be omitted from the third sentence of the figure legend.

Methods:

- Instead of citing the full OSF page for "protocols, data and code for the prevalence study and journal policy studies" using a single link, wouldn't it make sense to cite a specific DOI for each of these resources? The same thing hold for points in the text in which specific resources are cited (e.g. "The full search strategy is available on the OSF repository" could point to a direct link to the search strategy rather to the full OSF page).

I think this is optional, as the Readme files in the OSF are clear. But providing specific links to each resource would be more consistent with the authors' recommendation of providing pages for book citations, for example (in the sense of sparing the reader the trouble to search for a resource within a larger space).

- What is meant by "top journals" exactly? Are those the ones with the highest impact factor in the JCR in their specific fields? Although this would be my guess, it is not clear from the description.

- The data on whether papers were related to SARS-Cov2 sounded rather gratuitous, as Covid-19 was not mentioned anywhere in the introduction. If this data is to be kept in the paper (I personally don't think it adds much), the rationale for extracting this should be mentioned somewhere.

- Though this eventually became clear, I initially had a hard time to understand what was meant by "number of citations per shortcut". This could be made clearer when this variable is first introduced.

- The description of a probable shortcut states that "additional details are not provided in the following sentences or elsewhere in the methods sections". But what happens if the method is fully explained outside of the methods section (i.e. in the supplementary material or in a repository)? I was unsure how these cases were classified, so it's probably worth commenting explicitly on it.

- Electronic searches were performed using the terms "[journal name]", "journal citation reports ranking", "author guidelines", "journal policy", and "impact factor". I don't quite get what this search means to achieve. Why would one need to search for "impact factor" to look for policies?

Results:

Figure 2:

- The different areas have different mean numbers of methods citations per paper (being somewhat higher in Biology). Thus, showing the results for different categories in percentages as in Fig. 2A may cause misleading impressions - although there are still less "How" citations in Biology than in Neuroscience or Psychiatry when measuring absolute numbers, the actual difference is smaller (while that in "Who or what" citations is even larger). Having the bars represent absolute numbers (possibly still displaying the percentage within the bars) - with overall longer bars for Biology - would likely provide a more accurate impression of what's going on.

- It took me a while to understand the right panel in Fig. 2B. While the fact the two sides of the violin plots represent different data eventually becomes clear, wouldn't it make it easier on the reader to break the information for probable and possible citation into separate plots (especially as the left panel uses symmetric violin plots)?

Tables S5 and S6:

- Can't the information in these tables be included in the legend for Fig.2 and Fig.3 (as it is relatively short and essentially synthesizes the data in the figures)? This is optional, but would leave the information in one place instead of creating a lot of supplements.

Figure 5:

- Are the categories in Fig.5A and 5D mutually exclusive? It would seem to me that a journal could encouragd providing sufficient methodological details both in the author guidelines and as policy, and that they may encourage sharing methods in more than one place (i.e. repository or supplemental files). This is likely worth commenting on in the legend.

Discussion:

- I don't think the Germany and California examples mentioned in Box 1 are needed: there are plenty of places of the world with much worse access, and these particular examples are not particularly representative of difficulties faced by the world at large.

- While I agree with the recommendation to "make all methods publications open access", I don't think that there's any particular reason why methods papers are different from the rest of science (in the sense that they should be open access), so I'm not sure the recommendation really belongs here.

- The discussion about copyright issues described in the list of recommendations is long for an item in a list. Thus, it probably would fit better in the main text or in a box.

Table S7:

- I get the feeling that Table S7 would read better if lines and columns were reversed (i.e. methods as lines, features as columns), but it may be a matter of taste.

- Why are supplemental files and protocol journals deemed static while shortcut citations are not? This does not make much sense to me.

- I'd say supplemental files would generally be expected to have been peer reviewed. I agree that this is likely not always the case, but that probably depends more on the reviewer than on the journal (e.g. I don't know of journals that explicitly exempts supplementary material from the peer review process), so I'd remove "depending on the journal".

- The comment "protocols remain available over time" made for repositories stands for all categories - it makes sense when comparing a protocol repository to a lab notebook, not to the other forms of describing protocols. Thus, I'd probably not include it as an advantage here.

- I'd argue that both shortcut citations and supplemental files are "findable" for whoever's reading the paper (which is likely what matters here), so I'd be inclined to remove this category.

- Clinical journals are not the only one to publish protocols as articles (the systematic review community has a tradition of publishing protocols, for example).

Figure 6:

- In the last no/no option, describing the method in the main text (if it is simple enough to fit) should also be included as an alternative.

Reviewer #2:

[identifies herself as Emma Ganley]

Review of PBIOLOGY-D-23-00204R1, "Taking shortcuts: Great for travel, but not for reproducible methods sections".

------------------

Disclaimer/COI:

------------------

As disclosed to the editors, to be fully transparent I'll include this information here: I currently work at https://protocols.io and therefore have a very obvious positive bias about the value and importance of protocols and methods sharing as a part of the scientific research process. I have also participated in some working groups with Tracey Weissgerber. And less relevant here but in the interests of full disclosure I was previously employed in an editorial capacity at PLOS Biology.

------------------

Overall View and Recommendation:

------------------

This is an important meta-research study into the status of quo of how methods and protocols are cited in the scientific literature. The authors perform a detailed and thorough investigation of how researchers cite methods, the issues encountered by a reader who consults citations to understand methods that were used, and protocols followed, and how journal policies differ with respect to requirements or encouraged behaviour relating to methodological reporting.

This is important because methods sharing has lagged behind expectations around the sharing and availability of other research outputs like data, code, articles themselves etc. And yet insight into methods is critical for interpretation and reuse of data.

With Minor Revisions I would recommend this for publication in PLOS Biology. I don't have many comments and the first one is my subjective take on things, so will leave to editors/authors to decide whether to make changes or not in response.

------------------

Comments:

------------------

1. My one main comment is possibly a bit subjective, and the authors may have considered this already, but for the categories shown in Table 1 (for the reasons for citations in the methods sections), the difference between 'Who or What' and 'From where' isn't very clear. Both really seem to me to be ways to give credit:

- for the 'Who or what' for the wording used is: "give credit to the group that created the method, tool, resource or substance, …"

- and for the "From where" category the wording is that "The citation is used to show where a substance or organism was obtained from", and the example given is a lab as source of a mouse - this seems to be as much of a "Who" as a "From where".

Both categories include "substance" and seem to encompass resources/tools/materials - a mouse model could as easily fall into the Who or What as a Where from, I think these could easily be the same category.

In the results the numbers for the "From where" are quite small, I wonder if these categories might be better off all rolled into one? I appreciate that this would involve a rework of how results are reported and some of the assessment, but worth thinking about whether these are distinct categories (you could still show sub-types within as you already do for the Who or what section in Fig 2).

2. Very minor: Under the description for Case Series on P9, it's noted that an assessment was made to determine if the methodological description was "adequate" - I can infer what is meant by adequate, but it might be helpful to add a definition of what would qualify as adequate.

3. Very minor pedantic comment: The order of Neuroscience / Psychiatry / Biology in the Key for Figure 2 is different from the order that the categories are shown in for each of the figure parts where they're always shown in the order Neuro / Bio / Psych. Suggest authors switch the key order to be the same as the figure panels.

4. Journal policy assessment is interesting to read through, but I wonder if the authors might care to suggest an ideal wording for a journal policy that could then be adopted by journals who want to do better. Even if worded as encouragement rather than mandate this might very valuable and appreciated (of course as with all these things it may also be contentious, so for authors to decide if they want to go there or not…). Another option could have been to identify if there were any exemplary journals out there already, with or without the wording that they use?

5. While I'm well versed in what forking means for a methods-sharing platform like protocols.io, personal experience has made it clear that to a lot of researchers the terminology of forking it not as clear as it could be. The authors should consider including a short definition/description of forking so that there is an easier understanding of the value of being able to easily fork a protocol (maybe using an example e.g. easy creation of a similar protocol where tweaks are needed in order to apply it a different organ or cell type, or to use a new piece of equipment, and how this is facilitated if you can start with a forked copy of an existing protocol so that protocol creation from scratch is not needed).

6. I really like the citation pathway representations in Fig S2 - is it worth putting one example into the main text for a reader to see one without having to go to the Sup file. I think it's very illustrative of the entire article.

Reviewer #3:

[identifies himself as Prof. Lucas Helal, PhD]

Dear authors,

I congratulate you all for this very important piece of research I had the privilege to review. Please enface my comments as a way to improve your manuscript and not as a criticism. Also, I am going with some doubts throughout the text I hope you may have the opportunity to explain that in details.

### Comments ###

1. Title: for my sense, the title should be more technically informative, specially in terms of design. Please use the Murad et al (2017) reporting guideline for meta-research studies to guide your writing.

2. Abstract: it is a matter of style, but for the vast majority of the times, a structured abstract works better for readers (and referees) than a non-structured one. I strongly recommend a structured abstract for your manuscript, although not being an exigence of PLOS Biology journal.

3. Main text:

3.1. Structure of sub-studies:

- I'm up to believe your first study does not fit into a classification of a prevalence study (indeed, there is no prevalence study among study designs, but cross-sectional studies). To gather a true prevalence, your denominator should be all manuscripts of the studied population, which is not true. What you have in hands are relative frequencies. Therefore, for my sense, to claim fhe first study as a prevalence study is not correct at all. I would recommend "bibliometric analysis" OR "cross-sectional analysis". So, it serves also when presenting results - you cannot sustain prevalences in your report, rather than frequencies. It should be switched.

- As for your second study, I won't classify as a case series study as well. This is a clinical classification, although I understood the analogy. A term that is well accepted in the meta-research literature is a "cohort of studies", which I recommend for.

- Finally, your third study also does not fit into your classification for my understanding. What is a "journal policy study"? When you are cross-sectionally analyzing journal policies (which you did, not longitudinally), it is a cross-sectional study, but your unit of analysis are not patients - rather, studies. Then, I strongly suggest to revisit it.

3.2 Stratification of studies

- It wasn't clear for me. Did you take all of your studies, split them into quintiles of their JCR, and randomly selected a fraction of studies of each quintile? If yes, it does make sense for me, but I would love to hear from you if you think that JCR may be a proxy of editorial policies (word limit) and quality of reporting; and, how the potential role of the study design and the nature of intervention were accounted for? I explain: RCTs and pre-clinical studies usually requires more detailed methods explanation than other designs; or, for a given discipline (Psychiatry, for example), a trial published to pursue an FDA/EMA market approval is completely different of a trial in which the intervention is physical activity, for example (potentially). This should be highly acknowledged - if not possible to control, acknowledge as a limitation in your discussion.

3.3 Results

- Personally, I think that is clearer for the readers to be redundant in the main body and figures/tables when talking about results. I felt lack of clarity in results, specially in the explanation of the violin plots. Is that possible for you to use redundancy and write the results presented in figures/tables as whole text as well? Quite an analogous of the methods shortcut. We can't presume that all readers do know to interpret three-factors (or more) plots easily and nowadays to make science communicable to peers is more imperative than ever.

- I liked the way you structured the logical of presenting results. They are suitable in a timeline manner and make sense intrinsically.

3.4 Discussion

- I liked as well how you dedicated plenty of space for the future recommendations. We definitely should bring solutions rather than accuse the problems. This is easy. I congratulate you as a whole about this. My only minor point is that I felt some explanation how you determine "why" shortcut citations were used if a mixed-methods study was not conducted. Is it an inference (not statistical)? Did you gather any proxy? For my sense, this is a very difficult question to answer based only in a manuscript screening.

4. Final Considerations

I appreciated this piece of research. Personally, in times in which the discussion of journal standards for submission and etc. is very pertinent. I advocate for formats like this, although I disclose I don't recognize it as a potential conflict of interest to review this manuscript. My central final question for you is: among your all recommendations, how to deal *pragmatically* with the problem of shortcut methods citations while culture changing doesn't finish your cycle (which takes a lot)? We do know that hybrid journals will be still reluctant to expand room for words; we do know that, still nowadays, researchers are afraid to deposit supplemental material and/or data (by the way, congratulations for the data-sharing practice) or even don't know how to do it if they want either - and the list goes on. Theoretically all recommendations make a lot of sense for me, but the problem seems to be urgent and a pragmatic solution needed. Do you have something in mind?

Reviewer #4:

[identifies himself as Timothy M Errington]

This paper reports an exploratory study looking at shortcut citations, an approach used to cite a previous paper instead of describing the methods in detail. The authors found that citation used in methods sections were common with methodological shortcuts common among the 3 fields investigated. Case studies were also conducted to examine how methodological information was able to be located (or not) for several papers. Finally, the paper examined the journal polices around making methods sections complete and reproducible. Overall, this paper highlights a major challenge in how scientific information is communicated, specifically around the methods used in published research findings. The paper is very detailed and clearly communicated. My suggestions below are recommendations to the authors to consider when revising their paper.

(1) While this paper focuses on the way authors use citations in methods for papers published during March 2020, do the authors have any way to help readers understand how this relates to how methods were reported in the past? That is, is this a new phenomenon or something that's been entrenched in scientific methods sections over the years? This is beyond the scope of the study, so my question is only if there are other papers on this topic that might be considered in the discussion. If not, it might be relevant to also discuss that it is unknown to the authors how these findings compare over time because of a lack of identified literature on the topic.

(2) Did the authors look at whether the short-cut citations were self-citing? That is, how common are short-cut citations to one's prior work?

(3) I believe the reason for why citations in supplemental method were not examined was because of a logistical constraint. The authors discussed their subjective impression that many supplemental methods lacked details or described specific details (e.g., list of primers). However, do the authors believe that shortcut citations also exist in supplemental methods and/or repositories? And if so, what potential impact would this have on how readers should interpret the findings reported in this paper? Is it a 'best case' scenario and that the number and incidence of shortcut citations is likely to be higher than what's reported here?

(4) While the authors did not look at completeness of reporting or other aspects of methods reporting, something that shortcut citations reminds me of is another practice that occurs in methods sections where information is not fully detailed, specifically using the phrase 'following standard protocol' instead of writing out or properly linking to the methods used, especially since what is 'standard' varies by lab. While this might be beyond the scope of this paper, the authors might consider how this practice is similar or not to shortcut citations and how this might impact the recommendations proposed.

(5) The authors might also consider this initiative (https://elifesciences.org/inside-elife/c40ae742/elife-latest-request-detailed-protocols-easily-via-bio-protocol-integration) in addition to the one from PLOS One highlighted in the discussion.

(6) Figure 6 has a typo in the bottom left circle - it should be 'that does not meet' instead of 'the does not meet'

Decision Letter 2

Roland G Roberts

8 Feb 2024

Dear Tracey,

Thank you for your patience while we considered your revised manuscript "Taking shortcuts: Identifying strategies for using methodological shortcut citations responsibly" for publication as a Meta-Research Article at PLOS Biology. This revised version of your manuscript has been evaluated by the PLOS Biology editors and the Academic Editor.

Based on our Academic Editor's assessment of your revision, we are likely to accept this manuscript for publication, provided you satisfactorily address the following data and other policy-related requests.

IMPORTANT - please attend to the following:

a) Please change your title to something more explicit and informative. We suggest "Shortcut citations in the methods section: prevalence, problems and strategies for responsible use," but I'm happy to discuss this via email.

b) I noticed when writing a blurb that some of your major findings (or at least that many potential readers will find to be striking, IMHO) do not appear in the Abstract, namely that (study 1) analysis of >800 papers across three fields shows that >90% of papers contain at least one shortcut citation, and that (study 2) these cause serious problems in understanding the method. I wonder if these could be more explicitly highlighted in the Abstract? (the results of study 3 do seem to be described)

c) Many thanks for providing the underlying data and code in OSF. Please could you cite the location of the data clearly in all relevant main and supplementary Figure legends (Figs 2ABC, 3AB, 7ABCD, S2, I think), e.g. “The data underlying this Figure can be found in https://osf.io/XXXXX”

As you address these items, please take this last chance to review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the cover letter that accompanies your revised manuscript.

We expect to receive your revised manuscript within two weeks.

To submit your revision, please go to https://www.editorialmanager.com/pbiology/ and log in as an Author. Click the link labelled 'Submissions Needing Revision' to find your submission record. Your revised submission must include the following:

- a cover letter that should detail your responses to any editorial requests, if applicable, and whether changes have been made to the reference list

- a Response to Reviewers file that provides a detailed response to the reviewers' comments (if applicable, if not applicable please do not delete your existing 'Response to Reviewers' file.)

- a track-changes file indicating any changes that you have made to the manuscript.

NOTE: If Supporting Information files are included with your article, note that these are not copyedited and will be published as they are submitted. Please ensure that these files are legible and of high quality (at least 300 dpi) in an easily accessible file format. For this reason, please be aware that any references listed in an SI file will not be indexed. For more information, see our Supporting Information guidelines:

https://journals.plos.org/plosbiology/s/supporting-information

*Published Peer Review History*

Please note that you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out. Please see here for more details:

https://blogs.plos.org/plos/2019/05/plos-journals-now-open-for-published-peer-review/

*Press*

Should you, your institution's press office or the journal office choose to press release your paper, please ensure you have opted out of Early Article Posting on the submission form. We ask that you notify us as soon as possible if you or your institution is planning to press release the article.

*Protocols deposition*

To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols

Please do not hesitate to contact me should you have any questions.

Sincerely,

Roli

Roland Roberts, PhD

Senior Editor

rroberts@plos.org

PLOS Biology

------------------------------------------------------------------------

DATA NOT SHOWN?

- Please note that per journal policy, we do not allow the mention of "data not shown", "personal communication", "manuscript in preparation" or other references to data that is not publicly available or contained within this manuscript. Please either remove mention of these data or provide figures presenting the results and the data underlying the figure(s).

------------------------------------------------------------------------

Decision Letter 3

Roland G Roberts

26 Feb 2024

Dear Tracey,

Thank you for the submission of your revised Meta-Research Article "Shortcut citations in the methods section: frequency, problems and strategies for responsible reuse" for publication in PLOS Biology. On behalf of my colleagues and the Academic Editor, Cilene Lino de Oliveira, I'm pleased to say that we can in principle accept your manuscript for publication, provided you address any remaining formatting and reporting issues. These will be detailed in an email you should receive within 2-3 business days from our colleagues in the journal operations team; no action is required from you until then. Please note that we will not be able to formally accept your manuscript and schedule it for publication until you have completed any requested changes.

Please take a minute to log into Editorial Manager at http://www.editorialmanager.com/pbiology/, click the "Update My Information" link at the top of the page, and update your user information to ensure an efficient production process.

PRESS: We frequently collaborate with press offices. If your institution or institutions have a press office, please notify them about your upcoming paper at this point, to enable them to help maximise its impact. If the press office is planning to promote your findings, we would be grateful if they could coordinate with biologypress@plos.org. If you have previously opted in to the early version process, we ask that you notify us immediately of any press plans so that we may opt out on your behalf.

We also ask that you take this opportunity to read our Embargo Policy regarding the discussion, promotion and media coverage of work that is yet to be published by PLOS. As your manuscript is not yet published, it is bound by the conditions of our Embargo Policy. Please be aware that this policy is in place both to ensure that any press coverage of your article is fully substantiated and to provide a direct link between such coverage and the published work. For full details of our Embargo Policy, please visit http://www.plos.org/about/media-inquiries/embargo-policy/.

Thank you again for choosing PLOS Biology for publication and supporting Open Access publishing. We look forward to publishing your study. 

Sincerely, 

Roli

Roland G Roberts, PhD, PhD

Senior Editor

PLOS Biology

rroberts@plos.org

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Fig. Methodological citations study flow chart.

    This flow chart illustrates the journal and article screening process and shows the number of observations excluded and reasons for exclusion at each phase of screening. Data are available at https://osf.io/d2sa3/, in the methodological citations study folder [12].

    (TIF)

    pbio.3002562.s001.tif (2.9MB, tif)
    S2 Fig. Reasons for citing a resource in the methods section of a paper.

    The box plots illustrate the number of times that each type of citation was used, per article, for neuroscience (yellow), biology (blue), and psychiatry (red). The horizontal line within each box shows the median, whereas the top and bottom of the box show the 25th and 75th percentiles. Whiskers represent the furthest datapoint that is within 1.5* the interquartile range from the box. Dots above the whiskers show outliers. Data are available at https://osf.io/d2sa3/, in the methodological citations study folder [12].

    (TIF)

    pbio.3002562.s002.tif (4.2MB, tif)
    S3 Fig. Diagrams for individual articles in the shortcut citation chains study.

    These diagrams map the process of finding methodological details for each of the 15 papers in the shortcut citation chains study. Reviewers consulted resources cited in shortcut citations to find methodological details. The diagrams show the publication year and type of each cited resource and whether the resource was behind a paywall. Chains of shortcut citations occur when the cited source also uses a methodological shortcut citation to describe the method. Text on the diagram provides information describes problems encountered when searching for details about the cited method.

    (PDF)

    pbio.3002562.s003.pdf (213.1KB, pdf)
    S4 Fig. Flow diagram for journal policy study.

    This flow chart illustrates the journal screening process and shows the number of observations excluded and reasons for exclusion at each phase of screening. Data are available at https://osf.io/d2sa3/, in the journal policy study folder [12].

    (TIF)

    pbio.3002562.s004.tif (2.5MB, tif)
    S1 Table. Number of articles examined for each neuroscience journal.

    Values are n, or n (% of all articles). Screening was performed to exclude articles that were not full-length original research articles (e.g., reviews, editorials, perspectives, commentaries, letters to the editor, short communications), were not published in March 2020, or did not have a methods section. No issue indicates that the journal did not publish an issue or any articles in March 2020. Data are available at https://osf.io/d2sa3/, in the methodological citations study folder [12]. * Journals were included on both the neuroscience and psychiatry (S3 Table) lists.

    (DOCX)

    pbio.3002562.s005.docx (25.9KB, docx)
    S2 Table. Number of articles examined for each biology journal.

    Values are n, or n (% of all articles). Screening was performed to exclude articles that were not full-length original research articles (e.g., reviews, editorials, perspectives, commentaries, letters to the editor, short communications), were not published in March 2020, or did not have a methods section. No issue indicates that the journal did not publish an issue or any articles in March 2020. Data are available at https://osf.io/d2sa3/, in the methodological citations study folder [12].

    (DOCX)

    pbio.3002562.s006.docx (14.1KB, docx)
    S3 Table. Number of articles examined for each psychiatry journal.

    Values are n, or n (% of all articles). Screening was performed to exclude articles that were not full-length original research articles (e.g., reviews, editorials, perspectives, commentaries, letters to the editor, short communications), were not published in March 2020, or did not have a methods section. No issue indicates that the journal did not publish an issue or any articles in March 2020. Data are available at https://osf.io/d2sa3/, in the methodological citations study folder [12]. * Journals were included on both the neuroscience and psychiatry (S1 Table) lists.

    (DOCX)

    pbio.3002562.s007.docx (14.4KB, docx)
    S4 Table. Methods repositories used in neuroscience, biology, and psychiatry.

    Values are n (% of articles). Data are available at https://osf.io/d2sa3/, in the methodological citations study folder [12].

    (DOCX)

    pbio.3002562.s008.docx (13.5KB, docx)
    S5 Table. Age of the youngest, median, and oldest shortcut citations within a paper for each field.

    This table shows summary statistics for Fig 3A. IQR, interquartile range. Data are available at https://osf.io/d2sa3/, in the methodological citations study folder [12].

    (DOCX)

    pbio.3002562.s009.docx (25.4KB, docx)
    S6 Table. Comparing methods of sharing detailed protocols.

    JOVE, Journal of Visualized Experiments; N/A, not applicable; OA, open access. * protocols.io offers a partnership with PLOS ONE where authors can publish a methods article linked to their protocol.

    (DOCX)

    pbio.3002562.s010.docx (26.4KB, docx)
    Attachment

    Submitted filename: Shortcut_citations_R2R_20231014.docx

    pbio.3002562.s011.docx (47.8KB, docx)
    Attachment

    Submitted filename: Shortcut_citations_R2_ResponseToEditor_20240222.docx

    pbio.3002562.s012.docx (13.2KB, docx)

    Data Availability Statement

    The abstraction protocols, data and code for the methodological citations study and journal policy study were deposited on the Open Science Framework (RRID:SCR_003238) at https://osf.io/d2sa3/.


    Articles from PLOS Biology are provided here courtesy of PLOS

    RESOURCES